APAVOU Clément
af69360bf9
Add OPTForQuestionAnswering
( #19402 )
...
* Add `OPTForQuestionAnswering`
- added `OPTForQuestionAnswering` class based on `BloomForQuestionAnswering`
- added `OPTForQuestionAnswering` in common tests
- all common tests pass
- make fixup done
* added docstrings for OPTForQuestionAnswering
* Fix docstrings for OPTForQuestionAnswering
2022-10-10 09:30:59 -04:00
Aritra Roy Gosthipaty
ba71bf4cae
fix: renamed variable name ( #18850 )
...
The sequence_masked variable is actually the part of the sequence that is kept unmasked for the encoder. This commit renames the variable.
2022-10-10 09:26:36 -04:00
Ryan Chan
4824741c4c
Remove dependency of Roberta in Blenderbot ( #19411 )
...
* Remove dependency of Roberta in Blenderbot
* Move Copied from statements to each method of the Roberta classes
* Remove copied from line for mask_token.setter
* update output from example in docs
2022-10-10 09:25:22 -04:00
Mohit Sharma
3080bb4754
Add onnx support for VisionEncoderDecoder ( #19254 )
...
* Add onnx support for VisionEncoderDecoder
* Add onnx support for VisionEncoderDecoder
* Removed unused import
* Rename encoder hidden state
Co-authored-by: lewtun <lewis.c.tunstall@gmail.com>
* Update docstrings and removed redundant code
* Added test function for enc-dec models
* Update doc string text
Co-authored-by: lewtun <lewis.c.tunstall@gmail.com>
* fixed code style
Co-authored-by: lewtun <lewis.c.tunstall@gmail.com>
2022-10-10 09:20:19 -04:00
Lysandre Debut
298f6a98c2
Stop relying on huggingface_hub's private methods ( #19392 )
...
* Leverage hfh for move cache
* Style
2022-10-10 15:19:33 +02:00
wei zhao
7d5ce6802e
Fix typo in image-classification/README.md ( #19424 )
...
Fix link typo of the following content.
PyTorch version, Trainer
PyTorch version, no Trainer
2022-10-10 09:16:58 -04:00
Rak Alexey
c523a86929
fix marianMT convertion to onnx ( #19287 )
...
* fix marianMT convertion to onnx
* Update src/transformers/onnx/convert.py
Co-authored-by: lewtun <lewis.c.tunstall@gmail.com>
* Update src/transformers/onnx/convert.py
Co-authored-by: lewtun <lewis.c.tunstall@gmail.com>
Co-authored-by: lewtun <lewis.c.tunstall@gmail.com>
2022-10-10 09:11:29 -04:00
Darío Hereñú
3410705730
Fixed duplicated line (paragraph #83 ) Documentation: @sgugger ( #19436 )
...
* Fixed duplicated line (paragraph #83 ) @omarespejel @sgugger
* Datasets map denomination fixed (paragraph 42)
2022-10-10 09:08:34 -04:00
Darío Hereñú
83dc49b69b
Backtick fixed (paragraph 68) ( #19440 )
2022-10-10 08:47:14 -04:00
Druhin Abrol
1241a4993b
remove RobertaConfig inheritance from MarkupLMConfig ( #19404 )
...
* remove RobertaConfig inheritance from MarkupLMConfig
* Update src/transformers/models/markuplm/configuration_markuplm.py
fixed typo in docstring
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2022-10-10 08:44:59 -04:00
Matt
4107445a0f
Fix repo names for ESM tests ( #19451 )
2022-10-10 13:20:00 +01:00
Yih-Dar
cbb8a37929
Skip BloomEmbeddingTest.test_embeddings
for PyTorch < 1.10 ( #19261 )
...
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2022-10-10 10:05:30 +02:00
Yih-Dar
8b6bba54a7
Fix ViTMSNForImageClassification
doctest ( #19275 )
...
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2022-10-10 09:51:30 +02:00
Sylvain Gugger
d92e22d1f2
Remove ref to is_pipeline_test
2022-10-07 21:38:07 -04:00
Sylvain Gugger
9ac586b3c8
Rework pipeline tests ( #19366 )
...
* Rework pipeline tests
* Try to fix Flax tests
* Try to put it before
* Use a new decorator instead
* Remove ignore marker since it doesn't work
* Filter pipeline tests
* Woopsie
* Use the fitlered list
* Clean up and fake modif
* Remove init
* Revert fake modif
2022-10-07 18:01:58 -04:00
Alara Dirik
983451a13e
Improve and fix ImageSegmentationPipeline ( #19367 )
...
- Fixes the image segmentation pipeline test failures caused by changes to the postprocessing methods of supported models
- Updates the ImageSegmentationPipeline tests
- Improves docs, adds 'task' argument to optionally perform semantic, instance or panoptic segmentation
2022-10-07 23:34:41 +03:00
Vishwas
de4d71ea07
Removed Bert dependency from BertGeneration code base. ( #19370 )
...
* Copied all the code required from transformers.models.bert.modeling_bert to here
* Fixed styling issues
* Reformatted copied names with Model specific name.
* Reverted BertEncoder part as there is already a class called BertGenerationEncoder
* Added prefixes in missing places.
Co-authored-by: vishwaspai <vishwas.pai@emplay.net>
2022-10-07 13:45:24 -04:00
mustapha ajeghrir
34e0cc6d86
Make Camembert
TF version independent from Roberta
( #19364 )
...
* camembert tf version independent
* fixup
* fixup, all working
* remove comments
* Adding copied from roberta
Co-authored-by: Mustapha AJEGHRIR <mustapha.ajeghrir@kleegroup.com>
2022-10-07 13:42:24 -04:00
Blip blop
7418a48e34
Removed Bert
interdependency in tokenization_electra.py
( #19356 )
...
* Copied from BertTokenizer() in tokenization_bert
* Added BasicTokenizer and WordPieceTokenizer Class
* Update src/transformers/models/electra/tokenization_electra.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Added copied from comments for basicTokenizer and WordPieceTokenizer
* Updated the comments for the tokenizerClasses
* Update src/transformers/models/electra/tokenization_electra.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/electra/tokenization_electra.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Formatted tokenization_electra with `make style`
* Fix repo inconsistencies
* Update src/transformers/models/electra/tokenization_electra.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Set the logger
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2022-10-07 12:24:04 -04:00
Infrared1029
6ef16f2b67
Remove Dependency between Bart and LED (slow/fast) ( #19408 )
...
* removed dependency from bart(slow)
* removed dependency from bart(slow)
* adding copying comments (copied from bart to led)
* updated led docstring
* updated led docstring
* removed dependency from Bart (fast)
* replaced bart with LED in docstrings
* complying flake8
* added more copy comments
* fixing copying comments
* added comments back
* fix copy comments
* fixing copied from comments
* fixing copied from comments
2022-10-07 12:19:50 -04:00
Patrick von Platen
06514b3e1a
Clip device map ( #19409 )
...
* add first generation tutorial
* uP
* [Clip] Add text model to device map
2022-10-07 18:19:15 +02:00
harry7337
c2b83d540e
Removed Bert and XML Dependency from Herbert ( #19410 )
...
Co-authored-by: harry7337 <hari.8jan@gmail.com>
2022-10-07 11:49:09 -04:00
Ryan Chan
e6fc2016ad
Remove dependency of Bert from Squeezebert tokenizer ( #19403 )
...
* Remove dependency of Bert from Squeezebert tokenizer
* run style corrections
* update copies from BertTokenizers
* Update changes and style to Squeezebert files
* update copies for bert-fast
2022-10-07 11:32:55 -04:00
Arthur
994b7a4eea
update attention mask handling ( #19385 )
...
* update feature extractor params
* update attention mask handling
2022-10-07 16:54:08 +02:00
Dean Wyatte
a26d71d6ae
Export TensorFlow models to ONNX with dynamic input shapes ( #19255 )
...
* validate onnx models with a different input geometry than saved with
* only test working features for now
* simpler test skipping
* rm TODO
* expose batch_size/seq_length on vit
* skip certain name, feature, framework parameterizations known to fail validation
* Trigger CI
* Trigger CI
2022-10-07 10:53:03 -04:00
David Yang
5fef17f490
Copy BertTokenizer dependency into retribert tokenizer ( #19371 )
2022-10-07 10:14:00 -04:00
ddobokki
fa4bcd5274
edit: cast attention_mask to long in DataCollatorCTCWithPadding ( #19369 )
...
* edit: casting attention_mask to long in DataCollatorCTCWithPadding
* edit: casting attention_mask to long in DataCollatorCTCWithPadding
2022-10-07 10:05:48 -04:00
Amrit Sahu
e9a49babee
[WIP] Add ZeroShotObjectDetectionPipeline ( #18445 ) ( #18930 )
...
* Add ZeroShotObjectDetectionPipeline (#18445 )
* Add AutoModelForZeroShotObjectDetection task
This commit also adds the following
- Add explicit _processor method for ZeroShotObjectDetectionPipeline.
This is necessary as pipelines don't auto infer processors yet and
`OwlVitProcessor` wraps tokenizer and feature_extractor together, to
process multiple images at once
- Add auto tests and other tests for ZeroShotObjectDetectionPipeline
* Add AutoModelForZeroShotObjectDetection task
This commit also adds the following
- Add explicit _processor method for ZeroShotObjectDetectionPipeline.
This is necessary as pipelines don't auto infer processors yet and
`OwlVitProcessor` wraps tokenizer and feature_extractor together, to
process multiple images at once
- Add auto tests and other tests for ZeroShotObjectDetectionPipeline
* Add batching for ZeroShotObjectDetectionPipeline
* Fix doc-string ZeroShotObjectDetectionPipeline
* Fix output format: ZeroShotObjectDetectionPipeline
2022-10-07 10:00:19 -04:00
Omar Sanseviero
331ea019d7
Remove unneded words from audio-related feature extractors ( #19405 )
2022-10-07 15:52:52 +02:00
Sourab Mangrulkar
56af8df359
HF <-> megatron checkpoint reshaping and conversion for GPT ( #19317 )
...
* HF <-> megatron checkpoint conversion handling reshaping from different tensor and parallel sizes
* Apply suggestions from code review
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* addressing comments
* add doc strings and 🐛 fixes
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2022-10-07 19:16:55 +05:30
Thomas
41ec5d0ced
Added type hints for TF: TransfoXL ( #19380 )
...
* Added type hints for TF: TransfoXL
* Added type hints for TF: TransfoXL
* Change type hints for training
* Change type hints for training
2022-10-07 14:44:58 +01:00
h
b29ebdf4d8
removes prophet config dependencies from xlm-prophet ( #19400 )
2022-10-07 09:26:23 -04:00
Bibhabasu Mohapatra
e162cebfa3
add ONNX support for swin transformer ( #19390 )
...
* swin transformer onnx support
* Updated image dimensions as dynamic
Co-authored-by: lewtun <lewis.c.tunstall@gmail.com>
Co-authored-by: lewtun <lewis.c.tunstall@gmail.com>
2022-10-07 09:23:24 -04:00
IMvision12
969534af4b
Added Type hints for XLM TF ( #19333 )
...
* Update modeling_tf_xlm.py
* Updates
* Update src/transformers/models/xlm/modeling_tf_xlm.py
* Update src/transformers/models/xlm/modeling_tf_xlm.py
* Update src/transformers/models/xlm/modeling_tf_xlm.py
* Update src/transformers/models/xlm/modeling_tf_xlm.py
* Update src/transformers/models/xlm/modeling_tf_xlm.py
Co-authored-by: Matt <Rocketknight1@users.noreply.github.com>
2022-10-07 13:44:50 +01:00
Zachary Mueller
46fd04b481
Fix gather for metrics ( #19389 )
2022-10-07 08:36:05 -04:00
IMvision12
7e348aac96
Making ConvBert Tokenizer
independent from bert Tokenizer
( #19347 )
...
* ConvBert
* added comment
* Updated
* Final_updates
* Update tokenization_convbert.py
* Update tokenization_convbert_fast.py
* Update tokenization_convbert.py
* Update tokenization_convbert.py
* Update tokenization_convbert_fast.py
* Update tokenization_convbert.py
* Update tokenization_convbert_fast.py
* Updates
* Updates
* Updated
* Final Updates
2022-10-07 07:59:02 -04:00
Alara Dirik
ae3e3bc60a
fix docs example, add object_detection to DETR docs ( #19377 )
2022-10-07 00:02:26 +02:00
Ilaygoldman
ce2620194b
Change link of repojacking vulnerable link ( #19393 )
...
The link to https://github.com/vasudevgupta7/bigbird is vulnerable to repojacking (it redirects to the orignial project that changed name), you should change the link to the current name of the project. if you won't change the link, an attacker can open the linked repository and attacks users that trust your links
2022-10-06 23:06:39 +02:00
Alara Dirik
f0b490151e
🚨 🚨 🚨 Fix ViT parameter initialization ( #19341 )
...
This PR aims to rectify the discrepancy between the training performances of HF and Timm ViT implementations.
- Initializes torch and flax ViT dense layer weights with trunc_normal instead of normal (consistent with the TF implementation.
- Initializes cls_token and positional_embeddings with trunc_normal
- Updates DeiT copy to reflect the changes
2022-10-06 12:04:01 +03:00
Sylvain Gugger
7e7f62bfa7
Fix pipeline tests for Roberta-like tokenizers ( #19365 )
...
* Fix pipeline tests for Roberta-like tokenizers
* Fix fix
2022-10-05 17:48:14 -04:00
Alara Dirik
bad353cebf
Fix DETR segmentation postprocessing output ( #19363 )
...
Ensures post_process_instance_segmentation and post_process_panoptic_segmentation methods return a tensor of shape (target_height, target_width) filled with -1 values if no segment with score > threshold is found.
2022-10-06 00:16:36 +03:00
Arthur
45e14038f2
Add WhisperModel to transformers ( #19166 )
...
* simplify loop
* add featur extractor
* add model
* start conversion
* add dropout
* initial commit of test files
* copnversion for all models
* update processor for correct padding
* update feature extraction
* update integration test logits match
* fmnt: off for the logits
* on the fly mel bank
* small nit
* update test
* update tokenizer
* nit feature extraction
* update
* update tokenizer test
* adds logit processor and update tokenizer to get supress tokens
* style
* clean convert
* revert to original modeling tf utils
* Update
* update
* nit
* clean convert file
* update tests and nits
* quality
* slow generation test
* ffn_dim to allow customization
* update readme
* add to toctreee
* start fixing integration tests
* update tests and code
* fix feature extractor
* fix config tests common
* update code to fix tests
* fix feature exctractor
* nit feature extraction
* update test for new feature extractor
* style
* add absrtact
* large logits wioth custom decoder input ids
* wraap around is otrch available
* fix feature extractor
* correct logits for whisper small.en
* nit
* fix encoder_attentino_mask
* some fixes
* remove unnecessary inputs
* nits
* add normalizer file
* update etst tokenization
* fix attention mask not defined
* Add model to README
* Fix doc tests
* fix generate
* remove uncoder attention mask useless
* update test modeling whisper
* update condfig to add second non supress tokens
* nits on feature exrtactor
* nit for test tokenizers
* update etsts
* update tests
* update tokenization test
* fixup
* invalidated hf token. Clean convert openai to whisper
* fix logit tests
* fixup
* clean merge
* revert toc_tree changes
* remove useless LogitProcessor
* Update whisper .mdx
* update config file doc
* update configuration docstring
* update test tokenization
* update test tokenization
* update tokenization whisper
Added copied from where needed
* update feature extraction
* nit test name
* style
* quality
* remove get suppress tokens and update non_speech tokens global variables
* Update src/transformers/models/whisper/feature_extraction_whisper.py
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
* clean modeling whisper and test
Removed the attention mask arguments that are deprecated
* fix large test
* Add multilingual audio test, and translate test
* style
* fix larg multilingual test
* nits
* Update docs/source/en/model_doc/whisper.mdx
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
* add copied from for attention layer
* remove attention masks in doc
* add english normalizer
* update tokenization test
* remove copied from in whisper attention : no bias in k_proj only
* wrap around dependencies in english normalizer
* style
* correct import generation logits
* for now, wrap feature extractor with torch
* Update src/transformers/models/whisper/convert_openai_whisper_to_tfms.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update src/transformers/models/whisper/configuration_whisper.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update docs/source/en/model_doc/whisper.mdx
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* remove torch depencies for feature extraction and style
* fixup
* nit
* update logitds
* style
* nit
* nits and fix final tests
* add `is_more_itertools_available` to utils
* quality
* add begin supress tokens, supress tokens to generate args and config
* clean supressTokensLogitProcessor in generation logits
* Nit naming
* add supressTokensAtBegin
* udpate tests, supress tokens to None or correct values
* nit and style
* update RAG to fit test and generate_logit
* add copy pasted statment on english normalizer
* add arguments to config_common_kwargs
* Update src/transformers/generation_utils.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update src/transformers/generation_logits_process.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update src/transformers/models/whisper/configuration_whisper.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Apply suggestions from code review
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* revert changes based on reviews
* update doc and nits
* more nits
* last nits
* update test configuration common
* add BART name in decoder attention mask documentation
* Update src/transformers/models/whisper/modeling_whisper.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* style
* nit
* nit
* add english.json file to git
* nits on documentation
* nit
* nits
* last styling
* add main toctree file
* remove sentence piece dependency
* clean init file
* fix tokenizer that has no dependencies on sentencepiece
* update whisper init file, nit
* remove english.json file
* add get decoder prompt id
* revert changes and add forced logit processor
* nit
* clean normalizer
* remove protected
* update
* Update src/transformers/models/whisper/configuration_whisper.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* update based on review
* Update src/transformers/models/whisper/configuration_whisper.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* add batched tests
Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
Co-authored-by: NielsRogge <niels.rogge1@gmail.com>
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2022-10-05 22:28:31 +02:00
Alara Dirik
7598791c09
Fix MaskFormer failing postprocess tests ( #19354 )
...
Ensures post_process_instance_segmentation and post_process_panoptic_segmentation methods return a tensor of shape (target_height, target_width) filled with -1 values if no segment with score > threshold is found.
2022-10-05 23:25:58 +03:00
Zachary Mueller
ad98642a82
Fix gather for metrics ( #19360 )
2022-10-05 14:52:01 -04:00
Harsha
d9101b71bc
Removes Roberta and Bert config dependencies from Longformer ( #19343 )
...
* removes roberta and bert config dependencies from longformer
* adds copied from statements
* fixes style
* removes excessive comments and replace bert with longformer in a couple places
* fixes style
2022-10-05 13:50:15 -04:00
Paula Isabel
226b8ef063
correct typos in README ( #19304 )
2022-10-05 10:40:38 -07:00
Matt
071df6eb13
Call _set_save_spec() when creating TF models ( #19321 )
...
* Add a build_from_serving_sig_and_dummies method and replace all calls like model(model.dummy_inputs) with it.
* make fixup
* Remove the overridden save() as this is no longer necessary
* Also call _set_save_spec(), the last missing piece
* Ensure we set the save spec when loading from config too
* Turn this whole thing into a one-line PR
* Turn this whole thing into a one-line PR
* Turn this whole thing into a one-line PR
Co-authored-by: Your Name <you@example.com>
2022-10-05 18:03:49 +01:00
Sylvain Gugger
c875a96eb1
Test failing test while we resolve the issue. ( #19355 )
2022-10-05 12:23:48 -04:00
Younes Belkada
4cbc797b27
Change BloomConfig
docstring ( #19336 )
...
* change `BloomConfig` docstring
- slightly change the docstring of the `BloomConfig`
- Use correct default vocab size
- Use correct default `hidden_dim`, `n_head`
* Update src/transformers/models/bloom/configuration_bloom.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/bloom/configuration_bloom.py
Co-authored-by: SaulLu <55560583+SaulLu@users.noreply.github.com>
* make style
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Co-authored-by: SaulLu <55560583+SaulLu@users.noreply.github.com>
2022-10-05 18:12:13 +02:00
Harsha
e794ca5b16
Frees LongformerTokenizer of the Roberta dependency ( #19346 )
...
* copies over roberta tokenizer to longformertokenizer since they are both identical
* adds Copied from patterns to pass copy check
2022-10-05 11:49:14 -04:00