Sylvain Gugger
ee6e71e29c
Fix quality
2023-02-22 03:36:15 -05:00
Younes Belkada
24b930ad1d
[MBart
] Fix cross attention mask check ( #21730 )
...
fix typo
2023-02-22 09:21:25 +01:00
Aaron Gokaslan
5e8c8eb5ba
Apply ruff flake8-comprehensions ( #21694 )
2023-02-22 09:14:54 +01:00
Kashif Rasul
df06fb1f0b
Time series transformer: input projection and Std scaler ( #21020 )
...
* added loc and scale outputs from scalers
* fix typo
* fix tests
* fixed formatting
* initial StdScaler
* move scaling to optional str
* calculate std feature for scalers
* undid change as it does not help
* added StdScaler with weights
* added input projection layer and d_model hyperparam
* use linear proj
* add back layernorm_embedding
* add sin-cos pos embeddings
* updated scalers
* formatting
* fix type
* fixed test
* fix repeated_past_values cal.
* fix when keepdim=false
* fix default_scale
* backward compatibility of scaling config
* update integration test expected output
* fix style
* fix docs
* use the actual num_static_real_features in feature_dim cal
* clarified docs
* Update src/transformers/models/time_series_transformer/modeling_time_series_transformer.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update src/transformers/models/time_series_transformer/modeling_time_series_transformer.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update src/transformers/models/time_series_transformer/modeling_time_series_transformer.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* prediction_length is not optional
* fix for reviewer
* Update src/transformers/models/time_series_transformer/configuration_time_series_transformer.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* get rid of un-needed new lines
* fix doc
* remove unneeded new lines
* fix style
* static_categorical_features and static_real_features are optional
* fix integration test
* Update src/transformers/models/time_series_transformer/modeling_time_series_transformer.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* fixing docs for multivariate setting
* documentation for generate
---------
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2023-02-22 07:50:13 +01:00
mollerup23
bb5a2f2fc3
Adding type hints to call() functions in this file ( #21548 )
...
* Adding type hints to call() functions in this file
* make fixup
* Update src/transformers/models/marian/modeling_tf_marian.py
* Update src/transformers/models/marian/modeling_tf_marian.py
* Update src/transformers/models/marian/modeling_tf_marian.py
* Update src/transformers/models/marian/modeling_tf_marian.py
* Update src/transformers/models/marian/modeling_tf_marian.py
* Update src/transformers/models/marian/modeling_tf_marian.py
* Update src/transformers/models/marian/modeling_tf_marian.py
* Update src/transformers/models/marian/modeling_tf_marian.py
---------
Co-authored-by: Matt <rocketknight1@gmail.com>
Co-authored-by: Matt <Rocketknight1@users.noreply.github.com>
2023-02-21 16:28:33 +00:00
Maria Khalusova
78a53d59cb
Adding task guides to resources ( #21704 )
...
* added resources: links to task guides that support these models
* minor polishing
* conflict resolved
* link fix
* Update docs/source/en/model_doc/vision-encoder-decoder.mdx
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
---------
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2023-02-21 10:35:11 -05:00
Yih-Dar
03aaac3502
Fix TVLT (torch device issue) ( #21710 )
...
* fix tvlt ci
* fix tvlt ci
* fix tvlt ci
---------
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2023-02-21 11:37:49 +01:00
Yih-Dar
4c6346cc3e
Fix get_class_in_module
( #21709 )
...
Fix get_class_in_module
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2023-02-21 09:39:15 +01:00
Yih-Dar
ed6ceb7649
Fix typo in PROCESSOR_MAPPING_NAMES
and add tests ( #21703 )
...
* Add test
* Fix GITProcessor
* Update
---------
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2023-02-21 09:38:26 +01:00
Arthur
4deaa534f5
remove position ids and token type ids from forward args in docstring ( #21701 )
2023-02-21 07:01:36 +01:00
Ishan Jindal
c40e3581c7
Fix axial positional encoding calculations for reformer.mdx ( #21649 )
...
* Update reformer.mdx
Fix axial positional encoding calculations
* Update docs/source/en/model_doc/reformer.mdx
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
---------
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
2023-02-21 06:59:51 +01:00
Jonatan Kłosko
deafc24388
Add WhisperTokenizerFast ( #21222 )
...
* Add WhisperTokenizerFast
* Fixup
* Up
* Up
* Improve tests
* Update src/transformers/models/whisper/tokenization_whisper_fast.py
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* Keep stride in whisper pipelien test
* Remove unknown token special case
* Reduce vocabulary size in tests
* Fix vocab size assertion
* Sync copied changes from WhisperTokenizer
* Skip pipeline tests
* Update assertion
* Remove Whisper tokenizer dependency on sentencepiece
* Format
---------
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
2023-02-21 06:58:54 +01:00
Sylvain Gugger
8b3db33a76
Pass along revision in dynamic code fetch ( #21698 )
2023-02-20 21:21:42 +01:00
Arthur
4194e5f42b
Fix-rag-finetune-project-requirement ( #21697 )
...
pin pytorch lightning requirement
2023-02-20 17:23:39 +01:00
Alara Dirik
49ab16239c
Add EfficientNet ( #21563 )
...
* Add EfficientNet to transformers
2023-02-20 16:37:11 +03:00
Younes Belkada
c9a0671477
[bnb
] fix bnb
decoders bug ( #21688 )
...
* fix `bnb` decoders bug
* make fixup
2023-02-20 12:21:58 +00:00
tanreinama
f56174ac5b
add GPTSAN model (reopen) ( #21291 )
...
* add GPTSAN-Japanese
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN (update for review)
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* fix typo in comment text
* add GPTSAN
* add GPTSAN
* add GPTSAN
* add GPTSAN
* fix document and comments
* fix class name GPTSAN->GPTSan
* fix import and test for tokenizer
2023-02-20 11:25:27 +01:00
Sylvain Gugger
c87bbe1ff0
Fix quality
2023-02-20 03:27:09 -05:00
Morgan McGuire
011cc17a81
Fix for non-contiguous label tensors in VisonEncoderDecoder ( #21582 )
...
* add prints
* add shape
* add reshape
* clean up
2023-02-20 09:23:46 +01:00
Andy Ehrenberg
2840272c5f
add flax whisper implementation ( #20479 )
...
* add flax whisper implementation
* rever change to setup
* remove unused imports
* revert generation changes
* flax whisper docs
* docs
* import order
* import sorting
* isort
* add dummy objects
* doc formatting
* formatting
* remove trailing whitespaces
* fix flax whisper docs
* add generation logic to unlock flax whisper
* remove scans
* give credits to Flax Bart implementation
* remove unused imports
* add license
* remove assert
* more credits to Bart
* fix style
* formatting
* support left padding
* add flax whisper generation test
* remove copied from comments whenever not a full copy
* fix docstrings for logits processors
* revert change to FlaxForceTokensLogitsProcessor
* revert doc changes
* improve generation docs
* reorganize
* formatting
* cleanup docs
* add tests
* handle empty list case
* fix forced decoder ids in flax tests
* add flax whisper to inits
* upate dummy objects
* docs for FlaxAutoModelForSpeechSeq2Seq
* fix decoder_position_ids computation in pretrained model decode/__call__ fns
* add Copied from statements as necessary
* compute position_ids only in __call__ and decode methods of pretrained model subclasses
* improve readabilityof compute positional embeddings
* check dimensionality of input_features instead of hidden_states
* copied from statement for init_cache
* formatting
* fix copies
* fix copies
* pass attention mask to encoder layers
* fix decoder module outputs
* set dtype
Co-authored-by: Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
* smaller flax model for whisper test
* Update src/transformers/generation/flax_utils.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/whisper/modeling_flax_whisper.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update tests/models/whisper/test_modeling_flax_whisper.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* cleanup
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/whisper/modeling_flax_whisper.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* bias cleanup
* doc fix
* align style for force tokens processor
* readability
* fix input shape in tests
* revert FlaxGenerationMixin docstring
* formatting
* fix tests
* fix imports
* consistent encoder hidden states
* consistent hidden states
* input shapes
* typo
* partial class trick
* partial class for input shape
* base_class with correct input shape
* partial base classes
* match by name
* set main_input_name
* compare on names
* formatting
* remove unused import
* safer position ids computation
* safer position id computation
* Update src/transformers/models/whisper/modeling_flax_whisper.py
Co-authored-by: Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
* Update src/transformers/models/whisper/modeling_flax_whisper.py
Co-authored-by: Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
* remove identical inherited tests
* fix prompt ids in tests
* use generation config
* use jnp array
* better var names
* more explicit bias use
* import transformers
* formatting
* test formatting
* remove unused imports
* remove unused imports
* formatting
* isort
* docs
* fix ln orders for encoder hidden states
* whisper unique generation stuff
* flake
* use finfo for attention bias
* docs
* Update src/transformers/generation/flax_utils.py
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* docs
* add timestamp flax test
* jit for timestamps
* formatting
* clean up timestamps processor
* formatting
* remove if_true
* cleanup
---------
Co-authored-by: Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
2023-02-20 09:17:40 +01:00
AlexWertheim
7735e0406f
Enable PyTorch/XLA Fully Sharded Data Parallel (FSDP) ( #21406 )
...
* Reinserted import statement accidentally removed during rebasing.
* Added auto_wrap functionality, restructured XLA FSDP logic to more closely match PyTorch FSDP logic.
* Fixed flag descriptions; changed several instances of fsdp_ to xla_fsdp_; pass in auto_wrap_policy and auto_wrapper_callable directly to avoid lambda saving.
* Moved XLA FSDP logic to be adjacent to Fairscale FSDP logic in trainer.
* Formatted changes in accordance with HF style requirements.
* Added back in warning which was accidentally removed.
* - Merged XLA FSDP training arguments into `fsdp_config`
- Added `xla` boolean flag to `fsdp_config` to specify XLA FSDP wrapping
- Merged XLA FSDP wrapping logic into FSDP wrapping logic within trainer
class
* Cleaned up errors, moved argument to fsdp_config
- Set `xla` and `xla_fsdp_grad_ckpt` flags by default in fsdp_config
- Added missing colons following conditionals
- Moved `fsdp_transformer_layer_cls_to_wrap` to `fsdp_config`
- Modified `fsdp_transformer_layer_cls_to_wrap` to be list of strings,
not just one string
- Changed Fairscale FSDP logic to allow for set of layer classes to wrap
- Removed unnecessary checks for `xla_fsdp`
* Corrected small errors, improved layer class flag
- Correctly set default values for `xla` and `xla_fsdp_grad_ckpt`
arguments
- Made `fsdp_transformer_layer_cls_to_wrap` a list of strings instead of
a single string
- Added processing to ensure that `fsdp_transformer_layer_cls_to_wrap`
works as expected if passed as a single string
- Updated PyTorch FSDP logic to accept a list of layers to wrap, as done
with XLA FSDP
- Replaced instances of `getattr()` with `.get()` for dictionary
retrievals with default values, including when setting
`fsdp_min_num_params`
- Corrected `self.fsdp is not None` to `len(self.fsdp) > 0`
- Removed extraneous `xla_fsdp` argument descriptions from outside
`fsdp_config`
* Changed xla-fsdp-settings to be dictionary
- Modified xla-fsdp-settings to be entered directly as dictionary
instead of loaded through JSON file
- Made small style corrections
* Reverted unintentional local_rank TPU check
* Do not block XLA FSDP if local rank is -1
* Rebased and applied automatic formatting
- Rebased
- Applied automatic formatting changes via `make style`
* Applied automatic formatting with latest version of black
* Replaced expression with
* Reran black examples tests src utils
ruff examples tests src utils --fix
make autogenerate_code
make[1]: Entering directory '/usr/local/google/home/awertheim/HF-FSDP-PR/transformers'
make[1]: Leaving directory '/usr/local/google/home/awertheim/HF-FSDP-PR/transformers' after additional formatting changes
* Additionall automatic formatting changes
* Remove unnecessary whitespace characters from src/transformers/training_args.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
---------
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2023-02-20 09:06:23 +01:00
Yih-Dar
7f1cdf1895
Fix dynamic module import error ( #21646 )
...
* fix dynamic module import error
---------
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2023-02-17 21:22:39 +01:00
Younes Belkada
8a4c319d33
[BLIP
] update blip path on slow tests ( #21476 )
...
* update blip path
* Update tests/models/blip/test_modeling_blip.py
2023-02-17 18:26:36 +00:00
Younes Belkada
087fd5f368
[ImageProcessor
] Refactor default mean
& std
to OPENAI_CLIP_MEAN
& OPENAI_CLIP_STD
( #21425 )
...
* fix default value
* add the fix on other models
2023-02-17 18:57:05 +01:00
Joao Gante
005b515754
Generate: eta sampling numerical stability ( #21676 )
2023-02-17 17:09:37 +00:00
Yoshinari Fujinuma
bb6a664e14
Fix multi-gpu training error for LayoutLMv2 ( #21675 )
...
Co-authored-by: Yoshinari Fujinuma <fujinuy@amazon.com>
2023-02-17 17:04:11 +00:00
Younes Belkada
a8eb4f79f9
[CLAP
] Fix few broken things ( #21670 )
...
* add `is_longer`
* fix docstring
* fix config class
* fix loss
* fix all doctests
* fix order
* fix last failing tests
---------
Co-authored-by: arthur.zucker@gmail.com <arthur.zucker@gmail.com>
2023-02-17 11:32:14 +01:00
Younes Belkada
3668ec1716
[bnb
] Introducing BitsAndBytesConfig
( #21579 )
...
* v1 `BitsandbytesConfig`
- add v1
- add tests
- more user-friendly API
- add docs
* change to `BitsAndBytesConfig`
* replace logic
* changes
* make fixup
* quality
* make fixup
* fix doc
* fix test
* update toctree
* fix slow test
* add tips
* add warning
* change title
* oops
* Update docs/source/en/main_classes/quantization.mdx
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/utils/bitsandbytes.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* remove unused file
* adapt suggestion
- add also tests
- change logic
* update docs
* adapt suggestions
---------
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2023-02-17 09:44:01 +01:00
Steven Anton
f16d29b337
Adapt PerceiverIO Multimodal class to work with arbitrary modalities ( #20054 )
...
* * Properly register parameters in PerceiverMultimodalPreprocessor
* Adapt PerceiverTextPreprocessor to work with PerceiverMultimodalPreprocessor
* Change a few type hints
* Fix formatting; incorrect return type
* Return embeddings_wo_pos
---------
Co-authored-by: Steven Anton <antonstv@amazon.com>
2023-02-16 16:51:00 -05:00
Arthur
c236a62172
[CLAP] Add CLAP to the library ( #21370 )
...
* add model like clip
* update
* text model ok
* clap text works
* some refactor
- `CLAPVision` to `CLAPAudio`
- refactor kwargs of audio modules
* more refactor
* more refactor
* more refactor
* correct fusion
* more refactor
* new modules
* add basic processor
* fixup
* remove whisper copioed from
* audio logits match
* add doc
* correct filters mel and add maxlength
* style
* few fixes
* forward passes
* fixup
* fixup
* some clean up
* remove mels form the dictionnary
* pad after the repeat
* update padding when dsmaller
* fix padding
* style
* use swin patch merging
* use copied from swin
* processor with any tokenizer
* more copied from
* some clean up
* more refactor
* fix mel when rand_trunc
* style
* remove unused imports
* update processing
* remove image processing tests
* add testing fiel
* fixmodeling issues
* replace with `is_longer`
* clap in serialization
* more refactor
* `make fixup`
* make fixup
* fix feature extractor
* update test feature extractor
* `make fixup`
* clean up config
* more clean up
* more cleanup
* update tests
* refactor tests and inits
* removeCLAP vision config
* remove CLAP from image procssing auto and dummy vision objects
* update inits
* style
* re order classes in modeling clap
* Use roberta tokenizer as the other weights are not open sourced
* small cleaup
* remove tokenization CLAP
* processor tokenizr is roberta
* update feature extraction doc
* remove vclap from model zero shot
* update f_min and f_max to frequency_xx
* some changes
- fix modeling keys
- add `is_longer` in the forward pass
- make fixup
* make fixup
* consistent behavior ebtween rand_crop and fusion
* add numpy resize and bilinear and documentation
* move resizing to image utils
* clean feature extraction
* import resize from correct file
* resize in image transforms
* update
* style
* style
* nit
* remove unused arguments form the feature extractor
* style
* few fixes + make fixup
* oops
* fix more tests
* add zero shot audio classification pipeline
* update zeroshot classification pipeline
* fixup
* fix copies
* all CI tests pass
* make fixup + fix docs
* fix docs
* fix docs
* update tests pip;eline
* update zero shot pipeline
* update feature extraction clap
* update tokenization auto
* use nested simplify
* update pipeline tests
* Apply suggestions from code review
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* split in two lines
* fixes
* refactor
* clean up
* add integration tests
* update config docstring
* style
* update processor
* fix processor test
* fix feat extractor tests
* update docs
* Apply suggestions from code review
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* fix readmes
* fix tips
* Update src/transformers/models/auto/configuration_auto.py
* update doc and remove todo -> properly explained
* fix idx and typo
* typoe
* cleanup config
* cleanup tests, styles and doc
* ignore docstyle on image transform
* add conversion script
* remove the `clap` indx in favor of `CLAP`
* update __init
* nits
* Update src/transformers/pipelines/__init__.py
* fix bug
* clarifiy config
* fix copy
* fix init
* Apply suggestions from code review
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* fix model output
* fix comment
* make fixup
* make fixup
* rename to `Clap`
* replace to `Clap`
* replace to `Clap`
* repo consistency
* again repo-consistency
* make fixup
* Apply suggestions from code review
Co-authored-by: Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
* add config
* changes
* update conversion
* Apply suggestions from code review
Co-authored-by: Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
* remove unused function
* update based on code reviews
* style
* more comments
* cleanup
* clean up
* style
* apply suggestions
* Empty commit
* pipeline will be added in a different PR
* update calls to audio utils functions
* update pipeline init
* style
* style
* styling again
* use pad
* fix repo-consistency
* update utils and add doc for audio utils
* clean up resize by using torch. update inits accordingly
* style
* CLap's tokenizer is RobertA
* add audio utils to internal toctreee
* update totctree
* style
* update documentation and normalize naming accross audio utils and feature extraction clap
* style
* clean up
* update doc and typos
* fix doctest
* update modelin code, got rid of a lot of reshaping
* style on added doc audio utils
* update modeling clap
* style
* Apply suggestions from code review
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* docstringvariables with CLAP
* rename key
* update modeling CLAP
* update audio utils docstring
* update processing clap
* fix readmes
* fix toctree
* udpate configuration clap
* fix init
* make fixup
* fix
* fix
* update naming
* update
* update checkpoint path
* Apply suggestions from code review
* Major refactoring
* Update src/transformers/models/clap/configuration_clap.py
* merge
---------
Co-authored-by: younesbelkada <younesbelkada@gmail.com>
Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Co-authored-by: Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
2023-02-16 20:59:27 +01:00
Sylvain Gugger
6b0257de42
Sort deps alphabetically
2023-02-16 13:27:27 -05:00
Alissa
b0f0086fa4
Add OPT resources to the transformers documentation ( #21625 )
...
* Add resources to OPT
* Add additional resources for OPT
* Remove -{" "} after <PipelineTag pipeline="question-answering" />
* Change bitsnbytes to bitsandbytes
* Revert formatting
* Revert automatic format changes
* Remove - sign after <PipelineTag pipeline="question-answering" />
2023-02-16 12:44:28 -05:00
Stas Bekman
61d7fec87a
[bloom] gradient_checkpointing fix ( #21655 )
...
Update modeling_bloom.py
2023-02-16 08:57:19 -08:00
Connor Henderson
0f96c26de6
refactor: Make direct_transformers_import util ( #21652 )
...
* refactor: Make direct_import util
* edit direct import fn
* add docstring
* make import function specific to transformers only
* edit doc string
2023-02-16 11:32:32 -05:00
Jonatas Grosman
96d4fa46ed
[WhisperModel] fix bug in reshaping labels ( #21653 )
...
fix bug in reshaping labels
2023-02-16 16:00:46 +01:00
dependabot[bot]
fcfd4ec789
Bump werkzeug from 2.0.3 to 2.2.3 in /examples/research_projects/decision_transformer ( #21658 )
...
Bump werkzeug in /examples/research_projects/decision_transformer
Bumps [werkzeug](https://github.com/pallets/werkzeug ) from 2.0.3 to 2.2.3.
- [Release notes](https://github.com/pallets/werkzeug/releases )
- [Changelog](https://github.com/pallets/werkzeug/blob/main/CHANGES.rst )
- [Commits](https://github.com/pallets/werkzeug/compare/2.0.3...2.2.3 )
---
updated-dependencies:
- dependency-name: werkzeug
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2023-02-16 09:23:43 -05:00
Xiaoyang Chen
212c42a1e3
Update document of WhisperDecoderLayer ( #21621 )
...
* Update document of WhisperDecoderLayer
* Update modeling_mbart.py
* Update doc with utils/check_copies.py --fix_and_overwrite
* Update modeling_xlm_prophetnet.py
2023-02-16 09:19:59 -05:00
Jannis Vamvas
61abe3290b
[WIP] Move X-MOD models to facebook organization ( #21640 )
...
Move X-MOD models to facebook org
2023-02-16 09:18:25 -05:00
regisss
751f17aa48
Fix typos in contrastive-image-text example README ( #21665 )
2023-02-16 09:10:25 -05:00
Sylvain Gugger
9d1116e995
Update deprecated load_module ( #21651 )
2023-02-15 15:57:24 -05:00
Joao Gante
1567bef3b3
Generate: PT Dynamo without graph breaks in the main greedy/sample loop ( #21648 )
2023-02-15 20:16:46 +00:00
Steven Liu
7a5533b2c3
Refactor model summary ( #21408 )
...
* first draft of model summary
* restructure docs
* finish first draft
* ✨ minor reviews and edits
* apply feedbacks
* save important info, create new page for attention
* add attention doc to toctree
* ✨ few more minor fixes
2023-02-15 10:35:14 -08:00
Zineng Tang
a0e69a9375
Add TVLT ( #20725 )
...
* Update image_processing_tvlt.py
* Update modeling_tvlt.py
* Update
* Update modeling_tvlt.py
* Create tvlt.mdx
* Update configuration_tvlt.py
* Update modeling_tvlt.py
* Update test_modeling_tvlt.py
* Update modeling_tvlt.py
* Update modeling_tvlt.py
* Update image_processing_tvlt.py
* Update feature_extraction_tvlt.py
* Update tvlt models
* Update tests
* Update
* Update
* Update tests
* Update README_ko.md
* Update README_ja.md
* Update README_ko.md
* Update README_zh-hans.md
* Update docs/source/en/model_doc/tvlt.mdx
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update docs/source/en/model_doc/tvlt.mdx
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update src/transformers/models/tvlt/configuration_tvlt.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update src/transformers/models/tvlt/configuration_tvlt.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update src/transformers/models/tvlt/configuration_tvlt.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update src/transformers/models/tvlt/configuration_tvlt.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update tvlt.mdx
* Update modeling_tvlt.py
* Update configuration_tvlt.py
* Update modeling_tvlt.py
* Update modeling_tvlt.py
* Update modeling_tvlt.py
* Update modeling_tvlt.py
* Add files via upload
* Update model
* Update modeling_tvlt.py
* Update tvlt models
* Update src/transformers/models/tvlt/__init__.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/__init__.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/image_processing_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/image_processing_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Add files via upload
* Add files via upload
* Delete modeling_tvlt.py
* Delete feature_extraction_tvlt.py
* Delete configuration_tvlt.py
* Delete image_processing_tvlt.py
* Delete processing_tvlt.py
* Update tvlt
* Update src/transformers/models/tvlt/configuration_tvlt.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/tvlt/image_processing_tvlt.py
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* Update src/transformers/models/tvlt/configuration_tvlt.py
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* Update tests/models/tvlt/test_modeling_tvlt.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update tests/models/tvlt/test_modeling_tvlt.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/tvlt/modeling_tvlt.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/tvlt/modeling_tvlt.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update README.md
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update README_es.md
* Update README_hd.md
* Update README_ja.md
* Update README_ko.md
* Update README_zh-hans.md
* Update README_zh-hant.md
* Update index.mdx
* Update tvlt.mdx
* Update tvlt.mdx
* Update configuration_tvlt.py
* Update src/transformers/models/tvlt/image_processing_tvlt.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/tvlt/image_processing_tvlt.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/tvlt/image_processing_tvlt.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/tvlt/image_processing_tvlt.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update src/transformers/models/tvlt/modeling_tvlt.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
* Update modeling_tvlt.py
* Add files via upload
* Update tvlt.mdx
* Update modeling_auto.py
* Add files via upload
* Add files via upload
* Update dummy_pt_objects.py
* Update __init__.py
* Update feature_extraction_tvlt.py
* Update feature_extraction_tvlt.py
* Update image_processing_tvlt.py
* Update modeling_auto.py
* Update test_feature_extraction_tvlt.py
* Update test_processor_tvlt.py
* Update test_feature_extraction_tvlt.py
* Add files via upload
* Update test_image_processor_tvlt.py
* Update tests/models/tvlt/test_processor_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/modeling_tvlt.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update src/transformers/models/tvlt/processing_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/modeling_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/modeling_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update tests/models/tvlt/test_image_processor_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/modeling_tvlt.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update tests/models/tvlt/test_image_processor_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/modeling_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update tests/models/tvlt/test_image_processor_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update tests/models/tvlt/test_image_processor_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update tests/models/tvlt/test_modeling_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/modeling_tvlt.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update src/transformers/models/tvlt/modeling_tvlt.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update src/transformers/models/tvlt/configuration_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update tests/models/tvlt/test_feature_extraction_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/processing_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update docs/source/en/model_doc/tvlt.mdx
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/configuration_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/configuration_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/configuration_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/configuration_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/configuration_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/configuration_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/configuration_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/configuration_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/configuration_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/configuration_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/feature_extraction_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/configuration_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/feature_extraction_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/feature_extraction_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/feature_extraction_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update feature_extraction_tvlt.py
* Update feature_extraction_tvlt.py
* Update src/transformers/models/tvlt/image_processing_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/image_processing_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update image_processing_tvlt.py
* Update src/transformers/models/tvlt/modeling_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update modeling_tvlt.py
* Update modeling_tvlt.py
* Update modeling_tvlt.py
* Update test_image_processor_tvlt.py
* Update tests/models/tvlt/test_modeling_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update tests/models/tvlt/test_modeling_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update tests/models/tvlt/test_modeling_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update tests/models/tvlt/test_modeling_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update tests/models/tvlt/test_modeling_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update tests/models/tvlt/test_modeling_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update tests/models/tvlt/test_modeling_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update tests/models/tvlt/test_modeling_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update tests/models/tvlt/test_modeling_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Add files via upload
* Add files via upload
* Update modeling_tvlt.py
* Update modeling_tvlt.py
* Update modeling_tvlt.py
* Add files via upload
* Update docs/source/en/model_doc/tvlt.mdx
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update image_processing_tvlt.py
* Add files via upload
* Add files via upload
* Update tvlt.mdx
* Update docs/source/en/model_doc/tvlt.mdx
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update docs/source/en/model_doc/tvlt.mdx
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update src/transformers/models/tvlt/image_processing_tvlt.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update src/transformers/models/tvlt/image_processing_tvlt.py
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
* Update docs/source/en/model_doc/tvlt.mdx
Co-authored-by: Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
* Update docs/source/en/model_doc/tvlt.mdx
Co-authored-by: Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
* Update src/transformers/models/tvlt/configuration_tvlt.py
Co-authored-by: Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
* Add files via upload
* Add files via upload
* Add files via upload
* Add files via upload
* Update modeling_auto.py
* Update tvlt.mdx
* Update dummy_pt_objects.py
* Update feature_extraction_tvlt.py
* Update modeling_tvlt.py
* Update test_feature_extraction_tvlt.py
* Update test_image_processor_tvlt.py
* Update test_feature_extraction_tvlt.py
* Update modeling_tvlt.py
* Update dummy_pt_objects.py
* Update dummy_speech_objects.py
* Add files via upload
* Update README_hd.md
* Update modeling_tvlt.py
* Update modeling_tvlt.py
* Update modeling_tvlt.py
* Update modeling_tvlt.py
* Update modeling_tvlt.py
* Update modeling_tvlt.py
* Update test_modeling_tvlt.py
* Update src/transformers/models/tvlt/configuration_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/feature_extraction_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/image_processing_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/image_processing_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/modeling_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/image_processing_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/modeling_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update MAE processing
* Update modeling_tvlt.py
* Update modeling_tvlt.py
* Update modeling
* Update style
* Update src/transformers/models/tvlt/modeling_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/models/tvlt/modeling_tvlt.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update check_repo.py
* Update tvlt.mdx
* Update __init__.py
* Update tests
* Update tvlt models
* Update configuration_tvlt.py
* Update configuration_tvlt.py
* Update image_processing_tvlt.py
* Update dummy_pt_objects.py
* Add files via upload
* Update test_modeling_tvlt.py
* Update test_feature_extraction_tvlt.py
* Update test_feature_extraction_tvlt.py
* Update test_feature_extraction_tvlt.py
* Update test_feature_extraction_tvlt.py
* Update test_feature_extraction_tvlt.py
* Update test_feature_extraction_tvlt.py
---------
Co-authored-by: NielsRogge <48327001+NielsRogge@users.noreply.github.com>
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
Co-authored-by: Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
2023-02-15 18:10:30 +00:00
Bruno Alvisio
7bac51837b
Pass parent exception as context exception to provide clearer stack trace ( #21636 )
...
* Pass parent exception as context exception to provide clearer stack trace
* Update src/transformers/tokenization_utils_base.py
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
---------
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2023-02-15 11:34:02 -05:00
amyeroberts
3499c49c17
Skipping more high mem tests - Wav2Vec2 Hubert ( #21647 )
...
Skipping more tests
2023-02-15 16:00:50 +00:00
Susnato Dhar
0c9c8472e6
Add Ernie-M Model to huggingface ( #21349 )
...
* config and tokenization(fast too) changed and ErnieEncoder added
* Slow Tokenization Added
* Tokenizer(slow) is now working and Fast Tokenizer removed
* Added Config code
* Added Base Model and utils
* ErnieMModel is now working
* All added except tests
* All tests passed except ErnieUIEM
* All tests passed
* all fixes done
* all fixes done
* fixed MAP
* fixed check_code_quality
* fixed Build PR Documentation issue
* Added changes(comments) and also updated to the latest upstream/main
* Added fixup
* Added # Copied comments
* Added fixup
* Added more comments and some nits
* Added fixup
* Fixed README_hd.md
* Added more fixes
* ErnieMTokenizer (being sentencepiece) protected and other docs edited
* Added code_quality fix
* Fixed for
* Added more fix
* modified AZ
* ernie-m tokenization test added!
* attention mask part fixed(with 0->self.config.pad_token_id)
* applied make fixup
2023-02-15 09:24:56 -05:00
Bruno Alvisio
40ca13367e
Fix passing kwargs to TFBertTokenizer ( #21619 )
2023-02-15 09:18:48 -05:00
amyeroberts
fc28c006a6
Skip wav2vec2 hubert high mem tests ( #21643 )
...
* Skip high memory tests
* Skip high memory tests
* Remove unused import
2023-02-15 14:17:26 +00:00
Yih-Dar
e3d832ff87
Fix Blip-2 CI again ( #21637 )
...
* fix blip-2 ci
* fix blip-2 ci
---------
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2023-02-15 10:59:42 +01:00
Matthew McDermott
762dda44de
Remove extra "max_length
is reached." from InfNaNLogitsProcessor documentation ( #21634 )
...
* Fix typo in documentation.
* Remove trailing words typo in documentation.
2023-02-14 16:12:22 -05:00