dependabot[bot]
ea50b64bea
Bump pillow from 10.2.0 to 10.3.0 in /examples/research_projects/decision_transformer ( #31319 )
...
Bump pillow in /examples/research_projects/decision_transformer
Bumps [pillow](https://github.com/python-pillow/Pillow ) from 10.2.0 to 10.3.0.
- [Release notes](https://github.com/python-pillow/Pillow/releases )
- [Changelog](https://github.com/python-pillow/Pillow/blob/main/CHANGES.rst )
- [Commits](https://github.com/python-pillow/Pillow/compare/10.2.0...10.3.0 )
---
updated-dependencies:
- dependency-name: pillow
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-06-07 18:09:02 +01:00
Matt
065729a692
Remove ConversationalPipeline and Conversation object ( #31165 )
...
* Remove ConversationalPipeline and Conversation object, as they have been deprecated for some time and are due for removal
* Update not-doctested.txt
* Fix JA and ZH docs
* Fix JA and ZH docs some more
* Fix JA and ZH docs some more
2024-06-07 17:50:18 +01:00
dependabot[bot]
3a10058201
Bump transformers from 3.5.1 to 4.38.0 in /examples/research_projects/bert-loses-patience ( #31291 )
...
Bump transformers in /examples/research_projects/bert-loses-patience
Bumps [transformers](https://github.com/huggingface/transformers ) from 3.5.1 to 4.38.0.
- [Release notes](https://github.com/huggingface/transformers/releases )
- [Commits](https://github.com/huggingface/transformers/compare/v3.5.1...v4.38.0 )
---
updated-dependencies:
- dependency-name: transformers
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-06-07 16:45:54 +01:00
dependabot[bot]
e3f03789a9
Bump aiohttp from 3.9.0 to 3.9.4 in /examples/research_projects/decision_transformer ( #31317 )
...
Bump aiohttp in /examples/research_projects/decision_transformer
Bumps [aiohttp](https://github.com/aio-libs/aiohttp ) from 3.9.0 to 3.9.4.
- [Release notes](https://github.com/aio-libs/aiohttp/releases )
- [Changelog](https://github.com/aio-libs/aiohttp/blob/master/CHANGES.rst )
- [Commits](https://github.com/aio-libs/aiohttp/compare/v3.9.0...v3.9.4 )
---
updated-dependencies:
- dependency-name: aiohttp
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-06-07 16:43:57 +01:00
dependabot[bot]
48d35b2178
Bump tornado from 6.3.3 to 6.4.1 in /examples/research_projects/visual_bert ( #31298 )
...
Bump tornado in /examples/research_projects/visual_bert
Bumps [tornado](https://github.com/tornadoweb/tornado ) from 6.3.3 to 6.4.1.
- [Changelog](https://github.com/tornadoweb/tornado/blob/master/docs/releases.rst )
- [Commits](https://github.com/tornadoweb/tornado/compare/v6.3.3...v6.4.1 )
---
updated-dependencies:
- dependency-name: tornado
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-06-07 15:44:38 +01:00
조준래
60861fe1fd
Implement JSON dump conversion for torch_dtype in TrainingArguments ( #31224 )
...
* Implement JSON dump conversion for torch_dtype in TrainingArguments
* Add unit test for converting torch_dtype in TrainingArguments to JSON
* move unit test for converting torch_dtype into TrainerIntegrationTest class
* reformating using ruff
* convert dict_torch_dtype_to_str to private method _dict_torch_dtype_to_str
---------
Co-authored-by: jun.4 <jun.4@kakaobrain.com>
2024-06-07 15:43:34 +01:00
Benjamin Badger
ff689f57aa
Extend save_pretrained to offloaded models ( #27412 )
...
* added hidden subset
* debugged hidden subset contrastive search
* added contrastive search compression
* debugged compressed contrastive search
* memory reduction for contrastive search
* debugged mem red
* added low memory option feature
* debugged mem optmimization output stack
* debugged mem optmimization output stack
* debugged low mem
* added low mem cache
* fixed 2047 tensor view
* debugged 2042 past key val inputs
* reformatted tensors
* changed low mem output
* final clean
* removed subset hidden csearch
* fixed hidden device
* fixed hidden device
* changed compressor dtype
* removed hstate compression
* integrated csearch in generate
* test csearch integration into generation
exit()
* fixed csearch kwarg integration with generation
* final wrap and added doc
* Update src/transformers/generation/utils.py
Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com>
* Update src/transformers/generation/utils.py
Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com>
* Update src/transformers/generation/utils.py
Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com>
* added debug print
* direct hstate cat
* direct hstate cat
* direct hstate cat debug
* direct hstate cat debug
* expanded full hidden state stack
* expanded full hidden state stack
* matched dims for hstates
* matched dims for hstates
* logits fix
* equality test
* equality hidden debug
* debug
* added prints for debug
* added prints for debug
* equality check
* switched squeeze dim
* input format debug
* tracing top_k_ids
* removed trace
* added test context
* added jitter
* added jitter
* added jitter
* returned state
* rebuilt past key value reconstruction
* debugged
* cleaned traces
* added selection for pkv
* changed output to dict
* cleaned
* cleaned
* cleaned up contrastive search test
* moved low_memory kwarg
* debugged
* changed low mem test batch size to 1
* removed output
* debugged test input shape
* reformatted csearch test
* added trace
* removed unsqueeze on final forward pass
* replaced unsqueeze with view
* removed traces
* cleaned
* debugged model kwargs
* removed special models from test
* ran make quality
* Update src/transformers/generation/configuration_utils.py
Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com>
* Update src/transformers/generation/configuration_utils.py
Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com>
* refactored
* refactored
* refactored
* make fixup
* renamed flag sequential
* renamed flag sequential
* iterative onloading
* black style and test utils
* added traces for integrated test
* debugged
* added traces
* make style
* removed traces, make style
* included suggestions and added test
* debugged test
* added offload module check and make style
* is_accelerate_available and make style
* added test decorator
* changed test model and config spec
* added offload condition
* added lazy loading for each shard
* debugged
* modified sharding
* debugged
* added traces
* removed safe serialization
* no index overload;
* trace on safe save ptrs
* added ptr condition
* debugged
* debugged ptr
* moved module map init
* remake shard only for offloaded modules
* refactored
* debugged
* refactored
* debugged
* cleaned and make style
* cleaned and make style
* added trace
* sparse module map
* debugged
* removed module map conditional
* refactored
* debug
* debugged
* added traces
* added shard mem trace
* added shard mem trace
* removed underlying storage check
* refactored
* memory leak removal and make style
* cleaned
* swapped test decs and make style
* added mem checks and make style
* added free mem warning
* implemented some suggestions
* moved onloading to accelerate
* refactored for accelerate integration
* cleaned test
* make style
* debugged offload map name
* cleaned and make style
* replaced meta device check for sharding
* cleaned and make style
* implemented some suggestions
* more suggestions
* update warning
Co-authored-by: Marc Sun <57196510+SunMarc@users.noreply.github.com>
* more suggestions
* make style
* new make style
* Update src/transformers/modeling_utils.py
Co-authored-by: Marc Sun <57196510+SunMarc@users.noreply.github.com>
* Update src/transformers/modeling_utils.py
Co-authored-by: Marc Sun <57196510+SunMarc@users.noreply.github.com>
* Update src/transformers/modeling_utils.py
Co-authored-by: Marc Sun <57196510+SunMarc@users.noreply.github.com>
* Update src/transformers/modeling_utils.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
---------
Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com>
Co-authored-by: Marc Sun <57196510+SunMarc@users.noreply.github.com>
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
2024-06-07 07:50:35 -04:00
Cyril Vallez
8bcf9c8dd4
Fix jetmoe model ( #31279 )
...
* Fix jetmoe model
* Remove skip-tests
2024-06-07 11:51:41 +02:00
Danial Kurtumerov
f868cf731a
Fixed Wav2Vec2ProcessorWithLM decoding error ( #31188 )
...
* fix: wav2vec2_with_lm decoding error
Fixed an error where some language models could
not be loaded due to a decoding error, since it
was impossible to select the 'unigram_encoding'
value.
* fix: unexpected keyword argument
Fixed unexpected keyword argument caused by
passing kwargs directly to BeamSearchDecoderCTC.
* style: wav2vec2_with_lm
Changed single quotes to double quotes.
2024-06-07 11:50:07 +02:00
amyeroberts
bdf36dcd48
Enable HF pretrained backbones ( #31145 )
...
* Enable load HF or tim backbone checkpoints
* Fix up
* Fix test - pass in proper out_indices
* Update docs
* Fix tvp tests
* Fix doc examples
* Fix doc examples
* Try to resolve DPT backbone param init
* Don't conditionally set to None
* Add condition based on whether backbone is defined
* Address review comments
2024-06-06 22:02:38 +01:00
Jack Yang
a3d351c00f
Update text-to-speech.md ( #31269 )
...
SpeechBrain usage has changed
2024-06-06 21:59:22 +01:00
Alex Gorodnitskiy
3b4d3d09fd
Fix SwinLayer / DonutSwinLayer / ClapAudioLayer attention mask device ( #31295 )
...
Fix DonutSwinLayer attention mask device
2024-06-06 21:52:14 +01:00
dependabot[bot]
b6c9f47fd6
Bump transformers from 3.5.1 to 4.38.0 in /examples/research_projects/bertabs ( #31290 )
...
Bump transformers in /examples/research_projects/bertabs
Bumps [transformers](https://github.com/huggingface/transformers ) from 3.5.1 to 4.38.0.
- [Release notes](https://github.com/huggingface/transformers/releases )
- [Commits](https://github.com/huggingface/transformers/compare/v3.5.1...v4.38.0 )
---
updated-dependencies:
- dependency-name: transformers
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-06-06 16:13:18 +01:00
Vu Huy Nguyen
f9296249a3
Pipeline VQA: Add support for list of images and questions as pipeline input ( #31217 )
...
* Add list check for image and question
* Handle passing two lists and update docstring
* Add tests
* Add support for dataset
* Add test for dataset as input
* fixup
* fix unprotected import
* fix unprotected import
* fix import again
* fix param type
2024-06-06 14:50:45 +01:00
dependabot[bot]
4c82102523
Bump transformers from 4.19.0 to 4.38.0 in /examples/research_projects/codeparrot ( #31285 )
...
Bump transformers in /examples/research_projects/codeparrot
Bumps [transformers](https://github.com/huggingface/transformers ) from 4.19.0 to 4.38.0.
- [Release notes](https://github.com/huggingface/transformers/releases )
- [Commits](https://github.com/huggingface/transformers/compare/v4.19.0...v4.38.0 )
---
updated-dependencies:
- dependency-name: transformers
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-06-06 14:49:31 +01:00
amyeroberts
c53fcd8381
Mark MobileNetV1ModelTest::test_batching_equivalence as flaky ( #31258 )
...
* Mark MobileNetV1ModelTest::test_batching_equivalence as flaky
* Add link to issue
* woops
2024-06-06 14:47:58 +01:00
Omar Salman
681183974a
Enable dynamic resolution input for Beit ( #31053 )
...
* Initial attempt
* Updates: PR suggestions
* Interpolate the relative position bias when interpolate_pos_encoding is True
* Add slow tag for the added tests
* Add in DATA2VEC_VISION_INPUTS_DOCSTRING
2024-06-06 14:47:41 +01:00
Marc Sun
99895ae5e2
fix accelerate tests for roberta xl ( #31288 )
...
* fix accelerate tests for roberta xl
* style
2024-06-06 14:44:35 +01:00
Baole Ai
5ba8ac54f5
Fix _save_tpu: use _maybe_convert_to_cpu instead of to cpu. ( #31264 )
...
* Fix _save_tpu: use _maybe_convert_to_cpu instead of to cpu.
* fix lint
2024-06-06 09:42:55 -04:00
dependabot[bot]
14ff5dd962
Bump transformers from 3.5.1 to 4.38.0 in /examples/research_projects/bertology ( #31256 )
...
Bump transformers in /examples/research_projects/bertology
Bumps [transformers](https://github.com/huggingface/transformers ) from 3.5.1 to 4.38.0.
- [Release notes](https://github.com/huggingface/transformers/releases )
- [Commits](https://github.com/huggingface/transformers/compare/v3.5.1...v4.38.0 )
---
updated-dependencies:
- dependency-name: transformers
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-06-06 12:42:40 +01:00
Huazhong Ji
9e9679c022
fix: str
should be used not int
when setting env variables ( #31272 )
2024-06-06 12:41:31 +01:00
Lucain
9ef93fccad
Switch from cached_download
to hf_hub_download
in remaining occurrences ( #31284 )
...
Switch from hf_hub_url to hf_hub_download in remaining occurences
2024-06-06 12:05:59 +01:00
Raushan Turganbay
5fabd1e83b
Generation: fix handling of special tokens ( #31254 )
...
* fix special tokens in generatioon
* fix test
* add warning
* fix the check
* warn once
* fix
2024-06-06 15:21:32 +05:00
Raushan Turganbay
7729b77478
Make mamba use cache ( #31116 )
...
* make mamba use cache
* uss cache naming as in mamba
* fix musicgen
2024-06-06 13:37:29 +05:00
Zhiyuan Chen
f5c0fa9f6f
fix loading special_tokens_map_file ( #31012 )
2024-06-06 09:15:27 +02:00
Ranggi Hwang
9b85e405ab
[SwitchTransformer
] Significant performance improvement on MoE blocks ( #31173 )
...
* SwitchTransformer MoE layer performance improvement
* make fixup
* comments about shapes
* make fixup
2024-06-06 09:10:12 +02:00
graham
8177aa0e1a
no need for explicit EXTRA_TOKENS in processing_paligemma.py ( #31022 )
...
no need for explicit EXTRA_TOKENS
2024-06-06 08:41:41 +02:00
amyeroberts
940fde8daf
Skip failing JetMOE generation tests ( #31266 )
...
Skip failing tests for now
2024-06-05 19:06:46 +01:00
Cyril Vallez
bd5091df8d
Reduce by 2 the memory requirement in generate()
🔥 🔥 🔥 ( #30536 )
...
* Fix contrastive_search for new cache structure, and improve performance by removing inneficient torch.stack(torch.split(x, top_k, dim=0))
* Fix _contrastive_search for non-standard cache using ellipsis slicing
* Fix all outputs.logits memory leaks for all decoding strategies!
* Fix small error in _contrastive_search()
* Make all necessary change and revert for the new class
* Apply coding style
* Remove pipes in type hints for compatibility
* correct type hint
* apply style
* Use DynamicCache by default and solve conflicts
* Fix rebase issues
* Add `_supports_dynamic_cache_class` in models for models that support DynamicCache but not other caches to make DynamicCache the default for more models
* Create generation config to return legacy format by default, or to choose not to
* style
* Fix case when use_cache is False
* Remove default DynamicCache in assiste_decoding if assistant_model does not support it + fix _seen_tokens when cropping cache
* Update prepare_inputs_for_generation() for case with empty DynamicCache
* Correct return of args in _assisted_decoding
* Remove EfficientDynamicCache as it is no longer needed
* Correct mistake in generation config
* Move cache logic of assisted decoding to AssistedCandidateGenerator.__init__
* change DynamicCache function names from "split" to "batch_split" for readability + apply coding style
* Remove `_supports_dynamic_cache_class` attribute after rebase
* Correct missing line lost in conflict resolution during rebasing
* Add special case for Jamba
* Fix jamba test
* Coding style
* coding style
* Correct missing import in rebasing
* Simplify _validate_model_kwargs based on removal of _supports_dynamic_cache attribute
* Simplify code paths in _contrastive_search
* coding style
* Update docstrings of cache methods
* Update prepare_inputs_for_generation() -> past_key_values are always Cache objects
2024-06-05 17:05:01 +02:00
Yih-Dar
d6276f0fc5
Add condition to benchmark
job in push-important-models.yml
( #31259 )
...
fix
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2024-06-05 15:19:16 +02:00
Dhaivat Bhatt
b72752f068
Fix circular reference issue in CLIPTokenizerFast ( #31075 )
2024-06-05 14:01:13 +02:00
bastrob
464d986b6c
Add missing Flaubert tokenizer tests ( #30492 )
...
* add flaubert tokenization test, enrich inheritance in FlaubertTokenizer.
* fix quality code ci
* ensure parameter consistency
* fix ci
* fix copyright year and flatten vocab list.
* fix style
2024-06-05 13:52:16 +02:00
Huazhong Ji
41cf4097f7
enable deterministic mode for npu ( #31253 )
2024-06-05 07:35:35 -04:00
Vaibhav Srivastav
4a6024921f
doc: add info about wav2vec2 bert in older wav2vec2 models. ( #31120 )
...
* doc: add info about wav2vec2 bert in older wav2vec2 models.
* apply suggestions from review.
* forward contrib credits from review
---------
Co-authored-by: Sanchit Gandhi <sanchit-gandhi@users.noreply.github.com>
2024-06-05 11:56:11 +01:00
dependabot[bot]
c39aaea972
Bump transformers from 3.5.1 to 4.38.0 in /examples/research_projects/deebert ( #31244 )
...
Bump transformers in /examples/research_projects/deebert
Bumps [transformers](https://github.com/huggingface/transformers ) from 3.5.1 to 4.38.0.
- [Release notes](https://github.com/huggingface/transformers/releases )
- [Commits](https://github.com/huggingface/transformers/compare/v3.5.1...v4.38.0 )
---
updated-dependencies:
- dependency-name: transformers
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-06-05 11:12:58 +01:00
amyeroberts
54659048a2
Early labels validation ( #31240 )
...
* Move label validation checks - fail early
* Remove some formatting changes - add back labels change wav2vec2
2024-06-05 10:50:55 +01:00
Yih-Dar
03ea160937
Benchmark GitHub Actions workflow ( #31163 )
...
* benchmark workflow
* benchmark workflow
* benchmark workflow
* benchmark workflow
* build
* build
* build
* build
* build
* build
* build
* build
* build
* build
* build
* build
* build
* build
---------
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2024-06-05 10:39:00 +02:00
James Braza
63fb253df0
Fixing name 'torch' is not defined
in bitsandbytes
integration ( #31243 )
...
Fixed torch definition error
2024-06-05 08:00:30 +02:00
Yury Sulsky
66875ac070
Specify dtype=torch.bool to avoid xla error ( #31191 )
...
The StoppingCriteriaList allocates is_done without specifying dtype=torch.bool. On XLA this allocates a float tensor and causes a failure on the following line:
is_done = is_done | criteria(input_ids, scores, **kwargs)
by attempting to OR float with bool.
2024-06-05 07:50:54 +02:00
dependabot[bot]
8685b3c5d2
Bump transformers from 4.26.0 to 4.38.0 in /examples/research_projects/vqgan-clip ( #31242 )
...
Bump transformers in /examples/research_projects/vqgan-clip
Bumps [transformers](https://github.com/huggingface/transformers ) from 4.26.0 to 4.38.0.
- [Release notes](https://github.com/huggingface/transformers/releases )
- [Commits](https://github.com/huggingface/transformers/compare/v4.26.0...v4.38.0 )
---
updated-dependencies:
- dependency-name: transformers
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-06-04 22:11:45 +01:00
Yih-Dar
3714f3f86b
Upload (daily) CI results to Hub ( #31168 )
...
* build
* build
* build
* build
* fix
* fix
* fix
---------
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2024-06-04 21:20:54 +02:00
amyeroberts
99de3a844b
Move out common backbone config param validation ( #31144 )
...
* Move out common validation
* Add missing backbone config arguments
2024-06-04 18:15:37 +01:00
Younes Belkada
485d913dfb
Blip: Deprecate BlipModel
( #31235 )
...
* deprecate blip
* mention deprecation on docs
2024-06-04 18:29:45 +02:00
Yih-Dar
fd3238b4b0
Fix MistralIntegrationTest
( #31231 )
...
* fix
* fix
* fix
* fix
---------
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2024-06-04 18:04:08 +02:00
Manuel Faysse
2965b20459
add no split modules for xlmrobertaxl ( #31223 )
2024-06-04 15:46:19 +01:00
Jacklanda
821b772ab9
Add new line switch before logging ***** Running {description} ***** ( #31225 )
...
✨ Add new line switch before logging "***** Running {description} *****".
Signed-off-by: jacklanda <yonyonlau@gmail.com>
2024-06-04 13:38:17 +01:00
amyeroberts
4ba66fdb4c
Fix pipeline tests - torch imports ( #31227 )
...
* Fix pipeline tests - torch imports
* Frameowrk dependant float conversion
2024-06-04 12:30:23 +01:00
Chujie Zheng
6b22a8f2d8
fix bf16 issue in text classification pipeline ( #30996 )
...
* fix logits dtype
* Add bf16/fp16 tests for text_classification pipeline
* Update test_pipelines_text_classification.py
* fix
* fix
2024-06-04 11:20:48 +01:00
Kristen Pereira
de460e28e1
Add dynamic resolution input/interpolate position embedding to deit ( #31131 )
...
* Added interpolate pos encoding feature and test to deit
* Added interpolate pos encoding feature and test for deit TF model
* readded accidentally delted test for multi_gpu
* storing only patch_size instead of entire config and removed commented code
* Update modeling_tf_deit.py to remove extra line
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
---------
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
2024-06-04 10:29:01 +01:00
Raushan Turganbay
d64e4da713
Video-LLaVa: handle any number of frames ( #31221 )
...
video-llava can handle more frames
2024-06-04 14:20:03 +05:00