NielsRogge
63ffd56d02
Add SiglipForImageClassification and CLIPForImageClassification ( #28952 )
...
* First draft
* Add CLIPForImageClassification
* Remove scripts
* Fix doctests
2024-02-14 08:41:31 +01:00
Jonathan Tow
de6029a059
Add StableLM
( #28810 )
...
* Add `StableLM`
* fix(model): re-create from `huggingface-cli add-new-model-like persimmon`
* fix: re-add changes to address comments
* fix(readme): add links to paper
* fix(tokenization_auto): remove `GPTNeoXTokenizerFastFast` ref
* fix(tests): re-add `@slow` decorator to integration tests
* fix(tests): import slow...
* fix(readme_hd): remove whitespace edit
* fix(tokenizer): auto tokenizer tuple
* skip doctests for `modeling_stablelm`
2024-02-14 07:15:18 +01:00
Younes Belkada
164bdef8cc
ENH [AutoQuantizer
]: enhance trainer + not supported quant methods ( #28991 )
...
* enhance trainer + not support quant methods
* remove all old logic
* add version
2024-02-14 01:30:23 +01:00
Younes Belkada
1d12b8bc25
ENH: Do not pass warning message in case quantization_config
is in config but not passed as an arg ( #28988 )
...
* Update auto.py
* Update auto.py
* Update src/transformers/quantizers/auto.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* Update src/transformers/quantizers/auto.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
---------
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
2024-02-14 01:19:42 +01:00
amyeroberts
bd4b83e1ba
[DETR
] Update the processing to adapt masks & bboxes to reflect padding ( #28363 )
...
* Update the processing so bbox coords are adjusted for padding
* Just pad masks
* Tidy up, add tests
* Better tests
* Fix yolos and mark as slow for pycocotols
* Fix yolos - return_tensors
* Clarify padding and normalization behaviour
2024-02-13 18:27:06 +00:00
Aditya Kane
3de6a6b493
Update configuration_llama.py: fixed broken link ( #28946 )
...
* Update configuration_llama.py: fix broken link
* [Nit] Explicit redirection not required
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
---------
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
2024-02-13 13:02:07 +00:00
Joao Gante
3e70a207df
Static Cache: load models with MQA or GQA ( #28975 )
2024-02-13 09:58:19 +00:00
Hiroshi Matsuda
da20209dbc
Add sudachi_projection option to BertJapaneseTokenizer ( #28503 )
...
* add sudachi_projection option
* Upgrade sudachipy>=0.6.8
* add a test case for sudachi_projection
* Compatible with older versions of SudachiPy
* make fixup
* make style
* error message for unidic download
* revert jumanpp test cases
* format options for sudachi_projection
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* format options for sudachi_split_mode and sudachi_dict_type
* comment
* add tests for full_tokenizer kwargs
* pass projection arg directly
* require_sudachi_projection
* make style
* revert upgrade sudachipy
* check is_sudachi_projection_available()
* revert dependency_version_table and bugfix
* style format
* simply raise ImportError
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
* simply raise ImportError
---------
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
2024-02-13 04:47:20 +01:00
Arthur
b44567538b
[NllbTokenizer
] refactor with added tokens decoder ( #27717 )
...
* refactor with addedtokens decoder
* style
* get rid of lang code to id
* style
* keep some things for BC
* update tests
* add the mask token at the end of the vocab
* nits
* nits
* fix final tests
* style
* nits
* Update src/transformers/models/nllb/tokenization_nllb_fast.py
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
* nits
* style?
* Update src/transformers/convert_slow_tokenizer.py
* make it a tad bit more custom
* ruff please stop
Co-Authored by avidale
<dale.david@mail.ru>
* Update
Co-authored-by: avidale
<dale.david@mail.ru>
* Update
Co-authored-by: avidale <dale.david@mail.ru>
* oupts
* ouft
* nites
* test
* fix the remaining failing tests
* style
* fix failing test
* ficx other test
* temp dir + test the raw init
* update test
* style
---------
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
2024-02-13 03:49:20 +01:00
Klaus Hipp
d90acc1643
[i18n-de] Translate CONTRIBUTING.md to German ( #28954 )
...
* Translate contributing.md to German
* Fix formatting issues in contributing.md
* Address review comments
* Fix capitalization
2024-02-12 13:39:20 -08:00
NielsRogge
78ba9f4617
[Docs] Add video section ( #28958 )
...
Add video section
2024-02-12 19:50:31 +01:00
Klaus Hipp
fe3df9d5b3
[Docs] Add language identifiers to fenced code blocks ( #28955 )
...
Add language identifiers to code blocks
2024-02-12 10:48:31 -08:00
Yunxuan Xiao
c617f988f8
Clean up staging tmp checkpoint directory ( #28848 )
...
clean up remaining tmp checkpoint dir
Signed-off-by: woshiyyya <xiaoyunxuan1998@gmail.com>
2024-02-12 15:47:21 +00:00
JB (Don)
136cd893dc
Always initialize tied output_embeddings if it has a bias term ( #28947 )
...
Continue to initialize tied output_embeddings if it has a bias term
The bias term is not tied, and so will need to be initialized accordingly.
2024-02-12 15:47:08 +00:00
Alexey Fadeev
792819f6cf
Updated requirements for image-classification samples: datasets>=2.14.0 ( #28974 )
...
Updated datasets requirements. Need a package version >= 2.14.0
2024-02-12 14:57:25 +00:00
Joao Gante
e30bbb2685
Tests: tag test_save_load_fast_init_from_base
as flaky ( #28930 )
2024-02-12 14:43:34 +00:00
cmahmut
1709886eba
[pipelines
] updated docstring with vqa alias ( #28951 )
...
updated docstring with vqa alias
2024-02-12 14:34:08 +00:00
Kossai Sbai
cf4c20b9fb
Convert torch_dtype
as str
to actual torch data type (i.e. "float16" …to torch.float16
) ( #28208 )
...
* Convert torch_dtype as str to actual torch data type (i.e. "float16" to torch.float16)
* Check if passed torch_dtype is an attribute in torch
* Update src/transformers/pipelines/__init__.py
Check type via isinstance
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
---------
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
2024-02-12 14:04:53 +00:00
NielsRogge
ef5ab72f4b
[Docs] Update README and default pipelines ( #28864 )
...
* Update README and docs
* Update README
* Update README
2024-02-12 10:21:36 +01:00
NielsRogge
f278ef20ed
[Nougat] Fix pipeline ( #28242 )
...
* Fix pipeline
* Remove print statements
* Address comments
* Address issue
* Remove unused imports
2024-02-12 10:21:15 +01:00
Klaus Hipp
58e3d23e97
[i18n-de] Translate README.md to German ( #28933 )
...
* Translate README.md to German
* Add links to README_de.md
* Remove invisible characters in README
* Change to a formal tone and fix punctuation marks
2024-02-09 12:56:22 -08:00
Philip Blair
d123e661e4
Fix type annotations on neftune_noise_alpha and fsdp_config TrainingArguments parameters ( #28942 )
2024-02-09 15:42:01 +00:00
Yuki Watanabe
ebf3ea2788
Fix a wrong link to CONTRIBUTING.md section in PR template ( #28941 )
2024-02-09 15:10:47 +00:00
Karl Hajjar
de11e654c9
Fix max_position_embeddings default value for llama2 to 4096 #28241 ( #28754 )
...
* Changed max_position_embeddings default value from 2048 to 4096
* force push
* Fixed formatting issues. Fixed missing argument in write_model.
* Reverted to the default value 2048 in the Llama config. Added comments for the llama_version argument.
* Fixed issue with default value value of max_position_embeddings in docstring
* Updated help message for llama versions
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
---------
Co-authored-by: amyeroberts <22614925+amyeroberts@users.noreply.github.com>
2024-02-09 10:24:01 +00:00
Klaus Hipp
2749e479f3
[Docs] Fix broken links and syntax issues ( #28918 )
...
* Fix model documentation links in attention.md
* Fix external link syntax
* Fix target anchor names of section links
* Fix copyright statement comments
* Fix documentation headings
2024-02-08 14:13:35 -08:00
Raushan Turganbay
d628664688
Support batched input for decoder start ids ( #28887 )
...
* support batched input for decoder start ids
* Fix typos
Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com>
* minor changes
* fix: decoder_start_id as list
* empty commit
* empty commit
* empty commit
* empty commit
* empty commit
* empty commit
* empty commit
* empty commit
* empty commit
---------
Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com>
2024-02-08 16:00:53 +00:00
Raushan Turganbay
cc309fd406
pass kwargs in stopping criteria list ( #28927 )
2024-02-08 15:38:29 +00:00
vodkaslime
0b693e90e0
fix: torch.int32 instead of torch.torch.int32 ( #28883 )
2024-02-08 16:28:17 +01:00
Matt
693667b8ac
Remove dead TF loading code ( #28926 )
...
Remove dead code
2024-02-08 14:17:33 +00:00
Arthur
115ac94d06
[Core generation
] Adds support for static KV cache ( #27931 )
...
Co-authored-by: fxmarty <9808326+fxmarty@users.noreply.github.com>
Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com>
2024-02-08 11:50:34 +01:00
Javier
4b236aed76
Fix utf-8 yaml load for marian conversion to pytorch in Windows ( #28618 )
...
Fix utf-8 yaml in marian conversion
2024-02-08 08:23:15 +01:00
Klaus Hipp
33df036917
[Docs] Revert translation of '@slow' decorator ( #28912 )
2024-02-08 03:31:47 +01:00
Klaus Hipp
328ade855b
[Docs] Fix placement of tilde character ( #28913 )
...
Fix placement of tilde character
2024-02-07 17:19:39 -08:00
Huazhong Ji
5f96855761
Add npu device for pipeline ( #28885 )
...
add npu device for pipeline
Co-authored-by: unit_test <test@unit.com>
2024-02-07 17:27:01 +00:00
Yih-Dar
308d2b9004
Update the cache number ( #28905 )
...
* fix
* fix
* fix
---------
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2024-02-07 16:37:09 +01:00
Daniel Korat
abf8f54a01
⚠️ Raise Exception
when trying to generate 0 tokens ⚠️ ( #28621 )
...
* change warning to exception
* Update src/transformers/generation/utils.py
Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com>
* validate `max_new_tokens` > 0 in `GenerationConfig`
* fix truncation test parameterization in `TextGenerationPipelineTests`
---------
Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com>
2024-02-07 13:42:01 +01:00
Matt
349a6e8542
Fix Keras scheduler import so it works for older versions of Keras ( #28895 )
...
Fix our schedule import so it works for older versions of Keras
2024-02-07 12:28:24 +00:00
Sourab Mangrulkar
d9deddb4c1
fix Starcoder FA2 implementation ( #28891 )
2024-02-07 14:10:10 +05:30
Sai-Suraj-27
64d1518cbf
fix: Fixed the documentation for logging_first_step
by removing "evaluate" ( #28884 )
...
Fixed the documentation for logging_first_step by removing evaluate.
2024-02-07 08:46:36 +01:00
Klaus Hipp
1c31b7aa3b
[Docs] Add missing language options and fix broken links ( #28852 )
...
* Add missing entries to the language selector
* Add links to the Colab and AWS Studio notebooks for ONNX
* Use anchor links in CONTRIBUTING.md
* Fix broken hyperlinks due to spaces
* Fix links to OpenAI research articles
* Remove confusing footnote symbols from author names, as they are also considered invalid markup
2024-02-06 12:01:01 -08:00
Yih-Dar
40658be461
Hotfix - make torchaudio
get the correct version in torch_and_flax_job
( #28899 )
...
* check
* check
* check
---------
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2024-02-06 21:00:42 +01:00
Klaus Hipp
4830f26965
[Docs] Fix backticks in inline code and documentation links ( #28875 )
...
Fix backticks in code blocks and documentation links
2024-02-06 11:15:44 -08:00
Lucain
a1afec9e17
Explicit server error on gated model ( #28894 )
2024-02-06 17:45:20 +00:00
Yih-Dar
89439fea64
unpin torch ( #28892 )
...
* unpin torch
* check
* check
* check
---------
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2024-02-06 17:21:05 +01:00
Yih-Dar
76b4f666f5
Revert "[WIP] Hard error when ignoring tensors." ( #28898 )
...
Revert "[WIP] Hard error when ignoring tensors. (#27484 )"
This reverts commit 2da28c4b41
.
2024-02-06 17:18:30 +01:00
Yih-Dar
6529a5b5c1
Fix FastSpeech2ConformerModelTest
and skip it on CPU ( #28888 )
...
* fix
* fix
---------
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2024-02-06 11:05:23 +01:00
Sourab Mangrulkar
5346db1684
Raise error when using save_only_model
with load_best_model_at_end
for DeepSpeed/FSDP ( #28866 )
...
* Raise error when using `save_only_model` with `load_best_model_at_end` for DeepSpeed/FSDP
* Update trainer.py
2024-02-06 11:25:44 +05:30
Eran Hirsch
ee2a3400f2
Fix LongT5ForConditionalGeneration initialization of lm_head ( #28873 )
2024-02-06 04:24:20 +01:00
Klaus Hipp
1ea0bbd73c
[Docs] Update project names and links in awesome-transformers ( #28878 )
...
Update project names and repository links in awesome-transformers
2024-02-06 04:06:29 +01:00
dependabot[bot]
e83227d76e
Bump cryptography from 41.0.2 to 42.0.0 in /examples/research_projects/decision_transformer ( #28879 )
...
Bump cryptography in /examples/research_projects/decision_transformer
Bumps [cryptography](https://github.com/pyca/cryptography ) from 41.0.2 to 42.0.0.
- [Changelog](https://github.com/pyca/cryptography/blob/main/CHANGELOG.rst )
- [Commits](https://github.com/pyca/cryptography/compare/41.0.2...42.0.0 )
---
updated-dependencies:
- dependency-name: cryptography
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2024-02-06 03:53:08 +01:00