Raushan Turganbay
f8b88866f5
[VLMs] support passing embeds along with pixels ( #38467 )
...
* VLMs can work with embeds now
* update more models
* fix tests
* fix copies
* fixup
* fix
* style
* unskip tests
* fix copies
* fix tests
* style
* omni modality models
* qwen models had extra indentation
* fix some other tests
* fix copies
* fix test last time
* unrelated changes revert
* we can't rely only on embeds
* delete file
* de-flake mistral3
* fix qwen models
* fix style
* fix tests
* fix copies
* deflake the test
* modular reverted by fixes, fix again
* flaky test, overwritten
* fix copies
* style
2025-07-01 11:33:20 +00:00
Ayush Singh
20901f1d68
[typing] LlamaAttention return typehint ( #38998 )
...
* helo llama
* helo llama
* helo llama
* apply modular
* fix dia
---------
Co-authored-by: qubvel <qubvel@gmail.com>
2025-07-01 11:29:52 +01:00
Raushan Turganbay
7a25f8dfdb
[qwen2-vl] fix FA2 inference ( #39121 )
...
* fix FA2
* update is causal flag and remove mask for FA2
* update for FA2 with varlen path
* how the tests were passing with different devices?
* add comment and ref to the PR
* move mask preparation to base pretrained model
* seq len is the first dim, not second
* fix copies to fix GLM4V
2025-07-01 10:18:37 +00:00
Mehant Kammakomati
def9663239
feat: support indivisible shards for TP model loading and TPlizing. ( #37220 )
...
* feat: support uneven loading and sharding
resolve merge conflicts
Signed-off-by: Mehant Kammakomati <mehant.kammakomati2@ibm.com>
* fix: allow for empty tensor computations
Signed-off-by: Mehant Kammakomati <mehant.kammakomati2@ibm.com>
* test: add llama1b test case
Signed-off-by: Mehant Kammakomati <mehant.kammakomati2@ibm.com>
* due to q_proj colwise it has to be multi of 2
Signed-off-by: Mehant Kammakomati <mehant.kammakomati2@ibm.com>
* refactor: use slice API
Signed-off-by: Mehant Kammakomati <mehant.kammakomati2@ibm.com>
* refactor: use slice API
Signed-off-by: Mehant Kammakomati <mehant.kammakomati2@ibm.com>
* refactor: use slice API
Signed-off-by: Mehant Kammakomati <mehant.kammakomati2@ibm.com>
* refactor: use slice API
Signed-off-by: Mehant Kammakomati <mehant.kammakomati2@ibm.com>
---------
Signed-off-by: Mehant Kammakomati <mehant.kammakomati2@ibm.com>
2025-07-01 10:03:22 +00:00
jiqing-feng
06c4a4d499
fix caching_allocator_warmup with tie weights ( #39070 )
...
* fix caching_allocator_warmup with tie weights
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
* fix comment
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
---------
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
2025-07-01 11:32:20 +02:00
Raushan Turganbay
e435574721
🚨 Don't use cache in non-generative models ( #38751 )
...
* deprecate for 1 version
* style
* fix some tests
* fix esm
* skip for now, GC requires positional args but we have keyword args
* remove transpose for scores in modified models only
* skip fx trace tests
2025-07-01 09:08:21 +00:00
Cyril Vallez
dbc98328da
Several fixes for Gemma3n ( #39135 )
...
* remove the skips
* fix the epsilon to a small value (does not make sense otherwise)
* safeguard
* overload test_eager_matches_sdpa
* Update test_modeling_common.py
* skip appropriate tests
* correct no_split_layer
* fix all devices issue
* fix backward
* fix
2025-07-01 10:34:53 +02:00
BUI Van Tuan
d53518c5f2
Fix key mapping for VLMs ( #39029 )
...
* fix key mapping for VLMs
* use __mro__ instead
* update key mapping in save_pretrained
2025-07-01 09:47:53 +02:00
eustlb
3457e8e73e
[Whisper] update token timestamps tests ( #39126 )
...
Self-hosted runner (benchmark) / Benchmark (aws-g5-4xlarge-cache) (push) Waiting to run
Build documentation / build (push) Waiting to run
Slow tests on important models (on Push - A10) / Get all modified files (push) Waiting to run
Slow tests on important models (on Push - A10) / Slow & FA2 tests (push) Blocked by required conditions
Self-hosted runner (push-caller) / Check if setup was changed (push) Waiting to run
Self-hosted runner (push-caller) / build-docker-containers (push) Blocked by required conditions
Self-hosted runner (push-caller) / Trigger Push CI (push) Blocked by required conditions
Secret Leaks / trufflehog (push) Waiting to run
Update Transformers metadata / build_and_package (push) Waiting to run
* fixes
* update comment
* update for A10
* all a10
* all a10
* all a10
* all a10
---------
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2025-06-30 21:55:36 +02:00
Drew Ross
fe35eca7bd
Update BigBirdPegasus model card ( #39104 )
...
* Update igbird_pegasus.md
* Apply suggestions from code review
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
---------
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
2025-06-30 10:42:56 -07:00
Yao Matrix
29a3f5ed8c
switch default xpu tp backend to pytorch built-in XCCL from pytorch 2.8 ( #39024 )
...
* switch default xpu tp backend to pytorch built-in XCCL from pytorch 2.8
Signed-off-by: YAO Matrix <matrix.yao@intel.com>
* Update docs/source/en/perf_infer_gpu_multi.md
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
* Update perf_infer_gpu_multi.md
* Update perf_infer_gpu_multi.md
* Update perf_infer_gpu_multi.md
---------
Signed-off-by: YAO Matrix <matrix.yao@intel.com>
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
2025-06-30 08:54:05 -07:00
Vladimir Gutuev
9e0c865b8b
docs: correct two typos in awesome-transformers.md ( #39102 )
...
* docs(awesome-projects): fix typo “Itt leverages” → “It leverages” (#39101 )
closes #39101
* docs(awesome-projects): fix grammar “We provides” → “We provide” (#39101 )
closes #39101
2025-06-30 08:53:43 -07:00
jiqing-feng
03db2700ab
Enable XPU doc ( #38929 )
...
* fix example with dataset
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
* update torchao doc
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
* update torchao doc
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
* fix device type
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
* revert torchao change
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
* fix torchao doc
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
* revert torchao change
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
* update xpu torchao doc
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
* update chat_templating_multimodal.md
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
* use full name for int8
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
* revert int8 title
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
---------
Signed-off-by: jiqing-feng <jiqing.feng@intel.com>
Co-authored-by: Mohamed Mekkouri <93391238+MekkCyber@users.noreply.github.com>
2025-06-30 07:56:55 -07:00
Joao Gante
ea0ea392e5
Fix chat ( #39128 )
2025-06-30 13:47:48 +00:00
Lysandre Debut
ed36f8490e
Licenses ( #39127 )
...
Self-hosted runner (benchmark) / Benchmark (aws-g5-4xlarge-cache) (push) Waiting to run
Build documentation / build (push) Waiting to run
New model PR merged notification / Notify new model (push) Waiting to run
Slow tests on important models (on Push - A10) / Get all modified files (push) Waiting to run
Slow tests on important models (on Push - A10) / Slow & FA2 tests (push) Blocked by required conditions
Self-hosted runner (push-caller) / Check if setup was changed (push) Waiting to run
Self-hosted runner (push-caller) / build-docker-containers (push) Blocked by required conditions
Self-hosted runner (push-caller) / Trigger Push CI (push) Blocked by required conditions
Secret Leaks / trufflehog (push) Waiting to run
Update Transformers metadata / build_and_package (push) Waiting to run
* Licenses
* Licenses
2025-06-30 15:25:36 +02:00
Lysandre Debut
e8f90b5397
Split transformers chat
and transformers serve
( #38443 )
...
* Next token
* Split chat and serve
* Support both generation methods
* Style
* Generation Config
* temp
* temp
* Finalize serving.py
Co-authored-by: =?UTF-8?q?c=C3=A9lina?= <hanouticelina@gmail.com>
* Finalize chat.py
* Update src/transformers/commands/serving.py
Co-authored-by: célina <hanouticelina@gmail.com>
* Lucain's comments
Co-authored-by: Lucain <lucain@huggingface.co>
* Update
* Last comments on PR
* Better error handling
* Better error handling
* CI errors
* CI errors
* Add tests
* Fix tests
* Fix tests
* [chat] Split chat/serve (built on top of lysandre's PR) (#39031 )
* Next token
* Split chat and serve
* Support both generation methods
* Style
* Generation Config
* temp
* temp
* Finalize serving.py
Co-authored-by: =?UTF-8?q?c=C3=A9lina?= <hanouticelina@gmail.com>
* Finalize chat.py
* Update src/transformers/commands/serving.py
Co-authored-by: célina <hanouticelina@gmail.com>
* Lucain's comments
Co-authored-by: Lucain <lucain@huggingface.co>
* Update
* Last comments on PR
* Better error handling
* Better error handling
* CI errors
* CI errors
* Add tests
* Fix tests
* Fix tests
* streaming tool call
* abstract tool state; set tool start as eos
* todos
* server working on models without tools
* rm chat's deprecated flags
* chat defaults
* kv cache persists across calls
* add server docs
* link
* Update src/transformers/commands/serving.py
* Apply suggestions from code review
* i love merge conflicts
* solve multi turn with tiny-agents
* On the fly switching of the models
* Remove required positional arg
---------
Co-authored-by: Lysandre <hi@lysand.re>
Co-authored-by: =?UTF-8?q?c=C3=A9lina?= <hanouticelina@gmail.com>
Co-authored-by: Lucain <lucain@huggingface.co>
* Protect names
* Fix tests
---------
Co-authored-by: =?UTF-8?q?c=C3=A9lina?= <hanouticelina@gmail.com>
Co-authored-by: Lucain <lucain@huggingface.co>
Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com>
2025-06-30 15:10:53 +02:00
Yih-Dar
539c6c2fa8
All CI jobs with A10 ( #39119 )
...
all a10
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2025-06-30 14:23:27 +02:00
Ryan Mullins
ed9f252608
docs: Gemma 3n audio encoder ( #39087 )
...
Updating Gemma 3n docs and docstrings to clarify the relationship
between the newly trained audio encoder used in Gemma 3n and the USM
model from the original paper.
2025-06-30 14:10:51 +02:00
Yuxuan Zhang
4a79bf947d
Fix some bug for finetune and batch infer For GLM-4.1V ( #39090 )
...
* update
* 1
2025-06-30 12:16:22 +02:00
Yao Matrix
2100ee6545
fix UT failures on XPU w/ stock PyTorch 2.7 & 2.8 ( #39116 )
...
* fix UT failures on XPU w/ stock PyTorch 2.7 & 2.8
Signed-off-by: YAO Matrix <matrix.yao@intel.com>
* zamba2
Signed-off-by: YAO Matrix <matrix.yao@intel.com>
* xx
Signed-off-by: YAO Matrix <matrix.yao@intel.com>
* internvl
Signed-off-by: YAO Matrix <matrix.yao@intel.com>
* tp cases
Signed-off-by: YAO Matrix <matrix.yao@intel.com>
---------
Signed-off-by: YAO Matrix <matrix.yao@intel.com>
2025-06-30 11:49:03 +02:00
Yih-Dar
ccf2ca162e
skip some test_sdpa_can_dispatch_on_flash
( #39092 )
...
Self-hosted runner (benchmark) / Benchmark (aws-g5-4xlarge-cache) (push) Has been cancelled
Build documentation / build (push) Has been cancelled
Slow tests on important models (on Push - A10) / Get all modified files (push) Has been cancelled
Self-hosted runner (push-caller) / Check if setup was changed (push) Has been cancelled
Secret Leaks / trufflehog (push) Has been cancelled
Update Transformers metadata / build_and_package (push) Has been cancelled
Slow tests on important models (on Push - A10) / Slow & FA2 tests (push) Has been cancelled
Self-hosted runner (push-caller) / build-docker-containers (push) Has been cancelled
Self-hosted runner (push-caller) / Trigger Push CI (push) Has been cancelled
* fix
* fix
* fix
---------
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2025-06-27 23:08:14 +02:00
st81
a11f692895
Fixes the failing test test_is_split_into_words
in test_pipelines_token_classification.py
( #39079 )
...
Self-hosted runner (benchmark) / Benchmark (aws-g5-4xlarge-cache) (push) Waiting to run
Build documentation / build (push) Waiting to run
Slow tests on important models (on Push - A10) / Get all modified files (push) Waiting to run
Slow tests on important models (on Push - A10) / Slow & FA2 tests (push) Blocked by required conditions
Self-hosted runner (push-caller) / Check if setup was changed (push) Waiting to run
Self-hosted runner (push-caller) / build-docker-containers (push) Blocked by required conditions
Self-hosted runner (push-caller) / Trigger Push CI (push) Blocked by required conditions
Secret Leaks / trufflehog (push) Waiting to run
Update Transformers metadata / build_and_package (push) Waiting to run
New model PR merged notification / Notify new model (push) Has been cancelled
* Fix test pipelines token classification for is_split_into_words
* Fix incorrect import format
2025-06-27 19:25:32 +01:00
Sandeep Yadav
18143c76bf
Sandeepyadav1478/2025 06 19 deberta v2 model card update ( #38895 )
...
* [docs]: update deberta-v2.md model card
* chore: req updates
* chore: address code review feedback and update docs
* chore: review feedback and updates
* chore: model selection updates
* chores: quantizations review updates
2025-06-27 10:35:30 -07:00
Steven Liu
02a769b058
[fix] Add FastSpeech2ConformerWithHifiGan ( #38207 )
...
* add to mapping
* oops
* oops
* add to config_mapping_names
* revert
* fix?
* config-mapping-names
* fix?
* fix?
2025-06-27 09:38:21 -07:00
Benjamin Bossan
c2dc72bb5f
TST Fix PEFT integration test bitsandbytes config ( #39082 )
...
TST Fix PEFT integration test bitsandbytes config
The PEFT integration tests still used load_in_{4,8}_bit, which is
deprecated, moving to properly setting BitsAndBytesConfig. For 4bit,
also ensure that nf4 is being used to prevent
> RuntimeError: quant_type must be nf4 on CPU, got fp4
2025-06-27 18:33:11 +02:00
Matej Sirovatka
c8064bea9a
Fix: unprotected import of tp plugin ( #39083 )
2025-06-27 17:28:05 +02:00
farrosalferro
dd7dc4a4a2
Add Fast Image Processor for Chameleon ( #37140 )
...
* Add Fast Image Processor for Chameleon
* add warning to resize and move blend_rgba to convert_to_rgb
* Remove unrelated files
* Update image_processing_chameleon_fast to use auto_docstring
* fix equivalence test
---------
Co-authored-by: Yoni Gozlan <74535834+yonigozlan@users.noreply.github.com>
Co-authored-by: yonigozlan <yoni.gozlan@huggingface.co>
2025-06-27 15:26:57 +00:00
Yih-Dar
6d773fc3bc
fix dots1
tests ( #39088 )
...
fix
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2025-06-27 16:54:11 +02:00
Tijana Vukovic
c8764ab935
guard torch distributed check ( #39057 )
...
* guard torch distributed check
* Update src/transformers/pipelines/base.py
---------
Co-authored-by: Matt <Rocketknight1@users.noreply.github.com>
2025-06-27 14:49:47 +00:00
MinJu-Ha
49d9fd49bd
Add Fast Image Processor for mobileViT ( #37143 )
...
* Add image_processing_mobilevit_fast.py
* Fix copies
* update _preprocess for channel_flip
* Update for batched image processing
* Resolve merge conflicts with main
* Fix import order and remove trailing whitespace (ruff clean-up)
* Fix copy inconsistencies
* Add NotImplementedError for post_process_semantic_segmentation to satisfy repo checks
* Add auto_docstring
* Adjust style
* Update docs/source/en/model_doc/mobilevit.md
Co-authored-by: Yoni Gozlan <74535834+yonigozlan@users.noreply.github.com>
* Update src/transformers/models/mobilevit/image_processing_mobilevit_fast.py
Co-authored-by: Yoni Gozlan <74535834+yonigozlan@users.noreply.github.com>
* Update src/transformers/models/mobilevit/image_processing_mobilevit_fast.py
Co-authored-by: Yoni Gozlan <74535834+yonigozlan@users.noreply.github.com>
* Delete not used function
* test: add missing tests for and
* Add post_process_semantic_segmentation to mobilevit_fast.py
* Add preprocess function to image_processing_mobilebit_fast.py
* ruff check for formatting
* fix: modify preprocess method to handle BatchFeature correctly
* Remove logic for default value assignment
Co-authored-by: Yoni Gozlan <74535834+yonigozlan@users.noreply.github.com>
* Remove normalization adn RGB conversion logic not used in slow processor
Co-authored-by: Yoni Gozlan <74535834+yonigozlan@users.noreply.github.com>
* Simplify return_tensors logic using one-liner conditional expression
Co-authored-by: Yoni Gozlan <74535834+yonigozlan@users.noreply.github.com>
* Remove unused normalization and format parameters
Co-authored-by: Yoni Gozlan <74535834+yonigozlan@users.noreply.github.com>
* add **kwargs and remove default values in _preprocess
* add slow_fast equivalence tests for segmentation
* style: autoformat code with ruff
* Fix slow_fast equivalence test
* merge + remove skipped test
---------
Co-authored-by: Yoni Gozlan <74535834+yonigozlan@users.noreply.github.com>
Co-authored-by: yonigozlan <yoni.gozlan@huggingface.co>
2025-06-27 14:40:24 +00:00
Nahieli
4336ecd1ea
add fast image processor nougat ( #37661 )
...
* add fast image processor nougat
* test fixes
* docstring white space
* last fixes
* docstring_type
* tolerance unit test
* fix tolerance
* fix rtol
* remove traling white space
* remove white space
* note for tolerance unit test
* fix tests
* remove print
---------
Co-authored-by: yonigozlan <yoni.gozlan@huggingface.co>
Co-authored-by: Yoni Gozlan <74535834+yonigozlan@users.noreply.github.com>
2025-06-27 14:39:43 +00:00
Benjamin Bossan
0c35280e58
TST PEFT integration tests with pipeline generate ( #39086 )
...
Some PEFT integration tests involving text generation pipelines were
failing since #38129 because the base model is too small to generate
longer sequences. Setting max_new_tokens fixes this.
2025-06-27 15:58:10 +02:00
JINO ROHIT
993665a5ff
fixed typo for docstring in prepare_inputs method ( #39071 )
2025-06-27 13:57:56 +00:00
Yih-Dar
839893c86b
fix mistral3
tests ( #38989 )
...
* fix
* fix
* fix
* fix
* fix
---------
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2025-06-27 15:44:10 +02:00
eustlb
2b85b6ce19
[Whisper] 🚨 Fix pipeline word timestamp: timestamp token is end of token time !!! ( #36632 )
...
* timestamp token is end of token time !!!
* ensure correct alignment between tokens and timestamp tokens
* ignore input tokens for DTW computation
* use num_frames to avoid token timestamp hallucinations
* token timestamps test updates !
* num_frames: deprecate and use attention_mask instead
* avoid breaking change
* fix the pipeline usage for chunk approach
* make style
* better logging
* better logging
* make style
* update tests with correct values
2025-06-27 12:51:43 +00:00
eustlb
9c8d3a70b8
Pipeline: fix unnecessary warnings ( #35753 )
...
* return attention mask
* use correct model input name
* fix
* make
2025-06-27 14:32:03 +02:00
Yaswanth Gali
1750c518dd
✨ Add EoMT Model || 🚨 Fix Mask2Former loss calculation ( #37610 )
...
* Initial Commit
* up
* More changes
* up
* Only mask_logits mismatch
* close enough logits debug later
* fixes
* format
* Add dummy loss
* Close enough processing for semantic seg
* nit
* Added panoptic postprocessor
* refactor
* refactor
* finally fixed panoptic postprocessor
* temp update
* Refactor ForUniversalSegmentation class
* nits and config update
* Few fixes and inference matches
* change mapping
* Added training support but loss slightly off 🥲
* Loss is matching 😀
* update
* Initial tests skelton
* changes
* tests update
* more modular
* initial tests
* updates
* better docstrings
* changes
* proc tests passing :)
* Image processor update
* tiny change
* QOL changes
* Update test w.r.t latest attn refactor
* repo-consistency fixes
* up
* Image proc fix and integration tests :)
* docs update
* integration tests
* fix
* docs update 🥰
* minor fix
* Happy CI
* fix
* obvious refactoring
* refactoring w.r.t review
* Add fask image proc skelton
* Fast Image proc and cleanups
* Use more modular
* tests update
* Add more tests
* Nit
* QOL updates
* change init_weights to torch default
* add eager func coz of make style
* up
* changes
* typo fix
* Updates
* More deterministic tests
* More modular
* go more modular 🚀
* up
* dump
* add supprot for giant ckpts
* overhaul
* modular
* refactor
* instace seg is ready
* cleanup
* forgot this
* docs cleanup
* minor changes
* EoMT - > Eomt
* Happy CI
* remove redundant comment
* Change model references
* final change
* check annealing per block
* My other PR changes 😂
---------
Co-authored-by: Cyril Vallez <cyril.vallez@huggingface.co>
2025-06-27 14:18:18 +02:00
Yao Matrix
0106a50a6b
fix a bunch of XPU UT failures on stock PyTorch 2.7 and 2.8 ( #39069 )
...
* fix a bunch of XPU UT failures on stock PyTorch 2.7 and 2.8
Signed-off-by: YAO Matrix <matrix.yao@intel.com>
* qwen3
Signed-off-by: YAO Matrix <matrix.yao@intel.com>
* quanto
Signed-off-by: YAO Matrix <matrix.yao@intel.com>
* models
Signed-off-by: YAO Matrix <matrix.yao@intel.com>
* fix style
Signed-off-by: YAO Matrix <matrix.yao@intel.com>
* idefics2
Signed-off-by: YAO Matrix <matrix.yao@intel.com>
---------
Signed-off-by: YAO Matrix <matrix.yao@intel.com>
2025-06-27 14:01:53 +02:00
Mohamed Mekkouri
cb17103bd5
Uninstallling Flash attention from quantization docker ( #39078 )
...
Self-hosted runner (benchmark) / Benchmark (aws-g5-4xlarge-cache) (push) Waiting to run
Build documentation / build (push) Waiting to run
New model PR merged notification / Notify new model (push) Waiting to run
Slow tests on important models (on Push - A10) / Get all modified files (push) Waiting to run
Slow tests on important models (on Push - A10) / Slow & FA2 tests (push) Blocked by required conditions
Self-hosted runner (push-caller) / Check if setup was changed (push) Waiting to run
Self-hosted runner (push-caller) / build-docker-containers (push) Blocked by required conditions
Self-hosted runner (push-caller) / Trigger Push CI (push) Blocked by required conditions
Secret Leaks / trufflehog (push) Waiting to run
Update Transformers metadata / build_and_package (push) Waiting to run
* update
* revert
2025-06-27 13:51:46 +02:00
BUI Van Tuan
371c471113
Fix initialization of OneFormer ( #38901 )
...
* fix initialization of OneFormer
* remove redundant initializations
* remove redundant initializations
* remove redundant initializations
* keep BC
2025-06-27 12:39:37 +02:00
Yih-Dar
540a10848c
fix Gemma3nProcessorTest
( #39068 )
...
* fix
* fix
* oups forgot style
---------
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Co-authored-by: Cyril Vallez <cyril.vallez@gmail.com>
2025-06-27 12:28:10 +02:00
Yaswanth Gali
0d66ef7792
Cleanup Attention class for Siglip and dependent models ( #39040 )
...
* cleanup attention class
* More models
* more models
* Changes
* make style
* Should fix CI
* This should work 🙏
2025-06-27 12:14:09 +02:00
eustlb
1ccc73dee9
[Whisper] fix shape mismatch in tests ( #39074 )
...
fix shape mismatch
2025-06-27 09:27:42 +00:00
Steven Liu
a52478253b
[docs] Tensor parallelism ( #38241 )
...
Self-hosted runner (benchmark) / Benchmark (aws-g5-4xlarge-cache) (push) Waiting to run
Build documentation / build (push) Waiting to run
Slow tests on important models (on Push - A10) / Get all modified files (push) Waiting to run
Slow tests on important models (on Push - A10) / Slow & FA2 tests (push) Blocked by required conditions
Secret Leaks / trufflehog (push) Waiting to run
Update Transformers metadata / build_and_package (push) Waiting to run
* updates
* feedback
* badges
* fix?
* fix?
* fix?
* fix?
2025-06-26 14:40:45 -07:00
Steven Liu
84e8696cae
[docs] @auto_docstring ( #39011 )
...
* refactor
* feedback
2025-06-26 14:21:54 -07:00
Drew Ross
018855de63
Update PEGASUS-X model card ( #38971 )
...
* Update PEGASUS-X model card
* Add cache_implementation argument in quantization code example
* Update CLI example
* Apply suggestions from code review
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
* Remove TensorFlow and Flax badges
---------
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
2025-06-26 13:54:48 -07:00
Steven Liu
757c26fb40
[docs] Model contribution ( #38995 )
...
Self-hosted runner (benchmark) / Benchmark (aws-g5-4xlarge-cache) (push) Waiting to run
Build documentation / build (push) Waiting to run
New model PR merged notification / Notify new model (push) Waiting to run
Slow tests on important models (on Push - A10) / Get all modified files (push) Waiting to run
Slow tests on important models (on Push - A10) / Slow & FA2 tests (push) Blocked by required conditions
Self-hosted runner (push-caller) / Check if setup was changed (push) Waiting to run
Self-hosted runner (push-caller) / build-docker-containers (push) Blocked by required conditions
Self-hosted runner (push-caller) / Trigger Push CI (push) Blocked by required conditions
Secret Leaks / trufflehog (push) Waiting to run
Update Transformers metadata / build_and_package (push) Waiting to run
improve
2025-06-26 12:25:14 -07:00
Yih-Dar
b372bb5ed1
fix layoutlmv3
tests ( #39050 )
...
* fix
* fix
* fix
* fix
* fix
* fix
---------
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2025-06-26 20:07:17 +02:00
StevenBucaille
f171e7e884
Update SuperPoint model card ( #38896 )
...
* docs: first draft to more standard SuperPoint documentation
* Apply suggestions from code review
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
* docs: reverted changes on Auto classes
* docs: addressed the rest of the comments
* docs: remove outdated reference to keypoint detection task guide in SuperPoint documentation
* Update superpoint.md
---------
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
2025-06-26 10:13:06 -07:00
Yih-Dar
2f50230c59
fix t5gemma
tests ( #39052 )
...
* fix
* fix
* fix
* fix
* fix
---------
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2025-06-26 18:48:14 +02:00