Yih-Dar
3fb7e7bc01
Skip torchscript tests for 2 models ( #38643 )
...
Self-hosted runner (benchmark) / Benchmark (aws-g5-4xlarge-cache) (push) Waiting to run
Build documentation / build (push) Waiting to run
Slow tests on important models (on Push - A10) / Get all modified files (push) Waiting to run
Slow tests on important models (on Push - A10) / Slow & FA2 tests (push) Blocked by required conditions
Self-hosted runner (push-caller) / Check if setup was changed (push) Waiting to run
Self-hosted runner (push-caller) / build-docker-containers (push) Blocked by required conditions
Self-hosted runner (push-caller) / Trigger Push CI (push) Blocked by required conditions
Secret Leaks / trufflehog (push) Waiting to run
Update Transformers metadata / build_and_package (push) Waiting to run
fix
Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
2025-06-06 20:17:37 +02:00
Arthur
f5d45d89c4
🚨 Early-error 🚨 config will error out if output_attentions=True
and the attn implementation is wrong ( #38288 )
...
* Protect ParallelInterface
* early error out on output attention setting for no wraning in modeling
* modular update
* fixup
* update model tests
* update
* oups
* set model's config
* more cases
* ??
* properly fix
* fixup
* update
* last onces
* update
* fix?
* fix wrong merge commit
* fix hub test
* nits
* wow I am tired
* updates
* fix pipeline!
---------
Co-authored-by: Lysandre <hi@lysand.re>
2025-05-23 17:17:38 +02:00
cyyever
1e6b546ea6
Use Python 3.9 syntax in tests ( #37343 )
...
Signed-off-by: cyy <cyyever@outlook.com>
2025-04-08 14:12:08 +02:00
Matt
2d46a08b63
Purge unused ModelTester code ( #37085 )
...
* Purge correctly this time
* Remove more methods from recent PRs
* make fixup
2025-04-03 17:48:35 +01:00
Joao Gante
678885bbbd
[CI] Check test if the GenerationTesterMixin
inheritance is correct 🐛 🔫 ( #36180 )
2025-02-21 10:18:20 +00:00
Joao Gante
62c7ea0201
CI: avoid human error, automatically infer generative models ( #33212 )
...
* tmp commit
* move tests to the right class
* remove ALL all_generative_model_classes = ...
* skip tf roberta
* skip InstructBlipForConditionalGenerationDecoderOnlyTest
* videollava
* reduce diff
* reduce diff
* remove on vlms
* fix a few more
* manual rebase bits
* more manual rebase
* remove all manual generative model class test entries
* fix up to ernie
* a few more removals
* handle remaining cases
* recurrent gemma
* it's better here
* make fixup
* tf idefics is broken
* tf bert + generate is broken
* don't touch tf :()
* don't touch tf :(
* make fixup
* better comments for test skips
* revert tf changes
* remove empty line removal
* one more
* missing one
2025-02-13 16:27:11 +01:00
Arthur
b912f5ee43
use torch.testing.assertclose instead to get more details about error in cis ( #35659 )
...
* use torch.testing.assertclose instead to get more details about error in cis
* fix
* style
* test_all
* revert for I bert
* fixes and updates
* more image processing fixes
* more image processors
* fix mamba and co
* style
* less strick
* ok I won't be strict
* skip and be done
* up
2025-01-24 16:55:28 +01:00
eustlb
5f0f4b1b93
Patch moonshine ( #35731 )
...
* udpate expected logits for T4 runners
* update doc
* correct order of the args for better readability
* remove generate wrap
* convert modular
2025-01-20 16:19:29 +01:00
eustlb
5f087d1335
Add Moonshine ( #34784 )
...
* config draft
* full encoder forward
* full decoder forward
* fix sdpa and FA2
* fix sdpa and FA2
* moonshine model
* moonshine model forward
* fix attention with past_key_values
* add MoonshineForConditionalGeneration
* fix cache handling and causality for cross attention
* no causal attention mask for the encoder
* model addition (imports etc)
* small nit
* nits
* Update src/transformers/models/moonshine/convert_usefulsensors_to_hf.py
Co-authored-by: Joshua Lochner <admin@xenova.com>
* add rope_theta
* nits
* model doc
* Update src/transformers/models/auto/configuration_auto.py
Co-authored-by: Joshua Lochner <admin@xenova.com>
* imports
* add MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES
* updates modular
* make
* make fix-copies
* ruff check examples fix
* fix check_modular_conversion
* nit
* nits
* nits
* copied from -> imports
* imports fix
* integrate attention refacto
* modular edge case
* remove encoder
* convolutions params in config
* run modular_model_converter
* make
* Update docs/source/en/model_doc/moonshine.md
Co-authored-by: Joshua Lochner <admin@xenova.com>
* MoonshineModelTest
* correct typo
* make style
* integration tests
* make
* modular convert
* name conversion update (up_proj -> fc1 etc)
* update config
* update MLP
* update attention
* update encoder layer
* update decoder layer
* update convolutions parameters
* update encoder
* remove INPUTS_DOCSTRING
* update decoder
* update conditional generation
* update pretrained model
* imports
* modular converted
* update doc
* fix
* typo
* update doc
* update license
* update init
* split config in file
* two classes for MLP
* attention from GLM
* from GlmRotaryEmbedding
* split MLP
* apply arthur's review suggestions
* apply arthur's review suggestions
* apply arthur's review suggestions
* auto feature extractor
* convert modular
* fix + make
* convert modular
* make
* unsplit config
* use correct checkpoint
* wrap generate
* update tests
* typos
* make
* typo
* update doc
---------
Co-authored-by: Joshua Lochner <admin@xenova.com>
2025-01-10 11:00:54 +01:00