transformers/tests/models
Stas Bekman 6c811a322f
new model: IDEFICS via HuggingFaceM4 (#24796)
* rename

* restore

* mappings

* unedited tests+docs

* docs

* fixes

* fix auto-sync breakage

* cleanup

* wip

* wip

* add fetch_images

* remove einops dependency

* update

* fix

* fix

* fix

* fix

* fix

* re-add

* add batching

* rework

* fix

* improve

* add Leo as I am extending his work

* cleanup

* fix

* cleanup

* slow-test

* fix

* fix

* fixes

* deal with warning

* rename modified llama classes

* rework fetch_images

* alternative implementation

* cleanup

* strict version

* cleanup

* [`IDEFICS`] Fix idefics ci (#25056)

* Fix IDEFICS CI

* fix test file

* fixup

* some changes to make tests pass

* fix

* fixup

* Update src/transformers/models/idefics/configuration_idefics.py

Co-authored-by: Stas Bekman <stas00@users.noreply.github.com>

---------

Co-authored-by: Stas Bekman <stas00@users.noreply.github.com>

* remove compat checks

* style

* explain that Idefics is not for training from scratch

* require pt>=2.0

* fix idefics vision config (#25092)

* fix idefics vision config

* fixup

* clean

* Update src/transformers/models/idefics/configuration_idefics.py

---------

Co-authored-by: Stas Bekman <stas00@users.noreply.github.com>

* cleanup

* style

* cleanup

* Apply suggestions from code review

Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>

* upcase

* sequence of images

* handle the case with no images

* Update src/transformers/image_processing_utils.py

Co-authored-by: Victor SANH <victorsanh@gmail.com>

* support pure lm take 2

* support tokenizer options

* parameterize num_channels

* fix upcase

* s|IdeficsForCausalLM|IdeficsForVisionText2Text|g

* manual to one line

* addressing review

* unbreak

* remove clip dependency

* fix test

* consistency

* PIL import

* Idefics prefix

* Idefics prefix

* hack to make tests work

* style

* fix

* fix

* revert

* try/finally

* cleanup

* clean up

* move

* [`IDEFICS`] Fix idefics config refactor (#25149)

* refactor config

* nuke init weights

* more refactor

* oops

* remove visual question answering pipeline support

* Update src/transformers/models/idefics/clip.py

Co-authored-by: Stas Bekman <stas00@users.noreply.github.com>

* Update src/transformers/models/idefics/modeling_idefics.py

* cleanup

* mv clip.py vision.py

* tidyup

---------

Co-authored-by: Stas Bekman <stas00@users.noreply.github.com>
Co-authored-by: Stas Bekman <stas@stason.org>

* fix

* license

* condition on pt

* fix

* style

* fix

* rm torchvision dependency, allow custom transforms

* address review

* rework device arg

* add_eos_token

* s/transforms/transform/

* fix top level imports

* fix return value

* cleanup

* cleanup

* fix

* style

* license

* license

* Update src/transformers/models/idefics/image_processing_idefics.py

Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>

* add a wrapper to freeze vision layears

* tidyup

* use the correct std/mean settings

* parameterize values from config

* add tests/models/idefics/test_image_processing_idefics.py

* add test_processor_idefics.py

* cleanup

* cleanups

* fix

* fix

* move to the right group

* style

* Apply suggestions from code review

Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>

* add perceiver config

* reset

* missing arg docs

* Apply suggestions from code review

Co-authored-by: Leo Tronchon <leo.tronchon@gmail.com>

* address review comments

* inject automatic end of utterance tokens (#25218)

* inject automatic end of utterance tokens

* fix

* fix

* fix

* rework to not use the config

* not end_of_utterance_token at the end

* Update src/transformers/models/idefics/processing_idefics.py

Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>

* address review

* Apply suggestions from code review

Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com>

* Update src/transformers/image_processing_utils.py

Co-authored-by: Nicolas Patry <patry.nicolas@protonmail.com>

* [`Idefics`] add image_embeddings option in generate-related methods (#25442)

* add image_embeddings option in generate-related methods

* style

* rename image_embeddings and allow perceiver embeddings precomputation

* compute embeddings within generate

* make is_encoder_decoder= True the default in config

* nested if else fix

* better triple check

* switch if elif order for pixel values / img embeds

* update model_kwargs perceiver only at the end

* use _prepare_model_inputs instead of encoder_decoder logic

* fix comment typo

* fix config default for is_encoder_decoder

* style

* add typehints

* precompute in forward

* doc builder

* style

* pop instead of get image hidden states

* Trigger CI

* Update src/transformers/models/idefics/modeling_idefics.py

Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>

* Update src/transformers/models/idefics/modeling_idefics.py

Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>

* fix * + indentation + style

* simplify a bit the use_resampler logic using comments

* update diocstrings

* Trigger CI

---------

Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>

* fix rebase changes

* unbreak #25237 - to be fixed in follow up PRs

* is_composition = False

* no longer needed

---------

Co-authored-by: leot13 <leo.tronchon@gmail.com>
Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
Co-authored-by: Victor SANH <victorsanh@gmail.com>
Co-authored-by: Joao Gante <joaofranciscocardosogante@gmail.com>
Co-authored-by: Nicolas Patry <patry.nicolas@protonmail.com>
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
2023-08-18 14:12:28 -07:00
..
albert CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
align CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
altclip CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
audio_spectrogram_transformer CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
auto Remote code improvements (#23959) 2023-06-06 14:31:14 -04:00
autoformer Compute dropout_probability only in training mode (#24486) 2023-06-26 18:36:47 +02:00
bark Update Bark generation configs and tests (#25409) 2023-08-09 18:28:02 +02:00
bart CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
barthez Update quality tooling for formatting (#21480) 2023-02-06 18:10:56 -05:00
bartpho Update quality tooling for formatting (#21480) 2023-02-06 18:10:56 -05:00
beit Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
bert CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
bert_generation CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
bert_japanese Update quality tooling for formatting (#21480) 2023-02-06 18:10:56 -05:00
bertweet Update quality tooling for formatting (#21480) 2023-02-06 18:10:56 -05:00
big_bird Revert "Fix gradient checkpointing + fp16 autocast for most models" (#24420) 2023-06-22 16:11:27 +02:00
bigbird_pegasus CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
biogpt CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
bit Update old existing feature extractor references (#24552) 2023-06-29 10:17:36 +01:00
blenderbot CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
blenderbot_small CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
blip Input data format (#25464) 2023-08-16 17:45:02 +01:00
blip_2 CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
bloom Replaces calls to .cuda with .to(torch_device) in tests (#25571) 2023-08-18 12:40:40 +02:00
bridgetower Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
byt5 Update quality tooling for formatting (#21480) 2023-02-06 18:10:56 -05:00
camembert Better TF docstring types (#23477) 2023-05-24 13:52:52 +01:00
canine CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
chinese_clip Input data format (#25464) 2023-08-16 17:45:02 +01:00
clap CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
clip Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
clipseg Fix test_model_parallelism (#25359) 2023-08-08 10:48:45 +02:00
codegen CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
conditional_detr Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
convbert CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
convnext Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
convnextv2 Update ConvNextV2ModelIntegrationTest::test_inference_image_classification_head (#23402) 2023-05-16 23:35:11 +02:00
cpm Fix PipelineTests skip conditions (#22320) 2023-03-22 20:02:24 +01:00
cpmant CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
ctrl CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
cvt Make more test models smaller (#25005) 2023-07-24 10:08:47 -04:00
data2vec Fix test_model_parallelism (#25359) 2023-08-08 10:48:45 +02:00
deberta CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
deberta_v2 CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
decision_transformer 🔥Rework pipeline testing by removing PipelineTestCaseMeta 🚀 (#21516) 2023-02-28 19:40:57 +01:00
deformable_detr Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
deit Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
deta Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
detr Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
dinat Update old existing feature extractor references (#24552) 2023-06-29 10:17:36 +01:00
dinov2 CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
distilbert CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
dit Update old existing feature extractor references (#24552) 2023-06-29 10:17:36 +01:00
donut Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
dpr CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
dpt Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
efficientformer Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
efficientnet 🚨🚨🚨 Remove softmax for EfficientNetForImageClassification 🚨🚨🚨 (#25501) 2023-08-14 17:08:47 +01:00
electra CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
encodec Make more test models smaller (#25005) 2023-07-24 10:08:47 -04:00
encoder_decoder Move TF building to an actual build() method (#23760) 2023-06-06 18:30:51 +01:00
ernie CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
ernie_m CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
esm Fix test_model_parallelism (#25359) 2023-08-08 10:48:45 +02:00
falcon CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
flaubert CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
flava Input data format (#25464) 2023-08-16 17:45:02 +01:00
fnet CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
focalnet Update tiny models and pipeline tests (#23446) 2023-05-18 17:29:04 +02:00
fsmt update_pip_test_mapping (#22606) 2023-04-06 17:56:06 +02:00
funnel Big TF test cleanup (#24282) 2023-06-16 15:40:49 +01:00
git CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
glpn Input data format (#25464) 2023-08-16 17:45:02 +01:00
gpt_bigcode CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
gpt_neo CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
gpt_neox CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
gpt_neox_japanese CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
gpt_sw3 Add gpt-sw3 model to transformers (#20209) 2022-12-12 13:12:13 -05:00
gpt2 CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
gptj CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
gptsan_japanese CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
graphormer Make more test models smaller (#25005) 2023-07-24 10:08:47 -04:00
groupvit CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
herbert Update quality tooling for formatting (#21480) 2023-02-06 18:10:56 -05:00
hubert CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
ibert CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
idefics new model: IDEFICS via HuggingFaceM4 (#24796) 2023-08-18 14:12:28 -07:00
imagegpt Input data format (#25464) 2023-08-16 17:45:02 +01:00
informer Revert "Fix gradient checkpointing + fp16 autocast for most models" (#24420) 2023-06-22 16:11:27 +02:00
instructblip Update InstructBLIP & Align values after rescale update (#25209) 2023-08-03 11:01:10 +01:00
jukebox Replaces calls to .cuda with .to(torch_device) in tests (#25571) 2023-08-18 12:40:40 +02:00
layoutlm CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
layoutlmv2 [split_special_tokens] Add support for split_special_tokens argument to encode (#25081) 2023-08-18 13:26:27 +02:00
layoutlmv3 [split_special_tokens] Add support for split_special_tokens argument to encode (#25081) 2023-08-18 13:26:27 +02:00
layoutxlm [split_special_tokens] Add support for split_special_tokens argument to encode (#25081) 2023-08-18 13:26:27 +02:00
led Speed up TF tests by reducing hidden layer counts (#24595) 2023-06-30 16:30:33 +01:00
levit Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
lilt Revert "Fix gradient checkpointing + fp16 autocast for most models" (#24420) 2023-06-22 16:11:27 +02:00
llama 🚨🚨🚨 [SPM] Finish fix spm models 🚨🚨🚨 (#25224) 2023-08-17 17:08:05 +02:00
longformer Fix more offload edge cases (#25342) 2023-08-07 17:45:41 +02:00
longt5 CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
luke CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
lxmert Big TF test cleanup (#24282) 2023-06-16 15:40:49 +01:00
m2m_100 update_pip_test_mapping (#22606) 2023-04-06 17:56:06 +02:00
marian Marian: post-hack-fix correction (#25459) 2023-08-16 11:49:29 +01:00
markuplm [split_special_tokens] Add support for split_special_tokens argument to encode (#25081) 2023-08-18 13:26:27 +02:00
mask2former Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
maskformer Fix MaskFormerModelIntegrationTest OOM (#25544) 2023-08-16 18:11:24 +02:00
mbart CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
mbart50 Replace as_target context managers by direct calls (#18325) 2022-07-29 08:09:09 -04:00
mega CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
megatron_bert CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
megatron_gpt2 Move test model folders (#17034) 2022-05-03 14:42:02 +02:00
mgp_str Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
mluke Black preview (#17217) 2022-05-12 16:25:55 -04:00
mobilebert CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
mobilenet_v1 Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
mobilenet_v2 Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
mobilevit Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
mobilevitv2 Make more test models smaller (#25005) 2023-07-24 10:08:47 -04:00
mpnet CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
mpt Fix MPT CI (#25548) 2023-08-17 09:06:26 +02:00
mra CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
mt5 Better TF docstring types (#23477) 2023-05-24 13:52:52 +01:00
musicgen [MusicGen] Fix integration tests (#25169) 2023-07-28 18:50:15 +01:00
mvp CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
nat Update old existing feature extractor references (#24552) 2023-06-29 10:17:36 +01:00
nezha CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
nllb 🚨🚨🚨 [NLLB Tokenizer] Fix the prefix tokens 🚨🚨🚨 (#22313) 2023-04-04 14:53:06 +02:00
nllb_moe [NllbMoe] Update code to properly support loss computation (#25429) 2023-08-17 17:21:56 +02:00
nystromformer CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
oneformer Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
openai CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
opt Replaces calls to .cuda with .to(torch_device) in tests (#25571) 2023-08-18 12:40:40 +02:00
owlvit Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
pegasus CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
pegasus_x CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
perceiver Fix last models for common tests that are too big. (#25058) 2023-07-25 07:56:04 -04:00
phobert Update quality tooling for formatting (#21480) 2023-02-06 18:10:56 -05:00
pix2struct Input data format (#25464) 2023-08-16 17:45:02 +01:00
plbart CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
poolformer Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
prophetnet CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
pvt Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
qdqbert CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
rag Big TF test cleanup (#24282) 2023-06-16 15:40:49 +01:00
realm CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
reformer Generate: skip left-padding tests on old models (#23437) 2023-05-18 11:04:51 +01:00
regnet Update old existing feature extractor references (#24552) 2023-06-29 10:17:36 +01:00
rembert CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
resnet Update old existing feature extractor references (#24552) 2023-06-29 10:17:36 +01:00
roberta Fix test_model_parallelism (#25359) 2023-08-08 10:48:45 +02:00
roberta_prelayernorm Fix test_model_parallelism (#25359) 2023-08-08 10:48:45 +02:00
roc_bert CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
roformer CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
rwkv CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
sam Revert "Fix gradient checkpointing + fp16 autocast for most models" (#24420) 2023-06-22 16:11:27 +02:00
segformer Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
sew CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
sew_d CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
speech_encoder_decoder Update quality tooling for formatting (#21480) 2023-02-06 18:10:56 -05:00
speech_to_text Update some torchscript tests after #24505 (#24566) 2023-06-29 16:05:24 +02:00
speech_to_text_2 CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
speecht5 add generate method to SpeechT5ForTextToSpeech (#25233) 2023-08-03 14:12:07 +01:00
splinter CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
squeezebert CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
swiftformer Fix last models for common tests that are too big. (#25058) 2023-07-25 07:56:04 -04:00
swin Update old existing feature extractor references (#24552) 2023-06-29 10:17:36 +01:00
swin2sr Input data format (#25464) 2023-08-16 17:45:02 +01:00
swinv2 Update old existing feature extractor references (#24552) 2023-06-29 10:17:36 +01:00
switch_transformers Switch Transformers: remove overwritten beam sample test (#25458) 2023-08-11 13:16:01 +01:00
t5 🚨🚨🚨 [SPM] Finish fix spm models 🚨🚨🚨 (#25224) 2023-08-17 17:08:05 +02:00
table_transformer Fix last models for common tests that are too big. (#25058) 2023-07-25 07:56:04 -04:00
tapas CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
time_series_transformer Revert "Fix gradient checkpointing + fp16 autocast for most models" (#24420) 2023-06-22 16:11:27 +02:00
timesformer CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
timm_backbone Fix last models for common tests that are too big. (#25058) 2023-07-25 07:56:04 -04:00
transfo_xl CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
trocr CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
tvlt Input data format (#25464) 2023-08-16 17:45:02 +01:00
umt5 CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
unispeech CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
unispeech_sat CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
upernet Fix last models for common tests that are too big. (#25058) 2023-07-25 07:56:04 -04:00
videomae Input data format (#25464) 2023-08-16 17:45:02 +01:00
vilt Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
vision_encoder_decoder Update old existing feature extractor references (#24552) 2023-06-29 10:17:36 +01:00
vision_text_dual_encoder Fix VisionTextDualEncoderIntegrationTest (#24661) 2023-07-05 13:44:30 +02:00
visual_bert CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
vit Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
vit_hybrid fix vit hybrid test (#25543) 2023-08-16 17:02:57 +02:00
vit_mae CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
vit_msn CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
vivit Input data format (#25464) 2023-08-16 17:45:02 +01:00
wav2vec2 Mark flaky tests (#25463) 2023-08-11 15:26:45 +01:00
wav2vec2_conformer CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
wav2vec2_phoneme Move test model folders (#17034) 2022-05-03 14:42:02 +02:00
wav2vec2_with_lm Fix test_word_time_stamp_integration for Wav2Vec2ProcessorWithLMTest (#22800) 2023-04-17 12:41:55 +02:00
wavlm CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
whisper Skip test_beam_search_xla_generate_simple for T5 (#25566) 2023-08-17 15:30:46 +02:00
x_clip CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
xglm Replaces calls to .cuda with .to(torch_device) in tests (#25571) 2023-08-18 12:40:40 +02:00
xlm CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
xlm_prophetnet Update expected values in XLMProphetNetModelIntegrationTest (#21957) 2023-03-06 09:15:44 +01:00
xlm_roberta Better TF docstring types (#23477) 2023-05-24 13:52:52 +01:00
xlm_roberta_xl CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
xlnet Skip test_contrastive_generate for TFXLNet (#25574) 2023-08-17 18:56:34 +02:00
xmod CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
yolos Refactor image processor testers (#25450) 2023-08-11 11:30:18 +01:00
yoso CI with num_hidden_layers=2 🚀🚀🚀 (#25266) 2023-08-02 20:22:36 +02:00
__init__.py Move test model folders (#17034) 2022-05-03 14:42:02 +02:00