..
albert
auto
Prepare transformers for v0.8.0 huggingface-hub release ( #17716 )
2022-06-21 11:51:18 -04:00
bart
TF: BART compatible with XLA generation ( #17479 )
2022-06-20 11:07:46 +01:00
barthez
bartpho
beit
bert
Black preview ( #17217 )
2022-05-12 16:25:55 -04:00
bert_generation
Black preview ( #17217 )
2022-05-12 16:25:55 -04:00
bert_japanese
Black preview ( #17217 )
2022-05-12 16:25:55 -04:00
bertweet
big_bird
Use 5e-5 For BigBird PT/Flax equivalence tests ( #17780 )
2022-06-21 17:55:26 +02:00
bigbird_pegasus
Black preview ( #17217 )
2022-05-12 16:25:55 -04:00
blenderbot
fix train_new_from_iterator
in the case of byte-level tokenizers ( #17549 )
2022-06-08 15:30:41 +02:00
blenderbot_small
Fx support for multiple model architectures ( #17393 )
2022-05-31 10:02:55 +02:00
bloom
fix tolerance for a bloom slow test ( #17634 )
2022-06-14 18:14:12 +02:00
bort
byt5
camembert
canine
Black preview ( #17217 )
2022-05-12 16:25:55 -04:00
clip
Fx support for multiple model architectures ( #17393 )
2022-05-31 10:02:55 +02:00
convbert
convnext
has_attentions - consistent test skipping logic and tf tests ( #17495 )
2022-06-09 09:50:03 +02:00
cpm
ctrl
Fix CTRL tests ( #17508 )
2022-06-01 16:27:23 +02:00
cvt
has_attentions - consistent test skipping logic and tf tests ( #17495 )
2022-06-09 09:50:03 +02:00
data2vec
[Data2Vec] Speed up test ( #17660 )
2022-06-10 18:48:58 +02:00
deberta
fix train_new_from_iterator
in the case of byte-level tokenizers ( #17549 )
2022-06-08 15:30:41 +02:00
deberta_v2
Fx support for Deberta-v[1-2], Hubert and LXMERT ( #17539 )
2022-06-07 18:05:20 +02:00
decision_transformer
deit
detr
distilbert
dit
dpr
dpt
electra
encoder_decoder
flaubert
flava
has_attentions - consistent test skipping logic and tf tests ( #17495 )
2022-06-09 09:50:03 +02:00
fnet
Black preview ( #17217 )
2022-05-12 16:25:55 -04:00
fsmt
Not use -1e4 as attn mask ( #17306 )
2022-06-20 16:16:16 +02:00
funnel
glpn
gpt_neo
fix train_new_from_iterator
in the case of byte-level tokenizers ( #17549 )
2022-06-08 15:30:41 +02:00
gpt_neox
Fix cache for GPT-Neo-X ( #17764 )
2022-06-20 08:43:36 -04:00
gpt2
TF: BART compatible with XLA generation ( #17479 )
2022-06-20 11:07:46 +01:00
gptj
fix train_new_from_iterator
in the case of byte-level tokenizers ( #17549 )
2022-06-08 15:30:41 +02:00
herbert
hubert
Fx support for Deberta-v[1-2], Hubert and LXMERT ( #17539 )
2022-06-07 18:05:20 +02:00
ibert
fix train_new_from_iterator
in the case of byte-level tokenizers ( #17549 )
2022-06-08 15:30:41 +02:00
imagegpt
Enabling imageGPT
auto feature extractor. ( #16871 )
2022-05-24 12:30:46 +02:00
layoutlm
Fx support for multiple model architectures ( #17393 )
2022-05-31 10:02:55 +02:00
layoutlmv2
Add LayoutLMv3 ( #17060 )
2022-05-24 09:53:45 +02:00
layoutlmv3
Add LayoutLMv3 ( #17060 )
2022-05-24 09:53:45 +02:00
layoutxlm
Fix LayoutXLMProcessorTest ( #17506 )
2022-06-01 16:26:37 +02:00
led
fix train_new_from_iterator
in the case of byte-level tokenizers ( #17549 )
2022-06-08 15:30:41 +02:00
levit
Add skip logic for attentions test - Levit ( #17633 )
2022-06-10 12:46:30 +02:00
longformer
fix train_new_from_iterator
in the case of byte-level tokenizers ( #17549 )
2022-06-08 15:30:41 +02:00
longt5
[LongT5] disable model parallel test ( #17702 )
2022-06-14 17:27:39 +02:00
luke
Debug LukeForMaskedLM ( #17499 )
2022-06-01 10:03:06 -04:00
lxmert
Fx support for Deberta-v[1-2], Hubert and LXMERT ( #17539 )
2022-06-07 18:05:20 +02:00
m2m_100
Fx support for multiple model architectures ( #17393 )
2022-05-31 10:02:55 +02:00
marian
Fx support for multiple model architectures ( #17393 )
2022-05-31 10:02:55 +02:00
maskformer
mbart
Fx support for multiple model architectures ( #17393 )
2022-05-31 10:02:55 +02:00
mbart50
Black preview ( #17217 )
2022-05-12 16:25:55 -04:00
mctct
M-CTC-T Model ( #16402 )
2022-06-08 00:33:07 +02:00
megatron_bert
megatron_gpt2
mluke
Black preview ( #17217 )
2022-05-12 16:25:55 -04:00
mobilebert
Black preview ( #17217 )
2022-05-12 16:25:55 -04:00
mpnet
mt5
nystromformer
openai
opt
Add final_layer_norm to OPT model ( #17785 )
2022-06-21 20:26:36 +02:00
pegasus
Fx support for multiple model architectures ( #17393 )
2022-05-31 10:02:55 +02:00
perceiver
Black preview ( #17217 )
2022-05-12 16:25:55 -04:00
phobert
plbart
Fx support for multiple model architectures ( #17393 )
2022-05-31 10:02:55 +02:00
poolformer
has_attentions - consistent test skipping logic and tf tests ( #17495 )
2022-06-09 09:50:03 +02:00
prophetnet
Black preview ( #17217 )
2022-05-12 16:25:55 -04:00
qdqbert
rag
Avoid GPU OOM for a TF Rag test ( #17638 )
2022-06-10 18:50:29 +02:00
realm
Black preview ( #17217 )
2022-05-12 16:25:55 -04:00
reformer
Black preview ( #17217 )
2022-05-12 16:25:55 -04:00
regnet
has_attentions - consistent test skipping logic and tf tests ( #17495 )
2022-06-09 09:50:03 +02:00
rembert
resnet
has_attentions - consistent test skipping logic and tf tests ( #17495 )
2022-06-09 09:50:03 +02:00
retribert
fix retribert's test_torch_encode_plus_sent_to_model
( #17231 )
2022-05-17 14:33:13 +02:00
roberta
fix train_new_from_iterator
in the case of byte-level tokenizers ( #17549 )
2022-06-08 15:30:41 +02:00
roformer
segformer
sew
sew_d
speech_encoder_decoder
speech_to_text
Fx support for multiple model architectures ( #17393 )
2022-05-31 10:02:55 +02:00
speech_to_text_2
Fx support for multiple model architectures ( #17393 )
2022-05-31 10:02:55 +02:00
splinter
Add support for pretraining recurring span selection to Splinter ( #17247 )
2022-05-17 23:42:14 +02:00
squeezebert
swin
Fx support for multiple model architectures ( #17393 )
2022-05-31 10:02:55 +02:00
t5
TF: BART compatible with XLA generation ( #17479 )
2022-06-20 11:07:46 +01:00
tapas
Add magic method to our TF models to convert datasets with column inference ( #17160 )
2022-06-06 15:53:49 +01:00
tapex
trajectory_transformer
Add trajectory transformer ( #17141 )
2022-05-17 19:07:43 -04:00
transfo_xl
Add magic method to our TF models to convert datasets with column inference ( #17160 )
2022-06-06 15:53:49 +01:00
trocr
Fx support for multiple model architectures ( #17393 )
2022-05-31 10:02:55 +02:00
unispeech
unispeech_sat
van
has_attentions - consistent test skipping logic and tf tests ( #17495 )
2022-06-09 09:50:03 +02:00
vilt
Black preview ( #17217 )
2022-05-12 16:25:55 -04:00
vision_encoder_decoder
vision_text_dual_encoder
visual_bert
vit
ViT and Swin symbolic tracing with torch.fx ( #17182 )
2022-05-12 10:42:27 +02:00
vit_mae
[ViTMAE] Fix docstrings and variable names ( #17710 )
2022-06-21 15:56:00 +02:00
wav2vec2
Black preview ( #17217 )
2022-05-12 16:25:55 -04:00
wav2vec2_conformer
[Test] Fix W2V-Conformer integration test ( #17303 )
2022-05-17 18:20:36 +02:00
wav2vec2_phoneme
wav2vec2_with_lm
wavlm
xglm
Fx support for multiple model architectures ( #17393 )
2022-05-31 10:02:55 +02:00
xlm
xlm_prophetnet
Black preview ( #17217 )
2022-05-12 16:25:55 -04:00
xlm_roberta
Black preview ( #17217 )
2022-05-12 16:25:55 -04:00
xlm_roberta_xl
xlnet
Fx support for multiple model architectures ( #17393 )
2022-05-31 10:02:55 +02:00
yolos
yoso
fix train_new_from_iterator
in the case of byte-level tokenizers ( #17549 )
2022-06-08 15:30:41 +02:00
__init__.py