transformers/tests
Andreas Madsen b4b613b102
Implement Roberta PreLayerNorm (#20305)
* Copy RoBERTa

* formatting

* implement RoBERTa with prelayer normalization

* update test expectations

* add documentation

* add convertion script for DinkyTrain weights

* update checkpoint repo

Unfortunately the original checkpoints assumes a hacked roberta model

* add to RoBERTa-PreLayerNorm docs to toc

* run utils/check_copies.py

* lint files

* remove unused import

* fix check_repo reporting wrongly a test is missing

* fix import error, caused by rebase

* run make fix-copies

* add RobertaPreLayerNormConfig to ROBERTA_EMBEDDING_ADJUSMENT_CONFIGS

* Fix documentation <Facebook> -> Facebook

Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>

* fixup: Fix documentation <Facebook> -> Facebook

Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>

* Add missing Flax header

Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>

* expected_slice -> EXPECTED_SLICE

Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>

* update copies after rebase

* add missing copied from statements

* make fix-copies

* make prelayernorm explicit in code

* fix checkpoint path for the original implementation

* add flax integration tests

* improve docs

* update utils/documentation_tests.txt

* lint files

* Remove Copyright notice

Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>

* make fix-copies

* Remove EXPECTED_SLICE calculation comments

Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>

Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2022-12-19 09:30:17 +01:00
..
benchmark
deepspeed Fix tapas scatter (#20149) 2022-11-14 01:04:26 -05:00
extended Update self-push workflow (#17177) 2022-05-13 16:28:00 +02:00
fixtures add a warning in SpmConverter for sentencepiece's model using the byte fallback feature (#16629) 2022-04-11 11:06:10 +02:00
generation Generate: use GenerationConfig as the basis for .generate() parametrization (#20388) 2022-12-15 18:27:20 +00:00
mixed_int8 Add keep_in_fp32_modules support (#20683) 2022-12-13 11:59:57 +01:00
models Implement Roberta PreLayerNorm (#20305) 2022-12-19 09:30:17 +01:00
onnx Add RemBERT ONNX config (#20520) 2022-12-05 11:39:09 -05:00
optimization
pipelines Implement Roberta PreLayerNorm (#20305) 2022-12-19 09:30:17 +01:00
repo_utils Repo utils test (#19696) 2022-10-18 13:47:36 -04:00
sagemaker transformers-cli login => huggingface-cli login (#18490) 2022-08-06 09:42:55 +02:00
tokenization fix train_new_from_iterator in the case of byte-level tokenizers (#17549) 2022-06-08 15:30:41 +02:00
trainer Repurpose torchdynamo training args towards torch._dynamo (#20498) 2022-11-30 11:10:45 -05:00
utils Fix TF nightly tests (#20507) 2022-11-30 14:47:54 +00:00
__init__.py
test_configuration_common.py Add WhisperModel to transformers (#19166) 2022-10-05 22:28:31 +02:00
test_feature_extraction_common.py [Vision] .to function for ImageProcessors (#20536) 2022-12-05 19:10:54 +01:00
test_image_transforms.py Move convert_to_rgb to image_transforms module (#20784) 2022-12-15 18:47:04 +00:00
test_modeling_common.py [Tests] Improve test_attention_outputs (#20701) 2022-12-14 14:41:40 +01:00
test_modeling_flax_common.py Allow flax subfolder (#19902) 2022-10-26 18:33:23 +02:00
test_modeling_tf_common.py [Tests] Improve test_attention_outputs (#20701) 2022-12-14 14:41:40 +01:00
test_sequence_feature_extraction_common.py Some tests misusing assertTrue for comparisons fix (#16771) 2022-04-19 14:44:08 +02:00
test_tokenization_common.py 🚨 🚨 🚨 Fix Issue 15003: SentencePiece Tokenizers Not Adding Special Tokens in convert_tokens_to_string (#15775) 2022-11-02 15:45:38 -04:00