transformers/examples/research_projects
Loubna Ben Allal b48ac1a094
Fix CodeParrot training script (#17291)
* average loss over batches and accumulated steps for tracking

* fix layernorm weight decay

* use AdamW from Pytorch instead of Transformers

* add shuffling of sequences inside the batches

* add shuffling of sequences inside the batches

* add logging dir and reformat code

* fix lr tracking

* remove Mistral scaling

* keep Mistral scaling

* reformat code

* fix error

* fix error

* use shuffling function from Pytorch

* remove argument for shuffling batch sequences as it isn't optional

* update package versions and install accelerate from source

* remove unused package

* Update loss average over accumulated steps

Co-authored-by: Leandro von Werra <lvwerra@users.noreply.github.com>

* Update loss average over accumulated steps

Co-authored-by: Leandro von Werra <lvwerra@users.noreply.github.com>

* use one shuffle buffer argument

* compute avg_loss in one line

Co-authored-by: Loubna ben allal <loubnabenallal@gmail.com>
Co-authored-by: Leandro von Werra <lvwerra@users.noreply.github.com>
2022-05-23 12:55:35 +02:00
..
adversarial Black preview (#17217) 2022-05-12 16:25:55 -04:00
bert-loses-patience Black preview (#17217) 2022-05-12 16:25:55 -04:00
bertabs Black preview (#17217) 2022-05-12 16:25:55 -04:00
bertology Black preview (#17217) 2022-05-12 16:25:55 -04:00
codeparrot Fix CodeParrot training script (#17291) 2022-05-23 12:55:35 +02:00
decision_transformer Decision transformer gym (#15845) 2022-03-23 16:18:43 -04:00
deebert Black preview (#17217) 2022-05-12 16:25:55 -04:00
distillation Black preview (#17217) 2022-05-12 16:25:55 -04:00
fsner Black preview (#17217) 2022-05-12 16:25:55 -04:00
information-gain-filtration Fix style 2022-05-18 10:46:40 -04:00
jax-projects fixed bug in run_mlm_flax_stream.py (#17203) 2022-05-16 13:40:27 +02:00
longform-qa Black preview (#17217) 2022-05-12 16:25:55 -04:00
luke Black preview (#17217) 2022-05-12 16:25:55 -04:00
lxmert Black preview (#17217) 2022-05-12 16:25:55 -04:00
mlm_wwm Black preview (#17217) 2022-05-12 16:25:55 -04:00
mm-imdb Black preview (#17217) 2022-05-12 16:25:55 -04:00
movement-pruning Black preview (#17217) 2022-05-12 16:25:55 -04:00
onnx/summarization Black preview (#17217) 2022-05-12 16:25:55 -04:00
performer Black preview (#17217) 2022-05-12 16:25:55 -04:00
pplm Black preview (#17217) 2022-05-12 16:25:55 -04:00
quantization-qdqbert Black preview (#17217) 2022-05-12 16:25:55 -04:00
rag Black preview (#17217) 2022-05-12 16:25:55 -04:00
rag-end2end-retriever Black preview (#17217) 2022-05-12 16:25:55 -04:00
robust-speech-event Black preview (#17217) 2022-05-12 16:25:55 -04:00
self-training-text-classification Black preview (#17217) 2022-05-12 16:25:55 -04:00
seq2seq-distillation Black preview (#17217) 2022-05-12 16:25:55 -04:00
tapex Black preview (#17217) 2022-05-12 16:25:55 -04:00
visual_bert Black preview (#17217) 2022-05-12 16:25:55 -04:00
wav2vec2 Fix bug in Wav2Vec2 pretrain example (#17326) 2022-05-19 22:42:44 +02:00
xtreme-s Black preview (#17217) 2022-05-12 16:25:55 -04:00
zero-shot-distillation Updates the default branch from master to main (#16326) 2022-03-23 03:46:59 -04:00
README.md Reorganize examples (#9010) 2020-12-11 10:07:02 -05:00

Research projects

This folder contains various research projects using 🤗 Transformers. They are not maintained and require a specific version of 🤗 Transformers that is indicated in the requirements file of each folder. Updating them to the most recent version of the library will require some work.

To use any of them, just run the command

pip install -r requirements.txt

inside the folder of your choice.

If you need help with any of those, contact the author(s), indicated at the top of the README of each folder.