transformers/examples/research_projects
Jia LI 4868a830db
Jia multi gpu eval (#16428)
* add simple multi gpu complet

* add human_eval_multi_gpu

* use copy strategy to distribute across gpu, to avoid padding

* add doc string

* update code style

* use task id to arrange output

* truncate input to avoid zero pad

* Stop the copy mechanism

* update style

* restore copies to scale better in distributed mode

* update style

* replace human eval

* Apply suggestions from code review

1. Tokenize all input at the same time
2. use attention_mask to get the input length
3. other small fixes

Co-authored-by: Leandro von Werra <lvwerra@users.noreply.github.com>

* correct typo and update docstring

* update code style

* remove num sample division constraint

* remove max len calculation

* use accelerator.gather once to speed up

* use accelerate set_seed; update accelerate version

* correct gather bug

Co-authored-by: Leandro von Werra <lvwerra@users.noreply.github.com>
2022-04-11 11:24:32 +02:00
..
adversarial Update namespaces inside torch.utils.data to the latest. (#13167) 2021-08-19 14:29:51 +02:00
bert-loses-patience Fix tiny typo (#15884) 2022-03-02 15:37:05 +01:00
bertabs
bertology
codeparrot Jia multi gpu eval (#16428) 2022-04-11 11:24:32 +02:00
decision_transformer Decision transformer gym (#15845) 2022-03-23 16:18:43 -04:00
deebert remove extra white space from log format (#12360) 2021-06-25 13:20:14 -07:00
distillation Fix minor comment typos (#15740) 2022-02-21 12:41:27 +01:00
fsner Updates the default branch from master to main (#16326) 2022-03-23 03:46:59 -04:00
jax-projects [examples] max samples can't be bigger than the len of dataset (#16501) 2022-03-30 12:33:16 -07:00
longform-qa
luke Add (M)Luke model training for Token Classification in the examples (#14880) 2022-01-31 07:58:18 -05:00
lxmert Upgrade black to version ~=22.0 (#15565) 2022-02-09 09:28:57 -05:00
mlm_wwm [urls to hub] Replace outdated model tags with their now-canonical pipeline types (#14617) 2021-12-06 04:35:01 -05:00
mm-imdb Updates the default branch from master to main (#16326) 2022-03-23 03:46:59 -04:00
movement-pruning Updates the default branch from master to main (#16326) 2022-03-23 03:46:59 -04:00
onnx/summarization Upgrade black to version ~=22.0 (#15565) 2022-02-09 09:28:57 -05:00
performer [urls to hub] Replace outdated model tags with their now-canonical pipeline types (#14617) 2021-12-06 04:35:01 -05:00
pplm [research_projects] deal with security alerts (#15594) 2022-02-11 14:31:09 -05:00
quantization-qdqbert [examples] max samples can't be bigger than the len of dataset (#16501) 2022-03-30 12:33:16 -07:00
rag Updates the default branch from master to main (#16326) 2022-03-23 03:46:59 -04:00
rag-end2end-retriever Updates the default branch from master to main (#16326) 2022-03-23 03:46:59 -04:00
robust-speech-event Updates the default branch from master to main (#16326) 2022-03-23 03:46:59 -04:00
seq2seq-distillation Updates the default branch from master to main (#16326) 2022-03-23 03:46:59 -04:00
tapex Add TAPEX (#16473) 2022-04-08 10:57:51 +02:00
visual_bert Updates the default branch from master to main (#16326) 2022-03-23 03:46:59 -04:00
wav2vec2 [examples] max samples can't be bigger than the len of dataset (#16501) 2022-03-30 12:33:16 -07:00
xtreme-s [research] link to the XTREME-S paper (#16519) 2022-03-31 23:26:50 +04:00
zero-shot-distillation Updates the default branch from master to main (#16326) 2022-03-23 03:46:59 -04:00
README.md

Research projects

This folder contains various research projects using 🤗 Transformers. They are not maintained and require a specific version of 🤗 Transformers that is indicated in the requirements file of each folder. Updating them to the most recent version of the library will require some work.

To use any of them, just run the command

pip install -r requirements.txt

inside the folder of your choice.

If you need help with any of those, contact the author(s), indicated at the top of the README of each folder.