mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-19 20:48:22 +06:00
![]() * Add Luke training * Fix true label tags * Fix true label tags * Fix true label tags * Update the data collator for Luke * Some training refactor for Luke * Improve data collator for Luke * Fix import * Fix datasets concatenation * Add the --max_entity_length argument for Luke models * Remove unused code * Fix style issues * Fix style issues * Move the Luke training into a separate folder * Fix style * Fix naming * Fix filtering * Fix filtering * Fix filter * Update some preprocessing * Move luke to research_projects * Checkstyle * Address comments * Fix style |
||
---|---|---|
.. | ||
adversarial | ||
bert-loses-patience | ||
bertabs | ||
bertology | ||
codeparrot | ||
deebert | ||
distillation | ||
fsner | ||
jax-projects | ||
longform-qa | ||
luke | ||
lxmert | ||
mlm_wwm | ||
mm-imdb | ||
movement-pruning | ||
onnx/summarization | ||
performer | ||
pplm | ||
quantization-qdqbert | ||
rag | ||
rag-end2end-retriever | ||
robust-speech-event | ||
seq2seq-distillation | ||
visual_bert | ||
wav2vec2 | ||
zero-shot-distillation | ||
README.md |
Research projects
This folder contains various research projects using 🤗 Transformers. They are not maintained and require a specific version of 🤗 Transformers that is indicated in the requirements file of each folder. Updating them to the most recent version of the library will require some work.
To use any of them, just run the command
pip install -r requirements.txt
inside the folder of your choice.
If you need help with any of those, contact the author(s), indicated at the top of the README
of each folder.