mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-30 17:52:35 +06:00
![]() * move deepspeed to `lib_integrations.deepspeed` * more refactor * oops * fix slow tests * Fix docs * fix docs * addess feedback * address feedback * final modifs for PEFT * fixup * ok now * trigger CI * trigger CI again * Update docs/source/en/main_classes/deepspeed.md Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * import from `integrations` * address feedback * revert removal of `deepspeed` module * revert removal of `deepspeed` module * fix conflicts * ooops * oops * add deprecation warning * place it on the top * put `FutureWarning` * fix conflicts with not_doctested.txt * add back `bitsandbytes` module with a depr warning * fix * fix * fixup * oops * fix doctests --------- Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com> |
||
---|---|---|
.. | ||
adversarial | ||
bert-loses-patience | ||
bertabs | ||
bertology | ||
codeparrot | ||
decision_transformer | ||
deebert | ||
distillation | ||
fsner | ||
information-gain-filtration | ||
jax-projects | ||
layoutlmv3 | ||
longform-qa | ||
luke | ||
lxmert | ||
mlm_wwm | ||
mm-imdb | ||
movement-pruning | ||
onnx/summarization | ||
performer | ||
pplm | ||
quantization-qdqbert | ||
rag | ||
rag-end2end-retriever | ||
robust-speech-event | ||
self-training-text-classification | ||
seq2seq-distillation | ||
tapex | ||
visual_bert | ||
vqgan-clip | ||
wav2vec2 | ||
xtreme-s | ||
zero-shot-distillation | ||
README.md |
Research projects
This folder contains various research projects using 🤗 Transformers. They are not maintained and require a specific version of 🤗 Transformers that is indicated in the requirements file of each folder. Updating them to the most recent version of the library will require some work.
To use any of them, just run the command
pip install -r requirements.txt
inside the folder of your choice.
If you need help with any of those, contact the author(s), indicated at the top of the README
of each folder.