transformers/docs/source
RhuiDih 9cf4f2aa9a
Enhancing SFT Training Efficiency Using Packing and FlashAttention2 with Position IDs (#31629)
* add DataCollatorBatchFlattening

* Update data_collator.py

* change name

* new FA2 flow if position_ids is provided

* add comments

* minor fix

* minor fix data collator

* add test cases for models

* add test case for data collator

* remove extra code

* formating for ruff check and check_repo.py

* ruff format

ruff format tests src utils

* custom_init_isort.py
2024-07-23 15:56:41 +02:00
..
de Skip tests properly (#31308) 2024-06-26 21:59:08 +01:00
en Enhancing SFT Training Efficiency Using Packing and FlashAttention2 with Position IDs (#31629) 2024-07-23 15:56:41 +02:00
es [docs] Spanish translation of tokenizer_summary.md (#31154) 2024-06-03 16:52:23 -07:00
fr Add French version of run scripts tutorial (#31483) 2024-06-28 18:02:30 +02:00
hi More fixes for doctest (#30265) 2024-04-16 11:58:55 +02:00
it Docs / Quantization: Replace all occurences of load_in_8bit with bnb config (#31136) 2024-05-30 16:47:35 +02:00
ja Update CometCallback to allow reusing of the running experiment (#31366) 2024-07-05 08:13:46 +02:00
ko Update ko/_toctree.yml and remove custom_tools.md to reflect latest changes (#31969) 2024-07-22 08:27:13 -07:00
ms Remove old TF port docs (#30426) 2024-04-23 16:06:20 +01:00
pt Use HF_HUB_OFFLINE + fix has_file in offline mode (#31016) 2024-05-29 11:55:43 +01:00
te docs: fix broken link (#31370) 2024-06-12 11:33:00 +01:00
tr Translate index.md to Turkish (#27093) 2023-11-08 08:35:20 -05:00
zh Update CometCallback to allow reusing of the running experiment (#31366) 2024-07-05 08:13:46 +02:00
_config.py [#29174] ImportError Fix: Trainer with PyTorch requires accelerate>=0.20.1 Fix (#29888) 2024-04-08 14:21:16 +01:00