transformers/docs/source
bytebarde be3fd8a262
[Flash Attention 2] Add flash attention 2 for GPT-J (#28295)
* initial implementation of flash attention for gptj

* modify flash attention and overwrite test_flash_attn_2_generate_padding_right

* update flash attention support list

* remove the copy line in the `CodeGenBlock`

* address copy mechanism

* Update src/transformers/models/gptj/modeling_gptj.py

Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>

* Add GPTJ attention classes

* add expected outputs in the gptj test

* Ensure repo consistency with 'make fix-copies'

---------

Co-authored-by: Younes Belkada <49240599+younesbelkada@users.noreply.github.com>
Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
2024-03-13 08:43:00 +01:00
..
de Make torch xla available on GPU (#29334) 2024-03-11 14:07:16 +00:00
en [Flash Attention 2] Add flash attention 2 for GPT-J (#28295) 2024-03-13 08:43:00 +01:00
es [Docs] Spanish Translation -Torchscript md & Trainer md (#29310) 2024-03-04 13:57:51 -08:00
fr Update all references to canonical models (#29001) 2024-02-16 08:16:58 +01:00
hi Update all references to canonical models (#29001) 2024-02-16 08:16:58 +01:00
it Update all references to canonical models (#29001) 2024-02-16 08:16:58 +01:00
ja Make torch xla available on GPU (#29334) 2024-03-11 14:07:16 +00:00
ko Make torch xla available on GPU (#29334) 2024-03-11 14:07:16 +00:00
ms [Docs] Add missing language options and fix broken links (#28852) 2024-02-06 12:01:01 -08:00
pt Update all references to canonical models (#29001) 2024-02-16 08:16:58 +01:00
te Update all references to canonical models (#29001) 2024-02-16 08:16:58 +01:00
tr Translate index.md to Turkish (#27093) 2023-11-08 08:35:20 -05:00
zh v4.39 deprecations 🧼 (#29492) 2024-03-07 10:44:43 +00:00
_config.py [Styling] stylify using ruff (#27144) 2023-11-16 17:43:19 +01:00