transformers/docs/source
Cesare Campagnano d9050dc768
[LED] fix global_attention_mask not being passed for generation and docs clarification about grad checkpointing (#17112)
* [LED] fixed global_attention_mask not passed for generation + docs clarification for gradient checkpointing

* LED docs clarification

Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>

* [LED] gradient_checkpointing=True should be passed to TrainingArguments

Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>

* [LED] docs: remove wrong word

Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>

* [LED] docs fix typo

Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>

Co-authored-by: Patrick von Platen <patrick.v.platen@gmail.com>
2022-05-17 23:44:37 +02:00
..
en [LED] fix global_attention_mask not being passed for generation and docs clarification about grad checkpointing (#17112) 2022-05-17 23:44:37 +02:00
es Guide to create custom models in Spanish (#17158) 2022-05-13 16:19:29 -04:00
pt [ fast_tokenizers.mdx ] - Added translation to portuguese to tutorial (#17076) 2022-05-13 16:18:14 -04:00
_config.py Use doc builder styler (#16412) 2022-03-28 07:45:18 -04:00