mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-12 17:20:03 +06:00
![]() Docs: fix GaLore optimizer example Fix incorrect usage of GaLore optimizer in Transformers trainer code example. The GaLore optimizer uses low-rank gradient updates to reduce memory usage. GaLore is quite popular and is implemented by the authors in [https://github.com/jiaweizzhao/GaLore](https://github.com/jiaweizzhao/GaLore). A few months ago GaLore was added to the HuggingFace Transformers library in https://github.com/huggingface/transformers/pull/29588. Documentation of the Trainer module includes a few code examples of how to use GaLore. However, the `optim_targe_modules` argument to the `TrainingArguments` function is incorrect, as discussed in https://github.com/huggingface/transformers/pull/29588#issuecomment-2006289512. This pull request fixes this issue. |
||
---|---|---|
.. | ||
de | ||
en | ||
es | ||
fr | ||
hi | ||
it | ||
ja | ||
ko | ||
ms | ||
pt | ||
te | ||
tr | ||
zh | ||
_config.py |