transformers/.github
Sourab Mangrulkar a73b1d59a3
accelerate deepspeed and gradient accumulation integrate (#23236)
* mixed precision support via accelerate

* fix issues

* fix for the sharded ddp case

* fix flax and tf failing tests

* `refactor the place to create `Accelerator` object

* move ddp prep to accelerate

* fix 😅

* resolving comments

* move fsdp handling to accelerate

* fixex

* fix saving

* shift torch dynamo handling to accelerate

* shift deepspeed integration and save & load utils to accelerate

* fix accelerate launcher support

* oops

* fix 🐛

* save ckpt fix

* Trigger CI

* nasty 🐛 😅

* as deepspeed needs grad_acc fixes, transfer grad_acc to accelerate

* make tests happy

* quality 

* loss tracked needs to account for grad_acc

* fixing the deepspeed tests

* quality 

* 😅😅😅

* tests 😡

* quality 

* Trigger CI

* resolve comments and fix the issue with the previous merge from branch

* Trigger CI

* accelerate took over deepspeed integration

---------

Co-authored-by: Stas Bekman <stas@stason.org>
2023-05-31 15:16:22 +05:30
..
conda Fix the Conda package build (#16737) 2022-06-29 06:03:16 -04:00
ISSUE_TEMPLATE accelerate deepspeed and gradient accumulation integrate (#23236) 2023-05-31 15:16:22 +05:30
workflows Update workflow files (#23658) 2023-05-22 21:26:51 +02:00
PULL_REQUEST_TEMPLATE.md accelerate deepspeed and gradient accumulation integrate (#23236) 2023-05-31 15:16:22 +05:30