transformers/tests/generation
Matthew Hoffman 70b07d97cf
Default synced_gpus to True when using FullyShardedDataParallel (#33483)
* Default synced_gpus to True when using FullyShardedDataParallel

Fixes #30228

Related:

* https://github.com/pytorch/pytorch/issues/100069
* https://github.com/pytorch/pytorch/issues/123962

Similar to DeepSpeed ZeRO Stage 3, when using FSDP with multiple GPUs and differently sized data per rank, the ranks reach different synchronization points at the same time, leading to deadlock

To avoid this, we can automatically set synced_gpus to True if we detect that a PreTrainedModel is being managed by FSDP using _is_fsdp_managed_module, which was added in 2.0.0 for torch.compile: https://github.com/pytorch/pytorch/blob/v2.0.0/torch/distributed/fsdp/_dynamo_utils.py

* Remove test file

* ruff formatting

* ruff format

* Update copyright year

Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>

* Add test for FSDP-wrapped model generation

Before #33483, these tests would have hung for 10 minutes before crashing due to a timeout error

* Ruff format

* Move argparse import

* Remove barrier

I think this might cause more problems if one of the workers was killed

* Move import into function to decrease load time

https://github.com/huggingface/transformers/pull/33483#discussion_r1787972735

* Add test for accelerate and Trainer

https://github.com/huggingface/transformers/pull/33483#discussion_r1790309675

* Refactor imports

* Ruff format

* Use nullcontext

---------

Co-authored-by: Arthur <48595927+ArthurZucker@users.noreply.github.com>
2024-10-10 14:09:04 -04:00
..
__init__.py [Test refactor 1/5] Per-folder tests reorganization (#15725) 2022-02-23 15:46:28 -05:00
test_beam_constraints.py Generate: move generation_*.py src files into generation/*.py (#20096) 2022-11-09 15:34:08 +00:00
test_beam_search.py Time to Say Goodbye, torch 1.7 and 1.8 (#22291) 2023-03-21 19:22:01 +01:00
test_configuration_utils.py Fix some missing tests in circleci (#33559) 2024-09-20 20:58:51 +02:00
test_flax_logits_process.py Adding FlaxNoRepeatNGramLogitsProcessor (#29677) 2024-04-02 11:39:33 +02:00
test_flax_utils.py Add support for beam search's num_return_sequencs flag in flax (#23082) 2023-05-03 10:50:34 -04:00
test_framework_agnostic.py Generation: fix handling of special tokens (#31254) 2024-06-06 15:21:32 +05:00
test_fsdp.py Default synced_gpus to True when using FullyShardedDataParallel (#33483) 2024-10-10 14:09:04 -04:00
test_logits_process.py Pass device in Logits Processor's init (#29804) 2024-06-04 10:19:19 +05:00
test_stopping_criteria.py Dynamic number of speculative tokens in order to accelerate speculative decoding (#33258) 2024-09-11 14:22:28 +02:00
test_streamers.py Update all references to canonical models (#29001) 2024-02-16 08:16:58 +01:00
test_tf_logits_process.py fix: multilingual midel convert to tflite get wrong token (#32079) 2024-08-27 11:44:09 +02:00
test_tf_utils.py Revert workaround for TF safetensors loading (#30128) 2024-04-09 11:04:18 +01:00
test_utils.py Universal Assisted Generation: Assisted generation with any assistant model (by Intel Labs) (#33383) 2024-10-10 14:41:53 +02:00