mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-03 12:50:06 +06:00
🚨🚨🚨 Update min version of accelerate to 0.26.0 (#32627)
* Update min version of accelerate to 0.26.0 * dev-ci * update min version in import * remove useless check * dev-ci * style * dev-ci * dev-ci
This commit is contained in:
parent
13e645bb40
commit
fd06ad5438
2
setup.py
2
setup.py
@ -96,7 +96,7 @@ if stale_egg_info.exists():
|
||||
# 2. once modified, run: `make deps_table_update` to update src/transformers/dependency_versions_table.py
|
||||
_deps = [
|
||||
"Pillow>=10.0.1,<=15.0",
|
||||
"accelerate>=0.21.0",
|
||||
"accelerate>=0.26.0",
|
||||
"av==9.2.0", # Latest version of PyAV (10.0.0) has issues with audio stream.
|
||||
"beautifulsoup4",
|
||||
"codecarbon==1.2.0",
|
||||
|
@ -3,7 +3,7 @@
|
||||
# 2. run `make deps_table_update``
|
||||
deps = {
|
||||
"Pillow": "Pillow>=10.0.1,<=15.0",
|
||||
"accelerate": "accelerate>=0.21.0",
|
||||
"accelerate": "accelerate>=0.26.0",
|
||||
"av": "av==9.2.0",
|
||||
"beautifulsoup4": "beautifulsoup4",
|
||||
"codecarbon": "codecarbon==1.2.0",
|
||||
|
@ -4761,16 +4761,15 @@ class Trainer:
|
||||
fsdp_plugin.limit_all_gathers = self.args.fsdp_config.get(
|
||||
"limit_all_gathers", fsdp_plugin.limit_all_gathers
|
||||
)
|
||||
if is_accelerate_available("0.23.0"):
|
||||
fsdp_plugin.activation_checkpointing = self.args.fsdp_config.get(
|
||||
"activation_checkpointing", fsdp_plugin.activation_checkpointing
|
||||
fsdp_plugin.activation_checkpointing = self.args.fsdp_config.get(
|
||||
"activation_checkpointing", fsdp_plugin.activation_checkpointing
|
||||
)
|
||||
if fsdp_plugin.activation_checkpointing and self.args.gradient_checkpointing:
|
||||
raise ValueError(
|
||||
"The activation_checkpointing in FSDP config and the gradient_checkpointing in training arg "
|
||||
"can't be set to True simultaneously. Please use FSDP's activation_checkpointing logic "
|
||||
"when using FSDP."
|
||||
)
|
||||
if fsdp_plugin.activation_checkpointing and self.args.gradient_checkpointing:
|
||||
raise ValueError(
|
||||
"The activation_checkpointing in FSDP config and the gradient_checkpointing in training arg "
|
||||
"can't be set to True simultaneously. Please use FSDP's activation_checkpointing logic "
|
||||
"when using FSDP."
|
||||
)
|
||||
|
||||
if self.is_deepspeed_enabled and getattr(self.args, "hf_deepspeed_config", None) is None:
|
||||
self.propagate_args_to_deepspeed()
|
||||
|
@ -1915,10 +1915,8 @@ class TrainingArguments:
|
||||
for fsdp_option in self.fsdp:
|
||||
if fsdp_option.upper() in FSDP_SHARDING_STRATEGY:
|
||||
# set environment variable for FSDP sharding strategy
|
||||
os.environ[f"{prefix}SHARDING_STRATEGY"] = (
|
||||
str(FSDP_SHARDING_STRATEGY.index(fsdp_option.upper()) + 1)
|
||||
if is_accelerate_available("0.26.0")
|
||||
else fsdp_option.upper()
|
||||
os.environ[f"{prefix}SHARDING_STRATEGY"] = str(
|
||||
FSDP_SHARDING_STRATEGY.index(fsdp_option.upper()) + 1
|
||||
)
|
||||
elif fsdp_option == FSDPOption.OFFLOAD:
|
||||
os.environ[f"{prefix}OFFLOAD_PARAMS"] = "true"
|
||||
|
@ -87,7 +87,7 @@ FORCE_TF_AVAILABLE = os.environ.get("FORCE_TF_AVAILABLE", "AUTO").upper()
|
||||
# This is the version of torch required to run torch.fx features and torch.onnx with dictionary inputs.
|
||||
TORCH_FX_REQUIRED_VERSION = version.parse("1.10")
|
||||
|
||||
ACCELERATE_MIN_VERSION = "0.21.0"
|
||||
ACCELERATE_MIN_VERSION = "0.26.0"
|
||||
FSDP_MIN_VERSION = "1.12.0"
|
||||
XLA_FSDPV2_MIN_VERSION = "2.2.0"
|
||||
|
||||
|
@ -196,11 +196,7 @@ class TrainerIntegrationFSDP(TestCasePlus, TrainerIntegrationCommon):
|
||||
self.assertEqual(trainer.args.fsdp[0], sharding_strategy)
|
||||
self.assertEqual(trainer.args.fsdp[1], FSDPOption.OFFLOAD)
|
||||
self.assertEqual(trainer.args.fsdp[2], FSDPOption.AUTO_WRAP)
|
||||
fsdp_sharding_strategy = (
|
||||
str(FSDP_SHARDING_STRATEGY.index(sharding_strategy.upper()) + 1)
|
||||
if is_accelerate_available("0.26.0")
|
||||
else sharding_strategy.upper()
|
||||
)
|
||||
fsdp_sharding_strategy = str(FSDP_SHARDING_STRATEGY.index(sharding_strategy.upper()) + 1)
|
||||
self.assertEqual(os.environ[f"{prefix}SHARDING_STRATEGY"], fsdp_sharding_strategy)
|
||||
self.assertEqual(os.environ[f"{prefix}OFFLOAD_PARAMS"], "true")
|
||||
self.assertEqual(os.environ[f"{prefix}AUTO_WRAP_POLICY"], "TRANSFORMER_BASED_WRAP")
|
||||
|
Loading…
Reference in New Issue
Block a user