mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-19 20:48:22 +06:00
🚨🚨🚨 Update min version of accelerate to 0.26.0 (#32627)
* Update min version of accelerate to 0.26.0 * dev-ci * update min version in import * remove useless check * dev-ci * style * dev-ci * dev-ci
This commit is contained in:
parent
13e645bb40
commit
fd06ad5438
2
setup.py
2
setup.py
@ -96,7 +96,7 @@ if stale_egg_info.exists():
|
|||||||
# 2. once modified, run: `make deps_table_update` to update src/transformers/dependency_versions_table.py
|
# 2. once modified, run: `make deps_table_update` to update src/transformers/dependency_versions_table.py
|
||||||
_deps = [
|
_deps = [
|
||||||
"Pillow>=10.0.1,<=15.0",
|
"Pillow>=10.0.1,<=15.0",
|
||||||
"accelerate>=0.21.0",
|
"accelerate>=0.26.0",
|
||||||
"av==9.2.0", # Latest version of PyAV (10.0.0) has issues with audio stream.
|
"av==9.2.0", # Latest version of PyAV (10.0.0) has issues with audio stream.
|
||||||
"beautifulsoup4",
|
"beautifulsoup4",
|
||||||
"codecarbon==1.2.0",
|
"codecarbon==1.2.0",
|
||||||
|
@ -3,7 +3,7 @@
|
|||||||
# 2. run `make deps_table_update``
|
# 2. run `make deps_table_update``
|
||||||
deps = {
|
deps = {
|
||||||
"Pillow": "Pillow>=10.0.1,<=15.0",
|
"Pillow": "Pillow>=10.0.1,<=15.0",
|
||||||
"accelerate": "accelerate>=0.21.0",
|
"accelerate": "accelerate>=0.26.0",
|
||||||
"av": "av==9.2.0",
|
"av": "av==9.2.0",
|
||||||
"beautifulsoup4": "beautifulsoup4",
|
"beautifulsoup4": "beautifulsoup4",
|
||||||
"codecarbon": "codecarbon==1.2.0",
|
"codecarbon": "codecarbon==1.2.0",
|
||||||
|
@ -4761,16 +4761,15 @@ class Trainer:
|
|||||||
fsdp_plugin.limit_all_gathers = self.args.fsdp_config.get(
|
fsdp_plugin.limit_all_gathers = self.args.fsdp_config.get(
|
||||||
"limit_all_gathers", fsdp_plugin.limit_all_gathers
|
"limit_all_gathers", fsdp_plugin.limit_all_gathers
|
||||||
)
|
)
|
||||||
if is_accelerate_available("0.23.0"):
|
fsdp_plugin.activation_checkpointing = self.args.fsdp_config.get(
|
||||||
fsdp_plugin.activation_checkpointing = self.args.fsdp_config.get(
|
"activation_checkpointing", fsdp_plugin.activation_checkpointing
|
||||||
"activation_checkpointing", fsdp_plugin.activation_checkpointing
|
)
|
||||||
|
if fsdp_plugin.activation_checkpointing and self.args.gradient_checkpointing:
|
||||||
|
raise ValueError(
|
||||||
|
"The activation_checkpointing in FSDP config and the gradient_checkpointing in training arg "
|
||||||
|
"can't be set to True simultaneously. Please use FSDP's activation_checkpointing logic "
|
||||||
|
"when using FSDP."
|
||||||
)
|
)
|
||||||
if fsdp_plugin.activation_checkpointing and self.args.gradient_checkpointing:
|
|
||||||
raise ValueError(
|
|
||||||
"The activation_checkpointing in FSDP config and the gradient_checkpointing in training arg "
|
|
||||||
"can't be set to True simultaneously. Please use FSDP's activation_checkpointing logic "
|
|
||||||
"when using FSDP."
|
|
||||||
)
|
|
||||||
|
|
||||||
if self.is_deepspeed_enabled and getattr(self.args, "hf_deepspeed_config", None) is None:
|
if self.is_deepspeed_enabled and getattr(self.args, "hf_deepspeed_config", None) is None:
|
||||||
self.propagate_args_to_deepspeed()
|
self.propagate_args_to_deepspeed()
|
||||||
|
@ -1915,10 +1915,8 @@ class TrainingArguments:
|
|||||||
for fsdp_option in self.fsdp:
|
for fsdp_option in self.fsdp:
|
||||||
if fsdp_option.upper() in FSDP_SHARDING_STRATEGY:
|
if fsdp_option.upper() in FSDP_SHARDING_STRATEGY:
|
||||||
# set environment variable for FSDP sharding strategy
|
# set environment variable for FSDP sharding strategy
|
||||||
os.environ[f"{prefix}SHARDING_STRATEGY"] = (
|
os.environ[f"{prefix}SHARDING_STRATEGY"] = str(
|
||||||
str(FSDP_SHARDING_STRATEGY.index(fsdp_option.upper()) + 1)
|
FSDP_SHARDING_STRATEGY.index(fsdp_option.upper()) + 1
|
||||||
if is_accelerate_available("0.26.0")
|
|
||||||
else fsdp_option.upper()
|
|
||||||
)
|
)
|
||||||
elif fsdp_option == FSDPOption.OFFLOAD:
|
elif fsdp_option == FSDPOption.OFFLOAD:
|
||||||
os.environ[f"{prefix}OFFLOAD_PARAMS"] = "true"
|
os.environ[f"{prefix}OFFLOAD_PARAMS"] = "true"
|
||||||
|
@ -87,7 +87,7 @@ FORCE_TF_AVAILABLE = os.environ.get("FORCE_TF_AVAILABLE", "AUTO").upper()
|
|||||||
# This is the version of torch required to run torch.fx features and torch.onnx with dictionary inputs.
|
# This is the version of torch required to run torch.fx features and torch.onnx with dictionary inputs.
|
||||||
TORCH_FX_REQUIRED_VERSION = version.parse("1.10")
|
TORCH_FX_REQUIRED_VERSION = version.parse("1.10")
|
||||||
|
|
||||||
ACCELERATE_MIN_VERSION = "0.21.0"
|
ACCELERATE_MIN_VERSION = "0.26.0"
|
||||||
FSDP_MIN_VERSION = "1.12.0"
|
FSDP_MIN_VERSION = "1.12.0"
|
||||||
XLA_FSDPV2_MIN_VERSION = "2.2.0"
|
XLA_FSDPV2_MIN_VERSION = "2.2.0"
|
||||||
|
|
||||||
|
@ -196,11 +196,7 @@ class TrainerIntegrationFSDP(TestCasePlus, TrainerIntegrationCommon):
|
|||||||
self.assertEqual(trainer.args.fsdp[0], sharding_strategy)
|
self.assertEqual(trainer.args.fsdp[0], sharding_strategy)
|
||||||
self.assertEqual(trainer.args.fsdp[1], FSDPOption.OFFLOAD)
|
self.assertEqual(trainer.args.fsdp[1], FSDPOption.OFFLOAD)
|
||||||
self.assertEqual(trainer.args.fsdp[2], FSDPOption.AUTO_WRAP)
|
self.assertEqual(trainer.args.fsdp[2], FSDPOption.AUTO_WRAP)
|
||||||
fsdp_sharding_strategy = (
|
fsdp_sharding_strategy = str(FSDP_SHARDING_STRATEGY.index(sharding_strategy.upper()) + 1)
|
||||||
str(FSDP_SHARDING_STRATEGY.index(sharding_strategy.upper()) + 1)
|
|
||||||
if is_accelerate_available("0.26.0")
|
|
||||||
else sharding_strategy.upper()
|
|
||||||
)
|
|
||||||
self.assertEqual(os.environ[f"{prefix}SHARDING_STRATEGY"], fsdp_sharding_strategy)
|
self.assertEqual(os.environ[f"{prefix}SHARDING_STRATEGY"], fsdp_sharding_strategy)
|
||||||
self.assertEqual(os.environ[f"{prefix}OFFLOAD_PARAMS"], "true")
|
self.assertEqual(os.environ[f"{prefix}OFFLOAD_PARAMS"], "true")
|
||||||
self.assertEqual(os.environ[f"{prefix}AUTO_WRAP_POLICY"], "TRANSFORMER_BASED_WRAP")
|
self.assertEqual(os.environ[f"{prefix}AUTO_WRAP_POLICY"], "TRANSFORMER_BASED_WRAP")
|
||||||
|
Loading…
Reference in New Issue
Block a user