mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-31 02:02:21 +06:00
fix: Propagate lr_scheduler_kwargs
options to create LR Scheduler when LayerWiseDummyOptimizer is used (#34559)
fix: fix get_scheduler
This commit is contained in:
parent
8fb60bf6be
commit
ab65ba47ad
@ -549,6 +549,7 @@ def get_scheduler(
|
||||
optimizer=optimizer_dict[param],
|
||||
num_warmup_steps=num_warmup_steps,
|
||||
num_training_steps=num_training_steps,
|
||||
scheduler_specific_kwargs=scheduler_specific_kwargs,
|
||||
)
|
||||
|
||||
def scheduler_hook(param):
|
||||
|
Loading…
Reference in New Issue
Block a user