fix: Propagate lr_scheduler_kwargs options to create LR Scheduler when LayerWiseDummyOptimizer is used (#34559)

fix: fix get_scheduler
This commit is contained in:
Bongseok Lee 2025-05-13 20:56:45 +09:00 committed by GitHub
parent 8fb60bf6be
commit ab65ba47ad
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -549,6 +549,7 @@ def get_scheduler(
optimizer=optimizer_dict[param],
num_warmup_steps=num_warmup_steps,
num_training_steps=num_training_steps,
scheduler_specific_kwargs=scheduler_specific_kwargs,
)
def scheduler_hook(param):