mirror of
https://github.com/huggingface/transformers.git
synced 2025-08-02 19:21:31 +06:00
fix galore layerwise with frozen params (#29743)
This commit is contained in:
parent
8692aa88e2
commit
a1a7454107
@ -385,7 +385,8 @@ def get_scheduler(
|
||||
scheduler_dict[param].step()
|
||||
|
||||
for param in optimizer_dict.keys():
|
||||
param.register_post_accumulate_grad_hook(scheduler_hook)
|
||||
if param.requires_grad:
|
||||
param.register_post_accumulate_grad_hook(scheduler_hook)
|
||||
|
||||
return LayerWiseDummyScheduler()
|
||||
|
||||
|
@ -1303,7 +1303,8 @@ class Trainer:
|
||||
optimizer_dict[param].zero_grad()
|
||||
|
||||
for param in model.parameters():
|
||||
param.register_post_accumulate_grad_hook(optimizer_hook)
|
||||
if param.requires_grad:
|
||||
param.register_post_accumulate_grad_hook(optimizer_hook)
|
||||
|
||||
optimizer_cls = LayerWiseDummyOptimizer
|
||||
optimizer_kwargs.update({"optimizer_dict": optimizer_dict})
|
||||
|
Loading…
Reference in New Issue
Block a user