diff --git a/tests/test_trainer.py b/tests/test_trainer.py index f29a8a60fc7..ae5fe2a8195 100644 --- a/tests/test_trainer.py +++ b/tests/test_trainer.py @@ -992,7 +992,7 @@ class TrainerIntegrationTest(unittest.TestCase): # should be about half of fp16_init # perfect world: fp32_init/2 == fp16_eval self.assertAlmostEqual(fp16_eval, fp32_init / 2, delta=5_000) - + def test_no_wd_param_group(self): model = torch.nn.Sequential(TstLayer(128), torch.nn.ModuleList([TstLayer(128), TstLayer(128)])) trainer = Trainer(model=model) diff --git a/tests/test_trainer_utils.py b/tests/test_trainer_utils.py index 2d9d1d688fa..f56ef140e8e 100644 --- a/tests/test_trainer_utils.py +++ b/tests/test_trainer_utils.py @@ -30,7 +30,7 @@ if is_torch_available(): DistributedTensorGatherer, LabelSmoother, LengthGroupedSampler, - get_parameter_names + get_parameter_names, ) class TstLayer(torch.nn.Module):