From e6e5f1925764dbec2d5bdeec60a3216995bacc49 Mon Sep 17 00:00:00 2001 From: thomwolf Date: Tue, 18 Jun 2019 14:45:14 +0200 Subject: [PATCH] fix --- examples/run_squad.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/examples/run_squad.py b/examples/run_squad.py index fb3b4b7d34a..ce17b789e26 100644 --- a/examples/run_squad.py +++ b/examples/run_squad.py @@ -201,7 +201,7 @@ def main(): if args.do_train: if args.local_rank in [-1, 0]: - writer = SummaryWriter() + tb_writer = SummaryWriter() # Prepare data loader train_examples = read_squad_examples( input_file=args.train_file, is_training=True, version_2_with_negative=args.version_2_with_negative) @@ -302,8 +302,8 @@ def main(): loss.backward() if (step + 1) % args.gradient_accumulation_steps == 0: if args.local_rank in [-1, 0]: - writer.add_scalar('lr', optimizer.get_lr()[0], global_step) - writer.add_scalar('loss', loss.item(), global_step) + tb_writer.add_scalar('lr', optimizer.get_lr()[0], global_step) + tb_writer.add_scalar('loss', loss.item(), global_step) if args.fp16: # modify learning rate with special warm up BERT uses # if args.fp16 is False, BertAdam is used and handles this automatically