From a6f412da01d15cdd60e242e9765d4dfc175adb24 Mon Sep 17 00:00:00 2001 From: Christopher Goh Date: Wed, 7 Aug 2019 02:19:14 +0800 Subject: [PATCH] Fixed typo in migration guide --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index 38143d0b1fd..b86a5238c26 100644 --- a/README.md +++ b/README.md @@ -314,7 +314,7 @@ loss = outputs[0] # In pytorch-transformers you can also have access to the logits: loss, logits = outputs[:2] -# And even the attention weigths if you configure the model to output them (and other outputs too, see the docstrings and documentation) +# And even the attention weights if you configure the model to output them (and other outputs too, see the docstrings and documentation) model = BertForSequenceClassification.from_pretrained('bert-base-uncased', output_attentions=True) outputs = model(input_ids, labels=labels) loss, logits, attentions = outputs