erenup
4270d3da1b
fix a bug of evaluating
2019-08-19 16:38:52 +08:00
erenup
b8fde43868
a coding bug
2019-08-19 16:36:43 +08:00
Chi-Liang Liu
40acf6b52a
don't save model without training
2019-08-18 05:02:25 -04:00
erenup
47e9aea0fe
add args info to evaluate_result.txt
2019-08-18 17:00:53 +08:00
erenup
5582bc4b23
add multiple choice to robreta and xlnet, test on swag, roberta=0.82.28
...
, xlnet=0.80
2019-08-18 16:01:48 +08:00
wangfei
856a63da4d
Fix: save model/model.module
2019-08-18 11:03:47 +08:00
wangfei
1ef41b8337
Revert "Fix: save model/model.module"
...
This reverts commit 00e9c4cc96
.
2019-08-18 11:03:12 +08:00
wangfei
00e9c4cc96
Fix: save model/model.module
2019-08-18 11:02:02 +08:00
Christophe Bourguignat
189ff9b664
Update README after RoBERTa addition
2019-08-17 13:18:37 -04:00
erenup
e384ae2b9d
Merge remote-tracking branch 'huggingface/master'
...
merge huggingface/master to update
2019-08-17 12:05:57 +08:00
Jason Phang
d8923270e6
Correct truncation for RoBERTa in 2-input GLUE
2019-08-16 16:30:38 -04:00
Lysandre
5652f54ac2
Simplified data generator + better perplexity calculator
...
GPT-2 now obtains ~20 perplexity on WikiText-2
2019-08-16 13:49:56 -04:00
LysandreJik
7e7fc53da5
Fixing run_glue example with RoBERTa
2019-08-16 11:53:10 -04:00
LysandreJik
715534800a
BERT + RoBERTa masking tokens handling + GPU device update.
2019-08-16 10:10:21 -04:00
LysandreJik
339e556feb
CLM for BERT, beginning of CLM fot RoBERTa; still needs a better masking token mechanism.
2019-08-16 10:10:20 -04:00
LysandreJik
5c18825a18
Removed dataset limit
2019-08-16 10:10:20 -04:00
LysandreJik
3e3e145497
Added GPT to the generative fine-tuning.
2019-08-16 10:10:20 -04:00
LysandreJik
47975ed53e
Language Modeling fine-tuning using GPT-2.
2019-08-16 10:10:20 -04:00
LysandreJik
ab05280666
Order of strings in AutoModel/AutoTokenizer updated.
2019-08-16 09:53:26 -04:00
wangfei
b8ff56896c
Fix bug of multi-gpu training in lm finetuning
2019-08-16 12:11:05 +08:00
LysandreJik
9d0029e215
Added RoBERTa example to README
2019-08-15 17:17:35 -04:00
LysandreJik
83dba0b67b
Added RoBERTa tokenizer to AutoTokenizer
2019-08-15 17:07:07 -04:00
LysandreJik
e24e19ce3b
Added RoBERTa to AutoModel/AutoConfig
2019-08-15 14:02:11 -04:00
LysandreJik
fe02e45e48
Release: 1.1.0
2019-08-15 11:15:08 -04:00
Lysandre Debut
88efc65bac
Merge pull request #964 from huggingface/RoBERTa
...
RoBERTa: model conversion, inference, tests 🔥
2019-08-15 11:11:10 -04:00
LysandreJik
8308170156
Warning for RoBERTa sequences encoded without special tokens.
2019-08-15 10:29:04 -04:00
LysandreJik
572dcfd1db
Doc
2019-08-14 14:56:14 -04:00
Julien Chaumond
c4ef103447
[RoBERTa] First 4 authors
...
cf. https://github.com/huggingface/pytorch-transformers/pull/964#discussion_r313574354
Co-Authored-By: Myle Ott <myleott@fb.com>
2019-08-14 12:31:09 -04:00
Rabeeh KARIMI
3d47a7f8ab
loads the tokenizer for each checkpoint, to solve the reproducability issue
2019-08-14 10:58:26 +02:00
samvelyan
9ce36e3e4b
Re-implemented tokenize() iteratively in PreTrainedTokenizer.
2019-08-14 08:57:09 +00:00
LysandreJik
39f426be65
Added special tokens <pad> and <mask> to RoBERTa.
2019-08-13 15:19:50 -04:00
Julien Chaumond
baf08ca1d4
[RoBERTa] run_glue: correct pad_token + reorder labels
2019-08-13 12:51:15 -04:00
LysandreJik
3d87991f60
Fixed error with encoding
2019-08-13 12:00:24 -04:00
tuvuumass
ba4bce2581
fix issue #824
2019-08-13 11:26:27 -04:00
LysandreJik
634a3172d8
Added integration tests for sequence builders.
2019-08-12 15:14:15 -04:00
LysandreJik
22ac004a7c
Added documentation and changed parameters for special_tokens_sentences_pair.
2019-08-12 15:13:53 -04:00
Julien Chaumond
912fdff899
[RoBERTa] Update run_glue
for RoBERTa
2019-08-12 13:49:50 -04:00
Julien Chaumond
b3d83d68db
Fixup 9d0603148b
2019-08-12 12:28:55 -04:00
carefree0910
a7b4cfe919
Update README.md
...
I assume that it should test the `re-load` functionality after testing the `save` functionality, however I'm also surprised that nobody points this out after such a long time, so maybe I've misunderstood the purpose. This PR is just in case :)
2019-08-12 09:53:05 -04:00
erenup
b219029c45
refactoring old run_swag. This script is mainly refatored from run_squad in pytorch_transformers
2019-08-11 15:20:37 +08:00
thomwolf
aaedfc35a8
Merge branch 'master' of https://github.com/huggingface/pytorch-transformers
2019-08-10 20:04:37 +02:00
thomwolf
c683c3d5a5
fix #993
2019-08-10 20:04:35 +02:00
Kevin Trebing
7060766490
Corrected logger.error info
...
Signed-off-by: Kevin Trebing <Kevin.Trebing@gmx.net>
2019-08-09 19:36:44 -04:00
LysandreJik
75d5f98fd2
Roberta tokenization + fixed tests (py3 + py2).
2019-08-09 15:02:13 -04:00
LysandreJik
14e970c271
Tokenization encode/decode class-based sequence handling
2019-08-09 15:01:38 -04:00
LysandreJik
3566d27919
Clarified PreTrainedModel.from_pretrained warning messages in documentation.
2019-08-08 19:04:34 -04:00
LysandreJik
fbd746bd06
Updated test architecture
2019-08-08 18:21:34 -04:00
LysandreJik
6c41a8f5dc
Encode and Decode are back in the superclass. They now handle sentence pairs special tokens.
2019-08-08 18:20:32 -04:00
Julien Chaumond
e367ac469c
[RoBERTa] Re-apply 39d72bcc7b
...
cc @lysandrejik
2019-08-08 11:26:11 -04:00
Julien Chaumond
9d0603148b
[RoBERTa] RobertaForSequenceClassification + conversion
2019-08-08 11:24:54 -04:00