Guillem García Subies
|
bb04446285
|
Update tokenization_openai.py
|
2019-08-20 14:07:40 +02:00 |
|
Guillem García Subies
|
bfd75056b0
|
Update tokenization_xlm.py
|
2019-08-20 14:06:17 +02:00 |
|
erenup
|
fc74132598
|
add best steps to train
|
2019-08-20 19:06:41 +08:00 |
|
Thomas Wolf
|
933841d903
|
Merge pull request #1056 from Morizeyao/master
Swap of optimizer.step and scheduler.step for lm finetuning examples
|
2019-08-20 12:42:24 +02:00 |
|
thomwolf
|
6d0aa73981
|
fix #1034
|
2019-08-20 12:20:21 +02:00 |
|
Julien Chaumond
|
b0b9b8091b
|
minor typo
|
2019-08-20 11:33:46 +02:00 |
|
thomwolf
|
53c8f700f4
|
fix #808
|
2019-08-20 11:29:26 +02:00 |
|
thomwolf
|
901dde0e45
|
fix #1014
|
2019-08-20 11:05:51 +02:00 |
|
thomwolf
|
e239a4a20f
|
close #984
|
2019-08-20 11:02:00 +02:00 |
|
thomwolf
|
fecaed0ed4
|
add force_download option to from_pretrained methods
|
2019-08-20 10:56:12 +02:00 |
|
Duzeyao
|
d86b49ac86
|
swap optimizer.step and scheduler.step
|
2019-08-20 16:46:34 +08:00 |
|
Duzeyao
|
45ab8bf60e
|
Revert "Update finetune_on_pregenerated.py"
This reverts commit a1359b970c .
|
2019-08-20 16:40:39 +08:00 |
|
erenup
|
97c30b73d5
|
add test related code
|
2019-08-20 16:31:04 +08:00 |
|
erenup
|
d5e60e5b7a
|
add test related code
|
2019-08-20 16:25:50 +08:00 |
|
Zeyao Du
|
a1359b970c
|
Update finetune_on_pregenerated.py
|
2019-08-20 16:00:07 +08:00 |
|
Zeyao Du
|
28f7ca1f80
|
swap optimizer.step and scheduler.step
|
2019-08-20 15:58:42 +08:00 |
|
Peng Qi
|
a368b87791
|
Fix #1015
|
2019-08-19 13:07:00 -07:00 |
|
Lysandre
|
f94f1c6016
|
Distributed training + tokenizer agnostic mask token
|
2019-08-19 14:58:50 -04:00 |
|
Lysandre
|
c589862b78
|
Doc: loading from config alone does not load the model weights
|
2019-08-19 10:17:47 -04:00 |
|
Thomas Wolf
|
5a49b793d9
|
Merge pull request #1023 from tuvuumass/patch-1
fix issue #824
|
2019-08-19 15:31:46 +02:00 |
|
erenup
|
4270d3da1b
|
fix a bug of evaluating
|
2019-08-19 16:38:52 +08:00 |
|
erenup
|
b8fde43868
|
a coding bug
|
2019-08-19 16:36:43 +08:00 |
|
Chi-Liang Liu
|
40acf6b52a
|
don't save model without training
|
2019-08-18 05:02:25 -04:00 |
|
erenup
|
47e9aea0fe
|
add args info to evaluate_result.txt
|
2019-08-18 17:00:53 +08:00 |
|
erenup
|
5582bc4b23
|
add multiple choice to robreta and xlnet, test on swag, roberta=0.82.28
, xlnet=0.80
|
2019-08-18 16:01:48 +08:00 |
|
wangfei
|
856a63da4d
|
Fix: save model/model.module
|
2019-08-18 11:03:47 +08:00 |
|
wangfei
|
1ef41b8337
|
Revert "Fix: save model/model.module"
This reverts commit 00e9c4cc96 .
|
2019-08-18 11:03:12 +08:00 |
|
wangfei
|
00e9c4cc96
|
Fix: save model/model.module
|
2019-08-18 11:02:02 +08:00 |
|
Christophe Bourguignat
|
189ff9b664
|
Update README after RoBERTa addition
|
2019-08-17 13:18:37 -04:00 |
|
erenup
|
e384ae2b9d
|
Merge remote-tracking branch 'huggingface/master'
merge huggingface/master to update
|
2019-08-17 12:05:57 +08:00 |
|
Jason Phang
|
d8923270e6
|
Correct truncation for RoBERTa in 2-input GLUE
|
2019-08-16 16:30:38 -04:00 |
|
Lysandre
|
5652f54ac2
|
Simplified data generator + better perplexity calculator
GPT-2 now obtains ~20 perplexity on WikiText-2
|
2019-08-16 13:49:56 -04:00 |
|
LysandreJik
|
7e7fc53da5
|
Fixing run_glue example with RoBERTa
|
2019-08-16 11:53:10 -04:00 |
|
LysandreJik
|
715534800a
|
BERT + RoBERTa masking tokens handling + GPU device update.
|
2019-08-16 10:10:21 -04:00 |
|
LysandreJik
|
339e556feb
|
CLM for BERT, beginning of CLM fot RoBERTa; still needs a better masking token mechanism.
|
2019-08-16 10:10:20 -04:00 |
|
LysandreJik
|
5c18825a18
|
Removed dataset limit
|
2019-08-16 10:10:20 -04:00 |
|
LysandreJik
|
3e3e145497
|
Added GPT to the generative fine-tuning.
|
2019-08-16 10:10:20 -04:00 |
|
LysandreJik
|
47975ed53e
|
Language Modeling fine-tuning using GPT-2.
|
2019-08-16 10:10:20 -04:00 |
|
LysandreJik
|
ab05280666
|
Order of strings in AutoModel/AutoTokenizer updated.
|
2019-08-16 09:53:26 -04:00 |
|
wangfei
|
b8ff56896c
|
Fix bug of multi-gpu training in lm finetuning
|
2019-08-16 12:11:05 +08:00 |
|
LysandreJik
|
9d0029e215
|
Added RoBERTa example to README
|
2019-08-15 17:17:35 -04:00 |
|
LysandreJik
|
83dba0b67b
|
Added RoBERTa tokenizer to AutoTokenizer
|
2019-08-15 17:07:07 -04:00 |
|
LysandreJik
|
e24e19ce3b
|
Added RoBERTa to AutoModel/AutoConfig
|
2019-08-15 14:02:11 -04:00 |
|
LysandreJik
|
fe02e45e48
|
Release: 1.1.0
|
2019-08-15 11:15:08 -04:00 |
|
Lysandre Debut
|
88efc65bac
|
Merge pull request #964 from huggingface/RoBERTa
RoBERTa: model conversion, inference, tests 🔥
|
2019-08-15 11:11:10 -04:00 |
|
LysandreJik
|
8308170156
|
Warning for RoBERTa sequences encoded without special tokens.
|
2019-08-15 10:29:04 -04:00 |
|
LysandreJik
|
572dcfd1db
|
Doc
|
2019-08-14 14:56:14 -04:00 |
|
Julien Chaumond
|
c4ef103447
|
[RoBERTa] First 4 authors
cf. https://github.com/huggingface/pytorch-transformers/pull/964#discussion_r313574354
Co-Authored-By: Myle Ott <myleott@fb.com>
|
2019-08-14 12:31:09 -04:00 |
|
Rabeeh KARIMI
|
3d47a7f8ab
|
loads the tokenizer for each checkpoint, to solve the reproducability issue
|
2019-08-14 10:58:26 +02:00 |
|
samvelyan
|
9ce36e3e4b
|
Re-implemented tokenize() iteratively in PreTrainedTokenizer.
|
2019-08-14 08:57:09 +00:00 |
|