VictorSanh
|
fea921d382
|
add licensing
|
2019-08-28 04:45:39 +00:00 |
|
VictorSanh
|
da1e4e53fc
|
some fixes in train.py for loading previous checkpoint
|
2019-08-28 04:01:03 +00:00 |
|
VictorSanh
|
0d8f8848d5
|
add scripts/extract_for_distil.py
|
2019-08-28 04:00:19 +00:00 |
|
VictorSanh
|
7f2c384c80
|
add scripts/token_counts.py
|
2019-08-28 04:00:03 +00:00 |
|
VictorSanh
|
4d16b279e5
|
add scripts/binarized_data.py
|
2019-08-28 03:59:48 +00:00 |
|
LysandreJik
|
c513415b19
|
Dilbert tests from CommonTests
|
2019-08-27 23:59:00 -04:00 |
|
LysandreJik
|
778a263f09
|
GilBert added to AutoModels
|
2019-08-27 23:14:00 -04:00 |
|
VictorSanh
|
74d78beeb4
|
fix: add qa_dropout and seq_classif_dropout
|
2019-08-28 03:13:11 +00:00 |
|
VictorSanh
|
7f5d85347e
|
fix small typo
|
2019-08-28 02:44:51 +00:00 |
|
VictorSanh
|
906581ae3c
|
add s3 links for dilbert (+fix small typo)
|
2019-08-28 02:43:33 +00:00 |
|
VictorSanh
|
b247b0d880
|
add train.py for distillation
|
2019-08-28 02:12:47 +00:00 |
|
VictorSanh
|
780f183e55
|
add requirements
|
2019-08-28 01:39:52 +00:00 |
|
VictorSanh
|
e424d2e45d
|
add README
|
2019-08-28 01:10:10 +00:00 |
|
VictorSanh
|
1ae81e4aa1
|
add dataset. distiller, utils
|
2019-08-28 01:10:05 +00:00 |
|
VictorSanh
|
5d29f8e99b
|
fix bugs
|
2019-08-28 00:57:16 +00:00 |
|
VictorSanh
|
a8ad83040d
|
fix bugs
|
2019-08-28 00:45:33 +00:00 |
|
VictorSanh
|
60c984da6c
|
fix bugs
|
2019-08-27 22:25:55 +00:00 |
|
VictorSanh
|
42968138c8
|
wip wouf
|
2019-08-27 22:00:38 +00:00 |
|
VictorSanh
|
1d23240068
|
wip
|
2019-08-27 14:27:47 +00:00 |
|
LysandreJik
|
e08c01aa1a
|
fix #1102
|
2019-08-26 18:13:06 -04:00 |
|
Thomas Wolf
|
df9d6effae
|
Merge pull request #1081 from huggingface/fix_distributed_barrier_hang
Fix distributed barrier hang
|
2019-08-23 16:53:53 +02:00 |
|
Thomas Wolf
|
3f20dd7186
|
Merge pull request #1075 from abhishekraok/modeling_utils_config_None
reraise EnvironmentError in modeling_utils.py
|
2019-08-23 12:42:39 +02:00 |
|
Abhishek Rao
|
c603d099aa
|
reraise EnvironmentError in from_pretrained functions of Model and Tokenizer
|
2019-08-22 15:25:40 -07:00 |
|
VictorSanh
|
57272d5ddf
|
fix for glue
|
2019-08-22 00:25:49 -04:00 |
|
VictorSanh
|
b006a7a12f
|
fix for squad
|
2019-08-22 00:25:42 -04:00 |
|
Abhishek Rao
|
14eef67eb2
|
Fix at config rather than model
|
2019-08-21 15:48:43 -07:00 |
|
Abhishek Rao
|
296df2b18c
|
reraise exception
|
2019-08-21 15:29:30 -07:00 |
|
thomwolf
|
e00b4ff1de
|
fix #1017
|
2019-08-21 22:22:17 +02:00 |
|
Lysandre
|
2f9397139d
|
Added GPT-2 LARGE to Pre-trained Models documentation
|
2019-08-21 11:29:37 -04:00 |
|
VictorSanh
|
6f877d9daf
|
Update dev results on GLUE (bert-base-uncased) w/ median on 5 runs
|
2019-08-21 03:43:29 +00:00 |
|
Thomas Wolf
|
07681b6b58
|
Merge pull request #1064 from huggingface/gpt-2-large
Adding gpt-2 large (774M parameters) model
|
2019-08-21 03:05:56 +02:00 |
|
thomwolf
|
fdc487d8b3
|
Add max length
|
2019-08-21 02:35:01 +02:00 |
|
thomwolf
|
aa05dc8935
|
adding gpt-2 large
|
2019-08-21 02:29:34 +02:00 |
|
Thomas Wolf
|
e4515faf54
|
Merge pull request #1057 from huggingface/fixes
Add a few of typos corrections, bugs fixes and small improvements
|
2019-08-21 01:54:05 +02:00 |
|
Thomas Wolf
|
41789c6c3d
|
Merge pull request #1059 from GuillemGSubies/master
Better use of spacy tokenizer in open ai and xlm tokenizers
|
2019-08-21 01:53:48 +02:00 |
|
Thomas Wolf
|
260c86082d
|
Merge pull request #1027 from samvelyan/iterative_split_on_token
Re-implemented tokenize() iteratively in PreTrainedTokenizer.
|
2019-08-21 01:46:03 +02:00 |
|
Thomas Wolf
|
d30cbaf5dc
|
Merge branch 'master' into iterative_split_on_token
|
2019-08-21 01:33:02 +02:00 |
|
Thomas Wolf
|
9beaa85b07
|
Merge pull request #1055 from qipeng/run_squad_fix
Fix #1015 (tokenizer defaults to use_lower_case=True when loading from trained models)
|
2019-08-21 01:20:46 +02:00 |
|
Thomas Wolf
|
e753f249e1
|
Merge pull request #806 from wschin/fix-a-path
Fix a path so that a test can run on Windows
|
2019-08-21 01:14:40 +02:00 |
|
Peng Qi
|
3bffd2e8e5
|
more fixes
|
2019-08-20 10:59:28 -07:00 |
|
Thomas Wolf
|
c3619f5536
|
Merge pull request #1060 from CrafterKolyan/patch-1
Fix typo. configuratoin -> configuration
|
2019-08-20 17:39:06 +02:00 |
|
Thomas Wolf
|
3b56427a1e
|
Merge pull request #1040 from FeiWang96/multi_gpu
Fix bug of multi-gpu training in lm finetuning
|
2019-08-20 17:13:44 +02:00 |
|
thomwolf
|
43489756ad
|
adding proxies options for the from_pretrained methods
|
2019-08-20 16:59:11 +02:00 |
|
Nikolay Korolev
|
ad6e62cd82
|
Fix typo. configuratoin -> configuration
|
2019-08-20 15:43:06 +03:00 |
|
Guillem García Subies
|
388e3251fa
|
Update tokenization_xlm.py
|
2019-08-20 14:19:39 +02:00 |
|
Guillem García Subies
|
f5e2ed0fd8
|
Update tokenization_openai.py
|
2019-08-20 14:19:25 +02:00 |
|
Guillem García Subies
|
562b998366
|
Update tokenization_openai.py
|
2019-08-20 14:10:19 +02:00 |
|
Guillem García Subies
|
bb04446285
|
Update tokenization_openai.py
|
2019-08-20 14:07:40 +02:00 |
|
Guillem García Subies
|
bfd75056b0
|
Update tokenization_xlm.py
|
2019-08-20 14:06:17 +02:00 |
|
Thomas Wolf
|
933841d903
|
Merge pull request #1056 from Morizeyao/master
Swap of optimizer.step and scheduler.step for lm finetuning examples
|
2019-08-20 12:42:24 +02:00 |
|