Commit Graph

19383 Commits

Author SHA1 Message Date
LysandreJik
c3df2136e1 Added binary masking tests 2019-09-19 10:55:06 +02:00
LysandreJik
e391d4735e Tokenizers' encode function can output binary masks 2019-09-19 10:55:06 +02:00
sshleifer
119610b5c5 Merge branch 'master' into delete-n-special-doc 2019-09-19 01:35:01 -07:00
sshleifer
08e4ad5eea Remove documentation for unused kwarg 2019-09-18 16:35:01 -07:00
Erik Chan
f0340eccf9
Typo
Typo
2019-09-18 13:42:11 -07:00
Thomas Wolf
0d1dad6d53
Merge pull request #1004 from erenup/master
Refactoring old run_swag.py
2019-09-18 21:42:51 +02:00
erenup
8960988f35 fixed to find best dev acc 2019-09-19 01:10:05 +08:00
erenup
b57bfb5fa0
Merge pull request #3 from erenup/run_multiple_choice_merge
Run multiple choice merge
2019-09-18 21:45:04 +08:00
erenup
46ffc28329 Merge branch 'master' into run_multiple_choice_merge
# Please enter a commit message to explain why this merge is necessary,
# especially if it merges an updated upstream into a topic branch.
#
# Lines starting with '#' will be ignored, and an empty message aborts
# the commit.
2019-09-18 21:43:46 +08:00
Simon Layton
ec94f4e0f8
Fix fp16 masking in PoolerEndLogits
Necessary to run xlnet (at least in squad) with `--fp16 --fp16_opt_level="O2"`, otherwise loss is immediately `NaN` and fine-tuning cannot proceed.
2019-09-18 09:30:58 -04:00
erenup
15143fbad6 move run_multiple_choice.py and utils_multiple_choice.py to examples 2019-09-18 21:18:46 +08:00
erenup
3cd6289758 Merge remote-tracking branch 'huggingface/master' into run_multiple_choice_merge
# Conflicts:
#	examples/contrib/run_swag.py
2019-09-18 21:16:59 +08:00
erenup
36362cf086 move schedule.step after optimizer.step 2019-09-18 21:13:40 +08:00
thomwolf
3a527fa820 OpenAI GPT tests ok 2019-09-18 14:15:48 +02:00
thomwolf
556442afb3 hot fix 2019-09-18 14:12:41 +02:00
thomwolf
160b5d6080 fix xlm lang_embeddings loading 2019-09-18 14:10:20 +02:00
thomwolf
26497d1199 fix tests 2019-09-18 12:17:21 +02:00
thomwolf
6a083fd447 update pt-tf conversion script 2019-09-18 12:11:32 +02:00
thomwolf
f6969cc12b upgrade max model difference to 2e-2 (for transfo-xl adaptive softmax + inputs) 2019-09-18 11:12:02 +02:00
thomwolf
e768f2322a update run_openai_gpt to fix #1264 2019-09-18 10:07:47 +02:00
thomwolf
8334993915 clean up examples - updated to new keyword inputs - #1246 2019-09-18 10:01:27 +02:00
Julien Chaumond
62760baf46 tiny fixes 2019-09-17 18:29:15 -04:00
thomwolf
45de034bf8 fix #1223 2019-09-17 10:25:06 +02:00
erenup
5a81e79e25
Merge pull request #2 from erenup/run_multiple_choice_add_doc
Run multiple choice add doc
2019-09-16 22:39:54 +08:00
erenup
5882c442e5 add example usage 2019-09-16 22:38:08 +08:00
erenup
a9debaca3d fixed init_weight 2019-09-16 19:55:24 +08:00
thomwolf
c88f05163d fix typo in XLM models 2019-09-16 13:42:20 +02:00
erenup
982f181aa7 Merge remote-tracking branch 'origin/master' into run_multiple_choice_add_doc 2019-09-16 19:12:00 +08:00
erenup
84b9d1c423 Merge remote-tracking branch 'huggingface/master'
# Conflicts:
#	pytorch_transformers/__init__.py
2019-09-16 19:06:12 +08:00
erenup
603b470a3d add warnning info 2019-09-16 18:53:37 +08:00
erenup
4812a5a767 add doc string 2019-09-16 11:50:18 +08:00
thomwolf
4b956b2a6b add layer_norm_epsilon configuration for transformer xl 2019-09-13 17:09:20 +02:00
thomwolf
b97af8cce9 skip finetuned checkpoints 2019-09-13 16:43:49 +02:00
thomwolf
65c49bb27e adding TF 2.0 adaptive softmax with logits + loss outputs 2019-09-13 15:50:51 +02:00
thomwolf
39c38b2ea0 fix 2019-09-12 16:47:11 +02:00
thomwolf
dcddf498c8 fix bert layernorm 2019-09-12 16:46:32 +02:00
thomwolf
d3a3a0353c clean up cache after conversion 2019-09-12 16:42:52 +02:00
thomwolf
a84adddd1b convert all models 2019-09-12 13:14:07 +02:00
VictorSanh
32e1332acf [distil] fix once for all general logger for scripts 2019-09-11 14:19:07 +00:00
Thomas Wolf
b62abe87c9
Merge pull request #1249 from ziliwang/master
fixed: hard coding for max and min number will out of range in fp16, which will cause nan.
2019-09-11 15:53:28 +02:00
thomwolf
969d3ae95e XLMWithLMHead fixed - standardize conversion 2019-09-11 15:47:33 +02:00
thomwolf
646711e1e2 standardize scopes names - add conversion methods 2019-09-11 15:34:17 +02:00
thomwolf
4356f791a2 XLM passing tests 2019-09-11 11:49:54 +02:00
LysandreJik
11ac4b9555 [CI] Symbolic link for documentation 2019-09-11 10:13:44 +02:00
Zili Wang
8bdee1cb73 fixed: hard coding for max and min number will out of range in fp16, which will cause nan. 2019-09-11 15:41:53 +08:00
ziliwang
7424b2848f
Merge pull request #1 from huggingface/master
merege from original repo
2019-09-11 11:02:23 +08:00
VictorSanh
364920e216 fix small bug/typo 2019-09-10 21:45:01 +00:00
Thomas Wolf
23c23f5399
Merge pull request #1229 from SKRohit/master
changes in evaluate function in run_lm_finetuning.py
2019-09-10 22:16:45 +02:00
Thomas Wolf
99a54ac51c
Merge pull request #1233 from searchivarius/master
Fix to prevent crashing on assert len(tokens_b)>=1
2019-09-10 22:15:47 +02:00
Thomas Wolf
439b37b474
Merge pull request #1241 from mattolson93/patch-1
Fixing typo in gpt2 for doc site's class link
2019-09-10 22:14:18 +02:00