LysandreJik
c3df2136e1
Added binary masking tests
2019-09-19 10:55:06 +02:00
LysandreJik
e391d4735e
Tokenizers' encode function can output binary masks
2019-09-19 10:55:06 +02:00
sshleifer
119610b5c5
Merge branch 'master' into delete-n-special-doc
2019-09-19 01:35:01 -07:00
sshleifer
08e4ad5eea
Remove documentation for unused kwarg
2019-09-18 16:35:01 -07:00
Erik Chan
f0340eccf9
Typo
...
Typo
2019-09-18 13:42:11 -07:00
Thomas Wolf
0d1dad6d53
Merge pull request #1004 from erenup/master
...
Refactoring old run_swag.py
2019-09-18 21:42:51 +02:00
erenup
8960988f35
fixed to find best dev acc
2019-09-19 01:10:05 +08:00
erenup
b57bfb5fa0
Merge pull request #3 from erenup/run_multiple_choice_merge
...
Run multiple choice merge
2019-09-18 21:45:04 +08:00
erenup
46ffc28329
Merge branch 'master' into run_multiple_choice_merge
...
# Please enter a commit message to explain why this merge is necessary,
# especially if it merges an updated upstream into a topic branch.
#
# Lines starting with '#' will be ignored, and an empty message aborts
# the commit.
2019-09-18 21:43:46 +08:00
Simon Layton
ec94f4e0f8
Fix fp16 masking in PoolerEndLogits
...
Necessary to run xlnet (at least in squad) with `--fp16 --fp16_opt_level="O2"`, otherwise loss is immediately `NaN` and fine-tuning cannot proceed.
2019-09-18 09:30:58 -04:00
erenup
15143fbad6
move run_multiple_choice.py and utils_multiple_choice.py to examples
2019-09-18 21:18:46 +08:00
erenup
3cd6289758
Merge remote-tracking branch 'huggingface/master' into run_multiple_choice_merge
...
# Conflicts:
# examples/contrib/run_swag.py
2019-09-18 21:16:59 +08:00
erenup
36362cf086
move schedule.step after optimizer.step
2019-09-18 21:13:40 +08:00
thomwolf
3a527fa820
OpenAI GPT tests ok
2019-09-18 14:15:48 +02:00
thomwolf
556442afb3
hot fix
2019-09-18 14:12:41 +02:00
thomwolf
160b5d6080
fix xlm lang_embeddings loading
2019-09-18 14:10:20 +02:00
thomwolf
26497d1199
fix tests
2019-09-18 12:17:21 +02:00
thomwolf
6a083fd447
update pt-tf conversion script
2019-09-18 12:11:32 +02:00
thomwolf
f6969cc12b
upgrade max model difference to 2e-2 (for transfo-xl adaptive softmax + inputs)
2019-09-18 11:12:02 +02:00
thomwolf
e768f2322a
update run_openai_gpt to fix #1264
2019-09-18 10:07:47 +02:00
thomwolf
8334993915
clean up examples - updated to new keyword inputs - #1246
2019-09-18 10:01:27 +02:00
Julien Chaumond
62760baf46
tiny fixes
2019-09-17 18:29:15 -04:00
thomwolf
45de034bf8
fix #1223
2019-09-17 10:25:06 +02:00
erenup
5a81e79e25
Merge pull request #2 from erenup/run_multiple_choice_add_doc
...
Run multiple choice add doc
2019-09-16 22:39:54 +08:00
erenup
5882c442e5
add example usage
2019-09-16 22:38:08 +08:00
erenup
a9debaca3d
fixed init_weight
2019-09-16 19:55:24 +08:00
thomwolf
c88f05163d
fix typo in XLM models
2019-09-16 13:42:20 +02:00
erenup
982f181aa7
Merge remote-tracking branch 'origin/master' into run_multiple_choice_add_doc
2019-09-16 19:12:00 +08:00
erenup
84b9d1c423
Merge remote-tracking branch 'huggingface/master'
...
# Conflicts:
# pytorch_transformers/__init__.py
2019-09-16 19:06:12 +08:00
erenup
603b470a3d
add warnning info
2019-09-16 18:53:37 +08:00
erenup
4812a5a767
add doc string
2019-09-16 11:50:18 +08:00
thomwolf
4b956b2a6b
add layer_norm_epsilon configuration for transformer xl
2019-09-13 17:09:20 +02:00
thomwolf
b97af8cce9
skip finetuned checkpoints
2019-09-13 16:43:49 +02:00
thomwolf
65c49bb27e
adding TF 2.0 adaptive softmax with logits + loss outputs
2019-09-13 15:50:51 +02:00
thomwolf
39c38b2ea0
fix
2019-09-12 16:47:11 +02:00
thomwolf
dcddf498c8
fix bert layernorm
2019-09-12 16:46:32 +02:00
thomwolf
d3a3a0353c
clean up cache after conversion
2019-09-12 16:42:52 +02:00
thomwolf
a84adddd1b
convert all models
2019-09-12 13:14:07 +02:00
VictorSanh
32e1332acf
[distil] fix once for all general logger for scripts
2019-09-11 14:19:07 +00:00
Thomas Wolf
b62abe87c9
Merge pull request #1249 from ziliwang/master
...
fixed: hard coding for max and min number will out of range in fp16, which will cause nan.
2019-09-11 15:53:28 +02:00
thomwolf
969d3ae95e
XLMWithLMHead fixed - standardize conversion
2019-09-11 15:47:33 +02:00
thomwolf
646711e1e2
standardize scopes names - add conversion methods
2019-09-11 15:34:17 +02:00
thomwolf
4356f791a2
XLM passing tests
2019-09-11 11:49:54 +02:00
LysandreJik
11ac4b9555
[CI] Symbolic link for documentation
2019-09-11 10:13:44 +02:00
Zili Wang
8bdee1cb73
fixed: hard coding for max and min number will out of range in fp16, which will cause nan.
2019-09-11 15:41:53 +08:00
ziliwang
7424b2848f
Merge pull request #1 from huggingface/master
...
merege from original repo
2019-09-11 11:02:23 +08:00
VictorSanh
364920e216
fix small bug/typo
2019-09-10 21:45:01 +00:00
Thomas Wolf
23c23f5399
Merge pull request #1229 from SKRohit/master
...
changes in evaluate function in run_lm_finetuning.py
2019-09-10 22:16:45 +02:00
Thomas Wolf
99a54ac51c
Merge pull request #1233 from searchivarius/master
...
Fix to prevent crashing on assert len(tokens_b)>=1
2019-09-10 22:15:47 +02:00
Thomas Wolf
439b37b474
Merge pull request #1241 from mattolson93/patch-1
...
Fixing typo in gpt2 for doc site's class link
2019-09-10 22:14:18 +02:00