Thomas Wolf
1569610f2d
Merge pull request #1296 from danai-antoniou/add-duplicate-tokens-error
...
Added ValueError for duplicates in list of added tokens
2019-10-03 17:06:17 -04:00
drc10723
e1b2949ae6
DistillBert Documentation Code Example fixes
2019-10-03 15:51:33 -04:00
Simon Layton
899883644f
Fix test fails and warnings
...
Attention output was in bnij ordering instead of ijbn which everything
else will expect. This was an oversight on my part, and keeps the
attention inputs/outputs identical to the original code.
Also moved back from tensor slicing to index_select in rel_shift_bnij to
make the tracer happy.
2019-10-03 12:05:15 -04:00
VictorSanh
e2ae9c0b73
fix links in doc index
2019-10-03 11:42:21 -04:00
LysandreJik
aebd83230f
Update naming + remove f string in run_lm_finetuning example
2019-10-03 11:31:36 -04:00
LysandreJik
651bfb7ad5
always_truncate by default
2019-10-03 11:31:36 -04:00
LysandreJik
5ed50a93fb
LM finetuning won't mask special tokens anymore
2019-10-03 11:31:36 -04:00
LysandreJik
cc412edd42
Supports already existing special tokens
2019-10-03 11:31:36 -04:00
LysandreJik
2f259b228e
Sequence IDS
2019-10-03 11:31:36 -04:00
LysandreJik
7c789c337d
Always truncate argument in the encode method
2019-10-03 11:31:36 -04:00
Brian Ma
7af0777910
Update run_glue.py
...
add DistilBert model shortcut into ALL_MODELS
2019-10-03 15:31:11 +00:00
VictorSanh
c1689ac301
fix name
2019-10-03 10:56:39 -04:00
VictorSanh
4a790c40b1
update doc for distil*
2019-10-03 10:54:02 -04:00
VictorSanh
6be46a6e64
update links to new weights
2019-10-03 10:27:11 -04:00
VictorSanh
5f07d8f11a
prepare release
2019-10-03 10:27:11 -04:00
VictorSanh
35071007cb
incoming release 🔥 update links to arxiv preprint
2019-10-03 10:27:11 -04:00
VictorSanh
f1f23ad171
fix buf in convert_pt_chkpt_to_tf2
2019-10-03 10:27:11 -04:00
VictorSanh
2a91f6071f
upddate README - TODO updadte link to paper
2019-10-03 10:27:11 -04:00
VictorSanh
c51e533a5f
update train.py
2019-10-03 10:27:11 -04:00
VictorSanh
a76c3f9cb0
update requirements
2019-10-03 10:27:11 -04:00
VictorSanh
bb9c5ead54
update distiller
2019-10-03 10:27:11 -04:00
VictorSanh
a12ab0a8db
update binarized_data
2019-10-03 10:27:11 -04:00
VictorSanh
4d6dfbd376
update extract
2019-10-03 10:27:11 -04:00
VictorSanh
23edebc079
update extract_distilbert
2019-10-03 10:27:11 -04:00
VictorSanh
cbfcfce205
update token_counts
2019-10-03 10:27:11 -04:00
VictorSanh
19e4ebbe3f
grouped_batch_sampler
2019-10-03 10:27:11 -04:00
VictorSanh
594202a934
lm_seqs_dataset
2019-10-03 10:27:11 -04:00
VictorSanh
38084507c4
add distillation_configs
2019-10-03 10:27:11 -04:00
Simon Layton
9ffda216ec
Fix missed head transpose
2019-10-03 09:23:16 -04:00
erenup
b5d73976ad
Revert "fixing for roberta tokenizer decoding"
...
This reverts commit 22e7c4edaf
.
2019-10-03 20:48:17 +08:00
erenup
22e7c4edaf
fixing for roberta tokenizer decoding
2019-10-03 18:33:53 +08:00
Brian Ma
2195c0d5f9
Evaluation result.txt path changing #1286
2019-10-03 12:49:12 +08:00
LysandreJik
ebb32261b1
fix #1401
2019-10-02 17:52:56 -04:00
Simon Layton
d51b589404
Re-order attention head outputs for better perf
...
Significant performance boost over the original orderings
on an already somewhat optimised branch this gave me > 2x end-to-end
throughput on a squad xlnet fine-tuning task (batch 8, seq-length 612,
fp16)
2019-10-02 12:18:21 -04:00
Santiago Castro
63ed224b7c
initialy -> initially
2019-10-02 15:04:18 +00:00
danai-antoniou
a95158518d
Moved duplicate token check
2019-10-02 07:44:15 +01:00
danai-antoniou
d73957899a
Merge branch 'master' of https://github.com/danai-antoniou/pytorch-transformers into add-duplicate-tokens-error
2019-10-02 07:38:50 +01:00
Dima Veselov
cd69bc9c87
Fixed typo in docs README
2019-10-02 03:21:55 +03:00
thomwolf
391db836ab
fix #1260 - remove special logic for decoding pairs of sequence
2019-10-01 19:09:13 -04:00
Thomas Wolf
963529e29b
Merge pull request #1288 from echan00/master
...
Typo with LM Fine tuning script
2019-10-01 18:46:07 -04:00
thomwolf
f7978f70ec
use format instead of f-strings
2019-10-01 18:45:38 -04:00
Thomas Wolf
1e4a191366
Merge pull request #1284 from slayton58/pooler_end_logits_fp16_fix
...
Fix fp16 masking in PoolerEndLogits
2019-10-01 18:40:22 -04:00
thomwolf
c50783e388
Merge branch 'pooler_end_logits_fp16_fix' of https://github.com/slayton58/pytorch-transformers into pr/1284
2019-10-01 18:17:48 -04:00
DenysNahurnyi
6971556ab8
Fix syntax typo in README.md
2019-10-01 14:59:31 -04:00
Julien Chaumond
b350662955
overflowing_tokens do not really make sense here, let's just return a number
...
Co-Authored-By: Lysandre Debut <lysandre.debut@reseau.eseo.fr>
2019-09-30 16:37:09 -04:00
Julien Chaumond
f5bcde0b2f
[multiple-choice] Simplify and use tokenizer.encode_plus
2019-09-30 16:04:55 -04:00
Santosh Gupta
5c3b32d44d
Update README.md
...
Lines 183 - 200, fixed indentation. Line 198, replaced `tokenizer_class` with `BertTokenizer`, since `tokenizer_class` is not defined in the loop it belongs to.
2019-09-30 18:48:01 +00:00
VictorSanh
2dc8cb8734
fix unknown imports (*ForMultipleChoice) in run_multiple_choice
2019-09-29 19:51:01 -04:00
Tim Yagan
0a4ed7192e
Fixed critical css font-family issues
...
Fixed critical css font-family issues to ensure compatibility with multiple webbrowsers
2019-09-29 13:51:01 +02:00
Thomas Wolf
ae50ad91ea
Merge pull request #1362 from FeiWang96/doc
...
fix link
2019-09-28 10:26:42 +02:00