Commit Graph

15053 Commits

Author SHA1 Message Date
erenup
22e7c4edaf fixing for roberta tokenizer decoding 2019-10-03 18:33:53 +08:00
Brian Ma
2195c0d5f9 Evaluation result.txt path changing #1286 2019-10-03 12:49:12 +08:00
LysandreJik
ebb32261b1 fix #1401 2019-10-02 17:52:56 -04:00
Simon Layton
d51b589404 Re-order attention head outputs for better perf
Significant performance boost over the original orderings
on an already somewhat optimised branch this gave me > 2x end-to-end
throughput on a squad xlnet fine-tuning task (batch 8, seq-length 612,
fp16)
2019-10-02 12:18:21 -04:00
Santiago Castro
63ed224b7c initialy -> initially 2019-10-02 15:04:18 +00:00
danai-antoniou
a95158518d Moved duplicate token check 2019-10-02 07:44:15 +01:00
danai-antoniou
d73957899a Merge branch 'master' of https://github.com/danai-antoniou/pytorch-transformers into add-duplicate-tokens-error 2019-10-02 07:38:50 +01:00
Dima Veselov
cd69bc9c87
Fixed typo in docs README 2019-10-02 03:21:55 +03:00
thomwolf
391db836ab fix #1260 - remove special logic for decoding pairs of sequence 2019-10-01 19:09:13 -04:00
Thomas Wolf
963529e29b
Merge pull request #1288 from echan00/master
Typo with LM Fine tuning script
2019-10-01 18:46:07 -04:00
thomwolf
f7978f70ec use format instead of f-strings 2019-10-01 18:45:38 -04:00
Thomas Wolf
1e4a191366
Merge pull request #1284 from slayton58/pooler_end_logits_fp16_fix
Fix fp16 masking in PoolerEndLogits
2019-10-01 18:40:22 -04:00
thomwolf
c50783e388 Merge branch 'pooler_end_logits_fp16_fix' of https://github.com/slayton58/pytorch-transformers into pr/1284 2019-10-01 18:17:48 -04:00
DenysNahurnyi
6971556ab8 Fix syntax typo in README.md 2019-10-01 14:59:31 -04:00
Julien Chaumond
b350662955 overflowing_tokens do not really make sense here, let's just return a number
Co-Authored-By: Lysandre Debut <lysandre.debut@reseau.eseo.fr>
2019-09-30 16:37:09 -04:00
Julien Chaumond
f5bcde0b2f [multiple-choice] Simplify and use tokenizer.encode_plus 2019-09-30 16:04:55 -04:00
Santosh Gupta
5c3b32d44d Update README.md
Lines 183 - 200, fixed indentation. Line 198, replaced `tokenizer_class` with `BertTokenizer`, since `tokenizer_class` is not defined in the loop it belongs to.
2019-09-30 18:48:01 +00:00
VictorSanh
2dc8cb8734 fix unknown imports (*ForMultipleChoice) in run_multiple_choice 2019-09-29 19:51:01 -04:00
Tim Yagan
0a4ed7192e
Fixed critical css font-family issues
Fixed critical css font-family issues to ensure compatibility with multiple webbrowsers
2019-09-29 13:51:01 +02:00
Thomas Wolf
ae50ad91ea
Merge pull request #1362 from FeiWang96/doc
fix link
2019-09-28 10:26:42 +02:00
wangfei
60f791631b Fix link in readme 2019-09-28 16:20:17 +08:00
Ikuya Yamada
a6a6d9e638 fix padding_idx of RoBERTa model 2019-09-27 19:03:55 -04:00
Julien Chaumond
d8b641c839 6 -> 8 models 2019-09-27 17:22:01 -04:00
Julien Chaumond
c6acbdd50a Close #1304 2019-09-27 17:02:53 -04:00
Thomas Wolf
df7cd9e4e4
Merge pull request #1353 from wendingp/patch-1
Fix some typos
2019-09-27 23:00:34 +02:00
Thomas Wolf
6a17b3c51b
Merge pull request #1355 from agrinh/master
Fix tensorflow_dataset glue support
2019-09-27 22:59:54 +02:00
Thomas Wolf
04e9a6f512
Merge pull request #1359 from dennymarcels/patch-1
Update run_lm_finetuning.py
2019-09-27 22:58:19 +02:00
Denny
9478590630
Update run_lm_finetuning.py
The previous method, just as phrased, did not exist in the class.
2019-09-27 15:18:42 -03:00
Agrin Hilmkil
795b3e76ff Add docstring for processor method 2019-09-27 17:32:28 +02:00
Agrin Hilmkil
e31a472801 Fix tensorflow_dataset glue support
`glue_convert_examples_to_features` assumed that tensorflow_dataset
examples contains the features `'sentence1'` and `'sentence2'`. This
commit encapsulates the choice of features in the glue processor and
uses that to parse examples.
2019-09-27 17:16:02 +02:00
pj
4f2b6579bf
Fix some typos 2019-09-27 22:55:43 +08:00
Thomas Wolf
ca559826c4
Merge pull request #1349 from ogabrielluiz/master
Just some typos
2019-09-27 13:08:00 +02:00
Gabriel Luiz Freitas Almeida
d2de5b9d8c
Just some typos 2019-09-27 07:08:36 -03:00
Thomas Wolf
d83d295763
Merge pull request #1337 from mgrankin/fastdataset
faster dataset building
2019-09-27 10:35:12 +02:00
Thomas Wolf
f6de000305
Merge pull request #1346 from BramVanroy/documentation
Add small  note about the output of hidden states (closes #1332)
2019-09-27 10:30:07 +02:00
BramVanroy
15749bfc10 Add small note about the output of hidden states 2019-09-27 10:01:36 +02:00
thomwolf
da2e47ad15 clean up a little run_tf_glue 2019-09-27 09:41:15 +02:00
thomwolf
528c288fa9 clean up run_tf_glue 2019-09-27 09:40:29 +02:00
VictorSanh
702f589848 fix input in run_glue for distilbert 2019-09-27 00:20:14 -04:00
Julien Chaumond
22d2fded2c [docs] Fix doc auto-deploy
Co-Authored-By: Lysandre Debut <lysandre.debut@reseau.eseo.fr>
2019-09-26 18:22:45 -04:00
Julien Chaumond
fc9faa8a47 [docs] Doc tweaks
Co-Authored-By: Lysandre Debut <lysandre.debut@reseau.eseo.fr>
2019-09-26 18:19:51 -04:00
LysandreJik
ecfddc6034 Update RoBERTa and GPT-2 Tokenizer documentation (fix #1343) 2019-09-26 16:49:03 -04:00
LysandreJik
93f0c5fc72 Repository link in the documentation 2019-09-26 11:45:00 -04:00
thomwolf
6c3b131516 typo in readme/doc 2019-09-26 16:23:28 +02:00
thomwolf
f83b35b77d Merge branch 'master' of https://github.com/huggingface/pytorch-transformers 2019-09-26 16:14:23 +02:00
thomwolf
4e63c90720 update installation instructions in readme 2019-09-26 16:14:21 +02:00
LysandreJik
7e957237e4 [Doc] XLM + Torch in documentation 2019-09-26 10:08:56 -04:00
LysandreJik
302a4813a5 Doc building requirements [TF2] 2019-09-26 09:57:30 -04:00
mgrankin
f71a4577b8 faster dataset building 2019-09-26 16:53:13 +03:00
LysandreJik
a3e0dbba95 Doc building requirements [TF] 2019-09-26 09:51:14 -04:00