Thomas Wolf
0287d264e9
Merge pull request #1162 from huggingface/xlnet-bias
...
XLNet bias fix on resize embeddings (cf #1124 )
2019-09-02 23:14:04 +02:00
LysandreJik
7f522437bc
Updated documentation for LM finetuning script
2019-09-02 13:40:25 -04:00
LysandreJik
3fbf301bba
[CI] Updated resource size for python 3 tests
2019-09-02 12:35:14 -04:00
Julien Chaumond
2dcc5a1629
[doc] Add blurb about large-scale model downloads
...
cc @n1t0 @lysandrejik @thomwolf
2019-09-02 12:27:11 -04:00
Thomas Wolf
7b0c99add9
Merge pull request #1174 from huggingface/fix_byte_level_added_tokens
...
Fix byte-level BPE decoding error when using added tokens
2019-09-02 09:01:16 +02:00
LysandreJik
31d3373bc9
Appends space before special token
2019-09-01 21:07:00 -04:00
thomwolf
fede4ef45d
fixing #1133
2019-09-02 02:27:39 +02:00
Thomas Wolf
b6cd856b08
Merge pull request #1164 from stefan-it/master
...
distillation: fix ModuleNotFoundError error in token counts script
2019-09-02 02:00:07 +02:00
Thomas Wolf
ff7368eb6b
Merge pull request #1077 from huggingface/pruning-save-and-load
...
Pruning changes so that deleted heads are kept on save/load
2019-09-01 09:42:15 +02:00
LysandreJik
6ae0bb5291
XLM 100 different URLs
2019-08-31 14:46:31 -04:00
LysandreJik
819b468f70
Fixed XLM model url
2019-08-31 14:40:51 -04:00
LysandreJik
58b59a0c31
Random seed is accessible anywhere within the common tests
2019-08-31 13:17:08 -04:00
Stefan Schweter
a1c34bd286
distillation: fix ModuleNotFoundError error in token counts script
2019-08-31 12:21:38 +02:00
LysandreJik
ea86bef545
Check for None
2019-08-31 00:56:22 -04:00
LysandreJik
e0f867a9ba
XLNet bias fix on resize embeddings (cf #1124 )
2019-08-31 00:50:59 -04:00
LysandreJik
11600edc6e
Rebase on master + DistilBERT head pruning patch
2019-08-31 00:37:41 -04:00
LysandreJik
b6992b7b47
Applied patch to OpenAI GPT, RoBERTa, TransfoL, XLM and XLNet
2019-08-31 00:33:50 -04:00
thomwolf
bdb4409ed8
updated pruning logic with sets - Bert and GPT-2
2019-08-31 00:33:50 -04:00
LysandreJik
0c8e823b03
Added patch to remaining models
2019-08-31 00:33:50 -04:00
LysandreJik
0cd283522a
Attempt to fix head index
2019-08-31 00:33:50 -04:00
LysandreJik
c85b5db61a
Conditional append/init + fixed warning
2019-08-31 00:33:50 -04:00
LysandreJik
5c2b94c82a
Changed string so that Circle CI accepts the warning
2019-08-31 00:33:50 -04:00
LysandreJik
87747518e9
Blocks deletion from already deleted heads. Necessary integration test.
...
Now raises a warning when a head to be deleted already has been deleted. An integration test verifying the total pipeline (-> from config -> save model -> load model -> additional head pruning) has been added.
2019-08-31 00:33:50 -04:00
LysandreJik
719cb3738d
Pruning for GPT and GPT-2
2019-08-31 00:33:50 -04:00
LysandreJik
fc1fbae45d
XLM can be pruned
2019-08-31 00:33:50 -04:00
Lysandre
42e00cf9e1
Pruning saved to configuration first try
2019-08-31 00:33:50 -04:00
LysandreJik
d7a4c3252e
Fixed filename
2019-08-31 00:08:56 -04:00
LysandreJik
7f006cdd87
Set seed for head_masking test
2019-08-30 23:58:49 -04:00
Julien Chaumond
0fd0b674e6
[ci] legible output [skip ci]
2019-08-30 20:36:26 -04:00
Julien Chaumond
b65a994f59
[ci] decrease parallelism to increase success prob
2019-08-30 20:33:16 -04:00
Julien Chaumond
1d438f15b3
[XLNet] Use pytorch's layernorm like in BERT
...
See #1089
cc @thomwolf @lysandrejik
Also @dhpollack
2019-08-30 20:20:15 -04:00
Julien Chaumond
574c5b3a72
[RoBERTa] LayerNorm's eps is not a nn.Parameter so there's no point setting it on the model
...
Instead we correctly store it on the config
(regenerating the hosted config files)
cc @lysandrejik
2019-08-30 20:09:24 -04:00
LysandreJik
09363f2a8b
Fix documentation index
2019-08-30 19:48:32 -04:00
Thomas Wolf
51e980ce36
Merge pull request #1155 from anhnt170489/apex_fp16
...
Update apex fp16 implementation
2019-08-30 23:29:11 +02:00
Thomas Wolf
206c35e9a4
Merge pull request #1154 from ziliwang/master
...
fix: hard coding for max number
2019-08-30 23:23:08 +02:00
Thomas Wolf
f3d18c71ec
Merge pull request #1152 from epwalsh/fix-special-tokens
...
fix adding special tokens
2019-08-30 23:21:59 +02:00
Thomas Wolf
d483cd8e46
Merge pull request #1074 from huggingface/improved_testing
...
Shortcut to special tokens' ids - fix GPT2 & RoBERTa tokenizers - improved testing for GPT/GPT-2
2019-08-30 23:18:58 +02:00
Thomas Wolf
d2f21f08f5
Merge pull request #1092 from shijie-wu/xlm-tokenization
...
Added cleaned configuration properties for tokenizer with serialization - improve tokenization of XLM
2019-08-30 23:15:40 +02:00
Thomas Wolf
12b9cc9e26
Merge pull request #1110 from huggingface/automodels
...
Torch.hub now based on AutoModels - Updating AutoModels with AutoModelWithLMHead, Sequence Classification and Question Answering
2019-08-30 23:08:57 +02:00
thomwolf
bfe93a5a21
fix distilbert in auto tokenizer
2019-08-30 22:43:26 +02:00
thomwolf
256086bc69
clean up and simplify hubconf
2019-08-30 22:34:23 +02:00
thomwolf
80aa87d9a3
fix distilbert tokenizer
2019-08-30 22:24:23 +02:00
thomwolf
455a4c842c
add distilbert tokenizer
2019-08-30 22:20:51 +02:00
thomwolf
7a1f174a9d
update names of torch.hub to simpler names - update docstring
2019-08-30 22:20:44 +02:00
thomwolf
c665e0fcfe
Merge branch 'automodels' of https://github.com/huggingface/pytorch-transformers into automodels
2019-08-30 21:53:36 +02:00
LysandreJik
9b6e3b34d9
Docstrings
2019-08-30 14:09:02 -04:00
LysandreJik
dec8f4d6fd
Added DistilBERT models to all other AutoModels.
2019-08-30 13:52:18 -04:00
LysandreJik
bc29aa67a9
HubConf configuration
2019-08-30 12:48:55 -04:00
thomwolf
f35f612280
updating docstring for AutoModel
2019-08-30 12:48:55 -04:00
LysandreJik
7ca9653852
Pytorch Hub & AutoModels
2019-08-30 12:48:55 -04:00