Commit Graph

15053 Commits

Author SHA1 Message Date
Lysandre
b5d330d118 Fix #1784 2019-11-11 10:15:14 -05:00
eukaryote
90f6e73a35
Add DialoGPT support for Pytorch->TF 2019-11-09 16:46:19 +00:00
eukaryote
ef99852961
from_pretrained: convert DialoGPT format
DialoGPT checkpoints have "lm_head.decoder.weight" instead of "lm_head.weight". 

(see: https://www.reddit.com/r/MachineLearning/comments/dt5woy/p_dialogpt_state_of_the_art_conversational_model/f6vmwuy?utm_source=share&utm_medium=web2x)
2019-11-09 16:32:40 +00:00
Adrian Bauer
7a9aae1044 Fix run_bertology.py
Make imports and args.overwrite_cache match run_glue.py
2019-11-08 16:28:40 -05:00
thomwolf
268d4f2099 fix position biases + better tests 2019-11-08 16:41:55 +01:00
thomwolf
b4fcd59a5a add sentinels in tokenizer 2019-11-08 14:38:53 +01:00
thomwolf
15e53c4e87 maybe fix tests 2019-11-08 12:43:21 +01:00
thomwolf
f03c0c1423 adding models in readme and auto classes 2019-11-08 11:49:46 +01:00
thomwolf
4321c54125 fix tests 2019-11-08 11:49:32 +01:00
thomwolf
727a79b305 added TF2 model and tests - updated templates 2019-11-08 11:35:03 +01:00
Rémi Louf
cd286c2145 add condition around mask transformation 2019-11-08 11:31:16 +01:00
Rémi Louf
28d0ba35d7 only init encoder_attention_mask if stack is decoder
We currently initialize `encoder_attention_mask` when it is `None`,
whether the stack is that of an encoder or a decoder. Since this
may lead to bugs that are difficult to tracks down, I added a condition
that assesses whether the current stack is a decoder.
2019-11-08 11:22:19 +01:00
thomwolf
8fda532c3c fix python 2 sentencepiece tokenization 2019-11-07 17:09:50 +01:00
thomwolf
ba10065c4b update model, conversion script, tests and template 2019-11-07 15:55:36 +01:00
Diganta Misra
070dcf1c02
Added Mish Activation Function
Mish is a new activation function proposed here - https://arxiv.org/abs/1908.08681
It has seen some recent success and has been adopted in SpaCy, Thic, TensorFlow Addons and FastAI-dev. 
All benchmarks recorded till now (including against ReLU, Swish and GELU) is present in the repository - https://github.com/digantamisra98/Mish
Might be a good addition to experiment with especially in the Bert Model.
2019-11-07 03:45:43 +05:30
Julien Chaumond
1c542df7e5 Add RoBERTa-based GPT-2 Output Detector from OpenAI
converted from https://github.com/openai/gpt-2-output-dataset/tree/master/detector

Co-Authored-By: Lysandre Debut <lysandre.debut@reseau.eseo.fr>
Co-Authored-By: Jong Wook Kim <jongwook@nyu.edu>
Co-Authored-By: Jeff Wu <wuthefwasthat@gmail.com>
2019-11-06 16:26:31 -05:00
Julien Chaumond
2f3a421018 Fix other PyTorch models 2019-11-06 14:03:47 -05:00
Julien Chaumond
d5319793c4 Fix BERT 2019-11-06 14:03:47 -05:00
Julien Chaumond
27e015bd54 [tests] Flag to test on cuda 2019-11-06 14:03:47 -05:00
Julien Chaumond
13d9135fa5 [tests] get rid of warning
cf. https://docs.pytest.org/en/latest/example/simple.html
2019-11-06 14:03:47 -05:00
thomwolf
076a207935 adding tests and updating model 2019-11-06 11:52:50 +01:00
thomwolf
73f2c342f5 fixing template 2019-11-06 11:52:39 +01:00
thomwolf
3835e1e651 adding tokenizer 2019-11-06 11:52:29 +01:00
Julien Chaumond
f88c104d8f [run_tf_glue] Add comment for context 2019-11-05 19:56:43 -05:00
Julien Chaumond
30968d70af misc doc 2019-11-05 19:06:12 -05:00
Dom Hudson
de890ae67d Updating docblocks in optimizers.py 2019-11-05 17:31:29 -05:00
Lysandre
d7d36181fd GPT-2 XL 2019-11-05 13:31:58 -05:00
LysandreJik
151e4ab4e7 Fix CTRL past 2019-11-05 16:26:51 +00:00
thomwolf
88e5bef58f share position biases 2019-11-05 17:02:52 +01:00
thomwolf
568c0ffb7e adding T5 model 2019-11-05 16:40:29 +01:00
Julien Chaumond
7daacf00df
Merge pull request #1695 from huggingface/models_inputs_embeds
model forwards can take an inputs_embeds param
2019-11-05 09:55:28 -05:00
Clement
a44f112fb9
add authors for models 2019-11-05 08:48:26 -05:00
thomwolf
60a5babd57 adding files 2019-11-05 12:01:23 +01:00
Filip Povolny
124409d075 Make dummy inputs a property of TFPreTrainedModel. 2019-11-05 11:48:45 +01:00
Thomas Wolf
e99071f105
Merge pull request #1734 from orena1/patch-1
add progress bar to convert_examples_to_features
2019-11-05 11:34:20 +01:00
thomwolf
dfb61caf77 fix #1692 2019-11-05 11:25:13 +01:00
Thomas Wolf
ba973342e3
Merge pull request #1553 from WilliamTambellini/timeSquadInference
Add speed log to examples/run_squad.py
2019-11-05 11:13:12 +01:00
Filip Povolny
8df7dfd2a7 Make dummy inputs a local variable in TFPreTrainedModel. 2019-11-05 11:09:16 +01:00
Thomas Wolf
237fad339c
Merge pull request #1709 from oneraghavan/master
Fixing mode in evaluate during training
2019-11-05 10:55:33 +01:00
thomwolf
f1e4db2aa8 Fix #1686 2019-11-05 09:38:00 +01:00
Oren Amsalem
d7906165a3
add progress bar for convert_examples_to_features
It takes considerate amount of time (~10 min) to parse the examples to features, it is good to have a progress-bar to track this
2019-11-05 10:34:27 +02:00
Thomas Wolf
d2e2577dd3
Merge pull request #1723 from huggingface/fix-1623
Fix #1623
2019-11-05 08:36:30 +01:00
Julien Chaumond
00337e9687 [inputs_embeds] All PyTorch models 2019-11-05 00:39:18 +00:00
Julien Chaumond
9eddf44b7a docstring + check 2019-11-04 17:19:15 +00:00
Julien Chaumond
8e11de0e86 model forwards can take an inputs_embeds param 2019-11-04 16:56:26 +00:00
Lysandre
68f7064a3e Add model.train() line to ReadMe training example
Co-Authored-By: Santosh-Gupta <San.Gupta.ML@gmail.com>
2019-11-04 11:52:35 -05:00
thomwolf
8d6b9d717c fix #1532 and encode_plus 2019-11-04 17:07:51 +01:00
Thomas Wolf
c8f2712199
Merge pull request #1721 from huggingface/common_attributes
Add common getter and setter for input_embeddings & output_embeddings
2019-11-04 16:21:52 +01:00
thomwolf
89d6272898 Fix #1623 2019-11-04 16:21:12 +01:00
thomwolf
b340a910ed fix tests - flagged as slow all the tests downloading from AWS 2019-11-04 16:03:36 +01:00