Rémi Louf
4735c2af07
tweaks to the BeamSearch API
2019-12-09 20:37:55 -05:00
Rémi Louf
ba089c780b
share pretrained embeddings
2019-12-09 20:37:55 -05:00
Rémi Louf
9660ba1cbd
Add beam search
2019-12-09 20:37:55 -05:00
Rémi Louf
1c71ecc880
load the pretrained weights for encoder-decoder
...
We currently save the pretrained_weights of the encoder and decoder in
two separate directories `encoder` and `decoder`. However, for the
`from_pretrained` function to operate with automodels we need to
specify the type of model in the path to the weights.
The path to the encoder/decoder weights is handled by the
`PreTrainedEncoderDecoder` class in the `save_pretrained` function. Sice
there is no easy way to infer the type of model that was initialized for
the encoder and decoder we add a parameter `model_type` to the function.
This is not an ideal solution as it is error prone, and the model type
should be carried by the Model classes somehow.
This is a temporary fix that should be changed before merging.
2019-12-09 20:37:55 -05:00
Rémi Louf
07f4cd73f6
update function to add special tokens
...
Since I started my PR the `add_special_token_single_sequence` function
has been deprecated for another; I replaced it with the new function.
2019-12-09 20:37:55 -05:00
Pierric Cistac
5c877fe94a
fix albert links
2019-12-09 18:53:00 -05:00
Bilal Khan
79526f82f5
Remove unnecessary epoch variable
2019-12-09 16:24:35 -05:00
Bilal Khan
9626e0458c
Add functionality to continue training from last saved global_step
2019-12-09 16:24:35 -05:00
Bilal Khan
2d73591a18
Stop saving current epoch
2019-12-09 16:24:35 -05:00
Bilal Khan
0eb973b0d9
Use saved optimizer and scheduler states if available
2019-12-09 16:24:35 -05:00
Bilal Khan
a03fcf570d
Save tokenizer after each epoch to be able to resume training from a checkpoint
2019-12-09 16:24:35 -05:00
Bilal Khan
f71b1bb05a
Save optimizer state, scheduler state and current epoch
2019-12-09 16:24:35 -05:00
thomwolf
8e651f56b7
fix tf tests
2019-12-09 22:13:57 +01:00
thomwolf
808bb8da7e
fix transfo xl tests
2019-12-09 21:48:34 +01:00
thomwolf
b016dd16c9
fix tests on python 3.5
2019-12-09 21:38:07 +01:00
LysandreJik
2a4ef098d6
Add ALBERT and XLM to SQuAD script
2019-12-09 10:46:47 -05:00
Lysandre Debut
00c4e39581
Merge branch 'master' into squad-refactor
2019-12-09 10:41:15 -05:00
thomwolf
169fea6855
updating T5
2019-12-09 16:25:33 +01:00
Rémi Louf
3520be7824
create encoder attention mask from shape of hidden states
...
We currently create encoder attention masks (when they're not provided)
based on the shape of the inputs to the encoder. This is obviously
wrong; sequences can be of different lengths. We now create the encoder
attention mask based on the batch_size and sequence_length of the
encoder hidden states.
2019-12-09 11:19:45 +01:00
Aymeric Augustin
0cb163865a
Remove pytest dependency. ( #2093 )
2019-12-07 07:46:14 -05:00
Michael Watkins
2670b0d682
Fix bug which lowercases special tokens
2019-12-06 16:15:53 -05:00
Aymeric Augustin
35401fe50f
Remove dependency on pytest for running tests ( #2055 )
...
* Switch to plain unittest for skipping slow tests.
Add a RUN_SLOW environment variable for running them.
* Switch to plain unittest for PyTorch dependency.
* Switch to plain unittest for TensorFlow dependency.
* Avoid leaking open files in the test suite.
This prevents spurious warnings when running tests.
* Fix unicode warning on Python 2 when running tests.
The warning was:
UnicodeWarning: Unicode equal comparison failed to convert both arguments to Unicode - interpreting them as being unequal
* Support running PyTorch tests on a GPU.
Reverts 27e015bd
.
* Tests no longer require pytest.
* Make tests pass on cuda
2019-12-06 13:57:38 -05:00
Julien Chaumond
e4679cddce
[cli] Uploads: add progress bar ( #2078 )
...
* [cli] Uploads: add progress bar
see https://github.com/huggingface/transformers/pull/2044#discussion_r354057827 for context
* rename + documentation
* Add auto-referential comment
2019-12-06 11:56:23 -05:00
thomwolf
1d87b37d10
updating
2019-12-06 15:30:09 +01:00
Thomas Wolf
4cb9b60558
Merge pull request #2077 from patrickvonplaten/change_documentation_for_past_output_shape
...
corrected documentation for past tensor shape for ctrl and gpt2 model
2019-12-06 12:14:48 +01:00
Thomas Wolf
5482822a2b
Merge pull request #2046 from jplu/tf2-ner-example
...
Add NER TF2 example.
2019-12-06 12:12:22 +01:00
Thomas Wolf
fc1bb1f867
Merge pull request #2068 from huggingface/fix-2042
...
Nicer error message when Bert's input is missing batch size
2019-12-06 12:06:42 +01:00
Philipp Glock
21451ec6ba
handle string with only whitespaces as empty
2019-12-06 10:32:43 +01:00
Rémi Louf
f230d91b43
check the validity of links
...
We add a script and a CI workflow to check that all download links
present in the source code are valid.
2019-12-06 09:41:28 +01:00
patrickvonplaten
d0383e4daf
corrected documentation for past tensor shape for ctrl and gpt2 model
2019-12-06 01:24:22 +01:00
LysandreJik
e9217da5ff
Cleanup
...
Improve global visibility on the run_squad script, remove unused files and fixes related to XLNet.
2019-12-05 16:01:51 -05:00
LysandreJik
9ecd83dace
Patch evaluation for impossible values + cleanup
2019-12-05 14:44:57 -05:00
VictorSanh
35ff345fc9
update requirements
2019-12-05 12:07:04 -05:00
VictorSanh
552c44a9b1
release distilm-bert
2019-12-05 10:14:58 -05:00
Rosanne Liu
ee53de7aac
Pr for pplm ( #2060 )
...
* license
* changes
* ok
* Update paper link and commands to run
* pointer to uber repo
2019-12-05 09:20:07 -05:00
thomwolf
f8fb4335c9
clean up a little bit PT <=> TF conversion
2019-12-05 15:19:32 +01:00
Thomas Wolf
bebaa14039
Merge pull request #2045 from aaugustin/remove-dead-code
...
Remove dead code in tests.
2019-12-05 14:41:56 +01:00
thomwolf
18fb93530b
fixing #2042 - Nicer error message
2019-12-05 14:36:34 +01:00
thomwolf
2d5d86e037
fix #2031
2019-12-05 14:06:29 +01:00
Thomas Wolf
af077b15e2
Merge pull request #2065 from huggingface/fixing-camembert
...
Fixing camembert tokenization
2019-12-05 13:45:44 +01:00
thomwolf
3268ebd229
fix xlnet test
2019-12-05 13:35:29 +01:00
thomwolf
6c5297a423
Fixing camembert tokenization
2019-12-05 13:27:58 +01:00
Julien Plu
9200a759d7
Add few tests on the TF optimization file with some info in the documentation. Complete the README.
2019-12-05 12:56:43 +01:00
Thomas Wolf
1f179f095f
Merge pull request #2011 from AdityaSoni19031997/patch-1
...
typo fix on the docs as per Pytorch v1.1+
2019-12-05 12:39:04 +01:00
Thomas Wolf
1eaf44e713
Merge pull request #2007 from roskoN/xlnet_attention_fix
...
fixed XLNet attention output for both attention streams whenever target_mapping is provided
2019-12-05 12:32:39 +01:00
thomwolf
71e4693f08
fix #1968
2019-12-05 12:14:24 +01:00
Thomas Wolf
f9f395b21c
Merge pull request #1735 from ondewo/tf-do-not-use-gpu-on-import
...
Do not use GPU when importing transformers
2019-12-05 11:56:48 +01:00
thomwolf
75a97af6bc
fix #1450 - add doc
2019-12-05 11:26:55 +01:00
thomwolf
8b388827b5
fix #1920
2019-12-05 11:18:43 +01:00
Thomas Wolf
d425a4d60b
Merge pull request #1870 from alexzubiaga/xlnet-for-token-classification
...
XLNet for Token classification
2019-12-05 09:54:09 +01:00