Patrick von Platen
9c83b96e62
[Tests] Add Common Test for Training + Fix a couple of bugs ( #8415 )
...
* add training tests
* correct longformer
* fix docs
* fix some tests
* fix some more train tests
* remove ipdb
* fix multiple edge case model training
* fix funnel and prophetnet
* clean gpt models
* undo renaming of albert
2020-11-09 18:24:41 +01:00
Lysandre
a75c64d80c
Black 20 release
2020-08-26 17:20:22 +02:00
Patrick von Platen
505f2d749e
[Tests] fix attention masks in Tests ( #6621 )
...
* fix distilbert
* fix typo
2020-08-20 13:23:47 -04:00
Stas Bekman
5deed37f9f
cleanup torch unittests ( #6196 )
...
* improve unit tests
this is a sample of one test according to the request in https://github.com/huggingface/transformers/issues/5973
before I apply it to the rest
* batch 1
* batch 2
* batch 3
* batch 4
* batch 5
* style
* non-tf template
* last deletion of check_loss_output
2020-08-04 02:42:56 -04:00
Sylvain Gugger
d951c14ae4
Model output test ( #6155 )
...
* Use return_dict=True in all tests
* Formatting
2020-07-31 09:44:37 -04:00
Quentin Lhoest
4fedc1256c
Fix tests imports dpr ( #5576 )
...
* fix test imports
* fix max_length
* style
* fix tests
2020-07-07 16:35:12 +02:00
Quentin Lhoest
fbd8792195
Add DPR model ( #5279 )
...
* beginning of dpr modeling
* wip
* implement forward
* remove biencoder + better init weights
* export dpr model to embed model for nlp lib
* add new api
* remove old code
* make style
* fix dumb typo
* don't load bert weights
* docs
* docs
* style
* move the `k` parameter
* fix init_weights
* add pretrained configs
* minor
* update config names
* style
* better config
* style
* clean code based on PR comments
* change Dpr to DPR
* fix config
* switch encoder config to a dict
* style
* inheritance -> composition
* add messages in assert startements
* add dpr reader tokenizer
* one tokenizer per model
* fix base_model_prefix
* fix imports
* typo
* add convert script
* docs
* change tokenizers conf names
* style
* change tokenizers conf names
* minor
* minor
* fix wrong names
* minor
* remove unused convert functions
* rename convert script
* use return_tensors in tokenizers
* remove n_questions dim
* move generate logic to tokenizer
* style
* add docs
* docs
* quality
* docs
* add tests
* style
* add tokenization tests
* DPR full tests
* Stay true to the attention mask building
* update docs
* missing param in bert input docs
* docs
* style
Co-authored-by: Lysandre <lysandre.debut@reseau.eseo.fr>
2020-07-07 08:56:12 -04:00