Commit Graph

304 Commits

Author SHA1 Message Date
Lysandre Debut
88efc65bac
Merge pull request #964 from huggingface/RoBERTa
RoBERTa: model conversion, inference, tests 🔥
2019-08-15 11:11:10 -04:00
Julien Chaumond
c4ef103447 [RoBERTa] First 4 authors
cf. https://github.com/huggingface/pytorch-transformers/pull/964#discussion_r313574354

Co-Authored-By: Myle Ott <myleott@fb.com>
2019-08-14 12:31:09 -04:00
carefree0910
a7b4cfe919 Update README.md
I assume that it should test the `re-load` functionality after testing the `save` functionality, however I'm also surprised that nobody points this out after such a long time, so maybe I've misunderstood the purpose. This PR is just in case :)
2019-08-12 09:53:05 -04:00
LysandreJik
d2cc6b101e Merge branch 'master' into RoBERTa 2019-08-08 09:42:05 -04:00
Christopher Goh
a6f412da01 Fixed typo in migration guide 2019-08-07 02:19:14 +08:00
Thomas Wolf
d43dc48b34
Merge branch 'master' into auto_models 2019-08-05 19:17:35 +02:00
thomwolf
7223886dc9 fix #944 2019-08-05 17:16:56 +02:00
thomwolf
58830807d1 inidicate we only support pytorch 1.0.0+ now 2019-08-05 14:38:59 +02:00
thomwolf
328afb7097 cleaning up tokenizer tests structure (at last) - last remaining ppb refs 2019-08-05 14:08:56 +02:00
Julien Chaumond
05c083520a [RoBERTa] model conversion, inference, tests 🔥 2019-08-04 21:39:21 -04:00
thomwolf
009273dbdd big doc update [WIP] 2019-08-04 12:14:57 +02:00
Julien Chaumond
44dd941efb link to swift-coreml-transformers 2019-08-01 09:50:30 -04:00
Anthony MOI
f2a3eb987e
Fix small typos 2019-07-31 11:05:06 -04:00
Pierric Cistac
97091acb8c
Small spelling fix 2019-07-31 10:37:56 -04:00
Grégory Châtel
769bb643ce Fixing a broken link. 2019-07-31 10:22:41 -04:00
Thomas Wolf
fec76a481d
Update readme 2019-07-23 16:05:29 +02:00
thomwolf
ba52fe69d5 update breaking change section regarding from_pretrained keyword arguments 2019-07-23 15:10:02 +02:00
rish-16
2f869dc665 Fixed typo 2019-07-21 11:05:36 -04:00
Thomas Wolf
dbecfcf321
Merge pull request #815 from praateekmahajan/update-readme-link
Update Readme link for Fine Tune/Usage section
2019-07-18 18:30:32 +02:00
Peiqin Lin
acc48a0cc9 typos 2019-07-18 09:54:04 -04:00
Praateek Mahajan
0d46b17553
Update Readme
Incorrect link for `Quick tour: Fine-tuning/usage scripts`
2019-07-17 22:50:10 -07:00
thomwolf
c5b3d86a91 Merge branch 'master' of https://github.com/huggingface/pytorch-pretrained-BERT 2019-07-16 21:21:05 +02:00
thomwolf
6b70760204 typos 2019-07-16 21:21:03 +02:00
Thomas Wolf
b33a385091
update readme 2019-07-16 16:18:37 +02:00
thomwolf
6a72d9aa52 updated examples in readme 2019-07-16 16:09:29 +02:00
thomwolf
b59043bf8f update readme 2019-07-16 16:03:48 +02:00
thomwolf
edc79acb3b simpler quick tour 2019-07-16 16:02:32 +02:00
thomwolf
5c82d3488f indicate default evaluation in breaking changes 2019-07-16 15:45:58 +02:00
thomwolf
4acaa65068 model in evaluation mode by default after from_pretrained 2019-07-16 15:41:57 +02:00
thomwolf
1849aa7d39 update readme and pretrained model weight files 2019-07-16 15:11:29 +02:00
thomwolf
43e0e8fa04 updates to readme and doc 2019-07-16 13:56:47 +02:00
thomwolf
352e3ff998 added migration guide to readme 2019-07-16 09:03:49 +02:00
thomwolf
8ad7e5b4f2 indeed 2019-07-16 00:29:15 +02:00
thomwolf
064d0a0b76 update readme 2019-07-16 00:21:33 +02:00
thomwolf
3b8b0e01bb update readme 2019-07-16 00:12:55 +02:00
thomwolf
2397f958f9 updating examples and doc 2019-07-14 23:20:10 +02:00
thomwolf
6135de2fa3 readme update 2019-07-11 15:39:49 +02:00
thomwolf
e468192e2f Merge branch 'pytorch-transformers' into xlnet 2019-07-09 17:05:37 +02:00
LysandreJik
ab30651802 Hugging Face theme. 2019-07-08 16:05:26 -04:00
thomwolf
eb91f6437e update readme and setup 2019-07-05 12:30:15 +02:00
thomwolf
0231ba291e circle-ci 2019-07-05 11:59:04 +02:00
thomwolf
0bab55d5d5 [BIG] name change 2019-07-05 11:55:36 +02:00
thomwolf
93e9971c54 fix tests 2019-06-26 10:02:45 +02:00
thomwolf
e55d4c4ede various updates to conversion, models and examples 2019-06-26 00:57:53 +02:00
thomwolf
603c513b35 update main conversion script and readme 2019-06-25 10:45:07 +02:00
thomwolf
62d78aa37e updating GLUE utils for compatibility with XLNet 2019-06-24 14:36:11 +02:00
thomwolf
c304593d8f BERTology details in readme 2019-06-20 10:05:06 +02:00
thomwolf
34d706a0e1 pruning in bertology 2019-06-19 15:25:49 +02:00
thomwolf
dc8e0019b7 updating examples 2019-06-19 13:23:20 +02:00
thomwolf
68ab9599ce small fix and updates to readme 2019-06-19 09:38:38 +02:00
thomwolf
4d8c4337ae test barrier in distrib training 2019-06-18 22:41:28 +02:00
thomwolf
15ebd67d4e cache in run_classifier + various fixes to the examples 2019-06-18 15:58:22 +02:00
thomwolf
d82e5deeb1 set find_unused_parameters=True in DDP 2019-06-18 12:13:14 +02:00
thomwolf
f964753090 explanation on the current location of the caching folder 2019-06-18 11:36:28 +02:00
thomwolf
382e2d1e50 spliting config and weight files for bert also 2019-06-18 10:37:16 +02:00
thomwolf
4447f270b2 updating hub 2019-06-17 16:21:28 +02:00
thomwolf
33d3db5c43 updating head masking, readme and docstrings 2019-06-17 15:51:28 +02:00
thomwolf
34858ae1d9 adding bert whole words, bertgerman and gpt-2 medium models, head masking 2019-06-17 11:02:39 +02:00
timoeller
16af9ff7b0 Add German Bert model to code, update readme 2019-06-14 17:42:46 +02:00
Colanim
1eba8b9d96
Fix link in README 2019-05-30 14:01:46 +09:00
lukovnikov
331a46ff04 - replaced OpenAIGPTAdam with OpenAIAdam in docs 2019-04-25 16:04:37 +02:00
lukovnikov
704037ad51 - updated docs for new LR API
- added some images for illustration
- updated comments in optimization
2019-04-25 15:59:39 +02:00
thomwolf
18a8a15f78 improving GPT2 tokenization and adding tests 2019-04-16 17:00:55 +02:00
thomwolf
1135f2384a clean up logger in examples for distributed case 2019-04-15 15:22:40 +02:00
thomwolf
cc43307023 update readme 2019-04-15 15:06:10 +02:00
thomwolf
60ea6c59d2 added best practices for serialization in README and examples 2019-04-15 15:00:33 +02:00
thomwolf
20577d8a7c add configuration serialization to readme 2019-04-15 14:21:41 +02:00
thomwolf
b17963d82f update readme 2019-04-15 13:44:30 +02:00
Weixin Wang
f26ce6992e
Fix links in README 2019-04-02 17:20:32 +08:00
Sepehr Sameni
b588ff362a
fix lm_finetuning's link 2019-03-29 12:39:24 +04:30
Thomas Wolf
694e2117f3
Merge pull request #388 from ananyahjha93/master
Added remaining GLUE tasks to 'run_classifier.py'
2019-03-28 09:06:53 +01:00
Thomas Wolf
bbff03fbfc
Merge pull request #394 from desireevl/master
Minor change in README
2019-03-27 12:03:00 +01:00
thomwolf
34561e61a5 update main readme also 2019-03-27 12:00:04 +01:00
Ananya Harsh Jha
f471979167 added GLUE dev set results and details on how to run GLUE tasks 2019-03-21 15:38:30 -04:00
Desiree Vogt-Lee
d52f914e24
weigths to weights 2019-03-21 15:02:59 +10:00
Junjie Qian
d648a02203 Correct line number in README for classes 2019-03-08 16:28:03 -08:00
thomwolf
7cc35c3104 fix openai gpt example and updating readme 2019-03-06 11:43:21 +01:00
thomwolf
906b638efa updating readme 2019-03-06 10:24:19 +01:00
John Hewitt
e14c6b52e3 add BertTokenizer flag to skip basic tokenization 2019-02-26 20:11:24 -08:00
Joel Grus
8722e9eb3b finish updating docstrings 2019-02-23 06:31:59 -08:00
Stanislas Polu
ff22b3acc0 Few small nits in GPT-2's code examples 2019-02-21 09:15:27 +00:00
Tong Guo
09efcece75
Update README.md 2019-02-21 11:25:33 +08:00
Tony Lin
5b0e0b61f0
fix typo in readme 2019-02-19 20:34:18 +08:00
Davide Fiocco
0ae8eece55
MInor README typos corrected 2019-02-18 21:28:28 +01:00
sam-qordoba
1cb9c76ec5
Fix typo in GPT2Model code sample
Typo prevented code from running
2019-02-18 09:27:26 -08:00
Thomas Wolf
a25d056b7a
update readme 2019-02-18 15:30:11 +01:00
Thomas Wolf
517d7c8624
update readme 2019-02-18 14:39:55 +01:00
Thomas Wolf
ada22a1c9e
more details in GPT-2 usage example 2019-02-18 14:37:41 +01:00
Thomas Wolf
522733f6cb
readme typo fixes 2019-02-18 14:32:10 +01:00
thomwolf
d44db1145c update readme 2019-02-18 11:12:09 +01:00
Thomas Wolf
0e774e57a6
Update readme
Adding details on how to extract a full list of hidden states for the Transformer-XL
2019-02-14 08:39:58 +01:00
Thomas Wolf
4e56da38d9
Merge pull request #268 from wangxiaodiu/master
fixed a minor bug in README.md
2019-02-13 10:19:25 +01:00
thomwolf
67376c02e2 update readme for tokenizers 2019-02-13 10:11:11 +01:00
Liang Niu
e1b3cfb504 fixed a minor bug in README.md 2019-02-12 15:54:23 +04:00
Thomas Wolf
3c33499f87
fix typo in readme 2019-02-12 10:22:54 +01:00
thomwolf
1e71f11dec Release: 0.5.0 2019-02-11 14:16:27 +01:00
thomwolf
eebc8abbe2 clarify and unify model saving logic in examples 2019-02-11 14:04:19 +01:00
thomwolf
81c7e3ec9f fix typo in readme 2019-02-11 13:37:12 +01:00
thomwolf
884ca81d87 transposing the inputs of Transformer-XL to have a unified interface 2019-02-11 13:19:59 +01:00
thomwolf
32fea876bb add distant debugging to run_transfo_xl 2019-02-11 12:53:32 +01:00
thomwolf
b31ba23913 cuda on in the examples by default 2019-02-11 12:15:43 +01:00
thomwolf
2071a9b86e fix python 2.7 imports 2019-02-11 10:35:36 +01:00
thomwolf
b514a60c36 added tests for OpenAI GPT and Transformer-XL tokenizers 2019-02-11 10:17:16 +01:00
thomwolf
9f9909ea2f update readme 2019-02-09 16:59:21 +01:00
thomwolf
0c1a6f9b1d update readme 2019-02-08 22:32:25 +01:00
thomwolf
009b581316 updated readme 2019-02-07 23:15:05 +01:00
thomwolf
f99f2fb661 docstrings 2019-02-07 17:07:22 +01:00
Thomas Wolf
848aae49e1
Merge branch 'master' into python_2 2019-02-06 00:13:20 +01:00
thomwolf
ba37ddc5ce fix run_lm_modeling example command line 2019-02-06 00:07:08 +01:00
Girishkumar
0dd2b750ca
Minor update in README
Update links to classes in `modeling.py`
2019-01-30 23:49:15 +05:30
thomwolf
3a848111e6 update config, docstrings and readme to switch to seperated tokens and position embeddings 2019-01-29 11:00:11 +01:00
Davide Fiocco
35115eaf93
(very) minor update to README 2019-01-16 21:05:24 +01:00
nhatchan
8edc898f63 Fix documentation (missing backslashes)
This PR adds missing backslashes in LM Fine-tuning subsection in README.md.
2019-01-13 21:23:19 +09:00
thomwolf
e5c78c6684 update readme and few typos 2019-01-10 01:40:00 +01:00
thomwolf
fa5222c296 update readme 2019-01-10 01:25:28 +01:00
Thomas Wolf
c18bdb4433
Merge pull request #124 from deepset-ai/master
Add example for fine tuning BERT language model
2019-01-07 12:03:51 +01:00
Julien Chaumond
8da280ebbe Setup CI 2018-12-20 16:33:39 -05:00
tholor
e5fc98c542 add exemplary training data. update to nvidia apex. refactor 'item -> line in doc' mapping. add warning for unknown word. 2018-12-20 18:30:52 +01:00
tholor
67f4dd56a3 update readme for run_lm_finetuning 2018-12-19 09:22:37 +01:00
Julien Chaumond
d57763f582 Fix typos 2018-12-18 19:23:22 -05:00
Thomas Wolf
786cc41299
Typos in readme 2018-12-17 09:22:18 +01:00
Daniel Khashabi
8b1b93947f
Minor fix. 2018-12-14 14:10:36 -05:00
Thomas Wolf
8809eb6c93
update readme with information on NVIDIA's apex 2018-12-14 16:59:39 +01:00
thomwolf
d821358884 update readme 2018-12-14 15:15:17 +01:00
thomwolf
087798b7fa fix reloading model for evaluation in examples 2018-12-13 14:48:12 +01:00
thomwolf
0f544625f4 fix swag example for work with apex 2018-12-13 13:35:59 +01:00
thomwolf
4946c2c500 run_swag example in readme 2018-12-13 13:02:07 +01:00
Thomas Wolf
91aab2a6d3
Merge pull request #116 from FDecaYed/deyuf/fp16_with_apex
Change to use apex for better fp16 and multi-gpu support
2018-12-13 12:32:37 +01:00
Thomas Wolf
ffe9075f48
Merge pull request #96 from rodgzilla/multiple-choice-code
BertForMultipleChoice and Swag dataset example.
2018-12-13 12:05:11 +01:00
Grégory Châtel
dcb50eaa4b Swag example readme section update with gradient accumulation run. 2018-12-12 18:17:46 +01:00
Deyu Fu
c8ea286048 change to apex for better fp16 and multi-gpu support 2018-12-11 17:13:58 -08:00
Thomas Wolf
a3a3180c86
Bump up requirements to Python 3.6 2018-12-11 11:29:45 +01:00
Grégory Châtel
0876b77f7f Change to the README file to add SWAG results. 2018-12-10 15:34:19 +01:00
Davide Fiocco
c9f67e037c
Adding --do_lower_case for all uncased BERTs
I had missed those, it should make sense to use them
2018-12-07 20:40:56 +01:00
Grégory Châtel
150f3cd9fa Few typos in README.md 2018-12-06 19:22:07 +01:00
Grégory Châtel
4fa7892d64 Wrong line number link to modeling file. 2018-12-06 19:18:29 +01:00
Grégory Châtel
6a26e19ea3 Updating README.md with SWAG example informations. 2018-12-06 19:15:08 +01:00
Grégory Châtel
0a7c8bdcac Fixing badly formatted links. 2018-12-04 13:43:56 +01:00
Grégory Châtel
3113e967db Adding links to examples files. 2018-12-04 13:40:38 +01:00
Davide Fiocco
8a8aa59d8c
Update finetuning example adding --do_lower_case
Should be consistent with the fact that an uncased model is used
2018-12-01 01:00:05 +01:00
thomwolf
f9f3bdd60b update readme 2018-11-30 23:05:18 +01:00
thomwolf
52ff0590ff tup => tpu 2018-11-30 23:01:10 +01:00
thomwolf
296f006132 added BertForTokenClassification model 2018-11-30 13:56:53 +01:00
thomwolf
298107fed7 Added new bert models 2018-11-30 13:56:02 +01:00
Davide Fiocco
ec2c339b53
Updated quick-start example with BertForMaskedLM
As `convert_ids_to_tokens` returns a list, the code in the README currently throws an `AssertionError`, so I propose I quick fix.
2018-11-28 14:53:46 +01:00
thomwolf
05053d163c update cache_dir in readme and examples 2018-11-26 10:45:13 +01:00
thomwolf
029bdc0d50 fixing readme examples 2018-11-26 09:56:41 +01:00
Thomas Wolf
60e01ac427
fix link in readme 2018-11-21 12:08:30 +01:00
Thomas Wolf
fd32ebed81
Merge pull request #42 from weiyumou/master
Fixed UnicodeDecodeError: 'ascii' codec can't decode byte 0xc2
2018-11-20 10:09:50 +01:00
thomwolf
eed255a58d fixing CLI typo in readme 2018-11-20 10:02:57 +01:00