thomwolf
34d706a0e1
pruning in bertology
2019-06-19 15:25:49 +02:00
thomwolf
dc8e0019b7
updating examples
2019-06-19 13:23:20 +02:00
thomwolf
68ab9599ce
small fix and updates to readme
2019-06-19 09:38:38 +02:00
thomwolf
f7e2ac01ea
update barrier
2019-06-18 22:43:35 +02:00
thomwolf
4d8c4337ae
test barrier in distrib training
2019-06-18 22:41:28 +02:00
thomwolf
3359955622
updating run_classif
2019-06-18 22:23:10 +02:00
thomwolf
29b7b30eaa
updating evaluation on a single gpu
2019-06-18 22:20:21 +02:00
thomwolf
7d2001aa44
overwrite_output_dir
2019-06-18 22:13:30 +02:00
thomwolf
16a1f338c4
fixing
2019-06-18 17:06:31 +02:00
thomwolf
92e0ad5aba
no numpy
2019-06-18 17:00:52 +02:00
thomwolf
4e6edc3274
hop
2019-06-18 16:57:15 +02:00
thomwolf
f55b60b9ee
fixing again
2019-06-18 16:56:52 +02:00
thomwolf
8bd9118294
quick fix
2019-06-18 16:54:41 +02:00
thomwolf
3e847449ad
fix out_label_ids
2019-06-18 16:53:31 +02:00
thomwolf
aad3a54e9c
fix paths
2019-06-18 16:48:04 +02:00
thomwolf
40dbda6871
updating classification example
2019-06-18 16:45:52 +02:00
thomwolf
7388c83b60
update run_classifier for distributed eval
2019-06-18 16:32:49 +02:00
thomwolf
9727723243
fix pickle
2019-06-18 16:02:42 +02:00
thomwolf
9710b68dbc
fix pickles
2019-06-18 16:01:15 +02:00
thomwolf
15ebd67d4e
cache in run_classifier + various fixes to the examples
2019-06-18 15:58:22 +02:00
samuelbroscheit
94247ad6cb
Make num_train_optimization_steps int
2019-05-13 12:38:22 +02:00
samuel.broscheit
49a77ac16f
Clean up a little bit
2019-05-12 00:31:10 +02:00
samuel.broscheit
3bf3f9596f
Fixing the issues reported in https://github.com/huggingface/pytorch-pretrained-BERT/issues/556
...
Reason for issue was that optimzation steps where computed from example size, which is different from actual size of dataloader when an example is chunked into multiple instances.
Solution in this pull request is to compute num_optimization_steps directly from len(data_loader).
2019-05-12 00:13:45 +02:00
burcturkoglu
00c7fd2b79
Division to num_train_optimizer of global_step in lr_this_step is removed.
2019-05-09 10:57:03 +03:00
burcturkoglu
fa37b4da77
Merge branch 'master' of https://github.com/huggingface/pytorch-pretrained-BERT
2019-05-09 10:55:24 +03:00
burcturkoglu
5289b4b9e0
Division to num_train_optimizer of global_step in lr_this_step is removed.
2019-05-09 10:51:38 +03:00
Thomas Wolf
0198399d84
Merge pull request #570 from MottoX/fix-1
...
Create optimizer only when args.do_train is True
2019-05-08 16:07:50 +02:00
MottoX
18c8aef9d3
Fix documentation typo
2019-05-02 19:23:36 +08:00
MottoX
74dbba64bc
Prepare optimizer only when args.do_train is True
2019-05-02 19:09:29 +08:00
Thomas Wolf
2dee86319d
Merge pull request #527 from Mathieu-Prouveur/fix_value_training_loss
...
Update example files so that tr_loss is not affected by args.gradient…
2019-04-30 11:12:55 +02:00
Mathieu Prouveur
87b9ec3843
Fix tr_loss rescaling factor using global_step
2019-04-29 12:58:29 +02:00
Mathieu Prouveur
ed8fad7390
Update example files so that tr_loss is not affected by args.gradient_accumulation_step
2019-04-24 14:07:00 +02:00
thomwolf
d94c6b0144
fix training schedules in examples to match new API
2019-04-23 11:17:06 +02:00
Thomas Wolf
3d78e226e6
Merge pull request #489 from huggingface/tokenization_serialization
...
Better serialization for Tokenizers and Configuration classes - Also fix #466
2019-04-16 08:49:54 +02:00
thomwolf
3571187ef6
fix saving models in distributed setting examples
2019-04-15 16:43:56 +02:00
thomwolf
1135f2384a
clean up logger in examples for distributed case
2019-04-15 15:22:40 +02:00
thomwolf
60ea6c59d2
added best practices for serialization in README and examples
2019-04-15 15:00:33 +02:00
thomwolf
179a2c2ff6
update example to work with new serialization semantic
2019-04-15 14:33:23 +02:00
Jie Yang
c49ce3c722
fix tsv read error in Windows
2019-04-11 15:40:19 -04:00
Weixin Wang
d07db28f52
Fix typo in example code
...
Modify 'unambigiously' to 'unambiguously'
2019-03-31 01:20:18 +08:00
Ananya Harsh Jha
e5b63fb542
Merge branch 'master' of https://github.com/ananyahjha93/pytorch-pretrained-BERT
...
pull current master to local
2019-03-17 08:30:13 -04:00
Ananya Harsh Jha
8a4e90ff40
corrected folder creation error for MNLI-MM, verified GLUE results
2019-03-17 08:16:50 -04:00
Ananya Harsh Jha
e0bf01d9a9
added hack for mismatched MNLI
2019-03-16 14:10:48 -04:00
Ananya Harsh Jha
4c721c6b6a
added eval time metrics for GLUE tasks
2019-03-15 23:21:24 -04:00
Ananya Harsh Jha
043c8781ef
added code for all glue task processors
2019-03-14 04:24:04 -04:00
Yongbo Wang
22a465a91f
Simplify code, delete redundancy line
...
delete redundancy line `if args.train`, simplify code.
2019-03-13 09:42:06 +08:00
thomwolf
994d86609b
fixing PYTORCH_PRETRAINED_BERT_CACHE use in examples
2019-03-06 10:21:24 +01:00
thomwolf
5c85fc3977
fix typo - logger info
2019-03-06 10:05:21 +01:00
John Lehmann
0f96d4b1f7
Run classifier processor for SST-2.
2019-03-05 13:38:28 -06:00
thomwolf
eebc8abbe2
clarify and unify model saving logic in examples
2019-02-11 14:04:19 +01:00