Thomas Wolf
|
2fb8ddeeff
|
Merge pull request #392 from Rocketknight1/master
Add full language model fine-tuning
|
2019-03-27 12:02:36 +01:00 |
|
thomwolf
|
34561e61a5
|
update main readme also
|
2019-03-27 12:00:04 +01:00 |
|
thomwolf
|
361aff6de5
|
typos
|
2019-03-27 11:54:59 +01:00 |
|
thomwolf
|
cea8ba1d59
|
adjusted formating and some wording in the readme
|
2019-03-27 11:53:44 +01:00 |
|
Ikuya Yamada
|
0401317b23
|
Remove padding_idx from position_embeddings and token_type_embeddings
|
2019-03-26 21:56:35 +09:00 |
|
Matthew Carrigan
|
24e67fbf75
|
Minor README update
|
2019-03-25 12:33:30 +00:00 |
|
Matthew Carrigan
|
8d1d1ffde2
|
Corrected the displayed loss when gradient_accumulation_steps > 1
|
2019-03-25 12:15:19 +00:00 |
|
Catalin Voss
|
fda2f62395
|
Fix test failures due to old torch issue with non-contiguous view
|
2019-03-24 14:37:13 -07:00 |
|
Catalin Voss
|
0dd796e359
|
Also fix loss function issue with the double head models
|
2019-03-24 14:35:55 -07:00 |
|
Catalin Voss
|
472857c47f
|
Fix typo syntax err (sorry, c/p from my repo)
|
2019-03-24 14:14:49 -07:00 |
|
Catalin Voss
|
2e6f5ffb96
|
Fix GPT language model loss here as well
|
2019-03-24 14:14:44 -07:00 |
|
Catalin Voss
|
5938f31fa7
|
Fix c/p typo from my experiment code
|
2019-03-24 14:14:40 -07:00 |
|
Catalin Voss
|
7797d21b8d
|
Fix GPT2 language modeling loss computation
|
2019-03-24 14:14:35 -07:00 |
|
Ananya Harsh Jha
|
f471979167
|
added GLUE dev set results and details on how to run GLUE tasks
|
2019-03-21 15:38:30 -04:00 |
|
Matthew Carrigan
|
abb7d1ff6d
|
Added proper context management to ensure cleanup happens in the right
order.
|
2019-03-21 17:50:03 +00:00 |
|
Matthew Carrigan
|
06a30cfdf3
|
Added a --reduce_memory option to the training script to keep training
data on disc as a memmap rather than in memory
|
2019-03-21 17:04:12 +00:00 |
|
Matthew Carrigan
|
7d1ae644ef
|
Added a --reduce_memory option to the training script to keep training
data on disc as a memmap rather than in memory
|
2019-03-21 17:02:18 +00:00 |
|
Matthew Carrigan
|
2bba7f810e
|
Added a --reduce_memory option to shelve docs to disc instead of keeping them in memory.
|
2019-03-21 16:50:16 +00:00 |
|
Matthew Carrigan
|
8733ffcb5e
|
Removing a couple of other old unnecessary comments
|
2019-03-21 14:09:57 +00:00 |
|
Matthew Carrigan
|
8a861048dd
|
Fixed up the notes on a possible future low-memory path
|
2019-03-21 14:08:39 +00:00 |
|
Matthew Carrigan
|
a8a577ba93
|
Reduced memory usage for pregenerating the data a lot by writing it
out on the fly without shuffling - the Sampler in the finetuning script
will shuffle for us.
|
2019-03-21 14:05:52 +00:00 |
|
Matthew Carrigan
|
0ae59e662d
|
Reduced memory usage for pregenerating the data a lot by writing it
out on the fly without shuffling - the Sampler in the finetuning script
will shuffle for us.
|
2019-03-21 14:04:17 +00:00 |
|
Matthew Carrigan
|
6a9038ba53
|
Removed an old irrelevant comment
|
2019-03-21 13:36:41 +00:00 |
|
Yuqiang Xie
|
77944d1b31
|
add tqdm to the process of eval
Maybe better.
|
2019-03-21 20:59:33 +08:00 |
|
Desiree Vogt-Lee
|
d52f914e24
|
weigths to weights
|
2019-03-21 15:02:59 +10:00 |
|
Matthew Carrigan
|
29a392fbcf
|
Small README changes
|
2019-03-20 17:35:17 +00:00 |
|
Matthew Carrigan
|
832b2b0058
|
Adding README
|
2019-03-20 17:31:49 +00:00 |
|
Matthew Carrigan
|
934d3f4d2f
|
Syncing up argument names between the scripts
|
2019-03-20 17:23:23 +00:00 |
|
Matthew Carrigan
|
f19ba35b2b
|
Move old finetuning script into the new folder
|
2019-03-20 16:47:06 +00:00 |
|
Matthew Carrigan
|
7de5c6aa5e
|
PEP8 and formatting cleanups
|
2019-03-20 16:44:04 +00:00 |
|
Matthew Carrigan
|
1798e98e5a
|
Added final TODOs
|
2019-03-20 16:42:37 +00:00 |
|
Matthew Carrigan
|
c64c2fc4c2
|
Fixed embarrassing indentation problem
|
2019-03-20 15:42:57 +00:00 |
|
Matthew Carrigan
|
0540d360f2
|
Fixed logging
|
2019-03-20 15:36:51 +00:00 |
|
Matthew Carrigan
|
976554a472
|
First commit of the new LM finetuning
|
2019-03-20 14:23:51 +00:00 |
|
lukovnikov
|
262a9992d7
|
class weights
|
2019-03-18 18:29:12 +01:00 |
|
lukovnikov
|
19cc2c084e
|
same
|
2019-03-18 15:13:35 +01:00 |
|
lukovnikov
|
2283dcca5e
|
import revert
|
2019-03-18 13:40:12 +01:00 |
|
lukovnikov
|
b6c1cae67b
|
branches, optim cosine fix
|
2019-03-18 13:32:04 +01:00 |
|
lukovnikov
|
ef28b2c747
|
branches, optim cosine fix
|
2019-03-18 13:18:07 +01:00 |
|
lukovnikov
|
90430ae7ec
|
Merge remote-tracking branch 'origin/master'
# Conflicts:
# pytorch_pretrained_bert/optimization.py
|
2019-03-18 13:15:29 +01:00 |
|
lukovnikov
|
bed6408dcc
|
branches, optim cosine fix
|
2019-03-18 13:09:55 +01:00 |
|
Ananya Harsh Jha
|
e5b63fb542
|
Merge branch 'master' of https://github.com/ananyahjha93/pytorch-pretrained-BERT
pull current master to local
|
2019-03-17 08:30:13 -04:00 |
|
Ananya Harsh Jha
|
8a4e90ff40
|
corrected folder creation error for MNLI-MM, verified GLUE results
|
2019-03-17 08:16:50 -04:00 |
|
Ananya Harsh Jha
|
e0bf01d9a9
|
added hack for mismatched MNLI
|
2019-03-16 14:10:48 -04:00 |
|
Ananya Harsh Jha
|
4c721c6b6a
|
added eval time metrics for GLUE tasks
|
2019-03-15 23:21:24 -04:00 |
|
Thomas Wolf
|
f3e5404880
|
Merge pull request #381 from tseretelitornike/master
Added missing imports.
|
2019-03-15 12:54:40 +01:00 |
|
tseretelitornike
|
83857ffeaa
|
Added missing imports.
|
2019-03-15 12:45:48 +01:00 |
|
Thomas Wolf
|
d5c037c3ed
|
Merge pull request #380 from yongbowin/patch-3
typo in annotation
|
2019-03-14 15:56:40 +01:00 |
|
Yongbo Wang
|
d1e4fa98a9
|
typo in annotation
modify `heruistic` to `heuristic` in line 660, `charcter` to `character` in line 661.
|
2019-03-14 17:32:15 +08:00 |
|
Thomas Wolf
|
59e2bdd086
|
Merge pull request #379 from yongbowin/patch-2
typo
|
2019-03-14 10:17:18 +01:00 |
|