yzy5630
|
a1fe4ba9c9
|
use new API for save and load
|
2019-07-18 15:45:23 +08:00 |
|
yzy5630
|
a7ba27b1b4
|
add parser for adam
|
2019-07-18 08:52:51 +08:00 |
|
yzy5630
|
d6522e2873
|
change loss and optimizer to new API
|
2019-07-17 21:22:34 +08:00 |
|
yzy5630
|
123da5a2fa
|
fix errors for lm_finetuning examples
|
2019-07-17 09:56:07 +08:00 |
|
thomwolf
|
2397f958f9
|
updating examples and doc
|
2019-07-14 23:20:10 +02:00 |
|
thomwolf
|
0bab55d5d5
|
[BIG] name change
|
2019-07-05 11:55:36 +02:00 |
|
thomwolf
|
c41f2bad69
|
WIP XLM + refactoring
|
2019-07-03 22:54:39 +02:00 |
|
Oliver Guhr
|
5c08c8c273
|
adds the tokenizer + model config to the output
|
2019-06-11 13:46:33 +02:00 |
|
burcturkoglu
|
00c7fd2b79
|
Division to num_train_optimizer of global_step in lr_this_step is removed.
|
2019-05-09 10:57:03 +03:00 |
|
burcturkoglu
|
5289b4b9e0
|
Division to num_train_optimizer of global_step in lr_this_step is removed.
|
2019-05-09 10:51:38 +03:00 |
|
thomwolf
|
d94c6b0144
|
fix training schedules in examples to match new API
|
2019-04-23 11:17:06 +02:00 |
|
Matthew Carrigan
|
b8e2a9c584
|
Made --reduce_memory actually do something in finetune_on_pregenerated
|
2019-04-22 14:01:48 +01:00 |
|
thomwolf
|
4bc4c69af9
|
finetuning any BERT model - fixes #455
|
2019-04-11 16:57:59 +02:00 |
|
Matthew Carrigan
|
8d1d1ffde2
|
Corrected the displayed loss when gradient_accumulation_steps > 1
|
2019-03-25 12:15:19 +00:00 |
|
Matthew Carrigan
|
7d1ae644ef
|
Added a --reduce_memory option to the training script to keep training
data on disc as a memmap rather than in memory
|
2019-03-21 17:02:18 +00:00 |
|
Matthew Carrigan
|
8a861048dd
|
Fixed up the notes on a possible future low-memory path
|
2019-03-21 14:08:39 +00:00 |
|
Matthew Carrigan
|
a8a577ba93
|
Reduced memory usage for pregenerating the data a lot by writing it
out on the fly without shuffling - the Sampler in the finetuning script
will shuffle for us.
|
2019-03-21 14:05:52 +00:00 |
|
Matthew Carrigan
|
0ae59e662d
|
Reduced memory usage for pregenerating the data a lot by writing it
out on the fly without shuffling - the Sampler in the finetuning script
will shuffle for us.
|
2019-03-21 14:04:17 +00:00 |
|
Matthew Carrigan
|
7de5c6aa5e
|
PEP8 and formatting cleanups
|
2019-03-20 16:44:04 +00:00 |
|
Matthew Carrigan
|
1798e98e5a
|
Added final TODOs
|
2019-03-20 16:42:37 +00:00 |
|
Matthew Carrigan
|
c64c2fc4c2
|
Fixed embarrassing indentation problem
|
2019-03-20 15:42:57 +00:00 |
|
Matthew Carrigan
|
0540d360f2
|
Fixed logging
|
2019-03-20 15:36:51 +00:00 |
|
Matthew Carrigan
|
976554a472
|
First commit of the new LM finetuning
|
2019-03-20 14:23:51 +00:00 |
|