mirror of
https://github.com/huggingface/transformers.git
synced 2025-08-01 02:31:11 +06:00
update to BERT-large results
This commit is contained in:
parent
c4bfc646f5
commit
6d6b916f48
@ -206,7 +206,7 @@ Training with the previous hyper-parameters gave us the following results:
|
|||||||
|
|
||||||
The options we list above allow to fine-tune BERT-large rather easily on GPU(s) instead of the TPU used by the original implementation.
|
The options we list above allow to fine-tune BERT-large rather easily on GPU(s) instead of the TPU used by the original implementation.
|
||||||
|
|
||||||
For example, fine-tuning BERT-large on SQuAD can be done on a server with 4 k-80 (these are pretty old now) in 18 hours. Our results are similar to the TensorFlow implementation results:
|
For example, fine-tuning BERT-large on SQuAD can be done on a server with 4 k-80 (these are pretty old now) in 18 hours. Our results are similar to the TensorFlow implementation results (actually slightly higher):
|
||||||
```bash
|
```bash
|
||||||
{"exact_match": 84.56953642384106, "f1": 91.04028647786927}
|
{"exact_match": 84.56953642384106, "f1": 91.04028647786927}
|
||||||
```
|
```
|
||||||
|
Loading…
Reference in New Issue
Block a user