mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-31 10:12:23 +06:00
Created card for spanbert-finetuned-squadv1
This commit is contained in:
parent
11d8bcc9d7
commit
676adf8625
84
model_cards/mrm8488/spanbert-finetuned-squadv1/README.md
Normal file
84
model_cards/mrm8488/spanbert-finetuned-squadv1/README.md
Normal file
@ -0,0 +1,84 @@
|
||||
---
|
||||
language: english
|
||||
thumbnail:
|
||||
---
|
||||
|
||||
# SpanBERT (spanbert-base-cased) fine-tuned on SQuAD v1.1
|
||||
|
||||
|
||||
[SpanBERT](https://github.com/facebookresearch/SpanBERT) created by [Facebook Research](https://github.com/facebookresearch) and fine-tuned on [SQuAD 1.1](https://rajpurkar.github.io/SQuAD-explorer/) for **Q&A** downstream task.
|
||||
|
||||
## Details of SpanBERT
|
||||
|
||||
A pre-training method that is designed to better represent and predict spans of text.
|
||||
|
||||
[SpanBERT: Improving Pre-training by Representing and Predicting Spans](https://arxiv.org/abs/1907.10529)
|
||||
|
||||
## Details of the downstream task (Q&A) - Dataset
|
||||
|
||||
[SQuAD 1.1](https://rajpurkar.github.io/SQuAD-explorer/) contains 100,000+ question-answer pairs on 500+ articles.
|
||||
|
||||
| Dataset | Split | # samples |
|
||||
| -------- | ----- | --------- |
|
||||
| SQuAD1.1 | train | 87.7k |
|
||||
| SQuAD1.1 | eval | 10.6k |
|
||||
|
||||
## Model training
|
||||
|
||||
The model was trained on a Tesla P100 GPU and 25GB of RAM.
|
||||
The script for fine tuning can be found [here](https://github.com/huggingface/transformers/blob/master/examples/run_squad.py)
|
||||
|
||||
## Results:
|
||||
|
||||
| Metric | # Value |
|
||||
| ------ | --------- |
|
||||
| **EM** | **85.49** |
|
||||
| **F1** | **91.98** |
|
||||
|
||||
### Raw metrics:
|
||||
|
||||
```json
|
||||
{
|
||||
"exact": 85.49668874172185,
|
||||
"f1": 91.9845699540379,
|
||||
"total": 10570,
|
||||
"HasAns_exact": 85.49668874172185,
|
||||
"HasAns_f1": 91.9845699540379,
|
||||
"HasAns_total": 10570,
|
||||
"best_exact": 85.49668874172185,
|
||||
"best_exact_thresh": 0.0,
|
||||
"best_f1": 91.9845699540379,
|
||||
"best_f1_thresh": 0.0
|
||||
}
|
||||
```
|
||||
|
||||
## Comparison:
|
||||
|
||||
| Model | EM | F1 score |
|
||||
| ----------------------------------------------------------------------------------------- | --------- | --------- |
|
||||
| [SpanBert official repo](https://github.com/facebookresearch/SpanBERT#pre-trained-models) | - | 92.4\* |
|
||||
| [spanbert-finetuned-squadv1](https://huggingface.co/mrm8488/spanbert-finetuned-squadv1) | **85.49** | **91.98** |
|
||||
|
||||
## Model in action
|
||||
|
||||
Fast usage with **pipelines**:
|
||||
|
||||
```python
|
||||
from transformers import pipeline
|
||||
|
||||
qa_pipeline = pipeline(
|
||||
"question-answering",
|
||||
model="mrm8488/spanbert-finetuned-squadv1",
|
||||
tokenizer="mrm8488/spanbert-finetuned-squadv1"
|
||||
)
|
||||
|
||||
qa_pipeline({
|
||||
'context': "Manuel Romero has been working hardly in the repository hugginface/transformers lately",
|
||||
'question': "Who has been working hard for hugginface/transformers lately?"
|
||||
|
||||
})
|
||||
```
|
||||
|
||||
> Created by [Manuel Romero/@mrm8488](https://twitter.com/mrm8488) | [LinkedIn](https://www.linkedin.com/in/manuel-romero-cs/)
|
||||
|
||||
> Made with <span style="color: #e25555;">♥</span> in Spain
|
Loading…
Reference in New Issue
Block a user