mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-31 02:02:21 +06:00
[model_cards] xlnet_chinese_large & roberta_chinese_large
This commit is contained in:
parent
623ba0236d
commit
73d6a2f901
@ -1,3 +1,7 @@
|
||||
---
|
||||
language: chinese
|
||||
---
|
||||
|
||||
## albert_chinese_small
|
||||
|
||||
### Overview
|
||||
|
@ -1,3 +1,7 @@
|
||||
---
|
||||
language: chinese
|
||||
---
|
||||
|
||||
## albert_chinese_tiny
|
||||
|
||||
### Overview
|
||||
|
@ -1,3 +1,7 @@
|
||||
---
|
||||
language: chinese
|
||||
---
|
||||
|
||||
# Introduction
|
||||
This model was trained on TPU and the details are as follows:
|
||||
|
||||
|
@ -1,3 +1,7 @@
|
||||
---
|
||||
language: chinese
|
||||
---
|
||||
|
||||
## roberta_chinese_base
|
||||
|
||||
### Overview
|
||||
|
35
model_cards/clue/roberta_chinese_large/README.md
Normal file
35
model_cards/clue/roberta_chinese_large/README.md
Normal file
@ -0,0 +1,35 @@
|
||||
---
|
||||
language: chinese
|
||||
---
|
||||
|
||||
## roberta_chinese_large
|
||||
|
||||
### Overview
|
||||
|
||||
**Language model:** roberta-large
|
||||
**Model size:** 1.2G
|
||||
**Language:** Chinese
|
||||
**Training data:** [CLUECorpusSmall](https://github.com/CLUEbenchmark/CLUECorpus2020)
|
||||
**Eval data:** [CLUE dataset](https://github.com/CLUEbenchmark/CLUE)
|
||||
|
||||
### Results
|
||||
|
||||
For results on downstream tasks like text classification, please refer to [this repository](https://github.com/CLUEbenchmark/CLUE).
|
||||
|
||||
### Usage
|
||||
|
||||
**NOTE:** You have to call **BertTokenizer** instead of RobertaTokenizer !!!
|
||||
|
||||
```
|
||||
import torch
|
||||
from transformers import BertTokenizer, BertModel
|
||||
tokenizer = BertTokenizer.from_pretrained("clue/roberta_chinese_large")
|
||||
roberta = BertModel.from_pretrained("clue/roberta_chinese_large")
|
||||
```
|
||||
|
||||
### About CLUE benchmark
|
||||
|
||||
Organization of Language Understanding Evaluation benchmark for Chinese: tasks & datasets, baselines, pre-trained Chinese models, corpus and leaderboard.
|
||||
|
||||
Github: https://github.com/CLUEbenchmark
|
||||
Website: https://www.cluebenchmarks.com/
|
33
model_cards/clue/xlnet_chinese_large/README.md
Normal file
33
model_cards/clue/xlnet_chinese_large/README.md
Normal file
@ -0,0 +1,33 @@
|
||||
---
|
||||
language: chinese
|
||||
---
|
||||
|
||||
## xlnet_chinese_large
|
||||
|
||||
### Overview
|
||||
|
||||
**Language model:** xlnet-large
|
||||
**Model size:** 1.3G
|
||||
**Language:** Chinese
|
||||
**Training data:** [CLUECorpusSmall](https://github.com/CLUEbenchmark/CLUECorpus2020)
|
||||
**Eval data:** [CLUE dataset](https://github.com/CLUEbenchmark/CLUE)
|
||||
|
||||
### Results
|
||||
|
||||
For results on downstream tasks like text classification, please refer to [this repository](https://github.com/CLUEbenchmark/CLUE).
|
||||
|
||||
### Usage
|
||||
|
||||
```
|
||||
import torch
|
||||
from transformers import XLNetTokenizer,XLNetModel
|
||||
tokenizer = XLNetTokenizer.from_pretrained("clue/xlnet_chinese_large")
|
||||
xlnet = XLNetModel.from_pretrained("clue/xlnet_chinese_large")
|
||||
```
|
||||
|
||||
### About CLUE benchmark
|
||||
|
||||
Organization of Language Understanding Evaluation benchmark for Chinese: tasks & datasets, baselines, pre-trained Chinese models, corpus and leaderboard.
|
||||
|
||||
Github: https://github.com/CLUEbenchmark
|
||||
Website: https://www.cluebenchmarks.com/
|
Loading…
Reference in New Issue
Block a user