mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-31 02:02:21 +06:00
![]() * Create README.md * Update model_cards/uer/chinese_roberta_L-2_H-128/README.md Co-authored-by: Julien Chaumond <chaumond@gmail.com> |
||
---|---|---|
.. | ||
README.md |
language | datasets | |
---|---|---|
zh |
|
Chinese RoBERTa Miniatures
Model description
This is the set of 24 Chinese RoBERTa models pre-trained by UER-py.
You can download the 24 Chinese RoBERTa miniatures either from the UER-py Github page, or via HuggingFace from the links below:
H=128 | H=256 | H=512 | H=768 | |
---|---|---|---|---|
L=2 | 2/128 (BERT-Tiny) | [2/256] | [2/512] | [2/768] |
L=4 | [4/128] | [4/256 (BERT-Mini)] | [4/512 (BERT-Small)] | [4/768] |
L=6 | [6/128] | [6/256] | [6/512] | [6/768] |
L=8 | [8/128] | [8/256] | [8/512 (BERT-Medium)] | [8/768] |
L=10 | [10/128] | [10/256] | [10/512] | [10/768] |
L=12 | [12/128] | [12/256] | [12/512] | [12/768 (BERT-Base)] |
Training data
CLUECorpus2020 and CLUECorpusSmall are used as training corpus.
Training procedure
Training details can be found in UER-py.
BibTeX entry and citation info
@article{zhao2019uer,
title={UER: An Open-Source Toolkit for Pre-training Models},
author={Zhao, Zhe and Chen, Hui and Zhang, Jinbin and Zhao, Xin and Liu, Tao and Lu, Wei and Chen, Xi and Deng, Haotang and Ju, Qi and Du, Xiaoyong},
journal={EMNLP-IJCNLP 2019},
pages={241},
year={2019}
}