mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-24 23:08:57 +06:00
.. | ||
README.md |
language | |
---|---|
|
rubert-base-cased
RuBERT Russian, cased, 12‑layer, 768‑hidden, 12‑heads, 180M parameters
was trained on the Russian part of Wikipedia and news data. We used this training data to build a vocabulary of Russian subtokens and took a multilingual version of BERT‑base as an initialization for RuBERT[1].
1