ai-forever vmkhlv commited on
Commit
5192d06
1 Parent(s): 091f885

Update README.md (#4)

Browse files

- Update README.md (4824c79bf630104ee760caa7167ce530512ba09f)


Co-authored-by: Vladislav Mikhailov <[email protected]>

Files changed (1) hide show
  1. README.md +17 -3
README.md CHANGED
@@ -8,14 +8,28 @@ thumbnail: "https://github.com/sberbank-ai/model-zoo"
8
 
9
  ---
10
  # ruRoberta-large
11
- Model was trained by [SberDevices](https://sberdevices.ru/) team.
 
 
12
  * Task: `mask filling`
13
  * Type: `encoder`
14
- * Tokenizer: `bbpe`
15
  * Dict size: `50 257`
16
  * Num Parameters: `355 M`
17
  * Training Data Volume `250 GB`
18
 
19
  # Authors
20
  + NLP core team RnD [Telegram channel](https://t.me/nlpcoreteam):
21
- + Dmitry Zmitrovich
 
 
 
 
 
 
 
 
 
 
 
 
 
8
 
9
  ---
10
  # ruRoberta-large
11
+ The model architecture design, pretraining, and evaluation are documented in our preprint: [**A Family of Pretrained Transformer Language Models for Russian**](https://arxiv.org/abs/2309.10931).
12
+
13
+ The model is pretrained by the [SberDevices](https://sberdevices.ru/) team.
14
  * Task: `mask filling`
15
  * Type: `encoder`
16
+ * Tokenizer: `BBPE`
17
  * Dict size: `50 257`
18
  * Num Parameters: `355 M`
19
  * Training Data Volume `250 GB`
20
 
21
  # Authors
22
  + NLP core team RnD [Telegram channel](https://t.me/nlpcoreteam):
23
+ + Dmitry Zmitrovich
24
+
25
+ # Cite us
26
+ ```
27
+ @misc{zmitrovich2023family,
28
+ title={A Family of Pretrained Transformer Language Models for Russian},
29
+ author={Dmitry Zmitrovich and Alexander Abramov and Andrey Kalmykov and Maria Tikhonova and Ekaterina Taktasheva and Danil Astafurov and Mark Baushenko and Artem Snegirev and Tatiana Shavrina and Sergey Markov and Vladislav Mikhailov and Alena Fenogenova},
30
+ year={2023},
31
+ eprint={2309.10931},
32
+ archivePrefix={arXiv},
33
+ primaryClass={cs.CL}
34
+ }
35
+ ```