Commit
·
2bb66cb
1
Parent(s):
f397ab4
Update README.md
Browse files
README.md
CHANGED
@@ -6,11 +6,37 @@ pipeline_tag: text-generation
|
|
6 |
tags:
|
7 |
- PyTorch
|
8 |
- Transformers
|
|
|
|
|
|
|
|
|
9 |
thumbnail: "https://github.com/sberbank-ai/mgpt"
|
10 |
---
|
11 |
|
12 |
# Multilingual GPT model
|
13 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
14 |
Model was trained with sequence length 1024 using transformers lib by [SberDevices](https://sberdevices.ru/) team on 80B tokens for 3 epochs. After that model was finetuned 1 epoch with sequence length 2048.
|
15 |
|
16 |
Total training time was around n days on n GPUs for n context and few days on n GPUs for n context.
|
|
|
6 |
tags:
|
7 |
- PyTorch
|
8 |
- Transformers
|
9 |
+
- gpt3
|
10 |
+
- gpt2
|
11 |
+
- Deepspeed
|
12 |
+
- Megatron
|
13 |
thumbnail: "https://github.com/sberbank-ai/mgpt"
|
14 |
---
|
15 |
|
16 |
# Multilingual GPT model
|
17 |
|
18 |
+
We introduce family of autoregressive GPT-like models with 1.3 billion parameters trained on 60 languages from 25 language families using Wikipedia and Colossal Clean Crawled Corpus.
|
19 |
+
|
20 |
+
We reproduce the GPT-3 architecture using GPT-2 sources and the sparse attention mechanism, [Deepspeed](https://github.com/microsoft/DeepSpeed) and [Megatron]() frameworks allows us to effectively parallelize the training and inference steps. Resulting models show performance on par with the recently released [XGLM](https://arxiv.org/pdf/2112.10668.pdf) models at the same time covering more languages and enhance NLP possibilities for low resource languages.
|
21 |
+
|
22 |
+
## Code
|
23 |
+
The source code for the mGPT XL model is available on [Github](https://github.com/sberbank-ai/mgpt)
|
24 |
+
|
25 |
+
## Paper
|
26 |
+
[Arxiv preprint](https://arxiv.org/user)
|
27 |
+
|
28 |
+
Cite us:
|
29 |
+
```{
|
30 |
+
bibtex
|
31 |
+
}
|
32 |
+
```
|
33 |
+
|
34 |
+
## Languages
|
35 |
+
|
36 |
+
## Training Data Statistics
|
37 |
+
|
38 |
+
|
39 |
+
## Details
|
40 |
Model was trained with sequence length 1024 using transformers lib by [SberDevices](https://sberdevices.ru/) team on 80B tokens for 3 epochs. After that model was finetuned 1 epoch with sequence length 2048.
|
41 |
|
42 |
Total training time was around n days on n GPUs for n context and few days on n GPUs for n context.
|