Update README.md
Browse files
README.md
CHANGED
@@ -14,8 +14,9 @@ datasets:
|
|
14 |
# Model Card for Tamil-Mistral-7B-v0.1
|
15 |
|
16 |
The Tamil-Mistral-7B-v0.1 Large Language Model (LLM) is a pre-trained generative text model trained at the top of mistral base model 7 billion parameters. This is extends version of tokenization capability by increasing tamil tokens by 20k.
|
17 |
-
Additionally, it was Pretrained on 1.19 million
|
18 |
|
|
|
19 |
## Mistral model details
|
20 |
|
21 |
For full details of this model please read our [paper](https://arxiv.org/abs/2310.06825) and [release blog post](https://mistral.ai/news/announcing-mistral-7b/).
|
|
|
14 |
# Model Card for Tamil-Mistral-7B-v0.1
|
15 |
|
16 |
The Tamil-Mistral-7B-v0.1 Large Language Model (LLM) is a pre-trained generative text model trained at the top of mistral base model 7 billion parameters. This is extends version of tokenization capability by increasing tamil tokens by 20k.
|
17 |
+
Additionally, it was Pretrained on 1.19 million Tamil documents sourced from madlad-400 (Tamil) [MADLAD-400 (Multilingual Audited Dataset: Low-resource And Document-level)](https://arxiv.org/abs/2309.04662).
|
18 |
|
19 |
+
pretraining time: 145 hours (GPU NVIDIA RTX A6000 48GB)
|
20 |
## Mistral model details
|
21 |
|
22 |
For full details of this model please read our [paper](https://arxiv.org/abs/2310.06825) and [release blog post](https://mistral.ai/news/announcing-mistral-7b/).
|