Update README.md
Browse files
README.md
CHANGED
@@ -33,7 +33,7 @@ SEA-LION stands for <i>Southeast Asian Languages In One Network</i>.
|
|
33 |
|
34 |
## Model Details
|
35 |
### Model Description
|
36 |
-
The continued pre-training data for Llama3.1 8B CPT SEA-LIONv3 Base encompasses approximately 200B tokens.
|
37 |
|
38 |
For tokenisation, the model employs the default tokenizer used in Llama3.1 8B Instruct.
|
39 |
|
@@ -52,14 +52,13 @@ For more details on Llama3.1 8B CPT SEA-LIONv3 base benchmark performance, pleas
|
|
52 |
|
53 |
## Technical Specifications
|
54 |
### Infrastructure
|
55 |
-
Llama3.1 8B CPT SEA-LIONv3 was trained using [MosaicML Composer](https://github.com/mosaicml/composer)
|
56 |
-
on the following hardware:
|
57 |
|
58 |
-
| Training Details
|
59 |
-
|
60 |
-
|
|
61 |
-
| Nvidia
|
62 |
-
| Training Duration
|
63 |
|
64 |
### Configuration
|
65 |
| HyperParameter | Llama3.1 8B CPT SEA-LIONv3 |
|
@@ -69,7 +68,6 @@ on the following hardware:
|
|
69 |
| Scheduler | weight_stable_decay |
|
70 |
| Learning Rate | 1.0e-5 |
|
71 |
| Global Batch Size | 512 |
|
72 |
-
| Micro Batch Size | 1 |
|
73 |
|
74 |
## Data
|
75 |
Llama3.1 8B CPT SEA-LIONv3 base model was continued pre-trained on 200B tokens of the following data:
|
|
|
33 |
|
34 |
## Model Details
|
35 |
### Model Description
|
36 |
+
The continued pre-training data for Llama3.1 8B CPT SEA-LIONv3 Base encompasses approximately 200B tokens and includes the 11 official Southeast Asian languages: English, Chinese, Vietnamese, Indonesian, Thai, Tamil, Filipino, Malay, Khmer, Lao, Burmese.
|
37 |
|
38 |
For tokenisation, the model employs the default tokenizer used in Llama3.1 8B Instruct.
|
39 |
|
|
|
52 |
|
53 |
## Technical Specifications
|
54 |
### Infrastructure
|
55 |
+
Llama3.1 8B CPT SEA-LIONv3 was trained using [MosaicML Composer](https://github.com/mosaicml/composer) on the following hardware:
|
|
|
56 |
|
57 |
+
| Training Details | Llama3.1 8B CPT SEA-LIONv3 |
|
58 |
+
|-----------------------|:--------------------------:|
|
59 |
+
| AWS p5e.48xlarge | 8 instances |
|
60 |
+
| Nvidia H200 140GB GPU | 64 |
|
61 |
+
| Training Duration | 136 Hours |
|
62 |
|
63 |
### Configuration
|
64 |
| HyperParameter | Llama3.1 8B CPT SEA-LIONv3 |
|
|
|
68 |
| Scheduler | weight_stable_decay |
|
69 |
| Learning Rate | 1.0e-5 |
|
70 |
| Global Batch Size | 512 |
|
|
|
71 |
|
72 |
## Data
|
73 |
Llama3.1 8B CPT SEA-LIONv3 base model was continued pre-trained on 200B tokens of the following data:
|