tainc commited on
Commit
e6014de
1 Parent(s): 36cd657

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -9
README.md CHANGED
@@ -33,7 +33,7 @@ SEA-LION stands for <i>Southeast Asian Languages In One Network</i>.
33
 
34
  ## Model Details
35
  ### Model Description
36
- The continued pre-training data for Llama3.1 8B CPT SEA-LIONv3 Base encompasses approximately 200B tokens.
37
 
38
  For tokenisation, the model employs the default tokenizer used in Llama3.1 8B Instruct.
39
 
@@ -52,14 +52,13 @@ For more details on Llama3.1 8B CPT SEA-LIONv3 base benchmark performance, pleas
52
 
53
  ## Technical Specifications
54
  ### Infrastructure
55
- Llama3.1 8B CPT SEA-LIONv3 was trained using [MosaicML Composer](https://github.com/mosaicml/composer)
56
- on the following hardware:
57
 
58
- | Training Details | Llama3.1 8B CPT SEA-LIONv3 |
59
- |----------------------|:------------------------:|
60
- | SingTel HGX-100 | 8+1 instances |
61
- | Nvidia H100 80GB GPU | 64+8 |
62
- | Training Duration | 10 days |
63
 
64
  ### Configuration
65
  | HyperParameter | Llama3.1 8B CPT SEA-LIONv3 |
@@ -69,7 +68,6 @@ on the following hardware:
69
  | Scheduler | weight_stable_decay |
70
  | Learning Rate | 1.0e-5 |
71
  | Global Batch Size | 512 |
72
- | Micro Batch Size | 1 |
73
 
74
  ## Data
75
  Llama3.1 8B CPT SEA-LIONv3 base model was continued pre-trained on 200B tokens of the following data:
 
33
 
34
  ## Model Details
35
  ### Model Description
36
+ The continued pre-training data for Llama3.1 8B CPT SEA-LIONv3 Base encompasses approximately 200B tokens and includes the 11 official Southeast Asian languages: English, Chinese, Vietnamese, Indonesian, Thai, Tamil, Filipino, Malay, Khmer, Lao, Burmese.
37
 
38
  For tokenisation, the model employs the default tokenizer used in Llama3.1 8B Instruct.
39
 
 
52
 
53
  ## Technical Specifications
54
  ### Infrastructure
55
+ Llama3.1 8B CPT SEA-LIONv3 was trained using [MosaicML Composer](https://github.com/mosaicml/composer) on the following hardware:
 
56
 
57
+ | Training Details | Llama3.1 8B CPT SEA-LIONv3 |
58
+ |-----------------------|:--------------------------:|
59
+ | AWS p5e.48xlarge | 8 instances |
60
+ | Nvidia H200 140GB GPU | 64 |
61
+ | Training Duration | 136 Hours |
62
 
63
  ### Configuration
64
  | HyperParameter | Llama3.1 8B CPT SEA-LIONv3 |
 
68
  | Scheduler | weight_stable_decay |
69
  | Learning Rate | 1.0e-5 |
70
  | Global Batch Size | 512 |
 
71
 
72
  ## Data
73
  Llama3.1 8B CPT SEA-LIONv3 base model was continued pre-trained on 200B tokens of the following data: