SylvanL commited on
Commit
0ddb88d
·
verified ·
1 Parent(s): 20d8d15

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -3
README.md CHANGED
@@ -8,8 +8,11 @@ base_model:
8
  - Qwen/Qwen2-7B-Instruct
9
  ---
10
 
11
- 在2张V800-80G上, 基于Qwen/Qwen2-7B-Instruct上, 使用SylvanL/Traditional-Chinese-Medicine-Dataset-Pretrain进行了2个epoch的继续预训练.
12
 
 
 
 
13
 
14
  epoch 1:
15
  total FLOS 824852GF
@@ -17,14 +20,14 @@ train loss: 1.6586
17
  train_samples_per_second: 3.246
18
  train_steps_per_second: 0.014
19
  train_runtime: 37:01:33.17
20
- perplexity: 7.2646328377141005
21
 
22
  epoch 2:
23
  train loss: 1.3895
24
  train_samples_per_second: 3.11
25
  train_steps_per_second: 0.014
26
  train_runtime: 38:38:44.20
27
- perplexity:
28
 
29
  ```
30
  llamafactory-cli train \
 
8
  - Qwen/Qwen2-7B-Instruct
9
  ---
10
 
11
+ 在2张V800-80G上,
12
 
13
+ 基于Qwen/Qwen2-7B-Instruct,
14
+
15
+ 使用SylvanL/Traditional-Chinese-Medicine-Dataset-Pretrain进行了2个epoch的继续预训练(Continue Pre-train).
16
 
17
  epoch 1:
18
  total FLOS 824852GF
 
20
  train_samples_per_second: 3.246
21
  train_steps_per_second: 0.014
22
  train_runtime: 37:01:33.17
23
+ average perplexity: 7.2646328377141005
24
 
25
  epoch 2:
26
  train loss: 1.3895
27
  train_samples_per_second: 3.11
28
  train_steps_per_second: 0.014
29
  train_runtime: 38:38:44.20
30
+ average perplexity: 5.800355962033688
31
 
32
  ```
33
  llamafactory-cli train \