Safetensors
qwen2
reasoning
ptrdvn commited on
Commit
465d606
·
verified ·
1 Parent(s): af0a73e

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: other
4
+ base_model: deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
5
+ tags:
6
+ - llama-factory
7
+ - full
8
+ - generated_from_trainer
9
+ model-index:
10
+ - name: reasoning-multilingual-R1-Llama-70B-train
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # reasoning-multilingual-R1-Llama-70B-train
18
+
19
+ This model is a fine-tuned version of [deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B) on the reasoning-multilingual-R1-Llama-70B-train dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 0.9008
22
+
23
+ ## Model description
24
+
25
+ More information needed
26
+
27
+ ## Intended uses & limitations
28
+
29
+ More information needed
30
+
31
+ ## Training and evaluation data
32
+
33
+ More information needed
34
+
35
+ ## Training procedure
36
+
37
+ ### Training hyperparameters
38
+
39
+ The following hyperparameters were used during training:
40
+ - learning_rate: 1e-05
41
+ - train_batch_size: 1
42
+ - eval_batch_size: 1
43
+ - seed: 42
44
+ - distributed_type: multi-GPU
45
+ - num_devices: 8
46
+ - total_train_batch_size: 8
47
+ - total_eval_batch_size: 8
48
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
49
+ - lr_scheduler_type: cosine
50
+ - lr_scheduler_warmup_ratio: 0.01
51
+ - num_epochs: 1.0
52
+
53
+ ### Training results
54
+
55
+ | Training Loss | Epoch | Step | Validation Loss |
56
+ |:-------------:|:------:|:----:|:---------------:|
57
+ | 0.9946 | 0.1020 | 10 | 0.9808 |
58
+ | 1.042 | 0.2041 | 20 | 0.9511 |
59
+ | 1.1236 | 0.3061 | 30 | 0.9309 |
60
+ | 1.238 | 0.4082 | 40 | 0.9211 |
61
+ | 0.8178 | 0.5102 | 50 | 0.9119 |
62
+ | 0.951 | 0.6122 | 60 | 0.9066 |
63
+ | 0.7869 | 0.7143 | 70 | 0.9032 |
64
+ | 1.2626 | 0.8163 | 80 | 0.9012 |
65
+ | 1.2385 | 0.9184 | 90 | 0.9009 |
66
+
67
+
68
+ ### Framework versions
69
+
70
+ - Transformers 4.48.1
71
+ - Pytorch 2.5.1+cu124
72
+ - Datasets 3.1.0
73
+ - Tokenizers 0.21.0
all_results.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "eval_loss": 0.9007720947265625,
4
+ "eval_runtime": 0.2597,
5
+ "eval_samples_per_second": 30.809,
6
+ "eval_steps_per_second": 3.851,
7
+ "total_flos": 3.270639151271117e+16,
8
+ "train_loss": 1.0707874030483013,
9
+ "train_runtime": 199.5779,
10
+ "train_samples_per_second": 3.908,
11
+ "train_steps_per_second": 0.491
12
+ }
config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B",
3
+ "architectures": [
4
+ "Qwen2ForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "eos_token_id": 151643,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 1536,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 8960,
13
+ "max_position_embeddings": 131072,
14
+ "max_window_layers": 21,
15
+ "model_type": "qwen2",
16
+ "num_attention_heads": 12,
17
+ "num_hidden_layers": 28,
18
+ "num_key_value_heads": 2,
19
+ "rms_norm_eps": 1e-06,
20
+ "rope_scaling": null,
21
+ "rope_theta": 10000,
22
+ "sliding_window": null,
23
+ "tie_word_embeddings": false,
24
+ "torch_dtype": "bfloat16",
25
+ "transformers_version": "4.48.1",
26
+ "use_cache": false,
27
+ "use_mrope": false,
28
+ "use_sliding_window": false,
29
+ "vocab_size": 151936
30
+ }
eval_results.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "eval_loss": 0.9007720947265625,
4
+ "eval_runtime": 0.2597,
5
+ "eval_samples_per_second": 30.809,
6
+ "eval_steps_per_second": 3.851
7
+ }
generation_config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 151646,
4
+ "do_sample": true,
5
+ "eos_token_id": 151643,
6
+ "temperature": 0.6,
7
+ "top_p": 0.95,
8
+ "transformers_version": "4.48.1"
9
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9435a1147f867736d4fe643968003aa7901bea2c8ac0f3a2163bc38cddb115cc
3
+ size 3554214752
special_tokens_map.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ {
4
+ "content": "<|im_end|>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ }
10
+ ],
11
+ "bos_token": {
12
+ "content": "<|begin▁of▁sentence|>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false
17
+ },
18
+ "eos_token": {
19
+ "content": "<|end▁of▁sentence|>",
20
+ "lstrip": false,
21
+ "normalized": false,
22
+ "rstrip": false,
23
+ "single_word": false
24
+ },
25
+ "pad_token": {
26
+ "content": "<|end▁of▁sentence|>",
27
+ "lstrip": false,
28
+ "normalized": false,
29
+ "rstrip": false,
30
+ "single_word": false
31
+ }
32
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02643f00207dfc5ed248992486bde04314c21dca556bf65ce520690962b8db63
3
+ size 11422965
tokenizer_config.json ADDED
@@ -0,0 +1,208 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
+ "added_tokens_decoder": {
6
+ "151643": {
7
+ "content": "<|end▁of▁sentence|>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "151644": {
15
+ "content": "<|User|>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": false
21
+ },
22
+ "151645": {
23
+ "content": "<|Assistant|>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": false
29
+ },
30
+ "151646": {
31
+ "content": "<|begin▁of▁sentence|>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false,
36
+ "special": true
37
+ },
38
+ "151647": {
39
+ "content": "<|EOT|>",
40
+ "lstrip": false,
41
+ "normalized": false,
42
+ "rstrip": false,
43
+ "single_word": false,
44
+ "special": false
45
+ },
46
+ "151648": {
47
+ "content": "<think>",
48
+ "lstrip": false,
49
+ "normalized": false,
50
+ "rstrip": false,
51
+ "single_word": false,
52
+ "special": false
53
+ },
54
+ "151649": {
55
+ "content": "</think>",
56
+ "lstrip": false,
57
+ "normalized": false,
58
+ "rstrip": false,
59
+ "single_word": false,
60
+ "special": false
61
+ },
62
+ "151650": {
63
+ "content": "<|quad_start|>",
64
+ "lstrip": false,
65
+ "normalized": false,
66
+ "rstrip": false,
67
+ "single_word": false,
68
+ "special": true
69
+ },
70
+ "151651": {
71
+ "content": "<|quad_end|>",
72
+ "lstrip": false,
73
+ "normalized": false,
74
+ "rstrip": false,
75
+ "single_word": false,
76
+ "special": true
77
+ },
78
+ "151652": {
79
+ "content": "<|vision_start|>",
80
+ "lstrip": false,
81
+ "normalized": false,
82
+ "rstrip": false,
83
+ "single_word": false,
84
+ "special": true
85
+ },
86
+ "151653": {
87
+ "content": "<|vision_end|>",
88
+ "lstrip": false,
89
+ "normalized": false,
90
+ "rstrip": false,
91
+ "single_word": false,
92
+ "special": true
93
+ },
94
+ "151654": {
95
+ "content": "<|vision_pad|>",
96
+ "lstrip": false,
97
+ "normalized": false,
98
+ "rstrip": false,
99
+ "single_word": false,
100
+ "special": true
101
+ },
102
+ "151655": {
103
+ "content": "<|image_pad|>",
104
+ "lstrip": false,
105
+ "normalized": false,
106
+ "rstrip": false,
107
+ "single_word": false,
108
+ "special": true
109
+ },
110
+ "151656": {
111
+ "content": "<|video_pad|>",
112
+ "lstrip": false,
113
+ "normalized": false,
114
+ "rstrip": false,
115
+ "single_word": false,
116
+ "special": true
117
+ },
118
+ "151657": {
119
+ "content": "<tool_call>",
120
+ "lstrip": false,
121
+ "normalized": false,
122
+ "rstrip": false,
123
+ "single_word": false,
124
+ "special": false
125
+ },
126
+ "151658": {
127
+ "content": "</tool_call>",
128
+ "lstrip": false,
129
+ "normalized": false,
130
+ "rstrip": false,
131
+ "single_word": false,
132
+ "special": false
133
+ },
134
+ "151659": {
135
+ "content": "<|fim_prefix|>",
136
+ "lstrip": false,
137
+ "normalized": false,
138
+ "rstrip": false,
139
+ "single_word": false,
140
+ "special": false
141
+ },
142
+ "151660": {
143
+ "content": "<|fim_middle|>",
144
+ "lstrip": false,
145
+ "normalized": false,
146
+ "rstrip": false,
147
+ "single_word": false,
148
+ "special": false
149
+ },
150
+ "151661": {
151
+ "content": "<|fim_suffix|>",
152
+ "lstrip": false,
153
+ "normalized": false,
154
+ "rstrip": false,
155
+ "single_word": false,
156
+ "special": false
157
+ },
158
+ "151662": {
159
+ "content": "<|fim_pad|>",
160
+ "lstrip": false,
161
+ "normalized": false,
162
+ "rstrip": false,
163
+ "single_word": false,
164
+ "special": false
165
+ },
166
+ "151663": {
167
+ "content": "<|repo_name|>",
168
+ "lstrip": false,
169
+ "normalized": false,
170
+ "rstrip": false,
171
+ "single_word": false,
172
+ "special": false
173
+ },
174
+ "151664": {
175
+ "content": "<|file_sep|>",
176
+ "lstrip": false,
177
+ "normalized": false,
178
+ "rstrip": false,
179
+ "single_word": false,
180
+ "special": false
181
+ },
182
+ "151665": {
183
+ "content": "<|im_end|>",
184
+ "lstrip": false,
185
+ "normalized": false,
186
+ "rstrip": false,
187
+ "single_word": false,
188
+ "special": true
189
+ }
190
+ },
191
+ "additional_special_tokens": [
192
+ "<|im_end|>"
193
+ ],
194
+ "bos_token": "<|begin▁of▁sentence|>",
195
+ "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set ns = namespace(is_first=false, is_tool=false, is_output_first=true, system_prompt='') %}{%- for message in messages %}{%- if message['role'] == 'system' %}{% set ns.system_prompt = message['content'] %}{%- endif %}{%- endfor %}{{bos_token}}{{ns.system_prompt}}{%- for message in messages %}{%- if message['role'] == 'user' %}{%- set ns.is_tool = false -%}{{'<|User|>' + message['content']}}{%- endif %}{%- if message['role'] == 'assistant' and message['content'] is none %}{%- set ns.is_tool = false -%}{%- for tool in message['tool_calls']%}{%- if not ns.is_first %}{{'<|Assistant|><|tool▁calls▁begin|><|tool▁call▁begin|>' + tool['type'] + '<|tool▁sep|>' + tool['function']['name'] + '\\n' + '```json' + '\\n' + tool['function']['arguments'] + '\\n' + '```' + '<|tool▁call▁end|>'}}{%- set ns.is_first = true -%}{%- else %}{{'\\n' + '<|tool▁call▁begin|>' + tool['type'] + '<|tool▁sep|>' + tool['function']['name'] + '\\n' + '```json' + '\\n' + tool['function']['arguments'] + '\\n' + '```' + '<|tool▁call▁end|>'}}{{'<|tool▁calls▁end|><|end▁of▁sentence|>'}}{%- endif %}{%- endfor %}{%- endif %}{%- if message['role'] == 'assistant' and message['content'] is not none %}{%- if ns.is_tool %}{{'<|tool▁outputs▁end|>' + message['content'] + '<|end▁of▁sentence|>'}}{%- set ns.is_tool = false -%}{%- else %}{% set content = message['content'] %}{% if '</think>' in content %}{% set content = content.split('</think>')[-1] %}{% endif %}{{'<|Assistant|>' + content + '<|end▁of▁sentence|>'}}{%- endif %}{%- endif %}{%- if message['role'] == 'tool' %}{%- set ns.is_tool = true -%}{%- if ns.is_output_first %}{{'<|tool▁outputs▁begin|><|tool▁output▁begin|>' + message['content'] + '<|tool▁output▁end|>'}}{%- set ns.is_output_first = false %}{%- else %}{{'\\n<|tool▁output▁begin|>' + message['content'] + '<|tool▁output▁end|>'}}{%- endif %}{%- endif %}{%- endfor -%}{% if ns.is_tool %}{{'<|tool▁outputs▁end|>'}}{% endif %}{% if add_generation_prompt and not ns.is_tool %}{{'<|Assistant|>'}}{% endif %}",
196
+ "clean_up_tokenization_spaces": false,
197
+ "eos_token": "<|end▁of▁sentence|>",
198
+ "extra_special_tokens": {},
199
+ "legacy": true,
200
+ "model_max_length": 4500,
201
+ "pad_token": "<|end▁of▁sentence|>",
202
+ "padding_side": "right",
203
+ "sp_model_kwargs": {},
204
+ "split_special_tokens": false,
205
+ "tokenizer_class": "LlamaTokenizer",
206
+ "unk_token": null,
207
+ "use_default_system_prompt": false
208
+ }
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "total_flos": 3.270639151271117e+16,
4
+ "train_loss": 1.0707874030483013,
5
+ "train_runtime": 199.5779,
6
+ "train_samples_per_second": 3.908,
7
+ "train_steps_per_second": 0.491
8
+ }
trainer_log.jsonl ADDED
@@ -0,0 +1,108 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"current_steps": 1, "total_steps": 98, "loss": 1.1572, "lr": 1e-05, "epoch": 0.01020408163265306, "percentage": 1.02, "elapsed_time": "0:00:02", "remaining_time": "0:03:26"}
2
+ {"current_steps": 2, "total_steps": 98, "loss": 1.3198, "lr": 9.997377845227577e-06, "epoch": 0.02040816326530612, "percentage": 2.04, "elapsed_time": "0:00:04", "remaining_time": "0:03:12"}
3
+ {"current_steps": 3, "total_steps": 98, "loss": 1.0636, "lr": 9.98951413118856e-06, "epoch": 0.030612244897959183, "percentage": 3.06, "elapsed_time": "0:00:05", "remaining_time": "0:03:06"}
4
+ {"current_steps": 4, "total_steps": 98, "loss": 1.2853, "lr": 9.97641710583307e-06, "epoch": 0.04081632653061224, "percentage": 4.08, "elapsed_time": "0:00:07", "remaining_time": "0:03:02"}
5
+ {"current_steps": 5, "total_steps": 98, "loss": 1.0573, "lr": 9.958100506132127e-06, "epoch": 0.05102040816326531, "percentage": 5.1, "elapsed_time": "0:00:09", "remaining_time": "0:02:59"}
6
+ {"current_steps": 6, "total_steps": 98, "loss": 1.5486, "lr": 9.934583543669454e-06, "epoch": 0.061224489795918366, "percentage": 6.12, "elapsed_time": "0:00:11", "remaining_time": "0:02:56"}
7
+ {"current_steps": 7, "total_steps": 98, "loss": 0.8672, "lr": 9.905890884491196e-06, "epoch": 0.07142857142857142, "percentage": 7.14, "elapsed_time": "0:00:13", "remaining_time": "0:02:53"}
8
+ {"current_steps": 8, "total_steps": 98, "loss": 1.132, "lr": 9.872052623234632e-06, "epoch": 0.08163265306122448, "percentage": 8.16, "elapsed_time": "0:00:15", "remaining_time": "0:02:51"}
9
+ {"current_steps": 9, "total_steps": 98, "loss": 1.2419, "lr": 9.833104251563058e-06, "epoch": 0.09183673469387756, "percentage": 9.18, "elapsed_time": "0:00:17", "remaining_time": "0:02:49"}
10
+ {"current_steps": 10, "total_steps": 98, "loss": 0.9946, "lr": 9.789086620939936e-06, "epoch": 0.10204081632653061, "percentage": 10.2, "elapsed_time": "0:00:18", "remaining_time": "0:02:47"}
11
+ {"current_steps": 10, "total_steps": 98, "eval_loss": 0.9807827472686768, "epoch": 0.10204081632653061, "percentage": 10.2, "elapsed_time": "0:00:19", "remaining_time": "0:02:49"}
12
+ {"current_steps": 11, "total_steps": 98, "loss": 1.3435, "lr": 9.740045899781353e-06, "epoch": 0.11224489795918367, "percentage": 11.22, "elapsed_time": "0:00:21", "remaining_time": "0:02:47"}
13
+ {"current_steps": 12, "total_steps": 98, "loss": 0.7463, "lr": 9.68603352503172e-06, "epoch": 0.12244897959183673, "percentage": 12.24, "elapsed_time": "0:00:23", "remaining_time": "0:02:44"}
14
+ {"current_steps": 13, "total_steps": 98, "loss": 1.329, "lr": 9.627106148213521e-06, "epoch": 0.1326530612244898, "percentage": 13.27, "elapsed_time": "0:00:24", "remaining_time": "0:02:42"}
15
+ {"current_steps": 14, "total_steps": 98, "loss": 1.3815, "lr": 9.563325576007702e-06, "epoch": 0.14285714285714285, "percentage": 14.29, "elapsed_time": "0:00:26", "remaining_time": "0:02:40"}
16
+ {"current_steps": 15, "total_steps": 98, "loss": 1.0305, "lr": 9.494758705426978e-06, "epoch": 0.15306122448979592, "percentage": 15.31, "elapsed_time": "0:00:28", "remaining_time": "0:02:38"}
17
+ {"current_steps": 16, "total_steps": 98, "loss": 1.3317, "lr": 9.421477453650118e-06, "epoch": 0.16326530612244897, "percentage": 16.33, "elapsed_time": "0:00:30", "remaining_time": "0:02:36"}
18
+ {"current_steps": 17, "total_steps": 98, "loss": 1.323, "lr": 9.343558682590757e-06, "epoch": 0.17346938775510204, "percentage": 17.35, "elapsed_time": "0:00:32", "remaining_time": "0:02:34"}
19
+ {"current_steps": 18, "total_steps": 98, "loss": 1.5154, "lr": 9.261084118279846e-06, "epoch": 0.1836734693877551, "percentage": 18.37, "elapsed_time": "0:00:34", "remaining_time": "0:02:32"}
20
+ {"current_steps": 19, "total_steps": 98, "loss": 1.2371, "lr": 9.174140265146356e-06, "epoch": 0.19387755102040816, "percentage": 19.39, "elapsed_time": "0:00:36", "remaining_time": "0:02:30"}
21
+ {"current_steps": 20, "total_steps": 98, "loss": 1.042, "lr": 9.082818315286054e-06, "epoch": 0.20408163265306123, "percentage": 20.41, "elapsed_time": "0:00:38", "remaining_time": "0:02:28"}
22
+ {"current_steps": 20, "total_steps": 98, "eval_loss": 0.9510844945907593, "epoch": 0.20408163265306123, "percentage": 20.41, "elapsed_time": "0:00:38", "remaining_time": "0:02:29"}
23
+ {"current_steps": 21, "total_steps": 98, "loss": 1.07, "lr": 8.987214052813605e-06, "epoch": 0.21428571428571427, "percentage": 21.43, "elapsed_time": "0:00:40", "remaining_time": "0:02:27"}
24
+ {"current_steps": 22, "total_steps": 98, "loss": 1.2102, "lr": 8.887427753398249e-06, "epoch": 0.22448979591836735, "percentage": 22.45, "elapsed_time": "0:00:42", "remaining_time": "0:02:25"}
25
+ {"current_steps": 23, "total_steps": 98, "loss": 1.0913, "lr": 8.783564079088478e-06, "epoch": 0.23469387755102042, "percentage": 23.47, "elapsed_time": "0:00:43", "remaining_time": "0:02:23"}
26
+ {"current_steps": 24, "total_steps": 98, "loss": 0.8923, "lr": 8.675731968536004e-06, "epoch": 0.24489795918367346, "percentage": 24.49, "elapsed_time": "0:00:45", "remaining_time": "0:02:21"}
27
+ {"current_steps": 25, "total_steps": 98, "loss": 1.1723, "lr": 8.564044522734147e-06, "epoch": 0.25510204081632654, "percentage": 25.51, "elapsed_time": "0:00:47", "remaining_time": "0:02:19"}
28
+ {"current_steps": 26, "total_steps": 98, "loss": 1.1206, "lr": 8.448618886390523e-06, "epoch": 0.2653061224489796, "percentage": 26.53, "elapsed_time": "0:00:49", "remaining_time": "0:02:17"}
29
+ {"current_steps": 27, "total_steps": 98, "loss": 1.1261, "lr": 8.329576125058406e-06, "epoch": 0.2755102040816326, "percentage": 27.55, "elapsed_time": "0:00:51", "remaining_time": "0:02:15"}
30
+ {"current_steps": 28, "total_steps": 98, "loss": 1.2727, "lr": 8.207041098155701e-06, "epoch": 0.2857142857142857, "percentage": 28.57, "elapsed_time": "0:00:53", "remaining_time": "0:02:13"}
31
+ {"current_steps": 29, "total_steps": 98, "loss": 1.2197, "lr": 8.081142328004638e-06, "epoch": 0.29591836734693877, "percentage": 29.59, "elapsed_time": "0:00:55", "remaining_time": "0:02:11"}
32
+ {"current_steps": 30, "total_steps": 98, "loss": 1.1236, "lr": 7.952011865029614e-06, "epoch": 0.30612244897959184, "percentage": 30.61, "elapsed_time": "0:00:57", "remaining_time": "0:02:09"}
33
+ {"current_steps": 30, "total_steps": 98, "eval_loss": 0.9308969974517822, "epoch": 0.30612244897959184, "percentage": 30.61, "elapsed_time": "0:00:57", "remaining_time": "0:02:09"}
34
+ {"current_steps": 31, "total_steps": 98, "loss": 0.9098, "lr": 7.819785149254534e-06, "epoch": 0.3163265306122449, "percentage": 31.63, "elapsed_time": "0:00:59", "remaining_time": "0:02:07"}
35
+ {"current_steps": 32, "total_steps": 98, "loss": 1.0564, "lr": 7.68460086824492e-06, "epoch": 0.32653061224489793, "percentage": 32.65, "elapsed_time": "0:01:01", "remaining_time": "0:02:05"}
36
+ {"current_steps": 33, "total_steps": 98, "loss": 1.0174, "lr": 7.546600811643816e-06, "epoch": 0.336734693877551, "percentage": 33.67, "elapsed_time": "0:01:02", "remaining_time": "0:02:03"}
37
+ {"current_steps": 34, "total_steps": 98, "loss": 1.1875, "lr": 7.405929722454026e-06, "epoch": 0.3469387755102041, "percentage": 34.69, "elapsed_time": "0:01:04", "remaining_time": "0:02:01"}
38
+ {"current_steps": 35, "total_steps": 98, "loss": 0.9453, "lr": 7.262735145222696e-06, "epoch": 0.35714285714285715, "percentage": 35.71, "elapsed_time": "0:01:06", "remaining_time": "0:01:59"}
39
+ {"current_steps": 36, "total_steps": 98, "loss": 1.0147, "lr": 7.117167271287453e-06, "epoch": 0.3673469387755102, "percentage": 36.73, "elapsed_time": "0:01:08", "remaining_time": "0:01:57"}
40
+ {"current_steps": 37, "total_steps": 98, "loss": 0.995, "lr": 6.969378781246436e-06, "epoch": 0.37755102040816324, "percentage": 37.76, "elapsed_time": "0:01:10", "remaining_time": "0:01:56"}
41
+ {"current_steps": 38, "total_steps": 98, "loss": 0.9636, "lr": 6.819524684817439e-06, "epoch": 0.3877551020408163, "percentage": 38.78, "elapsed_time": "0:01:12", "remaining_time": "0:01:54"}
42
+ {"current_steps": 39, "total_steps": 98, "loss": 1.1511, "lr": 6.667762158254104e-06, "epoch": 0.3979591836734694, "percentage": 39.8, "elapsed_time": "0:01:14", "remaining_time": "0:01:52"}
43
+ {"current_steps": 40, "total_steps": 98, "loss": 1.238, "lr": 6.514250379489754e-06, "epoch": 0.40816326530612246, "percentage": 40.82, "elapsed_time": "0:01:16", "remaining_time": "0:01:50"}
44
+ {"current_steps": 40, "total_steps": 98, "eval_loss": 0.9210565090179443, "epoch": 0.40816326530612246, "percentage": 40.82, "elapsed_time": "0:01:16", "remaining_time": "0:01:50"}
45
+ {"current_steps": 41, "total_steps": 98, "loss": 0.927, "lr": 6.3591503611817155e-06, "epoch": 0.41836734693877553, "percentage": 41.84, "elapsed_time": "0:01:18", "remaining_time": "0:01:48"}
46
+ {"current_steps": 42, "total_steps": 98, "loss": 1.1018, "lr": 6.202624781831269e-06, "epoch": 0.42857142857142855, "percentage": 42.86, "elapsed_time": "0:01:20", "remaining_time": "0:01:46"}
47
+ {"current_steps": 43, "total_steps": 98, "loss": 0.938, "lr": 6.044837815156377e-06, "epoch": 0.4387755102040816, "percentage": 43.88, "elapsed_time": "0:01:21", "remaining_time": "0:01:44"}
48
+ {"current_steps": 44, "total_steps": 98, "loss": 1.0764, "lr": 5.885954957896115e-06, "epoch": 0.4489795918367347, "percentage": 44.9, "elapsed_time": "0:01:23", "remaining_time": "0:01:42"}
49
+ {"current_steps": 45, "total_steps": 98, "loss": 1.2437, "lr": 5.726142856227453e-06, "epoch": 0.45918367346938777, "percentage": 45.92, "elapsed_time": "0:01:25", "remaining_time": "0:01:40"}
50
+ {"current_steps": 46, "total_steps": 98, "loss": 1.0162, "lr": 5.5655691309764225e-06, "epoch": 0.46938775510204084, "percentage": 46.94, "elapsed_time": "0:01:27", "remaining_time": "0:01:38"}
51
+ {"current_steps": 47, "total_steps": 98, "loss": 1.1398, "lr": 5.404402201807022e-06, "epoch": 0.47959183673469385, "percentage": 47.96, "elapsed_time": "0:01:29", "remaining_time": "0:01:36"}
52
+ {"current_steps": 48, "total_steps": 98, "loss": 0.8257, "lr": 5.242811110572243e-06, "epoch": 0.4897959183673469, "percentage": 48.98, "elapsed_time": "0:01:31", "remaining_time": "0:01:35"}
53
+ {"current_steps": 49, "total_steps": 98, "loss": 0.9318, "lr": 5.080965344012509e-06, "epoch": 0.5, "percentage": 50.0, "elapsed_time": "0:01:33", "remaining_time": "0:01:33"}
54
+ {"current_steps": 50, "total_steps": 98, "loss": 0.8178, "lr": 4.919034655987493e-06, "epoch": 0.5102040816326531, "percentage": 51.02, "elapsed_time": "0:01:35", "remaining_time": "0:01:31"}
55
+ {"current_steps": 50, "total_steps": 98, "eval_loss": 0.9119246006011963, "epoch": 0.5102040816326531, "percentage": 51.02, "elapsed_time": "0:01:35", "remaining_time": "0:01:31"}
56
+ {"current_steps": 51, "total_steps": 98, "loss": 1.0659, "lr": 4.757188889427761e-06, "epoch": 0.5204081632653061, "percentage": 52.04, "elapsed_time": "0:01:37", "remaining_time": "0:01:29"}
57
+ {"current_steps": 52, "total_steps": 98, "loss": 1.0016, "lr": 4.59559779819298e-06, "epoch": 0.5306122448979592, "percentage": 53.06, "elapsed_time": "0:01:39", "remaining_time": "0:01:27"}
58
+ {"current_steps": 53, "total_steps": 98, "loss": 1.0541, "lr": 4.434430869023579e-06, "epoch": 0.5408163265306123, "percentage": 54.08, "elapsed_time": "0:01:40", "remaining_time": "0:01:25"}
59
+ {"current_steps": 54, "total_steps": 98, "loss": 0.8034, "lr": 4.27385714377255e-06, "epoch": 0.5510204081632653, "percentage": 55.1, "elapsed_time": "0:01:42", "remaining_time": "0:01:23"}
60
+ {"current_steps": 55, "total_steps": 98, "loss": 0.8833, "lr": 4.1140450421038865e-06, "epoch": 0.5612244897959183, "percentage": 56.12, "elapsed_time": "0:01:44", "remaining_time": "0:01:21"}
61
+ {"current_steps": 56, "total_steps": 98, "loss": 0.8837, "lr": 3.955162184843625e-06, "epoch": 0.5714285714285714, "percentage": 57.14, "elapsed_time": "0:01:46", "remaining_time": "0:01:19"}
62
+ {"current_steps": 57, "total_steps": 98, "loss": 1.0044, "lr": 3.7973752181687336e-06, "epoch": 0.5816326530612245, "percentage": 58.16, "elapsed_time": "0:01:48", "remaining_time": "0:01:17"}
63
+ {"current_steps": 58, "total_steps": 98, "loss": 0.9338, "lr": 3.6408496388182857e-06, "epoch": 0.5918367346938775, "percentage": 59.18, "elapsed_time": "0:01:50", "remaining_time": "0:01:16"}
64
+ {"current_steps": 59, "total_steps": 98, "loss": 1.1868, "lr": 3.4857496205102475e-06, "epoch": 0.6020408163265306, "percentage": 60.2, "elapsed_time": "0:01:52", "remaining_time": "0:01:14"}
65
+ {"current_steps": 60, "total_steps": 98, "loss": 0.951, "lr": 3.3322378417458985e-06, "epoch": 0.6122448979591837, "percentage": 61.22, "elapsed_time": "0:01:54", "remaining_time": "0:01:12"}
66
+ {"current_steps": 60, "total_steps": 98, "eval_loss": 0.9065544605255127, "epoch": 0.6122448979591837, "percentage": 61.22, "elapsed_time": "0:01:54", "remaining_time": "0:01:12"}
67
+ {"current_steps": 61, "total_steps": 98, "loss": 1.2871, "lr": 3.180475315182563e-06, "epoch": 0.6224489795918368, "percentage": 62.24, "elapsed_time": "0:01:56", "remaining_time": "0:01:10"}
68
+ {"current_steps": 62, "total_steps": 98, "loss": 0.9965, "lr": 3.0306212187535653e-06, "epoch": 0.6326530612244898, "percentage": 63.27, "elapsed_time": "0:01:58", "remaining_time": "0:01:08"}
69
+ {"current_steps": 63, "total_steps": 98, "loss": 0.8174, "lr": 2.882832728712551e-06, "epoch": 0.6428571428571429, "percentage": 64.29, "elapsed_time": "0:01:59", "remaining_time": "0:01:06"}
70
+ {"current_steps": 64, "total_steps": 98, "loss": 0.9641, "lr": 2.7372648547773063e-06, "epoch": 0.6530612244897959, "percentage": 65.31, "elapsed_time": "0:02:01", "remaining_time": "0:01:04"}
71
+ {"current_steps": 65, "total_steps": 98, "loss": 0.8579, "lr": 2.594070277545975e-06, "epoch": 0.6632653061224489, "percentage": 66.33, "elapsed_time": "0:02:03", "remaining_time": "0:01:02"}
72
+ {"current_steps": 66, "total_steps": 98, "loss": 0.872, "lr": 2.4533991883561868e-06, "epoch": 0.673469387755102, "percentage": 67.35, "elapsed_time": "0:02:05", "remaining_time": "0:01:00"}
73
+ {"current_steps": 67, "total_steps": 98, "loss": 1.1029, "lr": 2.315399131755081e-06, "epoch": 0.6836734693877551, "percentage": 68.37, "elapsed_time": "0:02:07", "remaining_time": "0:00:58"}
74
+ {"current_steps": 68, "total_steps": 98, "loss": 0.8727, "lr": 2.1802148507454675e-06, "epoch": 0.6938775510204082, "percentage": 69.39, "elapsed_time": "0:02:09", "remaining_time": "0:00:57"}
75
+ {"current_steps": 69, "total_steps": 98, "loss": 0.9211, "lr": 2.0479881349703885e-06, "epoch": 0.7040816326530612, "percentage": 70.41, "elapsed_time": "0:02:11", "remaining_time": "0:00:55"}
76
+ {"current_steps": 70, "total_steps": 98, "loss": 0.7869, "lr": 1.9188576719953635e-06, "epoch": 0.7142857142857143, "percentage": 71.43, "elapsed_time": "0:02:13", "remaining_time": "0:00:53"}
77
+ {"current_steps": 70, "total_steps": 98, "eval_loss": 0.9032285809516907, "epoch": 0.7142857142857143, "percentage": 71.43, "elapsed_time": "0:02:13", "remaining_time": "0:00:53"}
78
+ {"current_steps": 71, "total_steps": 98, "loss": 1.1762, "lr": 1.7929589018443016e-06, "epoch": 0.7244897959183674, "percentage": 72.45, "elapsed_time": "0:02:15", "remaining_time": "0:00:51"}
79
+ {"current_steps": 72, "total_steps": 98, "loss": 1.0489, "lr": 1.6704238749415958e-06, "epoch": 0.7346938775510204, "percentage": 73.47, "elapsed_time": "0:02:17", "remaining_time": "0:00:49"}
80
+ {"current_steps": 73, "total_steps": 98, "loss": 0.8249, "lr": 1.5513811136094786e-06, "epoch": 0.7448979591836735, "percentage": 74.49, "elapsed_time": "0:02:18", "remaining_time": "0:00:47"}
81
+ {"current_steps": 74, "total_steps": 98, "loss": 1.0011, "lr": 1.4359554772658551e-06, "epoch": 0.7551020408163265, "percentage": 75.51, "elapsed_time": "0:02:20", "remaining_time": "0:00:45"}
82
+ {"current_steps": 75, "total_steps": 98, "loss": 1.0774, "lr": 1.3242680314639995e-06, "epoch": 0.7653061224489796, "percentage": 76.53, "elapsed_time": "0:02:22", "remaining_time": "0:00:43"}
83
+ {"current_steps": 76, "total_steps": 98, "loss": 1.1492, "lr": 1.2164359209115235e-06, "epoch": 0.7755102040816326, "percentage": 77.55, "elapsed_time": "0:02:24", "remaining_time": "0:00:41"}
84
+ {"current_steps": 77, "total_steps": 98, "loss": 1.1329, "lr": 1.1125722466017547e-06, "epoch": 0.7857142857142857, "percentage": 78.57, "elapsed_time": "0:02:26", "remaining_time": "0:00:39"}
85
+ {"current_steps": 78, "total_steps": 98, "loss": 0.9375, "lr": 1.012785947186397e-06, "epoch": 0.7959183673469388, "percentage": 79.59, "elapsed_time": "0:02:28", "remaining_time": "0:00:38"}
86
+ {"current_steps": 79, "total_steps": 98, "loss": 1.0936, "lr": 9.171816847139447e-07, "epoch": 0.8061224489795918, "percentage": 80.61, "elapsed_time": "0:02:30", "remaining_time": "0:00:36"}
87
+ {"current_steps": 80, "total_steps": 98, "loss": 1.2626, "lr": 8.258597348536452e-07, "epoch": 0.8163265306122449, "percentage": 81.63, "elapsed_time": "0:02:32", "remaining_time": "0:00:34"}
88
+ {"current_steps": 80, "total_steps": 98, "eval_loss": 0.9012266397476196, "epoch": 0.8163265306122449, "percentage": 81.63, "elapsed_time": "0:02:32", "remaining_time": "0:00:34"}
89
+ {"current_steps": 81, "total_steps": 98, "loss": 1.3479, "lr": 7.389158817201541e-07, "epoch": 0.826530612244898, "percentage": 82.65, "elapsed_time": "0:02:34", "remaining_time": "0:00:32"}
90
+ {"current_steps": 82, "total_steps": 98, "loss": 0.9699, "lr": 6.564413174092443e-07, "epoch": 0.8367346938775511, "percentage": 83.67, "elapsed_time": "0:02:36", "remaining_time": "0:00:30"}
91
+ {"current_steps": 83, "total_steps": 98, "loss": 1.1757, "lr": 5.785225463498828e-07, "epoch": 0.8469387755102041, "percentage": 84.69, "elapsed_time": "0:02:37", "remaining_time": "0:00:28"}
92
+ {"current_steps": 84, "total_steps": 98, "loss": 1.1877, "lr": 5.05241294573024e-07, "epoch": 0.8571428571428571, "percentage": 85.71, "elapsed_time": "0:02:39", "remaining_time": "0:00:26"}
93
+ {"current_steps": 85, "total_steps": 98, "loss": 0.9878, "lr": 4.3667442399229985e-07, "epoch": 0.8673469387755102, "percentage": 86.73, "elapsed_time": "0:02:41", "remaining_time": "0:00:24"}
94
+ {"current_steps": 86, "total_steps": 98, "loss": 1.145, "lr": 3.728938517864794e-07, "epoch": 0.8775510204081632, "percentage": 87.76, "elapsed_time": "0:02:43", "remaining_time": "0:00:22"}
95
+ {"current_steps": 87, "total_steps": 98, "loss": 1.128, "lr": 3.1396647496828245e-07, "epoch": 0.8877551020408163, "percentage": 88.78, "elapsed_time": "0:02:45", "remaining_time": "0:00:20"}
96
+ {"current_steps": 88, "total_steps": 98, "loss": 0.9712, "lr": 2.599541002186479e-07, "epoch": 0.8979591836734694, "percentage": 89.8, "elapsed_time": "0:02:47", "remaining_time": "0:00:19"}
97
+ {"current_steps": 89, "total_steps": 98, "loss": 0.9278, "lr": 2.109133790600648e-07, "epoch": 0.9081632653061225, "percentage": 90.82, "elapsed_time": "0:02:49", "remaining_time": "0:00:17"}
98
+ {"current_steps": 90, "total_steps": 98, "loss": 1.2385, "lr": 1.6689574843694433e-07, "epoch": 0.9183673469387755, "percentage": 91.84, "elapsed_time": "0:02:51", "remaining_time": "0:00:15"}
99
+ {"current_steps": 90, "total_steps": 98, "eval_loss": 0.9008685350418091, "epoch": 0.9183673469387755, "percentage": 91.84, "elapsed_time": "0:02:51", "remaining_time": "0:00:15"}
100
+ {"current_steps": 91, "total_steps": 98, "loss": 1.0866, "lr": 1.2794737676536993e-07, "epoch": 0.9285714285714286, "percentage": 92.86, "elapsed_time": "0:02:53", "remaining_time": "0:00:13"}
101
+ {"current_steps": 92, "total_steps": 98, "loss": 1.0466, "lr": 9.410911550880474e-08, "epoch": 0.9387755102040817, "percentage": 93.88, "elapsed_time": "0:02:55", "remaining_time": "0:00:11"}
102
+ {"current_steps": 93, "total_steps": 98, "loss": 0.8466, "lr": 6.54164563305465e-08, "epoch": 0.9489795918367347, "percentage": 94.9, "elapsed_time": "0:02:56", "remaining_time": "0:00:09"}
103
+ {"current_steps": 94, "total_steps": 98, "loss": 1.1521, "lr": 4.189949386787462e-08, "epoch": 0.9591836734693877, "percentage": 95.92, "elapsed_time": "0:02:58", "remaining_time": "0:00:07"}
104
+ {"current_steps": 95, "total_steps": 98, "loss": 1.1505, "lr": 2.358289416693027e-08, "epoch": 0.9693877551020408, "percentage": 96.94, "elapsed_time": "0:03:00", "remaining_time": "0:00:05"}
105
+ {"current_steps": 96, "total_steps": 98, "loss": 0.9055, "lr": 1.0485868811441757e-08, "epoch": 0.9795918367346939, "percentage": 97.96, "elapsed_time": "0:03:02", "remaining_time": "0:00:03"}
106
+ {"current_steps": 97, "total_steps": 98, "loss": 1.0224, "lr": 2.6221547724253337e-09, "epoch": 0.9897959183673469, "percentage": 98.98, "elapsed_time": "0:03:04", "remaining_time": "0:00:01"}
107
+ {"current_steps": 98, "total_steps": 98, "loss": 0.9534, "lr": 0.0, "epoch": 1.0, "percentage": 100.0, "elapsed_time": "0:03:06", "remaining_time": "0:00:00"}
108
+ {"current_steps": 98, "total_steps": 98, "epoch": 1.0, "percentage": 100.0, "elapsed_time": "0:03:18", "remaining_time": "0:00:00"}
trainer_state.json ADDED
@@ -0,0 +1,800 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 1.0,
5
+ "eval_steps": 10,
6
+ "global_step": 98,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.01020408163265306,
13
+ "grad_norm": 1.5935372114181519,
14
+ "learning_rate": 1e-05,
15
+ "loss": 1.1572,
16
+ "step": 1
17
+ },
18
+ {
19
+ "epoch": 0.02040816326530612,
20
+ "grad_norm": 1.8600682020187378,
21
+ "learning_rate": 9.997377845227577e-06,
22
+ "loss": 1.3198,
23
+ "step": 2
24
+ },
25
+ {
26
+ "epoch": 0.030612244897959183,
27
+ "grad_norm": 1.662075400352478,
28
+ "learning_rate": 9.98951413118856e-06,
29
+ "loss": 1.0636,
30
+ "step": 3
31
+ },
32
+ {
33
+ "epoch": 0.04081632653061224,
34
+ "grad_norm": 1.6947352886199951,
35
+ "learning_rate": 9.97641710583307e-06,
36
+ "loss": 1.2853,
37
+ "step": 4
38
+ },
39
+ {
40
+ "epoch": 0.05102040816326531,
41
+ "grad_norm": 1.231564998626709,
42
+ "learning_rate": 9.958100506132127e-06,
43
+ "loss": 1.0573,
44
+ "step": 5
45
+ },
46
+ {
47
+ "epoch": 0.061224489795918366,
48
+ "grad_norm": 1.550899863243103,
49
+ "learning_rate": 9.934583543669454e-06,
50
+ "loss": 1.5486,
51
+ "step": 6
52
+ },
53
+ {
54
+ "epoch": 0.07142857142857142,
55
+ "grad_norm": 1.017148733139038,
56
+ "learning_rate": 9.905890884491196e-06,
57
+ "loss": 0.8672,
58
+ "step": 7
59
+ },
60
+ {
61
+ "epoch": 0.08163265306122448,
62
+ "grad_norm": 1.2291380167007446,
63
+ "learning_rate": 9.872052623234632e-06,
64
+ "loss": 1.132,
65
+ "step": 8
66
+ },
67
+ {
68
+ "epoch": 0.09183673469387756,
69
+ "grad_norm": 1.1650055646896362,
70
+ "learning_rate": 9.833104251563058e-06,
71
+ "loss": 1.2419,
72
+ "step": 9
73
+ },
74
+ {
75
+ "epoch": 0.10204081632653061,
76
+ "grad_norm": 0.9753865003585815,
77
+ "learning_rate": 9.789086620939936e-06,
78
+ "loss": 0.9946,
79
+ "step": 10
80
+ },
81
+ {
82
+ "epoch": 0.10204081632653061,
83
+ "eval_loss": 0.9807827472686768,
84
+ "eval_runtime": 0.2604,
85
+ "eval_samples_per_second": 30.717,
86
+ "eval_steps_per_second": 3.84,
87
+ "step": 10
88
+ },
89
+ {
90
+ "epoch": 0.11224489795918367,
91
+ "grad_norm": 1.191615343093872,
92
+ "learning_rate": 9.740045899781353e-06,
93
+ "loss": 1.3435,
94
+ "step": 11
95
+ },
96
+ {
97
+ "epoch": 0.12244897959183673,
98
+ "grad_norm": 0.9799769520759583,
99
+ "learning_rate": 9.68603352503172e-06,
100
+ "loss": 0.7463,
101
+ "step": 12
102
+ },
103
+ {
104
+ "epoch": 0.1326530612244898,
105
+ "grad_norm": 1.0710457563400269,
106
+ "learning_rate": 9.627106148213521e-06,
107
+ "loss": 1.329,
108
+ "step": 13
109
+ },
110
+ {
111
+ "epoch": 0.14285714285714285,
112
+ "grad_norm": 1.071432113647461,
113
+ "learning_rate": 9.563325576007702e-06,
114
+ "loss": 1.3815,
115
+ "step": 14
116
+ },
117
+ {
118
+ "epoch": 0.15306122448979592,
119
+ "grad_norm": 0.9442039132118225,
120
+ "learning_rate": 9.494758705426978e-06,
121
+ "loss": 1.0305,
122
+ "step": 15
123
+ },
124
+ {
125
+ "epoch": 0.16326530612244897,
126
+ "grad_norm": 1.022035837173462,
127
+ "learning_rate": 9.421477453650118e-06,
128
+ "loss": 1.3317,
129
+ "step": 16
130
+ },
131
+ {
132
+ "epoch": 0.17346938775510204,
133
+ "grad_norm": 1.0751738548278809,
134
+ "learning_rate": 9.343558682590757e-06,
135
+ "loss": 1.323,
136
+ "step": 17
137
+ },
138
+ {
139
+ "epoch": 0.1836734693877551,
140
+ "grad_norm": 1.0566898584365845,
141
+ "learning_rate": 9.261084118279846e-06,
142
+ "loss": 1.5154,
143
+ "step": 18
144
+ },
145
+ {
146
+ "epoch": 0.19387755102040816,
147
+ "grad_norm": 1.0009628534317017,
148
+ "learning_rate": 9.174140265146356e-06,
149
+ "loss": 1.2371,
150
+ "step": 19
151
+ },
152
+ {
153
+ "epoch": 0.20408163265306123,
154
+ "grad_norm": 0.9259373545646667,
155
+ "learning_rate": 9.082818315286054e-06,
156
+ "loss": 1.042,
157
+ "step": 20
158
+ },
159
+ {
160
+ "epoch": 0.20408163265306123,
161
+ "eval_loss": 0.9510844945907593,
162
+ "eval_runtime": 0.2601,
163
+ "eval_samples_per_second": 30.76,
164
+ "eval_steps_per_second": 3.845,
165
+ "step": 20
166
+ },
167
+ {
168
+ "epoch": 0.21428571428571427,
169
+ "grad_norm": 0.9078995585441589,
170
+ "learning_rate": 8.987214052813605e-06,
171
+ "loss": 1.07,
172
+ "step": 21
173
+ },
174
+ {
175
+ "epoch": 0.22448979591836735,
176
+ "grad_norm": 0.9820942282676697,
177
+ "learning_rate": 8.887427753398249e-06,
178
+ "loss": 1.2102,
179
+ "step": 22
180
+ },
181
+ {
182
+ "epoch": 0.23469387755102042,
183
+ "grad_norm": 1.0587691068649292,
184
+ "learning_rate": 8.783564079088478e-06,
185
+ "loss": 1.0913,
186
+ "step": 23
187
+ },
188
+ {
189
+ "epoch": 0.24489795918367346,
190
+ "grad_norm": 0.8137860894203186,
191
+ "learning_rate": 8.675731968536004e-06,
192
+ "loss": 0.8923,
193
+ "step": 24
194
+ },
195
+ {
196
+ "epoch": 0.25510204081632654,
197
+ "grad_norm": 0.9471170902252197,
198
+ "learning_rate": 8.564044522734147e-06,
199
+ "loss": 1.1723,
200
+ "step": 25
201
+ },
202
+ {
203
+ "epoch": 0.2653061224489796,
204
+ "grad_norm": 0.8885685205459595,
205
+ "learning_rate": 8.448618886390523e-06,
206
+ "loss": 1.1206,
207
+ "step": 26
208
+ },
209
+ {
210
+ "epoch": 0.2755102040816326,
211
+ "grad_norm": 0.8613781332969666,
212
+ "learning_rate": 8.329576125058406e-06,
213
+ "loss": 1.1261,
214
+ "step": 27
215
+ },
216
+ {
217
+ "epoch": 0.2857142857142857,
218
+ "grad_norm": 0.9407626986503601,
219
+ "learning_rate": 8.207041098155701e-06,
220
+ "loss": 1.2727,
221
+ "step": 28
222
+ },
223
+ {
224
+ "epoch": 0.29591836734693877,
225
+ "grad_norm": 0.8954651951789856,
226
+ "learning_rate": 8.081142328004638e-06,
227
+ "loss": 1.2197,
228
+ "step": 29
229
+ },
230
+ {
231
+ "epoch": 0.30612244897959184,
232
+ "grad_norm": 0.8864608407020569,
233
+ "learning_rate": 7.952011865029614e-06,
234
+ "loss": 1.1236,
235
+ "step": 30
236
+ },
237
+ {
238
+ "epoch": 0.30612244897959184,
239
+ "eval_loss": 0.9308969974517822,
240
+ "eval_runtime": 0.2599,
241
+ "eval_samples_per_second": 30.775,
242
+ "eval_steps_per_second": 3.847,
243
+ "step": 30
244
+ },
245
+ {
246
+ "epoch": 0.3163265306122449,
247
+ "grad_norm": 0.8076571226119995,
248
+ "learning_rate": 7.819785149254534e-06,
249
+ "loss": 0.9098,
250
+ "step": 31
251
+ },
252
+ {
253
+ "epoch": 0.32653061224489793,
254
+ "grad_norm": 0.8902921080589294,
255
+ "learning_rate": 7.68460086824492e-06,
256
+ "loss": 1.0564,
257
+ "step": 32
258
+ },
259
+ {
260
+ "epoch": 0.336734693877551,
261
+ "grad_norm": 0.9010928273200989,
262
+ "learning_rate": 7.546600811643816e-06,
263
+ "loss": 1.0174,
264
+ "step": 33
265
+ },
266
+ {
267
+ "epoch": 0.3469387755102041,
268
+ "grad_norm": 0.8672641515731812,
269
+ "learning_rate": 7.405929722454026e-06,
270
+ "loss": 1.1875,
271
+ "step": 34
272
+ },
273
+ {
274
+ "epoch": 0.35714285714285715,
275
+ "grad_norm": 0.7878830432891846,
276
+ "learning_rate": 7.262735145222696e-06,
277
+ "loss": 0.9453,
278
+ "step": 35
279
+ },
280
+ {
281
+ "epoch": 0.3673469387755102,
282
+ "grad_norm": 0.85221266746521,
283
+ "learning_rate": 7.117167271287453e-06,
284
+ "loss": 1.0147,
285
+ "step": 36
286
+ },
287
+ {
288
+ "epoch": 0.37755102040816324,
289
+ "grad_norm": 0.8591777682304382,
290
+ "learning_rate": 6.969378781246436e-06,
291
+ "loss": 0.995,
292
+ "step": 37
293
+ },
294
+ {
295
+ "epoch": 0.3877551020408163,
296
+ "grad_norm": 0.9585978388786316,
297
+ "learning_rate": 6.819524684817439e-06,
298
+ "loss": 0.9636,
299
+ "step": 38
300
+ },
301
+ {
302
+ "epoch": 0.3979591836734694,
303
+ "grad_norm": 0.9561256170272827,
304
+ "learning_rate": 6.667762158254104e-06,
305
+ "loss": 1.1511,
306
+ "step": 39
307
+ },
308
+ {
309
+ "epoch": 0.40816326530612246,
310
+ "grad_norm": 0.9076462388038635,
311
+ "learning_rate": 6.514250379489754e-06,
312
+ "loss": 1.238,
313
+ "step": 40
314
+ },
315
+ {
316
+ "epoch": 0.40816326530612246,
317
+ "eval_loss": 0.9210565090179443,
318
+ "eval_runtime": 0.26,
319
+ "eval_samples_per_second": 30.764,
320
+ "eval_steps_per_second": 3.846,
321
+ "step": 40
322
+ },
323
+ {
324
+ "epoch": 0.41836734693877553,
325
+ "grad_norm": 0.779888391494751,
326
+ "learning_rate": 6.3591503611817155e-06,
327
+ "loss": 0.927,
328
+ "step": 41
329
+ },
330
+ {
331
+ "epoch": 0.42857142857142855,
332
+ "grad_norm": 0.826004147529602,
333
+ "learning_rate": 6.202624781831269e-06,
334
+ "loss": 1.1018,
335
+ "step": 42
336
+ },
337
+ {
338
+ "epoch": 0.4387755102040816,
339
+ "grad_norm": 0.8641281127929688,
340
+ "learning_rate": 6.044837815156377e-06,
341
+ "loss": 0.938,
342
+ "step": 43
343
+ },
344
+ {
345
+ "epoch": 0.4489795918367347,
346
+ "grad_norm": 0.896023154258728,
347
+ "learning_rate": 5.885954957896115e-06,
348
+ "loss": 1.0764,
349
+ "step": 44
350
+ },
351
+ {
352
+ "epoch": 0.45918367346938777,
353
+ "grad_norm": 0.9072222709655762,
354
+ "learning_rate": 5.726142856227453e-06,
355
+ "loss": 1.2437,
356
+ "step": 45
357
+ },
358
+ {
359
+ "epoch": 0.46938775510204084,
360
+ "grad_norm": 0.7771997451782227,
361
+ "learning_rate": 5.5655691309764225e-06,
362
+ "loss": 1.0162,
363
+ "step": 46
364
+ },
365
+ {
366
+ "epoch": 0.47959183673469385,
367
+ "grad_norm": 0.8864902853965759,
368
+ "learning_rate": 5.404402201807022e-06,
369
+ "loss": 1.1398,
370
+ "step": 47
371
+ },
372
+ {
373
+ "epoch": 0.4897959183673469,
374
+ "grad_norm": 0.702143669128418,
375
+ "learning_rate": 5.242811110572243e-06,
376
+ "loss": 0.8257,
377
+ "step": 48
378
+ },
379
+ {
380
+ "epoch": 0.5,
381
+ "grad_norm": 0.7524011731147766,
382
+ "learning_rate": 5.080965344012509e-06,
383
+ "loss": 0.9318,
384
+ "step": 49
385
+ },
386
+ {
387
+ "epoch": 0.5102040816326531,
388
+ "grad_norm": 0.7233298420906067,
389
+ "learning_rate": 4.919034655987493e-06,
390
+ "loss": 0.8178,
391
+ "step": 50
392
+ },
393
+ {
394
+ "epoch": 0.5102040816326531,
395
+ "eval_loss": 0.9119246006011963,
396
+ "eval_runtime": 0.2595,
397
+ "eval_samples_per_second": 30.823,
398
+ "eval_steps_per_second": 3.853,
399
+ "step": 50
400
+ },
401
+ {
402
+ "epoch": 0.5204081632653061,
403
+ "grad_norm": 0.8432551622390747,
404
+ "learning_rate": 4.757188889427761e-06,
405
+ "loss": 1.0659,
406
+ "step": 51
407
+ },
408
+ {
409
+ "epoch": 0.5306122448979592,
410
+ "grad_norm": 0.772146999835968,
411
+ "learning_rate": 4.59559779819298e-06,
412
+ "loss": 1.0016,
413
+ "step": 52
414
+ },
415
+ {
416
+ "epoch": 0.5408163265306123,
417
+ "grad_norm": 0.8176573514938354,
418
+ "learning_rate": 4.434430869023579e-06,
419
+ "loss": 1.0541,
420
+ "step": 53
421
+ },
422
+ {
423
+ "epoch": 0.5510204081632653,
424
+ "grad_norm": 0.730097770690918,
425
+ "learning_rate": 4.27385714377255e-06,
426
+ "loss": 0.8034,
427
+ "step": 54
428
+ },
429
+ {
430
+ "epoch": 0.5612244897959183,
431
+ "grad_norm": 0.7190036177635193,
432
+ "learning_rate": 4.1140450421038865e-06,
433
+ "loss": 0.8833,
434
+ "step": 55
435
+ },
436
+ {
437
+ "epoch": 0.5714285714285714,
438
+ "grad_norm": 0.732845664024353,
439
+ "learning_rate": 3.955162184843625e-06,
440
+ "loss": 0.8837,
441
+ "step": 56
442
+ },
443
+ {
444
+ "epoch": 0.5816326530612245,
445
+ "grad_norm": 0.770910918712616,
446
+ "learning_rate": 3.7973752181687336e-06,
447
+ "loss": 1.0044,
448
+ "step": 57
449
+ },
450
+ {
451
+ "epoch": 0.5918367346938775,
452
+ "grad_norm": 0.7635073065757751,
453
+ "learning_rate": 3.6408496388182857e-06,
454
+ "loss": 0.9338,
455
+ "step": 58
456
+ },
457
+ {
458
+ "epoch": 0.6020408163265306,
459
+ "grad_norm": 0.8665493130683899,
460
+ "learning_rate": 3.4857496205102475e-06,
461
+ "loss": 1.1868,
462
+ "step": 59
463
+ },
464
+ {
465
+ "epoch": 0.6122448979591837,
466
+ "grad_norm": 0.7718795537948608,
467
+ "learning_rate": 3.3322378417458985e-06,
468
+ "loss": 0.951,
469
+ "step": 60
470
+ },
471
+ {
472
+ "epoch": 0.6122448979591837,
473
+ "eval_loss": 0.9065544605255127,
474
+ "eval_runtime": 0.2595,
475
+ "eval_samples_per_second": 30.831,
476
+ "eval_steps_per_second": 3.854,
477
+ "step": 60
478
+ },
479
+ {
480
+ "epoch": 0.6224489795918368,
481
+ "grad_norm": 0.932878851890564,
482
+ "learning_rate": 3.180475315182563e-06,
483
+ "loss": 1.2871,
484
+ "step": 61
485
+ },
486
+ {
487
+ "epoch": 0.6326530612244898,
488
+ "grad_norm": 0.8946093320846558,
489
+ "learning_rate": 3.0306212187535653e-06,
490
+ "loss": 0.9965,
491
+ "step": 62
492
+ },
493
+ {
494
+ "epoch": 0.6428571428571429,
495
+ "grad_norm": 0.7468050718307495,
496
+ "learning_rate": 2.882832728712551e-06,
497
+ "loss": 0.8174,
498
+ "step": 63
499
+ },
500
+ {
501
+ "epoch": 0.6530612244897959,
502
+ "grad_norm": 0.7200383543968201,
503
+ "learning_rate": 2.7372648547773063e-06,
504
+ "loss": 0.9641,
505
+ "step": 64
506
+ },
507
+ {
508
+ "epoch": 0.6632653061224489,
509
+ "grad_norm": 0.7331112623214722,
510
+ "learning_rate": 2.594070277545975e-06,
511
+ "loss": 0.8579,
512
+ "step": 65
513
+ },
514
+ {
515
+ "epoch": 0.673469387755102,
516
+ "grad_norm": 0.713594377040863,
517
+ "learning_rate": 2.4533991883561868e-06,
518
+ "loss": 0.872,
519
+ "step": 66
520
+ },
521
+ {
522
+ "epoch": 0.6836734693877551,
523
+ "grad_norm": 0.8468301296234131,
524
+ "learning_rate": 2.315399131755081e-06,
525
+ "loss": 1.1029,
526
+ "step": 67
527
+ },
528
+ {
529
+ "epoch": 0.6938775510204082,
530
+ "grad_norm": 0.7141689658164978,
531
+ "learning_rate": 2.1802148507454675e-06,
532
+ "loss": 0.8727,
533
+ "step": 68
534
+ },
535
+ {
536
+ "epoch": 0.7040816326530612,
537
+ "grad_norm": 0.7567382454872131,
538
+ "learning_rate": 2.0479881349703885e-06,
539
+ "loss": 0.9211,
540
+ "step": 69
541
+ },
542
+ {
543
+ "epoch": 0.7142857142857143,
544
+ "grad_norm": 0.7020692825317383,
545
+ "learning_rate": 1.9188576719953635e-06,
546
+ "loss": 0.7869,
547
+ "step": 70
548
+ },
549
+ {
550
+ "epoch": 0.7142857142857143,
551
+ "eval_loss": 0.9032285809516907,
552
+ "eval_runtime": 0.2594,
553
+ "eval_samples_per_second": 30.835,
554
+ "eval_steps_per_second": 3.854,
555
+ "step": 70
556
+ },
557
+ {
558
+ "epoch": 0.7244897959183674,
559
+ "grad_norm": 0.853243350982666,
560
+ "learning_rate": 1.7929589018443016e-06,
561
+ "loss": 1.1762,
562
+ "step": 71
563
+ },
564
+ {
565
+ "epoch": 0.7346938775510204,
566
+ "grad_norm": 0.8287662267684937,
567
+ "learning_rate": 1.6704238749415958e-06,
568
+ "loss": 1.0489,
569
+ "step": 72
570
+ },
571
+ {
572
+ "epoch": 0.7448979591836735,
573
+ "grad_norm": 0.6973975300788879,
574
+ "learning_rate": 1.5513811136094786e-06,
575
+ "loss": 0.8249,
576
+ "step": 73
577
+ },
578
+ {
579
+ "epoch": 0.7551020408163265,
580
+ "grad_norm": 0.8150553703308105,
581
+ "learning_rate": 1.4359554772658551e-06,
582
+ "loss": 1.0011,
583
+ "step": 74
584
+ },
585
+ {
586
+ "epoch": 0.7653061224489796,
587
+ "grad_norm": 0.7974684238433838,
588
+ "learning_rate": 1.3242680314639995e-06,
589
+ "loss": 1.0774,
590
+ "step": 75
591
+ },
592
+ {
593
+ "epoch": 0.7755102040816326,
594
+ "grad_norm": 0.8233014941215515,
595
+ "learning_rate": 1.2164359209115235e-06,
596
+ "loss": 1.1492,
597
+ "step": 76
598
+ },
599
+ {
600
+ "epoch": 0.7857142857142857,
601
+ "grad_norm": 0.8305495977401733,
602
+ "learning_rate": 1.1125722466017547e-06,
603
+ "loss": 1.1329,
604
+ "step": 77
605
+ },
606
+ {
607
+ "epoch": 0.7959183673469388,
608
+ "grad_norm": 0.7523290514945984,
609
+ "learning_rate": 1.012785947186397e-06,
610
+ "loss": 0.9375,
611
+ "step": 78
612
+ },
613
+ {
614
+ "epoch": 0.8061224489795918,
615
+ "grad_norm": 0.8073377013206482,
616
+ "learning_rate": 9.171816847139447e-07,
617
+ "loss": 1.0936,
618
+ "step": 79
619
+ },
620
+ {
621
+ "epoch": 0.8163265306122449,
622
+ "grad_norm": 0.8685086965560913,
623
+ "learning_rate": 8.258597348536452e-07,
624
+ "loss": 1.2626,
625
+ "step": 80
626
+ },
627
+ {
628
+ "epoch": 0.8163265306122449,
629
+ "eval_loss": 0.9012266397476196,
630
+ "eval_runtime": 0.2595,
631
+ "eval_samples_per_second": 30.833,
632
+ "eval_steps_per_second": 3.854,
633
+ "step": 80
634
+ },
635
+ {
636
+ "epoch": 0.826530612244898,
637
+ "grad_norm": 0.877649188041687,
638
+ "learning_rate": 7.389158817201541e-07,
639
+ "loss": 1.3479,
640
+ "step": 81
641
+ },
642
+ {
643
+ "epoch": 0.8367346938775511,
644
+ "grad_norm": 0.7652468085289001,
645
+ "learning_rate": 6.564413174092443e-07,
646
+ "loss": 0.9699,
647
+ "step": 82
648
+ },
649
+ {
650
+ "epoch": 0.8469387755102041,
651
+ "grad_norm": 0.8436093926429749,
652
+ "learning_rate": 5.785225463498828e-07,
653
+ "loss": 1.1757,
654
+ "step": 83
655
+ },
656
+ {
657
+ "epoch": 0.8571428571428571,
658
+ "grad_norm": 0.8732744455337524,
659
+ "learning_rate": 5.05241294573024e-07,
660
+ "loss": 1.1877,
661
+ "step": 84
662
+ },
663
+ {
664
+ "epoch": 0.8673469387755102,
665
+ "grad_norm": 0.7433239221572876,
666
+ "learning_rate": 4.3667442399229985e-07,
667
+ "loss": 0.9878,
668
+ "step": 85
669
+ },
670
+ {
671
+ "epoch": 0.8775510204081632,
672
+ "grad_norm": 0.8225981593132019,
673
+ "learning_rate": 3.728938517864794e-07,
674
+ "loss": 1.145,
675
+ "step": 86
676
+ },
677
+ {
678
+ "epoch": 0.8877551020408163,
679
+ "grad_norm": 0.8591734170913696,
680
+ "learning_rate": 3.1396647496828245e-07,
681
+ "loss": 1.128,
682
+ "step": 87
683
+ },
684
+ {
685
+ "epoch": 0.8979591836734694,
686
+ "grad_norm": 0.7502849102020264,
687
+ "learning_rate": 2.599541002186479e-07,
688
+ "loss": 0.9712,
689
+ "step": 88
690
+ },
691
+ {
692
+ "epoch": 0.9081632653061225,
693
+ "grad_norm": 0.7429087162017822,
694
+ "learning_rate": 2.109133790600648e-07,
695
+ "loss": 0.9278,
696
+ "step": 89
697
+ },
698
+ {
699
+ "epoch": 0.9183673469387755,
700
+ "grad_norm": 0.8890379667282104,
701
+ "learning_rate": 1.6689574843694433e-07,
702
+ "loss": 1.2385,
703
+ "step": 90
704
+ },
705
+ {
706
+ "epoch": 0.9183673469387755,
707
+ "eval_loss": 0.9008685350418091,
708
+ "eval_runtime": 0.2594,
709
+ "eval_samples_per_second": 30.837,
710
+ "eval_steps_per_second": 3.855,
711
+ "step": 90
712
+ },
713
+ {
714
+ "epoch": 0.9285714285714286,
715
+ "grad_norm": 0.9237527251243591,
716
+ "learning_rate": 1.2794737676536993e-07,
717
+ "loss": 1.0866,
718
+ "step": 91
719
+ },
720
+ {
721
+ "epoch": 0.9387755102040817,
722
+ "grad_norm": 0.8110918402671814,
723
+ "learning_rate": 9.410911550880474e-08,
724
+ "loss": 1.0466,
725
+ "step": 92
726
+ },
727
+ {
728
+ "epoch": 0.9489795918367347,
729
+ "grad_norm": 0.7134807109832764,
730
+ "learning_rate": 6.54164563305465e-08,
731
+ "loss": 0.8466,
732
+ "step": 93
733
+ },
734
+ {
735
+ "epoch": 0.9591836734693877,
736
+ "grad_norm": 0.8616638779640198,
737
+ "learning_rate": 4.189949386787462e-08,
738
+ "loss": 1.1521,
739
+ "step": 94
740
+ },
741
+ {
742
+ "epoch": 0.9693877551020408,
743
+ "grad_norm": 0.9501765966415405,
744
+ "learning_rate": 2.358289416693027e-08,
745
+ "loss": 1.1505,
746
+ "step": 95
747
+ },
748
+ {
749
+ "epoch": 0.9795918367346939,
750
+ "grad_norm": 0.7652553915977478,
751
+ "learning_rate": 1.0485868811441757e-08,
752
+ "loss": 0.9055,
753
+ "step": 96
754
+ },
755
+ {
756
+ "epoch": 0.9897959183673469,
757
+ "grad_norm": 0.862601637840271,
758
+ "learning_rate": 2.6221547724253337e-09,
759
+ "loss": 1.0224,
760
+ "step": 97
761
+ },
762
+ {
763
+ "epoch": 1.0,
764
+ "grad_norm": 0.712780773639679,
765
+ "learning_rate": 0.0,
766
+ "loss": 0.9534,
767
+ "step": 98
768
+ },
769
+ {
770
+ "epoch": 1.0,
771
+ "step": 98,
772
+ "total_flos": 3.270639151271117e+16,
773
+ "train_loss": 1.0707874030483013,
774
+ "train_runtime": 199.5779,
775
+ "train_samples_per_second": 3.908,
776
+ "train_steps_per_second": 0.491
777
+ }
778
+ ],
779
+ "logging_steps": 1,
780
+ "max_steps": 98,
781
+ "num_input_tokens_seen": 0,
782
+ "num_train_epochs": 1,
783
+ "save_steps": 98,
784
+ "stateful_callbacks": {
785
+ "TrainerControl": {
786
+ "args": {
787
+ "should_epoch_stop": false,
788
+ "should_evaluate": false,
789
+ "should_log": false,
790
+ "should_save": true,
791
+ "should_training_stop": true
792
+ },
793
+ "attributes": {}
794
+ }
795
+ },
796
+ "total_flos": 3.270639151271117e+16,
797
+ "train_batch_size": 1,
798
+ "trial_name": null,
799
+ "trial_params": null
800
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05ff4a2daceec38620085408ea7ead1959ffc0205553c71886db48306f253159
3
+ size 7416
training_eval_loss.png ADDED
training_loss.png ADDED