Minbyul commited on
Commit
319f07e
1 Parent(s): bb5cdad

Model save

Browse files
README.md CHANGED
@@ -2,15 +2,9 @@
2
  license: apache-2.0
3
  base_model: Minbyul/biomistral-7b-wo-kqa_golden-iter-dpo-step2
4
  tags:
5
- - alignment-handbook
6
  - trl
7
  - dpo
8
  - generated_from_trainer
9
- - trl
10
- - dpo
11
- - generated_from_trainer
12
- datasets:
13
- - HuggingFaceH4/ultrafeedback_binarized
14
  model-index:
15
  - name: biomistral-7b-wo-kqa_golden-iter-dpo-step3
16
  results: []
@@ -21,17 +15,7 @@ should probably proofread and complete it, then remove this comment. -->
21
 
22
  # biomistral-7b-wo-kqa_golden-iter-dpo-step3
23
 
24
- This model is a fine-tuned version of [Minbyul/biomistral-7b-wo-kqa_golden-iter-dpo-step2](https://huggingface.co/Minbyul/biomistral-7b-wo-kqa_golden-iter-dpo-step2) on the HuggingFaceH4/ultrafeedback_binarized dataset.
25
- It achieves the following results on the evaluation set:
26
- - Loss: 0.6829
27
- - Rewards/chosen: -0.0502
28
- - Rewards/rejected: -0.0800
29
- - Rewards/accuracies: 0.6300
30
- - Rewards/margins: 0.0298
31
- - Logps/rejected: -60.1185
32
- - Logps/chosen: -40.1264
33
- - Logits/rejected: -1.5228
34
- - Logits/chosen: -0.8710
35
 
36
  ## Model description
37
 
@@ -66,15 +50,11 @@ The following hyperparameters were used during training:
66
 
67
  ### Training results
68
 
69
- | Training Loss | Epoch | Step | Logits/chosen | Logits/rejected | Logps/chosen | Logps/rejected | Validation Loss | Rewards/accuracies | Rewards/chosen | Rewards/margins | Rewards/rejected |
70
- |:-------------:|:-----:|:----:|:-------------:|:---------------:|:------------:|:--------------:|:---------------:|:------------------:|:--------------:|:---------------:|:----------------:|
71
- | 0.6794 | 0.37 | 100 | -0.8266 | -1.4757 | -35.5860 | -53.0765 | 0.6906 | 0.5900 | -0.0048 | 0.0048 | -0.0096 |
72
- | 0.6555 | 0.74 | 200 | -0.8589 | -1.5130 | -39.0432 | -58.7210 | 0.6837 | 0.6400 | -0.0394 | 0.0267 | -0.0661 |
73
 
74
 
75
  ### Framework versions
76
 
77
- - Transformers 4.39.0.dev0
78
- - Pytorch 2.1.2
79
  - Datasets 2.14.6
80
  - Tokenizers 0.15.2
 
2
  license: apache-2.0
3
  base_model: Minbyul/biomistral-7b-wo-kqa_golden-iter-dpo-step2
4
  tags:
 
5
  - trl
6
  - dpo
7
  - generated_from_trainer
 
 
 
 
 
8
  model-index:
9
  - name: biomistral-7b-wo-kqa_golden-iter-dpo-step3
10
  results: []
 
15
 
16
  # biomistral-7b-wo-kqa_golden-iter-dpo-step3
17
 
18
+ This model is a fine-tuned version of [Minbyul/biomistral-7b-wo-kqa_golden-iter-dpo-step2](https://huggingface.co/Minbyul/biomistral-7b-wo-kqa_golden-iter-dpo-step2) on an unknown dataset.
 
 
 
 
 
 
 
 
 
 
19
 
20
  ## Model description
21
 
 
50
 
51
  ### Training results
52
 
 
 
 
 
53
 
54
 
55
  ### Framework versions
56
 
57
+ - Transformers 4.38.2
58
+ - Pytorch 2.1.2+cu121
59
  - Datasets 2.14.6
60
  - Tokenizers 0.15.2
all_results.json CHANGED
@@ -1,21 +1,8 @@
1
  {
2
- "epoch": 1.0,
3
- "eval_logits/chosen": -0.8710112571716309,
4
- "eval_logits/rejected": -1.522813320159912,
5
- "eval_logps/chosen": -40.126407623291016,
6
- "eval_logps/rejected": -60.11854934692383,
7
- "eval_loss": 0.6829047799110413,
8
- "eval_rewards/accuracies": 0.6299999952316284,
9
- "eval_rewards/chosen": -0.050214797258377075,
10
- "eval_rewards/margins": 0.02983289770781994,
11
- "eval_rewards/rejected": -0.08004770427942276,
12
- "eval_runtime": 65.4744,
13
- "eval_samples": 785,
14
- "eval_samples_per_second": 11.989,
15
- "eval_steps_per_second": 0.382,
16
- "train_loss": 0.1692200431117305,
17
- "train_runtime": 798.9647,
18
- "train_samples": 17264,
19
- "train_samples_per_second": 21.608,
20
- "train_steps_per_second": 0.338
21
  }
 
1
  {
2
+ "epoch": 0.99,
3
+ "train_loss": 0.6816894855905087,
4
+ "train_runtime": 1070.9433,
5
+ "train_samples": 6039,
6
+ "train_samples_per_second": 5.639,
7
+ "train_steps_per_second": 0.088
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  }
config.json CHANGED
@@ -20,7 +20,7 @@
20
  "sliding_window": 4096,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
- "transformers_version": "4.39.0.dev0",
24
- "use_cache": true,
25
  "vocab_size": 32000
26
  }
 
20
  "sliding_window": 4096,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
+ "transformers_version": "4.38.2",
24
+ "use_cache": false,
25
  "vocab_size": 32000
26
  }
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
- "transformers_version": "4.39.0.dev0"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
+ "transformers_version": "4.38.2"
6
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cce409cff4722b9a731610eee4285f50c9f2d5b36cf9e7d2481b206ae57402a9
3
  size 4943162336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75ddb36d2d5d2523812b20bb5a38e56c6ac47fb4338d8969c27b520509a2f32e
3
  size 4943162336
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8187324d72ff88a59d6bc9c26004792edd0b87978a47298a811599baa6429504
3
  size 4999819336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:105af67768a394930d011db6704a0f3abc9425e70d63e13d82d2c002e8efd97c
3
  size 4999819336
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a53d580d5ce1226972d503f03c4c98a758c0b21e66e24742d3e4d812d193cc5d
3
  size 4540516344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5421445557816c21ef40cf96ce4d7a8058cac224c2456363aac443ed153aca39
3
  size 4540516344
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 1.0,
3
- "train_loss": 0.1692200431117305,
4
- "train_runtime": 798.9647,
5
- "train_samples": 17264,
6
- "train_samples_per_second": 21.608,
7
- "train_steps_per_second": 0.338
8
  }
 
1
  {
2
+ "epoch": 0.99,
3
+ "train_loss": 0.6816894855905087,
4
+ "train_runtime": 1070.9433,
5
+ "train_samples": 6039,
6
+ "train_samples_per_second": 5.639,
7
+ "train_steps_per_second": 0.088
8
  }
trainer_state.json CHANGED
@@ -1,21 +1,21 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.0,
5
- "eval_steps": 100,
6
- "global_step": 270,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.0,
13
- "grad_norm": 41.70182769977995,
14
- "learning_rate": 3.703703703703703e-09,
15
- "logits/chosen": -1.5467987060546875,
16
- "logits/rejected": -1.308578372001648,
17
- "logps/chosen": -37.12041473388672,
18
- "logps/rejected": -61.68842315673828,
19
  "loss": 0.6931,
20
  "rewards/accuracies": 0.0,
21
  "rewards/chosen": 0.0,
@@ -24,457 +24,155 @@
24
  "step": 1
25
  },
26
  {
27
- "epoch": 0.04,
28
- "grad_norm": 37.047557637653945,
29
- "learning_rate": 3.7037037037037036e-08,
30
- "logits/chosen": -0.9453011751174927,
31
- "logits/rejected": -1.405588984489441,
32
- "logps/chosen": -28.646324157714844,
33
- "logps/rejected": -40.27184295654297,
34
- "loss": 0.693,
35
  "rewards/accuracies": 0.4930555522441864,
36
- "rewards/chosen": -0.00016396207502111793,
37
- "rewards/margins": 0.001127531286329031,
38
- "rewards/rejected": -0.0012914934195578098,
39
  "step": 10
40
  },
41
  {
42
- "epoch": 0.07,
43
- "grad_norm": 38.71986120422308,
44
- "learning_rate": 7.407407407407407e-08,
45
- "logits/chosen": -1.144275188446045,
46
- "logits/rejected": -1.4153748750686646,
47
- "logps/chosen": -38.298439025878906,
48
- "logps/rejected": -41.69558334350586,
49
- "loss": 0.6927,
50
- "rewards/accuracies": 0.5625,
51
- "rewards/chosen": 0.0009220183710567653,
52
- "rewards/margins": 0.0017043445259332657,
53
- "rewards/rejected": -0.0007823262130841613,
54
  "step": 20
55
  },
56
  {
57
- "epoch": 0.11,
58
- "grad_norm": 39.765817458266895,
59
- "learning_rate": 9.99623976252115e-08,
60
- "logits/chosen": -0.8312944173812866,
61
- "logits/rejected": -1.4380576610565186,
62
- "logps/chosen": -32.726478576660156,
63
- "logps/rejected": -43.810638427734375,
64
- "loss": 0.6925,
65
- "rewards/accuracies": 0.59375,
66
- "rewards/chosen": 0.001944520859979093,
67
- "rewards/margins": 0.0023826442193239927,
68
- "rewards/rejected": -0.00043812341755256057,
69
  "step": 30
70
  },
71
  {
72
- "epoch": 0.15,
73
- "grad_norm": 35.39090968218046,
74
- "learning_rate": 9.929548316723981e-08,
75
- "logits/chosen": -1.0166804790496826,
76
- "logits/rejected": -1.4055747985839844,
77
- "logps/chosen": -31.433521270751953,
78
- "logps/rejected": -46.84731674194336,
79
- "loss": 0.6915,
80
- "rewards/accuracies": 0.5874999761581421,
81
- "rewards/chosen": 0.0015541177708655596,
82
- "rewards/margins": 0.002054566517472267,
83
- "rewards/rejected": -0.0005004488630220294,
84
  "step": 40
85
  },
86
  {
87
- "epoch": 0.19,
88
- "grad_norm": 42.118488242592456,
89
- "learning_rate": 9.780577808869398e-08,
90
- "logits/chosen": -1.0024656057357788,
91
- "logits/rejected": -1.3138850927352905,
92
- "logps/chosen": -30.340801239013672,
93
- "logps/rejected": -43.08925247192383,
94
- "loss": 0.6905,
95
- "rewards/accuracies": 0.606249988079071,
96
- "rewards/chosen": 0.0033007909078150988,
97
- "rewards/margins": 0.003635294735431671,
98
- "rewards/rejected": -0.000334503740305081,
99
  "step": 50
100
  },
101
  {
102
- "epoch": 0.22,
103
- "grad_norm": 36.92382364235435,
104
- "learning_rate": 9.551814704830733e-08,
105
- "logits/chosen": -1.1217851638793945,
106
- "logits/rejected": -1.4166934490203857,
107
- "logps/chosen": -31.992273330688477,
108
- "logps/rejected": -46.66759490966797,
109
- "loss": 0.6886,
110
- "rewards/accuracies": 0.6625000238418579,
111
- "rewards/chosen": 0.008065198548138142,
112
- "rewards/margins": 0.010263124480843544,
113
- "rewards/rejected": -0.0021979263983666897,
114
  "step": 60
115
  },
116
  {
117
- "epoch": 0.26,
118
- "grad_norm": 39.24669749459966,
119
- "learning_rate": 9.247077288236486e-08,
120
- "logits/chosen": -0.9909595251083374,
121
- "logits/rejected": -1.514561653137207,
122
- "logps/chosen": -31.16408348083496,
123
- "logps/rejected": -57.032684326171875,
124
- "loss": 0.6863,
125
- "rewards/accuracies": 0.71875,
126
- "rewards/chosen": 0.009227448143064976,
127
- "rewards/margins": 0.01421318482607603,
128
- "rewards/rejected": -0.004985739476978779,
129
  "step": 70
130
  },
131
  {
132
- "epoch": 0.3,
133
- "grad_norm": 38.72478614258279,
134
- "learning_rate": 8.871451929520661e-08,
135
- "logits/chosen": -0.7630345225334167,
136
- "logits/rejected": -1.5050407648086548,
137
- "logps/chosen": -25.78152084350586,
138
- "logps/rejected": -38.437538146972656,
139
- "loss": 0.6839,
140
- "rewards/accuracies": 0.800000011920929,
141
- "rewards/chosen": 0.013479883782565594,
142
- "rewards/margins": 0.018440086394548416,
143
- "rewards/rejected": -0.004960204008966684,
144
  "step": 80
145
  },
146
  {
147
- "epoch": 0.33,
148
- "grad_norm": 35.63155458414573,
149
- "learning_rate": 8.431208189343668e-08,
150
- "logits/chosen": -0.8956977725028992,
151
- "logits/rejected": -1.463505506515503,
152
- "logps/chosen": -30.033924102783203,
153
- "logps/rejected": -56.11266326904297,
154
- "loss": 0.6827,
155
- "rewards/accuracies": 0.7749999761581421,
156
- "rewards/chosen": 0.013467254117131233,
157
- "rewards/margins": 0.022690553218126297,
158
- "rewards/rejected": -0.00922329816967249,
159
  "step": 90
160
  },
161
  {
162
- "epoch": 0.37,
163
- "grad_norm": 37.21064690820017,
164
- "learning_rate": 7.93369417339209e-08,
165
- "logits/chosen": -0.7459134459495544,
166
- "logits/rejected": -1.5061514377593994,
167
- "logps/chosen": -28.494699478149414,
168
- "logps/rejected": -49.45507049560547,
169
- "loss": 0.6794,
170
- "rewards/accuracies": 0.800000011920929,
171
- "rewards/chosen": 0.015813838690519333,
172
- "rewards/margins": 0.02535676583647728,
173
- "rewards/rejected": -0.009542929008603096,
174
- "step": 100
175
- },
176
- {
177
- "epoch": 0.37,
178
- "eval_logits/chosen": -0.8265538215637207,
179
- "eval_logits/rejected": -1.4756767749786377,
180
- "eval_logps/chosen": -35.586036682128906,
181
- "eval_logps/rejected": -53.076499938964844,
182
- "eval_loss": 0.6905515193939209,
183
- "eval_rewards/accuracies": 0.5899999737739563,
184
- "eval_rewards/chosen": -0.004811165854334831,
185
- "eval_rewards/margins": 0.004816049709916115,
186
- "eval_rewards/rejected": -0.009627215564250946,
187
- "eval_runtime": 65.669,
188
- "eval_samples_per_second": 11.954,
189
- "eval_steps_per_second": 0.381,
190
- "step": 100
191
- },
192
- {
193
- "epoch": 0.41,
194
- "grad_norm": 41.489362342224126,
195
- "learning_rate": 7.387213885189744e-08,
196
- "logits/chosen": -0.9582328796386719,
197
- "logits/rejected": -1.4757133722305298,
198
- "logps/chosen": -29.652179718017578,
199
- "logps/rejected": -51.44434356689453,
200
- "loss": 0.6783,
201
- "rewards/accuracies": 0.824999988079071,
202
- "rewards/chosen": 0.01538840215653181,
203
- "rewards/margins": 0.030315887182950974,
204
- "rewards/rejected": -0.014927484095096588,
205
- "step": 110
206
- },
207
- {
208
- "epoch": 0.44,
209
- "grad_norm": 34.811996422375785,
210
- "learning_rate": 6.800888624023552e-08,
211
- "logits/chosen": -1.1017180681228638,
212
- "logits/rejected": -1.2910267114639282,
213
- "logps/chosen": -34.17676544189453,
214
- "logps/rejected": -47.27839660644531,
215
- "loss": 0.6752,
216
- "rewards/accuracies": 0.8062499761581421,
217
- "rewards/chosen": 0.019322626292705536,
218
- "rewards/margins": 0.03619503602385521,
219
- "rewards/rejected": -0.01687241531908512,
220
- "step": 120
221
- },
222
- {
223
- "epoch": 0.48,
224
- "grad_norm": 36.517002002282,
225
- "learning_rate": 6.184504741390595e-08,
226
- "logits/chosen": -1.0513687133789062,
227
- "logits/rejected": -1.4773876667022705,
228
- "logps/chosen": -30.05930519104004,
229
- "logps/rejected": -49.12663269042969,
230
- "loss": 0.6722,
231
- "rewards/accuracies": 0.856249988079071,
232
- "rewards/chosen": 0.017959367483854294,
233
- "rewards/margins": 0.038190923631191254,
234
- "rewards/rejected": -0.02023155428469181,
235
- "step": 130
236
- },
237
- {
238
- "epoch": 0.52,
239
- "grad_norm": 45.92150266428131,
240
- "learning_rate": 5.5483502970626584e-08,
241
- "logits/chosen": -1.0855646133422852,
242
- "logits/rejected": -1.4738510847091675,
243
- "logps/chosen": -35.4333610534668,
244
- "logps/rejected": -48.60808563232422,
245
- "loss": 0.6713,
246
- "rewards/accuracies": 0.893750011920929,
247
- "rewards/chosen": 0.01625698432326317,
248
- "rewards/margins": 0.04006917029619217,
249
- "rewards/rejected": -0.023812180384993553,
250
- "step": 140
251
- },
252
- {
253
- "epoch": 0.56,
254
- "grad_norm": 52.17233360297021,
255
- "learning_rate": 4.9030433411408785e-08,
256
- "logits/chosen": -1.0967937707901,
257
- "logits/rejected": -1.4282653331756592,
258
- "logps/chosen": -27.93039321899414,
259
- "logps/rejected": -48.556251525878906,
260
- "loss": 0.6666,
261
- "rewards/accuracies": 0.875,
262
- "rewards/chosen": 0.01784222200512886,
263
- "rewards/margins": 0.05219421908259392,
264
- "rewards/rejected": -0.03435199707746506,
265
- "step": 150
266
- },
267
- {
268
- "epoch": 0.59,
269
- "grad_norm": 36.1012077969649,
270
- "learning_rate": 4.259354688243757e-08,
271
- "logits/chosen": -1.088585615158081,
272
- "logits/rejected": -1.4759548902511597,
273
- "logps/chosen": -30.474609375,
274
- "logps/rejected": -50.4084358215332,
275
- "loss": 0.668,
276
- "rewards/accuracies": 0.824999988079071,
277
- "rewards/chosen": 0.007844483479857445,
278
- "rewards/margins": 0.05492537096142769,
279
- "rewards/rejected": -0.04708089306950569,
280
- "step": 160
281
- },
282
- {
283
- "epoch": 0.63,
284
- "grad_norm": 39.4057901333892,
285
- "learning_rate": 3.628028141903493e-08,
286
- "logits/chosen": -1.1803226470947266,
287
- "logits/rejected": -1.41080641746521,
288
- "logps/chosen": -32.83087158203125,
289
- "logps/rejected": -45.69880676269531,
290
- "loss": 0.6612,
291
- "rewards/accuracies": 0.8187500238418579,
292
- "rewards/chosen": 0.01473814807832241,
293
- "rewards/margins": 0.06375907361507416,
294
- "rewards/rejected": -0.0490209236741066,
295
- "step": 170
296
- },
297
- {
298
- "epoch": 0.67,
299
- "grad_norm": 38.966283555722626,
300
- "learning_rate": 3.0196011698042156e-08,
301
- "logits/chosen": -1.0210983753204346,
302
- "logits/rejected": -1.4852098226547241,
303
- "logps/chosen": -29.679798126220703,
304
- "logps/rejected": -50.24894332885742,
305
- "loss": 0.6648,
306
- "rewards/accuracies": 0.8687499761581421,
307
- "rewards/chosen": 0.006426820065826178,
308
- "rewards/margins": 0.06615011394023895,
309
- "rewards/rejected": -0.059723298996686935,
310
- "step": 180
311
- },
312
- {
313
- "epoch": 0.7,
314
- "grad_norm": 40.704072746648535,
315
- "learning_rate": 2.444229022970634e-08,
316
- "logits/chosen": -0.9898138046264648,
317
- "logits/rejected": -1.4493308067321777,
318
- "logps/chosen": -29.71084976196289,
319
- "logps/rejected": -52.1125602722168,
320
- "loss": 0.6587,
321
- "rewards/accuracies": 0.8187500238418579,
322
- "rewards/chosen": 0.005389463622123003,
323
- "rewards/margins": 0.06682445108890533,
324
- "rewards/rejected": -0.06143498420715332,
325
- "step": 190
326
- },
327
- {
328
- "epoch": 0.74,
329
- "grad_norm": 48.9298479981709,
330
- "learning_rate": 1.9115152345327152e-08,
331
- "logits/chosen": -1.0230295658111572,
332
- "logits/rejected": -1.5533816814422607,
333
- "logps/chosen": -30.142093658447266,
334
- "logps/rejected": -58.23160934448242,
335
- "loss": 0.6555,
336
- "rewards/accuracies": 0.862500011920929,
337
- "rewards/chosen": 0.004340417217463255,
338
- "rewards/margins": 0.08527159690856934,
339
- "rewards/rejected": -0.08093118667602539,
340
- "step": 200
341
- },
342
- {
343
- "epoch": 0.74,
344
- "eval_logits/chosen": -0.8588806390762329,
345
- "eval_logits/rejected": -1.5130012035369873,
346
- "eval_logps/chosen": -39.04316711425781,
347
- "eval_logps/rejected": -58.72098159790039,
348
- "eval_loss": 0.6837000846862793,
349
- "eval_rewards/accuracies": 0.6399999856948853,
350
- "eval_rewards/chosen": -0.0393824502825737,
351
- "eval_rewards/margins": 0.026689546182751656,
352
- "eval_rewards/rejected": -0.0660719946026802,
353
- "eval_runtime": 65.4972,
354
- "eval_samples_per_second": 11.985,
355
- "eval_steps_per_second": 0.382,
356
- "step": 200
357
- },
358
- {
359
- "epoch": 0.78,
360
- "grad_norm": 38.64475256798845,
361
- "learning_rate": 1.4303513272105055e-08,
362
- "logits/chosen": -0.9661673307418823,
363
- "logits/rejected": -1.5310702323913574,
364
- "logps/chosen": -31.798084259033203,
365
- "logps/rejected": -55.2061882019043,
366
- "loss": 0.6543,
367
- "rewards/accuracies": 0.856249988079071,
368
- "rewards/chosen": 0.008609571494162083,
369
- "rewards/margins": 0.0721936896443367,
370
- "rewards/rejected": -0.06358411908149719,
371
- "step": 210
372
- },
373
- {
374
- "epoch": 0.81,
375
- "grad_norm": 46.11579146010557,
376
- "learning_rate": 1.0087684049605349e-08,
377
- "logits/chosen": -1.0623886585235596,
378
- "logits/rejected": -1.484731912612915,
379
- "logps/chosen": -33.05571365356445,
380
- "logps/rejected": -68.4007568359375,
381
- "loss": 0.6512,
382
- "rewards/accuracies": 0.84375,
383
- "rewards/chosen": 0.002279428532347083,
384
- "rewards/margins": 0.09988276660442352,
385
- "rewards/rejected": -0.09760333597660065,
386
- "step": 220
387
- },
388
- {
389
- "epoch": 0.85,
390
- "grad_norm": 42.366279277913364,
391
- "learning_rate": 6.538031058667609e-09,
392
- "logits/chosen": -0.9526311159133911,
393
- "logits/rejected": -1.5703321695327759,
394
- "logps/chosen": -32.05732727050781,
395
- "logps/rejected": -69.91355895996094,
396
- "loss": 0.6539,
397
- "rewards/accuracies": 0.8687499761581421,
398
- "rewards/chosen": -0.0006113927811384201,
399
- "rewards/margins": 0.10779329389333725,
400
- "rewards/rejected": -0.10840468108654022,
401
- "step": 230
402
- },
403
- {
404
- "epoch": 0.89,
405
- "grad_norm": 42.96686513067178,
406
- "learning_rate": 3.713801536555483e-09,
407
- "logits/chosen": -1.1161487102508545,
408
- "logits/rejected": -1.5149171352386475,
409
- "logps/chosen": -31.100393295288086,
410
- "logps/rejected": -60.296607971191406,
411
- "loss": 0.6558,
412
- "rewards/accuracies": 0.800000011920929,
413
- "rewards/chosen": -0.0010345360497012734,
414
- "rewards/margins": 0.09006427228450775,
415
- "rewards/rejected": -0.09109880030155182,
416
- "step": 240
417
- },
418
- {
419
- "epoch": 0.93,
420
- "grad_norm": 46.95737595476474,
421
- "learning_rate": 1.662134681666899e-09,
422
- "logits/chosen": -0.9997628927230835,
423
- "logits/rejected": -1.6180452108383179,
424
- "logps/chosen": -27.377477645874023,
425
- "logps/rejected": -54.080360412597656,
426
- "loss": 0.6512,
427
- "rewards/accuracies": 0.8999999761581421,
428
- "rewards/chosen": 0.0026857522316277027,
429
- "rewards/margins": 0.09887233376502991,
430
- "rewards/rejected": -0.0961865708231926,
431
- "step": 250
432
- },
433
- {
434
- "epoch": 0.96,
435
- "grad_norm": 42.2077659611999,
436
- "learning_rate": 4.172748534499448e-10,
437
- "logits/chosen": -1.158440113067627,
438
- "logits/rejected": -1.4339531660079956,
439
- "logps/chosen": -32.833595275878906,
440
- "logps/rejected": -55.285552978515625,
441
- "loss": 0.6543,
442
- "rewards/accuracies": 0.856249988079071,
443
- "rewards/chosen": -0.0024300869554281235,
444
- "rewards/margins": 0.09633214771747589,
445
- "rewards/rejected": -0.09876222908496857,
446
- "step": 260
447
- },
448
- {
449
- "epoch": 1.0,
450
- "grad_norm": 44.84738059556724,
451
- "learning_rate": 0.0,
452
- "logits/chosen": -1.0697628259658813,
453
- "logits/rejected": -1.563280463218689,
454
- "logps/chosen": -33.97581481933594,
455
- "logps/rejected": -64.04183197021484,
456
- "loss": 0.6482,
457
- "rewards/accuracies": 0.8187500238418579,
458
- "rewards/chosen": -0.0003928671940229833,
459
- "rewards/margins": 0.0867936760187149,
460
- "rewards/rejected": -0.0871865525841713,
461
- "step": 270
462
- },
463
- {
464
- "epoch": 1.0,
465
- "step": 270,
466
  "total_flos": 0.0,
467
- "train_loss": 0.1692200431117305,
468
- "train_runtime": 798.9647,
469
- "train_samples_per_second": 21.608,
470
- "train_steps_per_second": 0.338
471
  }
472
  ],
473
  "logging_steps": 10,
474
- "max_steps": 270,
475
  "num_input_tokens_seen": 0,
476
  "num_train_epochs": 1,
477
- "save_steps": 100,
478
  "total_flos": 0.0,
479
  "train_batch_size": 8,
480
  "trial_name": null,
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.9947089947089947,
5
+ "eval_steps": 500,
6
+ "global_step": 94,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.01,
13
+ "grad_norm": 25.005165779900267,
14
+ "learning_rate": 1e-08,
15
+ "logits/chosen": -1.9501205682754517,
16
+ "logits/rejected": -2.513594388961792,
17
+ "logps/chosen": -348.5884704589844,
18
+ "logps/rejected": -166.58517456054688,
19
  "loss": 0.6931,
20
  "rewards/accuracies": 0.0,
21
  "rewards/chosen": 0.0,
 
24
  "step": 1
25
  },
26
  {
27
+ "epoch": 0.11,
28
+ "grad_norm": 23.68906488321,
29
+ "learning_rate": 1e-07,
30
+ "logits/chosen": -2.640984535217285,
31
+ "logits/rejected": -2.219906806945801,
32
+ "logps/chosen": -213.95584106445312,
33
+ "logps/rejected": -198.17874145507812,
34
+ "loss": 0.6932,
35
  "rewards/accuracies": 0.4930555522441864,
36
+ "rewards/chosen": 0.0007135343039408326,
37
+ "rewards/margins": 0.0006534373969770968,
38
+ "rewards/rejected": 6.009703065501526e-05,
39
  "step": 10
40
  },
41
  {
42
+ "epoch": 0.21,
43
+ "grad_norm": 22.52231020802604,
44
+ "learning_rate": 9.65436874322102e-08,
45
+ "logits/chosen": -2.397062301635742,
46
+ "logits/rejected": -2.3303606510162354,
47
+ "logps/chosen": -257.8389587402344,
48
+ "logps/rejected": -222.90444946289062,
49
+ "loss": 0.6917,
50
+ "rewards/accuracies": 0.6187499761581421,
51
+ "rewards/chosen": 0.0030278589110821486,
52
+ "rewards/margins": 0.002860091160982847,
53
+ "rewards/rejected": 0.00016776802658569068,
54
  "step": 20
55
  },
56
  {
57
+ "epoch": 0.32,
58
+ "grad_norm": 24.931881082321315,
59
+ "learning_rate": 8.665259359149131e-08,
60
+ "logits/chosen": -2.534593105316162,
61
+ "logits/rejected": -2.4346184730529785,
62
+ "logps/chosen": -227.6776885986328,
63
+ "logps/rejected": -204.8966064453125,
64
+ "loss": 0.689,
65
+ "rewards/accuracies": 0.75,
66
+ "rewards/chosen": 0.00954088568687439,
67
+ "rewards/margins": 0.00823338981717825,
68
+ "rewards/rejected": 0.0013074951712042093,
69
  "step": 30
70
  },
71
  {
72
+ "epoch": 0.42,
73
+ "grad_norm": 24.657287539381663,
74
+ "learning_rate": 7.16941869558779e-08,
75
+ "logits/chosen": -2.358189105987549,
76
+ "logits/rejected": -2.440410614013672,
77
+ "logps/chosen": -230.328857421875,
78
+ "logps/rejected": -210.7056427001953,
79
+ "loss": 0.6846,
80
+ "rewards/accuracies": 0.800000011920929,
81
+ "rewards/chosen": 0.018901044502854347,
82
+ "rewards/margins": 0.01690484955906868,
83
+ "rewards/rejected": 0.0019961954094469547,
84
  "step": 40
85
  },
86
  {
87
+ "epoch": 0.53,
88
+ "grad_norm": 24.026917311824274,
89
+ "learning_rate": 5.373650467932121e-08,
90
+ "logits/chosen": -2.379296064376831,
91
+ "logits/rejected": -2.6683709621429443,
92
+ "logps/chosen": -233.249267578125,
93
+ "logps/rejected": -199.204833984375,
94
+ "loss": 0.6805,
95
+ "rewards/accuracies": 0.8187500238418579,
96
+ "rewards/chosen": 0.02693324163556099,
97
+ "rewards/margins": 0.024523768573999405,
98
+ "rewards/rejected": 0.002409472828730941,
99
  "step": 50
100
  },
101
  {
102
+ "epoch": 0.63,
103
+ "grad_norm": 23.069042390909715,
104
+ "learning_rate": 3.5262241279454787e-08,
105
+ "logits/chosen": -2.233121156692505,
106
+ "logits/rejected": -2.644624710083008,
107
+ "logps/chosen": -250.5662078857422,
108
+ "logps/rejected": -173.79067993164062,
109
+ "loss": 0.6762,
110
+ "rewards/accuracies": 0.90625,
111
+ "rewards/chosen": 0.03840692713856697,
112
+ "rewards/margins": 0.03592415899038315,
113
+ "rewards/rejected": 0.0024827648885548115,
114
  "step": 60
115
  },
116
  {
117
+ "epoch": 0.74,
118
+ "grad_norm": 23.60279896856753,
119
+ "learning_rate": 1.8825509907063325e-08,
120
+ "logits/chosen": -2.354218006134033,
121
+ "logits/rejected": -2.420661449432373,
122
+ "logps/chosen": -241.6086883544922,
123
+ "logps/rejected": -207.097900390625,
124
+ "loss": 0.6725,
125
+ "rewards/accuracies": 0.824999988079071,
126
+ "rewards/chosen": 0.04547766596078873,
127
+ "rewards/margins": 0.04097073897719383,
128
+ "rewards/rejected": 0.004506924655288458,
129
  "step": 70
130
  },
131
  {
132
+ "epoch": 0.85,
133
+ "grad_norm": 23.640093404183123,
134
+ "learning_rate": 6.698729810778064e-09,
135
+ "logits/chosen": -2.2659950256347656,
136
+ "logits/rejected": -2.467071056365967,
137
+ "logps/chosen": -243.5118408203125,
138
+ "logps/rejected": -205.4396209716797,
139
+ "loss": 0.6739,
140
+ "rewards/accuracies": 0.8812500238418579,
141
+ "rewards/chosen": 0.0479864627122879,
142
+ "rewards/margins": 0.04288201406598091,
143
+ "rewards/rejected": 0.005104447714984417,
144
  "step": 80
145
  },
146
  {
147
+ "epoch": 0.95,
148
+ "grad_norm": 22.97821730211384,
149
+ "learning_rate": 5.584586887435739e-10,
150
+ "logits/chosen": -2.3436455726623535,
151
+ "logits/rejected": -2.367281436920166,
152
+ "logps/chosen": -226.5314178466797,
153
+ "logps/rejected": -200.30862426757812,
154
+ "loss": 0.6745,
155
+ "rewards/accuracies": 0.8374999761581421,
156
+ "rewards/chosen": 0.043580617755651474,
157
+ "rewards/margins": 0.037381939589977264,
158
+ "rewards/rejected": 0.006198678631335497,
159
  "step": 90
160
  },
161
  {
162
+ "epoch": 0.99,
163
+ "step": 94,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
164
  "total_flos": 0.0,
165
+ "train_loss": 0.6816894855905087,
166
+ "train_runtime": 1070.9433,
167
+ "train_samples_per_second": 5.639,
168
+ "train_steps_per_second": 0.088
169
  }
170
  ],
171
  "logging_steps": 10,
172
+ "max_steps": 94,
173
  "num_input_tokens_seen": 0,
174
  "num_train_epochs": 1,
175
+ "save_steps": 500,
176
  "total_flos": 0.0,
177
  "train_batch_size": 8,
178
  "trial_name": null,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0e423e6281d937957b182db4274e151dd830ea7f134ddb99760f67ce8bed87a0
3
  size 6200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5966de10a9274f8fe235eed3be0e3d0ae289568d178ec8f7314f12dcf1ab1f8e
3
  size 6200