Namronaldo2004 commited on
Commit
07c543d
·
1 Parent(s): cbbcf00

Update fine-tuned model

Browse files
adapter_config.json CHANGED
@@ -22,10 +22,10 @@
22
  "target_modules": [
23
  "q_proj",
24
  "k_proj",
25
- "gate_proj",
26
  "down_proj",
27
- "up_proj",
28
  "v_proj",
 
 
29
  "o_proj"
30
  ],
31
  "task_type": " CAUSAL_LM",
 
22
  "target_modules": [
23
  "q_proj",
24
  "k_proj",
 
25
  "down_proj",
 
26
  "v_proj",
27
+ "gate_proj",
28
+ "up_proj",
29
  "o_proj"
30
  ],
31
  "task_type": " CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3d054996c8870886154c89eb4b321cffae3782503afa3093805cae5b92e89d63
3
  size 159967880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f33495d2a06946619fbcf27dfad787f5d2ecbaff7594e36553591b83763f5dc
3
  size 159967880
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:75b0df8cc32c0c3fdd058abda370de0bd80f3dfa6026c5b8d7404204aa249974
3
  size 852876198
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad2b6acd8c794a174246823de83003bf746dd08a56ee2debca38260dd8ede8c0
3
  size 852876198
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:34e497cead5b5d8cb6e8ac2be926d8e9aecf8f729139a14383abc9f0da4763cc
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f14f3fae9bd6d8bdd744584b0dede52dcf273aa344fff0056431563d7dff0e18
3
  size 14244
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4521b8db9cc205e54aa606d85e707c024abd2d8ad4a20bec4b2cff365dc59cdf
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e69e2b49ea642509f0c688c16fb190b7cf27dac0a18903a5e2d1467d0343d8b8
3
  size 1064
trainer_state.json CHANGED
@@ -3,189 +3,364 @@
3
  "best_model_checkpoint": null,
4
  "epoch": 5.0,
5
  "eval_steps": 500,
6
- "global_step": 25,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.2,
13
- "grad_norm": 0.6263333559036255,
14
- "learning_rate": 0.0001,
15
- "loss": 0.6512,
16
  "step": 1
17
  },
18
  {
19
- "epoch": 0.4,
20
- "grad_norm": 0.659807562828064,
21
- "learning_rate": 0.0002,
22
- "loss": 0.6388,
23
  "step": 2
24
  },
25
  {
26
- "epoch": 0.6,
27
- "grad_norm": 0.4718220829963684,
28
- "learning_rate": 0.00019906859460363307,
29
- "loss": 0.6469,
30
  "step": 3
31
  },
32
  {
33
- "epoch": 0.8,
34
- "grad_norm": 0.3931182324886322,
35
- "learning_rate": 0.00019629172873477995,
36
- "loss": 0.6385,
37
  "step": 4
38
  },
39
  {
40
- "epoch": 1.0,
41
- "grad_norm": 0.394342839717865,
42
- "learning_rate": 0.00019172113015054532,
43
- "loss": 0.6754,
44
  "step": 5
45
  },
46
  {
47
- "epoch": 1.2,
48
- "grad_norm": 0.4089839458465576,
49
- "learning_rate": 0.00018544194045464886,
50
- "loss": 0.5292,
51
  "step": 6
52
  },
53
  {
54
- "epoch": 1.4,
55
- "grad_norm": 0.39185038208961487,
56
- "learning_rate": 0.000177571129070442,
57
- "loss": 0.5756,
58
  "step": 7
59
  },
60
  {
61
- "epoch": 1.6,
62
- "grad_norm": 0.35863593220710754,
63
- "learning_rate": 0.00016825531432186543,
64
- "loss": 0.5311,
65
  "step": 8
66
  },
67
  {
68
- "epoch": 1.8,
69
- "grad_norm": 0.3710576891899109,
70
- "learning_rate": 0.00015766803221148673,
71
- "loss": 0.5056,
72
  "step": 9
73
  },
74
  {
75
- "epoch": 2.0,
76
- "grad_norm": 0.365957647562027,
77
- "learning_rate": 0.00014600650377311522,
78
- "loss": 0.4892,
79
  "step": 10
80
  },
81
  {
82
- "epoch": 2.2,
83
- "grad_norm": 0.34360572695732117,
84
- "learning_rate": 0.00013348796121709862,
85
- "loss": 0.4838,
86
  "step": 11
87
  },
88
  {
89
- "epoch": 2.4,
90
- "grad_norm": 0.33910414576530457,
91
- "learning_rate": 0.0001203456013052634,
92
- "loss": 0.4295,
93
  "step": 12
94
  },
95
  {
96
- "epoch": 2.6,
97
- "grad_norm": 0.31048351526260376,
98
- "learning_rate": 0.0001068242413364671,
99
- "loss": 0.4072,
100
  "step": 13
101
  },
102
  {
103
- "epoch": 2.8,
104
- "grad_norm": 0.3141072988510132,
105
- "learning_rate": 9.317575866353292e-05,
106
- "loss": 0.4244,
107
  "step": 14
108
  },
109
  {
110
- "epoch": 3.0,
111
- "grad_norm": 0.31501471996307373,
112
- "learning_rate": 7.965439869473664e-05,
113
- "loss": 0.4128,
114
  "step": 15
115
  },
116
  {
117
- "epoch": 3.2,
118
- "grad_norm": 0.2909337878227234,
119
- "learning_rate": 6.651203878290139e-05,
120
- "loss": 0.3821,
121
  "step": 16
122
  },
123
  {
124
- "epoch": 3.4,
125
- "grad_norm": 0.26387819647789,
126
- "learning_rate": 5.399349622688479e-05,
127
- "loss": 0.3776,
128
  "step": 17
129
  },
130
  {
131
- "epoch": 3.6,
132
- "grad_norm": 0.26343247294425964,
133
- "learning_rate": 4.2331967788513295e-05,
134
- "loss": 0.3619,
135
  "step": 18
136
  },
137
  {
138
- "epoch": 3.8,
139
- "grad_norm": 0.26532021164894104,
140
- "learning_rate": 3.174468567813461e-05,
141
- "loss": 0.3715,
142
  "step": 19
143
  },
144
  {
145
- "epoch": 4.0,
146
- "grad_norm": 0.27087175846099854,
147
- "learning_rate": 2.242887092955801e-05,
148
- "loss": 0.3683,
149
  "step": 20
150
  },
151
  {
152
- "epoch": 4.2,
153
- "grad_norm": 0.25199756026268005,
154
- "learning_rate": 1.4558059545351143e-05,
155
- "loss": 0.3523,
156
  "step": 21
157
  },
158
  {
159
- "epoch": 4.4,
160
- "grad_norm": 0.2494262307882309,
161
- "learning_rate": 8.278869849454718e-06,
162
- "loss": 0.333,
163
  "step": 22
164
  },
165
  {
166
- "epoch": 4.6,
167
- "grad_norm": 0.25421398878097534,
168
- "learning_rate": 3.7082712652200867e-06,
169
- "loss": 0.3289,
170
  "step": 23
171
  },
172
  {
173
- "epoch": 4.8,
174
- "grad_norm": 0.26278430223464966,
175
- "learning_rate": 9.314053963669245e-07,
176
- "loss": 0.3438,
177
  "step": 24
178
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
179
  {
180
  "epoch": 5.0,
181
- "grad_norm": 0.2513781785964966,
182
  "learning_rate": 0.0,
183
- "loss": 0.379,
184
- "step": 25
185
  }
186
  ],
187
  "logging_steps": 1,
188
- "max_steps": 25,
189
  "num_input_tokens_seen": 0,
190
  "num_train_epochs": 5,
191
  "save_steps": 500,
@@ -201,7 +376,7 @@
201
  "attributes": {}
202
  }
203
  },
204
- "total_flos": 3.082693846592717e+16,
205
  "train_batch_size": 1,
206
  "trial_name": null,
207
  "trial_params": null
 
3
  "best_model_checkpoint": null,
4
  "epoch": 5.0,
5
  "eval_steps": 500,
6
+ "global_step": 50,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.1,
13
+ "grad_norm": 0.47972142696380615,
14
+ "learning_rate": 6.666666666666667e-05,
15
+ "loss": 0.6289,
16
  "step": 1
17
  },
18
  {
19
+ "epoch": 0.2,
20
+ "grad_norm": 0.466259241104126,
21
+ "learning_rate": 0.00013333333333333334,
22
+ "loss": 0.622,
23
  "step": 2
24
  },
25
  {
26
+ "epoch": 0.3,
27
+ "grad_norm": 0.4508052170276642,
28
+ "learning_rate": 0.0002,
29
+ "loss": 0.6472,
30
  "step": 3
31
  },
32
  {
33
+ "epoch": 0.4,
34
+ "grad_norm": 0.39328619837760925,
35
+ "learning_rate": 0.00019977668786231534,
36
+ "loss": 0.6188,
37
  "step": 4
38
  },
39
  {
40
+ "epoch": 0.5,
41
+ "grad_norm": 0.34969136118888855,
42
+ "learning_rate": 0.000199107748815478,
43
+ "loss": 0.5904,
44
  "step": 5
45
  },
46
  {
47
+ "epoch": 0.6,
48
+ "grad_norm": 0.34362491965293884,
49
+ "learning_rate": 0.0001979961705036587,
50
+ "loss": 0.6218,
51
  "step": 6
52
  },
53
  {
54
+ "epoch": 0.7,
55
+ "grad_norm": 0.3137108385562897,
56
+ "learning_rate": 0.00019644691750543767,
57
+ "loss": 0.5773,
58
  "step": 7
59
  },
60
  {
61
+ "epoch": 0.8,
62
+ "grad_norm": 0.3064671754837036,
63
+ "learning_rate": 0.0001944669091607919,
64
+ "loss": 0.6056,
65
  "step": 8
66
  },
67
  {
68
+ "epoch": 0.9,
69
+ "grad_norm": 0.28781741857528687,
70
+ "learning_rate": 0.00019206498866764288,
71
+ "loss": 0.6026,
72
  "step": 9
73
  },
74
  {
75
+ "epoch": 1.0,
76
+ "grad_norm": 0.29755786061286926,
77
+ "learning_rate": 0.00018925188358598813,
78
+ "loss": 0.6226,
79
  "step": 10
80
  },
81
  {
82
+ "epoch": 1.1,
83
+ "grad_norm": 0.32020193338394165,
84
+ "learning_rate": 0.00018604015792601396,
85
+ "loss": 0.4575,
86
  "step": 11
87
  },
88
  {
89
+ "epoch": 1.2,
90
+ "grad_norm": 0.37006428837776184,
91
+ "learning_rate": 0.00018244415603417603,
92
+ "loss": 0.4834,
93
  "step": 12
94
  },
95
  {
96
+ "epoch": 1.3,
97
+ "grad_norm": 0.3829534947872162,
98
+ "learning_rate": 0.0001784799385278661,
99
+ "loss": 0.5147,
100
  "step": 13
101
  },
102
  {
103
+ "epoch": 1.4,
104
+ "grad_norm": 0.316731721162796,
105
+ "learning_rate": 0.00017416521056479577,
106
+ "loss": 0.4958,
107
  "step": 14
108
  },
109
  {
110
+ "epoch": 1.5,
111
+ "grad_norm": 0.29609623551368713,
112
+ "learning_rate": 0.00016951924276746425,
113
+ "loss": 0.491,
114
  "step": 15
115
  },
116
  {
117
+ "epoch": 1.6,
118
+ "grad_norm": 0.29552099108695984,
119
+ "learning_rate": 0.00016456278515588024,
120
+ "loss": 0.4543,
121
  "step": 16
122
  },
123
  {
124
+ "epoch": 1.7,
125
+ "grad_norm": 0.2840826213359833,
126
+ "learning_rate": 0.00015931797447293552,
127
+ "loss": 0.4774,
128
  "step": 17
129
  },
130
  {
131
+ "epoch": 1.8,
132
+ "grad_norm": 0.2735576033592224,
133
+ "learning_rate": 0.00015380823531633729,
134
+ "loss": 0.4513,
135
  "step": 18
136
  },
137
  {
138
+ "epoch": 1.9,
139
+ "grad_norm": 0.2856349050998688,
140
+ "learning_rate": 0.00014805817551866838,
141
+ "loss": 0.4556,
142
  "step": 19
143
  },
144
  {
145
+ "epoch": 2.0,
146
+ "grad_norm": 0.2851189970970154,
147
+ "learning_rate": 0.0001420934762428335,
148
+ "loss": 0.4646,
149
  "step": 20
150
  },
151
  {
152
+ "epoch": 2.1,
153
+ "grad_norm": 0.2509067952632904,
154
+ "learning_rate": 0.00013594077728375128,
155
+ "loss": 0.4102,
156
  "step": 21
157
  },
158
  {
159
+ "epoch": 2.2,
160
+ "grad_norm": 0.2623592019081116,
161
+ "learning_rate": 0.00012962755808856342,
162
+ "loss": 0.378,
163
  "step": 22
164
  },
165
  {
166
+ "epoch": 2.3,
167
+ "grad_norm": 0.267648845911026,
168
+ "learning_rate": 0.00012318201502675285,
169
+ "loss": 0.4115,
170
  "step": 23
171
  },
172
  {
173
+ "epoch": 2.4,
174
+ "grad_norm": 0.27410486340522766,
175
+ "learning_rate": 0.00011663293545831302,
176
+ "loss": 0.389,
177
  "step": 24
178
  },
179
+ {
180
+ "epoch": 2.5,
181
+ "grad_norm": 0.2685711979866028,
182
+ "learning_rate": 0.00011000956916240985,
183
+ "loss": 0.3922,
184
+ "step": 25
185
+ },
186
+ {
187
+ "epoch": 2.6,
188
+ "grad_norm": 0.2666301727294922,
189
+ "learning_rate": 0.00010334149770076747,
190
+ "loss": 0.3749,
191
+ "step": 26
192
+ },
193
+ {
194
+ "epoch": 2.7,
195
+ "grad_norm": 0.3079356551170349,
196
+ "learning_rate": 9.665850229923258e-05,
197
+ "loss": 0.3486,
198
+ "step": 27
199
+ },
200
+ {
201
+ "epoch": 2.8,
202
+ "grad_norm": 0.28150758147239685,
203
+ "learning_rate": 8.999043083759017e-05,
204
+ "loss": 0.3606,
205
+ "step": 28
206
+ },
207
+ {
208
+ "epoch": 2.9,
209
+ "grad_norm": 0.28707581758499146,
210
+ "learning_rate": 8.336706454168701e-05,
211
+ "loss": 0.3642,
212
+ "step": 29
213
+ },
214
+ {
215
+ "epoch": 3.0,
216
+ "grad_norm": 0.2959519028663635,
217
+ "learning_rate": 7.681798497324716e-05,
218
+ "loss": 0.3624,
219
+ "step": 30
220
+ },
221
+ {
222
+ "epoch": 3.1,
223
+ "grad_norm": 0.26570671796798706,
224
+ "learning_rate": 7.037244191143661e-05,
225
+ "loss": 0.3231,
226
+ "step": 31
227
+ },
228
+ {
229
+ "epoch": 3.2,
230
+ "grad_norm": 0.2674141526222229,
231
+ "learning_rate": 6.405922271624874e-05,
232
+ "loss": 0.3298,
233
+ "step": 32
234
+ },
235
+ {
236
+ "epoch": 3.3,
237
+ "grad_norm": 0.2791026532649994,
238
+ "learning_rate": 5.790652375716652e-05,
239
+ "loss": 0.3276,
240
+ "step": 33
241
+ },
242
+ {
243
+ "epoch": 3.4,
244
+ "grad_norm": 0.26298171281814575,
245
+ "learning_rate": 5.1941824481331626e-05,
246
+ "loss": 0.3048,
247
+ "step": 34
248
+ },
249
+ {
250
+ "epoch": 3.5,
251
+ "grad_norm": 0.269193559885025,
252
+ "learning_rate": 4.6191764683662744e-05,
253
+ "loss": 0.3216,
254
+ "step": 35
255
+ },
256
+ {
257
+ "epoch": 3.6,
258
+ "grad_norm": 0.26902416348457336,
259
+ "learning_rate": 4.0682025527064486e-05,
260
+ "loss": 0.2933,
261
+ "step": 36
262
+ },
263
+ {
264
+ "epoch": 3.7,
265
+ "grad_norm": 0.26919540762901306,
266
+ "learning_rate": 3.543721484411976e-05,
267
+ "loss": 0.3138,
268
+ "step": 37
269
+ },
270
+ {
271
+ "epoch": 3.8,
272
+ "grad_norm": 0.27092471718788147,
273
+ "learning_rate": 3.0480757232535772e-05,
274
+ "loss": 0.3184,
275
+ "step": 38
276
+ },
277
+ {
278
+ "epoch": 3.9,
279
+ "grad_norm": 0.26903486251831055,
280
+ "learning_rate": 2.5834789435204243e-05,
281
+ "loss": 0.3148,
282
+ "step": 39
283
+ },
284
+ {
285
+ "epoch": 4.0,
286
+ "grad_norm": 0.28141602873802185,
287
+ "learning_rate": 2.1520061472133902e-05,
288
+ "loss": 0.3125,
289
+ "step": 40
290
+ },
291
+ {
292
+ "epoch": 4.1,
293
+ "grad_norm": 0.25863194465637207,
294
+ "learning_rate": 1.7555843965823992e-05,
295
+ "loss": 0.2954,
296
+ "step": 41
297
+ },
298
+ {
299
+ "epoch": 4.2,
300
+ "grad_norm": 0.2740136682987213,
301
+ "learning_rate": 1.3959842073986085e-05,
302
+ "loss": 0.2992,
303
+ "step": 42
304
+ },
305
+ {
306
+ "epoch": 4.3,
307
+ "grad_norm": 0.2652461528778076,
308
+ "learning_rate": 1.0748116414011888e-05,
309
+ "loss": 0.2925,
310
+ "step": 43
311
+ },
312
+ {
313
+ "epoch": 4.4,
314
+ "grad_norm": 0.278138667345047,
315
+ "learning_rate": 7.935011332357112e-06,
316
+ "loss": 0.2816,
317
+ "step": 44
318
+ },
319
+ {
320
+ "epoch": 4.5,
321
+ "grad_norm": 0.2779091000556946,
322
+ "learning_rate": 5.533090839208133e-06,
323
+ "loss": 0.2726,
324
+ "step": 45
325
+ },
326
+ {
327
+ "epoch": 4.6,
328
+ "grad_norm": 0.2674082815647125,
329
+ "learning_rate": 3.5530824945623542e-06,
330
+ "loss": 0.2836,
331
+ "step": 46
332
+ },
333
+ {
334
+ "epoch": 4.7,
335
+ "grad_norm": 0.2792351245880127,
336
+ "learning_rate": 2.003829496341325e-06,
337
+ "loss": 0.2849,
338
+ "step": 47
339
+ },
340
+ {
341
+ "epoch": 4.8,
342
+ "grad_norm": 0.2667449414730072,
343
+ "learning_rate": 8.922511845219971e-07,
344
+ "loss": 0.2929,
345
+ "step": 48
346
+ },
347
+ {
348
+ "epoch": 4.9,
349
+ "grad_norm": 0.27168118953704834,
350
+ "learning_rate": 2.2331213768468363e-07,
351
+ "loss": 0.2739,
352
+ "step": 49
353
+ },
354
  {
355
  "epoch": 5.0,
356
+ "grad_norm": 0.27336716651916504,
357
  "learning_rate": 0.0,
358
+ "loss": 0.2936,
359
+ "step": 50
360
  }
361
  ],
362
  "logging_steps": 1,
363
+ "max_steps": 50,
364
  "num_input_tokens_seen": 0,
365
  "num_train_epochs": 5,
366
  "save_steps": 500,
 
376
  "attributes": {}
377
  }
378
  },
379
+ "total_flos": 6.031379393052672e+16,
380
  "train_batch_size": 1,
381
  "trial_name": null,
382
  "trial_params": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bdb67f7e08d706c1aba378a44137116de683468a035998cac2d26dc2c2a5fd26
3
  size 5240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f9069aab4af513f470da5848e8771d43df41e0e1100cb7715f6f428cceb308b
3
  size 5240