|
{ |
|
"best_metric": 0.7337000966072083, |
|
"best_model_checkpoint": "/home/ubuntu/utah/babylm-24/src/evaluation/results/finetune/DebertaV2-Base-10M_babylm-A/mnli/checkpoint-147264", |
|
"epoch": 5.0, |
|
"eval_steps": 500, |
|
"global_step": 245440, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.010185788787483703, |
|
"grad_norm": 3.972818613052368, |
|
"learning_rate": 2.9938885267275102e-05, |
|
"loss": 1.1058, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.020371577574967405, |
|
"grad_norm": 3.13022780418396, |
|
"learning_rate": 2.9877770534550196e-05, |
|
"loss": 1.0388, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.03055736636245111, |
|
"grad_norm": 5.194879055023193, |
|
"learning_rate": 2.9816655801825297e-05, |
|
"loss": 0.999, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.04074315514993481, |
|
"grad_norm": 6.858974933624268, |
|
"learning_rate": 2.975554106910039e-05, |
|
"loss": 0.9637, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.050928943937418515, |
|
"grad_norm": 7.074331760406494, |
|
"learning_rate": 2.9694426336375492e-05, |
|
"loss": 0.9697, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.06111473272490222, |
|
"grad_norm": 5.386518478393555, |
|
"learning_rate": 2.9633311603650586e-05, |
|
"loss": 0.9472, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.07130052151238592, |
|
"grad_norm": 5.582085609436035, |
|
"learning_rate": 2.9572196870925687e-05, |
|
"loss": 0.9273, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.08148631029986962, |
|
"grad_norm": 10.966410636901855, |
|
"learning_rate": 2.951108213820078e-05, |
|
"loss": 0.9316, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.09167209908735333, |
|
"grad_norm": 7.840672492980957, |
|
"learning_rate": 2.9449967405475882e-05, |
|
"loss": 0.9237, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.10185788787483703, |
|
"grad_norm": 7.442521095275879, |
|
"learning_rate": 2.9388852672750976e-05, |
|
"loss": 0.918, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.11204367666232073, |
|
"grad_norm": 8.379261016845703, |
|
"learning_rate": 2.9327737940026077e-05, |
|
"loss": 0.8798, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.12222946544980444, |
|
"grad_norm": 6.880769729614258, |
|
"learning_rate": 2.926662320730117e-05, |
|
"loss": 0.9045, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.13241525423728814, |
|
"grad_norm": 5.965817451477051, |
|
"learning_rate": 2.9205508474576272e-05, |
|
"loss": 0.8941, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.14260104302477183, |
|
"grad_norm": 9.67165756225586, |
|
"learning_rate": 2.914439374185137e-05, |
|
"loss": 0.878, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.15278683181225555, |
|
"grad_norm": 6.4934892654418945, |
|
"learning_rate": 2.9083279009126467e-05, |
|
"loss": 0.8723, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.16297262059973924, |
|
"grad_norm": 11.987072944641113, |
|
"learning_rate": 2.9022164276401565e-05, |
|
"loss": 0.8913, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.17315840938722293, |
|
"grad_norm": 9.218280792236328, |
|
"learning_rate": 2.8961049543676662e-05, |
|
"loss": 0.868, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.18334419817470665, |
|
"grad_norm": 6.697582244873047, |
|
"learning_rate": 2.8899934810951763e-05, |
|
"loss": 0.8618, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.19352998696219034, |
|
"grad_norm": 8.035455703735352, |
|
"learning_rate": 2.8838820078226858e-05, |
|
"loss": 0.8641, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.20371577574967406, |
|
"grad_norm": 6.890414237976074, |
|
"learning_rate": 2.877770534550196e-05, |
|
"loss": 0.8574, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.21390156453715775, |
|
"grad_norm": 9.001678466796875, |
|
"learning_rate": 2.8716590612777053e-05, |
|
"loss": 0.8547, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.22408735332464147, |
|
"grad_norm": 7.271191120147705, |
|
"learning_rate": 2.8655475880052154e-05, |
|
"loss": 0.8332, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.23427314211212516, |
|
"grad_norm": 9.849822998046875, |
|
"learning_rate": 2.8594361147327248e-05, |
|
"loss": 0.8447, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.24445893089960888, |
|
"grad_norm": 15.07829761505127, |
|
"learning_rate": 2.853324641460235e-05, |
|
"loss": 0.8251, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.25464471968709257, |
|
"grad_norm": 6.886602878570557, |
|
"learning_rate": 2.8472131681877446e-05, |
|
"loss": 0.8444, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.2648305084745763, |
|
"grad_norm": 9.94825267791748, |
|
"learning_rate": 2.8411016949152544e-05, |
|
"loss": 0.8504, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.27501629726205995, |
|
"grad_norm": 8.0740966796875, |
|
"learning_rate": 2.834990221642764e-05, |
|
"loss": 0.8182, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.28520208604954367, |
|
"grad_norm": 10.770129203796387, |
|
"learning_rate": 2.828878748370274e-05, |
|
"loss": 0.8373, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.2953878748370274, |
|
"grad_norm": 9.214078903198242, |
|
"learning_rate": 2.8227672750977836e-05, |
|
"loss": 0.8089, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.3055736636245111, |
|
"grad_norm": 14.231884956359863, |
|
"learning_rate": 2.8166558018252934e-05, |
|
"loss": 0.8337, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.31575945241199477, |
|
"grad_norm": 9.379185676574707, |
|
"learning_rate": 2.810544328552803e-05, |
|
"loss": 0.8315, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.3259452411994785, |
|
"grad_norm": 8.370843887329102, |
|
"learning_rate": 2.804432855280313e-05, |
|
"loss": 0.808, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.3361310299869622, |
|
"grad_norm": 7.201444149017334, |
|
"learning_rate": 2.7983213820078227e-05, |
|
"loss": 0.808, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 0.34631681877444587, |
|
"grad_norm": 10.696725845336914, |
|
"learning_rate": 2.7922099087353324e-05, |
|
"loss": 0.813, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.3565026075619296, |
|
"grad_norm": 6.316204071044922, |
|
"learning_rate": 2.7860984354628422e-05, |
|
"loss": 0.8213, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 0.3666883963494133, |
|
"grad_norm": 14.780086517333984, |
|
"learning_rate": 2.779986962190352e-05, |
|
"loss": 0.8165, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.376874185136897, |
|
"grad_norm": 11.34945011138916, |
|
"learning_rate": 2.773875488917862e-05, |
|
"loss": 0.7943, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 0.3870599739243807, |
|
"grad_norm": 7.136452674865723, |
|
"learning_rate": 2.7677640156453718e-05, |
|
"loss": 0.8017, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.3972457627118644, |
|
"grad_norm": 6.209742069244385, |
|
"learning_rate": 2.7616525423728815e-05, |
|
"loss": 0.7941, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 0.4074315514993481, |
|
"grad_norm": 8.923270225524902, |
|
"learning_rate": 2.7555410691003913e-05, |
|
"loss": 0.7826, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.41761734028683184, |
|
"grad_norm": 6.371203899383545, |
|
"learning_rate": 2.749429595827901e-05, |
|
"loss": 0.7938, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 0.4278031290743155, |
|
"grad_norm": 8.679354667663574, |
|
"learning_rate": 2.7433181225554108e-05, |
|
"loss": 0.7936, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.4379889178617992, |
|
"grad_norm": 5.962934970855713, |
|
"learning_rate": 2.7372066492829205e-05, |
|
"loss": 0.7924, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 0.44817470664928294, |
|
"grad_norm": 11.835200309753418, |
|
"learning_rate": 2.7310951760104303e-05, |
|
"loss": 0.7894, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 0.4583604954367666, |
|
"grad_norm": 9.62806510925293, |
|
"learning_rate": 2.72498370273794e-05, |
|
"loss": 0.7963, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 0.4685462842242503, |
|
"grad_norm": 7.6390509605407715, |
|
"learning_rate": 2.7188722294654498e-05, |
|
"loss": 0.783, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 0.47873207301173404, |
|
"grad_norm": 5.826275825500488, |
|
"learning_rate": 2.7127607561929596e-05, |
|
"loss": 0.7859, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 0.48891786179921776, |
|
"grad_norm": 5.199416637420654, |
|
"learning_rate": 2.7066492829204693e-05, |
|
"loss": 0.7989, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 0.4991036505867014, |
|
"grad_norm": 12.38355541229248, |
|
"learning_rate": 2.7005378096479794e-05, |
|
"loss": 0.7864, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 0.5092894393741851, |
|
"grad_norm": 7.700937271118164, |
|
"learning_rate": 2.694426336375489e-05, |
|
"loss": 0.7747, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 0.5194752281616688, |
|
"grad_norm": 10.8720703125, |
|
"learning_rate": 2.688314863102999e-05, |
|
"loss": 0.7974, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 0.5296610169491526, |
|
"grad_norm": 6.260867595672607, |
|
"learning_rate": 2.6822033898305083e-05, |
|
"loss": 0.7914, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 0.5398468057366362, |
|
"grad_norm": 11.113471984863281, |
|
"learning_rate": 2.6760919165580184e-05, |
|
"loss": 0.782, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 0.5500325945241199, |
|
"grad_norm": 7.497796058654785, |
|
"learning_rate": 2.6699804432855282e-05, |
|
"loss": 0.7642, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 0.5602183833116037, |
|
"grad_norm": 6.1969170570373535, |
|
"learning_rate": 2.663868970013038e-05, |
|
"loss": 0.7693, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 0.5704041720990873, |
|
"grad_norm": 9.965928077697754, |
|
"learning_rate": 2.6577574967405477e-05, |
|
"loss": 0.7681, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 0.5805899608865711, |
|
"grad_norm": 5.052292346954346, |
|
"learning_rate": 2.6516460234680575e-05, |
|
"loss": 0.7664, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 0.5907757496740548, |
|
"grad_norm": 6.7289276123046875, |
|
"learning_rate": 2.6455345501955672e-05, |
|
"loss": 0.7445, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 0.6009615384615384, |
|
"grad_norm": 8.478078842163086, |
|
"learning_rate": 2.639423076923077e-05, |
|
"loss": 0.7859, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 0.6111473272490222, |
|
"grad_norm": 6.2336106300354, |
|
"learning_rate": 2.6333116036505867e-05, |
|
"loss": 0.7598, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 0.6213331160365059, |
|
"grad_norm": 11.82295036315918, |
|
"learning_rate": 2.6272001303780965e-05, |
|
"loss": 0.7507, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 0.6315189048239895, |
|
"grad_norm": 8.545003890991211, |
|
"learning_rate": 2.6210886571056066e-05, |
|
"loss": 0.7769, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 0.6417046936114733, |
|
"grad_norm": 9.849065780639648, |
|
"learning_rate": 2.614977183833116e-05, |
|
"loss": 0.7781, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 0.651890482398957, |
|
"grad_norm": 7.641623020172119, |
|
"learning_rate": 2.608865710560626e-05, |
|
"loss": 0.7789, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 0.6620762711864406, |
|
"grad_norm": 16.397785186767578, |
|
"learning_rate": 2.6027542372881355e-05, |
|
"loss": 0.7542, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 0.6722620599739244, |
|
"grad_norm": 11.119671821594238, |
|
"learning_rate": 2.5966427640156456e-05, |
|
"loss": 0.7716, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 0.6824478487614081, |
|
"grad_norm": 12.466675758361816, |
|
"learning_rate": 2.590531290743155e-05, |
|
"loss": 0.7743, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 0.6926336375488917, |
|
"grad_norm": 8.151625633239746, |
|
"learning_rate": 2.584419817470665e-05, |
|
"loss": 0.7581, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 0.7028194263363755, |
|
"grad_norm": 7.743143558502197, |
|
"learning_rate": 2.5783083441981745e-05, |
|
"loss": 0.7599, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 0.7130052151238592, |
|
"grad_norm": 5.249680995941162, |
|
"learning_rate": 2.5721968709256846e-05, |
|
"loss": 0.7483, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 0.7231910039113429, |
|
"grad_norm": 6.738178730010986, |
|
"learning_rate": 2.566085397653194e-05, |
|
"loss": 0.7781, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 0.7333767926988266, |
|
"grad_norm": 6.7923102378845215, |
|
"learning_rate": 2.559973924380704e-05, |
|
"loss": 0.7464, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 0.7435625814863103, |
|
"grad_norm": 15.575695991516113, |
|
"learning_rate": 2.553862451108214e-05, |
|
"loss": 0.7571, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 0.753748370273794, |
|
"grad_norm": 13.128719329833984, |
|
"learning_rate": 2.5477509778357236e-05, |
|
"loss": 0.7625, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 0.7639341590612777, |
|
"grad_norm": 4.938621997833252, |
|
"learning_rate": 2.5416395045632337e-05, |
|
"loss": 0.7477, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 0.7741199478487614, |
|
"grad_norm": 8.46985912322998, |
|
"learning_rate": 2.535528031290743e-05, |
|
"loss": 0.7419, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 0.7843057366362451, |
|
"grad_norm": 18.881208419799805, |
|
"learning_rate": 2.5294165580182532e-05, |
|
"loss": 0.7403, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 0.7944915254237288, |
|
"grad_norm": 9.986156463623047, |
|
"learning_rate": 2.5233050847457626e-05, |
|
"loss": 0.7508, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 0.8046773142112125, |
|
"grad_norm": 11.625401496887207, |
|
"learning_rate": 2.5171936114732727e-05, |
|
"loss": 0.7426, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 0.8148631029986962, |
|
"grad_norm": 9.926522254943848, |
|
"learning_rate": 2.511082138200782e-05, |
|
"loss": 0.7365, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 0.8250488917861799, |
|
"grad_norm": 8.104277610778809, |
|
"learning_rate": 2.5049706649282922e-05, |
|
"loss": 0.7532, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 0.8352346805736637, |
|
"grad_norm": 5.898177623748779, |
|
"learning_rate": 2.4988591916558017e-05, |
|
"loss": 0.7657, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 0.8454204693611473, |
|
"grad_norm": 6.856070041656494, |
|
"learning_rate": 2.4927477183833118e-05, |
|
"loss": 0.7438, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 0.855606258148631, |
|
"grad_norm": 11.806628227233887, |
|
"learning_rate": 2.4866362451108212e-05, |
|
"loss": 0.7348, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 0.8657920469361148, |
|
"grad_norm": 12.08478832244873, |
|
"learning_rate": 2.4805247718383313e-05, |
|
"loss": 0.7277, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 0.8759778357235984, |
|
"grad_norm": 5.707648277282715, |
|
"learning_rate": 2.474413298565841e-05, |
|
"loss": 0.7458, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 0.8861636245110821, |
|
"grad_norm": 5.36006498336792, |
|
"learning_rate": 2.4683018252933508e-05, |
|
"loss": 0.7446, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 0.8963494132985659, |
|
"grad_norm": 6.603504180908203, |
|
"learning_rate": 2.4621903520208605e-05, |
|
"loss": 0.7278, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 0.9065352020860495, |
|
"grad_norm": 7.789953231811523, |
|
"learning_rate": 2.4560788787483703e-05, |
|
"loss": 0.7375, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 0.9167209908735332, |
|
"grad_norm": 11.424116134643555, |
|
"learning_rate": 2.4499674054758804e-05, |
|
"loss": 0.7297, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 0.926906779661017, |
|
"grad_norm": 6.7079620361328125, |
|
"learning_rate": 2.4438559322033898e-05, |
|
"loss": 0.7411, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 0.9370925684485006, |
|
"grad_norm": 11.931846618652344, |
|
"learning_rate": 2.4377444589309e-05, |
|
"loss": 0.749, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 0.9472783572359843, |
|
"grad_norm": 8.87076473236084, |
|
"learning_rate": 2.4316329856584093e-05, |
|
"loss": 0.7312, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 0.9574641460234681, |
|
"grad_norm": 12.787999153137207, |
|
"learning_rate": 2.4255215123859194e-05, |
|
"loss": 0.7244, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 0.9676499348109517, |
|
"grad_norm": 6.917387962341309, |
|
"learning_rate": 2.4194100391134288e-05, |
|
"loss": 0.733, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 0.9778357235984355, |
|
"grad_norm": 7.564793109893799, |
|
"learning_rate": 2.413298565840939e-05, |
|
"loss": 0.7253, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 0.9880215123859192, |
|
"grad_norm": 7.665174961090088, |
|
"learning_rate": 2.4071870925684483e-05, |
|
"loss": 0.7191, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 0.9982073011734028, |
|
"grad_norm": 9.235818862915039, |
|
"learning_rate": 2.4010756192959584e-05, |
|
"loss": 0.7288, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.6841890811920166, |
|
"eval_loss": 0.7356013059616089, |
|
"eval_runtime": 8.0037, |
|
"eval_samples_per_second": 613.213, |
|
"eval_steps_per_second": 76.714, |
|
"step": 49088 |
|
}, |
|
{ |
|
"epoch": 1.0083930899608866, |
|
"grad_norm": 9.232441902160645, |
|
"learning_rate": 2.3949641460234682e-05, |
|
"loss": 0.666, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 1.0185788787483703, |
|
"grad_norm": 5.098124980926514, |
|
"learning_rate": 2.388852672750978e-05, |
|
"loss": 0.6762, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 1.028764667535854, |
|
"grad_norm": 6.507536888122559, |
|
"learning_rate": 2.3827411994784877e-05, |
|
"loss": 0.6509, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 1.0389504563233376, |
|
"grad_norm": 12.516643524169922, |
|
"learning_rate": 2.3766297262059974e-05, |
|
"loss": 0.6569, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 1.0491362451108215, |
|
"grad_norm": 11.644853591918945, |
|
"learning_rate": 2.3705182529335072e-05, |
|
"loss": 0.6498, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 1.0593220338983051, |
|
"grad_norm": 15.303723335266113, |
|
"learning_rate": 2.364406779661017e-05, |
|
"loss": 0.6565, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 1.0695078226857888, |
|
"grad_norm": 11.212811470031738, |
|
"learning_rate": 2.3582953063885267e-05, |
|
"loss": 0.6492, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 1.0796936114732725, |
|
"grad_norm": 7.695071220397949, |
|
"learning_rate": 2.3521838331160365e-05, |
|
"loss": 0.6618, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 1.0898794002607561, |
|
"grad_norm": 8.234328269958496, |
|
"learning_rate": 2.3460723598435462e-05, |
|
"loss": 0.6709, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 1.1000651890482398, |
|
"grad_norm": 12.574545860290527, |
|
"learning_rate": 2.339960886571056e-05, |
|
"loss": 0.6518, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 1.1102509778357237, |
|
"grad_norm": 5.914109706878662, |
|
"learning_rate": 2.333849413298566e-05, |
|
"loss": 0.6494, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 1.1204367666232073, |
|
"grad_norm": 11.49263858795166, |
|
"learning_rate": 2.3277379400260758e-05, |
|
"loss": 0.6723, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 1.130622555410691, |
|
"grad_norm": 11.228019714355469, |
|
"learning_rate": 2.3216264667535856e-05, |
|
"loss": 0.6328, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 1.1408083441981747, |
|
"grad_norm": 14.478355407714844, |
|
"learning_rate": 2.3155149934810953e-05, |
|
"loss": 0.667, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 1.1509941329856583, |
|
"grad_norm": 18.817468643188477, |
|
"learning_rate": 2.309403520208605e-05, |
|
"loss": 0.6674, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 1.161179921773142, |
|
"grad_norm": 12.483678817749023, |
|
"learning_rate": 2.303292046936115e-05, |
|
"loss": 0.6677, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 1.1713657105606259, |
|
"grad_norm": 7.133495807647705, |
|
"learning_rate": 2.2971805736636246e-05, |
|
"loss": 0.642, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 1.1815514993481095, |
|
"grad_norm": 7.099282741546631, |
|
"learning_rate": 2.2910691003911343e-05, |
|
"loss": 0.664, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 1.1917372881355932, |
|
"grad_norm": 9.188867568969727, |
|
"learning_rate": 2.284957627118644e-05, |
|
"loss": 0.6703, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 1.2019230769230769, |
|
"grad_norm": 5.470687389373779, |
|
"learning_rate": 2.278846153846154e-05, |
|
"loss": 0.6575, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 1.2121088657105605, |
|
"grad_norm": 8.013532638549805, |
|
"learning_rate": 2.2727346805736636e-05, |
|
"loss": 0.6465, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 1.2222946544980444, |
|
"grad_norm": 8.846539497375488, |
|
"learning_rate": 2.2666232073011734e-05, |
|
"loss": 0.6682, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 1.232480443285528, |
|
"grad_norm": 11.2225341796875, |
|
"learning_rate": 2.260511734028683e-05, |
|
"loss": 0.653, |
|
"step": 60500 |
|
}, |
|
{ |
|
"epoch": 1.2426662320730117, |
|
"grad_norm": 12.363032341003418, |
|
"learning_rate": 2.254400260756193e-05, |
|
"loss": 0.6807, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 1.2528520208604954, |
|
"grad_norm": 8.01014232635498, |
|
"learning_rate": 2.248288787483703e-05, |
|
"loss": 0.6625, |
|
"step": 61500 |
|
}, |
|
{ |
|
"epoch": 1.263037809647979, |
|
"grad_norm": 20.960845947265625, |
|
"learning_rate": 2.2421773142112124e-05, |
|
"loss": 0.6524, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 1.2732235984354627, |
|
"grad_norm": 13.620718002319336, |
|
"learning_rate": 2.2360658409387225e-05, |
|
"loss": 0.6625, |
|
"step": 62500 |
|
}, |
|
{ |
|
"epoch": 1.2834093872229466, |
|
"grad_norm": 18.55965232849121, |
|
"learning_rate": 2.2299543676662322e-05, |
|
"loss": 0.6434, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 1.2935951760104303, |
|
"grad_norm": 12.919173240661621, |
|
"learning_rate": 2.223842894393742e-05, |
|
"loss": 0.652, |
|
"step": 63500 |
|
}, |
|
{ |
|
"epoch": 1.303780964797914, |
|
"grad_norm": 19.561386108398438, |
|
"learning_rate": 2.2177314211212517e-05, |
|
"loss": 0.6498, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 1.3139667535853976, |
|
"grad_norm": 10.529995918273926, |
|
"learning_rate": 2.2116199478487615e-05, |
|
"loss": 0.666, |
|
"step": 64500 |
|
}, |
|
{ |
|
"epoch": 1.3241525423728815, |
|
"grad_norm": 13.268050193786621, |
|
"learning_rate": 2.2055084745762713e-05, |
|
"loss": 0.6499, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 1.3343383311603652, |
|
"grad_norm": 6.3072285652160645, |
|
"learning_rate": 2.199397001303781e-05, |
|
"loss": 0.6798, |
|
"step": 65500 |
|
}, |
|
{ |
|
"epoch": 1.3445241199478488, |
|
"grad_norm": 5.304599761962891, |
|
"learning_rate": 2.1932855280312908e-05, |
|
"loss": 0.6559, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 1.3547099087353325, |
|
"grad_norm": 13.698792457580566, |
|
"learning_rate": 2.1871740547588005e-05, |
|
"loss": 0.6441, |
|
"step": 66500 |
|
}, |
|
{ |
|
"epoch": 1.3648956975228161, |
|
"grad_norm": 8.951632499694824, |
|
"learning_rate": 2.1810625814863103e-05, |
|
"loss": 0.6564, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 1.3750814863102998, |
|
"grad_norm": 5.84067440032959, |
|
"learning_rate": 2.17495110821382e-05, |
|
"loss": 0.6686, |
|
"step": 67500 |
|
}, |
|
{ |
|
"epoch": 1.3852672750977835, |
|
"grad_norm": 6.442080497741699, |
|
"learning_rate": 2.16883963494133e-05, |
|
"loss": 0.6517, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 1.3954530638852674, |
|
"grad_norm": 5.514212608337402, |
|
"learning_rate": 2.1627281616688395e-05, |
|
"loss": 0.6664, |
|
"step": 68500 |
|
}, |
|
{ |
|
"epoch": 1.405638852672751, |
|
"grad_norm": 14.150158882141113, |
|
"learning_rate": 2.1566166883963496e-05, |
|
"loss": 0.6561, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 1.4158246414602347, |
|
"grad_norm": 17.196884155273438, |
|
"learning_rate": 2.150505215123859e-05, |
|
"loss": 0.6513, |
|
"step": 69500 |
|
}, |
|
{ |
|
"epoch": 1.4260104302477183, |
|
"grad_norm": 6.181870937347412, |
|
"learning_rate": 2.144393741851369e-05, |
|
"loss": 0.65, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 1.436196219035202, |
|
"grad_norm": 14.865707397460938, |
|
"learning_rate": 2.1382822685788786e-05, |
|
"loss": 0.6596, |
|
"step": 70500 |
|
}, |
|
{ |
|
"epoch": 1.4463820078226859, |
|
"grad_norm": 15.808574676513672, |
|
"learning_rate": 2.1321707953063886e-05, |
|
"loss": 0.6559, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 1.4565677966101696, |
|
"grad_norm": 16.76003074645996, |
|
"learning_rate": 2.1260593220338984e-05, |
|
"loss": 0.6467, |
|
"step": 71500 |
|
}, |
|
{ |
|
"epoch": 1.4667535853976532, |
|
"grad_norm": 21.466825485229492, |
|
"learning_rate": 2.119947848761408e-05, |
|
"loss": 0.6579, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 1.4769393741851369, |
|
"grad_norm": 18.81052589416504, |
|
"learning_rate": 2.113836375488918e-05, |
|
"loss": 0.6699, |
|
"step": 72500 |
|
}, |
|
{ |
|
"epoch": 1.4871251629726205, |
|
"grad_norm": 10.334957122802734, |
|
"learning_rate": 2.1077249022164277e-05, |
|
"loss": 0.655, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 1.4973109517601042, |
|
"grad_norm": 14.836852073669434, |
|
"learning_rate": 2.1016134289439378e-05, |
|
"loss": 0.66, |
|
"step": 73500 |
|
}, |
|
{ |
|
"epoch": 1.5074967405475879, |
|
"grad_norm": 21.439233779907227, |
|
"learning_rate": 2.0955019556714472e-05, |
|
"loss": 0.6486, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 1.5176825293350718, |
|
"grad_norm": 21.09993553161621, |
|
"learning_rate": 2.0893904823989573e-05, |
|
"loss": 0.6432, |
|
"step": 74500 |
|
}, |
|
{ |
|
"epoch": 1.5278683181225554, |
|
"grad_norm": 8.926955223083496, |
|
"learning_rate": 2.0832790091264667e-05, |
|
"loss": 0.6418, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 1.538054106910039, |
|
"grad_norm": 3.551163911819458, |
|
"learning_rate": 2.0771675358539768e-05, |
|
"loss": 0.6418, |
|
"step": 75500 |
|
}, |
|
{ |
|
"epoch": 1.548239895697523, |
|
"grad_norm": 10.46932315826416, |
|
"learning_rate": 2.0710560625814862e-05, |
|
"loss": 0.657, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 1.5584256844850066, |
|
"grad_norm": 10.022995948791504, |
|
"learning_rate": 2.0649445893089963e-05, |
|
"loss": 0.6564, |
|
"step": 76500 |
|
}, |
|
{ |
|
"epoch": 1.5686114732724903, |
|
"grad_norm": 15.299884796142578, |
|
"learning_rate": 2.0588331160365057e-05, |
|
"loss": 0.6535, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 1.578797262059974, |
|
"grad_norm": 11.149444580078125, |
|
"learning_rate": 2.0527216427640158e-05, |
|
"loss": 0.6409, |
|
"step": 77500 |
|
}, |
|
{ |
|
"epoch": 1.5889830508474576, |
|
"grad_norm": 4.011179447174072, |
|
"learning_rate": 2.0466101694915252e-05, |
|
"loss": 0.6299, |
|
"step": 78000 |
|
}, |
|
{ |
|
"epoch": 1.5991688396349413, |
|
"grad_norm": 13.744772911071777, |
|
"learning_rate": 2.0404986962190353e-05, |
|
"loss": 0.6511, |
|
"step": 78500 |
|
}, |
|
{ |
|
"epoch": 1.609354628422425, |
|
"grad_norm": 11.129698753356934, |
|
"learning_rate": 2.0343872229465447e-05, |
|
"loss": 0.6349, |
|
"step": 79000 |
|
}, |
|
{ |
|
"epoch": 1.6195404172099086, |
|
"grad_norm": 16.331953048706055, |
|
"learning_rate": 2.0282757496740548e-05, |
|
"loss": 0.6517, |
|
"step": 79500 |
|
}, |
|
{ |
|
"epoch": 1.6297262059973925, |
|
"grad_norm": 12.816121101379395, |
|
"learning_rate": 2.0221642764015646e-05, |
|
"loss": 0.6409, |
|
"step": 80000 |
|
}, |
|
{ |
|
"epoch": 1.6399119947848761, |
|
"grad_norm": 13.881726264953613, |
|
"learning_rate": 2.0160528031290743e-05, |
|
"loss": 0.6606, |
|
"step": 80500 |
|
}, |
|
{ |
|
"epoch": 1.6500977835723598, |
|
"grad_norm": 14.215683937072754, |
|
"learning_rate": 2.0099413298565844e-05, |
|
"loss": 0.657, |
|
"step": 81000 |
|
}, |
|
{ |
|
"epoch": 1.6602835723598437, |
|
"grad_norm": 19.557342529296875, |
|
"learning_rate": 2.003829856584094e-05, |
|
"loss": 0.6451, |
|
"step": 81500 |
|
}, |
|
{ |
|
"epoch": 1.6704693611473274, |
|
"grad_norm": 17.83816909790039, |
|
"learning_rate": 1.997718383311604e-05, |
|
"loss": 0.6397, |
|
"step": 82000 |
|
}, |
|
{ |
|
"epoch": 1.680655149934811, |
|
"grad_norm": 8.789198875427246, |
|
"learning_rate": 1.9916069100391133e-05, |
|
"loss": 0.6514, |
|
"step": 82500 |
|
}, |
|
{ |
|
"epoch": 1.6908409387222947, |
|
"grad_norm": 11.361543655395508, |
|
"learning_rate": 1.9854954367666234e-05, |
|
"loss": 0.6466, |
|
"step": 83000 |
|
}, |
|
{ |
|
"epoch": 1.7010267275097783, |
|
"grad_norm": 15.630517959594727, |
|
"learning_rate": 1.979383963494133e-05, |
|
"loss": 0.6316, |
|
"step": 83500 |
|
}, |
|
{ |
|
"epoch": 1.711212516297262, |
|
"grad_norm": 10.55432415008545, |
|
"learning_rate": 1.973272490221643e-05, |
|
"loss": 0.6533, |
|
"step": 84000 |
|
}, |
|
{ |
|
"epoch": 1.7213983050847457, |
|
"grad_norm": 8.428609848022461, |
|
"learning_rate": 1.9671610169491524e-05, |
|
"loss": 0.6478, |
|
"step": 84500 |
|
}, |
|
{ |
|
"epoch": 1.7315840938722293, |
|
"grad_norm": 4.304576873779297, |
|
"learning_rate": 1.9610495436766625e-05, |
|
"loss": 0.6449, |
|
"step": 85000 |
|
}, |
|
{ |
|
"epoch": 1.7417698826597132, |
|
"grad_norm": 21.92937469482422, |
|
"learning_rate": 1.9549380704041722e-05, |
|
"loss": 0.6425, |
|
"step": 85500 |
|
}, |
|
{ |
|
"epoch": 1.7519556714471969, |
|
"grad_norm": 19.813392639160156, |
|
"learning_rate": 1.948826597131682e-05, |
|
"loss": 0.6238, |
|
"step": 86000 |
|
}, |
|
{ |
|
"epoch": 1.7621414602346805, |
|
"grad_norm": 6.831646919250488, |
|
"learning_rate": 1.9427151238591917e-05, |
|
"loss": 0.6553, |
|
"step": 86500 |
|
}, |
|
{ |
|
"epoch": 1.7723272490221644, |
|
"grad_norm": 11.058158874511719, |
|
"learning_rate": 1.9366036505867015e-05, |
|
"loss": 0.6528, |
|
"step": 87000 |
|
}, |
|
{ |
|
"epoch": 1.782513037809648, |
|
"grad_norm": 7.01440954208374, |
|
"learning_rate": 1.9304921773142112e-05, |
|
"loss": 0.6506, |
|
"step": 87500 |
|
}, |
|
{ |
|
"epoch": 1.7926988265971318, |
|
"grad_norm": 4.963765621185303, |
|
"learning_rate": 1.924380704041721e-05, |
|
"loss": 0.6412, |
|
"step": 88000 |
|
}, |
|
{ |
|
"epoch": 1.8028846153846154, |
|
"grad_norm": 38.63767623901367, |
|
"learning_rate": 1.9182692307692307e-05, |
|
"loss": 0.6629, |
|
"step": 88500 |
|
}, |
|
{ |
|
"epoch": 1.813070404172099, |
|
"grad_norm": 11.423843383789062, |
|
"learning_rate": 1.9121577574967405e-05, |
|
"loss": 0.673, |
|
"step": 89000 |
|
}, |
|
{ |
|
"epoch": 1.8232561929595827, |
|
"grad_norm": 11.73025894165039, |
|
"learning_rate": 1.9060462842242506e-05, |
|
"loss": 0.625, |
|
"step": 89500 |
|
}, |
|
{ |
|
"epoch": 1.8334419817470664, |
|
"grad_norm": 11.493837356567383, |
|
"learning_rate": 1.89993481095176e-05, |
|
"loss": 0.6658, |
|
"step": 90000 |
|
}, |
|
{ |
|
"epoch": 1.84362777053455, |
|
"grad_norm": 12.196702003479004, |
|
"learning_rate": 1.89382333767927e-05, |
|
"loss": 0.6451, |
|
"step": 90500 |
|
}, |
|
{ |
|
"epoch": 1.8538135593220337, |
|
"grad_norm": 9.690689086914062, |
|
"learning_rate": 1.8877118644067795e-05, |
|
"loss": 0.6282, |
|
"step": 91000 |
|
}, |
|
{ |
|
"epoch": 1.8639993481095176, |
|
"grad_norm": 13.679101943969727, |
|
"learning_rate": 1.8816003911342896e-05, |
|
"loss": 0.6328, |
|
"step": 91500 |
|
}, |
|
{ |
|
"epoch": 1.8741851368970013, |
|
"grad_norm": 14.264899253845215, |
|
"learning_rate": 1.8754889178617994e-05, |
|
"loss": 0.6355, |
|
"step": 92000 |
|
}, |
|
{ |
|
"epoch": 1.8843709256844852, |
|
"grad_norm": 9.039682388305664, |
|
"learning_rate": 1.869377444589309e-05, |
|
"loss": 0.6471, |
|
"step": 92500 |
|
}, |
|
{ |
|
"epoch": 1.8945567144719688, |
|
"grad_norm": 9.861899375915527, |
|
"learning_rate": 1.863265971316819e-05, |
|
"loss": 0.6549, |
|
"step": 93000 |
|
}, |
|
{ |
|
"epoch": 1.9047425032594525, |
|
"grad_norm": 14.948457717895508, |
|
"learning_rate": 1.8571544980443286e-05, |
|
"loss": 0.6436, |
|
"step": 93500 |
|
}, |
|
{ |
|
"epoch": 1.9149282920469362, |
|
"grad_norm": 15.613487243652344, |
|
"learning_rate": 1.8510430247718384e-05, |
|
"loss": 0.6355, |
|
"step": 94000 |
|
}, |
|
{ |
|
"epoch": 1.9251140808344198, |
|
"grad_norm": 14.105792045593262, |
|
"learning_rate": 1.844931551499348e-05, |
|
"loss": 0.6345, |
|
"step": 94500 |
|
}, |
|
{ |
|
"epoch": 1.9352998696219035, |
|
"grad_norm": 7.004807472229004, |
|
"learning_rate": 1.838820078226858e-05, |
|
"loss": 0.6456, |
|
"step": 95000 |
|
}, |
|
{ |
|
"epoch": 1.9454856584093871, |
|
"grad_norm": 10.506879806518555, |
|
"learning_rate": 1.8327086049543677e-05, |
|
"loss": 0.6303, |
|
"step": 95500 |
|
}, |
|
{ |
|
"epoch": 1.9556714471968708, |
|
"grad_norm": 11.595491409301758, |
|
"learning_rate": 1.8265971316818774e-05, |
|
"loss": 0.6509, |
|
"step": 96000 |
|
}, |
|
{ |
|
"epoch": 1.9658572359843545, |
|
"grad_norm": 11.29542064666748, |
|
"learning_rate": 1.820485658409387e-05, |
|
"loss": 0.6273, |
|
"step": 96500 |
|
}, |
|
{ |
|
"epoch": 1.9760430247718384, |
|
"grad_norm": 4.132607460021973, |
|
"learning_rate": 1.814374185136897e-05, |
|
"loss": 0.643, |
|
"step": 97000 |
|
}, |
|
{ |
|
"epoch": 1.986228813559322, |
|
"grad_norm": 9.745738983154297, |
|
"learning_rate": 1.8082627118644067e-05, |
|
"loss": 0.6656, |
|
"step": 97500 |
|
}, |
|
{ |
|
"epoch": 1.996414602346806, |
|
"grad_norm": 4.289452075958252, |
|
"learning_rate": 1.8021512385919164e-05, |
|
"loss": 0.6367, |
|
"step": 98000 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.7218826413154602, |
|
"eval_loss": 0.6872764825820923, |
|
"eval_runtime": 7.6175, |
|
"eval_samples_per_second": 644.307, |
|
"eval_steps_per_second": 80.604, |
|
"step": 98176 |
|
}, |
|
{ |
|
"epoch": 2.0066003911342896, |
|
"grad_norm": 19.260536193847656, |
|
"learning_rate": 1.7960397653194265e-05, |
|
"loss": 0.5568, |
|
"step": 98500 |
|
}, |
|
{ |
|
"epoch": 2.0167861799217732, |
|
"grad_norm": 8.347018241882324, |
|
"learning_rate": 1.7899282920469363e-05, |
|
"loss": 0.5344, |
|
"step": 99000 |
|
}, |
|
{ |
|
"epoch": 2.026971968709257, |
|
"grad_norm": 12.238682746887207, |
|
"learning_rate": 1.783816818774446e-05, |
|
"loss": 0.5685, |
|
"step": 99500 |
|
}, |
|
{ |
|
"epoch": 2.0371577574967406, |
|
"grad_norm": 18.42786979675293, |
|
"learning_rate": 1.7777053455019558e-05, |
|
"loss": 0.5539, |
|
"step": 100000 |
|
}, |
|
{ |
|
"epoch": 2.047343546284224, |
|
"grad_norm": 25.627717971801758, |
|
"learning_rate": 1.7715938722294655e-05, |
|
"loss": 0.5393, |
|
"step": 100500 |
|
}, |
|
{ |
|
"epoch": 2.057529335071708, |
|
"grad_norm": 15.942806243896484, |
|
"learning_rate": 1.7654823989569753e-05, |
|
"loss": 0.5191, |
|
"step": 101000 |
|
}, |
|
{ |
|
"epoch": 2.0677151238591915, |
|
"grad_norm": 7.591663360595703, |
|
"learning_rate": 1.759370925684485e-05, |
|
"loss": 0.5428, |
|
"step": 101500 |
|
}, |
|
{ |
|
"epoch": 2.077900912646675, |
|
"grad_norm": 7.183356285095215, |
|
"learning_rate": 1.7532594524119948e-05, |
|
"loss": 0.5475, |
|
"step": 102000 |
|
}, |
|
{ |
|
"epoch": 2.088086701434159, |
|
"grad_norm": 19.916288375854492, |
|
"learning_rate": 1.7471479791395046e-05, |
|
"loss": 0.5586, |
|
"step": 102500 |
|
}, |
|
{ |
|
"epoch": 2.098272490221643, |
|
"grad_norm": 9.07480239868164, |
|
"learning_rate": 1.7410365058670143e-05, |
|
"loss": 0.5564, |
|
"step": 103000 |
|
}, |
|
{ |
|
"epoch": 2.1084582790091266, |
|
"grad_norm": 7.647058010101318, |
|
"learning_rate": 1.734925032594524e-05, |
|
"loss": 0.543, |
|
"step": 103500 |
|
}, |
|
{ |
|
"epoch": 2.1186440677966103, |
|
"grad_norm": 8.871644973754883, |
|
"learning_rate": 1.728813559322034e-05, |
|
"loss": 0.5623, |
|
"step": 104000 |
|
}, |
|
{ |
|
"epoch": 2.128829856584094, |
|
"grad_norm": 9.516389846801758, |
|
"learning_rate": 1.7227020860495436e-05, |
|
"loss": 0.5464, |
|
"step": 104500 |
|
}, |
|
{ |
|
"epoch": 2.1390156453715776, |
|
"grad_norm": 13.605746269226074, |
|
"learning_rate": 1.7165906127770537e-05, |
|
"loss": 0.5425, |
|
"step": 105000 |
|
}, |
|
{ |
|
"epoch": 2.1492014341590613, |
|
"grad_norm": 14.016572952270508, |
|
"learning_rate": 1.710479139504563e-05, |
|
"loss": 0.5496, |
|
"step": 105500 |
|
}, |
|
{ |
|
"epoch": 2.159387222946545, |
|
"grad_norm": 14.719120979309082, |
|
"learning_rate": 1.7043676662320732e-05, |
|
"loss": 0.5591, |
|
"step": 106000 |
|
}, |
|
{ |
|
"epoch": 2.1695730117340286, |
|
"grad_norm": 14.922130584716797, |
|
"learning_rate": 1.6982561929595826e-05, |
|
"loss": 0.5462, |
|
"step": 106500 |
|
}, |
|
{ |
|
"epoch": 2.1797588005215123, |
|
"grad_norm": 10.74787425994873, |
|
"learning_rate": 1.6921447196870927e-05, |
|
"loss": 0.5699, |
|
"step": 107000 |
|
}, |
|
{ |
|
"epoch": 2.189944589308996, |
|
"grad_norm": 7.978420257568359, |
|
"learning_rate": 1.6860332464146024e-05, |
|
"loss": 0.5449, |
|
"step": 107500 |
|
}, |
|
{ |
|
"epoch": 2.2001303780964796, |
|
"grad_norm": 15.361347198486328, |
|
"learning_rate": 1.6799217731421122e-05, |
|
"loss": 0.5443, |
|
"step": 108000 |
|
}, |
|
{ |
|
"epoch": 2.2103161668839633, |
|
"grad_norm": 6.552661418914795, |
|
"learning_rate": 1.673810299869622e-05, |
|
"loss": 0.5663, |
|
"step": 108500 |
|
}, |
|
{ |
|
"epoch": 2.2205019556714474, |
|
"grad_norm": 6.750521659851074, |
|
"learning_rate": 1.6676988265971317e-05, |
|
"loss": 0.5488, |
|
"step": 109000 |
|
}, |
|
{ |
|
"epoch": 2.230687744458931, |
|
"grad_norm": 11.149799346923828, |
|
"learning_rate": 1.6615873533246415e-05, |
|
"loss": 0.5412, |
|
"step": 109500 |
|
}, |
|
{ |
|
"epoch": 2.2408735332464147, |
|
"grad_norm": 19.362638473510742, |
|
"learning_rate": 1.6554758800521512e-05, |
|
"loss": 0.5505, |
|
"step": 110000 |
|
}, |
|
{ |
|
"epoch": 2.2510593220338984, |
|
"grad_norm": 11.263615608215332, |
|
"learning_rate": 1.6493644067796613e-05, |
|
"loss": 0.5591, |
|
"step": 110500 |
|
}, |
|
{ |
|
"epoch": 2.261245110821382, |
|
"grad_norm": 8.90456771850586, |
|
"learning_rate": 1.6432529335071707e-05, |
|
"loss": 0.5491, |
|
"step": 111000 |
|
}, |
|
{ |
|
"epoch": 2.2714308996088657, |
|
"grad_norm": 11.962569236755371, |
|
"learning_rate": 1.6371414602346808e-05, |
|
"loss": 0.556, |
|
"step": 111500 |
|
}, |
|
{ |
|
"epoch": 2.2816166883963493, |
|
"grad_norm": 8.397544860839844, |
|
"learning_rate": 1.6310299869621902e-05, |
|
"loss": 0.5561, |
|
"step": 112000 |
|
}, |
|
{ |
|
"epoch": 2.291802477183833, |
|
"grad_norm": 16.376155853271484, |
|
"learning_rate": 1.6249185136897003e-05, |
|
"loss": 0.5435, |
|
"step": 112500 |
|
}, |
|
{ |
|
"epoch": 2.3019882659713167, |
|
"grad_norm": 9.14609432220459, |
|
"learning_rate": 1.6188070404172097e-05, |
|
"loss": 0.5457, |
|
"step": 113000 |
|
}, |
|
{ |
|
"epoch": 2.3121740547588003, |
|
"grad_norm": 26.53936767578125, |
|
"learning_rate": 1.61269556714472e-05, |
|
"loss": 0.5295, |
|
"step": 113500 |
|
}, |
|
{ |
|
"epoch": 2.322359843546284, |
|
"grad_norm": 17.222721099853516, |
|
"learning_rate": 1.6065840938722293e-05, |
|
"loss": 0.5177, |
|
"step": 114000 |
|
}, |
|
{ |
|
"epoch": 2.332545632333768, |
|
"grad_norm": 9.585100173950195, |
|
"learning_rate": 1.6004726205997394e-05, |
|
"loss": 0.5419, |
|
"step": 114500 |
|
}, |
|
{ |
|
"epoch": 2.3427314211212518, |
|
"grad_norm": 9.392489433288574, |
|
"learning_rate": 1.5943611473272488e-05, |
|
"loss": 0.5296, |
|
"step": 115000 |
|
}, |
|
{ |
|
"epoch": 2.3529172099087354, |
|
"grad_norm": 15.146162986755371, |
|
"learning_rate": 1.588249674054759e-05, |
|
"loss": 0.5518, |
|
"step": 115500 |
|
}, |
|
{ |
|
"epoch": 2.363102998696219, |
|
"grad_norm": 15.631641387939453, |
|
"learning_rate": 1.582138200782269e-05, |
|
"loss": 0.5584, |
|
"step": 116000 |
|
}, |
|
{ |
|
"epoch": 2.3732887874837028, |
|
"grad_norm": 17.875492095947266, |
|
"learning_rate": 1.5760267275097784e-05, |
|
"loss": 0.5601, |
|
"step": 116500 |
|
}, |
|
{ |
|
"epoch": 2.3834745762711864, |
|
"grad_norm": 7.7471184730529785, |
|
"learning_rate": 1.5699152542372885e-05, |
|
"loss": 0.549, |
|
"step": 117000 |
|
}, |
|
{ |
|
"epoch": 2.39366036505867, |
|
"grad_norm": 14.712841987609863, |
|
"learning_rate": 1.563803780964798e-05, |
|
"loss": 0.5293, |
|
"step": 117500 |
|
}, |
|
{ |
|
"epoch": 2.4038461538461537, |
|
"grad_norm": 9.229011535644531, |
|
"learning_rate": 1.557692307692308e-05, |
|
"loss": 0.5483, |
|
"step": 118000 |
|
}, |
|
{ |
|
"epoch": 2.4140319426336374, |
|
"grad_norm": 11.47548770904541, |
|
"learning_rate": 1.5515808344198174e-05, |
|
"loss": 0.5529, |
|
"step": 118500 |
|
}, |
|
{ |
|
"epoch": 2.424217731421121, |
|
"grad_norm": 29.980873107910156, |
|
"learning_rate": 1.5454693611473275e-05, |
|
"loss": 0.5489, |
|
"step": 119000 |
|
}, |
|
{ |
|
"epoch": 2.4344035202086047, |
|
"grad_norm": 10.478185653686523, |
|
"learning_rate": 1.539357887874837e-05, |
|
"loss": 0.5595, |
|
"step": 119500 |
|
}, |
|
{ |
|
"epoch": 2.444589308996089, |
|
"grad_norm": 12.978096008300781, |
|
"learning_rate": 1.533246414602347e-05, |
|
"loss": 0.5612, |
|
"step": 120000 |
|
}, |
|
{ |
|
"epoch": 2.4547750977835725, |
|
"grad_norm": 17.834806442260742, |
|
"learning_rate": 1.5271349413298564e-05, |
|
"loss": 0.5412, |
|
"step": 120500 |
|
}, |
|
{ |
|
"epoch": 2.464960886571056, |
|
"grad_norm": 25.001754760742188, |
|
"learning_rate": 1.5210234680573665e-05, |
|
"loss": 0.5673, |
|
"step": 121000 |
|
}, |
|
{ |
|
"epoch": 2.47514667535854, |
|
"grad_norm": 27.953767776489258, |
|
"learning_rate": 1.5149119947848761e-05, |
|
"loss": 0.5497, |
|
"step": 121500 |
|
}, |
|
{ |
|
"epoch": 2.4853324641460235, |
|
"grad_norm": 9.370855331420898, |
|
"learning_rate": 1.508800521512386e-05, |
|
"loss": 0.5344, |
|
"step": 122000 |
|
}, |
|
{ |
|
"epoch": 2.495518252933507, |
|
"grad_norm": 18.109821319580078, |
|
"learning_rate": 1.5026890482398956e-05, |
|
"loss": 0.5572, |
|
"step": 122500 |
|
}, |
|
{ |
|
"epoch": 2.505704041720991, |
|
"grad_norm": 20.435270309448242, |
|
"learning_rate": 1.4965775749674055e-05, |
|
"loss": 0.5631, |
|
"step": 123000 |
|
}, |
|
{ |
|
"epoch": 2.5158898305084745, |
|
"grad_norm": 10.656272888183594, |
|
"learning_rate": 1.4904661016949153e-05, |
|
"loss": 0.5344, |
|
"step": 123500 |
|
}, |
|
{ |
|
"epoch": 2.526075619295958, |
|
"grad_norm": 18.796552658081055, |
|
"learning_rate": 1.484354628422425e-05, |
|
"loss": 0.5464, |
|
"step": 124000 |
|
}, |
|
{ |
|
"epoch": 2.5362614080834422, |
|
"grad_norm": 9.630306243896484, |
|
"learning_rate": 1.4782431551499348e-05, |
|
"loss": 0.547, |
|
"step": 124500 |
|
}, |
|
{ |
|
"epoch": 2.5464471968709255, |
|
"grad_norm": 6.3599467277526855, |
|
"learning_rate": 1.4721316818774445e-05, |
|
"loss": 0.5577, |
|
"step": 125000 |
|
}, |
|
{ |
|
"epoch": 2.5566329856584096, |
|
"grad_norm": 8.967144966125488, |
|
"learning_rate": 1.4660202086049545e-05, |
|
"loss": 0.5524, |
|
"step": 125500 |
|
}, |
|
{ |
|
"epoch": 2.5668187744458932, |
|
"grad_norm": 16.690263748168945, |
|
"learning_rate": 1.4599087353324642e-05, |
|
"loss": 0.5546, |
|
"step": 126000 |
|
}, |
|
{ |
|
"epoch": 2.577004563233377, |
|
"grad_norm": 17.713790893554688, |
|
"learning_rate": 1.453797262059974e-05, |
|
"loss": 0.5686, |
|
"step": 126500 |
|
}, |
|
{ |
|
"epoch": 2.5871903520208606, |
|
"grad_norm": 7.970760822296143, |
|
"learning_rate": 1.4476857887874837e-05, |
|
"loss": 0.5273, |
|
"step": 127000 |
|
}, |
|
{ |
|
"epoch": 2.5973761408083442, |
|
"grad_norm": 8.869507789611816, |
|
"learning_rate": 1.4415743155149935e-05, |
|
"loss": 0.5472, |
|
"step": 127500 |
|
}, |
|
{ |
|
"epoch": 2.607561929595828, |
|
"grad_norm": 17.247589111328125, |
|
"learning_rate": 1.4354628422425032e-05, |
|
"loss": 0.5444, |
|
"step": 128000 |
|
}, |
|
{ |
|
"epoch": 2.6177477183833116, |
|
"grad_norm": 16.415138244628906, |
|
"learning_rate": 1.429351368970013e-05, |
|
"loss": 0.5445, |
|
"step": 128500 |
|
}, |
|
{ |
|
"epoch": 2.627933507170795, |
|
"grad_norm": 10.408729553222656, |
|
"learning_rate": 1.4232398956975227e-05, |
|
"loss": 0.5667, |
|
"step": 129000 |
|
}, |
|
{ |
|
"epoch": 2.638119295958279, |
|
"grad_norm": 53.77888488769531, |
|
"learning_rate": 1.4171284224250327e-05, |
|
"loss": 0.547, |
|
"step": 129500 |
|
}, |
|
{ |
|
"epoch": 2.648305084745763, |
|
"grad_norm": 12.244620323181152, |
|
"learning_rate": 1.4110169491525424e-05, |
|
"loss": 0.5428, |
|
"step": 130000 |
|
}, |
|
{ |
|
"epoch": 2.658490873533246, |
|
"grad_norm": 11.80124568939209, |
|
"learning_rate": 1.4049054758800522e-05, |
|
"loss": 0.5565, |
|
"step": 130500 |
|
}, |
|
{ |
|
"epoch": 2.6686766623207303, |
|
"grad_norm": 13.250449180603027, |
|
"learning_rate": 1.398794002607562e-05, |
|
"loss": 0.5423, |
|
"step": 131000 |
|
}, |
|
{ |
|
"epoch": 2.678862451108214, |
|
"grad_norm": 15.81285572052002, |
|
"learning_rate": 1.3926825293350719e-05, |
|
"loss": 0.5621, |
|
"step": 131500 |
|
}, |
|
{ |
|
"epoch": 2.6890482398956976, |
|
"grad_norm": 22.584035873413086, |
|
"learning_rate": 1.3865710560625816e-05, |
|
"loss": 0.5521, |
|
"step": 132000 |
|
}, |
|
{ |
|
"epoch": 2.6992340286831813, |
|
"grad_norm": 23.865680694580078, |
|
"learning_rate": 1.3804595827900914e-05, |
|
"loss": 0.551, |
|
"step": 132500 |
|
}, |
|
{ |
|
"epoch": 2.709419817470665, |
|
"grad_norm": 18.28876304626465, |
|
"learning_rate": 1.3743481095176011e-05, |
|
"loss": 0.5451, |
|
"step": 133000 |
|
}, |
|
{ |
|
"epoch": 2.7196056062581486, |
|
"grad_norm": 16.23969078063965, |
|
"learning_rate": 1.3682366362451109e-05, |
|
"loss": 0.5392, |
|
"step": 133500 |
|
}, |
|
{ |
|
"epoch": 2.7297913950456323, |
|
"grad_norm": 14.673959732055664, |
|
"learning_rate": 1.3621251629726206e-05, |
|
"loss": 0.5451, |
|
"step": 134000 |
|
}, |
|
{ |
|
"epoch": 2.739977183833116, |
|
"grad_norm": 8.019514083862305, |
|
"learning_rate": 1.3560136897001304e-05, |
|
"loss": 0.5568, |
|
"step": 134500 |
|
}, |
|
{ |
|
"epoch": 2.7501629726205996, |
|
"grad_norm": 13.495898246765137, |
|
"learning_rate": 1.3499022164276401e-05, |
|
"loss": 0.5734, |
|
"step": 135000 |
|
}, |
|
{ |
|
"epoch": 2.7603487614080837, |
|
"grad_norm": 7.548976421356201, |
|
"learning_rate": 1.3437907431551499e-05, |
|
"loss": 0.557, |
|
"step": 135500 |
|
}, |
|
{ |
|
"epoch": 2.770534550195567, |
|
"grad_norm": 20.049760818481445, |
|
"learning_rate": 1.3376792698826597e-05, |
|
"loss": 0.5658, |
|
"step": 136000 |
|
}, |
|
{ |
|
"epoch": 2.780720338983051, |
|
"grad_norm": 9.346122741699219, |
|
"learning_rate": 1.3315677966101694e-05, |
|
"loss": 0.535, |
|
"step": 136500 |
|
}, |
|
{ |
|
"epoch": 2.7909061277705347, |
|
"grad_norm": 15.080660820007324, |
|
"learning_rate": 1.3254563233376792e-05, |
|
"loss": 0.5538, |
|
"step": 137000 |
|
}, |
|
{ |
|
"epoch": 2.8010919165580184, |
|
"grad_norm": 11.485374450683594, |
|
"learning_rate": 1.3193448500651891e-05, |
|
"loss": 0.5419, |
|
"step": 137500 |
|
}, |
|
{ |
|
"epoch": 2.811277705345502, |
|
"grad_norm": 12.089446067810059, |
|
"learning_rate": 1.3132333767926988e-05, |
|
"loss": 0.5651, |
|
"step": 138000 |
|
}, |
|
{ |
|
"epoch": 2.8214634941329857, |
|
"grad_norm": 5.466490268707275, |
|
"learning_rate": 1.3071219035202088e-05, |
|
"loss": 0.5217, |
|
"step": 138500 |
|
}, |
|
{ |
|
"epoch": 2.8316492829204694, |
|
"grad_norm": 12.89148235321045, |
|
"learning_rate": 1.3010104302477185e-05, |
|
"loss": 0.5477, |
|
"step": 139000 |
|
}, |
|
{ |
|
"epoch": 2.841835071707953, |
|
"grad_norm": 5.610709190368652, |
|
"learning_rate": 1.2948989569752283e-05, |
|
"loss": 0.5377, |
|
"step": 139500 |
|
}, |
|
{ |
|
"epoch": 2.8520208604954367, |
|
"grad_norm": 26.2186222076416, |
|
"learning_rate": 1.288787483702738e-05, |
|
"loss": 0.5281, |
|
"step": 140000 |
|
}, |
|
{ |
|
"epoch": 2.8622066492829203, |
|
"grad_norm": 2.1066462993621826, |
|
"learning_rate": 1.2826760104302478e-05, |
|
"loss": 0.5387, |
|
"step": 140500 |
|
}, |
|
{ |
|
"epoch": 2.872392438070404, |
|
"grad_norm": 13.6646728515625, |
|
"learning_rate": 1.2765645371577575e-05, |
|
"loss": 0.5416, |
|
"step": 141000 |
|
}, |
|
{ |
|
"epoch": 2.8825782268578877, |
|
"grad_norm": 14.357284545898438, |
|
"learning_rate": 1.2704530638852673e-05, |
|
"loss": 0.5504, |
|
"step": 141500 |
|
}, |
|
{ |
|
"epoch": 2.8927640156453718, |
|
"grad_norm": 14.08674144744873, |
|
"learning_rate": 1.264341590612777e-05, |
|
"loss": 0.5691, |
|
"step": 142000 |
|
}, |
|
{ |
|
"epoch": 2.9029498044328554, |
|
"grad_norm": 20.460561752319336, |
|
"learning_rate": 1.2582301173402868e-05, |
|
"loss": 0.5624, |
|
"step": 142500 |
|
}, |
|
{ |
|
"epoch": 2.913135593220339, |
|
"grad_norm": 12.731225967407227, |
|
"learning_rate": 1.2521186440677966e-05, |
|
"loss": 0.5439, |
|
"step": 143000 |
|
}, |
|
{ |
|
"epoch": 2.9233213820078228, |
|
"grad_norm": 3.781522274017334, |
|
"learning_rate": 1.2460071707953065e-05, |
|
"loss": 0.5348, |
|
"step": 143500 |
|
}, |
|
{ |
|
"epoch": 2.9335071707953064, |
|
"grad_norm": 14.624204635620117, |
|
"learning_rate": 1.2398956975228162e-05, |
|
"loss": 0.5589, |
|
"step": 144000 |
|
}, |
|
{ |
|
"epoch": 2.94369295958279, |
|
"grad_norm": 18.551095962524414, |
|
"learning_rate": 1.233784224250326e-05, |
|
"loss": 0.532, |
|
"step": 144500 |
|
}, |
|
{ |
|
"epoch": 2.9538787483702738, |
|
"grad_norm": 17.094831466674805, |
|
"learning_rate": 1.2276727509778357e-05, |
|
"loss": 0.558, |
|
"step": 145000 |
|
}, |
|
{ |
|
"epoch": 2.9640645371577574, |
|
"grad_norm": 12.306907653808594, |
|
"learning_rate": 1.2215612777053455e-05, |
|
"loss": 0.542, |
|
"step": 145500 |
|
}, |
|
{ |
|
"epoch": 2.974250325945241, |
|
"grad_norm": 12.134025573730469, |
|
"learning_rate": 1.2154498044328553e-05, |
|
"loss": 0.5429, |
|
"step": 146000 |
|
}, |
|
{ |
|
"epoch": 2.9844361147327247, |
|
"grad_norm": 22.295795440673828, |
|
"learning_rate": 1.209338331160365e-05, |
|
"loss": 0.5632, |
|
"step": 146500 |
|
}, |
|
{ |
|
"epoch": 2.9946219035202084, |
|
"grad_norm": 17.749858856201172, |
|
"learning_rate": 1.2032268578878748e-05, |
|
"loss": 0.5402, |
|
"step": 147000 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.7337000966072083, |
|
"eval_loss": 0.7322831153869629, |
|
"eval_runtime": 8.3496, |
|
"eval_samples_per_second": 587.813, |
|
"eval_steps_per_second": 73.536, |
|
"step": 147264 |
|
}, |
|
{ |
|
"epoch": 3.0048076923076925, |
|
"grad_norm": 22.746931076049805, |
|
"learning_rate": 1.1971153846153847e-05, |
|
"loss": 0.5024, |
|
"step": 147500 |
|
}, |
|
{ |
|
"epoch": 3.014993481095176, |
|
"grad_norm": 19.812463760375977, |
|
"learning_rate": 1.1910039113428944e-05, |
|
"loss": 0.455, |
|
"step": 148000 |
|
}, |
|
{ |
|
"epoch": 3.02517926988266, |
|
"grad_norm": 8.608039855957031, |
|
"learning_rate": 1.1848924380704042e-05, |
|
"loss": 0.4529, |
|
"step": 148500 |
|
}, |
|
{ |
|
"epoch": 3.0353650586701435, |
|
"grad_norm": 21.67936897277832, |
|
"learning_rate": 1.178780964797914e-05, |
|
"loss": 0.467, |
|
"step": 149000 |
|
}, |
|
{ |
|
"epoch": 3.045550847457627, |
|
"grad_norm": 23.777921676635742, |
|
"learning_rate": 1.1726694915254239e-05, |
|
"loss": 0.4399, |
|
"step": 149500 |
|
}, |
|
{ |
|
"epoch": 3.055736636245111, |
|
"grad_norm": 5.8297600746154785, |
|
"learning_rate": 1.1665580182529336e-05, |
|
"loss": 0.4605, |
|
"step": 150000 |
|
}, |
|
{ |
|
"epoch": 3.0659224250325945, |
|
"grad_norm": 1.6593589782714844, |
|
"learning_rate": 1.1604465449804434e-05, |
|
"loss": 0.4668, |
|
"step": 150500 |
|
}, |
|
{ |
|
"epoch": 3.076108213820078, |
|
"grad_norm": 41.140750885009766, |
|
"learning_rate": 1.1543350717079531e-05, |
|
"loss": 0.4558, |
|
"step": 151000 |
|
}, |
|
{ |
|
"epoch": 3.086294002607562, |
|
"grad_norm": 14.093750953674316, |
|
"learning_rate": 1.1482235984354629e-05, |
|
"loss": 0.4645, |
|
"step": 151500 |
|
}, |
|
{ |
|
"epoch": 3.0964797913950455, |
|
"grad_norm": 25.050161361694336, |
|
"learning_rate": 1.1421121251629727e-05, |
|
"loss": 0.4549, |
|
"step": 152000 |
|
}, |
|
{ |
|
"epoch": 3.106665580182529, |
|
"grad_norm": 12.942317962646484, |
|
"learning_rate": 1.1360006518904824e-05, |
|
"loss": 0.4675, |
|
"step": 152500 |
|
}, |
|
{ |
|
"epoch": 3.1168513689700132, |
|
"grad_norm": 26.615070343017578, |
|
"learning_rate": 1.1298891786179922e-05, |
|
"loss": 0.4606, |
|
"step": 153000 |
|
}, |
|
{ |
|
"epoch": 3.127037157757497, |
|
"grad_norm": 35.78953552246094, |
|
"learning_rate": 1.123777705345502e-05, |
|
"loss": 0.4754, |
|
"step": 153500 |
|
}, |
|
{ |
|
"epoch": 3.1372229465449806, |
|
"grad_norm": 17.106658935546875, |
|
"learning_rate": 1.1176662320730117e-05, |
|
"loss": 0.46, |
|
"step": 154000 |
|
}, |
|
{ |
|
"epoch": 3.1474087353324642, |
|
"grad_norm": 5.81476354598999, |
|
"learning_rate": 1.1115547588005214e-05, |
|
"loss": 0.455, |
|
"step": 154500 |
|
}, |
|
{ |
|
"epoch": 3.157594524119948, |
|
"grad_norm": 13.205123901367188, |
|
"learning_rate": 1.1054432855280312e-05, |
|
"loss": 0.4443, |
|
"step": 155000 |
|
}, |
|
{ |
|
"epoch": 3.1677803129074316, |
|
"grad_norm": 44.934471130371094, |
|
"learning_rate": 1.0993318122555411e-05, |
|
"loss": 0.447, |
|
"step": 155500 |
|
}, |
|
{ |
|
"epoch": 3.1779661016949152, |
|
"grad_norm": 8.879769325256348, |
|
"learning_rate": 1.0932203389830509e-05, |
|
"loss": 0.4536, |
|
"step": 156000 |
|
}, |
|
{ |
|
"epoch": 3.188151890482399, |
|
"grad_norm": 22.281354904174805, |
|
"learning_rate": 1.0871088657105608e-05, |
|
"loss": 0.4541, |
|
"step": 156500 |
|
}, |
|
{ |
|
"epoch": 3.1983376792698825, |
|
"grad_norm": 46.504112243652344, |
|
"learning_rate": 1.0809973924380705e-05, |
|
"loss": 0.4467, |
|
"step": 157000 |
|
}, |
|
{ |
|
"epoch": 3.208523468057366, |
|
"grad_norm": 1.4721815586090088, |
|
"learning_rate": 1.0748859191655803e-05, |
|
"loss": 0.4474, |
|
"step": 157500 |
|
}, |
|
{ |
|
"epoch": 3.21870925684485, |
|
"grad_norm": 18.72897720336914, |
|
"learning_rate": 1.06877444589309e-05, |
|
"loss": 0.4465, |
|
"step": 158000 |
|
}, |
|
{ |
|
"epoch": 3.228895045632334, |
|
"grad_norm": 13.687529563903809, |
|
"learning_rate": 1.0626629726205998e-05, |
|
"loss": 0.4674, |
|
"step": 158500 |
|
}, |
|
{ |
|
"epoch": 3.2390808344198176, |
|
"grad_norm": 2.0822532176971436, |
|
"learning_rate": 1.0565514993481096e-05, |
|
"loss": 0.4516, |
|
"step": 159000 |
|
}, |
|
{ |
|
"epoch": 3.2492666232073013, |
|
"grad_norm": 15.965363502502441, |
|
"learning_rate": 1.0504400260756193e-05, |
|
"loss": 0.4582, |
|
"step": 159500 |
|
}, |
|
{ |
|
"epoch": 3.259452411994785, |
|
"grad_norm": 19.683805465698242, |
|
"learning_rate": 1.044328552803129e-05, |
|
"loss": 0.452, |
|
"step": 160000 |
|
}, |
|
{ |
|
"epoch": 3.2696382007822686, |
|
"grad_norm": 30.53873062133789, |
|
"learning_rate": 1.0382170795306388e-05, |
|
"loss": 0.4532, |
|
"step": 160500 |
|
}, |
|
{ |
|
"epoch": 3.2798239895697523, |
|
"grad_norm": 14.192473411560059, |
|
"learning_rate": 1.0321056062581486e-05, |
|
"loss": 0.4576, |
|
"step": 161000 |
|
}, |
|
{ |
|
"epoch": 3.290009778357236, |
|
"grad_norm": 19.519214630126953, |
|
"learning_rate": 1.0259941329856583e-05, |
|
"loss": 0.4531, |
|
"step": 161500 |
|
}, |
|
{ |
|
"epoch": 3.3001955671447196, |
|
"grad_norm": 11.308916091918945, |
|
"learning_rate": 1.0198826597131683e-05, |
|
"loss": 0.4374, |
|
"step": 162000 |
|
}, |
|
{ |
|
"epoch": 3.3103813559322033, |
|
"grad_norm": 22.54949378967285, |
|
"learning_rate": 1.013771186440678e-05, |
|
"loss": 0.4641, |
|
"step": 162500 |
|
}, |
|
{ |
|
"epoch": 3.320567144719687, |
|
"grad_norm": 27.339025497436523, |
|
"learning_rate": 1.0076597131681878e-05, |
|
"loss": 0.4603, |
|
"step": 163000 |
|
}, |
|
{ |
|
"epoch": 3.3307529335071706, |
|
"grad_norm": 10.614120483398438, |
|
"learning_rate": 1.0015482398956975e-05, |
|
"loss": 0.4482, |
|
"step": 163500 |
|
}, |
|
{ |
|
"epoch": 3.3409387222946547, |
|
"grad_norm": 19.92940902709961, |
|
"learning_rate": 9.954367666232073e-06, |
|
"loss": 0.4503, |
|
"step": 164000 |
|
}, |
|
{ |
|
"epoch": 3.3511245110821384, |
|
"grad_norm": 21.356348037719727, |
|
"learning_rate": 9.89325293350717e-06, |
|
"loss": 0.4542, |
|
"step": 164500 |
|
}, |
|
{ |
|
"epoch": 3.361310299869622, |
|
"grad_norm": 19.11351203918457, |
|
"learning_rate": 9.83213820078227e-06, |
|
"loss": 0.4653, |
|
"step": 165000 |
|
}, |
|
{ |
|
"epoch": 3.3714960886571057, |
|
"grad_norm": 15.949396133422852, |
|
"learning_rate": 9.771023468057367e-06, |
|
"loss": 0.4583, |
|
"step": 165500 |
|
}, |
|
{ |
|
"epoch": 3.3816818774445894, |
|
"grad_norm": 10.171916961669922, |
|
"learning_rate": 9.709908735332465e-06, |
|
"loss": 0.4747, |
|
"step": 166000 |
|
}, |
|
{ |
|
"epoch": 3.391867666232073, |
|
"grad_norm": 17.548404693603516, |
|
"learning_rate": 9.648794002607562e-06, |
|
"loss": 0.4252, |
|
"step": 166500 |
|
}, |
|
{ |
|
"epoch": 3.4020534550195567, |
|
"grad_norm": 26.116491317749023, |
|
"learning_rate": 9.58767926988266e-06, |
|
"loss": 0.4429, |
|
"step": 167000 |
|
}, |
|
{ |
|
"epoch": 3.4122392438070404, |
|
"grad_norm": 8.717327117919922, |
|
"learning_rate": 9.526564537157757e-06, |
|
"loss": 0.4752, |
|
"step": 167500 |
|
}, |
|
{ |
|
"epoch": 3.422425032594524, |
|
"grad_norm": 5.2715535163879395, |
|
"learning_rate": 9.465449804432857e-06, |
|
"loss": 0.4689, |
|
"step": 168000 |
|
}, |
|
{ |
|
"epoch": 3.4326108213820077, |
|
"grad_norm": 15.369921684265137, |
|
"learning_rate": 9.404335071707954e-06, |
|
"loss": 0.474, |
|
"step": 168500 |
|
}, |
|
{ |
|
"epoch": 3.4427966101694913, |
|
"grad_norm": 23.541234970092773, |
|
"learning_rate": 9.343220338983052e-06, |
|
"loss": 0.4717, |
|
"step": 169000 |
|
}, |
|
{ |
|
"epoch": 3.4529823989569755, |
|
"grad_norm": 25.710681915283203, |
|
"learning_rate": 9.28210560625815e-06, |
|
"loss": 0.459, |
|
"step": 169500 |
|
}, |
|
{ |
|
"epoch": 3.463168187744459, |
|
"grad_norm": 36.22392272949219, |
|
"learning_rate": 9.220990873533247e-06, |
|
"loss": 0.4374, |
|
"step": 170000 |
|
}, |
|
{ |
|
"epoch": 3.4733539765319428, |
|
"grad_norm": 17.486045837402344, |
|
"learning_rate": 9.159876140808344e-06, |
|
"loss": 0.4682, |
|
"step": 170500 |
|
}, |
|
{ |
|
"epoch": 3.4835397653194264, |
|
"grad_norm": 21.079090118408203, |
|
"learning_rate": 9.098761408083442e-06, |
|
"loss": 0.461, |
|
"step": 171000 |
|
}, |
|
{ |
|
"epoch": 3.49372555410691, |
|
"grad_norm": 47.44207763671875, |
|
"learning_rate": 9.03764667535854e-06, |
|
"loss": 0.428, |
|
"step": 171500 |
|
}, |
|
{ |
|
"epoch": 3.5039113428943938, |
|
"grad_norm": 10.158509254455566, |
|
"learning_rate": 8.976531942633637e-06, |
|
"loss": 0.4705, |
|
"step": 172000 |
|
}, |
|
{ |
|
"epoch": 3.5140971316818774, |
|
"grad_norm": 2.3025388717651367, |
|
"learning_rate": 8.915417209908735e-06, |
|
"loss": 0.4794, |
|
"step": 172500 |
|
}, |
|
{ |
|
"epoch": 3.524282920469361, |
|
"grad_norm": 7.106923580169678, |
|
"learning_rate": 8.854302477183832e-06, |
|
"loss": 0.4596, |
|
"step": 173000 |
|
}, |
|
{ |
|
"epoch": 3.5344687092568448, |
|
"grad_norm": 11.846363067626953, |
|
"learning_rate": 8.79318774445893e-06, |
|
"loss": 0.4548, |
|
"step": 173500 |
|
}, |
|
{ |
|
"epoch": 3.5446544980443284, |
|
"grad_norm": 23.35795783996582, |
|
"learning_rate": 8.73207301173403e-06, |
|
"loss": 0.4627, |
|
"step": 174000 |
|
}, |
|
{ |
|
"epoch": 3.554840286831812, |
|
"grad_norm": 51.494873046875, |
|
"learning_rate": 8.670958279009128e-06, |
|
"loss": 0.443, |
|
"step": 174500 |
|
}, |
|
{ |
|
"epoch": 3.565026075619296, |
|
"grad_norm": 17.30136489868164, |
|
"learning_rate": 8.609843546284226e-06, |
|
"loss": 0.4431, |
|
"step": 175000 |
|
}, |
|
{ |
|
"epoch": 3.5752118644067794, |
|
"grad_norm": 6.228327751159668, |
|
"learning_rate": 8.548728813559323e-06, |
|
"loss": 0.4494, |
|
"step": 175500 |
|
}, |
|
{ |
|
"epoch": 3.5853976531942635, |
|
"grad_norm": 7.742433071136475, |
|
"learning_rate": 8.48761408083442e-06, |
|
"loss": 0.4591, |
|
"step": 176000 |
|
}, |
|
{ |
|
"epoch": 3.595583441981747, |
|
"grad_norm": 17.392818450927734, |
|
"learning_rate": 8.426499348109518e-06, |
|
"loss": 0.4572, |
|
"step": 176500 |
|
}, |
|
{ |
|
"epoch": 3.605769230769231, |
|
"grad_norm": 9.273465156555176, |
|
"learning_rate": 8.365384615384616e-06, |
|
"loss": 0.4592, |
|
"step": 177000 |
|
}, |
|
{ |
|
"epoch": 3.6159550195567145, |
|
"grad_norm": 10.859349250793457, |
|
"learning_rate": 8.304269882659713e-06, |
|
"loss": 0.4507, |
|
"step": 177500 |
|
}, |
|
{ |
|
"epoch": 3.626140808344198, |
|
"grad_norm": 16.180192947387695, |
|
"learning_rate": 8.243155149934811e-06, |
|
"loss": 0.4615, |
|
"step": 178000 |
|
}, |
|
{ |
|
"epoch": 3.636326597131682, |
|
"grad_norm": 7.720341682434082, |
|
"learning_rate": 8.182040417209908e-06, |
|
"loss": 0.4614, |
|
"step": 178500 |
|
}, |
|
{ |
|
"epoch": 3.6465123859191655, |
|
"grad_norm": 10.102255821228027, |
|
"learning_rate": 8.120925684485006e-06, |
|
"loss": 0.4504, |
|
"step": 179000 |
|
}, |
|
{ |
|
"epoch": 3.656698174706649, |
|
"grad_norm": 9.698884010314941, |
|
"learning_rate": 8.059810951760104e-06, |
|
"loss": 0.4791, |
|
"step": 179500 |
|
}, |
|
{ |
|
"epoch": 3.666883963494133, |
|
"grad_norm": 13.645587921142578, |
|
"learning_rate": 7.998696219035203e-06, |
|
"loss": 0.4655, |
|
"step": 180000 |
|
}, |
|
{ |
|
"epoch": 3.677069752281617, |
|
"grad_norm": 6.0963358879089355, |
|
"learning_rate": 7.9375814863103e-06, |
|
"loss": 0.4522, |
|
"step": 180500 |
|
}, |
|
{ |
|
"epoch": 3.6872555410691, |
|
"grad_norm": 19.082395553588867, |
|
"learning_rate": 7.876466753585398e-06, |
|
"loss": 0.4606, |
|
"step": 181000 |
|
}, |
|
{ |
|
"epoch": 3.6974413298565842, |
|
"grad_norm": 21.68135643005371, |
|
"learning_rate": 7.815352020860495e-06, |
|
"loss": 0.475, |
|
"step": 181500 |
|
}, |
|
{ |
|
"epoch": 3.707627118644068, |
|
"grad_norm": 22.7216796875, |
|
"learning_rate": 7.754237288135593e-06, |
|
"loss": 0.4561, |
|
"step": 182000 |
|
}, |
|
{ |
|
"epoch": 3.7178129074315516, |
|
"grad_norm": 5.14304780960083, |
|
"learning_rate": 7.69312255541069e-06, |
|
"loss": 0.4561, |
|
"step": 182500 |
|
}, |
|
{ |
|
"epoch": 3.7279986962190352, |
|
"grad_norm": 12.879047393798828, |
|
"learning_rate": 7.63200782268579e-06, |
|
"loss": 0.4619, |
|
"step": 183000 |
|
}, |
|
{ |
|
"epoch": 3.738184485006519, |
|
"grad_norm": 20.963592529296875, |
|
"learning_rate": 7.570893089960887e-06, |
|
"loss": 0.4488, |
|
"step": 183500 |
|
}, |
|
{ |
|
"epoch": 3.7483702737940026, |
|
"grad_norm": 15.685653686523438, |
|
"learning_rate": 7.509778357235985e-06, |
|
"loss": 0.4515, |
|
"step": 184000 |
|
}, |
|
{ |
|
"epoch": 3.7585560625814862, |
|
"grad_norm": 14.296714782714844, |
|
"learning_rate": 7.4486636245110824e-06, |
|
"loss": 0.4461, |
|
"step": 184500 |
|
}, |
|
{ |
|
"epoch": 3.76874185136897, |
|
"grad_norm": 8.592365264892578, |
|
"learning_rate": 7.38754889178618e-06, |
|
"loss": 0.4571, |
|
"step": 185000 |
|
}, |
|
{ |
|
"epoch": 3.7789276401564535, |
|
"grad_norm": 32.55515670776367, |
|
"learning_rate": 7.326434159061278e-06, |
|
"loss": 0.4601, |
|
"step": 185500 |
|
}, |
|
{ |
|
"epoch": 3.7891134289439377, |
|
"grad_norm": 19.179519653320312, |
|
"learning_rate": 7.265319426336376e-06, |
|
"loss": 0.4591, |
|
"step": 186000 |
|
}, |
|
{ |
|
"epoch": 3.799299217731421, |
|
"grad_norm": 11.681869506835938, |
|
"learning_rate": 7.2042046936114735e-06, |
|
"loss": 0.4638, |
|
"step": 186500 |
|
}, |
|
{ |
|
"epoch": 3.809485006518905, |
|
"grad_norm": 17.548564910888672, |
|
"learning_rate": 7.143089960886571e-06, |
|
"loss": 0.4595, |
|
"step": 187000 |
|
}, |
|
{ |
|
"epoch": 3.8196707953063886, |
|
"grad_norm": 34.2673225402832, |
|
"learning_rate": 7.0819752281616686e-06, |
|
"loss": 0.4622, |
|
"step": 187500 |
|
}, |
|
{ |
|
"epoch": 3.8298565840938723, |
|
"grad_norm": 25.60137939453125, |
|
"learning_rate": 7.020860495436767e-06, |
|
"loss": 0.4653, |
|
"step": 188000 |
|
}, |
|
{ |
|
"epoch": 3.840042372881356, |
|
"grad_norm": 14.652885437011719, |
|
"learning_rate": 6.9597457627118645e-06, |
|
"loss": 0.4484, |
|
"step": 188500 |
|
}, |
|
{ |
|
"epoch": 3.8502281616688396, |
|
"grad_norm": 20.71872901916504, |
|
"learning_rate": 6.898631029986962e-06, |
|
"loss": 0.4717, |
|
"step": 189000 |
|
}, |
|
{ |
|
"epoch": 3.8604139504563233, |
|
"grad_norm": 6.4770989418029785, |
|
"learning_rate": 6.83751629726206e-06, |
|
"loss": 0.462, |
|
"step": 189500 |
|
}, |
|
{ |
|
"epoch": 3.870599739243807, |
|
"grad_norm": 21.029338836669922, |
|
"learning_rate": 6.776401564537158e-06, |
|
"loss": 0.4312, |
|
"step": 190000 |
|
}, |
|
{ |
|
"epoch": 3.8807855280312906, |
|
"grad_norm": 27.267345428466797, |
|
"learning_rate": 6.7152868318122556e-06, |
|
"loss": 0.4623, |
|
"step": 190500 |
|
}, |
|
{ |
|
"epoch": 3.8909713168187743, |
|
"grad_norm": 16.270328521728516, |
|
"learning_rate": 6.654172099087354e-06, |
|
"loss": 0.4594, |
|
"step": 191000 |
|
}, |
|
{ |
|
"epoch": 3.9011571056062584, |
|
"grad_norm": 15.952796936035156, |
|
"learning_rate": 6.5930573663624515e-06, |
|
"loss": 0.4478, |
|
"step": 191500 |
|
}, |
|
{ |
|
"epoch": 3.9113428943937416, |
|
"grad_norm": 24.082805633544922, |
|
"learning_rate": 6.531942633637549e-06, |
|
"loss": 0.4572, |
|
"step": 192000 |
|
}, |
|
{ |
|
"epoch": 3.9215286831812257, |
|
"grad_norm": 17.784955978393555, |
|
"learning_rate": 6.470827900912647e-06, |
|
"loss": 0.453, |
|
"step": 192500 |
|
}, |
|
{ |
|
"epoch": 3.9317144719687094, |
|
"grad_norm": 15.306330680847168, |
|
"learning_rate": 6.409713168187744e-06, |
|
"loss": 0.4516, |
|
"step": 193000 |
|
}, |
|
{ |
|
"epoch": 3.941900260756193, |
|
"grad_norm": 17.07221031188965, |
|
"learning_rate": 6.348598435462842e-06, |
|
"loss": 0.4513, |
|
"step": 193500 |
|
}, |
|
{ |
|
"epoch": 3.9520860495436767, |
|
"grad_norm": 1.9630801677703857, |
|
"learning_rate": 6.287483702737941e-06, |
|
"loss": 0.4339, |
|
"step": 194000 |
|
}, |
|
{ |
|
"epoch": 3.9622718383311604, |
|
"grad_norm": 23.103532791137695, |
|
"learning_rate": 6.2263689700130385e-06, |
|
"loss": 0.475, |
|
"step": 194500 |
|
}, |
|
{ |
|
"epoch": 3.972457627118644, |
|
"grad_norm": 9.09752082824707, |
|
"learning_rate": 6.165254237288136e-06, |
|
"loss": 0.4422, |
|
"step": 195000 |
|
}, |
|
{ |
|
"epoch": 3.9826434159061277, |
|
"grad_norm": 23.36123275756836, |
|
"learning_rate": 6.104139504563234e-06, |
|
"loss": 0.4578, |
|
"step": 195500 |
|
}, |
|
{ |
|
"epoch": 3.9928292046936114, |
|
"grad_norm": 15.44927978515625, |
|
"learning_rate": 6.043024771838331e-06, |
|
"loss": 0.4815, |
|
"step": 196000 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_accuracy": 0.7300326228141785, |
|
"eval_loss": 0.8482908010482788, |
|
"eval_runtime": 7.2526, |
|
"eval_samples_per_second": 676.722, |
|
"eval_steps_per_second": 84.659, |
|
"step": 196352 |
|
}, |
|
{ |
|
"epoch": 4.003014993481095, |
|
"grad_norm": 18.638896942138672, |
|
"learning_rate": 5.981910039113429e-06, |
|
"loss": 0.4162, |
|
"step": 196500 |
|
}, |
|
{ |
|
"epoch": 4.013200782268579, |
|
"grad_norm": 23.482778549194336, |
|
"learning_rate": 5.920795306388527e-06, |
|
"loss": 0.3857, |
|
"step": 197000 |
|
}, |
|
{ |
|
"epoch": 4.023386571056062, |
|
"grad_norm": 19.941246032714844, |
|
"learning_rate": 5.859680573663625e-06, |
|
"loss": 0.4055, |
|
"step": 197500 |
|
}, |
|
{ |
|
"epoch": 4.0335723598435465, |
|
"grad_norm": 55.201602935791016, |
|
"learning_rate": 5.798565840938722e-06, |
|
"loss": 0.4054, |
|
"step": 198000 |
|
}, |
|
{ |
|
"epoch": 4.04375814863103, |
|
"grad_norm": 9.683618545532227, |
|
"learning_rate": 5.7374511082138206e-06, |
|
"loss": 0.3772, |
|
"step": 198500 |
|
}, |
|
{ |
|
"epoch": 4.053943937418514, |
|
"grad_norm": 14.284856796264648, |
|
"learning_rate": 5.676336375488918e-06, |
|
"loss": 0.3761, |
|
"step": 199000 |
|
}, |
|
{ |
|
"epoch": 4.064129726205997, |
|
"grad_norm": 33.600894927978516, |
|
"learning_rate": 5.615221642764016e-06, |
|
"loss": 0.394, |
|
"step": 199500 |
|
}, |
|
{ |
|
"epoch": 4.074315514993481, |
|
"grad_norm": 4.879152774810791, |
|
"learning_rate": 5.554106910039114e-06, |
|
"loss": 0.3754, |
|
"step": 200000 |
|
}, |
|
{ |
|
"epoch": 4.084501303780965, |
|
"grad_norm": 4.289336681365967, |
|
"learning_rate": 5.492992177314212e-06, |
|
"loss": 0.3822, |
|
"step": 200500 |
|
}, |
|
{ |
|
"epoch": 4.094687092568448, |
|
"grad_norm": 13.225927352905273, |
|
"learning_rate": 5.431877444589309e-06, |
|
"loss": 0.385, |
|
"step": 201000 |
|
}, |
|
{ |
|
"epoch": 4.1048728813559325, |
|
"grad_norm": 30.11479949951172, |
|
"learning_rate": 5.370762711864407e-06, |
|
"loss": 0.4043, |
|
"step": 201500 |
|
}, |
|
{ |
|
"epoch": 4.115058670143416, |
|
"grad_norm": 30.226152420043945, |
|
"learning_rate": 5.309647979139504e-06, |
|
"loss": 0.3785, |
|
"step": 202000 |
|
}, |
|
{ |
|
"epoch": 4.1252444589309, |
|
"grad_norm": 57.69206619262695, |
|
"learning_rate": 5.248533246414602e-06, |
|
"loss": 0.3783, |
|
"step": 202500 |
|
}, |
|
{ |
|
"epoch": 4.135430247718383, |
|
"grad_norm": 0.5498570203781128, |
|
"learning_rate": 5.187418513689701e-06, |
|
"loss": 0.387, |
|
"step": 203000 |
|
}, |
|
{ |
|
"epoch": 4.145616036505867, |
|
"grad_norm": 1.2860363721847534, |
|
"learning_rate": 5.126303780964799e-06, |
|
"loss": 0.3788, |
|
"step": 203500 |
|
}, |
|
{ |
|
"epoch": 4.15580182529335, |
|
"grad_norm": 24.312036514282227, |
|
"learning_rate": 5.065189048239896e-06, |
|
"loss": 0.3958, |
|
"step": 204000 |
|
}, |
|
{ |
|
"epoch": 4.1659876140808345, |
|
"grad_norm": 31.0595645904541, |
|
"learning_rate": 5.004074315514994e-06, |
|
"loss": 0.368, |
|
"step": 204500 |
|
}, |
|
{ |
|
"epoch": 4.176173402868318, |
|
"grad_norm": 30.00829315185547, |
|
"learning_rate": 4.942959582790091e-06, |
|
"loss": 0.387, |
|
"step": 205000 |
|
}, |
|
{ |
|
"epoch": 4.186359191655802, |
|
"grad_norm": 5.656859874725342, |
|
"learning_rate": 4.881844850065189e-06, |
|
"loss": 0.3876, |
|
"step": 205500 |
|
}, |
|
{ |
|
"epoch": 4.196544980443286, |
|
"grad_norm": 28.364885330200195, |
|
"learning_rate": 4.820730117340287e-06, |
|
"loss": 0.3974, |
|
"step": 206000 |
|
}, |
|
{ |
|
"epoch": 4.206730769230769, |
|
"grad_norm": 21.450817108154297, |
|
"learning_rate": 4.759615384615385e-06, |
|
"loss": 0.4304, |
|
"step": 206500 |
|
}, |
|
{ |
|
"epoch": 4.216916558018253, |
|
"grad_norm": 26.55447769165039, |
|
"learning_rate": 4.698500651890482e-06, |
|
"loss": 0.3598, |
|
"step": 207000 |
|
}, |
|
{ |
|
"epoch": 4.2271023468057365, |
|
"grad_norm": 63.503299713134766, |
|
"learning_rate": 4.637385919165581e-06, |
|
"loss": 0.3972, |
|
"step": 207500 |
|
}, |
|
{ |
|
"epoch": 4.237288135593221, |
|
"grad_norm": 6.910649299621582, |
|
"learning_rate": 4.576271186440678e-06, |
|
"loss": 0.3788, |
|
"step": 208000 |
|
}, |
|
{ |
|
"epoch": 4.247473924380704, |
|
"grad_norm": 6.5545172691345215, |
|
"learning_rate": 4.515156453715776e-06, |
|
"loss": 0.3978, |
|
"step": 208500 |
|
}, |
|
{ |
|
"epoch": 4.257659713168188, |
|
"grad_norm": 17.815195083618164, |
|
"learning_rate": 4.454041720990874e-06, |
|
"loss": 0.4014, |
|
"step": 209000 |
|
}, |
|
{ |
|
"epoch": 4.267845501955671, |
|
"grad_norm": 33.210113525390625, |
|
"learning_rate": 4.392926988265972e-06, |
|
"loss": 0.3891, |
|
"step": 209500 |
|
}, |
|
{ |
|
"epoch": 4.278031290743155, |
|
"grad_norm": 33.96702194213867, |
|
"learning_rate": 4.331812255541069e-06, |
|
"loss": 0.3891, |
|
"step": 210000 |
|
}, |
|
{ |
|
"epoch": 4.2882170795306385, |
|
"grad_norm": 13.674240112304688, |
|
"learning_rate": 4.270697522816167e-06, |
|
"loss": 0.3743, |
|
"step": 210500 |
|
}, |
|
{ |
|
"epoch": 4.298402868318123, |
|
"grad_norm": 1.694655179977417, |
|
"learning_rate": 4.209582790091264e-06, |
|
"loss": 0.4072, |
|
"step": 211000 |
|
}, |
|
{ |
|
"epoch": 4.308588657105606, |
|
"grad_norm": 10.407938957214355, |
|
"learning_rate": 4.148468057366362e-06, |
|
"loss": 0.3788, |
|
"step": 211500 |
|
}, |
|
{ |
|
"epoch": 4.31877444589309, |
|
"grad_norm": 28.451736450195312, |
|
"learning_rate": 4.087353324641461e-06, |
|
"loss": 0.44, |
|
"step": 212000 |
|
}, |
|
{ |
|
"epoch": 4.328960234680574, |
|
"grad_norm": 40.655616760253906, |
|
"learning_rate": 4.026238591916559e-06, |
|
"loss": 0.3846, |
|
"step": 212500 |
|
}, |
|
{ |
|
"epoch": 4.339146023468057, |
|
"grad_norm": 33.640682220458984, |
|
"learning_rate": 3.965123859191656e-06, |
|
"loss": 0.3967, |
|
"step": 213000 |
|
}, |
|
{ |
|
"epoch": 4.349331812255541, |
|
"grad_norm": 77.51473236083984, |
|
"learning_rate": 3.904009126466754e-06, |
|
"loss": 0.384, |
|
"step": 213500 |
|
}, |
|
{ |
|
"epoch": 4.3595176010430245, |
|
"grad_norm": 15.64946174621582, |
|
"learning_rate": 3.842894393741851e-06, |
|
"loss": 0.3935, |
|
"step": 214000 |
|
}, |
|
{ |
|
"epoch": 4.369703389830509, |
|
"grad_norm": 38.03608703613281, |
|
"learning_rate": 3.7817796610169493e-06, |
|
"loss": 0.3933, |
|
"step": 214500 |
|
}, |
|
{ |
|
"epoch": 4.379889178617992, |
|
"grad_norm": 24.385404586791992, |
|
"learning_rate": 3.720664928292047e-06, |
|
"loss": 0.3879, |
|
"step": 215000 |
|
}, |
|
{ |
|
"epoch": 4.390074967405476, |
|
"grad_norm": 19.359683990478516, |
|
"learning_rate": 3.659550195567145e-06, |
|
"loss": 0.4027, |
|
"step": 215500 |
|
}, |
|
{ |
|
"epoch": 4.400260756192959, |
|
"grad_norm": 20.058656692504883, |
|
"learning_rate": 3.598435462842243e-06, |
|
"loss": 0.3947, |
|
"step": 216000 |
|
}, |
|
{ |
|
"epoch": 4.410446544980443, |
|
"grad_norm": 22.407522201538086, |
|
"learning_rate": 3.5373207301173403e-06, |
|
"loss": 0.3692, |
|
"step": 216500 |
|
}, |
|
{ |
|
"epoch": 4.4206323337679265, |
|
"grad_norm": 0.43725308775901794, |
|
"learning_rate": 3.476205997392438e-06, |
|
"loss": 0.3876, |
|
"step": 217000 |
|
}, |
|
{ |
|
"epoch": 4.430818122555411, |
|
"grad_norm": 6.114492416381836, |
|
"learning_rate": 3.4150912646675363e-06, |
|
"loss": 0.3812, |
|
"step": 217500 |
|
}, |
|
{ |
|
"epoch": 4.441003911342895, |
|
"grad_norm": 8.737237930297852, |
|
"learning_rate": 3.353976531942634e-06, |
|
"loss": 0.4218, |
|
"step": 218000 |
|
}, |
|
{ |
|
"epoch": 4.451189700130378, |
|
"grad_norm": 21.46657371520996, |
|
"learning_rate": 3.2928617992177314e-06, |
|
"loss": 0.3878, |
|
"step": 218500 |
|
}, |
|
{ |
|
"epoch": 4.461375488917862, |
|
"grad_norm": 28.781583786010742, |
|
"learning_rate": 3.2317470664928294e-06, |
|
"loss": 0.4014, |
|
"step": 219000 |
|
}, |
|
{ |
|
"epoch": 4.471561277705345, |
|
"grad_norm": 65.7254867553711, |
|
"learning_rate": 3.170632333767927e-06, |
|
"loss": 0.3846, |
|
"step": 219500 |
|
}, |
|
{ |
|
"epoch": 4.481747066492829, |
|
"grad_norm": 42.237693786621094, |
|
"learning_rate": 3.109517601043025e-06, |
|
"loss": 0.3945, |
|
"step": 220000 |
|
}, |
|
{ |
|
"epoch": 4.491932855280313, |
|
"grad_norm": 22.627384185791016, |
|
"learning_rate": 3.048402868318123e-06, |
|
"loss": 0.3846, |
|
"step": 220500 |
|
}, |
|
{ |
|
"epoch": 4.502118644067797, |
|
"grad_norm": 5.28376579284668, |
|
"learning_rate": 2.9872881355932204e-06, |
|
"loss": 0.3951, |
|
"step": 221000 |
|
}, |
|
{ |
|
"epoch": 4.51230443285528, |
|
"grad_norm": 29.31270980834961, |
|
"learning_rate": 2.926173402868318e-06, |
|
"loss": 0.387, |
|
"step": 221500 |
|
}, |
|
{ |
|
"epoch": 4.522490221642764, |
|
"grad_norm": 28.06836700439453, |
|
"learning_rate": 2.8650586701434163e-06, |
|
"loss": 0.3758, |
|
"step": 222000 |
|
}, |
|
{ |
|
"epoch": 4.532676010430247, |
|
"grad_norm": 19.28433609008789, |
|
"learning_rate": 2.803943937418514e-06, |
|
"loss": 0.4011, |
|
"step": 222500 |
|
}, |
|
{ |
|
"epoch": 4.542861799217731, |
|
"grad_norm": 28.01453399658203, |
|
"learning_rate": 2.7428292046936114e-06, |
|
"loss": 0.4123, |
|
"step": 223000 |
|
}, |
|
{ |
|
"epoch": 4.5530475880052155, |
|
"grad_norm": 22.563831329345703, |
|
"learning_rate": 2.6817144719687094e-06, |
|
"loss": 0.3901, |
|
"step": 223500 |
|
}, |
|
{ |
|
"epoch": 4.563233376792699, |
|
"grad_norm": 6.881279945373535, |
|
"learning_rate": 2.620599739243807e-06, |
|
"loss": 0.3641, |
|
"step": 224000 |
|
}, |
|
{ |
|
"epoch": 4.573419165580183, |
|
"grad_norm": 10.799131393432617, |
|
"learning_rate": 2.559485006518905e-06, |
|
"loss": 0.388, |
|
"step": 224500 |
|
}, |
|
{ |
|
"epoch": 4.583604954367666, |
|
"grad_norm": 8.748978614807129, |
|
"learning_rate": 2.498370273794003e-06, |
|
"loss": 0.3855, |
|
"step": 225000 |
|
}, |
|
{ |
|
"epoch": 4.59379074315515, |
|
"grad_norm": 30.88077163696289, |
|
"learning_rate": 2.4372555410691004e-06, |
|
"loss": 0.3846, |
|
"step": 225500 |
|
}, |
|
{ |
|
"epoch": 4.603976531942633, |
|
"grad_norm": 9.857905387878418, |
|
"learning_rate": 2.376140808344198e-06, |
|
"loss": 0.3887, |
|
"step": 226000 |
|
}, |
|
{ |
|
"epoch": 4.6141623207301175, |
|
"grad_norm": 18.84724235534668, |
|
"learning_rate": 2.3150260756192964e-06, |
|
"loss": 0.3981, |
|
"step": 226500 |
|
}, |
|
{ |
|
"epoch": 4.624348109517601, |
|
"grad_norm": 1.5743709802627563, |
|
"learning_rate": 2.253911342894394e-06, |
|
"loss": 0.4198, |
|
"step": 227000 |
|
}, |
|
{ |
|
"epoch": 4.634533898305085, |
|
"grad_norm": 1.2036515474319458, |
|
"learning_rate": 2.1927966101694915e-06, |
|
"loss": 0.3883, |
|
"step": 227500 |
|
}, |
|
{ |
|
"epoch": 4.644719687092568, |
|
"grad_norm": 25.730968475341797, |
|
"learning_rate": 2.1316818774445895e-06, |
|
"loss": 0.3791, |
|
"step": 228000 |
|
}, |
|
{ |
|
"epoch": 4.654905475880052, |
|
"grad_norm": 32.65802764892578, |
|
"learning_rate": 2.070567144719687e-06, |
|
"loss": 0.3772, |
|
"step": 228500 |
|
}, |
|
{ |
|
"epoch": 4.665091264667536, |
|
"grad_norm": 3.2796003818511963, |
|
"learning_rate": 2.009452411994785e-06, |
|
"loss": 0.3994, |
|
"step": 229000 |
|
}, |
|
{ |
|
"epoch": 4.675277053455019, |
|
"grad_norm": 12.903874397277832, |
|
"learning_rate": 1.948337679269883e-06, |
|
"loss": 0.3725, |
|
"step": 229500 |
|
}, |
|
{ |
|
"epoch": 4.6854628422425035, |
|
"grad_norm": 1.0021297931671143, |
|
"learning_rate": 1.8872229465449805e-06, |
|
"loss": 0.393, |
|
"step": 230000 |
|
}, |
|
{ |
|
"epoch": 4.695648631029987, |
|
"grad_norm": 5.547977447509766, |
|
"learning_rate": 1.8261082138200783e-06, |
|
"loss": 0.3907, |
|
"step": 230500 |
|
}, |
|
{ |
|
"epoch": 4.705834419817471, |
|
"grad_norm": 5.375828742980957, |
|
"learning_rate": 1.764993481095176e-06, |
|
"loss": 0.4023, |
|
"step": 231000 |
|
}, |
|
{ |
|
"epoch": 4.716020208604954, |
|
"grad_norm": 20.087383270263672, |
|
"learning_rate": 1.703878748370274e-06, |
|
"loss": 0.3652, |
|
"step": 231500 |
|
}, |
|
{ |
|
"epoch": 4.726205997392438, |
|
"grad_norm": 19.617931365966797, |
|
"learning_rate": 1.6427640156453715e-06, |
|
"loss": 0.3842, |
|
"step": 232000 |
|
}, |
|
{ |
|
"epoch": 4.736391786179921, |
|
"grad_norm": 22.911376953125, |
|
"learning_rate": 1.5816492829204695e-06, |
|
"loss": 0.3871, |
|
"step": 232500 |
|
}, |
|
{ |
|
"epoch": 4.7465775749674055, |
|
"grad_norm": 31.00206184387207, |
|
"learning_rate": 1.5205345501955673e-06, |
|
"loss": 0.3918, |
|
"step": 233000 |
|
}, |
|
{ |
|
"epoch": 4.756763363754889, |
|
"grad_norm": 39.066619873046875, |
|
"learning_rate": 1.4594198174706648e-06, |
|
"loss": 0.3743, |
|
"step": 233500 |
|
}, |
|
{ |
|
"epoch": 4.766949152542373, |
|
"grad_norm": 9.460115432739258, |
|
"learning_rate": 1.3983050847457628e-06, |
|
"loss": 0.388, |
|
"step": 234000 |
|
}, |
|
{ |
|
"epoch": 4.777134941329857, |
|
"grad_norm": 10.770241737365723, |
|
"learning_rate": 1.3371903520208605e-06, |
|
"loss": 0.3986, |
|
"step": 234500 |
|
}, |
|
{ |
|
"epoch": 4.78732073011734, |
|
"grad_norm": 29.36595916748047, |
|
"learning_rate": 1.2760756192959583e-06, |
|
"loss": 0.3836, |
|
"step": 235000 |
|
}, |
|
{ |
|
"epoch": 4.797506518904824, |
|
"grad_norm": 21.498857498168945, |
|
"learning_rate": 1.214960886571056e-06, |
|
"loss": 0.3839, |
|
"step": 235500 |
|
}, |
|
{ |
|
"epoch": 4.8076923076923075, |
|
"grad_norm": 2.3236119747161865, |
|
"learning_rate": 1.153846153846154e-06, |
|
"loss": 0.3983, |
|
"step": 236000 |
|
}, |
|
{ |
|
"epoch": 4.817878096479792, |
|
"grad_norm": 10.451170921325684, |
|
"learning_rate": 1.0927314211212516e-06, |
|
"loss": 0.3853, |
|
"step": 236500 |
|
}, |
|
{ |
|
"epoch": 4.828063885267275, |
|
"grad_norm": 18.35541343688965, |
|
"learning_rate": 1.0316166883963496e-06, |
|
"loss": 0.3779, |
|
"step": 237000 |
|
}, |
|
{ |
|
"epoch": 4.838249674054759, |
|
"grad_norm": 56.08652877807617, |
|
"learning_rate": 9.705019556714473e-07, |
|
"loss": 0.3608, |
|
"step": 237500 |
|
}, |
|
{ |
|
"epoch": 4.848435462842242, |
|
"grad_norm": 30.668987274169922, |
|
"learning_rate": 9.09387222946545e-07, |
|
"loss": 0.3901, |
|
"step": 238000 |
|
}, |
|
{ |
|
"epoch": 4.858621251629726, |
|
"grad_norm": 44.03152084350586, |
|
"learning_rate": 8.482724902216428e-07, |
|
"loss": 0.3742, |
|
"step": 238500 |
|
}, |
|
{ |
|
"epoch": 4.8688070404172095, |
|
"grad_norm": 46.421485900878906, |
|
"learning_rate": 7.871577574967406e-07, |
|
"loss": 0.4039, |
|
"step": 239000 |
|
}, |
|
{ |
|
"epoch": 4.878992829204694, |
|
"grad_norm": 41.14004898071289, |
|
"learning_rate": 7.260430247718384e-07, |
|
"loss": 0.4113, |
|
"step": 239500 |
|
}, |
|
{ |
|
"epoch": 4.889178617992178, |
|
"grad_norm": 3.9679319858551025, |
|
"learning_rate": 6.649282920469362e-07, |
|
"loss": 0.3856, |
|
"step": 240000 |
|
}, |
|
{ |
|
"epoch": 4.899364406779661, |
|
"grad_norm": 30.101577758789062, |
|
"learning_rate": 6.038135593220339e-07, |
|
"loss": 0.3742, |
|
"step": 240500 |
|
}, |
|
{ |
|
"epoch": 4.909550195567145, |
|
"grad_norm": 12.514345169067383, |
|
"learning_rate": 5.426988265971316e-07, |
|
"loss": 0.3795, |
|
"step": 241000 |
|
}, |
|
{ |
|
"epoch": 4.919735984354628, |
|
"grad_norm": 33.0556526184082, |
|
"learning_rate": 4.815840938722295e-07, |
|
"loss": 0.3865, |
|
"step": 241500 |
|
}, |
|
{ |
|
"epoch": 4.929921773142112, |
|
"grad_norm": 36.99669647216797, |
|
"learning_rate": 4.2046936114732726e-07, |
|
"loss": 0.4271, |
|
"step": 242000 |
|
}, |
|
{ |
|
"epoch": 4.9401075619295955, |
|
"grad_norm": 2.8075144290924072, |
|
"learning_rate": 3.59354628422425e-07, |
|
"loss": 0.3867, |
|
"step": 242500 |
|
}, |
|
{ |
|
"epoch": 4.95029335071708, |
|
"grad_norm": 0.9676657319068909, |
|
"learning_rate": 2.9823989569752284e-07, |
|
"loss": 0.3994, |
|
"step": 243000 |
|
}, |
|
{ |
|
"epoch": 4.960479139504563, |
|
"grad_norm": 26.419252395629883, |
|
"learning_rate": 2.3712516297262062e-07, |
|
"loss": 0.3956, |
|
"step": 243500 |
|
}, |
|
{ |
|
"epoch": 4.970664928292047, |
|
"grad_norm": 37.89094543457031, |
|
"learning_rate": 1.7601043024771838e-07, |
|
"loss": 0.3856, |
|
"step": 244000 |
|
}, |
|
{ |
|
"epoch": 4.98085071707953, |
|
"grad_norm": 20.763473510742188, |
|
"learning_rate": 1.1489569752281617e-07, |
|
"loss": 0.4161, |
|
"step": 244500 |
|
}, |
|
{ |
|
"epoch": 4.991036505867014, |
|
"grad_norm": 10.952351570129395, |
|
"learning_rate": 5.378096479791395e-08, |
|
"loss": 0.4074, |
|
"step": 245000 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_accuracy": 0.7310513257980347, |
|
"eval_loss": 1.0102216005325317, |
|
"eval_runtime": 7.6909, |
|
"eval_samples_per_second": 638.153, |
|
"eval_steps_per_second": 79.834, |
|
"step": 245440 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"step": 245440, |
|
"total_flos": 1.158449687808768e+17, |
|
"train_loss": 0.5704954994082295, |
|
"train_runtime": 14864.7666, |
|
"train_samples_per_second": 132.092, |
|
"train_steps_per_second": 16.512 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 245440, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 5, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 3, |
|
"early_stopping_threshold": 0.001 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 0 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.158449687808768e+17, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|