|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.9516041326808048, |
|
"eval_steps": 50, |
|
"global_step": 3500, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0027188689505165853, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7094, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.005437737901033171, |
|
"grad_norm": 0.609375, |
|
"learning_rate": 1e-05, |
|
"loss": 0.7117, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.008156606851549755, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 1.5e-05, |
|
"loss": 0.7047, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.010875475802066341, |
|
"grad_norm": 0.353515625, |
|
"learning_rate": 2e-05, |
|
"loss": 0.7059, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.013594344752582926, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 2.5e-05, |
|
"loss": 0.7016, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.013594344752582926, |
|
"eval_loss": 0.6993125081062317, |
|
"eval_runtime": 144.0611, |
|
"eval_samples_per_second": 3.471, |
|
"eval_steps_per_second": 0.111, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.01631321370309951, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 3e-05, |
|
"loss": 0.7023, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.019032082653616094, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 3.5e-05, |
|
"loss": 0.6961, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.021750951604132682, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 4e-05, |
|
"loss": 0.6957, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.024469820554649267, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 4.5e-05, |
|
"loss": 0.6902, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.027188689505165852, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 5e-05, |
|
"loss": 0.6828, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.027188689505165852, |
|
"eval_loss": 0.6763437390327454, |
|
"eval_runtime": 120.6338, |
|
"eval_samples_per_second": 4.145, |
|
"eval_steps_per_second": 0.133, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.029907558455682437, |
|
"grad_norm": 0.87890625, |
|
"learning_rate": 4.9860257126886535e-05, |
|
"loss": 0.6664, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.03262642740619902, |
|
"grad_norm": 1.125, |
|
"learning_rate": 4.972051425377306e-05, |
|
"loss": 0.6402, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.03534529635671561, |
|
"grad_norm": 1.7578125, |
|
"learning_rate": 4.9580771380659594e-05, |
|
"loss": 0.591, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.03806416530723219, |
|
"grad_norm": 1.4609375, |
|
"learning_rate": 4.944102850754612e-05, |
|
"loss": 0.5555, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.040783034257748776, |
|
"grad_norm": 2.71875, |
|
"learning_rate": 4.9301285634432645e-05, |
|
"loss": 0.5365, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.040783034257748776, |
|
"eval_loss": 0.5141562223434448, |
|
"eval_runtime": 107.3892, |
|
"eval_samples_per_second": 4.656, |
|
"eval_steps_per_second": 0.149, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.043501903208265365, |
|
"grad_norm": 2.40625, |
|
"learning_rate": 4.916154276131917e-05, |
|
"loss": 0.4957, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.046220772158781946, |
|
"grad_norm": 2.5625, |
|
"learning_rate": 4.9021799888205704e-05, |
|
"loss": 0.4451, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.048939641109298535, |
|
"grad_norm": 2.28125, |
|
"learning_rate": 4.888205701509223e-05, |
|
"loss": 0.4641, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.051658510059815116, |
|
"grad_norm": 1.6875, |
|
"learning_rate": 4.874231414197876e-05, |
|
"loss": 0.434, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.054377379010331704, |
|
"grad_norm": 2.671875, |
|
"learning_rate": 4.860257126886529e-05, |
|
"loss": 0.3877, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.054377379010331704, |
|
"eval_loss": 0.41612499952316284, |
|
"eval_runtime": 106.701, |
|
"eval_samples_per_second": 4.686, |
|
"eval_steps_per_second": 0.15, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.057096247960848286, |
|
"grad_norm": 2.015625, |
|
"learning_rate": 4.846282839575182e-05, |
|
"loss": 0.3903, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.059815116911364874, |
|
"grad_norm": 3.34375, |
|
"learning_rate": 4.8323085522638347e-05, |
|
"loss": 0.3504, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.06253398586188146, |
|
"grad_norm": 3.28125, |
|
"learning_rate": 4.818334264952488e-05, |
|
"loss": 0.3375, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.06525285481239804, |
|
"grad_norm": 2.203125, |
|
"learning_rate": 4.8043599776411405e-05, |
|
"loss": 0.3176, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.06797172376291463, |
|
"grad_norm": 2.78125, |
|
"learning_rate": 4.790385690329793e-05, |
|
"loss": 0.3045, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.06797172376291463, |
|
"eval_loss": 0.32106250524520874, |
|
"eval_runtime": 105.1814, |
|
"eval_samples_per_second": 4.754, |
|
"eval_steps_per_second": 0.152, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.07069059271343121, |
|
"grad_norm": 1.5703125, |
|
"learning_rate": 4.7764114030184464e-05, |
|
"loss": 0.3239, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.0734094616639478, |
|
"grad_norm": 1.84375, |
|
"learning_rate": 4.762437115707099e-05, |
|
"loss": 0.2965, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.07612833061446438, |
|
"grad_norm": 3.203125, |
|
"learning_rate": 4.748462828395752e-05, |
|
"loss": 0.3126, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.07884719956498097, |
|
"grad_norm": 2.5625, |
|
"learning_rate": 4.734488541084405e-05, |
|
"loss": 0.2654, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.08156606851549755, |
|
"grad_norm": 2.03125, |
|
"learning_rate": 4.720514253773058e-05, |
|
"loss": 0.3025, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.08156606851549755, |
|
"eval_loss": 0.2882343828678131, |
|
"eval_runtime": 101.5553, |
|
"eval_samples_per_second": 4.923, |
|
"eval_steps_per_second": 0.158, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.08428493746601413, |
|
"grad_norm": 2.296875, |
|
"learning_rate": 4.7065399664617106e-05, |
|
"loss": 0.2987, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.08700380641653073, |
|
"grad_norm": 1.296875, |
|
"learning_rate": 4.692565679150364e-05, |
|
"loss": 0.2868, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.08972267536704731, |
|
"grad_norm": 2.40625, |
|
"learning_rate": 4.6785913918390165e-05, |
|
"loss": 0.3149, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.09244154431756389, |
|
"grad_norm": 2.015625, |
|
"learning_rate": 4.664617104527669e-05, |
|
"loss": 0.231, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.09516041326808047, |
|
"grad_norm": 1.9140625, |
|
"learning_rate": 4.650642817216322e-05, |
|
"loss": 0.2977, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.09516041326808047, |
|
"eval_loss": 0.2687968611717224, |
|
"eval_runtime": 135.1857, |
|
"eval_samples_per_second": 3.699, |
|
"eval_steps_per_second": 0.118, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.09787928221859707, |
|
"grad_norm": 1.640625, |
|
"learning_rate": 4.636668529904975e-05, |
|
"loss": 0.2553, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.10059815116911365, |
|
"grad_norm": 2.171875, |
|
"learning_rate": 4.622694242593628e-05, |
|
"loss": 0.2471, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.10331702011963023, |
|
"grad_norm": 1.8203125, |
|
"learning_rate": 4.608719955282281e-05, |
|
"loss": 0.2753, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.10603588907014681, |
|
"grad_norm": 2.09375, |
|
"learning_rate": 4.594745667970934e-05, |
|
"loss": 0.2913, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.10875475802066341, |
|
"grad_norm": 1.8515625, |
|
"learning_rate": 4.5807713806595866e-05, |
|
"loss": 0.2674, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.10875475802066341, |
|
"eval_loss": 0.25398436188697815, |
|
"eval_runtime": 111.6096, |
|
"eval_samples_per_second": 4.48, |
|
"eval_steps_per_second": 0.143, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.11147362697117999, |
|
"grad_norm": 1.8828125, |
|
"learning_rate": 4.56679709334824e-05, |
|
"loss": 0.2763, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.11419249592169657, |
|
"grad_norm": 3.125, |
|
"learning_rate": 4.5528228060368925e-05, |
|
"loss": 0.2797, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.11691136487221315, |
|
"grad_norm": 1.5703125, |
|
"learning_rate": 4.538848518725545e-05, |
|
"loss": 0.252, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.11963023382272975, |
|
"grad_norm": 2.734375, |
|
"learning_rate": 4.5248742314141976e-05, |
|
"loss": 0.3126, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.12234910277324633, |
|
"grad_norm": 2.234375, |
|
"learning_rate": 4.510899944102851e-05, |
|
"loss": 0.2595, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.12234910277324633, |
|
"eval_loss": 0.2485624998807907, |
|
"eval_runtime": 117.7724, |
|
"eval_samples_per_second": 4.245, |
|
"eval_steps_per_second": 0.136, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.1250679717237629, |
|
"grad_norm": 1.7734375, |
|
"learning_rate": 4.4969256567915035e-05, |
|
"loss": 0.2341, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.1277868406742795, |
|
"grad_norm": 2.15625, |
|
"learning_rate": 4.482951369480157e-05, |
|
"loss": 0.2337, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.13050570962479607, |
|
"grad_norm": 1.1875, |
|
"learning_rate": 4.46897708216881e-05, |
|
"loss": 0.2757, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.13322457857531267, |
|
"grad_norm": 3.390625, |
|
"learning_rate": 4.4550027948574626e-05, |
|
"loss": 0.2219, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.13594344752582926, |
|
"grad_norm": 2.34375, |
|
"learning_rate": 4.441028507546116e-05, |
|
"loss": 0.2371, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.13594344752582926, |
|
"eval_loss": 0.23534375429153442, |
|
"eval_runtime": 104.6381, |
|
"eval_samples_per_second": 4.778, |
|
"eval_steps_per_second": 0.153, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.13866231647634583, |
|
"grad_norm": 3.984375, |
|
"learning_rate": 4.4270542202347684e-05, |
|
"loss": 0.2367, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.14138118542686243, |
|
"grad_norm": 1.8515625, |
|
"learning_rate": 4.413079932923421e-05, |
|
"loss": 0.2249, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.14410005437737902, |
|
"grad_norm": 1.6484375, |
|
"learning_rate": 4.3991056456120736e-05, |
|
"loss": 0.2861, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.1468189233278956, |
|
"grad_norm": 1.1953125, |
|
"learning_rate": 4.385131358300727e-05, |
|
"loss": 0.2504, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.14953779227841218, |
|
"grad_norm": 2.03125, |
|
"learning_rate": 4.3711570709893795e-05, |
|
"loss": 0.2458, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.14953779227841218, |
|
"eval_loss": 0.2247031182050705, |
|
"eval_runtime": 110.8821, |
|
"eval_samples_per_second": 4.509, |
|
"eval_steps_per_second": 0.144, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.15225666122892875, |
|
"grad_norm": 4.875, |
|
"learning_rate": 4.357182783678033e-05, |
|
"loss": 0.2205, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.15497553017944535, |
|
"grad_norm": 1.8828125, |
|
"learning_rate": 4.343208496366685e-05, |
|
"loss": 0.2121, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.15769439912996194, |
|
"grad_norm": 2.671875, |
|
"learning_rate": 4.3292342090553386e-05, |
|
"loss": 0.2286, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.1604132680804785, |
|
"grad_norm": 1.8828125, |
|
"learning_rate": 4.315259921743992e-05, |
|
"loss": 0.2234, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.1631321370309951, |
|
"grad_norm": 2.609375, |
|
"learning_rate": 4.3012856344326444e-05, |
|
"loss": 0.2261, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.1631321370309951, |
|
"eval_loss": 0.22798436880111694, |
|
"eval_runtime": 102.5588, |
|
"eval_samples_per_second": 4.875, |
|
"eval_steps_per_second": 0.156, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.1658510059815117, |
|
"grad_norm": 2.75, |
|
"learning_rate": 4.287311347121297e-05, |
|
"loss": 0.1937, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.16856987493202827, |
|
"grad_norm": 2.921875, |
|
"learning_rate": 4.2733370598099496e-05, |
|
"loss": 0.2187, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.17128874388254486, |
|
"grad_norm": 2.453125, |
|
"learning_rate": 4.259362772498603e-05, |
|
"loss": 0.2353, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.17400761283306146, |
|
"grad_norm": 2.671875, |
|
"learning_rate": 4.2453884851872554e-05, |
|
"loss": 0.2376, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.17672648178357803, |
|
"grad_norm": 1.8828125, |
|
"learning_rate": 4.231414197875909e-05, |
|
"loss": 0.2688, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.17672648178357803, |
|
"eval_loss": 0.21031250059604645, |
|
"eval_runtime": 102.6274, |
|
"eval_samples_per_second": 4.872, |
|
"eval_steps_per_second": 0.156, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.17944535073409462, |
|
"grad_norm": 1.34375, |
|
"learning_rate": 4.217439910564561e-05, |
|
"loss": 0.2137, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.1821642196846112, |
|
"grad_norm": 1.59375, |
|
"learning_rate": 4.2034656232532146e-05, |
|
"loss": 0.245, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.18488308863512778, |
|
"grad_norm": 2.140625, |
|
"learning_rate": 4.189491335941867e-05, |
|
"loss": 0.2625, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.18760195758564438, |
|
"grad_norm": 1.8203125, |
|
"learning_rate": 4.17551704863052e-05, |
|
"loss": 0.2325, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.19032082653616095, |
|
"grad_norm": 2.9375, |
|
"learning_rate": 4.161542761319172e-05, |
|
"loss": 0.2058, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.19032082653616095, |
|
"eval_loss": 0.20640625059604645, |
|
"eval_runtime": 103.0629, |
|
"eval_samples_per_second": 4.851, |
|
"eval_steps_per_second": 0.155, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.19303969548667754, |
|
"grad_norm": 1.328125, |
|
"learning_rate": 4.1475684740078256e-05, |
|
"loss": 0.222, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.19575856443719414, |
|
"grad_norm": 1.359375, |
|
"learning_rate": 4.133594186696479e-05, |
|
"loss": 0.2537, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.1984774333877107, |
|
"grad_norm": 1.1875, |
|
"learning_rate": 4.1196198993851314e-05, |
|
"loss": 0.2147, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.2011963023382273, |
|
"grad_norm": 2.09375, |
|
"learning_rate": 4.105645612073785e-05, |
|
"loss": 0.1797, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.2039151712887439, |
|
"grad_norm": 1.2109375, |
|
"learning_rate": 4.091671324762437e-05, |
|
"loss": 0.202, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.2039151712887439, |
|
"eval_loss": 0.1913750022649765, |
|
"eval_runtime": 103.5861, |
|
"eval_samples_per_second": 4.827, |
|
"eval_steps_per_second": 0.154, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.20663404023926046, |
|
"grad_norm": 2.328125, |
|
"learning_rate": 4.0776970374510905e-05, |
|
"loss": 0.2025, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.20935290918977706, |
|
"grad_norm": 0.828125, |
|
"learning_rate": 4.063722750139743e-05, |
|
"loss": 0.2231, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.21207177814029363, |
|
"grad_norm": 2.078125, |
|
"learning_rate": 4.049748462828396e-05, |
|
"loss": 0.2158, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.21479064709081022, |
|
"grad_norm": 1.890625, |
|
"learning_rate": 4.035774175517048e-05, |
|
"loss": 0.2059, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.21750951604132682, |
|
"grad_norm": 0.984375, |
|
"learning_rate": 4.0217998882057016e-05, |
|
"loss": 0.1792, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.21750951604132682, |
|
"eval_loss": 0.20859375596046448, |
|
"eval_runtime": 99.6459, |
|
"eval_samples_per_second": 5.018, |
|
"eval_steps_per_second": 0.161, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.22022838499184338, |
|
"grad_norm": 1.765625, |
|
"learning_rate": 4.007825600894354e-05, |
|
"loss": 0.2092, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.22294725394235998, |
|
"grad_norm": 1.7265625, |
|
"learning_rate": 3.9938513135830074e-05, |
|
"loss": 0.184, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.22566612289287658, |
|
"grad_norm": 1.578125, |
|
"learning_rate": 3.979877026271661e-05, |
|
"loss": 0.1796, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.22838499184339314, |
|
"grad_norm": 2.015625, |
|
"learning_rate": 3.965902738960313e-05, |
|
"loss": 0.219, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.23110386079390974, |
|
"grad_norm": 1.078125, |
|
"learning_rate": 3.9519284516489665e-05, |
|
"loss": 0.21, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.23110386079390974, |
|
"eval_loss": 0.18590624630451202, |
|
"eval_runtime": 100.5192, |
|
"eval_samples_per_second": 4.974, |
|
"eval_steps_per_second": 0.159, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.2338227297444263, |
|
"grad_norm": 1.1484375, |
|
"learning_rate": 3.937954164337619e-05, |
|
"loss": 0.1995, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.2365415986949429, |
|
"grad_norm": 2.375, |
|
"learning_rate": 3.923979877026272e-05, |
|
"loss": 0.2072, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.2392604676454595, |
|
"grad_norm": 2.0, |
|
"learning_rate": 3.910005589714924e-05, |
|
"loss": 0.1809, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.24197933659597606, |
|
"grad_norm": 3.203125, |
|
"learning_rate": 3.8960313024035775e-05, |
|
"loss": 0.2188, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.24469820554649266, |
|
"grad_norm": 1.8828125, |
|
"learning_rate": 3.88205701509223e-05, |
|
"loss": 0.2067, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.24469820554649266, |
|
"eval_loss": 0.18979687988758087, |
|
"eval_runtime": 109.9141, |
|
"eval_samples_per_second": 4.549, |
|
"eval_steps_per_second": 0.146, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.24741707449700925, |
|
"grad_norm": 2.421875, |
|
"learning_rate": 3.8680827277808834e-05, |
|
"loss": 0.2491, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.2501359434475258, |
|
"grad_norm": 3.015625, |
|
"learning_rate": 3.854108440469536e-05, |
|
"loss": 0.1773, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.2528548123980424, |
|
"grad_norm": 1.6953125, |
|
"learning_rate": 3.840134153158189e-05, |
|
"loss": 0.1618, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.255573681348559, |
|
"grad_norm": 1.9296875, |
|
"learning_rate": 3.826159865846842e-05, |
|
"loss": 0.2098, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.2582925502990756, |
|
"grad_norm": 3.625, |
|
"learning_rate": 3.812185578535495e-05, |
|
"loss": 0.2236, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.2582925502990756, |
|
"eval_loss": 0.18760937452316284, |
|
"eval_runtime": 105.2992, |
|
"eval_samples_per_second": 4.748, |
|
"eval_steps_per_second": 0.152, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.26101141924959215, |
|
"grad_norm": 1.6328125, |
|
"learning_rate": 3.798211291224148e-05, |
|
"loss": 0.228, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.26373028820010874, |
|
"grad_norm": 1.8515625, |
|
"learning_rate": 3.7842370039128e-05, |
|
"loss": 0.1924, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.26644915715062534, |
|
"grad_norm": 1.2421875, |
|
"learning_rate": 3.7702627166014535e-05, |
|
"loss": 0.1952, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.26916802610114193, |
|
"grad_norm": 2.140625, |
|
"learning_rate": 3.756288429290106e-05, |
|
"loss": 0.226, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.27188689505165853, |
|
"grad_norm": 1.3671875, |
|
"learning_rate": 3.7423141419787594e-05, |
|
"loss": 0.2095, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.27188689505165853, |
|
"eval_loss": 0.19670312106609344, |
|
"eval_runtime": 100.0411, |
|
"eval_samples_per_second": 4.998, |
|
"eval_steps_per_second": 0.16, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.27460576400217507, |
|
"grad_norm": 1.4765625, |
|
"learning_rate": 3.728339854667412e-05, |
|
"loss": 0.197, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.27732463295269166, |
|
"grad_norm": 1.703125, |
|
"learning_rate": 3.714365567356065e-05, |
|
"loss": 0.1979, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.28004350190320826, |
|
"grad_norm": 1.9375, |
|
"learning_rate": 3.700391280044718e-05, |
|
"loss": 0.2003, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.28276237085372485, |
|
"grad_norm": 1.6953125, |
|
"learning_rate": 3.686416992733371e-05, |
|
"loss": 0.1868, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.28548123980424145, |
|
"grad_norm": 1.53125, |
|
"learning_rate": 3.6724427054220237e-05, |
|
"loss": 0.198, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.28548123980424145, |
|
"eval_loss": 0.19362500309944153, |
|
"eval_runtime": 100.7072, |
|
"eval_samples_per_second": 4.965, |
|
"eval_steps_per_second": 0.159, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.28820010875475804, |
|
"grad_norm": 1.640625, |
|
"learning_rate": 3.658468418110676e-05, |
|
"loss": 0.2059, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.2909189777052746, |
|
"grad_norm": 1.1640625, |
|
"learning_rate": 3.6444941307993295e-05, |
|
"loss": 0.1568, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.2936378466557912, |
|
"grad_norm": 1.1328125, |
|
"learning_rate": 3.630519843487982e-05, |
|
"loss": 0.1812, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.2963567156063078, |
|
"grad_norm": 2.1875, |
|
"learning_rate": 3.6165455561766354e-05, |
|
"loss": 0.1819, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.29907558455682437, |
|
"grad_norm": 1.546875, |
|
"learning_rate": 3.602571268865288e-05, |
|
"loss": 0.1739, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.29907558455682437, |
|
"eval_loss": 0.18107812106609344, |
|
"eval_runtime": 101.4517, |
|
"eval_samples_per_second": 4.928, |
|
"eval_steps_per_second": 0.158, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.30179445350734097, |
|
"grad_norm": 1.15625, |
|
"learning_rate": 3.588596981553941e-05, |
|
"loss": 0.2062, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.3045133224578575, |
|
"grad_norm": 2.0, |
|
"learning_rate": 3.574622694242594e-05, |
|
"loss": 0.2333, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.3072321914083741, |
|
"grad_norm": 1.7734375, |
|
"learning_rate": 3.560648406931247e-05, |
|
"loss": 0.1878, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.3099510603588907, |
|
"grad_norm": 2.515625, |
|
"learning_rate": 3.5466741196198996e-05, |
|
"loss": 0.2058, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.3126699293094073, |
|
"grad_norm": 2.09375, |
|
"learning_rate": 3.532699832308552e-05, |
|
"loss": 0.206, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.3126699293094073, |
|
"eval_loss": 0.182679682970047, |
|
"eval_runtime": 99.428, |
|
"eval_samples_per_second": 5.029, |
|
"eval_steps_per_second": 0.161, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.3153887982599239, |
|
"grad_norm": 1.3125, |
|
"learning_rate": 3.518725544997205e-05, |
|
"loss": 0.2013, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.3181076672104405, |
|
"grad_norm": 1.75, |
|
"learning_rate": 3.504751257685858e-05, |
|
"loss": 0.2018, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.320826536160957, |
|
"grad_norm": 1.265625, |
|
"learning_rate": 3.4907769703745107e-05, |
|
"loss": 0.2129, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.3235454051114736, |
|
"grad_norm": 1.984375, |
|
"learning_rate": 3.476802683063164e-05, |
|
"loss": 0.1912, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.3262642740619902, |
|
"grad_norm": 2.171875, |
|
"learning_rate": 3.462828395751817e-05, |
|
"loss": 0.1822, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.3262642740619902, |
|
"eval_loss": 0.18774999678134918, |
|
"eval_runtime": 99.1074, |
|
"eval_samples_per_second": 5.045, |
|
"eval_steps_per_second": 0.161, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.3289831430125068, |
|
"grad_norm": 4.5625, |
|
"learning_rate": 3.44885410844047e-05, |
|
"loss": 0.1758, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.3317020119630234, |
|
"grad_norm": 1.9921875, |
|
"learning_rate": 3.434879821129123e-05, |
|
"loss": 0.2345, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.33442088091353994, |
|
"grad_norm": 1.3359375, |
|
"learning_rate": 3.4209055338177756e-05, |
|
"loss": 0.2056, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.33713974986405654, |
|
"grad_norm": 1.8515625, |
|
"learning_rate": 3.406931246506428e-05, |
|
"loss": 0.2017, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.33985861881457313, |
|
"grad_norm": 1.9375, |
|
"learning_rate": 3.392956959195081e-05, |
|
"loss": 0.1789, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.33985861881457313, |
|
"eval_loss": 0.18862499296665192, |
|
"eval_runtime": 100.4085, |
|
"eval_samples_per_second": 4.98, |
|
"eval_steps_per_second": 0.159, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.3425774877650897, |
|
"grad_norm": 1.3046875, |
|
"learning_rate": 3.378982671883734e-05, |
|
"loss": 0.1966, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.3452963567156063, |
|
"grad_norm": 2.234375, |
|
"learning_rate": 3.3650083845723866e-05, |
|
"loss": 0.201, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.3480152256661229, |
|
"grad_norm": 1.375, |
|
"learning_rate": 3.35103409726104e-05, |
|
"loss": 0.186, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.35073409461663946, |
|
"grad_norm": 2.3125, |
|
"learning_rate": 3.3370598099496925e-05, |
|
"loss": 0.1931, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.35345296356715605, |
|
"grad_norm": 1.6015625, |
|
"learning_rate": 3.323085522638346e-05, |
|
"loss": 0.1633, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.35345296356715605, |
|
"eval_loss": 0.17839062213897705, |
|
"eval_runtime": 97.1074, |
|
"eval_samples_per_second": 5.149, |
|
"eval_steps_per_second": 0.165, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.35617183251767265, |
|
"grad_norm": 1.4453125, |
|
"learning_rate": 3.309111235326999e-05, |
|
"loss": 0.1765, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.35889070146818924, |
|
"grad_norm": 3.640625, |
|
"learning_rate": 3.2951369480156516e-05, |
|
"loss": 0.1793, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.36160957041870584, |
|
"grad_norm": 1.34375, |
|
"learning_rate": 3.281162660704304e-05, |
|
"loss": 0.1976, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.3643284393692224, |
|
"grad_norm": 1.5, |
|
"learning_rate": 3.267188373392957e-05, |
|
"loss": 0.2001, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.367047308319739, |
|
"grad_norm": 0.9140625, |
|
"learning_rate": 3.25321408608161e-05, |
|
"loss": 0.1723, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.367047308319739, |
|
"eval_loss": 0.17289061844348907, |
|
"eval_runtime": 97.8393, |
|
"eval_samples_per_second": 5.11, |
|
"eval_steps_per_second": 0.164, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.36976617727025557, |
|
"grad_norm": 1.734375, |
|
"learning_rate": 3.2392397987702626e-05, |
|
"loss": 0.2236, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.37248504622077216, |
|
"grad_norm": 1.2109375, |
|
"learning_rate": 3.225265511458916e-05, |
|
"loss": 0.1914, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.37520391517128876, |
|
"grad_norm": 2.828125, |
|
"learning_rate": 3.2112912241475685e-05, |
|
"loss": 0.1715, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.37792278412180536, |
|
"grad_norm": 2.265625, |
|
"learning_rate": 3.197316936836222e-05, |
|
"loss": 0.1694, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.3806416530723219, |
|
"grad_norm": 1.671875, |
|
"learning_rate": 3.183342649524874e-05, |
|
"loss": 0.1828, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.3806416530723219, |
|
"eval_loss": 0.1787421852350235, |
|
"eval_runtime": 105.4565, |
|
"eval_samples_per_second": 4.741, |
|
"eval_steps_per_second": 0.152, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.3833605220228385, |
|
"grad_norm": 1.4296875, |
|
"learning_rate": 3.1693683622135276e-05, |
|
"loss": 0.224, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.3860793909733551, |
|
"grad_norm": 2.484375, |
|
"learning_rate": 3.15539407490218e-05, |
|
"loss": 0.1992, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.3887982599238717, |
|
"grad_norm": 3.0625, |
|
"learning_rate": 3.141419787590833e-05, |
|
"loss": 0.2106, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.3915171288743883, |
|
"grad_norm": 1.1875, |
|
"learning_rate": 3.127445500279486e-05, |
|
"loss": 0.1732, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.3942359978249048, |
|
"grad_norm": 1.5625, |
|
"learning_rate": 3.1134712129681386e-05, |
|
"loss": 0.1865, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.3942359978249048, |
|
"eval_loss": 0.1751093715429306, |
|
"eval_runtime": 97.1038, |
|
"eval_samples_per_second": 5.149, |
|
"eval_steps_per_second": 0.165, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.3969548667754214, |
|
"grad_norm": 1.8515625, |
|
"learning_rate": 3.099496925656792e-05, |
|
"loss": 0.2113, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.399673735725938, |
|
"grad_norm": 1.453125, |
|
"learning_rate": 3.0855226383454444e-05, |
|
"loss": 0.2173, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.4023926046764546, |
|
"grad_norm": 2.046875, |
|
"learning_rate": 3.071548351034098e-05, |
|
"loss": 0.1893, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.4051114736269712, |
|
"grad_norm": 2.0, |
|
"learning_rate": 3.05757406372275e-05, |
|
"loss": 0.1885, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.4078303425774878, |
|
"grad_norm": 1.4140625, |
|
"learning_rate": 3.0435997764114032e-05, |
|
"loss": 0.1642, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.4078303425774878, |
|
"eval_loss": 0.1752343773841858, |
|
"eval_runtime": 97.2789, |
|
"eval_samples_per_second": 5.14, |
|
"eval_steps_per_second": 0.164, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.41054921152800433, |
|
"grad_norm": 2.359375, |
|
"learning_rate": 3.0296254891000558e-05, |
|
"loss": 0.189, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.4132680804785209, |
|
"grad_norm": 1.5546875, |
|
"learning_rate": 3.015651201788709e-05, |
|
"loss": 0.1399, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.4159869494290375, |
|
"grad_norm": 0.9921875, |
|
"learning_rate": 3.0016769144773617e-05, |
|
"loss": 0.2088, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.4187058183795541, |
|
"grad_norm": 1.7421875, |
|
"learning_rate": 2.9877026271660146e-05, |
|
"loss": 0.2048, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.4214246873300707, |
|
"grad_norm": 1.1640625, |
|
"learning_rate": 2.973728339854668e-05, |
|
"loss": 0.1898, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.4214246873300707, |
|
"eval_loss": 0.17342187464237213, |
|
"eval_runtime": 95.7506, |
|
"eval_samples_per_second": 5.222, |
|
"eval_steps_per_second": 0.167, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.42414355628058725, |
|
"grad_norm": 2.71875, |
|
"learning_rate": 2.9597540525433204e-05, |
|
"loss": 0.2081, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.42686242523110385, |
|
"grad_norm": 2.34375, |
|
"learning_rate": 2.9457797652319734e-05, |
|
"loss": 0.1575, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.42958129418162044, |
|
"grad_norm": 3.21875, |
|
"learning_rate": 2.931805477920626e-05, |
|
"loss": 0.1849, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.43230016313213704, |
|
"grad_norm": 1.7421875, |
|
"learning_rate": 2.9178311906092792e-05, |
|
"loss": 0.2059, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.43501903208265363, |
|
"grad_norm": 1.2421875, |
|
"learning_rate": 2.9038569032979318e-05, |
|
"loss": 0.1641, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.43501903208265363, |
|
"eval_loss": 0.1657109409570694, |
|
"eval_runtime": 96.6429, |
|
"eval_samples_per_second": 5.174, |
|
"eval_steps_per_second": 0.166, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.4377379010331702, |
|
"grad_norm": 1.953125, |
|
"learning_rate": 2.889882615986585e-05, |
|
"loss": 0.1906, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.44045676998368677, |
|
"grad_norm": 1.6328125, |
|
"learning_rate": 2.8759083286752376e-05, |
|
"loss": 0.1907, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.44317563893420336, |
|
"grad_norm": 4.625, |
|
"learning_rate": 2.8619340413638906e-05, |
|
"loss": 0.2193, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.44589450788471996, |
|
"grad_norm": 1.515625, |
|
"learning_rate": 2.847959754052543e-05, |
|
"loss": 0.1714, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.44861337683523655, |
|
"grad_norm": 1.8125, |
|
"learning_rate": 2.8339854667411964e-05, |
|
"loss": 0.1771, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.44861337683523655, |
|
"eval_loss": 0.18095313012599945, |
|
"eval_runtime": 95.4318, |
|
"eval_samples_per_second": 5.239, |
|
"eval_steps_per_second": 0.168, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.45133224578575315, |
|
"grad_norm": 1.9296875, |
|
"learning_rate": 2.820011179429849e-05, |
|
"loss": 0.1917, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.4540511147362697, |
|
"grad_norm": 1.8046875, |
|
"learning_rate": 2.806036892118502e-05, |
|
"loss": 0.1915, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.4567699836867863, |
|
"grad_norm": 2.53125, |
|
"learning_rate": 2.7920626048071552e-05, |
|
"loss": 0.1436, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.4594888526373029, |
|
"grad_norm": 0.921875, |
|
"learning_rate": 2.7780883174958078e-05, |
|
"loss": 0.1683, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.4622077215878195, |
|
"grad_norm": 1.640625, |
|
"learning_rate": 2.764114030184461e-05, |
|
"loss": 0.1762, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.4622077215878195, |
|
"eval_loss": 0.1666562557220459, |
|
"eval_runtime": 96.3207, |
|
"eval_samples_per_second": 5.191, |
|
"eval_steps_per_second": 0.166, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.46492659053833607, |
|
"grad_norm": 2.484375, |
|
"learning_rate": 2.7501397428731136e-05, |
|
"loss": 0.1977, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.4676454594888526, |
|
"grad_norm": 2.171875, |
|
"learning_rate": 2.7361654555617665e-05, |
|
"loss": 0.2066, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.4703643284393692, |
|
"grad_norm": 2.5, |
|
"learning_rate": 2.722191168250419e-05, |
|
"loss": 0.1756, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.4730831973898858, |
|
"grad_norm": 1.5625, |
|
"learning_rate": 2.7082168809390724e-05, |
|
"loss": 0.1871, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.4758020663404024, |
|
"grad_norm": 1.8046875, |
|
"learning_rate": 2.694242593627725e-05, |
|
"loss": 0.2212, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.4758020663404024, |
|
"eval_loss": 0.15598437190055847, |
|
"eval_runtime": 94.5918, |
|
"eval_samples_per_second": 5.286, |
|
"eval_steps_per_second": 0.169, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.478520935290919, |
|
"grad_norm": 2.1875, |
|
"learning_rate": 2.680268306316378e-05, |
|
"loss": 0.1998, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.4812398042414356, |
|
"grad_norm": 2.046875, |
|
"learning_rate": 2.6662940190050305e-05, |
|
"loss": 0.1886, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.4839586731919521, |
|
"grad_norm": 1.5390625, |
|
"learning_rate": 2.6523197316936838e-05, |
|
"loss": 0.1857, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.4866775421424687, |
|
"grad_norm": 1.2421875, |
|
"learning_rate": 2.638345444382337e-05, |
|
"loss": 0.1601, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.4893964110929853, |
|
"grad_norm": 1.3984375, |
|
"learning_rate": 2.6243711570709896e-05, |
|
"loss": 0.1989, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.4893964110929853, |
|
"eval_loss": 0.17681249976158142, |
|
"eval_runtime": 97.019, |
|
"eval_samples_per_second": 5.154, |
|
"eval_steps_per_second": 0.165, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.4921152800435019, |
|
"grad_norm": 0.94140625, |
|
"learning_rate": 2.6103968697596425e-05, |
|
"loss": 0.1391, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.4948341489940185, |
|
"grad_norm": 2.765625, |
|
"learning_rate": 2.596422582448295e-05, |
|
"loss": 0.211, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.49755301794453505, |
|
"grad_norm": 2.859375, |
|
"learning_rate": 2.5824482951369484e-05, |
|
"loss": 0.1558, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.5002718868950516, |
|
"grad_norm": 2.796875, |
|
"learning_rate": 2.568474007825601e-05, |
|
"loss": 0.1764, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.5029907558455683, |
|
"grad_norm": 1.609375, |
|
"learning_rate": 2.554499720514254e-05, |
|
"loss": 0.1547, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.5029907558455683, |
|
"eval_loss": 0.1746249943971634, |
|
"eval_runtime": 95.3118, |
|
"eval_samples_per_second": 5.246, |
|
"eval_steps_per_second": 0.168, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.5057096247960848, |
|
"grad_norm": 2.1875, |
|
"learning_rate": 2.5405254332029065e-05, |
|
"loss": 0.1657, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.5084284937466014, |
|
"grad_norm": 1.828125, |
|
"learning_rate": 2.5265511458915597e-05, |
|
"loss": 0.1985, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.511147362697118, |
|
"grad_norm": 1.1875, |
|
"learning_rate": 2.5125768585802123e-05, |
|
"loss": 0.1828, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.5138662316476346, |
|
"grad_norm": 1.1953125, |
|
"learning_rate": 2.4986025712688656e-05, |
|
"loss": 0.1603, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.5165851005981512, |
|
"grad_norm": 1.8203125, |
|
"learning_rate": 2.484628283957518e-05, |
|
"loss": 0.1718, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.5165851005981512, |
|
"eval_loss": 0.1659843772649765, |
|
"eval_runtime": 117.6794, |
|
"eval_samples_per_second": 4.249, |
|
"eval_steps_per_second": 0.136, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.5193039695486678, |
|
"grad_norm": 0.7421875, |
|
"learning_rate": 2.470653996646171e-05, |
|
"loss": 0.1812, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.5220228384991843, |
|
"grad_norm": 1.4375, |
|
"learning_rate": 2.456679709334824e-05, |
|
"loss": 0.1624, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.524741707449701, |
|
"grad_norm": 1.90625, |
|
"learning_rate": 2.442705422023477e-05, |
|
"loss": 0.1344, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.5274605764002175, |
|
"grad_norm": 1.6015625, |
|
"learning_rate": 2.42873113471213e-05, |
|
"loss": 0.1823, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.5301794453507341, |
|
"grad_norm": 2.515625, |
|
"learning_rate": 2.4147568474007825e-05, |
|
"loss": 0.1742, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.5301794453507341, |
|
"eval_loss": 0.1723593771457672, |
|
"eval_runtime": 104.5426, |
|
"eval_samples_per_second": 4.783, |
|
"eval_steps_per_second": 0.153, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.5328983143012507, |
|
"grad_norm": 4.375, |
|
"learning_rate": 2.4007825600894354e-05, |
|
"loss": 0.2232, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.5356171832517672, |
|
"grad_norm": 1.6640625, |
|
"learning_rate": 2.3868082727780886e-05, |
|
"loss": 0.1532, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.5383360522022839, |
|
"grad_norm": 2.265625, |
|
"learning_rate": 2.3728339854667416e-05, |
|
"loss": 0.171, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.5410549211528004, |
|
"grad_norm": 1.4375, |
|
"learning_rate": 2.358859698155394e-05, |
|
"loss": 0.2081, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.5437737901033171, |
|
"grad_norm": 1.6640625, |
|
"learning_rate": 2.344885410844047e-05, |
|
"loss": 0.182, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.5437737901033171, |
|
"eval_loss": 0.1640625, |
|
"eval_runtime": 99.8495, |
|
"eval_samples_per_second": 5.008, |
|
"eval_steps_per_second": 0.16, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.5464926590538336, |
|
"grad_norm": 1.203125, |
|
"learning_rate": 2.3309111235327e-05, |
|
"loss": 0.1363, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 0.5492115280043501, |
|
"grad_norm": 1.8984375, |
|
"learning_rate": 2.316936836221353e-05, |
|
"loss": 0.1666, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.5519303969548668, |
|
"grad_norm": 1.6796875, |
|
"learning_rate": 2.302962548910006e-05, |
|
"loss": 0.1405, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 0.5546492659053833, |
|
"grad_norm": 1.3046875, |
|
"learning_rate": 2.2889882615986584e-05, |
|
"loss": 0.1768, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.5573681348559, |
|
"grad_norm": 3.140625, |
|
"learning_rate": 2.2750139742873114e-05, |
|
"loss": 0.2092, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.5573681348559, |
|
"eval_loss": 0.1575937569141388, |
|
"eval_runtime": 99.6448, |
|
"eval_samples_per_second": 5.018, |
|
"eval_steps_per_second": 0.161, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.5600870038064165, |
|
"grad_norm": 2.09375, |
|
"learning_rate": 2.2610396869759643e-05, |
|
"loss": 0.1888, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.5628058727569332, |
|
"grad_norm": 1.921875, |
|
"learning_rate": 2.2470653996646172e-05, |
|
"loss": 0.1454, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 0.5655247417074497, |
|
"grad_norm": 1.0703125, |
|
"learning_rate": 2.2330911123532698e-05, |
|
"loss": 0.1697, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.5682436106579662, |
|
"grad_norm": 3.171875, |
|
"learning_rate": 2.219116825041923e-05, |
|
"loss": 0.19, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 0.5709624796084829, |
|
"grad_norm": 1.546875, |
|
"learning_rate": 2.205142537730576e-05, |
|
"loss": 0.1925, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.5709624796084829, |
|
"eval_loss": 0.16167187690734863, |
|
"eval_runtime": 99.0392, |
|
"eval_samples_per_second": 5.049, |
|
"eval_steps_per_second": 0.162, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.5736813485589994, |
|
"grad_norm": 1.8828125, |
|
"learning_rate": 2.191168250419229e-05, |
|
"loss": 0.1822, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 0.5764002175095161, |
|
"grad_norm": 1.4609375, |
|
"learning_rate": 2.1771939631078815e-05, |
|
"loss": 0.1685, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.5791190864600326, |
|
"grad_norm": 1.0546875, |
|
"learning_rate": 2.1632196757965344e-05, |
|
"loss": 0.2201, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 0.5818379554105492, |
|
"grad_norm": 1.0625, |
|
"learning_rate": 2.1492453884851873e-05, |
|
"loss": 0.1572, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.5845568243610658, |
|
"grad_norm": 0.9140625, |
|
"learning_rate": 2.1352711011738403e-05, |
|
"loss": 0.1507, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.5845568243610658, |
|
"eval_loss": 0.15681250393390656, |
|
"eval_runtime": 96.4445, |
|
"eval_samples_per_second": 5.184, |
|
"eval_steps_per_second": 0.166, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.5872756933115824, |
|
"grad_norm": 2.9375, |
|
"learning_rate": 2.1212968138624932e-05, |
|
"loss": 0.1653, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.589994562262099, |
|
"grad_norm": 1.5625, |
|
"learning_rate": 2.1073225265511458e-05, |
|
"loss": 0.1815, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 0.5927134312126155, |
|
"grad_norm": 1.15625, |
|
"learning_rate": 2.0933482392397987e-05, |
|
"loss": 0.1157, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.5954323001631321, |
|
"grad_norm": 1.890625, |
|
"learning_rate": 2.0793739519284516e-05, |
|
"loss": 0.1535, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 0.5981511691136487, |
|
"grad_norm": 1.21875, |
|
"learning_rate": 2.0653996646171045e-05, |
|
"loss": 0.1547, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.5981511691136487, |
|
"eval_loss": 0.16277343034744263, |
|
"eval_runtime": 96.1155, |
|
"eval_samples_per_second": 5.202, |
|
"eval_steps_per_second": 0.166, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.6008700380641653, |
|
"grad_norm": 2.09375, |
|
"learning_rate": 2.0514253773057575e-05, |
|
"loss": 0.1582, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 0.6035889070146819, |
|
"grad_norm": 1.3203125, |
|
"learning_rate": 2.0374510899944104e-05, |
|
"loss": 0.1592, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.6063077759651985, |
|
"grad_norm": 2.6875, |
|
"learning_rate": 2.0234768026830633e-05, |
|
"loss": 0.1771, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 0.609026644915715, |
|
"grad_norm": 1.9921875, |
|
"learning_rate": 2.0095025153717162e-05, |
|
"loss": 0.169, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.6117455138662317, |
|
"grad_norm": 2.140625, |
|
"learning_rate": 1.995528228060369e-05, |
|
"loss": 0.1685, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.6117455138662317, |
|
"eval_loss": 0.16175781190395355, |
|
"eval_runtime": 99.1628, |
|
"eval_samples_per_second": 5.042, |
|
"eval_steps_per_second": 0.161, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.6144643828167482, |
|
"grad_norm": 1.234375, |
|
"learning_rate": 1.9815539407490218e-05, |
|
"loss": 0.165, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.6171832517672649, |
|
"grad_norm": 2.140625, |
|
"learning_rate": 1.9675796534376747e-05, |
|
"loss": 0.1928, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 0.6199021207177814, |
|
"grad_norm": 1.6484375, |
|
"learning_rate": 1.9536053661263276e-05, |
|
"loss": 0.2038, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.622620989668298, |
|
"grad_norm": 1.828125, |
|
"learning_rate": 1.9396310788149805e-05, |
|
"loss": 0.1871, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 0.6253398586188146, |
|
"grad_norm": 1.078125, |
|
"learning_rate": 1.9256567915036335e-05, |
|
"loss": 0.1733, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.6253398586188146, |
|
"eval_loss": 0.15731249749660492, |
|
"eval_runtime": 95.4546, |
|
"eval_samples_per_second": 5.238, |
|
"eval_steps_per_second": 0.168, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.6280587275693311, |
|
"grad_norm": 1.40625, |
|
"learning_rate": 1.911682504192286e-05, |
|
"loss": 0.1706, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 0.6307775965198478, |
|
"grad_norm": 3.078125, |
|
"learning_rate": 1.897708216880939e-05, |
|
"loss": 0.1856, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.6334964654703643, |
|
"grad_norm": 1.375, |
|
"learning_rate": 1.8837339295695922e-05, |
|
"loss": 0.1852, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 0.636215334420881, |
|
"grad_norm": 2.3125, |
|
"learning_rate": 1.869759642258245e-05, |
|
"loss": 0.1748, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.6389342033713975, |
|
"grad_norm": 1.2890625, |
|
"learning_rate": 1.8557853549468977e-05, |
|
"loss": 0.1576, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.6389342033713975, |
|
"eval_loss": 0.15459375083446503, |
|
"eval_runtime": 94.7107, |
|
"eval_samples_per_second": 5.279, |
|
"eval_steps_per_second": 0.169, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.641653072321914, |
|
"grad_norm": 1.4921875, |
|
"learning_rate": 1.8418110676355507e-05, |
|
"loss": 0.1657, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.6443719412724307, |
|
"grad_norm": 1.8359375, |
|
"learning_rate": 1.8278367803242036e-05, |
|
"loss": 0.1725, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 0.6470908102229472, |
|
"grad_norm": 1.90625, |
|
"learning_rate": 1.8138624930128565e-05, |
|
"loss": 0.1667, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.6498096791734639, |
|
"grad_norm": 2.203125, |
|
"learning_rate": 1.7998882057015094e-05, |
|
"loss": 0.1468, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 0.6525285481239804, |
|
"grad_norm": 2.3125, |
|
"learning_rate": 1.785913918390162e-05, |
|
"loss": 0.1852, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.6525285481239804, |
|
"eval_loss": 0.14464062452316284, |
|
"eval_runtime": 109.621, |
|
"eval_samples_per_second": 4.561, |
|
"eval_steps_per_second": 0.146, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.655247417074497, |
|
"grad_norm": 2.015625, |
|
"learning_rate": 1.771939631078815e-05, |
|
"loss": 0.1942, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 0.6579662860250136, |
|
"grad_norm": 1.9453125, |
|
"learning_rate": 1.757965343767468e-05, |
|
"loss": 0.1513, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.6606851549755302, |
|
"grad_norm": 1.1875, |
|
"learning_rate": 1.7439910564561208e-05, |
|
"loss": 0.1445, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 0.6634040239260468, |
|
"grad_norm": 3.171875, |
|
"learning_rate": 1.7300167691447737e-05, |
|
"loss": 0.1874, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.6661228928765633, |
|
"grad_norm": 2.5625, |
|
"learning_rate": 1.7160424818334266e-05, |
|
"loss": 0.1697, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.6661228928765633, |
|
"eval_loss": 0.14278124272823334, |
|
"eval_runtime": 99.7567, |
|
"eval_samples_per_second": 5.012, |
|
"eval_steps_per_second": 0.16, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.6688417618270799, |
|
"grad_norm": 1.6171875, |
|
"learning_rate": 1.7020681945220796e-05, |
|
"loss": 0.1505, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.6715606307775965, |
|
"grad_norm": 1.1875, |
|
"learning_rate": 1.6880939072107325e-05, |
|
"loss": 0.2015, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 0.6742794997281131, |
|
"grad_norm": 1.5625, |
|
"learning_rate": 1.6741196198993854e-05, |
|
"loss": 0.1782, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.6769983686786297, |
|
"grad_norm": 3.140625, |
|
"learning_rate": 1.660145332588038e-05, |
|
"loss": 0.159, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 0.6797172376291463, |
|
"grad_norm": 1.265625, |
|
"learning_rate": 1.646171045276691e-05, |
|
"loss": 0.2079, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.6797172376291463, |
|
"eval_loss": 0.15110155940055847, |
|
"eval_runtime": 101.2462, |
|
"eval_samples_per_second": 4.938, |
|
"eval_steps_per_second": 0.158, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.6824361065796629, |
|
"grad_norm": 2.046875, |
|
"learning_rate": 1.632196757965344e-05, |
|
"loss": 0.1347, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 0.6851549755301795, |
|
"grad_norm": 1.140625, |
|
"learning_rate": 1.6182224706539968e-05, |
|
"loss": 0.1575, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.687873844480696, |
|
"grad_norm": 1.8046875, |
|
"learning_rate": 1.6042481833426497e-05, |
|
"loss": 0.1613, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 0.6905927134312126, |
|
"grad_norm": 2.015625, |
|
"learning_rate": 1.5902738960313023e-05, |
|
"loss": 0.1625, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.6933115823817292, |
|
"grad_norm": 3.421875, |
|
"learning_rate": 1.5762996087199552e-05, |
|
"loss": 0.1748, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.6933115823817292, |
|
"eval_loss": 0.15357030928134918, |
|
"eval_runtime": 102.5541, |
|
"eval_samples_per_second": 4.875, |
|
"eval_steps_per_second": 0.156, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.6960304513322458, |
|
"grad_norm": 2.3125, |
|
"learning_rate": 1.562325321408608e-05, |
|
"loss": 0.1767, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.6987493202827624, |
|
"grad_norm": 1.734375, |
|
"learning_rate": 1.5483510340972614e-05, |
|
"loss": 0.1507, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 0.7014681892332789, |
|
"grad_norm": 2.0625, |
|
"learning_rate": 1.534376746785914e-05, |
|
"loss": 0.13, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.7041870581837956, |
|
"grad_norm": 1.3671875, |
|
"learning_rate": 1.5204024594745669e-05, |
|
"loss": 0.1576, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 0.7069059271343121, |
|
"grad_norm": 1.0546875, |
|
"learning_rate": 1.5064281721632198e-05, |
|
"loss": 0.173, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.7069059271343121, |
|
"eval_loss": 0.15370312333106995, |
|
"eval_runtime": 102.0222, |
|
"eval_samples_per_second": 4.901, |
|
"eval_steps_per_second": 0.157, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.7096247960848288, |
|
"grad_norm": 0.62890625, |
|
"learning_rate": 1.4924538848518726e-05, |
|
"loss": 0.1421, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 0.7123436650353453, |
|
"grad_norm": 1.8125, |
|
"learning_rate": 1.4784795975405255e-05, |
|
"loss": 0.1548, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.7150625339858618, |
|
"grad_norm": 0.96484375, |
|
"learning_rate": 1.4645053102291784e-05, |
|
"loss": 0.1422, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 0.7177814029363785, |
|
"grad_norm": 1.8125, |
|
"learning_rate": 1.4505310229178312e-05, |
|
"loss": 0.1577, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.720500271886895, |
|
"grad_norm": 1.046875, |
|
"learning_rate": 1.4365567356064841e-05, |
|
"loss": 0.1542, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.720500271886895, |
|
"eval_loss": 0.16090625524520874, |
|
"eval_runtime": 105.0015, |
|
"eval_samples_per_second": 4.762, |
|
"eval_steps_per_second": 0.152, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.7232191408374117, |
|
"grad_norm": 3.53125, |
|
"learning_rate": 1.4225824482951369e-05, |
|
"loss": 0.1655, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.7259380097879282, |
|
"grad_norm": 0.96484375, |
|
"learning_rate": 1.4086081609837898e-05, |
|
"loss": 0.1508, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 0.7286568787384448, |
|
"grad_norm": 1.9609375, |
|
"learning_rate": 1.3946338736724427e-05, |
|
"loss": 0.1455, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.7313757476889614, |
|
"grad_norm": 2.171875, |
|
"learning_rate": 1.3806595863610958e-05, |
|
"loss": 0.1408, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 0.734094616639478, |
|
"grad_norm": 1.125, |
|
"learning_rate": 1.3666852990497486e-05, |
|
"loss": 0.1562, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.734094616639478, |
|
"eval_loss": 0.15767186880111694, |
|
"eval_runtime": 98.7176, |
|
"eval_samples_per_second": 5.065, |
|
"eval_steps_per_second": 0.162, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.7368134855899946, |
|
"grad_norm": 0.875, |
|
"learning_rate": 1.3527110117384015e-05, |
|
"loss": 0.1621, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 0.7395323545405111, |
|
"grad_norm": 1.578125, |
|
"learning_rate": 1.3387367244270544e-05, |
|
"loss": 0.1673, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 0.7422512234910277, |
|
"grad_norm": 1.1640625, |
|
"learning_rate": 1.3247624371157072e-05, |
|
"loss": 0.1708, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 0.7449700924415443, |
|
"grad_norm": 1.4765625, |
|
"learning_rate": 1.3107881498043601e-05, |
|
"loss": 0.1295, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 0.7476889613920609, |
|
"grad_norm": 1.9765625, |
|
"learning_rate": 1.2968138624930128e-05, |
|
"loss": 0.1518, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.7476889613920609, |
|
"eval_loss": 0.1566796898841858, |
|
"eval_runtime": 99.834, |
|
"eval_samples_per_second": 5.008, |
|
"eval_steps_per_second": 0.16, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.7504078303425775, |
|
"grad_norm": 3.34375, |
|
"learning_rate": 1.2828395751816658e-05, |
|
"loss": 0.2034, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.7531266992930941, |
|
"grad_norm": 1.9765625, |
|
"learning_rate": 1.2688652878703185e-05, |
|
"loss": 0.1786, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 0.7558455682436107, |
|
"grad_norm": 1.5, |
|
"learning_rate": 1.2548910005589715e-05, |
|
"loss": 0.1941, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 0.7585644371941273, |
|
"grad_norm": 2.109375, |
|
"learning_rate": 1.2409167132476244e-05, |
|
"loss": 0.1586, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 0.7612833061446438, |
|
"grad_norm": 2.59375, |
|
"learning_rate": 1.2269424259362773e-05, |
|
"loss": 0.1914, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.7612833061446438, |
|
"eval_loss": 0.14214062690734863, |
|
"eval_runtime": 99.6815, |
|
"eval_samples_per_second": 5.016, |
|
"eval_steps_per_second": 0.161, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.7640021750951604, |
|
"grad_norm": 2.6875, |
|
"learning_rate": 1.2129681386249302e-05, |
|
"loss": 0.1537, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 0.766721044045677, |
|
"grad_norm": 1.5234375, |
|
"learning_rate": 1.198993851313583e-05, |
|
"loss": 0.1528, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.7694399129961936, |
|
"grad_norm": 1.5234375, |
|
"learning_rate": 1.1850195640022359e-05, |
|
"loss": 0.1504, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 0.7721587819467102, |
|
"grad_norm": 1.875, |
|
"learning_rate": 1.1710452766908888e-05, |
|
"loss": 0.1599, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 0.7748776508972267, |
|
"grad_norm": 2.21875, |
|
"learning_rate": 1.1570709893795418e-05, |
|
"loss": 0.1833, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.7748776508972267, |
|
"eval_loss": 0.15465624630451202, |
|
"eval_runtime": 98.484, |
|
"eval_samples_per_second": 5.077, |
|
"eval_steps_per_second": 0.162, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.7775965198477434, |
|
"grad_norm": 2.796875, |
|
"learning_rate": 1.1430967020681945e-05, |
|
"loss": 0.1537, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 0.7803153887982599, |
|
"grad_norm": 2.796875, |
|
"learning_rate": 1.1291224147568474e-05, |
|
"loss": 0.1916, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 0.7830342577487766, |
|
"grad_norm": 1.875, |
|
"learning_rate": 1.1151481274455004e-05, |
|
"loss": 0.1445, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 0.7857531266992931, |
|
"grad_norm": 3.0625, |
|
"learning_rate": 1.1011738401341531e-05, |
|
"loss": 0.187, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 0.7884719956498096, |
|
"grad_norm": 1.34375, |
|
"learning_rate": 1.0871995528228062e-05, |
|
"loss": 0.1492, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.7884719956498096, |
|
"eval_loss": 0.14948438107967377, |
|
"eval_runtime": 117.0767, |
|
"eval_samples_per_second": 4.271, |
|
"eval_steps_per_second": 0.137, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.7911908646003263, |
|
"grad_norm": 1.15625, |
|
"learning_rate": 1.073225265511459e-05, |
|
"loss": 0.211, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 0.7939097335508428, |
|
"grad_norm": 1.2109375, |
|
"learning_rate": 1.0592509782001119e-05, |
|
"loss": 0.1555, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 0.7966286025013595, |
|
"grad_norm": 1.578125, |
|
"learning_rate": 1.0452766908887646e-05, |
|
"loss": 0.1666, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 0.799347471451876, |
|
"grad_norm": 1.8046875, |
|
"learning_rate": 1.0313024035774176e-05, |
|
"loss": 0.1936, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 0.8020663404023926, |
|
"grad_norm": 2.421875, |
|
"learning_rate": 1.0173281162660705e-05, |
|
"loss": 0.1484, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.8020663404023926, |
|
"eval_loss": 0.1609531193971634, |
|
"eval_runtime": 104.0405, |
|
"eval_samples_per_second": 4.806, |
|
"eval_steps_per_second": 0.154, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.8047852093529092, |
|
"grad_norm": 1.34375, |
|
"learning_rate": 1.0033538289547234e-05, |
|
"loss": 0.1312, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 0.8075040783034257, |
|
"grad_norm": 1.265625, |
|
"learning_rate": 9.893795416433763e-06, |
|
"loss": 0.1239, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 0.8102229472539424, |
|
"grad_norm": 2.265625, |
|
"learning_rate": 9.754052543320291e-06, |
|
"loss": 0.1733, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 0.8129418162044589, |
|
"grad_norm": 1.359375, |
|
"learning_rate": 9.61430967020682e-06, |
|
"loss": 0.1964, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 0.8156606851549756, |
|
"grad_norm": 1.46875, |
|
"learning_rate": 9.474566797093348e-06, |
|
"loss": 0.1463, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.8156606851549756, |
|
"eval_loss": 0.15464062988758087, |
|
"eval_runtime": 103.2751, |
|
"eval_samples_per_second": 4.841, |
|
"eval_steps_per_second": 0.155, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.8183795541054921, |
|
"grad_norm": 1.84375, |
|
"learning_rate": 9.334823923979877e-06, |
|
"loss": 0.1547, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 0.8210984230560087, |
|
"grad_norm": 1.7265625, |
|
"learning_rate": 9.195081050866406e-06, |
|
"loss": 0.1783, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 0.8238172920065253, |
|
"grad_norm": 1.609375, |
|
"learning_rate": 9.055338177752935e-06, |
|
"loss": 0.1595, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 0.8265361609570419, |
|
"grad_norm": 1.296875, |
|
"learning_rate": 8.915595304639463e-06, |
|
"loss": 0.1625, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 0.8292550299075585, |
|
"grad_norm": 1.671875, |
|
"learning_rate": 8.775852431525992e-06, |
|
"loss": 0.1821, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.8292550299075585, |
|
"eval_loss": 0.15335936844348907, |
|
"eval_runtime": 100.4729, |
|
"eval_samples_per_second": 4.976, |
|
"eval_steps_per_second": 0.159, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.831973898858075, |
|
"grad_norm": 3.0625, |
|
"learning_rate": 8.636109558412521e-06, |
|
"loss": 0.1855, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 0.8346927678085916, |
|
"grad_norm": 1.328125, |
|
"learning_rate": 8.496366685299049e-06, |
|
"loss": 0.1513, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 0.8374116367591082, |
|
"grad_norm": 1.0859375, |
|
"learning_rate": 8.35662381218558e-06, |
|
"loss": 0.1566, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 0.8401305057096248, |
|
"grad_norm": 1.90625, |
|
"learning_rate": 8.216880939072108e-06, |
|
"loss": 0.1611, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 0.8428493746601414, |
|
"grad_norm": 1.9453125, |
|
"learning_rate": 8.077138065958637e-06, |
|
"loss": 0.1432, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.8428493746601414, |
|
"eval_loss": 0.15146875381469727, |
|
"eval_runtime": 100.1544, |
|
"eval_samples_per_second": 4.992, |
|
"eval_steps_per_second": 0.16, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.845568243610658, |
|
"grad_norm": 1.8359375, |
|
"learning_rate": 7.937395192845164e-06, |
|
"loss": 0.1737, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 0.8482871125611745, |
|
"grad_norm": 1.0703125, |
|
"learning_rate": 7.797652319731694e-06, |
|
"loss": 0.1423, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 0.8510059815116912, |
|
"grad_norm": 2.046875, |
|
"learning_rate": 7.657909446618223e-06, |
|
"loss": 0.1262, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 0.8537248504622077, |
|
"grad_norm": 0.8515625, |
|
"learning_rate": 7.518166573504752e-06, |
|
"loss": 0.1402, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 0.8564437194127243, |
|
"grad_norm": 1.3125, |
|
"learning_rate": 7.3784237003912805e-06, |
|
"loss": 0.1803, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.8564437194127243, |
|
"eval_loss": 0.13735155761241913, |
|
"eval_runtime": 98.6763, |
|
"eval_samples_per_second": 5.067, |
|
"eval_steps_per_second": 0.162, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.8591625883632409, |
|
"grad_norm": 1.2265625, |
|
"learning_rate": 7.238680827277809e-06, |
|
"loss": 0.1801, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 0.8618814573137574, |
|
"grad_norm": 3.9375, |
|
"learning_rate": 7.098937954164338e-06, |
|
"loss": 0.1819, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 0.8646003262642741, |
|
"grad_norm": 1.875, |
|
"learning_rate": 6.9591950810508665e-06, |
|
"loss": 0.1458, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 0.8673191952147906, |
|
"grad_norm": 2.9375, |
|
"learning_rate": 6.819452207937395e-06, |
|
"loss": 0.168, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 0.8700380641653073, |
|
"grad_norm": 1.9609375, |
|
"learning_rate": 6.679709334823925e-06, |
|
"loss": 0.1505, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.8700380641653073, |
|
"eval_loss": 0.14777344465255737, |
|
"eval_runtime": 99.0119, |
|
"eval_samples_per_second": 5.05, |
|
"eval_steps_per_second": 0.162, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.8727569331158238, |
|
"grad_norm": 9.0, |
|
"learning_rate": 6.539966461710453e-06, |
|
"loss": 0.1582, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 0.8754758020663403, |
|
"grad_norm": 1.890625, |
|
"learning_rate": 6.400223588596982e-06, |
|
"loss": 0.176, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 0.878194671016857, |
|
"grad_norm": 1.25, |
|
"learning_rate": 6.26048071548351e-06, |
|
"loss": 0.1346, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 0.8809135399673735, |
|
"grad_norm": 1.953125, |
|
"learning_rate": 6.1207378423700394e-06, |
|
"loss": 0.1424, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 0.8836324089178902, |
|
"grad_norm": 1.8671875, |
|
"learning_rate": 5.980994969256568e-06, |
|
"loss": 0.1413, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.8836324089178902, |
|
"eval_loss": 0.14879687130451202, |
|
"eval_runtime": 95.0133, |
|
"eval_samples_per_second": 5.262, |
|
"eval_steps_per_second": 0.168, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.8863512778684067, |
|
"grad_norm": 1.8359375, |
|
"learning_rate": 5.841252096143097e-06, |
|
"loss": 0.1432, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 0.8890701468189234, |
|
"grad_norm": 2.203125, |
|
"learning_rate": 5.701509223029626e-06, |
|
"loss": 0.1352, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 0.8917890157694399, |
|
"grad_norm": 1.578125, |
|
"learning_rate": 5.561766349916155e-06, |
|
"loss": 0.1572, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 0.8945078847199565, |
|
"grad_norm": 2.078125, |
|
"learning_rate": 5.422023476802683e-06, |
|
"loss": 0.1746, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 0.8972267536704731, |
|
"grad_norm": 1.5703125, |
|
"learning_rate": 5.282280603689212e-06, |
|
"loss": 0.1812, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.8972267536704731, |
|
"eval_loss": 0.1419062465429306, |
|
"eval_runtime": 95.5103, |
|
"eval_samples_per_second": 5.235, |
|
"eval_steps_per_second": 0.168, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.8999456226209896, |
|
"grad_norm": 1.2265625, |
|
"learning_rate": 5.142537730575741e-06, |
|
"loss": 0.1398, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 0.9026644915715063, |
|
"grad_norm": 2.28125, |
|
"learning_rate": 5.002794857462269e-06, |
|
"loss": 0.143, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 0.9053833605220228, |
|
"grad_norm": 2.34375, |
|
"learning_rate": 4.863051984348798e-06, |
|
"loss": 0.1623, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 0.9081022294725394, |
|
"grad_norm": 3.21875, |
|
"learning_rate": 4.723309111235328e-06, |
|
"loss": 0.1893, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 0.910821098423056, |
|
"grad_norm": 1.5703125, |
|
"learning_rate": 4.583566238121856e-06, |
|
"loss": 0.1617, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.910821098423056, |
|
"eval_loss": 0.15714062750339508, |
|
"eval_runtime": 94.6634, |
|
"eval_samples_per_second": 5.282, |
|
"eval_steps_per_second": 0.169, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.9135399673735726, |
|
"grad_norm": 1.7734375, |
|
"learning_rate": 4.443823365008385e-06, |
|
"loss": 0.1508, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 0.9162588363240892, |
|
"grad_norm": 5.09375, |
|
"learning_rate": 4.304080491894914e-06, |
|
"loss": 0.1355, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 0.9189777052746058, |
|
"grad_norm": 2.09375, |
|
"learning_rate": 4.164337618781442e-06, |
|
"loss": 0.1457, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 0.9216965742251223, |
|
"grad_norm": 1.296875, |
|
"learning_rate": 4.024594745667971e-06, |
|
"loss": 0.1391, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 0.924415443175639, |
|
"grad_norm": 2.046875, |
|
"learning_rate": 3.8848518725545e-06, |
|
"loss": 0.1853, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.924415443175639, |
|
"eval_loss": 0.14689843356609344, |
|
"eval_runtime": 109.943, |
|
"eval_samples_per_second": 4.548, |
|
"eval_steps_per_second": 0.146, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.9271343121261555, |
|
"grad_norm": 2.265625, |
|
"learning_rate": 3.7451089994410285e-06, |
|
"loss": 0.1904, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 0.9298531810766721, |
|
"grad_norm": 1.34375, |
|
"learning_rate": 3.6053661263275578e-06, |
|
"loss": 0.1342, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 0.9325720500271887, |
|
"grad_norm": 2.15625, |
|
"learning_rate": 3.465623253214086e-06, |
|
"loss": 0.1583, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 0.9352909189777052, |
|
"grad_norm": 1.859375, |
|
"learning_rate": 3.325880380100615e-06, |
|
"loss": 0.1763, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 0.9380097879282219, |
|
"grad_norm": 1.703125, |
|
"learning_rate": 3.1861375069871442e-06, |
|
"loss": 0.1367, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.9380097879282219, |
|
"eval_loss": 0.1392187476158142, |
|
"eval_runtime": 96.262, |
|
"eval_samples_per_second": 5.194, |
|
"eval_steps_per_second": 0.166, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.9407286568787384, |
|
"grad_norm": 1.15625, |
|
"learning_rate": 3.0463946338736726e-06, |
|
"loss": 0.184, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 0.9434475258292551, |
|
"grad_norm": 1.2734375, |
|
"learning_rate": 2.9066517607602015e-06, |
|
"loss": 0.1794, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 0.9461663947797716, |
|
"grad_norm": 2.9375, |
|
"learning_rate": 2.7669088876467303e-06, |
|
"loss": 0.1705, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 0.9488852637302883, |
|
"grad_norm": 1.0078125, |
|
"learning_rate": 2.627166014533259e-06, |
|
"loss": 0.1846, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 0.9516041326808048, |
|
"grad_norm": 0.97265625, |
|
"learning_rate": 2.487423141419788e-06, |
|
"loss": 0.1917, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.9516041326808048, |
|
"eval_loss": 0.14869531989097595, |
|
"eval_runtime": 96.4372, |
|
"eval_samples_per_second": 5.185, |
|
"eval_steps_per_second": 0.166, |
|
"step": 3500 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 3678, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2.05068095973888e+18, |
|
"train_batch_size": 32, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|