{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 100, "global_step": 1626, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0018450184501845018, "grad_norm": 9.194052941983164, "learning_rate": 1.226993865030675e-07, "loss": 1.1392, "step": 1 }, { "epoch": 0.00922509225092251, "grad_norm": 8.728469464225432, "learning_rate": 6.134969325153375e-07, "loss": 1.1321, "step": 5 }, { "epoch": 0.01845018450184502, "grad_norm": 5.066035045474869, "learning_rate": 1.226993865030675e-06, "loss": 1.0802, "step": 10 }, { "epoch": 0.027675276752767528, "grad_norm": 6.151048691792626, "learning_rate": 1.8404907975460124e-06, "loss": 1.0186, "step": 15 }, { "epoch": 0.03690036900369004, "grad_norm": 2.030218046940431, "learning_rate": 2.45398773006135e-06, "loss": 1.0181, "step": 20 }, { "epoch": 0.046125461254612546, "grad_norm": 1.7169054577646434, "learning_rate": 3.0674846625766875e-06, "loss": 0.9867, "step": 25 }, { "epoch": 0.055350553505535055, "grad_norm": 1.414702551784086, "learning_rate": 3.680981595092025e-06, "loss": 0.9848, "step": 30 }, { "epoch": 0.06457564575645756, "grad_norm": 1.471062511929668, "learning_rate": 4.294478527607362e-06, "loss": 0.975, "step": 35 }, { "epoch": 0.07380073800738007, "grad_norm": 1.9876641303020315, "learning_rate": 4.9079754601227e-06, "loss": 0.9616, "step": 40 }, { "epoch": 0.08302583025830258, "grad_norm": 1.7086605102377759, "learning_rate": 5.521472392638038e-06, "loss": 0.9716, "step": 45 }, { "epoch": 0.09225092250922509, "grad_norm": 2.202769359683669, "learning_rate": 6.134969325153375e-06, "loss": 0.9766, "step": 50 }, { "epoch": 0.1014760147601476, "grad_norm": 1.6222357117334487, "learning_rate": 6.748466257668712e-06, "loss": 0.9929, "step": 55 }, { "epoch": 0.11070110701107011, "grad_norm": 2.161648398755977, "learning_rate": 7.36196319018405e-06, "loss": 0.9774, "step": 60 }, { "epoch": 0.11992619926199262, "grad_norm": 1.7198404521131392, "learning_rate": 7.975460122699386e-06, "loss": 0.9743, "step": 65 }, { "epoch": 0.12915129151291513, "grad_norm": 2.5936580446065594, "learning_rate": 8.588957055214725e-06, "loss": 0.9878, "step": 70 }, { "epoch": 0.13837638376383765, "grad_norm": 2.188257188915145, "learning_rate": 9.202453987730062e-06, "loss": 0.9568, "step": 75 }, { "epoch": 0.14760147601476015, "grad_norm": 1.7531151641523148, "learning_rate": 9.8159509202454e-06, "loss": 0.9789, "step": 80 }, { "epoch": 0.15682656826568267, "grad_norm": 1.8091240872427208, "learning_rate": 1.0429447852760737e-05, "loss": 0.9678, "step": 85 }, { "epoch": 0.16605166051660517, "grad_norm": 2.098514635540621, "learning_rate": 1.1042944785276076e-05, "loss": 0.9617, "step": 90 }, { "epoch": 0.1752767527675277, "grad_norm": 2.4275494428488607, "learning_rate": 1.1656441717791411e-05, "loss": 0.9676, "step": 95 }, { "epoch": 0.18450184501845018, "grad_norm": 2.0637923302738095, "learning_rate": 1.226993865030675e-05, "loss": 0.9681, "step": 100 }, { "epoch": 0.18450184501845018, "eval_loss": 0.9788174629211426, "eval_runtime": 515.1712, "eval_samples_per_second": 29.796, "eval_steps_per_second": 0.116, "step": 100 }, { "epoch": 0.1937269372693727, "grad_norm": 2.069416549180579, "learning_rate": 1.2883435582822085e-05, "loss": 0.9528, "step": 105 }, { "epoch": 0.2029520295202952, "grad_norm": 2.2916715973700024, "learning_rate": 1.3496932515337424e-05, "loss": 0.9696, "step": 110 }, { "epoch": 0.21217712177121772, "grad_norm": 2.062468142825091, "learning_rate": 1.4110429447852763e-05, "loss": 0.9747, "step": 115 }, { "epoch": 0.22140221402214022, "grad_norm": 1.7271367882138293, "learning_rate": 1.47239263803681e-05, "loss": 0.9786, "step": 120 }, { "epoch": 0.23062730627306274, "grad_norm": 1.9545058702706481, "learning_rate": 1.5337423312883436e-05, "loss": 0.9758, "step": 125 }, { "epoch": 0.23985239852398524, "grad_norm": 1.9400595646067775, "learning_rate": 1.5950920245398772e-05, "loss": 0.9829, "step": 130 }, { "epoch": 0.24907749077490776, "grad_norm": 1.865861850010034, "learning_rate": 1.656441717791411e-05, "loss": 0.9915, "step": 135 }, { "epoch": 0.25830258302583026, "grad_norm": 1.9529698824708406, "learning_rate": 1.717791411042945e-05, "loss": 0.9831, "step": 140 }, { "epoch": 0.26752767527675275, "grad_norm": 1.8749039852563243, "learning_rate": 1.7791411042944788e-05, "loss": 0.9842, "step": 145 }, { "epoch": 0.2767527675276753, "grad_norm": 1.4867806820095497, "learning_rate": 1.8404907975460123e-05, "loss": 0.9859, "step": 150 }, { "epoch": 0.2859778597785978, "grad_norm": 2.1169911338934644, "learning_rate": 1.9018404907975462e-05, "loss": 0.9771, "step": 155 }, { "epoch": 0.2952029520295203, "grad_norm": 1.5398155481235816, "learning_rate": 1.96319018404908e-05, "loss": 0.9817, "step": 160 }, { "epoch": 0.3044280442804428, "grad_norm": 1.7130250807487832, "learning_rate": 1.9999907776750355e-05, "loss": 0.9997, "step": 165 }, { "epoch": 0.31365313653136534, "grad_norm": 2.1366171045520383, "learning_rate": 1.9998870284726968e-05, "loss": 1.0004, "step": 170 }, { "epoch": 0.32287822878228783, "grad_norm": 2.01400597362679, "learning_rate": 1.9996680141616956e-05, "loss": 0.9937, "step": 175 }, { "epoch": 0.33210332103321033, "grad_norm": 1.925569839756876, "learning_rate": 1.9993337599895925e-05, "loss": 0.9939, "step": 180 }, { "epoch": 0.3413284132841328, "grad_norm": 1.4590251035585917, "learning_rate": 1.998884304488584e-05, "loss": 0.9982, "step": 185 }, { "epoch": 0.3505535055350554, "grad_norm": 1.531094729781709, "learning_rate": 1.998319699471061e-05, "loss": 0.9925, "step": 190 }, { "epoch": 0.35977859778597787, "grad_norm": 1.9624667257758441, "learning_rate": 1.997640010023634e-05, "loss": 0.9765, "step": 195 }, { "epoch": 0.36900369003690037, "grad_norm": 1.6943282928766075, "learning_rate": 1.9968453144996345e-05, "loss": 0.9962, "step": 200 }, { "epoch": 0.36900369003690037, "eval_loss": 1.0030262470245361, "eval_runtime": 518.1283, "eval_samples_per_second": 29.626, "eval_steps_per_second": 0.116, "step": 200 }, { "epoch": 0.37822878228782286, "grad_norm": 3.0469311990676857, "learning_rate": 1.9959357045100764e-05, "loss": 0.9947, "step": 205 }, { "epoch": 0.3874538745387454, "grad_norm": 2.2551326104892864, "learning_rate": 1.9949112849131005e-05, "loss": 1.0023, "step": 210 }, { "epoch": 0.3966789667896679, "grad_norm": 1.8771683509279502, "learning_rate": 1.993772173801884e-05, "loss": 0.9934, "step": 215 }, { "epoch": 0.4059040590405904, "grad_norm": 1.8016877222967922, "learning_rate": 1.992518502491028e-05, "loss": 0.9807, "step": 220 }, { "epoch": 0.4151291512915129, "grad_norm": 1.4456497466009737, "learning_rate": 1.9911504155014187e-05, "loss": 0.9926, "step": 225 }, { "epoch": 0.42435424354243545, "grad_norm": 1.5156073716841811, "learning_rate": 1.989668070543569e-05, "loss": 0.9766, "step": 230 }, { "epoch": 0.43357933579335795, "grad_norm": 1.3959824735787207, "learning_rate": 1.9880716384994355e-05, "loss": 0.9964, "step": 235 }, { "epoch": 0.44280442804428044, "grad_norm": 1.4724192694561282, "learning_rate": 1.9863613034027224e-05, "loss": 0.9942, "step": 240 }, { "epoch": 0.45202952029520294, "grad_norm": 2.064409139190994, "learning_rate": 1.9845372624176646e-05, "loss": 1.0103, "step": 245 }, { "epoch": 0.4612546125461255, "grad_norm": 2.190902421105104, "learning_rate": 1.982599725816299e-05, "loss": 1.0075, "step": 250 }, { "epoch": 0.470479704797048, "grad_norm": 1.9443583169417478, "learning_rate": 1.9805489169542245e-05, "loss": 0.9971, "step": 255 }, { "epoch": 0.4797047970479705, "grad_norm": 1.553791831408308, "learning_rate": 1.978385072244857e-05, "loss": 0.9992, "step": 260 }, { "epoch": 0.488929889298893, "grad_norm": 1.4174068635451451, "learning_rate": 1.9761084411321706e-05, "loss": 0.9793, "step": 265 }, { "epoch": 0.4981549815498155, "grad_norm": 1.4969414214930414, "learning_rate": 1.9737192860619477e-05, "loss": 0.9791, "step": 270 }, { "epoch": 0.507380073800738, "grad_norm": 1.4025421975340602, "learning_rate": 1.971217882451521e-05, "loss": 0.9796, "step": 275 }, { "epoch": 0.5166051660516605, "grad_norm": 1.4448369862138994, "learning_rate": 1.9686045186580258e-05, "loss": 0.9884, "step": 280 }, { "epoch": 0.525830258302583, "grad_norm": 2.0639483249182464, "learning_rate": 1.9658794959451583e-05, "loss": 0.9831, "step": 285 }, { "epoch": 0.5350553505535055, "grad_norm": 1.6048970102781592, "learning_rate": 1.9630431284484447e-05, "loss": 0.9849, "step": 290 }, { "epoch": 0.544280442804428, "grad_norm": 1.4540480684938577, "learning_rate": 1.960095743139033e-05, "loss": 0.9796, "step": 295 }, { "epoch": 0.5535055350553506, "grad_norm": 1.424947900669971, "learning_rate": 1.957037679785994e-05, "loss": 0.9917, "step": 300 }, { "epoch": 0.5535055350553506, "eval_loss": 1.0008341073989868, "eval_runtime": 513.1068, "eval_samples_per_second": 29.916, "eval_steps_per_second": 0.117, "step": 300 }, { "epoch": 0.5627306273062731, "grad_norm": 1.2480517696242786, "learning_rate": 1.953869290917158e-05, "loss": 0.9943, "step": 305 }, { "epoch": 0.5719557195571956, "grad_norm": 1.191133450390735, "learning_rate": 1.9505909417784758e-05, "loss": 0.9899, "step": 310 }, { "epoch": 0.5811808118081181, "grad_norm": 1.1766418475997753, "learning_rate": 1.9472030102919102e-05, "loss": 0.9883, "step": 315 }, { "epoch": 0.5904059040590406, "grad_norm": 1.2121897211885717, "learning_rate": 1.9437058870118745e-05, "loss": 1.0037, "step": 320 }, { "epoch": 0.5996309963099631, "grad_norm": 1.2903187102851559, "learning_rate": 1.940099975080207e-05, "loss": 0.9892, "step": 325 }, { "epoch": 0.6088560885608856, "grad_norm": 1.4260318993897811, "learning_rate": 1.9363856901796984e-05, "loss": 0.9896, "step": 330 }, { "epoch": 0.6180811808118081, "grad_norm": 1.324489901337969, "learning_rate": 1.9325634604861728e-05, "loss": 0.9978, "step": 335 }, { "epoch": 0.6273062730627307, "grad_norm": 1.275426852454915, "learning_rate": 1.9286337266191295e-05, "loss": 0.993, "step": 340 }, { "epoch": 0.6365313653136532, "grad_norm": 1.329213272796139, "learning_rate": 1.9245969415909464e-05, "loss": 0.9879, "step": 345 }, { "epoch": 0.6457564575645757, "grad_norm": 1.4085398606096227, "learning_rate": 1.9204535707546602e-05, "loss": 0.9869, "step": 350 }, { "epoch": 0.6549815498154982, "grad_norm": 1.1848936755869721, "learning_rate": 1.916204091750321e-05, "loss": 0.9726, "step": 355 }, { "epoch": 0.6642066420664207, "grad_norm": 1.2968309154541056, "learning_rate": 1.9118489944499287e-05, "loss": 0.9902, "step": 360 }, { "epoch": 0.6734317343173432, "grad_norm": 1.2286246913756114, "learning_rate": 1.907388780900964e-05, "loss": 0.9811, "step": 365 }, { "epoch": 0.6826568265682657, "grad_norm": 1.2591567733071325, "learning_rate": 1.902823965268513e-05, "loss": 0.9858, "step": 370 }, { "epoch": 0.6918819188191881, "grad_norm": 1.4378514619406175, "learning_rate": 1.8981550737759932e-05, "loss": 0.9828, "step": 375 }, { "epoch": 0.7011070110701108, "grad_norm": 1.497308547977116, "learning_rate": 1.8933826446444933e-05, "loss": 0.9892, "step": 380 }, { "epoch": 0.7103321033210332, "grad_norm": 1.1745393096620436, "learning_rate": 1.888507228030729e-05, "loss": 0.9859, "step": 385 }, { "epoch": 0.7195571955719557, "grad_norm": 1.2233160586824499, "learning_rate": 1.8835293859636177e-05, "loss": 0.9763, "step": 390 }, { "epoch": 0.7287822878228782, "grad_norm": 1.3127902536541989, "learning_rate": 1.8784496922794947e-05, "loss": 0.981, "step": 395 }, { "epoch": 0.7380073800738007, "grad_norm": 1.3089866347753676, "learning_rate": 1.873268732555957e-05, "loss": 0.9652, "step": 400 }, { "epoch": 0.7380073800738007, "eval_loss": 0.993894636631012, "eval_runtime": 513.7147, "eval_samples_per_second": 29.88, "eval_steps_per_second": 0.117, "step": 400 }, { "epoch": 0.7472324723247232, "grad_norm": 1.3088189271034285, "learning_rate": 1.8679871040443632e-05, "loss": 1.0048, "step": 405 }, { "epoch": 0.7564575645756457, "grad_norm": 1.2954577066196238, "learning_rate": 1.8626054156009807e-05, "loss": 0.9927, "step": 410 }, { "epoch": 0.7656826568265682, "grad_norm": 1.317981053662398, "learning_rate": 1.8571242876167995e-05, "loss": 0.9752, "step": 415 }, { "epoch": 0.7749077490774908, "grad_norm": 1.4156756831610378, "learning_rate": 1.851544351946014e-05, "loss": 0.9945, "step": 420 }, { "epoch": 0.7841328413284133, "grad_norm": 1.1285773664771428, "learning_rate": 1.845866251833183e-05, "loss": 0.9708, "step": 425 }, { "epoch": 0.7933579335793358, "grad_norm": 1.2640468813011223, "learning_rate": 1.8400906418390808e-05, "loss": 0.9757, "step": 430 }, { "epoch": 0.8025830258302583, "grad_norm": 1.288546177133416, "learning_rate": 1.834218187765237e-05, "loss": 0.976, "step": 435 }, { "epoch": 0.8118081180811808, "grad_norm": 1.3086160465192265, "learning_rate": 1.8282495665771864e-05, "loss": 0.9761, "step": 440 }, { "epoch": 0.8210332103321033, "grad_norm": 1.1919282548241303, "learning_rate": 1.8221854663264294e-05, "loss": 0.9718, "step": 445 }, { "epoch": 0.8302583025830258, "grad_norm": 1.2454331164701038, "learning_rate": 1.8160265860711134e-05, "loss": 0.9842, "step": 450 }, { "epoch": 0.8394833948339483, "grad_norm": 1.183454477783249, "learning_rate": 1.8097736357954487e-05, "loss": 0.9705, "step": 455 }, { "epoch": 0.8487084870848709, "grad_norm": 1.1394535207411802, "learning_rate": 1.8034273363278615e-05, "loss": 0.9751, "step": 460 }, { "epoch": 0.8579335793357934, "grad_norm": 1.1866949984179949, "learning_rate": 1.7969884192578977e-05, "loss": 0.9749, "step": 465 }, { "epoch": 0.8671586715867159, "grad_norm": 1.299660479182102, "learning_rate": 1.7904576268518886e-05, "loss": 0.9598, "step": 470 }, { "epoch": 0.8763837638376384, "grad_norm": 1.2221383874437446, "learning_rate": 1.783835711967382e-05, "loss": 0.9842, "step": 475 }, { "epoch": 0.8856088560885609, "grad_norm": 1.2535423952991984, "learning_rate": 1.7771234379663545e-05, "loss": 0.9641, "step": 480 }, { "epoch": 0.8948339483394834, "grad_norm": 1.4654400132426395, "learning_rate": 1.770321578627213e-05, "loss": 0.9784, "step": 485 }, { "epoch": 0.9040590405904059, "grad_norm": 1.3747052246285973, "learning_rate": 1.763430918055595e-05, "loss": 0.9694, "step": 490 }, { "epoch": 0.9132841328413284, "grad_norm": 1.1551950486505687, "learning_rate": 1.756452250593979e-05, "loss": 0.9727, "step": 495 }, { "epoch": 0.922509225092251, "grad_norm": 1.128236535385729, "learning_rate": 1.7493863807301116e-05, "loss": 0.9666, "step": 500 }, { "epoch": 0.922509225092251, "eval_loss": 0.9816026091575623, "eval_runtime": 517.2137, "eval_samples_per_second": 29.678, "eval_steps_per_second": 0.116, "step": 500 }, { "epoch": 0.9317343173431735, "grad_norm": 1.230218009681161, "learning_rate": 1.74223412300427e-05, "loss": 0.9769, "step": 505 }, { "epoch": 0.940959409594096, "grad_norm": 1.1847589898088133, "learning_rate": 1.7349963019153638e-05, "loss": 0.9628, "step": 510 }, { "epoch": 0.9501845018450185, "grad_norm": 1.2246308831747907, "learning_rate": 1.7276737518258865e-05, "loss": 0.9602, "step": 515 }, { "epoch": 0.959409594095941, "grad_norm": 1.1390750572317663, "learning_rate": 1.7202673168657318e-05, "loss": 0.9627, "step": 520 }, { "epoch": 0.9686346863468634, "grad_norm": 1.1728205351456946, "learning_rate": 1.7127778508348858e-05, "loss": 0.9714, "step": 525 }, { "epoch": 0.977859778597786, "grad_norm": 1.2796699310011739, "learning_rate": 1.7052062171050008e-05, "loss": 0.967, "step": 530 }, { "epoch": 0.9870848708487084, "grad_norm": 1.1205342517216532, "learning_rate": 1.6975532885198678e-05, "loss": 0.9663, "step": 535 }, { "epoch": 0.996309963099631, "grad_norm": 1.185279277131673, "learning_rate": 1.6898199472947972e-05, "loss": 0.9581, "step": 540 }, { "epoch": 1.0055350553505535, "grad_norm": 3.007398366081561, "learning_rate": 1.6820070849149174e-05, "loss": 0.8519, "step": 545 }, { "epoch": 1.014760147601476, "grad_norm": 2.1038299784593337, "learning_rate": 1.6741156020324086e-05, "loss": 0.7509, "step": 550 }, { "epoch": 1.0239852398523985, "grad_norm": 1.5701183943228265, "learning_rate": 1.6661464083626734e-05, "loss": 0.7453, "step": 555 }, { "epoch": 1.033210332103321, "grad_norm": 1.2911074026361753, "learning_rate": 1.6581004225794715e-05, "loss": 0.7391, "step": 560 }, { "epoch": 1.0424354243542435, "grad_norm": 1.5938907876285198, "learning_rate": 1.649978572209012e-05, "loss": 0.7347, "step": 565 }, { "epoch": 1.051660516605166, "grad_norm": 1.3495506131008623, "learning_rate": 1.6417817935230318e-05, "loss": 0.7396, "step": 570 }, { "epoch": 1.0608856088560885, "grad_norm": 1.2781771587882627, "learning_rate": 1.6335110314308654e-05, "loss": 0.7305, "step": 575 }, { "epoch": 1.070110701107011, "grad_norm": 1.5798733908227265, "learning_rate": 1.6251672393705155e-05, "loss": 0.7365, "step": 580 }, { "epoch": 1.0793357933579335, "grad_norm": 1.416304183876239, "learning_rate": 1.6167513791987423e-05, "loss": 0.7373, "step": 585 }, { "epoch": 1.088560885608856, "grad_norm": 1.3677150489575043, "learning_rate": 1.6082644210801846e-05, "loss": 0.7299, "step": 590 }, { "epoch": 1.0977859778597785, "grad_norm": 1.3506677105351055, "learning_rate": 1.5997073433755187e-05, "loss": 0.7426, "step": 595 }, { "epoch": 1.1070110701107012, "grad_norm": 1.461155474048458, "learning_rate": 1.5910811325286768e-05, "loss": 0.7366, "step": 600 }, { "epoch": 1.1070110701107012, "eval_loss": 0.9852360486984253, "eval_runtime": 516.2338, "eval_samples_per_second": 29.735, "eval_steps_per_second": 0.116, "step": 600 }, { "epoch": 1.1162361623616237, "grad_norm": 1.2999195127889172, "learning_rate": 1.582386782953129e-05, "loss": 0.7351, "step": 605 }, { "epoch": 1.1254612546125462, "grad_norm": 1.5599221554130673, "learning_rate": 1.5736252969172522e-05, "loss": 0.7335, "step": 610 }, { "epoch": 1.1346863468634687, "grad_norm": 1.30824219510555, "learning_rate": 1.5647976844287884e-05, "loss": 0.7321, "step": 615 }, { "epoch": 1.1439114391143912, "grad_norm": 1.3590431139669035, "learning_rate": 1.5559049631184136e-05, "loss": 0.7294, "step": 620 }, { "epoch": 1.1531365313653137, "grad_norm": 1.5685872513743657, "learning_rate": 1.5469481581224274e-05, "loss": 0.7372, "step": 625 }, { "epoch": 1.1623616236162362, "grad_norm": 1.4194329169102744, "learning_rate": 1.5379283019645757e-05, "loss": 0.7423, "step": 630 }, { "epoch": 1.1715867158671587, "grad_norm": 1.8516238628155155, "learning_rate": 1.5288464344370267e-05, "loss": 0.7389, "step": 635 }, { "epoch": 1.1808118081180812, "grad_norm": 1.3787465939384576, "learning_rate": 1.5197036024805018e-05, "loss": 0.7277, "step": 640 }, { "epoch": 1.1900369003690037, "grad_norm": 1.2679935699299498, "learning_rate": 1.5105008600635888e-05, "loss": 0.7251, "step": 645 }, { "epoch": 1.1992619926199262, "grad_norm": 1.3661565990701046, "learning_rate": 1.5012392680612408e-05, "loss": 0.7348, "step": 650 }, { "epoch": 1.2084870848708487, "grad_norm": 1.380476117633752, "learning_rate": 1.4919198941324813e-05, "loss": 0.733, "step": 655 }, { "epoch": 1.2177121771217712, "grad_norm": 1.301175007422796, "learning_rate": 1.4825438125973263e-05, "loss": 0.7331, "step": 660 }, { "epoch": 1.2269372693726937, "grad_norm": 1.3531205842843421, "learning_rate": 1.4731121043129392e-05, "loss": 0.7379, "step": 665 }, { "epoch": 1.2361623616236161, "grad_norm": 1.444864127952419, "learning_rate": 1.4636258565490304e-05, "loss": 0.739, "step": 670 }, { "epoch": 1.2453874538745389, "grad_norm": 1.2863648775710423, "learning_rate": 1.4540861628625207e-05, "loss": 0.7368, "step": 675 }, { "epoch": 1.2546125461254611, "grad_norm": 1.2200332099647682, "learning_rate": 1.444494122971476e-05, "loss": 0.7343, "step": 680 }, { "epoch": 1.2638376383763839, "grad_norm": 1.3714375121406106, "learning_rate": 1.4348508426283342e-05, "loss": 0.7391, "step": 685 }, { "epoch": 1.2730627306273063, "grad_norm": 1.2638691361743832, "learning_rate": 1.4251574334924395e-05, "loss": 0.7397, "step": 690 }, { "epoch": 1.2822878228782288, "grad_norm": 1.4011111864399106, "learning_rate": 1.4154150130018867e-05, "loss": 0.7374, "step": 695 }, { "epoch": 1.2915129151291513, "grad_norm": 1.2912923761278596, "learning_rate": 1.4056247042447096e-05, "loss": 0.7228, "step": 700 }, { "epoch": 1.2915129151291513, "eval_loss": 0.9835454225540161, "eval_runtime": 517.9285, "eval_samples_per_second": 29.637, "eval_steps_per_second": 0.116, "step": 700 }, { "epoch": 1.3007380073800738, "grad_norm": 1.5854901671726367, "learning_rate": 1.3957876358294115e-05, "loss": 0.7296, "step": 705 }, { "epoch": 1.3099630996309963, "grad_norm": 1.38846996136312, "learning_rate": 1.385904941754862e-05, "loss": 0.7257, "step": 710 }, { "epoch": 1.3191881918819188, "grad_norm": 1.5297133474564781, "learning_rate": 1.375977761279571e-05, "loss": 0.7352, "step": 715 }, { "epoch": 1.3284132841328413, "grad_norm": 1.287259224142701, "learning_rate": 1.366007238790358e-05, "loss": 0.7301, "step": 720 }, { "epoch": 1.3376383763837638, "grad_norm": 1.2884194224179173, "learning_rate": 1.3559945236704286e-05, "loss": 0.7383, "step": 725 }, { "epoch": 1.3468634686346863, "grad_norm": 1.3779553004575515, "learning_rate": 1.3459407701668762e-05, "loss": 0.7313, "step": 730 }, { "epoch": 1.3560885608856088, "grad_norm": 1.5349656095564503, "learning_rate": 1.3358471372576229e-05, "loss": 0.7334, "step": 735 }, { "epoch": 1.3653136531365313, "grad_norm": 1.3570612666553503, "learning_rate": 1.3257147885178125e-05, "loss": 0.7253, "step": 740 }, { "epoch": 1.3745387453874538, "grad_norm": 1.3514442377769267, "learning_rate": 1.3155448919856792e-05, "loss": 0.7375, "step": 745 }, { "epoch": 1.3837638376383765, "grad_norm": 1.338752928401098, "learning_rate": 1.3053386200278963e-05, "loss": 0.7349, "step": 750 }, { "epoch": 1.3929889298892988, "grad_norm": 1.3943704063449442, "learning_rate": 1.2950971492044272e-05, "loss": 0.7338, "step": 755 }, { "epoch": 1.4022140221402215, "grad_norm": 1.3567491078204894, "learning_rate": 1.2848216601328958e-05, "loss": 0.7385, "step": 760 }, { "epoch": 1.4114391143911438, "grad_norm": 1.2556919848553412, "learning_rate": 1.2745133373524855e-05, "loss": 0.7457, "step": 765 }, { "epoch": 1.4206642066420665, "grad_norm": 1.3027608934231716, "learning_rate": 1.2641733691873884e-05, "loss": 0.7342, "step": 770 }, { "epoch": 1.429889298892989, "grad_norm": 1.2668132369825373, "learning_rate": 1.2538029476098175e-05, "loss": 0.7317, "step": 775 }, { "epoch": 1.4391143911439115, "grad_norm": 1.2498842281077402, "learning_rate": 1.2434032681025986e-05, "loss": 0.732, "step": 780 }, { "epoch": 1.448339483394834, "grad_norm": 1.221148464370588, "learning_rate": 1.2329755295213568e-05, "loss": 0.7168, "step": 785 }, { "epoch": 1.4575645756457565, "grad_norm": 1.2029873246463332, "learning_rate": 1.2225209339563144e-05, "loss": 0.7299, "step": 790 }, { "epoch": 1.466789667896679, "grad_norm": 1.2769506053242343, "learning_rate": 1.2120406865937174e-05, "loss": 0.7385, "step": 795 }, { "epoch": 1.4760147601476015, "grad_norm": 1.5254063393209267, "learning_rate": 1.2015359955769021e-05, "loss": 0.7319, "step": 800 }, { "epoch": 1.4760147601476015, "eval_loss": 0.9644125699996948, "eval_runtime": 512.8317, "eval_samples_per_second": 29.932, "eval_steps_per_second": 0.117, "step": 800 }, { "epoch": 1.485239852398524, "grad_norm": 1.4657220418578245, "learning_rate": 1.1910080718670246e-05, "loss": 0.7234, "step": 805 }, { "epoch": 1.4944649446494465, "grad_norm": 1.3333083489866098, "learning_rate": 1.1804581291034615e-05, "loss": 0.7314, "step": 810 }, { "epoch": 1.503690036900369, "grad_norm": 1.3111534531304956, "learning_rate": 1.169887383463906e-05, "loss": 0.7212, "step": 815 }, { "epoch": 1.5129151291512914, "grad_norm": 1.2536260067392955, "learning_rate": 1.1592970535241668e-05, "loss": 0.723, "step": 820 }, { "epoch": 1.5221402214022142, "grad_norm": 1.239943596383526, "learning_rate": 1.1486883601176944e-05, "loss": 0.7315, "step": 825 }, { "epoch": 1.5313653136531364, "grad_norm": 1.188861248391431, "learning_rate": 1.1380625261948458e-05, "loss": 0.7301, "step": 830 }, { "epoch": 1.5405904059040592, "grad_norm": 1.247650108627454, "learning_rate": 1.127420776681905e-05, "loss": 0.7202, "step": 835 }, { "epoch": 1.5498154981549814, "grad_norm": 1.4048683840262912, "learning_rate": 1.1167643383398746e-05, "loss": 0.7247, "step": 840 }, { "epoch": 1.5590405904059041, "grad_norm": 1.2897015340446114, "learning_rate": 1.1060944396230583e-05, "loss": 0.7311, "step": 845 }, { "epoch": 1.5682656826568264, "grad_norm": 1.21939417183643, "learning_rate": 1.0954123105374468e-05, "loss": 0.7249, "step": 850 }, { "epoch": 1.5774907749077491, "grad_norm": 1.2309319468475195, "learning_rate": 1.0847191824989252e-05, "loss": 0.7298, "step": 855 }, { "epoch": 1.5867158671586716, "grad_norm": 1.2218109998078897, "learning_rate": 1.0740162881913165e-05, "loss": 0.7223, "step": 860 }, { "epoch": 1.5959409594095941, "grad_norm": 1.4183791452745522, "learning_rate": 1.0633048614242817e-05, "loss": 0.7359, "step": 865 }, { "epoch": 1.6051660516605166, "grad_norm": 1.2210289040303786, "learning_rate": 1.0525861369910877e-05, "loss": 0.7302, "step": 870 }, { "epoch": 1.6143911439114391, "grad_norm": 1.3175608261808258, "learning_rate": 1.0418613505262623e-05, "loss": 0.7226, "step": 875 }, { "epoch": 1.6236162361623616, "grad_norm": 1.3018239201611663, "learning_rate": 1.0311317383631532e-05, "loss": 0.7227, "step": 880 }, { "epoch": 1.632841328413284, "grad_norm": 1.1647552351758403, "learning_rate": 1.0203985373914056e-05, "loss": 0.7204, "step": 885 }, { "epoch": 1.6420664206642066, "grad_norm": 1.210717925144679, "learning_rate": 1.0096629849143757e-05, "loss": 0.7115, "step": 890 }, { "epoch": 1.651291512915129, "grad_norm": 1.1959081633999162, "learning_rate": 9.989263185064974e-06, "loss": 0.7164, "step": 895 }, { "epoch": 1.6605166051660518, "grad_norm": 1.1679984043624778, "learning_rate": 9.881897758706155e-06, "loss": 0.7177, "step": 900 }, { "epoch": 1.6605166051660518, "eval_loss": 0.9529369473457336, "eval_runtime": 516.4151, "eval_samples_per_second": 29.724, "eval_steps_per_second": 0.116, "step": 900 }, { "epoch": 1.669741697416974, "grad_norm": 1.1784785526719634, "learning_rate": 9.77454594695308e-06, "loss": 0.7274, "step": 905 }, { "epoch": 1.6789667896678968, "grad_norm": 1.1964871209199903, "learning_rate": 9.667220125122044e-06, "loss": 0.7119, "step": 910 }, { "epoch": 1.688191881918819, "grad_norm": 1.173031357661576, "learning_rate": 9.559932665533291e-06, "loss": 0.7134, "step": 915 }, { "epoch": 1.6974169741697418, "grad_norm": 1.2312863536042935, "learning_rate": 9.452695936084728e-06, "loss": 0.7144, "step": 920 }, { "epoch": 1.706642066420664, "grad_norm": 1.2013984113686338, "learning_rate": 9.345522298826177e-06, "loss": 0.7146, "step": 925 }, { "epoch": 1.7158671586715868, "grad_norm": 1.1285995450468198, "learning_rate": 9.238424108534333e-06, "loss": 0.7126, "step": 930 }, { "epoch": 1.725092250922509, "grad_norm": 1.1727971825533714, "learning_rate": 9.131413711288485e-06, "loss": 0.7173, "step": 935 }, { "epoch": 1.7343173431734318, "grad_norm": 1.198238879588798, "learning_rate": 9.024503443047318e-06, "loss": 0.7186, "step": 940 }, { "epoch": 1.7435424354243543, "grad_norm": 1.2092538734459182, "learning_rate": 8.917705628226823e-06, "loss": 0.7064, "step": 945 }, { "epoch": 1.7527675276752768, "grad_norm": 1.1850959753551464, "learning_rate": 8.81103257827957e-06, "loss": 0.7196, "step": 950 }, { "epoch": 1.7619926199261993, "grad_norm": 1.1849846233150378, "learning_rate": 8.704496590275479e-06, "loss": 0.7181, "step": 955 }, { "epoch": 1.7712177121771218, "grad_norm": 1.1192440025321218, "learning_rate": 8.598109945484208e-06, "loss": 0.7127, "step": 960 }, { "epoch": 1.7804428044280443, "grad_norm": 1.185810311236685, "learning_rate": 8.491884907959426e-06, "loss": 0.7092, "step": 965 }, { "epoch": 1.7896678966789668, "grad_norm": 1.1653670987242044, "learning_rate": 8.385833723125006e-06, "loss": 0.7115, "step": 970 }, { "epoch": 1.7988929889298892, "grad_norm": 1.2928934171032893, "learning_rate": 8.279968616363417e-06, "loss": 0.7116, "step": 975 }, { "epoch": 1.8081180811808117, "grad_norm": 1.1749460908752425, "learning_rate": 8.174301791606384e-06, "loss": 0.7159, "step": 980 }, { "epoch": 1.8173431734317345, "grad_norm": 1.2968530721458553, "learning_rate": 8.06884542992806e-06, "loss": 0.7022, "step": 985 }, { "epoch": 1.8265682656826567, "grad_norm": 1.214409149915767, "learning_rate": 7.963611688140814e-06, "loss": 0.705, "step": 990 }, { "epoch": 1.8357933579335795, "grad_norm": 1.1751136227927774, "learning_rate": 7.858612697393792e-06, "loss": 0.7166, "step": 995 }, { "epoch": 1.8450184501845017, "grad_norm": 1.2707314516132002, "learning_rate": 7.753860561774495e-06, "loss": 0.7095, "step": 1000 }, { "epoch": 1.8450184501845017, "eval_loss": 0.9393758773803711, "eval_runtime": 524.5955, "eval_samples_per_second": 29.261, "eval_steps_per_second": 0.114, "step": 1000 }, { "epoch": 1.8542435424354244, "grad_norm": 1.2737022554438457, "learning_rate": 7.649367356913422e-06, "loss": 0.7133, "step": 1005 }, { "epoch": 1.8634686346863467, "grad_norm": 1.2146494230865963, "learning_rate": 7.545145128592009e-06, "loss": 0.7162, "step": 1010 }, { "epoch": 1.8726937269372694, "grad_norm": 1.2563305762066708, "learning_rate": 7.441205891354037e-06, "loss": 0.7128, "step": 1015 }, { "epoch": 1.881918819188192, "grad_norm": 1.2400110075293442, "learning_rate": 7.337561627120591e-06, "loss": 0.7059, "step": 1020 }, { "epoch": 1.8911439114391144, "grad_norm": 1.2653437150866325, "learning_rate": 7.234224283808832e-06, "loss": 0.7058, "step": 1025 }, { "epoch": 1.900369003690037, "grad_norm": 1.1646651085367645, "learning_rate": 7.131205773954636e-06, "loss": 0.706, "step": 1030 }, { "epoch": 1.9095940959409594, "grad_norm": 1.1518551233990397, "learning_rate": 7.028517973339361e-06, "loss": 0.7138, "step": 1035 }, { "epoch": 1.918819188191882, "grad_norm": 1.223360815231687, "learning_rate": 6.926172719620827e-06, "loss": 0.697, "step": 1040 }, { "epoch": 1.9280442804428044, "grad_norm": 1.2198079984824493, "learning_rate": 6.824181810968675e-06, "loss": 0.7004, "step": 1045 }, { "epoch": 1.937269372693727, "grad_norm": 1.176959664107674, "learning_rate": 6.722557004704322e-06, "loss": 0.7082, "step": 1050 }, { "epoch": 1.9464944649446494, "grad_norm": 1.1844320699248965, "learning_rate": 6.62131001594558e-06, "loss": 0.7043, "step": 1055 }, { "epoch": 1.9557195571955721, "grad_norm": 1.148753422424237, "learning_rate": 6.520452516256157e-06, "loss": 0.6949, "step": 1060 }, { "epoch": 1.9649446494464944, "grad_norm": 1.1572577267352544, "learning_rate": 6.419996132300203e-06, "loss": 0.7071, "step": 1065 }, { "epoch": 1.974169741697417, "grad_norm": 1.2001014830908205, "learning_rate": 6.319952444501984e-06, "loss": 0.7103, "step": 1070 }, { "epoch": 1.9833948339483394, "grad_norm": 1.4841715888010063, "learning_rate": 6.220332985710936e-06, "loss": 0.694, "step": 1075 }, { "epoch": 1.992619926199262, "grad_norm": 1.4256755997357629, "learning_rate": 6.121149239872151e-06, "loss": 0.6964, "step": 1080 }, { "epoch": 2.0018450184501844, "grad_norm": 4.270149025567802, "learning_rate": 6.0224126407025616e-06, "loss": 0.6543, "step": 1085 }, { "epoch": 2.011070110701107, "grad_norm": 2.6490744221351044, "learning_rate": 5.924134570372863e-06, "loss": 0.4529, "step": 1090 }, { "epoch": 2.0202952029520294, "grad_norm": 2.2645999605838227, "learning_rate": 5.826326358195391e-06, "loss": 0.4559, "step": 1095 }, { "epoch": 2.029520295202952, "grad_norm": 1.5705400512864462, "learning_rate": 5.728999279318131e-06, "loss": 0.4465, "step": 1100 }, { "epoch": 2.029520295202952, "eval_loss": 0.9917108416557312, "eval_runtime": 517.7798, "eval_samples_per_second": 29.646, "eval_steps_per_second": 0.116, "step": 1100 }, { "epoch": 2.0387453874538743, "grad_norm": 1.6254518927847355, "learning_rate": 5.632164553424904e-06, "loss": 0.4353, "step": 1105 }, { "epoch": 2.047970479704797, "grad_norm": 14.583137561537578, "learning_rate": 5.5358333434420054e-06, "loss": 0.4424, "step": 1110 }, { "epoch": 2.0571955719557193, "grad_norm": 1.447005279720627, "learning_rate": 5.440016754251364e-06, "loss": 0.4423, "step": 1115 }, { "epoch": 2.066420664206642, "grad_norm": 1.4595204240426687, "learning_rate": 5.344725831410369e-06, "loss": 0.4384, "step": 1120 }, { "epoch": 2.0756457564575648, "grad_norm": 1.3190598016289843, "learning_rate": 5.24997155987859e-06, "loss": 0.4368, "step": 1125 }, { "epoch": 2.084870848708487, "grad_norm": 1.322338946677976, "learning_rate": 5.155764862751427e-06, "loss": 0.4392, "step": 1130 }, { "epoch": 2.0940959409594098, "grad_norm": 1.3472757392525208, "learning_rate": 5.062116600000933e-06, "loss": 0.4297, "step": 1135 }, { "epoch": 2.103321033210332, "grad_norm": 1.2895577097092337, "learning_rate": 4.969037567223881e-06, "loss": 0.4413, "step": 1140 }, { "epoch": 2.1125461254612548, "grad_norm": 1.3471090116973288, "learning_rate": 4.876538494397274e-06, "loss": 0.4317, "step": 1145 }, { "epoch": 2.121771217712177, "grad_norm": 1.3092628602239211, "learning_rate": 4.784630044641435e-06, "loss": 0.4509, "step": 1150 }, { "epoch": 2.1309963099630997, "grad_norm": 1.344809966917295, "learning_rate": 4.6933228129907395e-06, "loss": 0.4375, "step": 1155 }, { "epoch": 2.140221402214022, "grad_norm": 1.3014430618254322, "learning_rate": 4.602627325172279e-06, "loss": 0.4424, "step": 1160 }, { "epoch": 2.1494464944649447, "grad_norm": 1.3672933559982345, "learning_rate": 4.512554036392448e-06, "loss": 0.4419, "step": 1165 }, { "epoch": 2.158671586715867, "grad_norm": 1.3446667993737584, "learning_rate": 4.423113330131708e-06, "loss": 0.4303, "step": 1170 }, { "epoch": 2.1678966789667897, "grad_norm": 1.3257443131859206, "learning_rate": 4.33431551694758e-06, "loss": 0.4369, "step": 1175 }, { "epoch": 2.177121771217712, "grad_norm": 1.3655737456565726, "learning_rate": 4.246170833286075e-06, "loss": 0.4293, "step": 1180 }, { "epoch": 2.1863468634686347, "grad_norm": 1.3298593125645854, "learning_rate": 4.1586894403016576e-06, "loss": 0.439, "step": 1185 }, { "epoch": 2.195571955719557, "grad_norm": 1.32505780264794, "learning_rate": 4.071881422685877e-06, "loss": 0.4285, "step": 1190 }, { "epoch": 2.2047970479704797, "grad_norm": 1.3004312804341762, "learning_rate": 3.985756787504837e-06, "loss": 0.4353, "step": 1195 }, { "epoch": 2.2140221402214024, "grad_norm": 1.3177561620055287, "learning_rate": 3.9003254630455775e-06, "loss": 0.4341, "step": 1200 }, { "epoch": 2.2140221402214024, "eval_loss": 0.9978848695755005, "eval_runtime": 514.7843, "eval_samples_per_second": 29.818, "eval_steps_per_second": 0.117, "step": 1200 }, { "epoch": 2.2232472324723247, "grad_norm": 1.3438896554856818, "learning_rate": 3.815597297671578e-06, "loss": 0.4336, "step": 1205 }, { "epoch": 2.2324723247232474, "grad_norm": 1.2896295540334282, "learning_rate": 3.731582058687462e-06, "loss": 0.435, "step": 1210 }, { "epoch": 2.2416974169741697, "grad_norm": 1.358035688644123, "learning_rate": 3.6482894312130146e-06, "loss": 0.4324, "step": 1215 }, { "epoch": 2.2509225092250924, "grad_norm": 1.312197292051631, "learning_rate": 3.565729017066729e-06, "loss": 0.4315, "step": 1220 }, { "epoch": 2.2601476014760147, "grad_norm": 1.3227121347141655, "learning_rate": 3.483910333658913e-06, "loss": 0.4364, "step": 1225 }, { "epoch": 2.2693726937269374, "grad_norm": 1.3256090212374516, "learning_rate": 3.402842812894529e-06, "loss": 0.4356, "step": 1230 }, { "epoch": 2.2785977859778597, "grad_norm": 1.317549750635349, "learning_rate": 3.3225358000859287e-06, "loss": 0.4349, "step": 1235 }, { "epoch": 2.2878228782287824, "grad_norm": 1.2612830347481554, "learning_rate": 3.2429985528755127e-06, "loss": 0.4306, "step": 1240 }, { "epoch": 2.2970479704797047, "grad_norm": 1.3450073317730427, "learning_rate": 3.1642402401685557e-06, "loss": 0.4361, "step": 1245 }, { "epoch": 2.3062730627306274, "grad_norm": 1.3431835139445107, "learning_rate": 3.0862699410762043e-06, "loss": 0.4393, "step": 1250 }, { "epoch": 2.3154981549815496, "grad_norm": 1.3379126436430948, "learning_rate": 3.0090966438688774e-06, "loss": 0.4306, "step": 1255 }, { "epoch": 2.3247232472324724, "grad_norm": 1.2809064467748859, "learning_rate": 2.9327292449401067e-06, "loss": 0.4416, "step": 1260 }, { "epoch": 2.3339483394833946, "grad_norm": 1.3548015164880183, "learning_rate": 2.8571765477809645e-06, "loss": 0.4338, "step": 1265 }, { "epoch": 2.3431734317343174, "grad_norm": 1.320665427008479, "learning_rate": 2.7824472619652386e-06, "loss": 0.4361, "step": 1270 }, { "epoch": 2.35239852398524, "grad_norm": 1.3096646770487193, "learning_rate": 2.7085500021453838e-06, "loss": 0.4294, "step": 1275 }, { "epoch": 2.3616236162361623, "grad_norm": 1.2800372167523524, "learning_rate": 2.635493287059464e-06, "loss": 0.4299, "step": 1280 }, { "epoch": 2.3708487084870846, "grad_norm": 1.303993086907089, "learning_rate": 2.563285538549104e-06, "loss": 0.4361, "step": 1285 }, { "epoch": 2.3800738007380073, "grad_norm": 1.2720280407092956, "learning_rate": 2.491935080588658e-06, "loss": 0.4384, "step": 1290 }, { "epoch": 2.38929889298893, "grad_norm": 1.2941980810201439, "learning_rate": 2.421450138325625e-06, "loss": 0.4306, "step": 1295 }, { "epoch": 2.3985239852398523, "grad_norm": 1.2949495993502738, "learning_rate": 2.351838837132464e-06, "loss": 0.432, "step": 1300 }, { "epoch": 2.3985239852398523, "eval_loss": 0.9954376816749573, "eval_runtime": 519.9495, "eval_samples_per_second": 29.522, "eval_steps_per_second": 0.115, "step": 1300 }, { "epoch": 2.407749077490775, "grad_norm": 1.3018815365771563, "learning_rate": 2.283109201669936e-06, "loss": 0.4357, "step": 1305 }, { "epoch": 2.4169741697416973, "grad_norm": 1.2956106687686837, "learning_rate": 2.2152691549620155e-06, "loss": 0.4283, "step": 1310 }, { "epoch": 2.42619926199262, "grad_norm": 1.287230882437174, "learning_rate": 2.148326517482543e-06, "loss": 0.4303, "step": 1315 }, { "epoch": 2.4354243542435423, "grad_norm": 1.2592322120333668, "learning_rate": 2.0822890062537106e-06, "loss": 0.4366, "step": 1320 }, { "epoch": 2.444649446494465, "grad_norm": 1.3039469988205457, "learning_rate": 2.01716423395644e-06, "loss": 0.4317, "step": 1325 }, { "epoch": 2.4538745387453873, "grad_norm": 1.282772824972497, "learning_rate": 1.9529597080528207e-06, "loss": 0.4272, "step": 1330 }, { "epoch": 2.46309963099631, "grad_norm": 1.3227463435260074, "learning_rate": 1.8896828299206494e-06, "loss": 0.4256, "step": 1335 }, { "epoch": 2.4723247232472323, "grad_norm": 1.3607936617452498, "learning_rate": 1.8273408940002202e-06, "loss": 0.4389, "step": 1340 }, { "epoch": 2.481549815498155, "grad_norm": 1.2740801988744865, "learning_rate": 1.7659410869534466e-06, "loss": 0.4247, "step": 1345 }, { "epoch": 2.4907749077490777, "grad_norm": 1.2544315701192987, "learning_rate": 1.7054904868353717e-06, "loss": 0.4256, "step": 1350 }, { "epoch": 2.5, "grad_norm": 1.31550558585801, "learning_rate": 1.6459960622782466e-06, "loss": 0.428, "step": 1355 }, { "epoch": 2.5092250922509223, "grad_norm": 1.3030144767834306, "learning_rate": 1.587464671688187e-06, "loss": 0.4217, "step": 1360 }, { "epoch": 2.518450184501845, "grad_norm": 1.261812680015863, "learning_rate": 1.5299030624545563e-06, "loss": 0.4381, "step": 1365 }, { "epoch": 2.5276752767527677, "grad_norm": 1.3015065571944802, "learning_rate": 1.4733178701721262e-06, "loss": 0.4337, "step": 1370 }, { "epoch": 2.53690036900369, "grad_norm": 1.2805139778312684, "learning_rate": 1.4177156178761508e-06, "loss": 0.4313, "step": 1375 }, { "epoch": 2.5461254612546127, "grad_norm": 1.3271791125805354, "learning_rate": 1.363102715290402e-06, "loss": 0.4314, "step": 1380 }, { "epoch": 2.555350553505535, "grad_norm": 1.3155240192251205, "learning_rate": 1.3094854580882599e-06, "loss": 0.4298, "step": 1385 }, { "epoch": 2.5645756457564577, "grad_norm": 1.2884517504542843, "learning_rate": 1.2568700271669676e-06, "loss": 0.4315, "step": 1390 }, { "epoch": 2.57380073800738, "grad_norm": 1.2601572769871257, "learning_rate": 1.2052624879351105e-06, "loss": 0.4341, "step": 1395 }, { "epoch": 2.5830258302583027, "grad_norm": 1.283042988722646, "learning_rate": 1.1546687896133924e-06, "loss": 0.4301, "step": 1400 }, { "epoch": 2.5830258302583027, "eval_loss": 0.9943162798881531, "eval_runtime": 513.9906, "eval_samples_per_second": 29.864, "eval_steps_per_second": 0.117, "step": 1400 }, { "epoch": 2.592250922509225, "grad_norm": 1.269448040169663, "learning_rate": 1.1050947645488419e-06, "loss": 0.424, "step": 1405 }, { "epoch": 2.6014760147601477, "grad_norm": 1.291108826010762, "learning_rate": 1.0565461275424504e-06, "loss": 0.4288, "step": 1410 }, { "epoch": 2.61070110701107, "grad_norm": 1.246075371329031, "learning_rate": 1.0090284751903989e-06, "loss": 0.4308, "step": 1415 }, { "epoch": 2.6199261992619927, "grad_norm": 1.268331381912208, "learning_rate": 9.625472852388739e-07, "loss": 0.4274, "step": 1420 }, { "epoch": 2.6291512915129154, "grad_norm": 1.2558980878489436, "learning_rate": 9.171079159526186e-07, "loss": 0.4263, "step": 1425 }, { "epoch": 2.6383763837638377, "grad_norm": 1.2507458001549574, "learning_rate": 8.727156054972374e-07, "loss": 0.4364, "step": 1430 }, { "epoch": 2.64760147601476, "grad_norm": 1.2344093421817917, "learning_rate": 8.29375471335343e-07, "loss": 0.43, "step": 1435 }, { "epoch": 2.6568265682656826, "grad_norm": 1.2520176453134155, "learning_rate": 7.870925096366366e-07, "loss": 0.4298, "step": 1440 }, { "epoch": 2.6660516605166054, "grad_norm": 1.2874930933327957, "learning_rate": 7.458715947019468e-07, "loss": 0.4262, "step": 1445 }, { "epoch": 2.6752767527675276, "grad_norm": 1.2682188739552445, "learning_rate": 7.057174784013432e-07, "loss": 0.4339, "step": 1450 }, { "epoch": 2.6845018450184504, "grad_norm": 1.2828645340804818, "learning_rate": 6.666347896263326e-07, "loss": 0.4274, "step": 1455 }, { "epoch": 2.6937269372693726, "grad_norm": 1.2595258026091076, "learning_rate": 6.286280337562656e-07, "loss": 0.4303, "step": 1460 }, { "epoch": 2.7029520295202953, "grad_norm": 1.24521822647123, "learning_rate": 5.917015921389569e-07, "loss": 0.4288, "step": 1465 }, { "epoch": 2.7121771217712176, "grad_norm": 1.232445478302712, "learning_rate": 5.558597215856065e-07, "loss": 0.4285, "step": 1470 }, { "epoch": 2.7214022140221403, "grad_norm": 1.216057817991593, "learning_rate": 5.211065538800952e-07, "loss": 0.4208, "step": 1475 }, { "epoch": 2.7306273062730626, "grad_norm": 1.288524367589534, "learning_rate": 4.874460953026705e-07, "loss": 0.4255, "step": 1480 }, { "epoch": 2.7398523985239853, "grad_norm": 1.2332155213343263, "learning_rate": 4.548822261681107e-07, "loss": 0.423, "step": 1485 }, { "epoch": 2.7490774907749076, "grad_norm": 1.2278878382563285, "learning_rate": 4.2341870037841516e-07, "loss": 0.4291, "step": 1490 }, { "epoch": 2.7583025830258303, "grad_norm": 1.262898121860552, "learning_rate": 3.930591449900578e-07, "loss": 0.4247, "step": 1495 }, { "epoch": 2.767527675276753, "grad_norm": 1.2437619506416164, "learning_rate": 3.638070597958665e-07, "loss": 0.4361, "step": 1500 }, { "epoch": 2.767527675276753, "eval_loss": 0.9930853247642517, "eval_runtime": 516.1928, "eval_samples_per_second": 29.737, "eval_steps_per_second": 0.116, "step": 1500 }, { "epoch": 2.7767527675276753, "grad_norm": 1.2468366513522777, "learning_rate": 3.356658169215743e-07, "loss": 0.4282, "step": 1505 }, { "epoch": 2.7859778597785976, "grad_norm": 1.2336029324910027, "learning_rate": 3.0863866043708393e-07, "loss": 0.4267, "step": 1510 }, { "epoch": 2.7952029520295203, "grad_norm": 1.3330748292636831, "learning_rate": 2.8272870598250677e-07, "loss": 0.4281, "step": 1515 }, { "epoch": 2.804428044280443, "grad_norm": 1.2486193575900169, "learning_rate": 2.5793894040898384e-07, "loss": 0.4224, "step": 1520 }, { "epoch": 2.8136531365313653, "grad_norm": 1.235394179484528, "learning_rate": 2.3427222143438065e-07, "loss": 0.4184, "step": 1525 }, { "epoch": 2.8228782287822876, "grad_norm": 1.2913244981868073, "learning_rate": 2.117312773138458e-07, "loss": 0.4238, "step": 1530 }, { "epoch": 2.8321033210332103, "grad_norm": 1.2580451640594703, "learning_rate": 1.903187065253076e-07, "loss": 0.4274, "step": 1535 }, { "epoch": 2.841328413284133, "grad_norm": 1.262849856657073, "learning_rate": 1.7003697746992398e-07, "loss": 0.4242, "step": 1540 }, { "epoch": 2.8505535055350553, "grad_norm": 1.2336423601103856, "learning_rate": 1.5088842818752892e-07, "loss": 0.4338, "step": 1545 }, { "epoch": 2.859778597785978, "grad_norm": 1.279029201429549, "learning_rate": 1.3287526608711132e-07, "loss": 0.4247, "step": 1550 }, { "epoch": 2.8690036900369003, "grad_norm": 1.2569044993333771, "learning_rate": 1.1599956769234533e-07, "loss": 0.4167, "step": 1555 }, { "epoch": 2.878228782287823, "grad_norm": 1.229520630461672, "learning_rate": 1.0026327840221728e-07, "loss": 0.4182, "step": 1560 }, { "epoch": 2.8874538745387452, "grad_norm": 1.255986608003343, "learning_rate": 8.566821226675514e-08, "loss": 0.4294, "step": 1565 }, { "epoch": 2.896678966789668, "grad_norm": 1.2895814979486142, "learning_rate": 7.22160517779169e-08, "loss": 0.429, "step": 1570 }, { "epoch": 2.9059040590405907, "grad_norm": 1.2790109206046127, "learning_rate": 5.99083476756357e-08, "loss": 0.4261, "step": 1575 }, { "epoch": 2.915129151291513, "grad_norm": 1.2194809596900478, "learning_rate": 4.87465187690439e-08, "loss": 0.4211, "step": 1580 }, { "epoch": 2.9243542435424352, "grad_norm": 1.2665552740156838, "learning_rate": 3.873185177292737e-08, "loss": 0.4251, "step": 1585 }, { "epoch": 2.933579335793358, "grad_norm": 1.2812371627035533, "learning_rate": 2.9865501159387355e-08, "loss": 0.4282, "step": 1590 }, { "epoch": 2.9428044280442807, "grad_norm": 1.2399165066075877, "learning_rate": 2.214848902475808e-08, "loss": 0.4341, "step": 1595 }, { "epoch": 2.952029520295203, "grad_norm": 1.2154194504631015, "learning_rate": 1.558170497178213e-08, "loss": 0.4256, "step": 1600 }, { "epoch": 2.952029520295203, "eval_loss": 0.9934021830558777, "eval_runtime": 525.4852, "eval_samples_per_second": 29.211, "eval_steps_per_second": 0.114, "step": 1600 }, { "epoch": 2.961254612546125, "grad_norm": 1.2717521820081574, "learning_rate": 1.0165906007056914e-08, "loss": 0.4323, "step": 1605 }, { "epoch": 2.970479704797048, "grad_norm": 1.2491830905746684, "learning_rate": 5.901716453770023e-09, "loss": 0.4271, "step": 1610 }, { "epoch": 2.9797047970479706, "grad_norm": 1.2521953436091506, "learning_rate": 2.7896278797256983e-09, "loss": 0.4256, "step": 1615 }, { "epoch": 2.988929889298893, "grad_norm": 1.2335508198968657, "learning_rate": 8.299990406823721e-10, "loss": 0.4342, "step": 1620 }, { "epoch": 2.9981549815498156, "grad_norm": 1.2480273735451688, "learning_rate": 2.3055838990204693e-11, "loss": 0.4266, "step": 1625 }, { "epoch": 3.0, "step": 1626, "total_flos": 1361805280542720.0, "train_loss": 0.713569560815634, "train_runtime": 59769.2599, "train_samples_per_second": 6.961, "train_steps_per_second": 0.027 } ], "logging_steps": 5, "max_steps": 1626, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1361805280542720.0, "train_batch_size": 16, "trial_name": null, "trial_params": null }