diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,7395 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 9.990375360923965, + "eval_steps": 500, + "global_step": 5190, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0019249278152069298, + "grad_norm": 9.235594749450684, + "learning_rate": 3.8535645472061657e-07, + "loss": 2.3328, + "step": 1 + }, + { + "epoch": 0.009624639076034648, + "grad_norm": 9.342337608337402, + "learning_rate": 1.9267822736030827e-06, + "loss": 2.3107, + "step": 5 + }, + { + "epoch": 0.019249278152069296, + "grad_norm": 8.154550552368164, + "learning_rate": 3.853564547206165e-06, + "loss": 2.3049, + "step": 10 + }, + { + "epoch": 0.028873917228103944, + "grad_norm": 5.9875688552856445, + "learning_rate": 5.780346820809249e-06, + "loss": 2.1949, + "step": 15 + }, + { + "epoch": 0.03849855630413859, + "grad_norm": 2.7122750282287598, + "learning_rate": 7.70712909441233e-06, + "loss": 2.0383, + "step": 20 + }, + { + "epoch": 0.04812319538017324, + "grad_norm": 1.6343287229537964, + "learning_rate": 9.633911368015415e-06, + "loss": 1.9244, + "step": 25 + }, + { + "epoch": 0.05774783445620789, + "grad_norm": 0.805985152721405, + "learning_rate": 1.1560693641618498e-05, + "loss": 1.8037, + "step": 30 + }, + { + "epoch": 0.06737247353224254, + "grad_norm": 0.685213029384613, + "learning_rate": 1.348747591522158e-05, + "loss": 1.7133, + "step": 35 + }, + { + "epoch": 0.07699711260827719, + "grad_norm": 0.5439901351928711, + "learning_rate": 1.541425818882466e-05, + "loss": 1.6271, + "step": 40 + }, + { + "epoch": 0.08662175168431184, + "grad_norm": 0.5319092273712158, + "learning_rate": 1.7341040462427746e-05, + "loss": 1.5405, + "step": 45 + }, + { + "epoch": 0.09624639076034648, + "grad_norm": 0.5163573026657104, + "learning_rate": 1.926782273603083e-05, + "loss": 1.4612, + "step": 50 + }, + { + "epoch": 0.10587102983638114, + "grad_norm": 0.4213581085205078, + "learning_rate": 2.119460500963391e-05, + "loss": 1.3647, + "step": 55 + }, + { + "epoch": 0.11549566891241578, + "grad_norm": 0.37413254380226135, + "learning_rate": 2.3121387283236996e-05, + "loss": 1.3279, + "step": 60 + }, + { + "epoch": 0.12512030798845045, + "grad_norm": 0.3393540680408478, + "learning_rate": 2.504816955684008e-05, + "loss": 1.2962, + "step": 65 + }, + { + "epoch": 0.1347449470644851, + "grad_norm": 0.3041280210018158, + "learning_rate": 2.697495183044316e-05, + "loss": 1.2746, + "step": 70 + }, + { + "epoch": 0.14436958614051973, + "grad_norm": 0.29960623383522034, + "learning_rate": 2.8901734104046245e-05, + "loss": 1.2432, + "step": 75 + }, + { + "epoch": 0.15399422521655437, + "grad_norm": 0.28563690185546875, + "learning_rate": 3.082851637764932e-05, + "loss": 1.224, + "step": 80 + }, + { + "epoch": 0.16361886429258904, + "grad_norm": 0.3082931339740753, + "learning_rate": 3.275529865125241e-05, + "loss": 1.2034, + "step": 85 + }, + { + "epoch": 0.17324350336862368, + "grad_norm": 0.3015296757221222, + "learning_rate": 3.468208092485549e-05, + "loss": 1.186, + "step": 90 + }, + { + "epoch": 0.18286814244465832, + "grad_norm": 0.330247163772583, + "learning_rate": 3.660886319845858e-05, + "loss": 1.1795, + "step": 95 + }, + { + "epoch": 0.19249278152069296, + "grad_norm": 0.30705899000167847, + "learning_rate": 3.853564547206166e-05, + "loss": 1.171, + "step": 100 + }, + { + "epoch": 0.20211742059672763, + "grad_norm": 0.3239520192146301, + "learning_rate": 4.046242774566474e-05, + "loss": 1.1629, + "step": 105 + }, + { + "epoch": 0.21174205967276227, + "grad_norm": 0.31190788745880127, + "learning_rate": 4.238921001926782e-05, + "loss": 1.1507, + "step": 110 + }, + { + "epoch": 0.22136669874879691, + "grad_norm": 0.3129926025867462, + "learning_rate": 4.43159922928709e-05, + "loss": 1.1597, + "step": 115 + }, + { + "epoch": 0.23099133782483156, + "grad_norm": 0.32413914799690247, + "learning_rate": 4.624277456647399e-05, + "loss": 1.1507, + "step": 120 + }, + { + "epoch": 0.24061597690086622, + "grad_norm": 0.41083359718322754, + "learning_rate": 4.816955684007707e-05, + "loss": 1.1259, + "step": 125 + }, + { + "epoch": 0.2502406159769009, + "grad_norm": 0.3095736801624298, + "learning_rate": 5.009633911368016e-05, + "loss": 1.124, + "step": 130 + }, + { + "epoch": 0.2598652550529355, + "grad_norm": 0.3358061611652374, + "learning_rate": 5.2023121387283234e-05, + "loss": 1.1299, + "step": 135 + }, + { + "epoch": 0.2694898941289702, + "grad_norm": 0.37028777599334717, + "learning_rate": 5.394990366088632e-05, + "loss": 1.1085, + "step": 140 + }, + { + "epoch": 0.2791145332050048, + "grad_norm": 0.3638240396976471, + "learning_rate": 5.58766859344894e-05, + "loss": 1.1139, + "step": 145 + }, + { + "epoch": 0.28873917228103946, + "grad_norm": 0.3208532929420471, + "learning_rate": 5.780346820809249e-05, + "loss": 1.0867, + "step": 150 + }, + { + "epoch": 0.2983638113570741, + "grad_norm": 0.325976699590683, + "learning_rate": 5.973025048169557e-05, + "loss": 1.0794, + "step": 155 + }, + { + "epoch": 0.30798845043310874, + "grad_norm": 0.3301510214805603, + "learning_rate": 6.165703275529865e-05, + "loss": 1.0811, + "step": 160 + }, + { + "epoch": 0.3176130895091434, + "grad_norm": 0.35519587993621826, + "learning_rate": 6.358381502890174e-05, + "loss": 1.076, + "step": 165 + }, + { + "epoch": 0.3272377285851781, + "grad_norm": 0.38242989778518677, + "learning_rate": 6.551059730250482e-05, + "loss": 1.0774, + "step": 170 + }, + { + "epoch": 0.3368623676612127, + "grad_norm": 0.3178574740886688, + "learning_rate": 6.74373795761079e-05, + "loss": 1.0678, + "step": 175 + }, + { + "epoch": 0.34648700673724736, + "grad_norm": 0.2955685257911682, + "learning_rate": 6.936416184971098e-05, + "loss": 1.0741, + "step": 180 + }, + { + "epoch": 0.35611164581328203, + "grad_norm": 0.3037715554237366, + "learning_rate": 7.129094412331408e-05, + "loss": 1.0649, + "step": 185 + }, + { + "epoch": 0.36573628488931664, + "grad_norm": 0.3199213445186615, + "learning_rate": 7.321772639691716e-05, + "loss": 1.0635, + "step": 190 + }, + { + "epoch": 0.3753609239653513, + "grad_norm": 0.317488431930542, + "learning_rate": 7.514450867052023e-05, + "loss": 1.0526, + "step": 195 + }, + { + "epoch": 0.3849855630413859, + "grad_norm": 0.3228258490562439, + "learning_rate": 7.707129094412332e-05, + "loss": 1.064, + "step": 200 + }, + { + "epoch": 0.3946102021174206, + "grad_norm": 0.2934040129184723, + "learning_rate": 7.89980732177264e-05, + "loss": 1.0544, + "step": 205 + }, + { + "epoch": 0.40423484119345526, + "grad_norm": 0.32170167565345764, + "learning_rate": 8.092485549132948e-05, + "loss": 1.0508, + "step": 210 + }, + { + "epoch": 0.4138594802694899, + "grad_norm": 0.29049986600875854, + "learning_rate": 8.285163776493256e-05, + "loss": 1.0611, + "step": 215 + }, + { + "epoch": 0.42348411934552455, + "grad_norm": 0.31131693720817566, + "learning_rate": 8.477842003853564e-05, + "loss": 1.0581, + "step": 220 + }, + { + "epoch": 0.4331087584215592, + "grad_norm": 0.2872338891029358, + "learning_rate": 8.670520231213874e-05, + "loss": 1.0512, + "step": 225 + }, + { + "epoch": 0.44273339749759383, + "grad_norm": 0.3063661754131317, + "learning_rate": 8.86319845857418e-05, + "loss": 1.0508, + "step": 230 + }, + { + "epoch": 0.4523580365736285, + "grad_norm": 0.30761733651161194, + "learning_rate": 9.05587668593449e-05, + "loss": 1.0549, + "step": 235 + }, + { + "epoch": 0.4619826756496631, + "grad_norm": 0.2758205533027649, + "learning_rate": 9.248554913294798e-05, + "loss": 1.0446, + "step": 240 + }, + { + "epoch": 0.4716073147256978, + "grad_norm": 0.3492432236671448, + "learning_rate": 9.441233140655106e-05, + "loss": 1.0511, + "step": 245 + }, + { + "epoch": 0.48123195380173245, + "grad_norm": 0.27041804790496826, + "learning_rate": 9.633911368015414e-05, + "loss": 1.0275, + "step": 250 + }, + { + "epoch": 0.49085659287776706, + "grad_norm": 0.2999095916748047, + "learning_rate": 9.826589595375723e-05, + "loss": 1.0433, + "step": 255 + }, + { + "epoch": 0.5004812319538018, + "grad_norm": 0.297323614358902, + "learning_rate": 0.00010019267822736032, + "loss": 1.0416, + "step": 260 + }, + { + "epoch": 0.5101058710298364, + "grad_norm": 0.3357987403869629, + "learning_rate": 0.00010211946050096339, + "loss": 1.0374, + "step": 265 + }, + { + "epoch": 0.519730510105871, + "grad_norm": 0.2953435778617859, + "learning_rate": 0.00010404624277456647, + "loss": 1.0352, + "step": 270 + }, + { + "epoch": 0.5293551491819056, + "grad_norm": 0.32853737473487854, + "learning_rate": 0.00010597302504816958, + "loss": 1.0529, + "step": 275 + }, + { + "epoch": 0.5389797882579404, + "grad_norm": 0.28152966499328613, + "learning_rate": 0.00010789980732177264, + "loss": 1.0427, + "step": 280 + }, + { + "epoch": 0.548604427333975, + "grad_norm": 0.2928714454174042, + "learning_rate": 0.00010982658959537572, + "loss": 1.0375, + "step": 285 + }, + { + "epoch": 0.5582290664100096, + "grad_norm": 0.29662230610847473, + "learning_rate": 0.0001117533718689788, + "loss": 1.0326, + "step": 290 + }, + { + "epoch": 0.5678537054860443, + "grad_norm": 0.2677823305130005, + "learning_rate": 0.00011368015414258189, + "loss": 1.0477, + "step": 295 + }, + { + "epoch": 0.5774783445620789, + "grad_norm": 0.2860727906227112, + "learning_rate": 0.00011560693641618498, + "loss": 1.0272, + "step": 300 + }, + { + "epoch": 0.5871029836381135, + "grad_norm": 0.2599497437477112, + "learning_rate": 0.00011753371868978806, + "loss": 1.0364, + "step": 305 + }, + { + "epoch": 0.5967276227141483, + "grad_norm": 0.26607978343963623, + "learning_rate": 0.00011946050096339114, + "loss": 1.0338, + "step": 310 + }, + { + "epoch": 0.6063522617901829, + "grad_norm": 0.2653907239437103, + "learning_rate": 0.00012138728323699422, + "loss": 1.0274, + "step": 315 + }, + { + "epoch": 0.6159769008662175, + "grad_norm": 0.2570829689502716, + "learning_rate": 0.0001233140655105973, + "loss": 1.0349, + "step": 320 + }, + { + "epoch": 0.6256015399422522, + "grad_norm": 0.2542014420032501, + "learning_rate": 0.00012524084778420039, + "loss": 1.0306, + "step": 325 + }, + { + "epoch": 0.6352261790182868, + "grad_norm": 0.2354612797498703, + "learning_rate": 0.00012716763005780348, + "loss": 1.0336, + "step": 330 + }, + { + "epoch": 0.6448508180943214, + "grad_norm": 0.26090219616889954, + "learning_rate": 0.00012909441233140655, + "loss": 1.0319, + "step": 335 + }, + { + "epoch": 0.6544754571703562, + "grad_norm": 0.2287357598543167, + "learning_rate": 0.00013102119460500964, + "loss": 1.0228, + "step": 340 + }, + { + "epoch": 0.6641000962463908, + "grad_norm": 0.2653840184211731, + "learning_rate": 0.0001329479768786127, + "loss": 1.019, + "step": 345 + }, + { + "epoch": 0.6737247353224254, + "grad_norm": 0.25462430715560913, + "learning_rate": 0.0001348747591522158, + "loss": 1.0289, + "step": 350 + }, + { + "epoch": 0.6833493743984601, + "grad_norm": 0.24566137790679932, + "learning_rate": 0.0001368015414258189, + "loss": 1.0286, + "step": 355 + }, + { + "epoch": 0.6929740134744947, + "grad_norm": 0.24448491632938385, + "learning_rate": 0.00013872832369942197, + "loss": 1.0195, + "step": 360 + }, + { + "epoch": 0.7025986525505293, + "grad_norm": 0.2303464114665985, + "learning_rate": 0.00014065510597302506, + "loss": 1.0328, + "step": 365 + }, + { + "epoch": 0.7122232916265641, + "grad_norm": 0.2552158832550049, + "learning_rate": 0.00014258188824662816, + "loss": 1.0366, + "step": 370 + }, + { + "epoch": 0.7218479307025987, + "grad_norm": 0.22079892456531525, + "learning_rate": 0.00014450867052023122, + "loss": 1.024, + "step": 375 + }, + { + "epoch": 0.7314725697786333, + "grad_norm": 0.32242512702941895, + "learning_rate": 0.00014643545279383432, + "loss": 1.033, + "step": 380 + }, + { + "epoch": 0.7410972088546679, + "grad_norm": 0.2999092936515808, + "learning_rate": 0.00014836223506743738, + "loss": 1.0204, + "step": 385 + }, + { + "epoch": 0.7507218479307026, + "grad_norm": 0.26794490218162537, + "learning_rate": 0.00015028901734104045, + "loss": 1.0305, + "step": 390 + }, + { + "epoch": 0.7603464870067372, + "grad_norm": 0.33896663784980774, + "learning_rate": 0.00015221579961464357, + "loss": 1.0308, + "step": 395 + }, + { + "epoch": 0.7699711260827719, + "grad_norm": 0.22748759388923645, + "learning_rate": 0.00015414258188824664, + "loss": 1.0197, + "step": 400 + }, + { + "epoch": 0.7795957651588066, + "grad_norm": 0.23324738442897797, + "learning_rate": 0.0001560693641618497, + "loss": 1.0131, + "step": 405 + }, + { + "epoch": 0.7892204042348412, + "grad_norm": 0.24805064499378204, + "learning_rate": 0.0001579961464354528, + "loss": 1.0094, + "step": 410 + }, + { + "epoch": 0.7988450433108758, + "grad_norm": 0.24965739250183105, + "learning_rate": 0.00015992292870905587, + "loss": 1.0203, + "step": 415 + }, + { + "epoch": 0.8084696823869105, + "grad_norm": 0.22509600222110748, + "learning_rate": 0.00016184971098265897, + "loss": 1.0265, + "step": 420 + }, + { + "epoch": 0.8180943214629451, + "grad_norm": 0.2149883359670639, + "learning_rate": 0.00016377649325626206, + "loss": 1.0171, + "step": 425 + }, + { + "epoch": 0.8277189605389798, + "grad_norm": 0.24780240654945374, + "learning_rate": 0.00016570327552986513, + "loss": 1.0144, + "step": 430 + }, + { + "epoch": 0.8373435996150145, + "grad_norm": 0.2780991494655609, + "learning_rate": 0.00016763005780346822, + "loss": 1.0145, + "step": 435 + }, + { + "epoch": 0.8469682386910491, + "grad_norm": 0.22135606408119202, + "learning_rate": 0.0001695568400770713, + "loss": 1.0187, + "step": 440 + }, + { + "epoch": 0.8565928777670837, + "grad_norm": 0.20605282485485077, + "learning_rate": 0.00017148362235067438, + "loss": 1.0197, + "step": 445 + }, + { + "epoch": 0.8662175168431184, + "grad_norm": 0.24270793795585632, + "learning_rate": 0.00017341040462427748, + "loss": 1.0106, + "step": 450 + }, + { + "epoch": 0.875842155919153, + "grad_norm": 0.24285346269607544, + "learning_rate": 0.00017533718689788055, + "loss": 1.0242, + "step": 455 + }, + { + "epoch": 0.8854667949951877, + "grad_norm": 0.21814145147800446, + "learning_rate": 0.0001772639691714836, + "loss": 1.0176, + "step": 460 + }, + { + "epoch": 0.8950914340712224, + "grad_norm": 0.22261013090610504, + "learning_rate": 0.00017919075144508673, + "loss": 1.0099, + "step": 465 + }, + { + "epoch": 0.904716073147257, + "grad_norm": 0.21424554288387299, + "learning_rate": 0.0001811175337186898, + "loss": 1.0097, + "step": 470 + }, + { + "epoch": 0.9143407122232916, + "grad_norm": 0.2335994988679886, + "learning_rate": 0.00018304431599229287, + "loss": 1.0179, + "step": 475 + }, + { + "epoch": 0.9239653512993262, + "grad_norm": 0.20568034052848816, + "learning_rate": 0.00018497109826589596, + "loss": 1.0067, + "step": 480 + }, + { + "epoch": 0.933589990375361, + "grad_norm": 0.20264984667301178, + "learning_rate": 0.00018689788053949903, + "loss": 1.0147, + "step": 485 + }, + { + "epoch": 0.9432146294513956, + "grad_norm": 0.2133115977048874, + "learning_rate": 0.00018882466281310213, + "loss": 1.0071, + "step": 490 + }, + { + "epoch": 0.9528392685274302, + "grad_norm": 0.2007424235343933, + "learning_rate": 0.00019075144508670522, + "loss": 1.0095, + "step": 495 + }, + { + "epoch": 0.9624639076034649, + "grad_norm": 0.20568867027759552, + "learning_rate": 0.0001926782273603083, + "loss": 1.0113, + "step": 500 + }, + { + "epoch": 0.9720885466794995, + "grad_norm": 0.19897951185703278, + "learning_rate": 0.00019460500963391138, + "loss": 1.0129, + "step": 505 + }, + { + "epoch": 0.9817131857555341, + "grad_norm": 0.21554742753505707, + "learning_rate": 0.00019653179190751445, + "loss": 1.0107, + "step": 510 + }, + { + "epoch": 0.9913378248315688, + "grad_norm": 0.20981793105602264, + "learning_rate": 0.00019845857418111754, + "loss": 1.0008, + "step": 515 + }, + { + "epoch": 0.9990375360923965, + "eval_loss": 2.1032063961029053, + "eval_runtime": 0.7869, + "eval_samples_per_second": 13.979, + "eval_steps_per_second": 2.542, + "step": 519 + }, + { + "epoch": 1.0009624639076036, + "grad_norm": 0.22678163647651672, + "learning_rate": 0.0001999999773822188, + "loss": 1.0012, + "step": 520 + }, + { + "epoch": 1.0105871029836382, + "grad_norm": 0.2608613967895508, + "learning_rate": 0.00019999918576095053, + "loss": 0.9875, + "step": 525 + }, + { + "epoch": 1.0202117420596728, + "grad_norm": 0.2601936459541321, + "learning_rate": 0.0001999972632608527, + "loss": 0.9805, + "step": 530 + }, + { + "epoch": 1.0298363811357074, + "grad_norm": 0.21544857323169708, + "learning_rate": 0.00019999420990366674, + "loss": 0.9805, + "step": 535 + }, + { + "epoch": 1.039461020211742, + "grad_norm": 0.20171190798282623, + "learning_rate": 0.00019999002572392255, + "loss": 0.9798, + "step": 540 + }, + { + "epoch": 1.0490856592877766, + "grad_norm": 0.2205726057291031, + "learning_rate": 0.0001999847107689386, + "loss": 0.9805, + "step": 545 + }, + { + "epoch": 1.0587102983638113, + "grad_norm": 0.20397739112377167, + "learning_rate": 0.0001999782650988211, + "loss": 0.9952, + "step": 550 + }, + { + "epoch": 1.068334937439846, + "grad_norm": 0.207752525806427, + "learning_rate": 0.00019997068878646333, + "loss": 0.9786, + "step": 555 + }, + { + "epoch": 1.0779595765158807, + "grad_norm": 0.2041793167591095, + "learning_rate": 0.0001999619819175449, + "loss": 0.9951, + "step": 560 + }, + { + "epoch": 1.0875842155919153, + "grad_norm": 0.19135500490665436, + "learning_rate": 0.00019995214459053075, + "loss": 0.9912, + "step": 565 + }, + { + "epoch": 1.09720885466795, + "grad_norm": 0.2038804590702057, + "learning_rate": 0.00019994117691667004, + "loss": 0.9821, + "step": 570 + }, + { + "epoch": 1.1068334937439845, + "grad_norm": 0.21948496997356415, + "learning_rate": 0.00019992907901999484, + "loss": 0.9933, + "step": 575 + }, + { + "epoch": 1.1164581328200192, + "grad_norm": 0.21123313903808594, + "learning_rate": 0.0001999158510373189, + "loss": 0.9723, + "step": 580 + }, + { + "epoch": 1.126082771896054, + "grad_norm": 0.2110896110534668, + "learning_rate": 0.00019990149311823588, + "loss": 0.9789, + "step": 585 + }, + { + "epoch": 1.1357074109720886, + "grad_norm": 0.20370599627494812, + "learning_rate": 0.00019988600542511766, + "loss": 0.9902, + "step": 590 + }, + { + "epoch": 1.1453320500481232, + "grad_norm": 0.19531656801700592, + "learning_rate": 0.00019986938813311284, + "loss": 0.9846, + "step": 595 + }, + { + "epoch": 1.1549566891241578, + "grad_norm": 0.2497565895318985, + "learning_rate": 0.00019985164143014432, + "loss": 0.9864, + "step": 600 + }, + { + "epoch": 1.1645813282001924, + "grad_norm": 0.2870050072669983, + "learning_rate": 0.00019983276551690745, + "loss": 0.9851, + "step": 605 + }, + { + "epoch": 1.174205967276227, + "grad_norm": 0.20774626731872559, + "learning_rate": 0.0001998127606068677, + "loss": 0.9819, + "step": 610 + }, + { + "epoch": 1.1838306063522617, + "grad_norm": 0.2567305266857147, + "learning_rate": 0.00019979162692625817, + "loss": 0.9754, + "step": 615 + }, + { + "epoch": 1.1934552454282965, + "grad_norm": 0.1896723359823227, + "learning_rate": 0.00019976936471407717, + "loss": 0.9762, + "step": 620 + }, + { + "epoch": 1.2030798845043311, + "grad_norm": 0.19382244348526, + "learning_rate": 0.00019974597422208533, + "loss": 0.9783, + "step": 625 + }, + { + "epoch": 1.2127045235803657, + "grad_norm": 0.19210918247699738, + "learning_rate": 0.00019972145571480295, + "loss": 0.9778, + "step": 630 + }, + { + "epoch": 1.2223291626564003, + "grad_norm": 0.2057211995124817, + "learning_rate": 0.00019969580946950695, + "loss": 0.9632, + "step": 635 + }, + { + "epoch": 1.231953801732435, + "grad_norm": 0.23469866812229156, + "learning_rate": 0.0001996690357762276, + "loss": 0.9824, + "step": 640 + }, + { + "epoch": 1.2415784408084698, + "grad_norm": 0.19450876116752625, + "learning_rate": 0.00019964113493774538, + "loss": 0.9788, + "step": 645 + }, + { + "epoch": 1.2512030798845044, + "grad_norm": 0.18963035941123962, + "learning_rate": 0.00019961210726958758, + "loss": 0.9854, + "step": 650 + }, + { + "epoch": 1.260827718960539, + "grad_norm": 0.2049696296453476, + "learning_rate": 0.00019958195310002457, + "loss": 0.9901, + "step": 655 + }, + { + "epoch": 1.2704523580365736, + "grad_norm": 0.18745918571949005, + "learning_rate": 0.00019955067277006633, + "loss": 0.9772, + "step": 660 + }, + { + "epoch": 1.2800769971126083, + "grad_norm": 0.1893537938594818, + "learning_rate": 0.00019951826663345827, + "loss": 0.9862, + "step": 665 + }, + { + "epoch": 1.2897016361886429, + "grad_norm": 0.18441106379032135, + "learning_rate": 0.00019948473505667757, + "loss": 0.9836, + "step": 670 + }, + { + "epoch": 1.2993262752646775, + "grad_norm": 0.21260684728622437, + "learning_rate": 0.00019945007841892884, + "loss": 0.9878, + "step": 675 + }, + { + "epoch": 1.3089509143407123, + "grad_norm": 0.19159361720085144, + "learning_rate": 0.00019941429711213982, + "loss": 1.0004, + "step": 680 + }, + { + "epoch": 1.318575553416747, + "grad_norm": 0.19893284142017365, + "learning_rate": 0.000199377391540957, + "loss": 0.9728, + "step": 685 + }, + { + "epoch": 1.3282001924927815, + "grad_norm": 0.2625219225883484, + "learning_rate": 0.00019933936212274115, + "loss": 0.9815, + "step": 690 + }, + { + "epoch": 1.3378248315688162, + "grad_norm": 0.20059077441692352, + "learning_rate": 0.00019930020928756232, + "loss": 0.9869, + "step": 695 + }, + { + "epoch": 1.3474494706448508, + "grad_norm": 0.19443583488464355, + "learning_rate": 0.00019925993347819532, + "loss": 0.9852, + "step": 700 + }, + { + "epoch": 1.3570741097208856, + "grad_norm": 0.19254858791828156, + "learning_rate": 0.00019921853515011438, + "loss": 0.9768, + "step": 705 + }, + { + "epoch": 1.36669874879692, + "grad_norm": 0.1973366141319275, + "learning_rate": 0.0001991760147714883, + "loss": 0.9865, + "step": 710 + }, + { + "epoch": 1.3763233878729548, + "grad_norm": 0.2019069492816925, + "learning_rate": 0.00019913237282317495, + "loss": 0.9701, + "step": 715 + }, + { + "epoch": 1.3859480269489894, + "grad_norm": 0.20254430174827576, + "learning_rate": 0.0001990876097987159, + "loss": 0.9867, + "step": 720 + }, + { + "epoch": 1.395572666025024, + "grad_norm": 0.2121659815311432, + "learning_rate": 0.00019904172620433078, + "loss": 0.9688, + "step": 725 + }, + { + "epoch": 1.4051973051010587, + "grad_norm": 0.2147083729505539, + "learning_rate": 0.00019899472255891176, + "loss": 0.9802, + "step": 730 + }, + { + "epoch": 1.4148219441770933, + "grad_norm": 0.21038152277469635, + "learning_rate": 0.0001989465993940174, + "loss": 0.9759, + "step": 735 + }, + { + "epoch": 1.4244465832531281, + "grad_norm": 0.21153226494789124, + "learning_rate": 0.00019889735725386683, + "loss": 0.9735, + "step": 740 + }, + { + "epoch": 1.4340712223291627, + "grad_norm": 0.2074025273323059, + "learning_rate": 0.00019884699669533347, + "loss": 0.9913, + "step": 745 + }, + { + "epoch": 1.4436958614051973, + "grad_norm": 0.21015384793281555, + "learning_rate": 0.00019879551828793892, + "loss": 0.9737, + "step": 750 + }, + { + "epoch": 1.453320500481232, + "grad_norm": 0.21345528960227966, + "learning_rate": 0.0001987429226138463, + "loss": 0.9675, + "step": 755 + }, + { + "epoch": 1.4629451395572666, + "grad_norm": 0.21284109354019165, + "learning_rate": 0.0001986892102678538, + "loss": 0.9787, + "step": 760 + }, + { + "epoch": 1.4725697786333012, + "grad_norm": 0.19105084240436554, + "learning_rate": 0.0001986343818573879, + "loss": 0.9714, + "step": 765 + }, + { + "epoch": 1.4821944177093358, + "grad_norm": 0.18031322956085205, + "learning_rate": 0.0001985784380024966, + "loss": 0.965, + "step": 770 + }, + { + "epoch": 1.4918190567853706, + "grad_norm": 0.19423770904541016, + "learning_rate": 0.00019852137933584215, + "loss": 0.9743, + "step": 775 + }, + { + "epoch": 1.5014436958614052, + "grad_norm": 0.1923457533121109, + "learning_rate": 0.0001984632065026943, + "loss": 0.9872, + "step": 780 + }, + { + "epoch": 1.5110683349374399, + "grad_norm": 0.1957743912935257, + "learning_rate": 0.0001984039201609226, + "loss": 0.9799, + "step": 785 + }, + { + "epoch": 1.5206929740134745, + "grad_norm": 0.17838570475578308, + "learning_rate": 0.0001983435209809892, + "loss": 0.9765, + "step": 790 + }, + { + "epoch": 1.530317613089509, + "grad_norm": 0.1872684508562088, + "learning_rate": 0.00019828200964594123, + "loss": 0.9768, + "step": 795 + }, + { + "epoch": 1.539942252165544, + "grad_norm": 0.19497379660606384, + "learning_rate": 0.00019821938685140298, + "loss": 0.9686, + "step": 800 + }, + { + "epoch": 1.5495668912415783, + "grad_norm": 0.18703444302082062, + "learning_rate": 0.00019815565330556816, + "loss": 0.9785, + "step": 805 + }, + { + "epoch": 1.5591915303176132, + "grad_norm": 0.18727166950702667, + "learning_rate": 0.00019809080972919181, + "loss": 0.9748, + "step": 810 + }, + { + "epoch": 1.5688161693936478, + "grad_norm": 0.19498740136623383, + "learning_rate": 0.00019802485685558222, + "loss": 0.975, + "step": 815 + }, + { + "epoch": 1.5784408084696824, + "grad_norm": 0.1881551891565323, + "learning_rate": 0.00019795779543059248, + "loss": 0.9749, + "step": 820 + }, + { + "epoch": 1.588065447545717, + "grad_norm": 0.17449571192264557, + "learning_rate": 0.00019788962621261226, + "loss": 0.9676, + "step": 825 + }, + { + "epoch": 1.5976900866217516, + "grad_norm": 0.1892375349998474, + "learning_rate": 0.0001978203499725591, + "loss": 0.9722, + "step": 830 + }, + { + "epoch": 1.6073147256977864, + "grad_norm": 0.19908097386360168, + "learning_rate": 0.00019774996749386968, + "loss": 0.9674, + "step": 835 + }, + { + "epoch": 1.6169393647738208, + "grad_norm": 0.17946478724479675, + "learning_rate": 0.00019767847957249108, + "loss": 0.9741, + "step": 840 + }, + { + "epoch": 1.6265640038498557, + "grad_norm": 0.17460967600345612, + "learning_rate": 0.0001976058870168716, + "loss": 0.9726, + "step": 845 + }, + { + "epoch": 1.6361886429258903, + "grad_norm": 0.17595893144607544, + "learning_rate": 0.0001975321906479518, + "loss": 0.9783, + "step": 850 + }, + { + "epoch": 1.645813282001925, + "grad_norm": 0.18718552589416504, + "learning_rate": 0.00019745739129915508, + "loss": 0.9746, + "step": 855 + }, + { + "epoch": 1.6554379210779597, + "grad_norm": 0.18480895459651947, + "learning_rate": 0.00019738148981637835, + "loss": 0.9675, + "step": 860 + }, + { + "epoch": 1.6650625601539941, + "grad_norm": 0.1780669093132019, + "learning_rate": 0.00019730448705798239, + "loss": 0.9648, + "step": 865 + }, + { + "epoch": 1.674687199230029, + "grad_norm": 0.17525720596313477, + "learning_rate": 0.00019722638389478217, + "loss": 0.9911, + "step": 870 + }, + { + "epoch": 1.6843118383060636, + "grad_norm": 0.1761050969362259, + "learning_rate": 0.00019714718121003705, + "loss": 0.9745, + "step": 875 + }, + { + "epoch": 1.6939364773820982, + "grad_norm": 0.193415105342865, + "learning_rate": 0.00019706687989944072, + "loss": 0.9669, + "step": 880 + }, + { + "epoch": 1.7035611164581328, + "grad_norm": 0.18116651475429535, + "learning_rate": 0.00019698548087111102, + "loss": 0.9573, + "step": 885 + }, + { + "epoch": 1.7131857555341674, + "grad_norm": 0.17790788412094116, + "learning_rate": 0.0001969029850455799, + "loss": 0.9738, + "step": 890 + }, + { + "epoch": 1.7228103946102022, + "grad_norm": 0.18143677711486816, + "learning_rate": 0.00019681939335578275, + "loss": 0.9641, + "step": 895 + }, + { + "epoch": 1.7324350336862366, + "grad_norm": 0.1727439910173416, + "learning_rate": 0.00019673470674704801, + "loss": 0.9612, + "step": 900 + }, + { + "epoch": 1.7420596727622715, + "grad_norm": 0.17776042222976685, + "learning_rate": 0.00019664892617708642, + "loss": 0.9704, + "step": 905 + }, + { + "epoch": 1.751684311838306, + "grad_norm": 0.1788305640220642, + "learning_rate": 0.00019656205261598016, + "loss": 0.9822, + "step": 910 + }, + { + "epoch": 1.7613089509143407, + "grad_norm": 0.18292832374572754, + "learning_rate": 0.00019647408704617192, + "loss": 0.981, + "step": 915 + }, + { + "epoch": 1.7709335899903753, + "grad_norm": 0.1903613954782486, + "learning_rate": 0.00019638503046245383, + "loss": 0.9815, + "step": 920 + }, + { + "epoch": 1.78055822906641, + "grad_norm": 0.18801650404930115, + "learning_rate": 0.00019629488387195614, + "loss": 0.9723, + "step": 925 + }, + { + "epoch": 1.7901828681424448, + "grad_norm": 0.19215719401836395, + "learning_rate": 0.0001962036482941359, + "loss": 0.9785, + "step": 930 + }, + { + "epoch": 1.7998075072184792, + "grad_norm": 0.1913854032754898, + "learning_rate": 0.00019611132476076527, + "loss": 0.9661, + "step": 935 + }, + { + "epoch": 1.809432146294514, + "grad_norm": 0.19718807935714722, + "learning_rate": 0.00019601791431592006, + "loss": 0.9791, + "step": 940 + }, + { + "epoch": 1.8190567853705486, + "grad_norm": 0.18217253684997559, + "learning_rate": 0.00019592341801596787, + "loss": 0.9575, + "step": 945 + }, + { + "epoch": 1.8286814244465832, + "grad_norm": 0.17967750132083893, + "learning_rate": 0.00019582783692955605, + "loss": 0.9637, + "step": 950 + }, + { + "epoch": 1.838306063522618, + "grad_norm": 0.17850783467292786, + "learning_rate": 0.00019573117213759957, + "loss": 0.9605, + "step": 955 + }, + { + "epoch": 1.8479307025986524, + "grad_norm": 0.19147521257400513, + "learning_rate": 0.00019563342473326913, + "loss": 0.9818, + "step": 960 + }, + { + "epoch": 1.8575553416746873, + "grad_norm": 0.17569051682949066, + "learning_rate": 0.00019553459582197835, + "loss": 0.9642, + "step": 965 + }, + { + "epoch": 1.867179980750722, + "grad_norm": 0.18762874603271484, + "learning_rate": 0.00019543468652137157, + "loss": 0.9744, + "step": 970 + }, + { + "epoch": 1.8768046198267565, + "grad_norm": 0.17426376044750214, + "learning_rate": 0.00019533369796131118, + "loss": 0.9725, + "step": 975 + }, + { + "epoch": 1.8864292589027911, + "grad_norm": 0.21174634993076324, + "learning_rate": 0.00019523163128386465, + "loss": 0.9793, + "step": 980 + }, + { + "epoch": 1.8960538979788257, + "grad_norm": 0.19356350600719452, + "learning_rate": 0.00019512848764329188, + "loss": 0.9632, + "step": 985 + }, + { + "epoch": 1.9056785370548606, + "grad_norm": 0.1924716979265213, + "learning_rate": 0.00019502426820603192, + "loss": 0.9791, + "step": 990 + }, + { + "epoch": 1.915303176130895, + "grad_norm": 0.20623841881752014, + "learning_rate": 0.00019491897415068997, + "loss": 0.9678, + "step": 995 + }, + { + "epoch": 1.9249278152069298, + "grad_norm": 0.1916794627904892, + "learning_rate": 0.00019481260666802386, + "loss": 0.9677, + "step": 1000 + }, + { + "epoch": 1.9345524542829644, + "grad_norm": 0.17562657594680786, + "learning_rate": 0.00019470516696093072, + "loss": 0.9818, + "step": 1005 + }, + { + "epoch": 1.944177093358999, + "grad_norm": 0.17680735886096954, + "learning_rate": 0.00019459665624443342, + "loss": 0.9789, + "step": 1010 + }, + { + "epoch": 1.9538017324350336, + "grad_norm": 0.17583592236042023, + "learning_rate": 0.00019448707574566657, + "loss": 0.9758, + "step": 1015 + }, + { + "epoch": 1.9634263715110682, + "grad_norm": 0.18164704740047455, + "learning_rate": 0.00019437642670386304, + "loss": 0.9596, + "step": 1020 + }, + { + "epoch": 1.973051010587103, + "grad_norm": 0.16976359486579895, + "learning_rate": 0.0001942647103703395, + "loss": 0.9725, + "step": 1025 + }, + { + "epoch": 1.9826756496631375, + "grad_norm": 0.16864246129989624, + "learning_rate": 0.00019415192800848263, + "loss": 0.9788, + "step": 1030 + }, + { + "epoch": 1.9923002887391723, + "grad_norm": 0.17933247983455658, + "learning_rate": 0.00019403808089373472, + "loss": 0.9747, + "step": 1035 + }, + { + "epoch": 2.0, + "eval_loss": 2.1443910598754883, + "eval_runtime": 0.7795, + "eval_samples_per_second": 14.112, + "eval_steps_per_second": 2.566, + "step": 1039 + }, + { + "epoch": 2.001924927815207, + "grad_norm": 0.17538660764694214, + "learning_rate": 0.00019392317031357908, + "loss": 0.9576, + "step": 1040 + }, + { + "epoch": 2.0115495668912415, + "grad_norm": 0.18830431997776031, + "learning_rate": 0.00019380719756752584, + "loss": 0.9117, + "step": 1045 + }, + { + "epoch": 2.0211742059672764, + "grad_norm": 0.18357954919338226, + "learning_rate": 0.00019369016396709681, + "loss": 0.9149, + "step": 1050 + }, + { + "epoch": 2.0307988450433108, + "grad_norm": 0.19075176119804382, + "learning_rate": 0.000193572070835811, + "loss": 0.9114, + "step": 1055 + }, + { + "epoch": 2.0404234841193456, + "grad_norm": 0.19288337230682373, + "learning_rate": 0.0001934529195091695, + "loss": 0.9061, + "step": 1060 + }, + { + "epoch": 2.05004812319538, + "grad_norm": 0.1923680603504181, + "learning_rate": 0.00019333271133464047, + "loss": 0.9165, + "step": 1065 + }, + { + "epoch": 2.059672762271415, + "grad_norm": 0.19743940234184265, + "learning_rate": 0.00019321144767164367, + "loss": 0.9115, + "step": 1070 + }, + { + "epoch": 2.0692974013474497, + "grad_norm": 0.18134470283985138, + "learning_rate": 0.00019308912989153548, + "loss": 0.9117, + "step": 1075 + }, + { + "epoch": 2.078922040423484, + "grad_norm": 0.19912441074848175, + "learning_rate": 0.00019296575937759292, + "loss": 0.9139, + "step": 1080 + }, + { + "epoch": 2.088546679499519, + "grad_norm": 0.20187345147132874, + "learning_rate": 0.00019284133752499848, + "loss": 0.9233, + "step": 1085 + }, + { + "epoch": 2.0981713185755533, + "grad_norm": 0.19697998464107513, + "learning_rate": 0.00019271586574082393, + "loss": 0.9189, + "step": 1090 + }, + { + "epoch": 2.107795957651588, + "grad_norm": 0.1886579543352127, + "learning_rate": 0.0001925893454440147, + "loss": 0.9157, + "step": 1095 + }, + { + "epoch": 2.1174205967276225, + "grad_norm": 0.1850527673959732, + "learning_rate": 0.00019246177806537377, + "loss": 0.9173, + "step": 1100 + }, + { + "epoch": 2.1270452358036573, + "grad_norm": 0.19263537228107452, + "learning_rate": 0.00019233316504754523, + "loss": 0.9213, + "step": 1105 + }, + { + "epoch": 2.136669874879692, + "grad_norm": 0.18643324077129364, + "learning_rate": 0.00019220350784499837, + "loss": 0.9281, + "step": 1110 + }, + { + "epoch": 2.1462945139557266, + "grad_norm": 0.20145340263843536, + "learning_rate": 0.00019207280792401098, + "loss": 0.9154, + "step": 1115 + }, + { + "epoch": 2.1559191530317614, + "grad_norm": 0.20724299550056458, + "learning_rate": 0.00019194106676265283, + "loss": 0.9216, + "step": 1120 + }, + { + "epoch": 2.165543792107796, + "grad_norm": 0.20987021923065186, + "learning_rate": 0.000191808285850769, + "loss": 0.9191, + "step": 1125 + }, + { + "epoch": 2.1751684311838306, + "grad_norm": 0.19462813436985016, + "learning_rate": 0.00019167446668996285, + "loss": 0.9206, + "step": 1130 + }, + { + "epoch": 2.1847930702598655, + "grad_norm": 0.18060922622680664, + "learning_rate": 0.00019153961079357935, + "loss": 0.9194, + "step": 1135 + }, + { + "epoch": 2.1944177093359, + "grad_norm": 0.19130302965641022, + "learning_rate": 0.00019140371968668767, + "loss": 0.9209, + "step": 1140 + }, + { + "epoch": 2.2040423484119347, + "grad_norm": 0.1925574392080307, + "learning_rate": 0.00019126679490606404, + "loss": 0.915, + "step": 1145 + }, + { + "epoch": 2.213666987487969, + "grad_norm": 0.18374784290790558, + "learning_rate": 0.00019112883800017448, + "loss": 0.9266, + "step": 1150 + }, + { + "epoch": 2.223291626564004, + "grad_norm": 0.1928727775812149, + "learning_rate": 0.0001909898505291571, + "loss": 0.9177, + "step": 1155 + }, + { + "epoch": 2.2329162656400383, + "grad_norm": 0.19703041017055511, + "learning_rate": 0.00019084983406480462, + "loss": 0.9129, + "step": 1160 + }, + { + "epoch": 2.242540904716073, + "grad_norm": 0.19135095179080963, + "learning_rate": 0.00019070879019054645, + "loss": 0.9204, + "step": 1165 + }, + { + "epoch": 2.252165543792108, + "grad_norm": 0.18242081999778748, + "learning_rate": 0.00019056672050143087, + "loss": 0.9158, + "step": 1170 + }, + { + "epoch": 2.2617901828681424, + "grad_norm": 0.19838295876979828, + "learning_rate": 0.00019042362660410706, + "loss": 0.9282, + "step": 1175 + }, + { + "epoch": 2.271414821944177, + "grad_norm": 0.1942119151353836, + "learning_rate": 0.0001902795101168068, + "loss": 0.9224, + "step": 1180 + }, + { + "epoch": 2.2810394610202116, + "grad_norm": 0.1880965530872345, + "learning_rate": 0.00019013437266932615, + "loss": 0.919, + "step": 1185 + }, + { + "epoch": 2.2906641000962464, + "grad_norm": 0.18855926394462585, + "learning_rate": 0.00018998821590300713, + "loss": 0.9314, + "step": 1190 + }, + { + "epoch": 2.300288739172281, + "grad_norm": 0.20218202471733093, + "learning_rate": 0.00018984104147071917, + "loss": 0.9209, + "step": 1195 + }, + { + "epoch": 2.3099133782483157, + "grad_norm": 0.19384799897670746, + "learning_rate": 0.00018969285103684032, + "loss": 0.9147, + "step": 1200 + }, + { + "epoch": 2.3195380173243505, + "grad_norm": 0.1903255134820938, + "learning_rate": 0.00018954364627723843, + "loss": 0.9178, + "step": 1205 + }, + { + "epoch": 2.329162656400385, + "grad_norm": 0.180522158741951, + "learning_rate": 0.00018939342887925234, + "loss": 0.9215, + "step": 1210 + }, + { + "epoch": 2.3387872954764197, + "grad_norm": 0.1928156316280365, + "learning_rate": 0.00018924220054167257, + "loss": 0.9274, + "step": 1215 + }, + { + "epoch": 2.348411934552454, + "grad_norm": 0.19860059022903442, + "learning_rate": 0.00018908996297472235, + "loss": 0.9281, + "step": 1220 + }, + { + "epoch": 2.358036573628489, + "grad_norm": 0.19085602462291718, + "learning_rate": 0.00018893671790003804, + "loss": 0.9288, + "step": 1225 + }, + { + "epoch": 2.3676612127045233, + "grad_norm": 0.20947015285491943, + "learning_rate": 0.00018878246705064994, + "loss": 0.9245, + "step": 1230 + }, + { + "epoch": 2.377285851780558, + "grad_norm": 0.2144593894481659, + "learning_rate": 0.00018862721217096243, + "loss": 0.9122, + "step": 1235 + }, + { + "epoch": 2.386910490856593, + "grad_norm": 0.2063259780406952, + "learning_rate": 0.00018847095501673438, + "loss": 0.915, + "step": 1240 + }, + { + "epoch": 2.3965351299326274, + "grad_norm": 0.19159218668937683, + "learning_rate": 0.0001883136973550592, + "loss": 0.9172, + "step": 1245 + }, + { + "epoch": 2.4061597690086622, + "grad_norm": 0.19970135390758514, + "learning_rate": 0.00018815544096434503, + "loss": 0.9356, + "step": 1250 + }, + { + "epoch": 2.4157844080846966, + "grad_norm": 0.19337432086467743, + "learning_rate": 0.00018799618763429445, + "loss": 0.9284, + "step": 1255 + }, + { + "epoch": 2.4254090471607315, + "grad_norm": 0.19304610788822174, + "learning_rate": 0.00018783593916588432, + "loss": 0.9278, + "step": 1260 + }, + { + "epoch": 2.4350336862367663, + "grad_norm": 0.18972693383693695, + "learning_rate": 0.00018767469737134538, + "loss": 0.9251, + "step": 1265 + }, + { + "epoch": 2.4446583253128007, + "grad_norm": 0.19995278120040894, + "learning_rate": 0.0001875124640741418, + "loss": 0.9231, + "step": 1270 + }, + { + "epoch": 2.4542829643888355, + "grad_norm": 0.1899886578321457, + "learning_rate": 0.00018734924110895055, + "loss": 0.9289, + "step": 1275 + }, + { + "epoch": 2.46390760346487, + "grad_norm": 0.1865253895521164, + "learning_rate": 0.0001871850303216406, + "loss": 0.9141, + "step": 1280 + }, + { + "epoch": 2.4735322425409048, + "grad_norm": 0.205548956990242, + "learning_rate": 0.00018701983356925214, + "loss": 0.92, + "step": 1285 + }, + { + "epoch": 2.4831568816169396, + "grad_norm": 0.20036041736602783, + "learning_rate": 0.00018685365271997544, + "loss": 0.9274, + "step": 1290 + }, + { + "epoch": 2.492781520692974, + "grad_norm": 0.20605804026126862, + "learning_rate": 0.00018668648965312982, + "loss": 0.9262, + "step": 1295 + }, + { + "epoch": 2.502406159769009, + "grad_norm": 0.19875019788742065, + "learning_rate": 0.00018651834625914247, + "loss": 0.9288, + "step": 1300 + }, + { + "epoch": 2.512030798845043, + "grad_norm": 0.20208601653575897, + "learning_rate": 0.00018634922443952693, + "loss": 0.9246, + "step": 1305 + }, + { + "epoch": 2.521655437921078, + "grad_norm": 0.20923365652561188, + "learning_rate": 0.00018617912610686155, + "loss": 0.9285, + "step": 1310 + }, + { + "epoch": 2.5312800769971124, + "grad_norm": 0.21708457171916962, + "learning_rate": 0.00018600805318476807, + "loss": 0.9244, + "step": 1315 + }, + { + "epoch": 2.5409047160731473, + "grad_norm": 0.19935211539268494, + "learning_rate": 0.00018583600760788967, + "loss": 0.9261, + "step": 1320 + }, + { + "epoch": 2.550529355149182, + "grad_norm": 0.19352373480796814, + "learning_rate": 0.00018566299132186925, + "loss": 0.9203, + "step": 1325 + }, + { + "epoch": 2.5601539942252165, + "grad_norm": 0.18096321821212769, + "learning_rate": 0.00018548900628332726, + "loss": 0.924, + "step": 1330 + }, + { + "epoch": 2.5697786333012513, + "grad_norm": 0.20240572094917297, + "learning_rate": 0.0001853140544598397, + "loss": 0.9242, + "step": 1335 + }, + { + "epoch": 2.5794032723772857, + "grad_norm": 0.18877889215946198, + "learning_rate": 0.00018513813782991578, + "loss": 0.9101, + "step": 1340 + }, + { + "epoch": 2.5890279114533206, + "grad_norm": 0.1912551075220108, + "learning_rate": 0.00018496125838297572, + "loss": 0.9201, + "step": 1345 + }, + { + "epoch": 2.598652550529355, + "grad_norm": 0.19026340544223785, + "learning_rate": 0.0001847834181193279, + "loss": 0.9356, + "step": 1350 + }, + { + "epoch": 2.60827718960539, + "grad_norm": 0.19470341503620148, + "learning_rate": 0.00018460461905014664, + "loss": 0.9213, + "step": 1355 + }, + { + "epoch": 2.6179018286814246, + "grad_norm": 0.1977526694536209, + "learning_rate": 0.00018442486319744926, + "loss": 0.9292, + "step": 1360 + }, + { + "epoch": 2.627526467757459, + "grad_norm": 0.19127926230430603, + "learning_rate": 0.00018424415259407317, + "loss": 0.9283, + "step": 1365 + }, + { + "epoch": 2.637151106833494, + "grad_norm": 0.18255840241909027, + "learning_rate": 0.00018406248928365295, + "loss": 0.9179, + "step": 1370 + }, + { + "epoch": 2.6467757459095282, + "grad_norm": 0.18344487249851227, + "learning_rate": 0.00018387987532059725, + "loss": 0.9397, + "step": 1375 + }, + { + "epoch": 2.656400384985563, + "grad_norm": 0.1913861185312271, + "learning_rate": 0.00018369631277006555, + "loss": 0.9248, + "step": 1380 + }, + { + "epoch": 2.6660250240615975, + "grad_norm": 0.1795121282339096, + "learning_rate": 0.00018351180370794479, + "loss": 0.9223, + "step": 1385 + }, + { + "epoch": 2.6756496631376323, + "grad_norm": 0.19478866457939148, + "learning_rate": 0.00018332635022082582, + "loss": 0.9282, + "step": 1390 + }, + { + "epoch": 2.685274302213667, + "grad_norm": 0.1917424350976944, + "learning_rate": 0.00018313995440598002, + "loss": 0.9228, + "step": 1395 + }, + { + "epoch": 2.6948989412897015, + "grad_norm": 0.18964500725269318, + "learning_rate": 0.00018295261837133532, + "loss": 0.928, + "step": 1400 + }, + { + "epoch": 2.7045235803657364, + "grad_norm": 0.19044145941734314, + "learning_rate": 0.00018276434423545253, + "loss": 0.926, + "step": 1405 + }, + { + "epoch": 2.714148219441771, + "grad_norm": 0.1876942664384842, + "learning_rate": 0.0001825751341275013, + "loss": 0.9224, + "step": 1410 + }, + { + "epoch": 2.7237728585178056, + "grad_norm": 0.19307979941368103, + "learning_rate": 0.00018238499018723614, + "loss": 0.9322, + "step": 1415 + }, + { + "epoch": 2.73339749759384, + "grad_norm": 0.1879437267780304, + "learning_rate": 0.00018219391456497216, + "loss": 0.9104, + "step": 1420 + }, + { + "epoch": 2.743022136669875, + "grad_norm": 0.2002253383398056, + "learning_rate": 0.00018200190942156062, + "loss": 0.9266, + "step": 1425 + }, + { + "epoch": 2.7526467757459097, + "grad_norm": 0.1822872757911682, + "learning_rate": 0.00018180897692836483, + "loss": 0.9245, + "step": 1430 + }, + { + "epoch": 2.762271414821944, + "grad_norm": 0.1884424090385437, + "learning_rate": 0.0001816151192672352, + "loss": 0.9273, + "step": 1435 + }, + { + "epoch": 2.771896053897979, + "grad_norm": 0.1969207227230072, + "learning_rate": 0.00018142033863048485, + "loss": 0.929, + "step": 1440 + }, + { + "epoch": 2.7815206929740137, + "grad_norm": 0.1919521689414978, + "learning_rate": 0.0001812246372208647, + "loss": 0.9213, + "step": 1445 + }, + { + "epoch": 2.791145332050048, + "grad_norm": 0.18795301020145416, + "learning_rate": 0.00018102801725153862, + "loss": 0.9281, + "step": 1450 + }, + { + "epoch": 2.8007699711260825, + "grad_norm": 0.19035767018795013, + "learning_rate": 0.00018083048094605825, + "loss": 0.9264, + "step": 1455 + }, + { + "epoch": 2.8103946102021173, + "grad_norm": 0.181080624461174, + "learning_rate": 0.0001806320305383381, + "loss": 0.926, + "step": 1460 + }, + { + "epoch": 2.820019249278152, + "grad_norm": 0.18840213119983673, + "learning_rate": 0.00018043266827263003, + "loss": 0.9327, + "step": 1465 + }, + { + "epoch": 2.8296438883541866, + "grad_norm": 0.18549908697605133, + "learning_rate": 0.0001802323964034981, + "loss": 0.9345, + "step": 1470 + }, + { + "epoch": 2.8392685274302214, + "grad_norm": 0.18507707118988037, + "learning_rate": 0.00018003121719579294, + "loss": 0.9243, + "step": 1475 + }, + { + "epoch": 2.8488931665062562, + "grad_norm": 0.19053645431995392, + "learning_rate": 0.0001798291329246261, + "loss": 0.9136, + "step": 1480 + }, + { + "epoch": 2.8585178055822906, + "grad_norm": 0.18798498809337616, + "learning_rate": 0.00017962614587534444, + "loss": 0.9296, + "step": 1485 + }, + { + "epoch": 2.8681424446583255, + "grad_norm": 0.19244647026062012, + "learning_rate": 0.00017942225834350424, + "loss": 0.9212, + "step": 1490 + }, + { + "epoch": 2.87776708373436, + "grad_norm": 0.18958385288715363, + "learning_rate": 0.00017921747263484518, + "loss": 0.9204, + "step": 1495 + }, + { + "epoch": 2.8873917228103947, + "grad_norm": 0.1872030794620514, + "learning_rate": 0.00017901179106526434, + "loss": 0.9167, + "step": 1500 + }, + { + "epoch": 2.897016361886429, + "grad_norm": 0.1842317432165146, + "learning_rate": 0.00017880521596079003, + "loss": 0.9295, + "step": 1505 + }, + { + "epoch": 2.906641000962464, + "grad_norm": 0.1908566802740097, + "learning_rate": 0.00017859774965755534, + "loss": 0.933, + "step": 1510 + }, + { + "epoch": 2.9162656400384988, + "grad_norm": 0.17877928912639618, + "learning_rate": 0.0001783893945017719, + "loss": 0.9209, + "step": 1515 + }, + { + "epoch": 2.925890279114533, + "grad_norm": 0.19019804894924164, + "learning_rate": 0.00017818015284970328, + "loss": 0.9298, + "step": 1520 + }, + { + "epoch": 2.935514918190568, + "grad_norm": 0.17898397147655487, + "learning_rate": 0.0001779700270676382, + "loss": 0.9149, + "step": 1525 + }, + { + "epoch": 2.9451395572666024, + "grad_norm": 0.19317851960659027, + "learning_rate": 0.0001777590195318641, + "loss": 0.9268, + "step": 1530 + }, + { + "epoch": 2.954764196342637, + "grad_norm": 0.1835252344608307, + "learning_rate": 0.00017754713262863985, + "loss": 0.9156, + "step": 1535 + }, + { + "epoch": 2.9643888354186716, + "grad_norm": 0.18219447135925293, + "learning_rate": 0.00017733436875416917, + "loss": 0.928, + "step": 1540 + }, + { + "epoch": 2.9740134744947064, + "grad_norm": 0.19455976784229279, + "learning_rate": 0.00017712073031457331, + "loss": 0.9358, + "step": 1545 + }, + { + "epoch": 2.9836381135707413, + "grad_norm": 0.19101083278656006, + "learning_rate": 0.0001769062197258637, + "loss": 0.919, + "step": 1550 + }, + { + "epoch": 2.9932627526467757, + "grad_norm": 0.1850951611995697, + "learning_rate": 0.00017669083941391502, + "loss": 0.9289, + "step": 1555 + }, + { + "epoch": 2.9990375360923966, + "eval_loss": 2.251723289489746, + "eval_runtime": 0.7901, + "eval_samples_per_second": 13.922, + "eval_steps_per_second": 2.531, + "step": 1558 + }, + { + "epoch": 3.0028873917228105, + "grad_norm": 0.1839417815208435, + "learning_rate": 0.00017647459181443739, + "loss": 0.9099, + "step": 1560 + }, + { + "epoch": 3.012512030798845, + "grad_norm": 0.21318542957305908, + "learning_rate": 0.0001762574793729491, + "loss": 0.8622, + "step": 1565 + }, + { + "epoch": 3.0221366698748797, + "grad_norm": 0.20732618868350983, + "learning_rate": 0.00017603950454474877, + "loss": 0.8502, + "step": 1570 + }, + { + "epoch": 3.0317613089509146, + "grad_norm": 0.20737336575984955, + "learning_rate": 0.00017582066979488764, + "loss": 0.8479, + "step": 1575 + }, + { + "epoch": 3.041385948026949, + "grad_norm": 0.2138897329568863, + "learning_rate": 0.00017560097759814172, + "loss": 0.8517, + "step": 1580 + }, + { + "epoch": 3.051010587102984, + "grad_norm": 0.20526482164859772, + "learning_rate": 0.00017538043043898376, + "loss": 0.8548, + "step": 1585 + }, + { + "epoch": 3.060635226179018, + "grad_norm": 0.21120765805244446, + "learning_rate": 0.00017515903081155525, + "loss": 0.8531, + "step": 1590 + }, + { + "epoch": 3.070259865255053, + "grad_norm": 0.20420415699481964, + "learning_rate": 0.00017493678121963807, + "loss": 0.8607, + "step": 1595 + }, + { + "epoch": 3.0798845043310874, + "grad_norm": 0.2265135943889618, + "learning_rate": 0.00017471368417662627, + "loss": 0.8638, + "step": 1600 + }, + { + "epoch": 3.0895091434071222, + "grad_norm": 0.2099863588809967, + "learning_rate": 0.00017448974220549764, + "loss": 0.8648, + "step": 1605 + }, + { + "epoch": 3.099133782483157, + "grad_norm": 0.2183115929365158, + "learning_rate": 0.00017426495783878508, + "loss": 0.8554, + "step": 1610 + }, + { + "epoch": 3.1087584215591915, + "grad_norm": 0.2061695158481598, + "learning_rate": 0.00017403933361854814, + "loss": 0.8561, + "step": 1615 + }, + { + "epoch": 3.1183830606352263, + "grad_norm": 0.21093107759952545, + "learning_rate": 0.0001738128720963442, + "loss": 0.8639, + "step": 1620 + }, + { + "epoch": 3.1280076997112607, + "grad_norm": 0.22155196964740753, + "learning_rate": 0.0001735855758331994, + "loss": 0.8687, + "step": 1625 + }, + { + "epoch": 3.1376323387872955, + "grad_norm": 0.21988868713378906, + "learning_rate": 0.0001733574473995801, + "loss": 0.8571, + "step": 1630 + }, + { + "epoch": 3.14725697786333, + "grad_norm": 0.20397303998470306, + "learning_rate": 0.00017312848937536338, + "loss": 0.8556, + "step": 1635 + }, + { + "epoch": 3.1568816169393648, + "grad_norm": 0.21777671575546265, + "learning_rate": 0.00017289870434980824, + "loss": 0.8657, + "step": 1640 + }, + { + "epoch": 3.1665062560153996, + "grad_norm": 0.20753996074199677, + "learning_rate": 0.00017266809492152597, + "loss": 0.8578, + "step": 1645 + }, + { + "epoch": 3.176130895091434, + "grad_norm": 0.22726857662200928, + "learning_rate": 0.00017243666369845103, + "loss": 0.8713, + "step": 1650 + }, + { + "epoch": 3.185755534167469, + "grad_norm": 0.20830857753753662, + "learning_rate": 0.00017220441329781147, + "loss": 0.8621, + "step": 1655 + }, + { + "epoch": 3.195380173243503, + "grad_norm": 0.21678543090820312, + "learning_rate": 0.00017197134634609924, + "loss": 0.8589, + "step": 1660 + }, + { + "epoch": 3.205004812319538, + "grad_norm": 0.21865533292293549, + "learning_rate": 0.00017173746547904063, + "loss": 0.872, + "step": 1665 + }, + { + "epoch": 3.214629451395573, + "grad_norm": 0.20973502099514008, + "learning_rate": 0.0001715027733415664, + "loss": 0.8624, + "step": 1670 + }, + { + "epoch": 3.2242540904716073, + "grad_norm": 0.21278487145900726, + "learning_rate": 0.00017126727258778187, + "loss": 0.8693, + "step": 1675 + }, + { + "epoch": 3.233878729547642, + "grad_norm": 0.2145373523235321, + "learning_rate": 0.00017103096588093686, + "loss": 0.8665, + "step": 1680 + }, + { + "epoch": 3.2435033686236765, + "grad_norm": 0.21175837516784668, + "learning_rate": 0.00017079385589339568, + "loss": 0.8592, + "step": 1685 + }, + { + "epoch": 3.2531280076997113, + "grad_norm": 0.21969176828861237, + "learning_rate": 0.00017055594530660678, + "loss": 0.8686, + "step": 1690 + }, + { + "epoch": 3.2627526467757457, + "grad_norm": 0.23275814950466156, + "learning_rate": 0.00017031723681107256, + "loss": 0.8643, + "step": 1695 + }, + { + "epoch": 3.2723772858517806, + "grad_norm": 0.22712193429470062, + "learning_rate": 0.0001700777331063188, + "loss": 0.8774, + "step": 1700 + }, + { + "epoch": 3.2820019249278154, + "grad_norm": 0.2357400804758072, + "learning_rate": 0.0001698374369008643, + "loss": 0.8654, + "step": 1705 + }, + { + "epoch": 3.29162656400385, + "grad_norm": 0.21586911380290985, + "learning_rate": 0.00016959635091219011, + "loss": 0.8682, + "step": 1710 + }, + { + "epoch": 3.3012512030798846, + "grad_norm": 0.20854496955871582, + "learning_rate": 0.00016935447786670875, + "loss": 0.872, + "step": 1715 + }, + { + "epoch": 3.310875842155919, + "grad_norm": 0.22415196895599365, + "learning_rate": 0.00016911182049973364, + "loss": 0.8691, + "step": 1720 + }, + { + "epoch": 3.320500481231954, + "grad_norm": 0.21514172852039337, + "learning_rate": 0.00016886838155544785, + "loss": 0.8662, + "step": 1725 + }, + { + "epoch": 3.3301251203079882, + "grad_norm": 0.21508009731769562, + "learning_rate": 0.0001686241637868734, + "loss": 0.8677, + "step": 1730 + }, + { + "epoch": 3.339749759384023, + "grad_norm": 0.21434170007705688, + "learning_rate": 0.00016837916995583965, + "loss": 0.8691, + "step": 1735 + }, + { + "epoch": 3.349374398460058, + "grad_norm": 0.21920685470104218, + "learning_rate": 0.00016813340283295265, + "loss": 0.8632, + "step": 1740 + }, + { + "epoch": 3.3589990375360923, + "grad_norm": 0.20799002051353455, + "learning_rate": 0.00016788686519756337, + "loss": 0.8711, + "step": 1745 + }, + { + "epoch": 3.368623676612127, + "grad_norm": 0.22760187089443207, + "learning_rate": 0.00016763955983773642, + "loss": 0.8716, + "step": 1750 + }, + { + "epoch": 3.3782483156881615, + "grad_norm": 0.20473913848400116, + "learning_rate": 0.00016739148955021853, + "loss": 0.8672, + "step": 1755 + }, + { + "epoch": 3.3878729547641964, + "grad_norm": 0.2237493246793747, + "learning_rate": 0.00016714265714040688, + "loss": 0.8711, + "step": 1760 + }, + { + "epoch": 3.3974975938402308, + "grad_norm": 0.21266481280326843, + "learning_rate": 0.00016689306542231754, + "loss": 0.8581, + "step": 1765 + }, + { + "epoch": 3.4071222329162656, + "grad_norm": 0.21926787495613098, + "learning_rate": 0.00016664271721855323, + "loss": 0.8647, + "step": 1770 + }, + { + "epoch": 3.4167468719923004, + "grad_norm": 0.21556758880615234, + "learning_rate": 0.00016639161536027196, + "loss": 0.8627, + "step": 1775 + }, + { + "epoch": 3.426371511068335, + "grad_norm": 0.22477813065052032, + "learning_rate": 0.00016613976268715458, + "loss": 0.8734, + "step": 1780 + }, + { + "epoch": 3.4359961501443697, + "grad_norm": 0.22144025564193726, + "learning_rate": 0.00016588716204737281, + "loss": 0.8633, + "step": 1785 + }, + { + "epoch": 3.445620789220404, + "grad_norm": 0.21546606719493866, + "learning_rate": 0.00016563381629755713, + "loss": 0.87, + "step": 1790 + }, + { + "epoch": 3.455245428296439, + "grad_norm": 0.21200338006019592, + "learning_rate": 0.00016537972830276424, + "loss": 0.8749, + "step": 1795 + }, + { + "epoch": 3.4648700673724737, + "grad_norm": 0.21702003479003906, + "learning_rate": 0.00016512490093644491, + "loss": 0.8736, + "step": 1800 + }, + { + "epoch": 3.474494706448508, + "grad_norm": 0.20890291035175323, + "learning_rate": 0.00016486933708041138, + "loss": 0.8658, + "step": 1805 + }, + { + "epoch": 3.484119345524543, + "grad_norm": 0.21432092785835266, + "learning_rate": 0.0001646130396248047, + "loss": 0.8671, + "step": 1810 + }, + { + "epoch": 3.4937439846005773, + "grad_norm": 0.21486730873584747, + "learning_rate": 0.0001643560114680621, + "loss": 0.8624, + "step": 1815 + }, + { + "epoch": 3.503368623676612, + "grad_norm": 0.2079630345106125, + "learning_rate": 0.0001640982555168843, + "loss": 0.8623, + "step": 1820 + }, + { + "epoch": 3.512993262752647, + "grad_norm": 0.21051821112632751, + "learning_rate": 0.00016383977468620252, + "loss": 0.8694, + "step": 1825 + }, + { + "epoch": 3.5226179018286814, + "grad_norm": 0.22331751883029938, + "learning_rate": 0.00016358057189914553, + "loss": 0.8867, + "step": 1830 + }, + { + "epoch": 3.5322425409047162, + "grad_norm": 0.21272289752960205, + "learning_rate": 0.00016332065008700666, + "loss": 0.8643, + "step": 1835 + }, + { + "epoch": 3.5418671799807506, + "grad_norm": 0.2075881063938141, + "learning_rate": 0.00016306001218921055, + "loss": 0.8758, + "step": 1840 + }, + { + "epoch": 3.5514918190567855, + "grad_norm": 0.21468383073806763, + "learning_rate": 0.00016279866115328012, + "loss": 0.8743, + "step": 1845 + }, + { + "epoch": 3.56111645813282, + "grad_norm": 0.20136167109012604, + "learning_rate": 0.00016253659993480284, + "loss": 0.874, + "step": 1850 + }, + { + "epoch": 3.5707410972088547, + "grad_norm": 0.2094564139842987, + "learning_rate": 0.00016227383149739776, + "loss": 0.8798, + "step": 1855 + }, + { + "epoch": 3.5803657362848895, + "grad_norm": 0.21963797509670258, + "learning_rate": 0.00016201035881268166, + "loss": 0.8751, + "step": 1860 + }, + { + "epoch": 3.589990375360924, + "grad_norm": 0.22210368514060974, + "learning_rate": 0.00016174618486023565, + "loss": 0.8709, + "step": 1865 + }, + { + "epoch": 3.5996150144369587, + "grad_norm": 0.22002506256103516, + "learning_rate": 0.00016148131262757134, + "loss": 0.8724, + "step": 1870 + }, + { + "epoch": 3.609239653512993, + "grad_norm": 0.21862515807151794, + "learning_rate": 0.0001612157451100971, + "loss": 0.8715, + "step": 1875 + }, + { + "epoch": 3.618864292589028, + "grad_norm": 0.21481823921203613, + "learning_rate": 0.0001609494853110843, + "loss": 0.8727, + "step": 1880 + }, + { + "epoch": 3.6284889316650624, + "grad_norm": 0.21671965718269348, + "learning_rate": 0.00016068253624163307, + "loss": 0.8695, + "step": 1885 + }, + { + "epoch": 3.638113570741097, + "grad_norm": 0.22262564301490784, + "learning_rate": 0.00016041490092063852, + "loss": 0.8707, + "step": 1890 + }, + { + "epoch": 3.647738209817132, + "grad_norm": 0.21777838468551636, + "learning_rate": 0.0001601465823747565, + "loss": 0.8719, + "step": 1895 + }, + { + "epoch": 3.6573628488931664, + "grad_norm": 0.2157593071460724, + "learning_rate": 0.00015987758363836932, + "loss": 0.8649, + "step": 1900 + }, + { + "epoch": 3.6669874879692013, + "grad_norm": 0.21907728910446167, + "learning_rate": 0.00015960790775355159, + "loss": 0.8727, + "step": 1905 + }, + { + "epoch": 3.6766121270452357, + "grad_norm": 0.2181127518415451, + "learning_rate": 0.00015933755777003552, + "loss": 0.8642, + "step": 1910 + }, + { + "epoch": 3.6862367661212705, + "grad_norm": 0.21002036333084106, + "learning_rate": 0.0001590665367451768, + "loss": 0.8853, + "step": 1915 + }, + { + "epoch": 3.695861405197305, + "grad_norm": 0.21628259122371674, + "learning_rate": 0.0001587948477439198, + "loss": 0.8781, + "step": 1920 + }, + { + "epoch": 3.7054860442733397, + "grad_norm": 0.21025903522968292, + "learning_rate": 0.00015852249383876285, + "loss": 0.8788, + "step": 1925 + }, + { + "epoch": 3.7151106833493746, + "grad_norm": 0.21036125719547272, + "learning_rate": 0.00015824947810972378, + "loss": 0.8769, + "step": 1930 + }, + { + "epoch": 3.724735322425409, + "grad_norm": 0.20949947834014893, + "learning_rate": 0.00015797580364430473, + "loss": 0.8689, + "step": 1935 + }, + { + "epoch": 3.734359961501444, + "grad_norm": 0.22593073546886444, + "learning_rate": 0.00015770147353745754, + "loss": 0.8763, + "step": 1940 + }, + { + "epoch": 3.7439846005774786, + "grad_norm": 0.22361914813518524, + "learning_rate": 0.00015742649089154858, + "loss": 0.8743, + "step": 1945 + }, + { + "epoch": 3.753609239653513, + "grad_norm": 0.21210341155529022, + "learning_rate": 0.00015715085881632366, + "loss": 0.8754, + "step": 1950 + }, + { + "epoch": 3.7632338787295474, + "grad_norm": 0.21233123540878296, + "learning_rate": 0.00015687458042887298, + "loss": 0.8823, + "step": 1955 + }, + { + "epoch": 3.7728585178055822, + "grad_norm": 0.20900115370750427, + "learning_rate": 0.00015659765885359572, + "loss": 0.8601, + "step": 1960 + }, + { + "epoch": 3.782483156881617, + "grad_norm": 0.20850348472595215, + "learning_rate": 0.0001563200972221649, + "loss": 0.8748, + "step": 1965 + }, + { + "epoch": 3.7921077959576515, + "grad_norm": 0.2235669642686844, + "learning_rate": 0.00015604189867349182, + "loss": 0.8767, + "step": 1970 + }, + { + "epoch": 3.8017324350336863, + "grad_norm": 0.20681613683700562, + "learning_rate": 0.00015576306635369053, + "loss": 0.87, + "step": 1975 + }, + { + "epoch": 3.811357074109721, + "grad_norm": 0.2126859724521637, + "learning_rate": 0.00015548360341604244, + "loss": 0.8767, + "step": 1980 + }, + { + "epoch": 3.8209817131857555, + "grad_norm": 0.21969568729400635, + "learning_rate": 0.00015520351302096043, + "loss": 0.8619, + "step": 1985 + }, + { + "epoch": 3.83060635226179, + "grad_norm": 0.20034681260585785, + "learning_rate": 0.0001549227983359533, + "loss": 0.879, + "step": 1990 + }, + { + "epoch": 3.8402309913378248, + "grad_norm": 0.22048155963420868, + "learning_rate": 0.00015464146253558987, + "loss": 0.8704, + "step": 1995 + }, + { + "epoch": 3.8498556304138596, + "grad_norm": 0.2217637300491333, + "learning_rate": 0.00015435950880146297, + "loss": 0.874, + "step": 2000 + }, + { + "epoch": 3.859480269489894, + "grad_norm": 0.2207387238740921, + "learning_rate": 0.00015407694032215375, + "loss": 0.871, + "step": 2005 + }, + { + "epoch": 3.869104908565929, + "grad_norm": 0.21759381890296936, + "learning_rate": 0.00015379376029319526, + "loss": 0.881, + "step": 2010 + }, + { + "epoch": 3.8787295476419636, + "grad_norm": 0.21979306638240814, + "learning_rate": 0.00015350997191703662, + "loss": 0.8707, + "step": 2015 + }, + { + "epoch": 3.888354186717998, + "grad_norm": 0.2088766098022461, + "learning_rate": 0.0001532255784030066, + "loss": 0.8715, + "step": 2020 + }, + { + "epoch": 3.897978825794033, + "grad_norm": 0.23208843171596527, + "learning_rate": 0.00015294058296727746, + "loss": 0.867, + "step": 2025 + }, + { + "epoch": 3.9076034648700673, + "grad_norm": 0.211493119597435, + "learning_rate": 0.00015265498883282848, + "loss": 0.8746, + "step": 2030 + }, + { + "epoch": 3.917228103946102, + "grad_norm": 0.2072470784187317, + "learning_rate": 0.00015236879922940952, + "loss": 0.8815, + "step": 2035 + }, + { + "epoch": 3.9268527430221365, + "grad_norm": 0.2107774019241333, + "learning_rate": 0.0001520820173935046, + "loss": 0.8762, + "step": 2040 + }, + { + "epoch": 3.9364773820981713, + "grad_norm": 0.22592873871326447, + "learning_rate": 0.00015179464656829526, + "loss": 0.8781, + "step": 2045 + }, + { + "epoch": 3.946102021174206, + "grad_norm": 0.210884690284729, + "learning_rate": 0.00015150669000362372, + "loss": 0.8759, + "step": 2050 + }, + { + "epoch": 3.9557266602502406, + "grad_norm": 0.22325028479099274, + "learning_rate": 0.00015121815095595631, + "loss": 0.8759, + "step": 2055 + }, + { + "epoch": 3.9653512993262754, + "grad_norm": 0.20822718739509583, + "learning_rate": 0.0001509290326883466, + "loss": 0.8743, + "step": 2060 + }, + { + "epoch": 3.97497593840231, + "grad_norm": 0.22340907156467438, + "learning_rate": 0.00015063933847039856, + "loss": 0.8768, + "step": 2065 + }, + { + "epoch": 3.9846005774783446, + "grad_norm": 0.21545882523059845, + "learning_rate": 0.0001503490715782294, + "loss": 0.8737, + "step": 2070 + }, + { + "epoch": 3.994225216554379, + "grad_norm": 0.21250423789024353, + "learning_rate": 0.00015005823529443268, + "loss": 0.8818, + "step": 2075 + }, + { + "epoch": 4.0, + "eval_loss": 2.463193893432617, + "eval_runtime": 0.7794, + "eval_samples_per_second": 14.113, + "eval_steps_per_second": 2.566, + "step": 2078 + }, + { + "epoch": 4.003849855630414, + "grad_norm": 0.20480164885520935, + "learning_rate": 0.00014976683290804116, + "loss": 0.8452, + "step": 2080 + }, + { + "epoch": 4.013474494706449, + "grad_norm": 0.24909119307994843, + "learning_rate": 0.00014947486771448956, + "loss": 0.799, + "step": 2085 + }, + { + "epoch": 4.023099133782483, + "grad_norm": 0.2511972188949585, + "learning_rate": 0.00014918234301557732, + "loss": 0.7996, + "step": 2090 + }, + { + "epoch": 4.0327237728585175, + "grad_norm": 0.2290249615907669, + "learning_rate": 0.00014888926211943128, + "loss": 0.7821, + "step": 2095 + }, + { + "epoch": 4.042348411934553, + "grad_norm": 0.2516845762729645, + "learning_rate": 0.0001485956283404682, + "loss": 0.8027, + "step": 2100 + }, + { + "epoch": 4.051973051010587, + "grad_norm": 0.23257547616958618, + "learning_rate": 0.00014830144499935742, + "loss": 0.8051, + "step": 2105 + }, + { + "epoch": 4.0615976900866215, + "grad_norm": 0.23365622758865356, + "learning_rate": 0.00014800671542298312, + "loss": 0.8046, + "step": 2110 + }, + { + "epoch": 4.071222329162657, + "grad_norm": 0.25398579239845276, + "learning_rate": 0.00014771144294440682, + "loss": 0.7998, + "step": 2115 + }, + { + "epoch": 4.080846968238691, + "grad_norm": 0.25395774841308594, + "learning_rate": 0.00014741563090282965, + "loss": 0.7896, + "step": 2120 + }, + { + "epoch": 4.090471607314726, + "grad_norm": 0.23397642374038696, + "learning_rate": 0.00014711928264355466, + "loss": 0.7982, + "step": 2125 + }, + { + "epoch": 4.10009624639076, + "grad_norm": 0.24863800406455994, + "learning_rate": 0.0001468224015179488, + "loss": 0.8004, + "step": 2130 + }, + { + "epoch": 4.109720885466795, + "grad_norm": 0.24272161722183228, + "learning_rate": 0.00014652499088340523, + "loss": 0.7949, + "step": 2135 + }, + { + "epoch": 4.11934552454283, + "grad_norm": 0.24697747826576233, + "learning_rate": 0.00014622705410330522, + "loss": 0.792, + "step": 2140 + }, + { + "epoch": 4.128970163618864, + "grad_norm": 0.2412373572587967, + "learning_rate": 0.0001459285945469802, + "loss": 0.7999, + "step": 2145 + }, + { + "epoch": 4.138594802694899, + "grad_norm": 0.255993127822876, + "learning_rate": 0.0001456296155896736, + "loss": 0.7965, + "step": 2150 + }, + { + "epoch": 4.148219441770934, + "grad_norm": 0.23746897280216217, + "learning_rate": 0.00014533012061250264, + "loss": 0.8057, + "step": 2155 + }, + { + "epoch": 4.157844080846968, + "grad_norm": 0.24358995258808136, + "learning_rate": 0.00014503011300242023, + "loss": 0.8073, + "step": 2160 + }, + { + "epoch": 4.1674687199230025, + "grad_norm": 0.2651350498199463, + "learning_rate": 0.0001447295961521765, + "loss": 0.7961, + "step": 2165 + }, + { + "epoch": 4.177093358999038, + "grad_norm": 0.25750139355659485, + "learning_rate": 0.0001444285734602806, + "loss": 0.7961, + "step": 2170 + }, + { + "epoch": 4.186717998075072, + "grad_norm": 0.2478739321231842, + "learning_rate": 0.00014412704833096217, + "loss": 0.7955, + "step": 2175 + }, + { + "epoch": 4.196342637151107, + "grad_norm": 0.26254719495773315, + "learning_rate": 0.00014382502417413276, + "loss": 0.7929, + "step": 2180 + }, + { + "epoch": 4.205967276227142, + "grad_norm": 0.25435176491737366, + "learning_rate": 0.00014352250440534747, + "loss": 0.8052, + "step": 2185 + }, + { + "epoch": 4.215591915303176, + "grad_norm": 0.24811629951000214, + "learning_rate": 0.00014321949244576617, + "loss": 0.7989, + "step": 2190 + }, + { + "epoch": 4.225216554379211, + "grad_norm": 0.2621951103210449, + "learning_rate": 0.00014291599172211485, + "loss": 0.8092, + "step": 2195 + }, + { + "epoch": 4.234841193455245, + "grad_norm": 0.2780658006668091, + "learning_rate": 0.0001426120056666469, + "loss": 0.8058, + "step": 2200 + }, + { + "epoch": 4.24446583253128, + "grad_norm": 0.233393132686615, + "learning_rate": 0.0001423075377171043, + "loss": 0.8049, + "step": 2205 + }, + { + "epoch": 4.254090471607315, + "grad_norm": 0.26003360748291016, + "learning_rate": 0.00014200259131667858, + "loss": 0.8091, + "step": 2210 + }, + { + "epoch": 4.263715110683349, + "grad_norm": 0.25277137756347656, + "learning_rate": 0.00014169716991397214, + "loss": 0.8126, + "step": 2215 + }, + { + "epoch": 4.273339749759384, + "grad_norm": 0.23928789794445038, + "learning_rate": 0.00014139127696295912, + "loss": 0.8044, + "step": 2220 + }, + { + "epoch": 4.282964388835419, + "grad_norm": 0.254984587430954, + "learning_rate": 0.00014108491592294627, + "loss": 0.8036, + "step": 2225 + }, + { + "epoch": 4.292589027911453, + "grad_norm": 0.2602671682834625, + "learning_rate": 0.000140778090258534, + "loss": 0.8147, + "step": 2230 + }, + { + "epoch": 4.3022136669874875, + "grad_norm": 0.24539902806282043, + "learning_rate": 0.000140470803439577, + "loss": 0.8078, + "step": 2235 + }, + { + "epoch": 4.311838306063523, + "grad_norm": 0.24983367323875427, + "learning_rate": 0.00014016305894114516, + "loss": 0.8089, + "step": 2240 + }, + { + "epoch": 4.321462945139557, + "grad_norm": 0.2500509023666382, + "learning_rate": 0.0001398548602434842, + "loss": 0.8053, + "step": 2245 + }, + { + "epoch": 4.331087584215592, + "grad_norm": 0.24786844849586487, + "learning_rate": 0.00013954621083197628, + "loss": 0.8091, + "step": 2250 + }, + { + "epoch": 4.340712223291627, + "grad_norm": 0.2504083216190338, + "learning_rate": 0.00013923711419710076, + "loss": 0.8122, + "step": 2255 + }, + { + "epoch": 4.350336862367661, + "grad_norm": 0.24594616889953613, + "learning_rate": 0.0001389275738343944, + "loss": 0.8142, + "step": 2260 + }, + { + "epoch": 4.359961501443696, + "grad_norm": 0.25497034192085266, + "learning_rate": 0.00013861759324441223, + "loss": 0.8102, + "step": 2265 + }, + { + "epoch": 4.369586140519731, + "grad_norm": 0.26248982548713684, + "learning_rate": 0.00013830717593268764, + "loss": 0.8106, + "step": 2270 + }, + { + "epoch": 4.379210779595765, + "grad_norm": 0.24808135628700256, + "learning_rate": 0.00013799632540969286, + "loss": 0.8069, + "step": 2275 + }, + { + "epoch": 4.3888354186718, + "grad_norm": 0.2534014582633972, + "learning_rate": 0.00013768504519079923, + "loss": 0.8166, + "step": 2280 + }, + { + "epoch": 4.398460057747834, + "grad_norm": 0.24292294681072235, + "learning_rate": 0.0001373733387962376, + "loss": 0.8072, + "step": 2285 + }, + { + "epoch": 4.408084696823869, + "grad_norm": 0.24815544486045837, + "learning_rate": 0.00013706120975105822, + "loss": 0.8189, + "step": 2290 + }, + { + "epoch": 4.417709335899904, + "grad_norm": 0.24199172854423523, + "learning_rate": 0.00013674866158509117, + "loss": 0.8084, + "step": 2295 + }, + { + "epoch": 4.427333974975938, + "grad_norm": 0.26282939314842224, + "learning_rate": 0.00013643569783290622, + "loss": 0.8103, + "step": 2300 + }, + { + "epoch": 4.4369586140519734, + "grad_norm": 0.2644505202770233, + "learning_rate": 0.00013612232203377307, + "loss": 0.8106, + "step": 2305 + }, + { + "epoch": 4.446583253128008, + "grad_norm": 0.250636488199234, + "learning_rate": 0.0001358085377316211, + "loss": 0.823, + "step": 2310 + }, + { + "epoch": 4.456207892204042, + "grad_norm": 0.2760376036167145, + "learning_rate": 0.00013549434847499945, + "loss": 0.8109, + "step": 2315 + }, + { + "epoch": 4.465832531280077, + "grad_norm": 0.24669407308101654, + "learning_rate": 0.00013517975781703688, + "loss": 0.8135, + "step": 2320 + }, + { + "epoch": 4.475457170356112, + "grad_norm": 0.24369503557682037, + "learning_rate": 0.00013486476931540145, + "loss": 0.8083, + "step": 2325 + }, + { + "epoch": 4.485081809432146, + "grad_norm": 0.2656605839729309, + "learning_rate": 0.00013454938653226047, + "loss": 0.8082, + "step": 2330 + }, + { + "epoch": 4.494706448508181, + "grad_norm": 0.24139179289340973, + "learning_rate": 0.0001342336130342401, + "loss": 0.8046, + "step": 2335 + }, + { + "epoch": 4.504331087584216, + "grad_norm": 0.2464561015367508, + "learning_rate": 0.00013391745239238508, + "loss": 0.8205, + "step": 2340 + }, + { + "epoch": 4.51395572666025, + "grad_norm": 0.25290533900260925, + "learning_rate": 0.0001336009081821183, + "loss": 0.8135, + "step": 2345 + }, + { + "epoch": 4.523580365736285, + "grad_norm": 0.2681277096271515, + "learning_rate": 0.00013328398398320036, + "loss": 0.8111, + "step": 2350 + }, + { + "epoch": 4.53320500481232, + "grad_norm": 0.24826329946517944, + "learning_rate": 0.00013296668337968903, + "loss": 0.8161, + "step": 2355 + }, + { + "epoch": 4.542829643888354, + "grad_norm": 0.26754263043403625, + "learning_rate": 0.000132649009959899, + "loss": 0.8103, + "step": 2360 + }, + { + "epoch": 4.552454282964389, + "grad_norm": 0.2548888027667999, + "learning_rate": 0.00013233096731636088, + "loss": 0.8114, + "step": 2365 + }, + { + "epoch": 4.562078922040423, + "grad_norm": 0.2608910799026489, + "learning_rate": 0.00013201255904578095, + "loss": 0.8104, + "step": 2370 + }, + { + "epoch": 4.5717035611164585, + "grad_norm": 0.2469130903482437, + "learning_rate": 0.00013169378874900017, + "loss": 0.8084, + "step": 2375 + }, + { + "epoch": 4.581328200192493, + "grad_norm": 0.26305124163627625, + "learning_rate": 0.0001313746600309538, + "loss": 0.8198, + "step": 2380 + }, + { + "epoch": 4.590952839268527, + "grad_norm": 0.2730869650840759, + "learning_rate": 0.00013105517650063026, + "loss": 0.823, + "step": 2385 + }, + { + "epoch": 4.600577478344562, + "grad_norm": 0.25265151262283325, + "learning_rate": 0.0001307353417710306, + "loss": 0.8166, + "step": 2390 + }, + { + "epoch": 4.610202117420597, + "grad_norm": 0.2525179386138916, + "learning_rate": 0.00013041515945912753, + "loss": 0.8142, + "step": 2395 + }, + { + "epoch": 4.619826756496631, + "grad_norm": 0.2585461735725403, + "learning_rate": 0.00013009463318582447, + "loss": 0.8112, + "step": 2400 + }, + { + "epoch": 4.629451395572666, + "grad_norm": 0.25699469447135925, + "learning_rate": 0.00012977376657591474, + "loss": 0.815, + "step": 2405 + }, + { + "epoch": 4.639076034648701, + "grad_norm": 0.2651076316833496, + "learning_rate": 0.00012945256325804048, + "loss": 0.8215, + "step": 2410 + }, + { + "epoch": 4.648700673724735, + "grad_norm": 0.2517280876636505, + "learning_rate": 0.0001291310268646515, + "loss": 0.8126, + "step": 2415 + }, + { + "epoch": 4.65832531280077, + "grad_norm": 0.25369200110435486, + "learning_rate": 0.00012880916103196448, + "loss": 0.811, + "step": 2420 + }, + { + "epoch": 4.667949951876805, + "grad_norm": 0.2509647011756897, + "learning_rate": 0.0001284869693999216, + "loss": 0.8144, + "step": 2425 + }, + { + "epoch": 4.6775745909528395, + "grad_norm": 0.25037845969200134, + "learning_rate": 0.00012816445561214946, + "loss": 0.8145, + "step": 2430 + }, + { + "epoch": 4.687199230028874, + "grad_norm": 0.24885617196559906, + "learning_rate": 0.000127841623315918, + "loss": 0.815, + "step": 2435 + }, + { + "epoch": 4.696823869104908, + "grad_norm": 0.26731571555137634, + "learning_rate": 0.0001275184761620989, + "loss": 0.8151, + "step": 2440 + }, + { + "epoch": 4.7064485081809435, + "grad_norm": 0.24980269372463226, + "learning_rate": 0.00012719501780512476, + "loss": 0.8189, + "step": 2445 + }, + { + "epoch": 4.716073147256978, + "grad_norm": 0.26535722613334656, + "learning_rate": 0.0001268712519029474, + "loss": 0.8205, + "step": 2450 + }, + { + "epoch": 4.725697786333012, + "grad_norm": 0.24587014317512512, + "learning_rate": 0.00012654718211699674, + "loss": 0.8127, + "step": 2455 + }, + { + "epoch": 4.735322425409047, + "grad_norm": 0.26100653409957886, + "learning_rate": 0.00012622281211213915, + "loss": 0.8146, + "step": 2460 + }, + { + "epoch": 4.744947064485082, + "grad_norm": 0.24849233031272888, + "learning_rate": 0.00012589814555663626, + "loss": 0.8107, + "step": 2465 + }, + { + "epoch": 4.754571703561116, + "grad_norm": 0.2688848078250885, + "learning_rate": 0.0001255731861221033, + "loss": 0.8209, + "step": 2470 + }, + { + "epoch": 4.764196342637151, + "grad_norm": 0.2500625550746918, + "learning_rate": 0.00012524793748346758, + "loss": 0.815, + "step": 2475 + }, + { + "epoch": 4.773820981713186, + "grad_norm": 0.2789405882358551, + "learning_rate": 0.00012492240331892716, + "loss": 0.8196, + "step": 2480 + }, + { + "epoch": 4.78344562078922, + "grad_norm": 0.25875410437583923, + "learning_rate": 0.00012459658730990891, + "loss": 0.8196, + "step": 2485 + }, + { + "epoch": 4.793070259865255, + "grad_norm": 0.25247231125831604, + "learning_rate": 0.00012427049314102707, + "loss": 0.8242, + "step": 2490 + }, + { + "epoch": 4.80269489894129, + "grad_norm": 0.2572121024131775, + "learning_rate": 0.00012394412450004164, + "loss": 0.8215, + "step": 2495 + }, + { + "epoch": 4.8123195380173245, + "grad_norm": 0.25512033700942993, + "learning_rate": 0.0001236174850778165, + "loss": 0.8163, + "step": 2500 + }, + { + "epoch": 4.821944177093359, + "grad_norm": 0.25790128111839294, + "learning_rate": 0.0001232905785682778, + "loss": 0.8119, + "step": 2505 + }, + { + "epoch": 4.831568816169393, + "grad_norm": 0.26126110553741455, + "learning_rate": 0.00012296340866837222, + "loss": 0.8133, + "step": 2510 + }, + { + "epoch": 4.8411934552454285, + "grad_norm": 0.2542867362499237, + "learning_rate": 0.00012263597907802493, + "loss": 0.818, + "step": 2515 + }, + { + "epoch": 4.850818094321463, + "grad_norm": 0.2690134048461914, + "learning_rate": 0.00012230829350009804, + "loss": 0.8106, + "step": 2520 + }, + { + "epoch": 4.860442733397497, + "grad_norm": 0.25750601291656494, + "learning_rate": 0.00012198035564034856, + "loss": 0.8125, + "step": 2525 + }, + { + "epoch": 4.870067372473533, + "grad_norm": 0.2467714548110962, + "learning_rate": 0.00012165216920738651, + "loss": 0.8172, + "step": 2530 + }, + { + "epoch": 4.879692011549567, + "grad_norm": 0.25768086314201355, + "learning_rate": 0.000121323737912633, + "loss": 0.8186, + "step": 2535 + }, + { + "epoch": 4.889316650625601, + "grad_norm": 0.2579788863658905, + "learning_rate": 0.00012099506547027826, + "loss": 0.8124, + "step": 2540 + }, + { + "epoch": 4.898941289701636, + "grad_norm": 0.250635027885437, + "learning_rate": 0.00012066615559723961, + "loss": 0.8185, + "step": 2545 + }, + { + "epoch": 4.908565928777671, + "grad_norm": 0.24465559422969818, + "learning_rate": 0.00012033701201311945, + "loss": 0.8246, + "step": 2550 + }, + { + "epoch": 4.9181905678537055, + "grad_norm": 0.24917738139629364, + "learning_rate": 0.00012000763844016321, + "loss": 0.8112, + "step": 2555 + }, + { + "epoch": 4.92781520692974, + "grad_norm": 0.24168001115322113, + "learning_rate": 0.00011967803860321726, + "loss": 0.8169, + "step": 2560 + }, + { + "epoch": 4.937439846005775, + "grad_norm": 0.2604310214519501, + "learning_rate": 0.0001193482162296867, + "loss": 0.8092, + "step": 2565 + }, + { + "epoch": 4.9470644850818095, + "grad_norm": 0.2558085024356842, + "learning_rate": 0.00011901817504949331, + "loss": 0.8226, + "step": 2570 + }, + { + "epoch": 4.956689124157844, + "grad_norm": 0.2698078155517578, + "learning_rate": 0.00011868791879503324, + "loss": 0.8147, + "step": 2575 + }, + { + "epoch": 4.966313763233879, + "grad_norm": 0.268557608127594, + "learning_rate": 0.00011835745120113508, + "loss": 0.8039, + "step": 2580 + }, + { + "epoch": 4.975938402309914, + "grad_norm": 0.25237688422203064, + "learning_rate": 0.00011802677600501725, + "loss": 0.8129, + "step": 2585 + }, + { + "epoch": 4.985563041385948, + "grad_norm": 0.24979138374328613, + "learning_rate": 0.00011769589694624601, + "loss": 0.8222, + "step": 2590 + }, + { + "epoch": 4.995187680461982, + "grad_norm": 0.282382071018219, + "learning_rate": 0.00011736481776669306, + "loss": 0.8109, + "step": 2595 + }, + { + "epoch": 4.999037536092397, + "eval_loss": 2.708376407623291, + "eval_runtime": 0.7926, + "eval_samples_per_second": 13.879, + "eval_steps_per_second": 2.523, + "step": 2597 + }, + { + "epoch": 5.004812319538018, + "grad_norm": 0.23464234173297882, + "learning_rate": 0.00011703354221049318, + "loss": 0.78, + "step": 2600 + }, + { + "epoch": 5.014436958614052, + "grad_norm": 0.29451891779899597, + "learning_rate": 0.0001167020740240021, + "loss": 0.7451, + "step": 2605 + }, + { + "epoch": 5.024061597690086, + "grad_norm": 0.26757895946502686, + "learning_rate": 0.00011637041695575383, + "loss": 0.7358, + "step": 2610 + }, + { + "epoch": 5.033686236766122, + "grad_norm": 0.27909424901008606, + "learning_rate": 0.00011603857475641846, + "loss": 0.7299, + "step": 2615 + }, + { + "epoch": 5.043310875842156, + "grad_norm": 0.27936622500419617, + "learning_rate": 0.0001157065511787598, + "loss": 0.7264, + "step": 2620 + }, + { + "epoch": 5.0529355149181905, + "grad_norm": 0.2764790952205658, + "learning_rate": 0.0001153743499775927, + "loss": 0.7414, + "step": 2625 + }, + { + "epoch": 5.062560153994225, + "grad_norm": 0.28827911615371704, + "learning_rate": 0.00011504197490974085, + "loss": 0.7344, + "step": 2630 + }, + { + "epoch": 5.07218479307026, + "grad_norm": 0.29319116473197937, + "learning_rate": 0.0001147094297339941, + "loss": 0.7419, + "step": 2635 + }, + { + "epoch": 5.0818094321462945, + "grad_norm": 0.27190330624580383, + "learning_rate": 0.0001143767182110661, + "loss": 0.7384, + "step": 2640 + }, + { + "epoch": 5.091434071222329, + "grad_norm": 0.28567731380462646, + "learning_rate": 0.00011404384410355167, + "loss": 0.7428, + "step": 2645 + }, + { + "epoch": 5.101058710298364, + "grad_norm": 0.27502113580703735, + "learning_rate": 0.00011371081117588417, + "loss": 0.751, + "step": 2650 + }, + { + "epoch": 5.110683349374399, + "grad_norm": 0.2895454168319702, + "learning_rate": 0.00011337762319429326, + "loss": 0.7389, + "step": 2655 + }, + { + "epoch": 5.120307988450433, + "grad_norm": 0.28590232133865356, + "learning_rate": 0.00011304428392676194, + "loss": 0.7351, + "step": 2660 + }, + { + "epoch": 5.129932627526467, + "grad_norm": 0.29666268825531006, + "learning_rate": 0.00011271079714298405, + "loss": 0.7437, + "step": 2665 + }, + { + "epoch": 5.139557266602503, + "grad_norm": 0.2858620584011078, + "learning_rate": 0.00011237716661432181, + "loss": 0.7393, + "step": 2670 + }, + { + "epoch": 5.149181905678537, + "grad_norm": 0.29355934262275696, + "learning_rate": 0.00011204339611376291, + "loss": 0.7429, + "step": 2675 + }, + { + "epoch": 5.1588065447545715, + "grad_norm": 0.31132546067237854, + "learning_rate": 0.00011170948941587805, + "loss": 0.7477, + "step": 2680 + }, + { + "epoch": 5.168431183830607, + "grad_norm": 0.2698726952075958, + "learning_rate": 0.00011137545029677809, + "loss": 0.7453, + "step": 2685 + }, + { + "epoch": 5.178055822906641, + "grad_norm": 0.2867010533809662, + "learning_rate": 0.0001110412825340715, + "loss": 0.7375, + "step": 2690 + }, + { + "epoch": 5.1876804619826755, + "grad_norm": 0.2847628593444824, + "learning_rate": 0.00011070698990682156, + "loss": 0.7492, + "step": 2695 + }, + { + "epoch": 5.19730510105871, + "grad_norm": 0.29182639718055725, + "learning_rate": 0.00011037257619550352, + "loss": 0.7399, + "step": 2700 + }, + { + "epoch": 5.206929740134745, + "grad_norm": 0.30024299025535583, + "learning_rate": 0.0001100380451819621, + "loss": 0.7509, + "step": 2705 + }, + { + "epoch": 5.21655437921078, + "grad_norm": 0.2791791260242462, + "learning_rate": 0.00010970340064936853, + "loss": 0.7515, + "step": 2710 + }, + { + "epoch": 5.226179018286814, + "grad_norm": 0.3051394522190094, + "learning_rate": 0.00010936864638217776, + "loss": 0.7458, + "step": 2715 + }, + { + "epoch": 5.235803657362849, + "grad_norm": 0.2900809049606323, + "learning_rate": 0.00010903378616608573, + "loss": 0.7433, + "step": 2720 + }, + { + "epoch": 5.245428296438884, + "grad_norm": 0.2843543291091919, + "learning_rate": 0.00010869882378798663, + "loss": 0.7454, + "step": 2725 + }, + { + "epoch": 5.255052935514918, + "grad_norm": 0.30490627884864807, + "learning_rate": 0.0001083637630359299, + "loss": 0.7461, + "step": 2730 + }, + { + "epoch": 5.264677574590952, + "grad_norm": 0.28951317071914673, + "learning_rate": 0.00010802860769907748, + "loss": 0.7496, + "step": 2735 + }, + { + "epoch": 5.274302213666988, + "grad_norm": 0.2910211980342865, + "learning_rate": 0.00010769336156766101, + "loss": 0.7465, + "step": 2740 + }, + { + "epoch": 5.283926852743022, + "grad_norm": 0.2923705279827118, + "learning_rate": 0.00010735802843293888, + "loss": 0.7409, + "step": 2745 + }, + { + "epoch": 5.2935514918190565, + "grad_norm": 0.2950255274772644, + "learning_rate": 0.0001070226120871534, + "loss": 0.7471, + "step": 2750 + }, + { + "epoch": 5.303176130895092, + "grad_norm": 0.29950594902038574, + "learning_rate": 0.00010668711632348787, + "loss": 0.7417, + "step": 2755 + }, + { + "epoch": 5.312800769971126, + "grad_norm": 0.28509971499443054, + "learning_rate": 0.0001063515449360238, + "loss": 0.7515, + "step": 2760 + }, + { + "epoch": 5.3224254090471605, + "grad_norm": 0.3036702871322632, + "learning_rate": 0.00010601590171969782, + "loss": 0.7395, + "step": 2765 + }, + { + "epoch": 5.332050048123195, + "grad_norm": 0.2864932119846344, + "learning_rate": 0.00010568019047025893, + "loss": 0.7473, + "step": 2770 + }, + { + "epoch": 5.34167468719923, + "grad_norm": 0.29944750666618347, + "learning_rate": 0.00010534441498422552, + "loss": 0.7454, + "step": 2775 + }, + { + "epoch": 5.351299326275265, + "grad_norm": 0.2880357503890991, + "learning_rate": 0.00010500857905884233, + "loss": 0.7455, + "step": 2780 + }, + { + "epoch": 5.360923965351299, + "grad_norm": 0.2973332107067108, + "learning_rate": 0.00010467268649203774, + "loss": 0.7607, + "step": 2785 + }, + { + "epoch": 5.370548604427334, + "grad_norm": 0.28307193517684937, + "learning_rate": 0.00010433674108238059, + "loss": 0.7522, + "step": 2790 + }, + { + "epoch": 5.380173243503369, + "grad_norm": 0.29455214738845825, + "learning_rate": 0.00010400074662903729, + "loss": 0.748, + "step": 2795 + }, + { + "epoch": 5.389797882579403, + "grad_norm": 0.2844898998737335, + "learning_rate": 0.00010366470693172896, + "loss": 0.7523, + "step": 2800 + }, + { + "epoch": 5.399422521655438, + "grad_norm": 0.29002171754837036, + "learning_rate": 0.0001033286257906883, + "loss": 0.7493, + "step": 2805 + }, + { + "epoch": 5.409047160731473, + "grad_norm": 0.2987057566642761, + "learning_rate": 0.00010299250700661678, + "loss": 0.7436, + "step": 2810 + }, + { + "epoch": 5.418671799807507, + "grad_norm": 0.2843535244464874, + "learning_rate": 0.00010265635438064145, + "loss": 0.7418, + "step": 2815 + }, + { + "epoch": 5.4282964388835415, + "grad_norm": 0.2849405109882355, + "learning_rate": 0.00010232017171427223, + "loss": 0.7489, + "step": 2820 + }, + { + "epoch": 5.437921077959577, + "grad_norm": 0.2880149781703949, + "learning_rate": 0.00010198396280935866, + "loss": 0.7563, + "step": 2825 + }, + { + "epoch": 5.447545717035611, + "grad_norm": 0.31955039501190186, + "learning_rate": 0.00010164773146804696, + "loss": 0.7544, + "step": 2830 + }, + { + "epoch": 5.457170356111646, + "grad_norm": 0.32376500964164734, + "learning_rate": 0.00010131148149273723, + "loss": 0.7469, + "step": 2835 + }, + { + "epoch": 5.466794995187681, + "grad_norm": 0.2932649254798889, + "learning_rate": 0.00010097521668604015, + "loss": 0.7548, + "step": 2840 + }, + { + "epoch": 5.476419634263715, + "grad_norm": 0.29111912846565247, + "learning_rate": 0.00010063894085073424, + "loss": 0.7517, + "step": 2845 + }, + { + "epoch": 5.48604427333975, + "grad_norm": 0.30052244663238525, + "learning_rate": 0.0001003026577897227, + "loss": 0.7468, + "step": 2850 + }, + { + "epoch": 5.495668912415784, + "grad_norm": 0.27394363284111023, + "learning_rate": 9.99663713059904e-05, + "loss": 0.7545, + "step": 2855 + }, + { + "epoch": 5.505293551491819, + "grad_norm": 0.29920995235443115, + "learning_rate": 9.9630085202561e-05, + "loss": 0.7578, + "step": 2860 + }, + { + "epoch": 5.514918190567854, + "grad_norm": 0.301736056804657, + "learning_rate": 9.929380328245378e-05, + "loss": 0.7474, + "step": 2865 + }, + { + "epoch": 5.524542829643888, + "grad_norm": 0.29692158102989197, + "learning_rate": 9.895752934864078e-05, + "loss": 0.7529, + "step": 2870 + }, + { + "epoch": 5.534167468719923, + "grad_norm": 0.28971490263938904, + "learning_rate": 9.862126720400364e-05, + "loss": 0.7533, + "step": 2875 + }, + { + "epoch": 5.543792107795958, + "grad_norm": 0.2907530665397644, + "learning_rate": 9.828502065129076e-05, + "loss": 0.7488, + "step": 2880 + }, + { + "epoch": 5.553416746871992, + "grad_norm": 0.28844624757766724, + "learning_rate": 9.794879349307419e-05, + "loss": 0.7534, + "step": 2885 + }, + { + "epoch": 5.563041385948027, + "grad_norm": 0.29487764835357666, + "learning_rate": 9.761258953170667e-05, + "loss": 0.7503, + "step": 2890 + }, + { + "epoch": 5.572666025024062, + "grad_norm": 0.29256966710090637, + "learning_rate": 9.72764125692785e-05, + "loss": 0.7516, + "step": 2895 + }, + { + "epoch": 5.582290664100096, + "grad_norm": 0.2992061376571655, + "learning_rate": 9.694026640757481e-05, + "loss": 0.7544, + "step": 2900 + }, + { + "epoch": 5.591915303176131, + "grad_norm": 0.28604987263679504, + "learning_rate": 9.660415484803226e-05, + "loss": 0.7484, + "step": 2905 + }, + { + "epoch": 5.601539942252166, + "grad_norm": 0.28531113266944885, + "learning_rate": 9.626808169169634e-05, + "loss": 0.7437, + "step": 2910 + }, + { + "epoch": 5.6111645813282, + "grad_norm": 0.2726121246814728, + "learning_rate": 9.593205073917817e-05, + "loss": 0.7589, + "step": 2915 + }, + { + "epoch": 5.620789220404235, + "grad_norm": 0.29796725511550903, + "learning_rate": 9.559606579061154e-05, + "loss": 0.7524, + "step": 2920 + }, + { + "epoch": 5.630413859480269, + "grad_norm": 0.3006713390350342, + "learning_rate": 9.526013064561006e-05, + "loss": 0.765, + "step": 2925 + }, + { + "epoch": 5.640038498556304, + "grad_norm": 0.30542224645614624, + "learning_rate": 9.492424910322413e-05, + "loss": 0.7545, + "step": 2930 + }, + { + "epoch": 5.649663137632339, + "grad_norm": 0.2783224284648895, + "learning_rate": 9.458842496189789e-05, + "loss": 0.7493, + "step": 2935 + }, + { + "epoch": 5.659287776708373, + "grad_norm": 0.3057067394256592, + "learning_rate": 9.425266201942645e-05, + "loss": 0.7668, + "step": 2940 + }, + { + "epoch": 5.668912415784408, + "grad_norm": 0.29461607336997986, + "learning_rate": 9.391696407291269e-05, + "loss": 0.7544, + "step": 2945 + }, + { + "epoch": 5.678537054860443, + "grad_norm": 0.2968499958515167, + "learning_rate": 9.358133491872453e-05, + "loss": 0.7508, + "step": 2950 + }, + { + "epoch": 5.688161693936477, + "grad_norm": 0.3040287494659424, + "learning_rate": 9.324577835245197e-05, + "loss": 0.7618, + "step": 2955 + }, + { + "epoch": 5.6977863330125125, + "grad_norm": 0.29871127009391785, + "learning_rate": 9.291029816886405e-05, + "loss": 0.7537, + "step": 2960 + }, + { + "epoch": 5.707410972088547, + "grad_norm": 0.2989570200443268, + "learning_rate": 9.257489816186606e-05, + "loss": 0.7472, + "step": 2965 + }, + { + "epoch": 5.717035611164581, + "grad_norm": 0.2932529151439667, + "learning_rate": 9.223958212445656e-05, + "loss": 0.7488, + "step": 2970 + }, + { + "epoch": 5.726660250240616, + "grad_norm": 0.29275083541870117, + "learning_rate": 9.190435384868448e-05, + "loss": 0.7532, + "step": 2975 + }, + { + "epoch": 5.736284889316651, + "grad_norm": 0.2811647057533264, + "learning_rate": 9.156921712560626e-05, + "loss": 0.7597, + "step": 2980 + }, + { + "epoch": 5.745909528392685, + "grad_norm": 0.2994243800640106, + "learning_rate": 9.123417574524307e-05, + "loss": 0.7494, + "step": 2985 + }, + { + "epoch": 5.75553416746872, + "grad_norm": 0.29861563444137573, + "learning_rate": 9.089923349653776e-05, + "loss": 0.7513, + "step": 2990 + }, + { + "epoch": 5.765158806544754, + "grad_norm": 0.27614521980285645, + "learning_rate": 9.056439416731223e-05, + "loss": 0.7498, + "step": 2995 + }, + { + "epoch": 5.774783445620789, + "grad_norm": 0.29117491841316223, + "learning_rate": 9.02296615442243e-05, + "loss": 0.7571, + "step": 3000 + }, + { + "epoch": 5.784408084696824, + "grad_norm": 0.32449835538864136, + "learning_rate": 8.989503941272522e-05, + "loss": 0.7575, + "step": 3005 + }, + { + "epoch": 5.794032723772858, + "grad_norm": 0.2920955419540405, + "learning_rate": 8.956053155701661e-05, + "loss": 0.7445, + "step": 3010 + }, + { + "epoch": 5.803657362848893, + "grad_norm": 0.3088265061378479, + "learning_rate": 8.922614176000783e-05, + "loss": 0.7534, + "step": 3015 + }, + { + "epoch": 5.813282001924928, + "grad_norm": 0.3056049942970276, + "learning_rate": 8.889187380327312e-05, + "loss": 0.7548, + "step": 3020 + }, + { + "epoch": 5.822906641000962, + "grad_norm": 0.28941500186920166, + "learning_rate": 8.855773146700872e-05, + "loss": 0.7485, + "step": 3025 + }, + { + "epoch": 5.8325312800769975, + "grad_norm": 0.2886408269405365, + "learning_rate": 8.82237185299904e-05, + "loss": 0.7422, + "step": 3030 + }, + { + "epoch": 5.842155919153032, + "grad_norm": 0.2928673028945923, + "learning_rate": 8.788983876953051e-05, + "loss": 0.7551, + "step": 3035 + }, + { + "epoch": 5.851780558229066, + "grad_norm": 0.3021661043167114, + "learning_rate": 8.755609596143534e-05, + "loss": 0.7445, + "step": 3040 + }, + { + "epoch": 5.861405197305101, + "grad_norm": 0.2965797781944275, + "learning_rate": 8.722249387996237e-05, + "loss": 0.7502, + "step": 3045 + }, + { + "epoch": 5.871029836381136, + "grad_norm": 0.3059804141521454, + "learning_rate": 8.688903629777762e-05, + "loss": 0.7544, + "step": 3050 + }, + { + "epoch": 5.88065447545717, + "grad_norm": 0.2819983661174774, + "learning_rate": 8.655572698591297e-05, + "loss": 0.7611, + "step": 3055 + }, + { + "epoch": 5.890279114533205, + "grad_norm": 0.297858327627182, + "learning_rate": 8.62225697137236e-05, + "loss": 0.7526, + "step": 3060 + }, + { + "epoch": 5.89990375360924, + "grad_norm": 0.2882884740829468, + "learning_rate": 8.588956824884523e-05, + "loss": 0.762, + "step": 3065 + }, + { + "epoch": 5.909528392685274, + "grad_norm": 0.31062471866607666, + "learning_rate": 8.555672635715162e-05, + "loss": 0.7537, + "step": 3070 + }, + { + "epoch": 5.919153031761309, + "grad_norm": 0.30393049120903015, + "learning_rate": 8.522404780271186e-05, + "loss": 0.75, + "step": 3075 + }, + { + "epoch": 5.928777670837343, + "grad_norm": 0.2902856469154358, + "learning_rate": 8.489153634774796e-05, + "loss": 0.7459, + "step": 3080 + }, + { + "epoch": 5.9384023099133785, + "grad_norm": 0.2876073718070984, + "learning_rate": 8.455919575259217e-05, + "loss": 0.7541, + "step": 3085 + }, + { + "epoch": 5.948026948989413, + "grad_norm": 0.3035559058189392, + "learning_rate": 8.422702977564453e-05, + "loss": 0.7564, + "step": 3090 + }, + { + "epoch": 5.957651588065447, + "grad_norm": 0.2893913984298706, + "learning_rate": 8.389504217333039e-05, + "loss": 0.749, + "step": 3095 + }, + { + "epoch": 5.9672762271414825, + "grad_norm": 0.2977910041809082, + "learning_rate": 8.356323670005772e-05, + "loss": 0.7509, + "step": 3100 + }, + { + "epoch": 5.976900866217517, + "grad_norm": 0.27759596705436707, + "learning_rate": 8.3231617108175e-05, + "loss": 0.7623, + "step": 3105 + }, + { + "epoch": 5.986525505293551, + "grad_norm": 0.30392059683799744, + "learning_rate": 8.290018714792852e-05, + "loss": 0.7565, + "step": 3110 + }, + { + "epoch": 5.996150144369587, + "grad_norm": 0.2790631949901581, + "learning_rate": 8.256895056742006e-05, + "loss": 0.7513, + "step": 3115 + }, + { + "epoch": 6.0, + "eval_loss": 2.9357750415802, + "eval_runtime": 0.7796, + "eval_samples_per_second": 14.111, + "eval_steps_per_second": 2.566, + "step": 3117 + }, + { + "epoch": 6.005774783445621, + "grad_norm": 0.27328184247016907, + "learning_rate": 8.223791111256447e-05, + "loss": 0.7169, + "step": 3120 + }, + { + "epoch": 6.015399422521655, + "grad_norm": 0.3284066319465637, + "learning_rate": 8.190707252704736e-05, + "loss": 0.6924, + "step": 3125 + }, + { + "epoch": 6.02502406159769, + "grad_norm": 0.30812135338783264, + "learning_rate": 8.157643855228267e-05, + "loss": 0.6785, + "step": 3130 + }, + { + "epoch": 6.034648700673725, + "grad_norm": 0.3338078558444977, + "learning_rate": 8.12460129273705e-05, + "loss": 0.6847, + "step": 3135 + }, + { + "epoch": 6.044273339749759, + "grad_norm": 0.3224867284297943, + "learning_rate": 8.091579938905474e-05, + "loss": 0.6756, + "step": 3140 + }, + { + "epoch": 6.053897978825794, + "grad_norm": 0.317451149225235, + "learning_rate": 8.05858016716808e-05, + "loss": 0.6758, + "step": 3145 + }, + { + "epoch": 6.063522617901829, + "grad_norm": 0.29282692074775696, + "learning_rate": 8.025602350715332e-05, + "loss": 0.687, + "step": 3150 + }, + { + "epoch": 6.0731472569778635, + "grad_norm": 0.3204721510410309, + "learning_rate": 7.992646862489417e-05, + "loss": 0.6808, + "step": 3155 + }, + { + "epoch": 6.082771896053898, + "grad_norm": 0.3063673675060272, + "learning_rate": 7.959714075180008e-05, + "loss": 0.6764, + "step": 3160 + }, + { + "epoch": 6.092396535129932, + "grad_norm": 0.3125745356082916, + "learning_rate": 7.926804361220055e-05, + "loss": 0.6852, + "step": 3165 + }, + { + "epoch": 6.102021174205968, + "grad_norm": 0.31588083505630493, + "learning_rate": 7.893918092781583e-05, + "loss": 0.6805, + "step": 3170 + }, + { + "epoch": 6.111645813282002, + "grad_norm": 0.3146851062774658, + "learning_rate": 7.861055641771459e-05, + "loss": 0.6862, + "step": 3175 + }, + { + "epoch": 6.121270452358036, + "grad_norm": 0.33888891339302063, + "learning_rate": 7.828217379827215e-05, + "loss": 0.6943, + "step": 3180 + }, + { + "epoch": 6.130895091434072, + "grad_norm": 0.33557072281837463, + "learning_rate": 7.79540367831283e-05, + "loss": 0.6936, + "step": 3185 + }, + { + "epoch": 6.140519730510106, + "grad_norm": 0.33382484316825867, + "learning_rate": 7.762614908314521e-05, + "loss": 0.6935, + "step": 3190 + }, + { + "epoch": 6.15014436958614, + "grad_norm": 0.31766244769096375, + "learning_rate": 7.729851440636575e-05, + "loss": 0.6927, + "step": 3195 + }, + { + "epoch": 6.159769008662175, + "grad_norm": 0.3161802291870117, + "learning_rate": 7.69711364579712e-05, + "loss": 0.6902, + "step": 3200 + }, + { + "epoch": 6.16939364773821, + "grad_norm": 0.31405240297317505, + "learning_rate": 7.664401894023967e-05, + "loss": 0.6824, + "step": 3205 + }, + { + "epoch": 6.1790182868142445, + "grad_norm": 0.31615492701530457, + "learning_rate": 7.6317165552504e-05, + "loss": 0.6893, + "step": 3210 + }, + { + "epoch": 6.188642925890279, + "grad_norm": 0.3123544454574585, + "learning_rate": 7.59905799911101e-05, + "loss": 0.6788, + "step": 3215 + }, + { + "epoch": 6.198267564966314, + "grad_norm": 0.3448927402496338, + "learning_rate": 7.566426594937503e-05, + "loss": 0.6829, + "step": 3220 + }, + { + "epoch": 6.2078922040423485, + "grad_norm": 0.2904527187347412, + "learning_rate": 7.533822711754515e-05, + "loss": 0.6953, + "step": 3225 + }, + { + "epoch": 6.217516843118383, + "grad_norm": 0.31403473019599915, + "learning_rate": 7.501246718275471e-05, + "loss": 0.6819, + "step": 3230 + }, + { + "epoch": 6.227141482194417, + "grad_norm": 0.31581783294677734, + "learning_rate": 7.468698982898382e-05, + "loss": 0.6838, + "step": 3235 + }, + { + "epoch": 6.236766121270453, + "grad_norm": 0.3196973204612732, + "learning_rate": 7.436179873701688e-05, + "loss": 0.687, + "step": 3240 + }, + { + "epoch": 6.246390760346487, + "grad_norm": 0.3196184039115906, + "learning_rate": 7.403689758440115e-05, + "loss": 0.6897, + "step": 3245 + }, + { + "epoch": 6.256015399422521, + "grad_norm": 0.32126832008361816, + "learning_rate": 7.371229004540481e-05, + "loss": 0.6954, + "step": 3250 + }, + { + "epoch": 6.265640038498557, + "grad_norm": 0.3566059470176697, + "learning_rate": 7.338797979097571e-05, + "loss": 0.698, + "step": 3255 + }, + { + "epoch": 6.275264677574591, + "grad_norm": 0.3231862783432007, + "learning_rate": 7.306397048869977e-05, + "loss": 0.6864, + "step": 3260 + }, + { + "epoch": 6.2848893166506254, + "grad_norm": 0.3360905945301056, + "learning_rate": 7.274026580275937e-05, + "loss": 0.6981, + "step": 3265 + }, + { + "epoch": 6.29451395572666, + "grad_norm": 0.30905240774154663, + "learning_rate": 7.241686939389214e-05, + "loss": 0.6839, + "step": 3270 + }, + { + "epoch": 6.304138594802695, + "grad_norm": 0.29758358001708984, + "learning_rate": 7.20937849193493e-05, + "loss": 0.6899, + "step": 3275 + }, + { + "epoch": 6.3137632338787295, + "grad_norm": 0.32738837599754333, + "learning_rate": 7.177101603285458e-05, + "loss": 0.6907, + "step": 3280 + }, + { + "epoch": 6.323387872954764, + "grad_norm": 0.30813169479370117, + "learning_rate": 7.144856638456272e-05, + "loss": 0.6919, + "step": 3285 + }, + { + "epoch": 6.333012512030799, + "grad_norm": 0.340621680021286, + "learning_rate": 7.112643962101817e-05, + "loss": 0.6884, + "step": 3290 + }, + { + "epoch": 6.342637151106834, + "grad_norm": 0.3451749384403229, + "learning_rate": 7.080463938511405e-05, + "loss": 0.6937, + "step": 3295 + }, + { + "epoch": 6.352261790182868, + "grad_norm": 0.32087814807891846, + "learning_rate": 7.048316931605062e-05, + "loss": 0.6929, + "step": 3300 + }, + { + "epoch": 6.361886429258902, + "grad_norm": 0.30795004963874817, + "learning_rate": 7.016203304929451e-05, + "loss": 0.6983, + "step": 3305 + }, + { + "epoch": 6.371511068334938, + "grad_norm": 0.3312138617038727, + "learning_rate": 6.984123421653733e-05, + "loss": 0.6845, + "step": 3310 + }, + { + "epoch": 6.381135707410972, + "grad_norm": 0.3371661901473999, + "learning_rate": 6.952077644565469e-05, + "loss": 0.6898, + "step": 3315 + }, + { + "epoch": 6.390760346487006, + "grad_norm": 0.3481803834438324, + "learning_rate": 6.920066336066524e-05, + "loss": 0.6912, + "step": 3320 + }, + { + "epoch": 6.400384985563042, + "grad_norm": 0.32163578271865845, + "learning_rate": 6.888089858168949e-05, + "loss": 0.6901, + "step": 3325 + }, + { + "epoch": 6.410009624639076, + "grad_norm": 0.3223172128200531, + "learning_rate": 6.85614857249091e-05, + "loss": 0.6944, + "step": 3330 + }, + { + "epoch": 6.4196342637151105, + "grad_norm": 0.30212926864624023, + "learning_rate": 6.824242840252588e-05, + "loss": 0.7016, + "step": 3335 + }, + { + "epoch": 6.429258902791146, + "grad_norm": 0.32831230759620667, + "learning_rate": 6.79237302227209e-05, + "loss": 0.6869, + "step": 3340 + }, + { + "epoch": 6.43888354186718, + "grad_norm": 0.3248232305049896, + "learning_rate": 6.76053947896138e-05, + "loss": 0.6945, + "step": 3345 + }, + { + "epoch": 6.4485081809432145, + "grad_norm": 0.3347261846065521, + "learning_rate": 6.728742570322181e-05, + "loss": 0.6911, + "step": 3350 + }, + { + "epoch": 6.458132820019249, + "grad_norm": 0.3434222936630249, + "learning_rate": 6.69698265594194e-05, + "loss": 0.7001, + "step": 3355 + }, + { + "epoch": 6.467757459095284, + "grad_norm": 0.31891781091690063, + "learning_rate": 6.66526009498972e-05, + "loss": 0.6961, + "step": 3360 + }, + { + "epoch": 6.477382098171319, + "grad_norm": 0.32785654067993164, + "learning_rate": 6.633575246212175e-05, + "loss": 0.6986, + "step": 3365 + }, + { + "epoch": 6.487006737247353, + "grad_norm": 0.3148154020309448, + "learning_rate": 6.601928467929472e-05, + "loss": 0.6857, + "step": 3370 + }, + { + "epoch": 6.496631376323388, + "grad_norm": 0.3220577836036682, + "learning_rate": 6.570320118031232e-05, + "loss": 0.6933, + "step": 3375 + }, + { + "epoch": 6.506256015399423, + "grad_norm": 0.3030003309249878, + "learning_rate": 6.538750553972509e-05, + "loss": 0.6963, + "step": 3380 + }, + { + "epoch": 6.515880654475457, + "grad_norm": 0.32863059639930725, + "learning_rate": 6.507220132769723e-05, + "loss": 0.6929, + "step": 3385 + }, + { + "epoch": 6.5255052935514914, + "grad_norm": 0.35064488649368286, + "learning_rate": 6.475729210996637e-05, + "loss": 0.6864, + "step": 3390 + }, + { + "epoch": 6.535129932627527, + "grad_norm": 0.32089149951934814, + "learning_rate": 6.444278144780325e-05, + "loss": 0.6858, + "step": 3395 + }, + { + "epoch": 6.544754571703561, + "grad_norm": 0.4273422658443451, + "learning_rate": 6.41286728979712e-05, + "loss": 0.6968, + "step": 3400 + }, + { + "epoch": 6.5543792107795955, + "grad_norm": 0.33466604351997375, + "learning_rate": 6.38149700126863e-05, + "loss": 0.6966, + "step": 3405 + }, + { + "epoch": 6.564003849855631, + "grad_norm": 0.3052511513233185, + "learning_rate": 6.350167633957698e-05, + "loss": 0.6983, + "step": 3410 + }, + { + "epoch": 6.573628488931665, + "grad_norm": 0.3621208071708679, + "learning_rate": 6.318879542164385e-05, + "loss": 0.6986, + "step": 3415 + }, + { + "epoch": 6.5832531280077, + "grad_norm": 0.32712018489837646, + "learning_rate": 6.287633079721986e-05, + "loss": 0.6927, + "step": 3420 + }, + { + "epoch": 6.592877767083735, + "grad_norm": 0.3064589202404022, + "learning_rate": 6.256428599993e-05, + "loss": 0.6995, + "step": 3425 + }, + { + "epoch": 6.602502406159769, + "grad_norm": 0.3126335144042969, + "learning_rate": 6.225266455865157e-05, + "loss": 0.6985, + "step": 3430 + }, + { + "epoch": 6.612127045235804, + "grad_norm": 0.35115116834640503, + "learning_rate": 6.194146999747419e-05, + "loss": 0.6918, + "step": 3435 + }, + { + "epoch": 6.621751684311838, + "grad_norm": 0.32435253262519836, + "learning_rate": 6.163070583565993e-05, + "loss": 0.6988, + "step": 3440 + }, + { + "epoch": 6.631376323387873, + "grad_norm": 0.3202888071537018, + "learning_rate": 6.13203755876035e-05, + "loss": 0.6895, + "step": 3445 + }, + { + "epoch": 6.641000962463908, + "grad_norm": 0.3102019131183624, + "learning_rate": 6.1010482762792585e-05, + "loss": 0.6923, + "step": 3450 + }, + { + "epoch": 6.650625601539942, + "grad_norm": 0.3367016911506653, + "learning_rate": 6.070103086576802e-05, + "loss": 0.6915, + "step": 3455 + }, + { + "epoch": 6.6602502406159765, + "grad_norm": 0.3353261351585388, + "learning_rate": 6.039202339608432e-05, + "loss": 0.687, + "step": 3460 + }, + { + "epoch": 6.669874879692012, + "grad_norm": 0.30828601121902466, + "learning_rate": 6.0083463848269995e-05, + "loss": 0.6934, + "step": 3465 + }, + { + "epoch": 6.679499518768046, + "grad_norm": 0.3269566595554352, + "learning_rate": 5.977535571178809e-05, + "loss": 0.6967, + "step": 3470 + }, + { + "epoch": 6.6891241578440805, + "grad_norm": 0.339278906583786, + "learning_rate": 5.946770247099661e-05, + "loss": 0.691, + "step": 3475 + }, + { + "epoch": 6.698748796920116, + "grad_norm": 0.33345827460289, + "learning_rate": 5.9160507605109275e-05, + "loss": 0.7039, + "step": 3480 + }, + { + "epoch": 6.70837343599615, + "grad_norm": 0.318852037191391, + "learning_rate": 5.885377458815609e-05, + "loss": 0.7019, + "step": 3485 + }, + { + "epoch": 6.717998075072185, + "grad_norm": 0.3394601047039032, + "learning_rate": 5.8547506888944007e-05, + "loss": 0.6881, + "step": 3490 + }, + { + "epoch": 6.72762271414822, + "grad_norm": 0.32474079728126526, + "learning_rate": 5.824170797101787e-05, + "loss": 0.6879, + "step": 3495 + }, + { + "epoch": 6.737247353224254, + "grad_norm": 0.325595885515213, + "learning_rate": 5.7936381292621e-05, + "loss": 0.6951, + "step": 3500 + }, + { + "epoch": 6.746871992300289, + "grad_norm": 0.3558216989040375, + "learning_rate": 5.763153030665629e-05, + "loss": 0.6947, + "step": 3505 + }, + { + "epoch": 6.756496631376323, + "grad_norm": 0.3530566692352295, + "learning_rate": 5.7327158460647065e-05, + "loss": 0.6986, + "step": 3510 + }, + { + "epoch": 6.766121270452358, + "grad_norm": 0.33962172269821167, + "learning_rate": 5.702326919669817e-05, + "loss": 0.6964, + "step": 3515 + }, + { + "epoch": 6.775745909528393, + "grad_norm": 0.3108658790588379, + "learning_rate": 5.671986595145693e-05, + "loss": 0.6923, + "step": 3520 + }, + { + "epoch": 6.785370548604427, + "grad_norm": 0.32073214650154114, + "learning_rate": 5.64169521560743e-05, + "loss": 0.6792, + "step": 3525 + }, + { + "epoch": 6.7949951876804615, + "grad_norm": 0.3249306380748749, + "learning_rate": 5.611453123616618e-05, + "loss": 0.7013, + "step": 3530 + }, + { + "epoch": 6.804619826756497, + "grad_norm": 0.333997905254364, + "learning_rate": 5.581260661177463e-05, + "loss": 0.6923, + "step": 3535 + }, + { + "epoch": 6.814244465832531, + "grad_norm": 0.3433645963668823, + "learning_rate": 5.551118169732901e-05, + "loss": 0.7014, + "step": 3540 + }, + { + "epoch": 6.823869104908566, + "grad_norm": 0.3301408886909485, + "learning_rate": 5.521025990160772e-05, + "loss": 0.6966, + "step": 3545 + }, + { + "epoch": 6.833493743984601, + "grad_norm": 0.341169148683548, + "learning_rate": 5.4909844627699255e-05, + "loss": 0.6963, + "step": 3550 + }, + { + "epoch": 6.843118383060635, + "grad_norm": 0.31754934787750244, + "learning_rate": 5.460993927296407e-05, + "loss": 0.6996, + "step": 3555 + }, + { + "epoch": 6.85274302213667, + "grad_norm": 0.3002949655056, + "learning_rate": 5.4310547228995936e-05, + "loss": 0.6946, + "step": 3560 + }, + { + "epoch": 6.862367661212705, + "grad_norm": 0.3369508981704712, + "learning_rate": 5.4011671881583656e-05, + "loss": 0.6902, + "step": 3565 + }, + { + "epoch": 6.871992300288739, + "grad_norm": 0.3112001419067383, + "learning_rate": 5.371331661067284e-05, + "loss": 0.6935, + "step": 3570 + }, + { + "epoch": 6.881616939364774, + "grad_norm": 0.3145786225795746, + "learning_rate": 5.341548479032745e-05, + "loss": 0.7027, + "step": 3575 + }, + { + "epoch": 6.891241578440808, + "grad_norm": 0.32883113622665405, + "learning_rate": 5.311817978869198e-05, + "loss": 0.6928, + "step": 3580 + }, + { + "epoch": 6.900866217516843, + "grad_norm": 0.3237265646457672, + "learning_rate": 5.2821404967953114e-05, + "loss": 0.6865, + "step": 3585 + }, + { + "epoch": 6.910490856592878, + "grad_norm": 0.32935890555381775, + "learning_rate": 5.2525163684301806e-05, + "loss": 0.687, + "step": 3590 + }, + { + "epoch": 6.920115495668912, + "grad_norm": 0.342359721660614, + "learning_rate": 5.222945928789533e-05, + "loss": 0.691, + "step": 3595 + }, + { + "epoch": 6.929740134744947, + "grad_norm": 0.3421998620033264, + "learning_rate": 5.193429512281926e-05, + "loss": 0.6863, + "step": 3600 + }, + { + "epoch": 6.939364773820982, + "grad_norm": 0.33589935302734375, + "learning_rate": 5.1639674527049855e-05, + "loss": 0.6916, + "step": 3605 + }, + { + "epoch": 6.948989412897016, + "grad_norm": 0.3499864637851715, + "learning_rate": 5.134560083241624e-05, + "loss": 0.6878, + "step": 3610 + }, + { + "epoch": 6.958614051973051, + "grad_norm": 0.3289993405342102, + "learning_rate": 5.105207736456257e-05, + "loss": 0.6976, + "step": 3615 + }, + { + "epoch": 6.968238691049086, + "grad_norm": 0.32949408888816833, + "learning_rate": 5.0759107442910715e-05, + "loss": 0.6949, + "step": 3620 + }, + { + "epoch": 6.97786333012512, + "grad_norm": 0.3234226703643799, + "learning_rate": 5.046669438062238e-05, + "loss": 0.6958, + "step": 3625 + }, + { + "epoch": 6.987487969201155, + "grad_norm": 0.3094496726989746, + "learning_rate": 5.0174841484561953e-05, + "loss": 0.6938, + "step": 3630 + }, + { + "epoch": 6.99711260827719, + "grad_norm": 0.31556159257888794, + "learning_rate": 4.988355205525893e-05, + "loss": 0.7004, + "step": 3635 + }, + { + "epoch": 6.999037536092397, + "eval_loss": 3.276942253112793, + "eval_runtime": 0.7888, + "eval_samples_per_second": 13.945, + "eval_steps_per_second": 2.535, + "step": 3636 + }, + { + "epoch": 7.006737247353224, + "grad_norm": 0.26794806122779846, + "learning_rate": 4.959282938687061e-05, + "loss": 0.6482, + "step": 3640 + }, + { + "epoch": 7.016361886429259, + "grad_norm": 0.3672392666339874, + "learning_rate": 4.9302676767144926e-05, + "loss": 0.6471, + "step": 3645 + }, + { + "epoch": 7.025986525505293, + "grad_norm": 0.2901393175125122, + "learning_rate": 4.901309747738305e-05, + "loss": 0.654, + "step": 3650 + }, + { + "epoch": 7.035611164581328, + "grad_norm": 0.3516036868095398, + "learning_rate": 4.872409479240259e-05, + "loss": 0.6452, + "step": 3655 + }, + { + "epoch": 7.045235803657363, + "grad_norm": 0.3640913665294647, + "learning_rate": 4.843567198050031e-05, + "loss": 0.6369, + "step": 3660 + }, + { + "epoch": 7.054860442733397, + "grad_norm": 0.2963874936103821, + "learning_rate": 4.814783230341531e-05, + "loss": 0.6353, + "step": 3665 + }, + { + "epoch": 7.0644850818094325, + "grad_norm": 0.3295438587665558, + "learning_rate": 4.786057901629209e-05, + "loss": 0.6398, + "step": 3670 + }, + { + "epoch": 7.074109720885467, + "grad_norm": 0.3382556736469269, + "learning_rate": 4.757391536764366e-05, + "loss": 0.6452, + "step": 3675 + }, + { + "epoch": 7.083734359961501, + "grad_norm": 0.3277692496776581, + "learning_rate": 4.728784459931495e-05, + "loss": 0.637, + "step": 3680 + }, + { + "epoch": 7.0933589990375365, + "grad_norm": 0.3565356433391571, + "learning_rate": 4.700236994644609e-05, + "loss": 0.6379, + "step": 3685 + }, + { + "epoch": 7.102983638113571, + "grad_norm": 0.35193830728530884, + "learning_rate": 4.671749463743572e-05, + "loss": 0.6512, + "step": 3690 + }, + { + "epoch": 7.112608277189605, + "grad_norm": 0.32000118494033813, + "learning_rate": 4.64332218939047e-05, + "loss": 0.6445, + "step": 3695 + }, + { + "epoch": 7.12223291626564, + "grad_norm": 0.33006584644317627, + "learning_rate": 4.61495549306594e-05, + "loss": 0.6381, + "step": 3700 + }, + { + "epoch": 7.131857555341675, + "grad_norm": 0.3775092661380768, + "learning_rate": 4.586649695565563e-05, + "loss": 0.6331, + "step": 3705 + }, + { + "epoch": 7.141482194417709, + "grad_norm": 0.3325980007648468, + "learning_rate": 4.558405116996214e-05, + "loss": 0.6436, + "step": 3710 + }, + { + "epoch": 7.151106833493744, + "grad_norm": 0.3391129970550537, + "learning_rate": 4.530222076772456e-05, + "loss": 0.6415, + "step": 3715 + }, + { + "epoch": 7.160731472569779, + "grad_norm": 0.31919702887535095, + "learning_rate": 4.5021008936129216e-05, + "loss": 0.6441, + "step": 3720 + }, + { + "epoch": 7.170356111645813, + "grad_norm": 0.3420950770378113, + "learning_rate": 4.4740418855367005e-05, + "loss": 0.6524, + "step": 3725 + }, + { + "epoch": 7.179980750721848, + "grad_norm": 0.354056179523468, + "learning_rate": 4.4460453698597623e-05, + "loss": 0.6476, + "step": 3730 + }, + { + "epoch": 7.189605389797882, + "grad_norm": 0.31593650579452515, + "learning_rate": 4.418111663191354e-05, + "loss": 0.6473, + "step": 3735 + }, + { + "epoch": 7.1992300288739175, + "grad_norm": 0.33761167526245117, + "learning_rate": 4.390241081430423e-05, + "loss": 0.6402, + "step": 3740 + }, + { + "epoch": 7.208854667949952, + "grad_norm": 0.35358771681785583, + "learning_rate": 4.362433939762046e-05, + "loss": 0.6471, + "step": 3745 + }, + { + "epoch": 7.218479307025986, + "grad_norm": 0.32182127237319946, + "learning_rate": 4.3346905526538574e-05, + "loss": 0.6408, + "step": 3750 + }, + { + "epoch": 7.228103946102022, + "grad_norm": 0.3282702565193176, + "learning_rate": 4.307011233852505e-05, + "loss": 0.642, + "step": 3755 + }, + { + "epoch": 7.237728585178056, + "grad_norm": 0.33513620495796204, + "learning_rate": 4.279396296380097e-05, + "loss": 0.6391, + "step": 3760 + }, + { + "epoch": 7.24735322425409, + "grad_norm": 0.33494138717651367, + "learning_rate": 4.2518460525306524e-05, + "loss": 0.6401, + "step": 3765 + }, + { + "epoch": 7.256977863330125, + "grad_norm": 0.33716508746147156, + "learning_rate": 4.2243608138665906e-05, + "loss": 0.6499, + "step": 3770 + }, + { + "epoch": 7.26660250240616, + "grad_norm": 0.3404597043991089, + "learning_rate": 4.19694089121518e-05, + "loss": 0.6385, + "step": 3775 + }, + { + "epoch": 7.276227141482194, + "grad_norm": 0.32999253273010254, + "learning_rate": 4.169586594665048e-05, + "loss": 0.6433, + "step": 3780 + }, + { + "epoch": 7.285851780558229, + "grad_norm": 0.3411442041397095, + "learning_rate": 4.142298233562664e-05, + "loss": 0.6422, + "step": 3785 + }, + { + "epoch": 7.295476419634264, + "grad_norm": 0.3550765812397003, + "learning_rate": 4.115076116508837e-05, + "loss": 0.6458, + "step": 3790 + }, + { + "epoch": 7.3051010587102985, + "grad_norm": 0.3416723608970642, + "learning_rate": 4.08792055135524e-05, + "loss": 0.6456, + "step": 3795 + }, + { + "epoch": 7.314725697786333, + "grad_norm": 0.35609087347984314, + "learning_rate": 4.0608318452009e-05, + "loss": 0.6533, + "step": 3800 + }, + { + "epoch": 7.324350336862367, + "grad_norm": 0.332507461309433, + "learning_rate": 4.033810304388759e-05, + "loss": 0.6282, + "step": 3805 + }, + { + "epoch": 7.3339749759384025, + "grad_norm": 0.34344714879989624, + "learning_rate": 4.006856234502191e-05, + "loss": 0.633, + "step": 3810 + }, + { + "epoch": 7.343599615014437, + "grad_norm": 0.3543119430541992, + "learning_rate": 3.9799699403615457e-05, + "loss": 0.6417, + "step": 3815 + }, + { + "epoch": 7.353224254090471, + "grad_norm": 0.3393097221851349, + "learning_rate": 3.953151726020713e-05, + "loss": 0.6337, + "step": 3820 + }, + { + "epoch": 7.362848893166507, + "grad_norm": 0.34601929783821106, + "learning_rate": 3.926401894763663e-05, + "loss": 0.6514, + "step": 3825 + }, + { + "epoch": 7.372473532242541, + "grad_norm": 0.3476494550704956, + "learning_rate": 3.89972074910104e-05, + "loss": 0.6381, + "step": 3830 + }, + { + "epoch": 7.382098171318575, + "grad_norm": 0.3308873474597931, + "learning_rate": 3.8731085907667345e-05, + "loss": 0.6523, + "step": 3835 + }, + { + "epoch": 7.39172281039461, + "grad_norm": 0.33746767044067383, + "learning_rate": 3.846565720714451e-05, + "loss": 0.6386, + "step": 3840 + }, + { + "epoch": 7.401347449470645, + "grad_norm": 0.33146432042121887, + "learning_rate": 3.820092439114339e-05, + "loss": 0.6505, + "step": 3845 + }, + { + "epoch": 7.410972088546679, + "grad_norm": 0.34075871109962463, + "learning_rate": 3.793689045349575e-05, + "loss": 0.6292, + "step": 3850 + }, + { + "epoch": 7.420596727622714, + "grad_norm": 0.3384300172328949, + "learning_rate": 3.7673558380129735e-05, + "loss": 0.649, + "step": 3855 + }, + { + "epoch": 7.430221366698749, + "grad_norm": 0.35409146547317505, + "learning_rate": 3.741093114903631e-05, + "loss": 0.6401, + "step": 3860 + }, + { + "epoch": 7.4398460057747835, + "grad_norm": 0.3388952314853668, + "learning_rate": 3.7149011730235394e-05, + "loss": 0.646, + "step": 3865 + }, + { + "epoch": 7.449470644850818, + "grad_norm": 0.3542778789997101, + "learning_rate": 3.688780308574238e-05, + "loss": 0.6367, + "step": 3870 + }, + { + "epoch": 7.459095283926853, + "grad_norm": 0.33730167150497437, + "learning_rate": 3.66273081695346e-05, + "loss": 0.655, + "step": 3875 + }, + { + "epoch": 7.468719923002888, + "grad_norm": 0.3402201533317566, + "learning_rate": 3.6367529927517855e-05, + "loss": 0.6327, + "step": 3880 + }, + { + "epoch": 7.478344562078922, + "grad_norm": 0.3543342649936676, + "learning_rate": 3.610847129749323e-05, + "loss": 0.6534, + "step": 3885 + }, + { + "epoch": 7.487969201154956, + "grad_norm": 0.3624216914176941, + "learning_rate": 3.585013520912377e-05, + "loss": 0.6393, + "step": 3890 + }, + { + "epoch": 7.497593840230992, + "grad_norm": 0.3448854386806488, + "learning_rate": 3.559252458390142e-05, + "loss": 0.6473, + "step": 3895 + }, + { + "epoch": 7.507218479307026, + "grad_norm": 0.3260321021080017, + "learning_rate": 3.533564233511394e-05, + "loss": 0.635, + "step": 3900 + }, + { + "epoch": 7.51684311838306, + "grad_norm": 0.36959561705589294, + "learning_rate": 3.507949136781189e-05, + "loss": 0.6454, + "step": 3905 + }, + { + "epoch": 7.526467757459095, + "grad_norm": 0.3395916223526001, + "learning_rate": 3.482407457877598e-05, + "loss": 0.6491, + "step": 3910 + }, + { + "epoch": 7.53609239653513, + "grad_norm": 0.3479905426502228, + "learning_rate": 3.456939485648406e-05, + "loss": 0.638, + "step": 3915 + }, + { + "epoch": 7.5457170356111645, + "grad_norm": 0.3783397674560547, + "learning_rate": 3.4315455081078696e-05, + "loss": 0.6446, + "step": 3920 + }, + { + "epoch": 7.555341674687199, + "grad_norm": 0.34621936082839966, + "learning_rate": 3.4062258124334434e-05, + "loss": 0.64, + "step": 3925 + }, + { + "epoch": 7.564966313763234, + "grad_norm": 0.34806111454963684, + "learning_rate": 3.3809806849625314e-05, + "loss": 0.641, + "step": 3930 + }, + { + "epoch": 7.5745909528392685, + "grad_norm": 0.33737459778785706, + "learning_rate": 3.355810411189264e-05, + "loss": 0.6389, + "step": 3935 + }, + { + "epoch": 7.584215591915303, + "grad_norm": 0.36518171429634094, + "learning_rate": 3.330715275761257e-05, + "loss": 0.6448, + "step": 3940 + }, + { + "epoch": 7.593840230991338, + "grad_norm": 0.3364472985267639, + "learning_rate": 3.305695562476393e-05, + "loss": 0.6378, + "step": 3945 + }, + { + "epoch": 7.603464870067373, + "grad_norm": 0.345920592546463, + "learning_rate": 3.280751554279622e-05, + "loss": 0.634, + "step": 3950 + }, + { + "epoch": 7.613089509143407, + "grad_norm": 0.33815324306488037, + "learning_rate": 3.255883533259741e-05, + "loss": 0.6452, + "step": 3955 + }, + { + "epoch": 7.622714148219442, + "grad_norm": 0.34798070788383484, + "learning_rate": 3.2310917806462274e-05, + "loss": 0.6433, + "step": 3960 + }, + { + "epoch": 7.632338787295477, + "grad_norm": 0.34050893783569336, + "learning_rate": 3.2063765768060475e-05, + "loss": 0.6505, + "step": 3965 + }, + { + "epoch": 7.641963426371511, + "grad_norm": 0.3409608006477356, + "learning_rate": 3.1817382012404854e-05, + "loss": 0.6515, + "step": 3970 + }, + { + "epoch": 7.651588065447545, + "grad_norm": 0.3448992371559143, + "learning_rate": 3.157176932581983e-05, + "loss": 0.6355, + "step": 3975 + }, + { + "epoch": 7.661212704523581, + "grad_norm": 0.3314208984375, + "learning_rate": 3.132693048590988e-05, + "loss": 0.647, + "step": 3980 + }, + { + "epoch": 7.670837343599615, + "grad_norm": 0.34806132316589355, + "learning_rate": 3.108286826152818e-05, + "loss": 0.6377, + "step": 3985 + }, + { + "epoch": 7.6804619826756495, + "grad_norm": 0.3525891900062561, + "learning_rate": 3.083958541274518e-05, + "loss": 0.6326, + "step": 3990 + }, + { + "epoch": 7.690086621751684, + "grad_norm": 0.36846107244491577, + "learning_rate": 3.059708469081754e-05, + "loss": 0.6327, + "step": 3995 + }, + { + "epoch": 7.699711260827719, + "grad_norm": 0.33311864733695984, + "learning_rate": 3.035536883815696e-05, + "loss": 0.6379, + "step": 4000 + }, + { + "epoch": 7.709335899903754, + "grad_norm": 0.3615313172340393, + "learning_rate": 3.0114440588299033e-05, + "loss": 0.6522, + "step": 4005 + }, + { + "epoch": 7.718960538979788, + "grad_norm": 0.33901557326316833, + "learning_rate": 2.9874302665872544e-05, + "loss": 0.6495, + "step": 4010 + }, + { + "epoch": 7.728585178055823, + "grad_norm": 0.3336678743362427, + "learning_rate": 2.963495778656853e-05, + "loss": 0.6583, + "step": 4015 + }, + { + "epoch": 7.738209817131858, + "grad_norm": 0.38028064370155334, + "learning_rate": 2.9396408657109608e-05, + "loss": 0.6365, + "step": 4020 + }, + { + "epoch": 7.747834456207892, + "grad_norm": 0.3507869243621826, + "learning_rate": 2.9158657975219385e-05, + "loss": 0.6466, + "step": 4025 + }, + { + "epoch": 7.757459095283927, + "grad_norm": 0.3580639660358429, + "learning_rate": 2.8921708429591797e-05, + "loss": 0.6472, + "step": 4030 + }, + { + "epoch": 7.767083734359962, + "grad_norm": 0.3309887945652008, + "learning_rate": 2.8685562699860957e-05, + "loss": 0.6476, + "step": 4035 + }, + { + "epoch": 7.776708373435996, + "grad_norm": 0.3457421064376831, + "learning_rate": 2.8450223456570668e-05, + "loss": 0.6414, + "step": 4040 + }, + { + "epoch": 7.7863330125120305, + "grad_norm": 0.33013686537742615, + "learning_rate": 2.8215693361144324e-05, + "loss": 0.6535, + "step": 4045 + }, + { + "epoch": 7.795957651588066, + "grad_norm": 0.32177311182022095, + "learning_rate": 2.798197506585464e-05, + "loss": 0.6487, + "step": 4050 + }, + { + "epoch": 7.8055822906641, + "grad_norm": 0.3439447283744812, + "learning_rate": 2.774907121379393e-05, + "loss": 0.6354, + "step": 4055 + }, + { + "epoch": 7.8152069297401345, + "grad_norm": 0.34718647599220276, + "learning_rate": 2.751698443884394e-05, + "loss": 0.6504, + "step": 4060 + }, + { + "epoch": 7.824831568816169, + "grad_norm": 0.34381964802742004, + "learning_rate": 2.7285717365646256e-05, + "loss": 0.6453, + "step": 4065 + }, + { + "epoch": 7.834456207892204, + "grad_norm": 0.34925544261932373, + "learning_rate": 2.7055272609572568e-05, + "loss": 0.6484, + "step": 4070 + }, + { + "epoch": 7.844080846968239, + "grad_norm": 0.34031766653060913, + "learning_rate": 2.6825652776695076e-05, + "loss": 0.6462, + "step": 4075 + }, + { + "epoch": 7.853705486044273, + "grad_norm": 0.3397299349308014, + "learning_rate": 2.6596860463756935e-05, + "loss": 0.6444, + "step": 4080 + }, + { + "epoch": 7.863330125120308, + "grad_norm": 0.348021537065506, + "learning_rate": 2.636889825814307e-05, + "loss": 0.6389, + "step": 4085 + }, + { + "epoch": 7.872954764196343, + "grad_norm": 0.3368039131164551, + "learning_rate": 2.6141768737850814e-05, + "loss": 0.6453, + "step": 4090 + }, + { + "epoch": 7.882579403272377, + "grad_norm": 0.34815698862075806, + "learning_rate": 2.5915474471460732e-05, + "loss": 0.6474, + "step": 4095 + }, + { + "epoch": 7.892204042348412, + "grad_norm": 0.3499961793422699, + "learning_rate": 2.5690018018107642e-05, + "loss": 0.6436, + "step": 4100 + }, + { + "epoch": 7.901828681424447, + "grad_norm": 0.3426460921764374, + "learning_rate": 2.5465401927451537e-05, + "loss": 0.6437, + "step": 4105 + }, + { + "epoch": 7.911453320500481, + "grad_norm": 0.3375738561153412, + "learning_rate": 2.524162873964896e-05, + "loss": 0.6394, + "step": 4110 + }, + { + "epoch": 7.9210779595765155, + "grad_norm": 0.34224507212638855, + "learning_rate": 2.501870098532412e-05, + "loss": 0.6524, + "step": 4115 + }, + { + "epoch": 7.930702598652551, + "grad_norm": 0.3286498785018921, + "learning_rate": 2.4796621185540348e-05, + "loss": 0.6507, + "step": 4120 + }, + { + "epoch": 7.940327237728585, + "grad_norm": 0.36504673957824707, + "learning_rate": 2.4575391851771477e-05, + "loss": 0.6389, + "step": 4125 + }, + { + "epoch": 7.94995187680462, + "grad_norm": 0.3325868546962738, + "learning_rate": 2.4355015485873644e-05, + "loss": 0.6402, + "step": 4130 + }, + { + "epoch": 7.959576515880655, + "grad_norm": 0.35220691561698914, + "learning_rate": 2.4135494580056737e-05, + "loss": 0.6553, + "step": 4135 + }, + { + "epoch": 7.969201154956689, + "grad_norm": 0.3708426058292389, + "learning_rate": 2.3916831616856473e-05, + "loss": 0.6518, + "step": 4140 + }, + { + "epoch": 7.978825794032724, + "grad_norm": 0.34426939487457275, + "learning_rate": 2.3699029069106115e-05, + "loss": 0.6505, + "step": 4145 + }, + { + "epoch": 7.988450433108758, + "grad_norm": 0.3554341793060303, + "learning_rate": 2.348208939990866e-05, + "loss": 0.6497, + "step": 4150 + }, + { + "epoch": 7.998075072184793, + "grad_norm": 0.3434050381183624, + "learning_rate": 2.3266015062608838e-05, + "loss": 0.6466, + "step": 4155 + }, + { + "epoch": 8.0, + "eval_loss": 3.694774627685547, + "eval_runtime": 0.7787, + "eval_samples_per_second": 14.127, + "eval_steps_per_second": 2.569, + "step": 4156 + }, + { + "epoch": 8.007699711260829, + "grad_norm": 0.28248271346092224, + "learning_rate": 2.3050808500765487e-05, + "loss": 0.6121, + "step": 4160 + }, + { + "epoch": 8.017324350336862, + "grad_norm": 0.36666032671928406, + "learning_rate": 2.2836472148123878e-05, + "loss": 0.6176, + "step": 4165 + }, + { + "epoch": 8.026948989412897, + "grad_norm": 0.32897964119911194, + "learning_rate": 2.2623008428588177e-05, + "loss": 0.6079, + "step": 4170 + }, + { + "epoch": 8.03657362848893, + "grad_norm": 0.32618117332458496, + "learning_rate": 2.24104197561941e-05, + "loss": 0.6043, + "step": 4175 + }, + { + "epoch": 8.046198267564966, + "grad_norm": 0.3435162305831909, + "learning_rate": 2.2198708535081446e-05, + "loss": 0.6082, + "step": 4180 + }, + { + "epoch": 8.055822906641001, + "grad_norm": 0.3350038528442383, + "learning_rate": 2.198787715946712e-05, + "loss": 0.6098, + "step": 4185 + }, + { + "epoch": 8.065447545717035, + "grad_norm": 0.3771952986717224, + "learning_rate": 2.1777928013617908e-05, + "loss": 0.6137, + "step": 4190 + }, + { + "epoch": 8.07507218479307, + "grad_norm": 0.3174493610858917, + "learning_rate": 2.1568863471823642e-05, + "loss": 0.6169, + "step": 4195 + }, + { + "epoch": 8.084696823869105, + "grad_norm": 0.33214735984802246, + "learning_rate": 2.1360685898370146e-05, + "loss": 0.6066, + "step": 4200 + }, + { + "epoch": 8.094321462945139, + "grad_norm": 0.3336653411388397, + "learning_rate": 2.1153397647512763e-05, + "loss": 0.6073, + "step": 4205 + }, + { + "epoch": 8.103946102021174, + "grad_norm": 0.32206472754478455, + "learning_rate": 2.0947001063449457e-05, + "loss": 0.6, + "step": 4210 + }, + { + "epoch": 8.11357074109721, + "grad_norm": 0.3184707760810852, + "learning_rate": 2.074149848029453e-05, + "loss": 0.6065, + "step": 4215 + }, + { + "epoch": 8.123195380173243, + "grad_norm": 0.3209008276462555, + "learning_rate": 2.0536892222052128e-05, + "loss": 0.608, + "step": 4220 + }, + { + "epoch": 8.132820019249278, + "grad_norm": 0.34929510951042175, + "learning_rate": 2.0333184602589962e-05, + "loss": 0.6125, + "step": 4225 + }, + { + "epoch": 8.142444658325314, + "grad_norm": 0.34042608737945557, + "learning_rate": 2.01303779256131e-05, + "loss": 0.6094, + "step": 4230 + }, + { + "epoch": 8.152069297401347, + "grad_norm": 0.33042535185813904, + "learning_rate": 1.992847448463798e-05, + "loss": 0.6122, + "step": 4235 + }, + { + "epoch": 8.161693936477382, + "grad_norm": 0.3154657781124115, + "learning_rate": 1.9727476562966508e-05, + "loss": 0.6141, + "step": 4240 + }, + { + "epoch": 8.171318575553416, + "grad_norm": 0.33518335223197937, + "learning_rate": 1.952738643366011e-05, + "loss": 0.6139, + "step": 4245 + }, + { + "epoch": 8.180943214629451, + "grad_norm": 0.3391817510128021, + "learning_rate": 1.9328206359514155e-05, + "loss": 0.6106, + "step": 4250 + }, + { + "epoch": 8.190567853705486, + "grad_norm": 0.33157217502593994, + "learning_rate": 1.9129938593032227e-05, + "loss": 0.6051, + "step": 4255 + }, + { + "epoch": 8.20019249278152, + "grad_norm": 0.3601199686527252, + "learning_rate": 1.8932585376400803e-05, + "loss": 0.6127, + "step": 4260 + }, + { + "epoch": 8.209817131857555, + "grad_norm": 0.3452966511249542, + "learning_rate": 1.8736148941463795e-05, + "loss": 0.6162, + "step": 4265 + }, + { + "epoch": 8.21944177093359, + "grad_norm": 0.3637758791446686, + "learning_rate": 1.854063150969737e-05, + "loss": 0.6232, + "step": 4270 + }, + { + "epoch": 8.229066410009624, + "grad_norm": 0.3771421015262604, + "learning_rate": 1.834603529218475e-05, + "loss": 0.6066, + "step": 4275 + }, + { + "epoch": 8.23869104908566, + "grad_norm": 0.338925302028656, + "learning_rate": 1.81523624895913e-05, + "loss": 0.6155, + "step": 4280 + }, + { + "epoch": 8.248315688161695, + "grad_norm": 0.3534870147705078, + "learning_rate": 1.7959615292139544e-05, + "loss": 0.614, + "step": 4285 + }, + { + "epoch": 8.257940327237728, + "grad_norm": 0.33125004172325134, + "learning_rate": 1.7767795879584504e-05, + "loss": 0.6175, + "step": 4290 + }, + { + "epoch": 8.267564966313763, + "grad_norm": 0.3411141037940979, + "learning_rate": 1.7576906421188967e-05, + "loss": 0.6114, + "step": 4295 + }, + { + "epoch": 8.277189605389799, + "grad_norm": 0.3340323865413666, + "learning_rate": 1.738694907569901e-05, + "loss": 0.6233, + "step": 4300 + }, + { + "epoch": 8.286814244465832, + "grad_norm": 0.3233914375305176, + "learning_rate": 1.7197925991319486e-05, + "loss": 0.6082, + "step": 4305 + }, + { + "epoch": 8.296438883541867, + "grad_norm": 0.3364531099796295, + "learning_rate": 1.7009839305689855e-05, + "loss": 0.6049, + "step": 4310 + }, + { + "epoch": 8.306063522617901, + "grad_norm": 0.34157273173332214, + "learning_rate": 1.682269114585996e-05, + "loss": 0.6141, + "step": 4315 + }, + { + "epoch": 8.315688161693936, + "grad_norm": 0.33447617292404175, + "learning_rate": 1.6636483628265942e-05, + "loss": 0.6093, + "step": 4320 + }, + { + "epoch": 8.325312800769971, + "grad_norm": 0.33221328258514404, + "learning_rate": 1.6451218858706374e-05, + "loss": 0.6073, + "step": 4325 + }, + { + "epoch": 8.334937439846005, + "grad_norm": 0.32823801040649414, + "learning_rate": 1.626689893231832e-05, + "loss": 0.6069, + "step": 4330 + }, + { + "epoch": 8.34456207892204, + "grad_norm": 0.3583478629589081, + "learning_rate": 1.60835259335538e-05, + "loss": 0.6171, + "step": 4335 + }, + { + "epoch": 8.354186717998076, + "grad_norm": 0.33178088068962097, + "learning_rate": 1.5901101936156136e-05, + "loss": 0.6066, + "step": 4340 + }, + { + "epoch": 8.363811357074109, + "grad_norm": 0.3466804623603821, + "learning_rate": 1.5719629003136506e-05, + "loss": 0.6023, + "step": 4345 + }, + { + "epoch": 8.373435996150144, + "grad_norm": 0.357316792011261, + "learning_rate": 1.5539109186750544e-05, + "loss": 0.6059, + "step": 4350 + }, + { + "epoch": 8.38306063522618, + "grad_norm": 0.3246915340423584, + "learning_rate": 1.5359544528475323e-05, + "loss": 0.6231, + "step": 4355 + }, + { + "epoch": 8.392685274302213, + "grad_norm": 0.3579736649990082, + "learning_rate": 1.5180937058986033e-05, + "loss": 0.617, + "step": 4360 + }, + { + "epoch": 8.402309913378248, + "grad_norm": 0.33767664432525635, + "learning_rate": 1.5003288798133198e-05, + "loss": 0.6135, + "step": 4365 + }, + { + "epoch": 8.411934552454284, + "grad_norm": 0.34384191036224365, + "learning_rate": 1.4826601754919755e-05, + "loss": 0.6045, + "step": 4370 + }, + { + "epoch": 8.421559191530317, + "grad_norm": 0.34475091099739075, + "learning_rate": 1.4650877927478357e-05, + "loss": 0.611, + "step": 4375 + }, + { + "epoch": 8.431183830606352, + "grad_norm": 0.3544045686721802, + "learning_rate": 1.4476119303048707e-05, + "loss": 0.6048, + "step": 4380 + }, + { + "epoch": 8.440808469682388, + "grad_norm": 0.3278457820415497, + "learning_rate": 1.43023278579552e-05, + "loss": 0.6216, + "step": 4385 + }, + { + "epoch": 8.450433108758421, + "grad_norm": 0.33195823431015015, + "learning_rate": 1.4129505557584511e-05, + "loss": 0.6106, + "step": 4390 + }, + { + "epoch": 8.460057747834457, + "grad_norm": 0.32435399293899536, + "learning_rate": 1.3957654356363349e-05, + "loss": 0.6142, + "step": 4395 + }, + { + "epoch": 8.46968238691049, + "grad_norm": 0.34540995955467224, + "learning_rate": 1.3786776197736417e-05, + "loss": 0.6112, + "step": 4400 + }, + { + "epoch": 8.479307025986525, + "grad_norm": 0.3274092972278595, + "learning_rate": 1.3616873014144327e-05, + "loss": 0.6151, + "step": 4405 + }, + { + "epoch": 8.48893166506256, + "grad_norm": 0.3616076409816742, + "learning_rate": 1.3447946727001881e-05, + "loss": 0.6167, + "step": 4410 + }, + { + "epoch": 8.498556304138594, + "grad_norm": 0.32997846603393555, + "learning_rate": 1.3279999246676256e-05, + "loss": 0.611, + "step": 4415 + }, + { + "epoch": 8.50818094321463, + "grad_norm": 0.34430432319641113, + "learning_rate": 1.3113032472465426e-05, + "loss": 0.613, + "step": 4420 + }, + { + "epoch": 8.517805582290665, + "grad_norm": 0.35246655344963074, + "learning_rate": 1.2947048292576636e-05, + "loss": 0.6133, + "step": 4425 + }, + { + "epoch": 8.527430221366698, + "grad_norm": 0.3330981433391571, + "learning_rate": 1.2782048584105166e-05, + "loss": 0.615, + "step": 4430 + }, + { + "epoch": 8.537054860442733, + "grad_norm": 0.33830517530441284, + "learning_rate": 1.2618035213012924e-05, + "loss": 0.6175, + "step": 4435 + }, + { + "epoch": 8.546679499518769, + "grad_norm": 0.3427278399467468, + "learning_rate": 1.2455010034107527e-05, + "loss": 0.6111, + "step": 4440 + }, + { + "epoch": 8.556304138594802, + "grad_norm": 0.3526034355163574, + "learning_rate": 1.2292974891021236e-05, + "loss": 0.6135, + "step": 4445 + }, + { + "epoch": 8.565928777670837, + "grad_norm": 0.3584502935409546, + "learning_rate": 1.2131931616190118e-05, + "loss": 0.6143, + "step": 4450 + }, + { + "epoch": 8.575553416746873, + "grad_norm": 0.32676076889038086, + "learning_rate": 1.1971882030833248e-05, + "loss": 0.6092, + "step": 4455 + }, + { + "epoch": 8.585178055822906, + "grad_norm": 0.3570641279220581, + "learning_rate": 1.181282794493227e-05, + "loss": 0.6101, + "step": 4460 + }, + { + "epoch": 8.594802694898942, + "grad_norm": 0.35699462890625, + "learning_rate": 1.165477115721083e-05, + "loss": 0.6116, + "step": 4465 + }, + { + "epoch": 8.604427333974975, + "grad_norm": 0.3642681837081909, + "learning_rate": 1.1497713455114212e-05, + "loss": 0.6204, + "step": 4470 + }, + { + "epoch": 8.61405197305101, + "grad_norm": 0.34195858240127563, + "learning_rate": 1.1341656614789208e-05, + "loss": 0.6105, + "step": 4475 + }, + { + "epoch": 8.623676612127046, + "grad_norm": 0.3449951410293579, + "learning_rate": 1.1186602401063917e-05, + "loss": 0.6061, + "step": 4480 + }, + { + "epoch": 8.63330125120308, + "grad_norm": 0.3435938358306885, + "learning_rate": 1.1032552567427912e-05, + "loss": 0.6097, + "step": 4485 + }, + { + "epoch": 8.642925890279114, + "grad_norm": 0.3187827169895172, + "learning_rate": 1.0879508856012366e-05, + "loss": 0.6022, + "step": 4490 + }, + { + "epoch": 8.65255052935515, + "grad_norm": 0.3434700667858124, + "learning_rate": 1.0727472997570243e-05, + "loss": 0.6116, + "step": 4495 + }, + { + "epoch": 8.662175168431183, + "grad_norm": 0.34856435656547546, + "learning_rate": 1.0576446711456933e-05, + "loss": 0.605, + "step": 4500 + }, + { + "epoch": 8.671799807507218, + "grad_norm": 0.3594229817390442, + "learning_rate": 1.0426431705610606e-05, + "loss": 0.6133, + "step": 4505 + }, + { + "epoch": 8.681424446583254, + "grad_norm": 0.3380817174911499, + "learning_rate": 1.0277429676533023e-05, + "loss": 0.6073, + "step": 4510 + }, + { + "epoch": 8.691049085659287, + "grad_norm": 0.3276160955429077, + "learning_rate": 1.012944230927031e-05, + "loss": 0.6021, + "step": 4515 + }, + { + "epoch": 8.700673724735323, + "grad_norm": 0.34987348318099976, + "learning_rate": 9.9824712773939e-06, + "loss": 0.617, + "step": 4520 + }, + { + "epoch": 8.710298363811358, + "grad_norm": 0.3415302336215973, + "learning_rate": 9.83651824298164e-06, + "loss": 0.6111, + "step": 4525 + }, + { + "epoch": 8.719923002887391, + "grad_norm": 0.34866005182266235, + "learning_rate": 9.69158485659889e-06, + "loss": 0.603, + "step": 4530 + }, + { + "epoch": 8.729547641963427, + "grad_norm": 0.36085546016693115, + "learning_rate": 9.547672757280001e-06, + "loss": 0.6042, + "step": 4535 + }, + { + "epoch": 8.739172281039462, + "grad_norm": 0.36267852783203125, + "learning_rate": 9.40478357250969e-06, + "loss": 0.6127, + "step": 4540 + }, + { + "epoch": 8.748796920115495, + "grad_norm": 0.36462917923927307, + "learning_rate": 9.262918918204643e-06, + "loss": 0.6123, + "step": 4545 + }, + { + "epoch": 8.75842155919153, + "grad_norm": 0.34768378734588623, + "learning_rate": 9.122080398695299e-06, + "loss": 0.6048, + "step": 4550 + }, + { + "epoch": 8.768046198267564, + "grad_norm": 0.330387681722641, + "learning_rate": 8.982269606707593e-06, + "loss": 0.6165, + "step": 4555 + }, + { + "epoch": 8.7776708373436, + "grad_norm": 0.3596397936344147, + "learning_rate": 8.843488123345044e-06, + "loss": 0.6072, + "step": 4560 + }, + { + "epoch": 8.787295476419635, + "grad_norm": 0.35082703828811646, + "learning_rate": 8.705737518070888e-06, + "loss": 0.6185, + "step": 4565 + }, + { + "epoch": 8.796920115495668, + "grad_norm": 0.33255165815353394, + "learning_rate": 8.569019348690189e-06, + "loss": 0.6099, + "step": 4570 + }, + { + "epoch": 8.806544754571703, + "grad_norm": 0.3488062620162964, + "learning_rate": 8.433335161332412e-06, + "loss": 0.6056, + "step": 4575 + }, + { + "epoch": 8.816169393647739, + "grad_norm": 0.35131949186325073, + "learning_rate": 8.298686490433771e-06, + "loss": 0.6102, + "step": 4580 + }, + { + "epoch": 8.825794032723772, + "grad_norm": 0.37358999252319336, + "learning_rate": 8.165074858719989e-06, + "loss": 0.6103, + "step": 4585 + }, + { + "epoch": 8.835418671799808, + "grad_norm": 0.35089996457099915, + "learning_rate": 8.032501777189017e-06, + "loss": 0.6112, + "step": 4590 + }, + { + "epoch": 8.845043310875843, + "grad_norm": 0.35341012477874756, + "learning_rate": 7.900968745093996e-06, + "loss": 0.6089, + "step": 4595 + }, + { + "epoch": 8.854667949951876, + "grad_norm": 0.3873613178730011, + "learning_rate": 7.770477249926256e-06, + "loss": 0.6111, + "step": 4600 + }, + { + "epoch": 8.864292589027912, + "grad_norm": 0.34750309586524963, + "learning_rate": 7.641028767398472e-06, + "loss": 0.616, + "step": 4605 + }, + { + "epoch": 8.873917228103947, + "grad_norm": 0.32477355003356934, + "learning_rate": 7.512624761428066e-06, + "loss": 0.6089, + "step": 4610 + }, + { + "epoch": 8.88354186717998, + "grad_norm": 0.35710757970809937, + "learning_rate": 7.385266684120573e-06, + "loss": 0.61, + "step": 4615 + }, + { + "epoch": 8.893166506256016, + "grad_norm": 0.34388595819473267, + "learning_rate": 7.258955975753279e-06, + "loss": 0.6076, + "step": 4620 + }, + { + "epoch": 8.90279114533205, + "grad_norm": 0.32944580912590027, + "learning_rate": 7.133694064758867e-06, + "loss": 0.606, + "step": 4625 + }, + { + "epoch": 8.912415784408084, + "grad_norm": 0.3470548093318939, + "learning_rate": 7.0094823677092856e-06, + "loss": 0.6015, + "step": 4630 + }, + { + "epoch": 8.92204042348412, + "grad_norm": 0.3423613905906677, + "learning_rate": 6.886322289299763e-06, + "loss": 0.6155, + "step": 4635 + }, + { + "epoch": 8.931665062560153, + "grad_norm": 0.35634317994117737, + "learning_rate": 6.764215222332914e-06, + "loss": 0.6146, + "step": 4640 + }, + { + "epoch": 8.941289701636189, + "grad_norm": 0.33485671877861023, + "learning_rate": 6.643162547702931e-06, + "loss": 0.6135, + "step": 4645 + }, + { + "epoch": 8.950914340712224, + "grad_norm": 0.35238829255104065, + "learning_rate": 6.523165634380046e-06, + "loss": 0.6044, + "step": 4650 + }, + { + "epoch": 8.960538979788257, + "grad_norm": 0.3438652753829956, + "learning_rate": 6.404225839394973e-06, + "loss": 0.6107, + "step": 4655 + }, + { + "epoch": 8.970163618864293, + "grad_norm": 0.352061003446579, + "learning_rate": 6.286344507823638e-06, + "loss": 0.6164, + "step": 4660 + }, + { + "epoch": 8.979788257940328, + "grad_norm": 0.3431857228279114, + "learning_rate": 6.169522972771924e-06, + "loss": 0.6144, + "step": 4665 + }, + { + "epoch": 8.989412897016361, + "grad_norm": 0.32378876209259033, + "learning_rate": 6.053762555360587e-06, + "loss": 0.6162, + "step": 4670 + }, + { + "epoch": 8.999037536092397, + "grad_norm": 0.36266306042671204, + "learning_rate": 5.939064564710373e-06, + "loss": 0.6132, + "step": 4675 + }, + { + "epoch": 8.999037536092397, + "eval_loss": 3.9708144664764404, + "eval_runtime": 0.7877, + "eval_samples_per_second": 13.964, + "eval_steps_per_second": 2.539, + "step": 4675 + }, + { + "epoch": 9.008662175168432, + "grad_norm": 0.3068545162677765, + "learning_rate": 5.825430297927092e-06, + "loss": 0.5915, + "step": 4680 + }, + { + "epoch": 9.018286814244465, + "grad_norm": 0.3031752407550812, + "learning_rate": 5.712861040087092e-06, + "loss": 0.586, + "step": 4685 + }, + { + "epoch": 9.0279114533205, + "grad_norm": 0.33787086606025696, + "learning_rate": 5.601358064222639e-06, + "loss": 0.5911, + "step": 4690 + }, + { + "epoch": 9.037536092396536, + "grad_norm": 0.35586461424827576, + "learning_rate": 5.49092263130756e-06, + "loss": 0.5828, + "step": 4695 + }, + { + "epoch": 9.04716073147257, + "grad_norm": 0.3516261875629425, + "learning_rate": 5.381555990242959e-06, + "loss": 0.5847, + "step": 4700 + }, + { + "epoch": 9.056785370548605, + "grad_norm": 0.34338730573654175, + "learning_rate": 5.273259377843087e-06, + "loss": 0.6036, + "step": 4705 + }, + { + "epoch": 9.066410009624638, + "grad_norm": 0.3557838499546051, + "learning_rate": 5.166034018821364e-06, + "loss": 0.5939, + "step": 4710 + }, + { + "epoch": 9.076034648700674, + "grad_norm": 0.31932586431503296, + "learning_rate": 5.059881125776589e-06, + "loss": 0.6016, + "step": 4715 + }, + { + "epoch": 9.085659287776709, + "grad_norm": 0.3272048532962799, + "learning_rate": 4.9548018991790846e-06, + "loss": 0.5909, + "step": 4720 + }, + { + "epoch": 9.095283926852742, + "grad_norm": 0.3446064889431, + "learning_rate": 4.850797527357287e-06, + "loss": 0.5827, + "step": 4725 + }, + { + "epoch": 9.104908565928778, + "grad_norm": 0.32635557651519775, + "learning_rate": 4.747869186484177e-06, + "loss": 0.5921, + "step": 4730 + }, + { + "epoch": 9.114533205004813, + "grad_norm": 0.31974223256111145, + "learning_rate": 4.64601804056406e-06, + "loss": 0.5932, + "step": 4735 + }, + { + "epoch": 9.124157844080846, + "grad_norm": 0.3654205799102783, + "learning_rate": 4.545245241419349e-06, + "loss": 0.5995, + "step": 4740 + }, + { + "epoch": 9.133782483156882, + "grad_norm": 0.35849812626838684, + "learning_rate": 4.445551928677594e-06, + "loss": 0.5995, + "step": 4745 + }, + { + "epoch": 9.143407122232917, + "grad_norm": 0.3359050750732422, + "learning_rate": 4.346939229758529e-06, + "loss": 0.5982, + "step": 4750 + }, + { + "epoch": 9.15303176130895, + "grad_norm": 0.33533555269241333, + "learning_rate": 4.2494082598613875e-06, + "loss": 0.6007, + "step": 4755 + }, + { + "epoch": 9.162656400384986, + "grad_norm": 0.3292589783668518, + "learning_rate": 4.152960121952209e-06, + "loss": 0.5974, + "step": 4760 + }, + { + "epoch": 9.172281039461021, + "grad_norm": 0.34592679142951965, + "learning_rate": 4.057595906751466e-06, + "loss": 0.5922, + "step": 4765 + }, + { + "epoch": 9.181905678537055, + "grad_norm": 0.34907424449920654, + "learning_rate": 3.963316692721663e-06, + "loss": 0.6007, + "step": 4770 + }, + { + "epoch": 9.19153031761309, + "grad_norm": 0.3478921949863434, + "learning_rate": 3.870123546055149e-06, + "loss": 0.5882, + "step": 4775 + }, + { + "epoch": 9.201154956689123, + "grad_norm": 0.3408016860485077, + "learning_rate": 3.7780175206620915e-06, + "loss": 0.595, + "step": 4780 + }, + { + "epoch": 9.210779595765159, + "grad_norm": 0.33491307497024536, + "learning_rate": 3.686999658158474e-06, + "loss": 0.5951, + "step": 4785 + }, + { + "epoch": 9.220404234841194, + "grad_norm": 0.3383229672908783, + "learning_rate": 3.597070987854456e-06, + "loss": 0.5966, + "step": 4790 + }, + { + "epoch": 9.230028873917227, + "grad_norm": 0.3315028250217438, + "learning_rate": 3.508232526742583e-06, + "loss": 0.5959, + "step": 4795 + }, + { + "epoch": 9.239653512993263, + "grad_norm": 0.30691462755203247, + "learning_rate": 3.420485279486385e-06, + "loss": 0.5853, + "step": 4800 + }, + { + "epoch": 9.249278152069298, + "grad_norm": 0.34303727746009827, + "learning_rate": 3.333830238409019e-06, + "loss": 0.5973, + "step": 4805 + }, + { + "epoch": 9.258902791145331, + "grad_norm": 0.3458213210105896, + "learning_rate": 3.248268383481934e-06, + "loss": 0.5978, + "step": 4810 + }, + { + "epoch": 9.268527430221367, + "grad_norm": 0.3539816737174988, + "learning_rate": 3.163800682313933e-06, + "loss": 0.5958, + "step": 4815 + }, + { + "epoch": 9.278152069297402, + "grad_norm": 0.3442062735557556, + "learning_rate": 3.080428090140142e-06, + "loss": 0.6022, + "step": 4820 + }, + { + "epoch": 9.287776708373435, + "grad_norm": 0.3180767893791199, + "learning_rate": 2.9981515498112456e-06, + "loss": 0.5955, + "step": 4825 + }, + { + "epoch": 9.29740134744947, + "grad_norm": 0.34698548913002014, + "learning_rate": 2.91697199178278e-06, + "loss": 0.5947, + "step": 4830 + }, + { + "epoch": 9.307025986525506, + "grad_norm": 0.3273780047893524, + "learning_rate": 2.8368903341046583e-06, + "loss": 0.5998, + "step": 4835 + }, + { + "epoch": 9.31665062560154, + "grad_norm": 0.31761637330055237, + "learning_rate": 2.757907482410771e-06, + "loss": 0.5841, + "step": 4840 + }, + { + "epoch": 9.326275264677575, + "grad_norm": 0.3708135783672333, + "learning_rate": 2.680024329908737e-06, + "loss": 0.5953, + "step": 4845 + }, + { + "epoch": 9.33589990375361, + "grad_norm": 0.309467613697052, + "learning_rate": 2.603241757369812e-06, + "loss": 0.5969, + "step": 4850 + }, + { + "epoch": 9.345524542829644, + "grad_norm": 0.32634660601615906, + "learning_rate": 2.5275606331189416e-06, + "loss": 0.602, + "step": 4855 + }, + { + "epoch": 9.355149181905679, + "grad_norm": 0.33582308888435364, + "learning_rate": 2.452981813024868e-06, + "loss": 0.5875, + "step": 4860 + }, + { + "epoch": 9.364773820981712, + "grad_norm": 0.3333386182785034, + "learning_rate": 2.379506140490595e-06, + "loss": 0.5986, + "step": 4865 + }, + { + "epoch": 9.374398460057748, + "grad_norm": 0.35826408863067627, + "learning_rate": 2.3071344464436595e-06, + "loss": 0.6015, + "step": 4870 + }, + { + "epoch": 9.384023099133783, + "grad_norm": 0.334588885307312, + "learning_rate": 2.235867549326931e-06, + "loss": 0.5942, + "step": 4875 + }, + { + "epoch": 9.393647738209816, + "grad_norm": 0.3338033854961395, + "learning_rate": 2.165706255089217e-06, + "loss": 0.5991, + "step": 4880 + }, + { + "epoch": 9.403272377285852, + "grad_norm": 0.3354242742061615, + "learning_rate": 2.0966513571761827e-06, + "loss": 0.5991, + "step": 4885 + }, + { + "epoch": 9.412897016361887, + "grad_norm": 0.34545251727104187, + "learning_rate": 2.028703636521434e-06, + "loss": 0.6058, + "step": 4890 + }, + { + "epoch": 9.42252165543792, + "grad_norm": 0.33035480976104736, + "learning_rate": 1.961863861537594e-06, + "loss": 0.5981, + "step": 4895 + }, + { + "epoch": 9.432146294513956, + "grad_norm": 0.33753854036331177, + "learning_rate": 1.8961327881076963e-06, + "loss": 0.5944, + "step": 4900 + }, + { + "epoch": 9.441770933589991, + "grad_norm": 0.34246233105659485, + "learning_rate": 1.8315111595765932e-06, + "loss": 0.5931, + "step": 4905 + }, + { + "epoch": 9.451395572666025, + "grad_norm": 0.33052095770835876, + "learning_rate": 1.767999706742529e-06, + "loss": 0.5986, + "step": 4910 + }, + { + "epoch": 9.46102021174206, + "grad_norm": 0.35342252254486084, + "learning_rate": 1.7055991478489464e-06, + "loss": 0.5938, + "step": 4915 + }, + { + "epoch": 9.470644850818095, + "grad_norm": 0.33293551206588745, + "learning_rate": 1.6443101885762812e-06, + "loss": 0.5917, + "step": 4920 + }, + { + "epoch": 9.480269489894129, + "grad_norm": 0.3331868648529053, + "learning_rate": 1.5841335220340593e-06, + "loss": 0.5951, + "step": 4925 + }, + { + "epoch": 9.489894128970164, + "grad_norm": 0.35304731130599976, + "learning_rate": 1.525069828753012e-06, + "loss": 0.602, + "step": 4930 + }, + { + "epoch": 9.499518768046197, + "grad_norm": 0.3421652019023895, + "learning_rate": 1.4671197766773615e-06, + "loss": 0.5966, + "step": 4935 + }, + { + "epoch": 9.509143407122233, + "grad_norm": 0.3255125880241394, + "learning_rate": 1.4102840211573264e-06, + "loss": 0.5944, + "step": 4940 + }, + { + "epoch": 9.518768046198268, + "grad_norm": 0.34258726239204407, + "learning_rate": 1.3545632049416502e-06, + "loss": 0.5889, + "step": 4945 + }, + { + "epoch": 9.528392685274301, + "grad_norm": 0.3264661729335785, + "learning_rate": 1.2999579581703947e-06, + "loss": 0.5954, + "step": 4950 + }, + { + "epoch": 9.538017324350337, + "grad_norm": 0.3256395161151886, + "learning_rate": 1.2464688983677697e-06, + "loss": 0.5907, + "step": 4955 + }, + { + "epoch": 9.547641963426372, + "grad_norm": 0.32232365012168884, + "learning_rate": 1.1940966304351265e-06, + "loss": 0.5949, + "step": 4960 + }, + { + "epoch": 9.557266602502406, + "grad_norm": 0.32586029171943665, + "learning_rate": 1.1428417466442076e-06, + "loss": 0.5885, + "step": 4965 + }, + { + "epoch": 9.56689124157844, + "grad_norm": 0.3531622886657715, + "learning_rate": 1.0927048266303419e-06, + "loss": 0.6064, + "step": 4970 + }, + { + "epoch": 9.576515880654476, + "grad_norm": 0.34918224811553955, + "learning_rate": 1.0436864373859712e-06, + "loss": 0.6043, + "step": 4975 + }, + { + "epoch": 9.58614051973051, + "grad_norm": 0.3377608358860016, + "learning_rate": 9.95787133254189e-07, + "loss": 0.5869, + "step": 4980 + }, + { + "epoch": 9.595765158806545, + "grad_norm": 0.32988688349723816, + "learning_rate": 9.490074559225015e-07, + "loss": 0.5957, + "step": 4985 + }, + { + "epoch": 9.60538979788258, + "grad_norm": 0.3335455656051636, + "learning_rate": 9.033479344166873e-07, + "loss": 0.5901, + "step": 4990 + }, + { + "epoch": 9.615014436958614, + "grad_norm": 0.34015801548957825, + "learning_rate": 8.588090850948027e-07, + "loss": 0.5956, + "step": 4995 + }, + { + "epoch": 9.624639076034649, + "grad_norm": 0.32440024614334106, + "learning_rate": 8.153914116413752e-07, + "loss": 0.6035, + "step": 5000 + }, + { + "epoch": 9.634263715110684, + "grad_norm": 0.33188602328300476, + "learning_rate": 7.730954050616746e-07, + "loss": 0.6025, + "step": 5005 + }, + { + "epoch": 9.643888354186718, + "grad_norm": 0.33264580368995667, + "learning_rate": 7.319215436761839e-07, + "loss": 0.5973, + "step": 5010 + }, + { + "epoch": 9.653512993262753, + "grad_norm": 0.342488557100296, + "learning_rate": 6.918702931151711e-07, + "loss": 0.5914, + "step": 5015 + }, + { + "epoch": 9.663137632338787, + "grad_norm": 0.33260515332221985, + "learning_rate": 6.529421063134478e-07, + "loss": 0.5964, + "step": 5020 + }, + { + "epoch": 9.672762271414822, + "grad_norm": 0.358557790517807, + "learning_rate": 6.151374235051966e-07, + "loss": 0.6021, + "step": 5025 + }, + { + "epoch": 9.682386910490857, + "grad_norm": 0.341327965259552, + "learning_rate": 5.784566722190965e-07, + "loss": 0.5911, + "step": 5030 + }, + { + "epoch": 9.69201154956689, + "grad_norm": 0.31675535440444946, + "learning_rate": 5.429002672733274e-07, + "loss": 0.6015, + "step": 5035 + }, + { + "epoch": 9.701636188642926, + "grad_norm": 0.31824976205825806, + "learning_rate": 5.084686107710513e-07, + "loss": 0.599, + "step": 5040 + }, + { + "epoch": 9.711260827718961, + "grad_norm": 0.3493671715259552, + "learning_rate": 4.751620920957489e-07, + "loss": 0.596, + "step": 5045 + }, + { + "epoch": 9.720885466794995, + "grad_norm": 0.34269365668296814, + "learning_rate": 4.429810879068463e-07, + "loss": 0.5969, + "step": 5050 + }, + { + "epoch": 9.73051010587103, + "grad_norm": 0.3367815613746643, + "learning_rate": 4.1192596213548427e-07, + "loss": 0.5885, + "step": 5055 + }, + { + "epoch": 9.740134744947065, + "grad_norm": 0.34025177359580994, + "learning_rate": 3.81997065980344e-07, + "loss": 0.6051, + "step": 5060 + }, + { + "epoch": 9.749759384023099, + "grad_norm": 0.3241323232650757, + "learning_rate": 3.5319473790373924e-07, + "loss": 0.5914, + "step": 5065 + }, + { + "epoch": 9.759384023099134, + "grad_norm": 0.3496091961860657, + "learning_rate": 3.2551930362776373e-07, + "loss": 0.5962, + "step": 5070 + }, + { + "epoch": 9.769008662175168, + "grad_norm": 0.38736647367477417, + "learning_rate": 2.989710761305942e-07, + "loss": 0.5941, + "step": 5075 + }, + { + "epoch": 9.778633301251203, + "grad_norm": 0.33493003249168396, + "learning_rate": 2.7355035564294865e-07, + "loss": 0.5971, + "step": 5080 + }, + { + "epoch": 9.788257940327238, + "grad_norm": 0.3347594738006592, + "learning_rate": 2.4925742964471144e-07, + "loss": 0.605, + "step": 5085 + }, + { + "epoch": 9.797882579403272, + "grad_norm": 0.3406401574611664, + "learning_rate": 2.2609257286169138e-07, + "loss": 0.5876, + "step": 5090 + }, + { + "epoch": 9.807507218479307, + "grad_norm": 0.34672555327415466, + "learning_rate": 2.0405604726246864e-07, + "loss": 0.5979, + "step": 5095 + }, + { + "epoch": 9.817131857555342, + "grad_norm": 0.3294496238231659, + "learning_rate": 1.8314810205547483e-07, + "loss": 0.584, + "step": 5100 + }, + { + "epoch": 9.826756496631376, + "grad_norm": 0.33348360657691956, + "learning_rate": 1.633689736861732e-07, + "loss": 0.5945, + "step": 5105 + }, + { + "epoch": 9.836381135707411, + "grad_norm": 0.32033050060272217, + "learning_rate": 1.4471888583436067e-07, + "loss": 0.5952, + "step": 5110 + }, + { + "epoch": 9.846005774783446, + "grad_norm": 0.35708528757095337, + "learning_rate": 1.2719804941163648e-07, + "loss": 0.5936, + "step": 5115 + }, + { + "epoch": 9.85563041385948, + "grad_norm": 0.34551671147346497, + "learning_rate": 1.108066625590487e-07, + "loss": 0.5902, + "step": 5120 + }, + { + "epoch": 9.865255052935515, + "grad_norm": 0.32259657979011536, + "learning_rate": 9.554491064484028e-08, + "loss": 0.5976, + "step": 5125 + }, + { + "epoch": 9.87487969201155, + "grad_norm": 0.3397790491580963, + "learning_rate": 8.141296626231754e-08, + "loss": 0.6072, + "step": 5130 + }, + { + "epoch": 9.884504331087584, + "grad_norm": 0.3509461283683777, + "learning_rate": 6.841098922797384e-08, + "loss": 0.6013, + "step": 5135 + }, + { + "epoch": 9.894128970163619, + "grad_norm": 0.3350575268268585, + "learning_rate": 5.653912657959115e-08, + "loss": 0.6012, + "step": 5140 + }, + { + "epoch": 9.903753609239654, + "grad_norm": 0.3318527042865753, + "learning_rate": 4.579751257466347e-08, + "loss": 0.6048, + "step": 5145 + }, + { + "epoch": 9.913378248315688, + "grad_norm": 0.37916940450668335, + "learning_rate": 3.618626868879815e-08, + "loss": 0.6068, + "step": 5150 + }, + { + "epoch": 9.923002887391723, + "grad_norm": 0.3555992841720581, + "learning_rate": 2.7705503614416928e-08, + "loss": 0.588, + "step": 5155 + }, + { + "epoch": 9.932627526467758, + "grad_norm": 0.32007142901420593, + "learning_rate": 2.0355313259468046e-08, + "loss": 0.5954, + "step": 5160 + }, + { + "epoch": 9.942252165543792, + "grad_norm": 0.3455217182636261, + "learning_rate": 1.4135780746382665e-08, + "loss": 0.5917, + "step": 5165 + }, + { + "epoch": 9.951876804619827, + "grad_norm": 0.32252103090286255, + "learning_rate": 9.046976411108965e-09, + "loss": 0.5889, + "step": 5170 + }, + { + "epoch": 9.96150144369586, + "grad_norm": 0.3604857921600342, + "learning_rate": 5.0889578023238794e-09, + "loss": 0.5959, + "step": 5175 + }, + { + "epoch": 9.971126082771896, + "grad_norm": 0.33323296904563904, + "learning_rate": 2.261769680789172e-09, + "loss": 0.5918, + "step": 5180 + }, + { + "epoch": 9.980750721847931, + "grad_norm": 0.33578982949256897, + "learning_rate": 5.654440188296306e-10, + "loss": 0.6011, + "step": 5185 + }, + { + "epoch": 9.990375360923965, + "grad_norm": 0.34376034140586853, + "learning_rate": 0.0, + "loss": 0.5965, + "step": 5190 + }, + { + "epoch": 9.990375360923965, + "eval_loss": 4.085933685302734, + "eval_runtime": 0.8044, + "eval_samples_per_second": 13.675, + "eval_steps_per_second": 2.486, + "step": 5190 + }, + { + "epoch": 9.990375360923965, + "step": 5190, + "total_flos": 7.743588771836199e+18, + "train_loss": 0.8018066772835792, + "train_runtime": 21791.6644, + "train_samples_per_second": 7.627, + "train_steps_per_second": 0.238 + } + ], + "logging_steps": 5, + "max_steps": 5190, + "num_input_tokens_seen": 0, + "num_train_epochs": 10, + "save_steps": 100, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 7.743588771836199e+18, + "train_batch_size": 4, + "trial_name": null, + "trial_params": null +}