{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.6299212598425197, "eval_steps": 20, "global_step": 2000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00031496062992125983, "grad_norm": NaN, "learning_rate": 1e-05, "loss": 0.6279, "step": 1 }, { "epoch": 0.0006299212598425197, "grad_norm": NaN, "learning_rate": 1e-05, "loss": 0.7073, "step": 2 }, { "epoch": 0.0009448818897637795, "grad_norm": Infinity, "learning_rate": 1e-05, "loss": 0.4724, "step": 3 }, { "epoch": 0.0012598425196850393, "grad_norm": Infinity, "learning_rate": 1e-05, "loss": 0.8548, "step": 4 }, { "epoch": 0.0015748031496062992, "grad_norm": 241.3546600341797, "learning_rate": 9.99999842519685e-06, "loss": 0.9812, "step": 5 }, { "epoch": 0.001889763779527559, "grad_norm": 68.62373352050781, "learning_rate": 9.999996850393701e-06, "loss": 0.6851, "step": 6 }, { "epoch": 0.002204724409448819, "grad_norm": 45.99556350708008, "learning_rate": 9.999995275590552e-06, "loss": 0.7491, "step": 7 }, { "epoch": 0.0025196850393700786, "grad_norm": 27.5998592376709, "learning_rate": 9.999993700787403e-06, "loss": 0.6691, "step": 8 }, { "epoch": 0.0028346456692913387, "grad_norm": 53.23931121826172, "learning_rate": 9.999992125984252e-06, "loss": 0.7518, "step": 9 }, { "epoch": 0.0031496062992125984, "grad_norm": 21.858654022216797, "learning_rate": 9.999990551181103e-06, "loss": 0.6986, "step": 10 }, { "epoch": 0.0034645669291338585, "grad_norm": 38.456905364990234, "learning_rate": 9.999988976377953e-06, "loss": 0.6296, "step": 11 }, { "epoch": 0.003779527559055118, "grad_norm": 36.96352005004883, "learning_rate": 9.999987401574804e-06, "loss": 0.6674, "step": 12 }, { "epoch": 0.004094488188976378, "grad_norm": 50.633941650390625, "learning_rate": 9.999985826771655e-06, "loss": 0.6214, "step": 13 }, { "epoch": 0.004409448818897638, "grad_norm": 90.81790161132812, "learning_rate": 9.999984251968506e-06, "loss": 0.7667, "step": 14 }, { "epoch": 0.004724409448818898, "grad_norm": 157.15757751464844, "learning_rate": 9.999982677165355e-06, "loss": 1.0109, "step": 15 }, { "epoch": 0.005039370078740157, "grad_norm": 57.94607925415039, "learning_rate": 9.999981102362206e-06, "loss": 0.6623, "step": 16 }, { "epoch": 0.005354330708661417, "grad_norm": 96.26383972167969, "learning_rate": 9.999979527559057e-06, "loss": 0.7511, "step": 17 }, { "epoch": 0.005669291338582677, "grad_norm": 63.17537307739258, "learning_rate": 9.999977952755906e-06, "loss": 0.7292, "step": 18 }, { "epoch": 0.005984251968503937, "grad_norm": 28.436891555786133, "learning_rate": 9.999976377952757e-06, "loss": 0.6612, "step": 19 }, { "epoch": 0.006299212598425197, "grad_norm": 46.60204315185547, "learning_rate": 9.999974803149607e-06, "loss": 0.471, "step": 20 }, { "epoch": 0.006299212598425197, "eval_loss": 0.6495372653007507, "eval_runtime": 323.1698, "eval_samples_per_second": 0.362, "eval_steps_per_second": 0.362, "step": 20 }, { "epoch": 0.006614173228346456, "grad_norm": 34.931636810302734, "learning_rate": 9.999973228346457e-06, "loss": 0.6546, "step": 21 }, { "epoch": 0.006929133858267717, "grad_norm": 37.24106216430664, "learning_rate": 9.999971653543308e-06, "loss": 0.5786, "step": 22 }, { "epoch": 0.007244094488188977, "grad_norm": 49.713348388671875, "learning_rate": 9.99997007874016e-06, "loss": 0.605, "step": 23 }, { "epoch": 0.007559055118110236, "grad_norm": 54.44386291503906, "learning_rate": 9.999968503937009e-06, "loss": 0.3443, "step": 24 }, { "epoch": 0.007874015748031496, "grad_norm": 79.05270385742188, "learning_rate": 9.999966929133858e-06, "loss": 0.6105, "step": 25 }, { "epoch": 0.008188976377952756, "grad_norm": 77.4219741821289, "learning_rate": 9.999965354330709e-06, "loss": 0.864, "step": 26 }, { "epoch": 0.008503937007874015, "grad_norm": 137.48190307617188, "learning_rate": 9.99996377952756e-06, "loss": 1.1885, "step": 27 }, { "epoch": 0.008818897637795276, "grad_norm": 156.0934600830078, "learning_rate": 9.999962204724411e-06, "loss": 1.1261, "step": 28 }, { "epoch": 0.009133858267716535, "grad_norm": 117.86957550048828, "learning_rate": 9.99996062992126e-06, "loss": 0.6776, "step": 29 }, { "epoch": 0.009448818897637795, "grad_norm": 107.25189971923828, "learning_rate": 9.999959055118111e-06, "loss": 0.7123, "step": 30 }, { "epoch": 0.009763779527559056, "grad_norm": 53.983299255371094, "learning_rate": 9.99995748031496e-06, "loss": 0.6213, "step": 31 }, { "epoch": 0.010078740157480314, "grad_norm": 60.340389251708984, "learning_rate": 9.999955905511812e-06, "loss": 0.7145, "step": 32 }, { "epoch": 0.010393700787401575, "grad_norm": 76.8556137084961, "learning_rate": 9.999954330708663e-06, "loss": 0.6812, "step": 33 }, { "epoch": 0.010708661417322834, "grad_norm": 45.225807189941406, "learning_rate": 9.999952755905514e-06, "loss": 0.6361, "step": 34 }, { "epoch": 0.011023622047244094, "grad_norm": 65.44268798828125, "learning_rate": 9.999951181102363e-06, "loss": 0.6421, "step": 35 }, { "epoch": 0.011338582677165355, "grad_norm": 42.70692825317383, "learning_rate": 9.999949606299212e-06, "loss": 0.6809, "step": 36 }, { "epoch": 0.011653543307086614, "grad_norm": 43.51832580566406, "learning_rate": 9.999948031496063e-06, "loss": 0.6612, "step": 37 }, { "epoch": 0.011968503937007874, "grad_norm": 37.09170913696289, "learning_rate": 9.999946456692914e-06, "loss": 0.7119, "step": 38 }, { "epoch": 0.012283464566929133, "grad_norm": 29.477069854736328, "learning_rate": 9.999944881889765e-06, "loss": 0.6933, "step": 39 }, { "epoch": 0.012598425196850394, "grad_norm": 44.96734619140625, "learning_rate": 9.999943307086614e-06, "loss": 0.7328, "step": 40 }, { "epoch": 0.012598425196850394, "eval_loss": 0.6671837568283081, "eval_runtime": 309.18, "eval_samples_per_second": 0.378, "eval_steps_per_second": 0.378, "step": 40 }, { "epoch": 0.012913385826771654, "grad_norm": 19.817779541015625, "learning_rate": 9.999941732283465e-06, "loss": 0.6673, "step": 41 }, { "epoch": 0.013228346456692913, "grad_norm": 23.819435119628906, "learning_rate": 9.999940157480316e-06, "loss": 0.6966, "step": 42 }, { "epoch": 0.013543307086614173, "grad_norm": 21.729511260986328, "learning_rate": 9.999938582677167e-06, "loss": 0.6791, "step": 43 }, { "epoch": 0.013858267716535434, "grad_norm": 18.30646514892578, "learning_rate": 9.999937007874017e-06, "loss": 0.6786, "step": 44 }, { "epoch": 0.014173228346456693, "grad_norm": 11.647773742675781, "learning_rate": 9.999935433070866e-06, "loss": 0.6956, "step": 45 }, { "epoch": 0.014488188976377953, "grad_norm": 15.509359359741211, "learning_rate": 9.999933858267717e-06, "loss": 0.6616, "step": 46 }, { "epoch": 0.014803149606299212, "grad_norm": 20.342838287353516, "learning_rate": 9.999932283464568e-06, "loss": 0.6184, "step": 47 }, { "epoch": 0.015118110236220473, "grad_norm": 15.10333251953125, "learning_rate": 9.999930708661419e-06, "loss": 0.6671, "step": 48 }, { "epoch": 0.015433070866141733, "grad_norm": 22.806962966918945, "learning_rate": 9.999929133858268e-06, "loss": 0.7039, "step": 49 }, { "epoch": 0.015748031496062992, "grad_norm": 22.755117416381836, "learning_rate": 9.99992755905512e-06, "loss": 0.6512, "step": 50 }, { "epoch": 0.016062992125984252, "grad_norm": 30.809261322021484, "learning_rate": 9.999925984251969e-06, "loss": 0.7041, "step": 51 }, { "epoch": 0.016377952755905513, "grad_norm": 30.779508590698242, "learning_rate": 9.99992440944882e-06, "loss": 0.6321, "step": 52 }, { "epoch": 0.01669291338582677, "grad_norm": 41.0311279296875, "learning_rate": 9.99992283464567e-06, "loss": 0.7655, "step": 53 }, { "epoch": 0.01700787401574803, "grad_norm": 38.755794525146484, "learning_rate": 9.999921259842522e-06, "loss": 0.5663, "step": 54 }, { "epoch": 0.01732283464566929, "grad_norm": 76.18267059326172, "learning_rate": 9.99991968503937e-06, "loss": 0.817, "step": 55 }, { "epoch": 0.01763779527559055, "grad_norm": 64.29126739501953, "learning_rate": 9.99991811023622e-06, "loss": 0.7239, "step": 56 }, { "epoch": 0.017952755905511812, "grad_norm": 64.24166107177734, "learning_rate": 9.999916535433071e-06, "loss": 0.5583, "step": 57 }, { "epoch": 0.01826771653543307, "grad_norm": 35.76716232299805, "learning_rate": 9.999914960629922e-06, "loss": 0.6473, "step": 58 }, { "epoch": 0.01858267716535433, "grad_norm": 31.879194259643555, "learning_rate": 9.999913385826773e-06, "loss": 0.5875, "step": 59 }, { "epoch": 0.01889763779527559, "grad_norm": 43.807613372802734, "learning_rate": 9.999911811023622e-06, "loss": 0.6888, "step": 60 }, { "epoch": 0.01889763779527559, "eval_loss": 0.6864338517189026, "eval_runtime": 308.0649, "eval_samples_per_second": 0.38, "eval_steps_per_second": 0.38, "step": 60 }, { "epoch": 0.01921259842519685, "grad_norm": 68.39881896972656, "learning_rate": 9.999910236220473e-06, "loss": 0.7587, "step": 61 }, { "epoch": 0.01952755905511811, "grad_norm": 43.71537780761719, "learning_rate": 9.999908661417323e-06, "loss": 0.6054, "step": 62 }, { "epoch": 0.01984251968503937, "grad_norm": 38.36960220336914, "learning_rate": 9.999907086614175e-06, "loss": 0.527, "step": 63 }, { "epoch": 0.02015748031496063, "grad_norm": 53.25741958618164, "learning_rate": 9.999905511811025e-06, "loss": 0.8724, "step": 64 }, { "epoch": 0.02047244094488189, "grad_norm": 38.55160140991211, "learning_rate": 9.999903937007874e-06, "loss": 0.6693, "step": 65 }, { "epoch": 0.02078740157480315, "grad_norm": 36.05056381225586, "learning_rate": 9.999902362204725e-06, "loss": 0.6024, "step": 66 }, { "epoch": 0.02110236220472441, "grad_norm": 83.4520492553711, "learning_rate": 9.999900787401576e-06, "loss": 0.6391, "step": 67 }, { "epoch": 0.021417322834645668, "grad_norm": 23.22808837890625, "learning_rate": 9.999899212598427e-06, "loss": 0.6275, "step": 68 }, { "epoch": 0.021732283464566928, "grad_norm": 42.370445251464844, "learning_rate": 9.999897637795276e-06, "loss": 0.3825, "step": 69 }, { "epoch": 0.02204724409448819, "grad_norm": 38.64667892456055, "learning_rate": 9.999896062992127e-06, "loss": 0.6354, "step": 70 }, { "epoch": 0.02236220472440945, "grad_norm": 44.61943054199219, "learning_rate": 9.999894488188977e-06, "loss": 0.6041, "step": 71 }, { "epoch": 0.02267716535433071, "grad_norm": 36.52523422241211, "learning_rate": 9.999892913385828e-06, "loss": 0.5188, "step": 72 }, { "epoch": 0.022992125984251967, "grad_norm": 38.4240608215332, "learning_rate": 9.999891338582679e-06, "loss": 0.612, "step": 73 }, { "epoch": 0.023307086614173227, "grad_norm": 112.46929168701172, "learning_rate": 9.99988976377953e-06, "loss": 0.9316, "step": 74 }, { "epoch": 0.023622047244094488, "grad_norm": 91.35350799560547, "learning_rate": 9.999888188976379e-06, "loss": 0.4754, "step": 75 }, { "epoch": 0.02393700787401575, "grad_norm": 136.8651123046875, "learning_rate": 9.999886614173228e-06, "loss": 0.8443, "step": 76 }, { "epoch": 0.02425196850393701, "grad_norm": 64.04878997802734, "learning_rate": 9.999885039370079e-06, "loss": 0.5659, "step": 77 }, { "epoch": 0.024566929133858266, "grad_norm": 127.41741180419922, "learning_rate": 9.99988346456693e-06, "loss": 0.5924, "step": 78 }, { "epoch": 0.024881889763779527, "grad_norm": 88.72442626953125, "learning_rate": 9.999881889763781e-06, "loss": 0.6118, "step": 79 }, { "epoch": 0.025196850393700787, "grad_norm": 91.45403289794922, "learning_rate": 9.99988031496063e-06, "loss": 0.7566, "step": 80 }, { "epoch": 0.025196850393700787, "eval_loss": 0.620968222618103, "eval_runtime": 308.6891, "eval_samples_per_second": 0.379, "eval_steps_per_second": 0.379, "step": 80 }, { "epoch": 0.025511811023622048, "grad_norm": 40.077823638916016, "learning_rate": 9.999878740157481e-06, "loss": 0.3255, "step": 81 }, { "epoch": 0.025826771653543308, "grad_norm": 86.21344757080078, "learning_rate": 9.99987716535433e-06, "loss": 0.5621, "step": 82 }, { "epoch": 0.02614173228346457, "grad_norm": 102.60726165771484, "learning_rate": 9.999875590551182e-06, "loss": 0.8275, "step": 83 }, { "epoch": 0.026456692913385826, "grad_norm": 118.94241333007812, "learning_rate": 9.999874015748033e-06, "loss": 0.5316, "step": 84 }, { "epoch": 0.026771653543307086, "grad_norm": 44.944576263427734, "learning_rate": 9.999872440944882e-06, "loss": 0.6057, "step": 85 }, { "epoch": 0.027086614173228347, "grad_norm": 43.299503326416016, "learning_rate": 9.999870866141733e-06, "loss": 0.3967, "step": 86 }, { "epoch": 0.027401574803149607, "grad_norm": 67.91696166992188, "learning_rate": 9.999869291338584e-06, "loss": 0.5094, "step": 87 }, { "epoch": 0.027716535433070868, "grad_norm": 54.101783752441406, "learning_rate": 9.999867716535435e-06, "loss": 0.7315, "step": 88 }, { "epoch": 0.028031496062992125, "grad_norm": 72.56822204589844, "learning_rate": 9.999866141732284e-06, "loss": 0.7512, "step": 89 }, { "epoch": 0.028346456692913385, "grad_norm": 74.34241485595703, "learning_rate": 9.999864566929135e-06, "loss": 0.6363, "step": 90 }, { "epoch": 0.028661417322834646, "grad_norm": 63.87611770629883, "learning_rate": 9.999862992125984e-06, "loss": 0.7365, "step": 91 }, { "epoch": 0.028976377952755906, "grad_norm": 90.9892807006836, "learning_rate": 9.999861417322835e-06, "loss": 0.8111, "step": 92 }, { "epoch": 0.029291338582677167, "grad_norm": 51.74814987182617, "learning_rate": 9.999859842519686e-06, "loss": 0.4766, "step": 93 }, { "epoch": 0.029606299212598424, "grad_norm": 48.99016571044922, "learning_rate": 9.999858267716537e-06, "loss": 0.4001, "step": 94 }, { "epoch": 0.029921259842519685, "grad_norm": 35.5272216796875, "learning_rate": 9.999856692913387e-06, "loss": 0.4208, "step": 95 }, { "epoch": 0.030236220472440945, "grad_norm": 53.10519790649414, "learning_rate": 9.999855118110236e-06, "loss": 0.7011, "step": 96 }, { "epoch": 0.030551181102362206, "grad_norm": 51.19492721557617, "learning_rate": 9.999853543307087e-06, "loss": 0.5737, "step": 97 }, { "epoch": 0.030866141732283466, "grad_norm": 40.27799987792969, "learning_rate": 9.999851968503938e-06, "loss": 0.4889, "step": 98 }, { "epoch": 0.031181102362204723, "grad_norm": 53.10594940185547, "learning_rate": 9.999850393700789e-06, "loss": 0.2341, "step": 99 }, { "epoch": 0.031496062992125984, "grad_norm": 64.59747314453125, "learning_rate": 9.999848818897638e-06, "loss": 0.5309, "step": 100 }, { "epoch": 0.031496062992125984, "eval_loss": 0.6238653659820557, "eval_runtime": 308.8997, "eval_samples_per_second": 0.379, "eval_steps_per_second": 0.379, "step": 100 }, { "epoch": 0.03181102362204724, "grad_norm": 56.20338821411133, "learning_rate": 9.99984724409449e-06, "loss": 0.5554, "step": 101 }, { "epoch": 0.032125984251968505, "grad_norm": 78.87137603759766, "learning_rate": 9.999845669291339e-06, "loss": 0.6822, "step": 102 }, { "epoch": 0.03244094488188976, "grad_norm": 76.1572036743164, "learning_rate": 9.99984409448819e-06, "loss": 0.4211, "step": 103 }, { "epoch": 0.032755905511811026, "grad_norm": 53.714942932128906, "learning_rate": 9.99984251968504e-06, "loss": 0.2694, "step": 104 }, { "epoch": 0.03307086614173228, "grad_norm": 100.88641357421875, "learning_rate": 9.99984094488189e-06, "loss": 0.6587, "step": 105 }, { "epoch": 0.03338582677165354, "grad_norm": 59.46120071411133, "learning_rate": 9.999839370078741e-06, "loss": 0.3308, "step": 106 }, { "epoch": 0.033700787401574804, "grad_norm": 142.22496032714844, "learning_rate": 9.99983779527559e-06, "loss": 0.9014, "step": 107 }, { "epoch": 0.03401574803149606, "grad_norm": 116.70782470703125, "learning_rate": 9.999836220472441e-06, "loss": 0.7636, "step": 108 }, { "epoch": 0.034330708661417325, "grad_norm": 122.21369171142578, "learning_rate": 9.999834645669292e-06, "loss": 0.6502, "step": 109 }, { "epoch": 0.03464566929133858, "grad_norm": 146.69210815429688, "learning_rate": 9.999833070866143e-06, "loss": 1.088, "step": 110 }, { "epoch": 0.03496062992125984, "grad_norm": 90.0801010131836, "learning_rate": 9.999831496062992e-06, "loss": 0.5323, "step": 111 }, { "epoch": 0.0352755905511811, "grad_norm": 64.70466613769531, "learning_rate": 9.999829921259843e-06, "loss": 0.4768, "step": 112 }, { "epoch": 0.03559055118110236, "grad_norm": 43.21613311767578, "learning_rate": 9.999828346456694e-06, "loss": 0.5867, "step": 113 }, { "epoch": 0.035905511811023624, "grad_norm": 98.97393798828125, "learning_rate": 9.999826771653545e-06, "loss": 0.9022, "step": 114 }, { "epoch": 0.03622047244094488, "grad_norm": 49.0715446472168, "learning_rate": 9.999825196850395e-06, "loss": 0.4157, "step": 115 }, { "epoch": 0.03653543307086614, "grad_norm": 49.2851676940918, "learning_rate": 9.999823622047244e-06, "loss": 0.4487, "step": 116 }, { "epoch": 0.0368503937007874, "grad_norm": 37.42869567871094, "learning_rate": 9.999822047244095e-06, "loss": 0.576, "step": 117 }, { "epoch": 0.03716535433070866, "grad_norm": 43.0858154296875, "learning_rate": 9.999820472440946e-06, "loss": 0.6744, "step": 118 }, { "epoch": 0.037480314960629924, "grad_norm": 51.691558837890625, "learning_rate": 9.999818897637797e-06, "loss": 0.4399, "step": 119 }, { "epoch": 0.03779527559055118, "grad_norm": 48.11525344848633, "learning_rate": 9.999817322834646e-06, "loss": 0.7208, "step": 120 }, { "epoch": 0.03779527559055118, "eval_loss": 0.6145237684249878, "eval_runtime": 297.9679, "eval_samples_per_second": 0.393, "eval_steps_per_second": 0.393, "step": 120 }, { "epoch": 0.03811023622047244, "grad_norm": 59.46645736694336, "learning_rate": 9.999815748031497e-06, "loss": 0.6338, "step": 121 }, { "epoch": 0.0384251968503937, "grad_norm": 41.5179443359375, "learning_rate": 9.999814173228347e-06, "loss": 0.6118, "step": 122 }, { "epoch": 0.03874015748031496, "grad_norm": 30.39054298400879, "learning_rate": 9.999812598425198e-06, "loss": 0.589, "step": 123 }, { "epoch": 0.03905511811023622, "grad_norm": 47.73324966430664, "learning_rate": 9.999811023622049e-06, "loss": 0.4502, "step": 124 }, { "epoch": 0.03937007874015748, "grad_norm": 63.116180419921875, "learning_rate": 9.999809448818898e-06, "loss": 0.5543, "step": 125 }, { "epoch": 0.03968503937007874, "grad_norm": 48.51982879638672, "learning_rate": 9.999807874015749e-06, "loss": 0.3345, "step": 126 }, { "epoch": 0.04, "grad_norm": 70.61180114746094, "learning_rate": 9.999806299212598e-06, "loss": 0.408, "step": 127 }, { "epoch": 0.04031496062992126, "grad_norm": 100.2572021484375, "learning_rate": 9.999804724409449e-06, "loss": 0.5952, "step": 128 }, { "epoch": 0.04062992125984252, "grad_norm": 120.8499984741211, "learning_rate": 9.9998031496063e-06, "loss": 0.5751, "step": 129 }, { "epoch": 0.04094488188976378, "grad_norm": 116.19609069824219, "learning_rate": 9.999801574803151e-06, "loss": 0.9086, "step": 130 }, { "epoch": 0.041259842519685036, "grad_norm": 203.10231018066406, "learning_rate": 9.9998e-06, "loss": 0.907, "step": 131 }, { "epoch": 0.0415748031496063, "grad_norm": 90.6951904296875, "learning_rate": 9.999798425196851e-06, "loss": 0.4064, "step": 132 }, { "epoch": 0.04188976377952756, "grad_norm": 67.41472625732422, "learning_rate": 9.999796850393702e-06, "loss": 0.4828, "step": 133 }, { "epoch": 0.04220472440944882, "grad_norm": 137.20547485351562, "learning_rate": 9.999795275590553e-06, "loss": 0.7428, "step": 134 }, { "epoch": 0.04251968503937008, "grad_norm": 290.8530578613281, "learning_rate": 9.999793700787403e-06, "loss": 0.8161, "step": 135 }, { "epoch": 0.042834645669291335, "grad_norm": 149.33602905273438, "learning_rate": 9.999792125984252e-06, "loss": 0.5009, "step": 136 }, { "epoch": 0.0431496062992126, "grad_norm": 114.50115203857422, "learning_rate": 9.999790551181103e-06, "loss": 1.0042, "step": 137 }, { "epoch": 0.043464566929133856, "grad_norm": 101.2292251586914, "learning_rate": 9.999788976377954e-06, "loss": 0.4708, "step": 138 }, { "epoch": 0.04377952755905512, "grad_norm": 91.65269470214844, "learning_rate": 9.999787401574805e-06, "loss": 0.6723, "step": 139 }, { "epoch": 0.04409448818897638, "grad_norm": 54.82379913330078, "learning_rate": 9.999785826771654e-06, "loss": 0.4811, "step": 140 }, { "epoch": 0.04409448818897638, "eval_loss": 0.5761768221855164, "eval_runtime": 296.7123, "eval_samples_per_second": 0.394, "eval_steps_per_second": 0.394, "step": 140 }, { "epoch": 0.044409448818897634, "grad_norm": 30.785768508911133, "learning_rate": 9.999784251968505e-06, "loss": 0.2963, "step": 141 }, { "epoch": 0.0447244094488189, "grad_norm": 44.203250885009766, "learning_rate": 9.999782677165354e-06, "loss": 0.6864, "step": 142 }, { "epoch": 0.045039370078740155, "grad_norm": 67.31315612792969, "learning_rate": 9.999781102362205e-06, "loss": 0.6552, "step": 143 }, { "epoch": 0.04535433070866142, "grad_norm": 36.43077087402344, "learning_rate": 9.999779527559056e-06, "loss": 0.6163, "step": 144 }, { "epoch": 0.04566929133858268, "grad_norm": 81.87039184570312, "learning_rate": 9.999777952755906e-06, "loss": 1.0169, "step": 145 }, { "epoch": 0.045984251968503934, "grad_norm": 32.918399810791016, "learning_rate": 9.999776377952757e-06, "loss": 0.4143, "step": 146 }, { "epoch": 0.0462992125984252, "grad_norm": 40.705284118652344, "learning_rate": 9.999774803149606e-06, "loss": 0.5362, "step": 147 }, { "epoch": 0.046614173228346455, "grad_norm": 25.83769416809082, "learning_rate": 9.999773228346457e-06, "loss": 0.4726, "step": 148 }, { "epoch": 0.04692913385826772, "grad_norm": 51.383758544921875, "learning_rate": 9.999771653543308e-06, "loss": 0.6378, "step": 149 }, { "epoch": 0.047244094488188976, "grad_norm": 59.23312759399414, "learning_rate": 9.999770078740159e-06, "loss": 0.4571, "step": 150 }, { "epoch": 0.04755905511811023, "grad_norm": 32.09741973876953, "learning_rate": 9.999768503937008e-06, "loss": 0.3173, "step": 151 }, { "epoch": 0.0478740157480315, "grad_norm": 40.37042999267578, "learning_rate": 9.99976692913386e-06, "loss": 0.6127, "step": 152 }, { "epoch": 0.048188976377952754, "grad_norm": 50.806793212890625, "learning_rate": 9.999765354330709e-06, "loss": 0.6617, "step": 153 }, { "epoch": 0.04850393700787402, "grad_norm": 42.13128662109375, "learning_rate": 9.99976377952756e-06, "loss": 0.4311, "step": 154 }, { "epoch": 0.048818897637795275, "grad_norm": 51.54093933105469, "learning_rate": 9.99976220472441e-06, "loss": 0.7501, "step": 155 }, { "epoch": 0.04913385826771653, "grad_norm": 71.40542602539062, "learning_rate": 9.99976062992126e-06, "loss": 1.0623, "step": 156 }, { "epoch": 0.049448818897637796, "grad_norm": 23.955883026123047, "learning_rate": 9.999759055118111e-06, "loss": 0.118, "step": 157 }, { "epoch": 0.04976377952755905, "grad_norm": 46.93206024169922, "learning_rate": 9.999757480314962e-06, "loss": 0.6759, "step": 158 }, { "epoch": 0.05007874015748032, "grad_norm": 40.86898422241211, "learning_rate": 9.999755905511813e-06, "loss": 0.5849, "step": 159 }, { "epoch": 0.050393700787401574, "grad_norm": 54.818450927734375, "learning_rate": 9.999754330708662e-06, "loss": 0.4963, "step": 160 }, { "epoch": 0.050393700787401574, "eval_loss": 0.646256148815155, "eval_runtime": 299.1934, "eval_samples_per_second": 0.391, "eval_steps_per_second": 0.391, "step": 160 }, { "epoch": 0.05070866141732284, "grad_norm": 39.67280578613281, "learning_rate": 9.999752755905513e-06, "loss": 0.6, "step": 161 }, { "epoch": 0.051023622047244095, "grad_norm": 44.99142074584961, "learning_rate": 9.999751181102362e-06, "loss": 0.4204, "step": 162 }, { "epoch": 0.05133858267716535, "grad_norm": 47.1932373046875, "learning_rate": 9.999749606299213e-06, "loss": 0.5264, "step": 163 }, { "epoch": 0.051653543307086616, "grad_norm": 59.98406219482422, "learning_rate": 9.999748031496064e-06, "loss": 0.5378, "step": 164 }, { "epoch": 0.05196850393700787, "grad_norm": 54.76002883911133, "learning_rate": 9.999746456692914e-06, "loss": 0.4653, "step": 165 }, { "epoch": 0.05228346456692914, "grad_norm": 65.97516632080078, "learning_rate": 9.999744881889765e-06, "loss": 0.493, "step": 166 }, { "epoch": 0.052598425196850394, "grad_norm": 74.50453186035156, "learning_rate": 9.999743307086614e-06, "loss": 0.5024, "step": 167 }, { "epoch": 0.05291338582677165, "grad_norm": 79.50423431396484, "learning_rate": 9.999741732283465e-06, "loss": 0.683, "step": 168 }, { "epoch": 0.053228346456692915, "grad_norm": 30.747211456298828, "learning_rate": 9.999740157480316e-06, "loss": 0.1244, "step": 169 }, { "epoch": 0.05354330708661417, "grad_norm": 127.78273010253906, "learning_rate": 9.999738582677167e-06, "loss": 1.0797, "step": 170 }, { "epoch": 0.053858267716535436, "grad_norm": 33.981021881103516, "learning_rate": 9.999737007874016e-06, "loss": 0.1601, "step": 171 }, { "epoch": 0.054173228346456694, "grad_norm": 89.67857360839844, "learning_rate": 9.999735433070867e-06, "loss": 0.5669, "step": 172 }, { "epoch": 0.05448818897637795, "grad_norm": 44.71755599975586, "learning_rate": 9.999733858267717e-06, "loss": 0.2556, "step": 173 }, { "epoch": 0.054803149606299215, "grad_norm": 63.42751693725586, "learning_rate": 9.999732283464568e-06, "loss": 0.353, "step": 174 }, { "epoch": 0.05511811023622047, "grad_norm": 145.7510986328125, "learning_rate": 9.999730708661419e-06, "loss": 0.6861, "step": 175 }, { "epoch": 0.055433070866141736, "grad_norm": 86.73828125, "learning_rate": 9.999729133858268e-06, "loss": 0.7471, "step": 176 }, { "epoch": 0.05574803149606299, "grad_norm": 78.96038055419922, "learning_rate": 9.999727559055119e-06, "loss": 0.5997, "step": 177 }, { "epoch": 0.05606299212598425, "grad_norm": 77.76657104492188, "learning_rate": 9.999725984251968e-06, "loss": 0.4139, "step": 178 }, { "epoch": 0.056377952755905514, "grad_norm": 56.66273880004883, "learning_rate": 9.999724409448819e-06, "loss": 0.3804, "step": 179 }, { "epoch": 0.05669291338582677, "grad_norm": 45.65275192260742, "learning_rate": 9.99972283464567e-06, "loss": 0.2486, "step": 180 }, { "epoch": 0.05669291338582677, "eval_loss": 0.6303219795227051, "eval_runtime": 297.0224, "eval_samples_per_second": 0.394, "eval_steps_per_second": 0.394, "step": 180 }, { "epoch": 0.057007874015748035, "grad_norm": 110.07376098632812, "learning_rate": 9.999721259842521e-06, "loss": 0.9565, "step": 181 }, { "epoch": 0.05732283464566929, "grad_norm": 83.31719970703125, "learning_rate": 9.99971968503937e-06, "loss": 0.7403, "step": 182 }, { "epoch": 0.05763779527559055, "grad_norm": 74.98153686523438, "learning_rate": 9.999718110236221e-06, "loss": 0.9278, "step": 183 }, { "epoch": 0.05795275590551181, "grad_norm": 86.86173248291016, "learning_rate": 9.999716535433072e-06, "loss": 0.5827, "step": 184 }, { "epoch": 0.05826771653543307, "grad_norm": 33.1976318359375, "learning_rate": 9.999714960629922e-06, "loss": 0.3653, "step": 185 }, { "epoch": 0.058582677165354334, "grad_norm": 80.26583862304688, "learning_rate": 9.999713385826773e-06, "loss": 0.4716, "step": 186 }, { "epoch": 0.05889763779527559, "grad_norm": 58.74989700317383, "learning_rate": 9.999711811023622e-06, "loss": 0.3977, "step": 187 }, { "epoch": 0.05921259842519685, "grad_norm": 75.13705444335938, "learning_rate": 9.999710236220473e-06, "loss": 0.595, "step": 188 }, { "epoch": 0.05952755905511811, "grad_norm": 44.060882568359375, "learning_rate": 9.999708661417324e-06, "loss": 0.3714, "step": 189 }, { "epoch": 0.05984251968503937, "grad_norm": 35.92017364501953, "learning_rate": 9.999707086614175e-06, "loss": 0.4465, "step": 190 }, { "epoch": 0.06015748031496063, "grad_norm": 35.966800689697266, "learning_rate": 9.999705511811024e-06, "loss": 0.4875, "step": 191 }, { "epoch": 0.06047244094488189, "grad_norm": 48.5458869934082, "learning_rate": 9.999703937007875e-06, "loss": 0.6907, "step": 192 }, { "epoch": 0.06078740157480315, "grad_norm": 38.40484619140625, "learning_rate": 9.999702362204725e-06, "loss": 0.3482, "step": 193 }, { "epoch": 0.06110236220472441, "grad_norm": 87.23228454589844, "learning_rate": 9.999700787401576e-06, "loss": 0.4824, "step": 194 }, { "epoch": 0.06141732283464567, "grad_norm": 62.46897888183594, "learning_rate": 9.999699212598427e-06, "loss": 0.3294, "step": 195 }, { "epoch": 0.06173228346456693, "grad_norm": 33.665218353271484, "learning_rate": 9.999697637795276e-06, "loss": 0.222, "step": 196 }, { "epoch": 0.06204724409448819, "grad_norm": 91.13434600830078, "learning_rate": 9.999696062992127e-06, "loss": 0.9194, "step": 197 }, { "epoch": 0.06236220472440945, "grad_norm": 31.107872009277344, "learning_rate": 9.999694488188976e-06, "loss": 0.2354, "step": 198 }, { "epoch": 0.06267716535433071, "grad_norm": 99.7812728881836, "learning_rate": 9.999692913385827e-06, "loss": 0.6351, "step": 199 }, { "epoch": 0.06299212598425197, "grad_norm": 41.42717361450195, "learning_rate": 9.999691338582678e-06, "loss": 0.1414, "step": 200 }, { "epoch": 0.06299212598425197, "eval_loss": 0.6573231220245361, "eval_runtime": 302.1543, "eval_samples_per_second": 0.387, "eval_steps_per_second": 0.387, "step": 200 }, { "epoch": 0.06330708661417322, "grad_norm": 61.020408630371094, "learning_rate": 9.999689763779529e-06, "loss": 0.6242, "step": 201 }, { "epoch": 0.06362204724409448, "grad_norm": 70.84980010986328, "learning_rate": 9.999688188976378e-06, "loss": 0.1907, "step": 202 }, { "epoch": 0.06393700787401575, "grad_norm": 174.39080810546875, "learning_rate": 9.99968661417323e-06, "loss": 0.9131, "step": 203 }, { "epoch": 0.06425196850393701, "grad_norm": 47.28941345214844, "learning_rate": 9.99968503937008e-06, "loss": 0.2425, "step": 204 }, { "epoch": 0.06456692913385827, "grad_norm": 51.628211975097656, "learning_rate": 9.99968346456693e-06, "loss": 0.2873, "step": 205 }, { "epoch": 0.06488188976377952, "grad_norm": 63.0713996887207, "learning_rate": 9.99968188976378e-06, "loss": 0.6194, "step": 206 }, { "epoch": 0.06519685039370078, "grad_norm": 164.83543395996094, "learning_rate": 9.99968031496063e-06, "loss": 0.5043, "step": 207 }, { "epoch": 0.06551181102362205, "grad_norm": 83.96135711669922, "learning_rate": 9.999678740157481e-06, "loss": 0.6309, "step": 208 }, { "epoch": 0.06582677165354331, "grad_norm": 81.93275451660156, "learning_rate": 9.999677165354332e-06, "loss": 0.4175, "step": 209 }, { "epoch": 0.06614173228346457, "grad_norm": 129.6193389892578, "learning_rate": 9.999675590551183e-06, "loss": 0.5378, "step": 210 }, { "epoch": 0.06645669291338582, "grad_norm": 137.78428649902344, "learning_rate": 9.999674015748032e-06, "loss": 1.2212, "step": 211 }, { "epoch": 0.06677165354330708, "grad_norm": 42.24091339111328, "learning_rate": 9.999672440944883e-06, "loss": 0.173, "step": 212 }, { "epoch": 0.06708661417322835, "grad_norm": 68.79737091064453, "learning_rate": 9.999670866141732e-06, "loss": 0.1975, "step": 213 }, { "epoch": 0.06740157480314961, "grad_norm": 125.35755920410156, "learning_rate": 9.999669291338583e-06, "loss": 0.7453, "step": 214 }, { "epoch": 0.06771653543307087, "grad_norm": 56.881229400634766, "learning_rate": 9.999667716535434e-06, "loss": 0.3073, "step": 215 }, { "epoch": 0.06803149606299212, "grad_norm": 121.92823028564453, "learning_rate": 9.999666141732284e-06, "loss": 1.0795, "step": 216 }, { "epoch": 0.06834645669291338, "grad_norm": 44.86691665649414, "learning_rate": 9.999664566929135e-06, "loss": 0.4312, "step": 217 }, { "epoch": 0.06866141732283465, "grad_norm": 142.288330078125, "learning_rate": 9.999662992125984e-06, "loss": 0.6631, "step": 218 }, { "epoch": 0.06897637795275591, "grad_norm": 200.1629180908203, "learning_rate": 9.999661417322835e-06, "loss": 0.5105, "step": 219 }, { "epoch": 0.06929133858267716, "grad_norm": 83.08853149414062, "learning_rate": 9.999659842519686e-06, "loss": 0.6537, "step": 220 }, { "epoch": 0.06929133858267716, "eval_loss": 0.7070333957672119, "eval_runtime": 307.2171, "eval_samples_per_second": 0.381, "eval_steps_per_second": 0.381, "step": 220 }, { "epoch": 0.06960629921259842, "grad_norm": 131.77316284179688, "learning_rate": 9.999658267716537e-06, "loss": 0.7263, "step": 221 }, { "epoch": 0.06992125984251968, "grad_norm": 103.4114761352539, "learning_rate": 9.999656692913386e-06, "loss": 1.0884, "step": 222 }, { "epoch": 0.07023622047244095, "grad_norm": 68.44525909423828, "learning_rate": 9.999655118110237e-06, "loss": 0.6249, "step": 223 }, { "epoch": 0.0705511811023622, "grad_norm": 61.6135139465332, "learning_rate": 9.999653543307087e-06, "loss": 0.5604, "step": 224 }, { "epoch": 0.07086614173228346, "grad_norm": 86.59762573242188, "learning_rate": 9.999651968503938e-06, "loss": 0.5263, "step": 225 }, { "epoch": 0.07118110236220472, "grad_norm": 42.36429214477539, "learning_rate": 9.999650393700789e-06, "loss": 0.4644, "step": 226 }, { "epoch": 0.07149606299212598, "grad_norm": 38.47148132324219, "learning_rate": 9.999648818897638e-06, "loss": 0.4286, "step": 227 }, { "epoch": 0.07181102362204725, "grad_norm": 23.091997146606445, "learning_rate": 9.999647244094489e-06, "loss": 0.3635, "step": 228 }, { "epoch": 0.0721259842519685, "grad_norm": 48.05474090576172, "learning_rate": 9.99964566929134e-06, "loss": 0.3296, "step": 229 }, { "epoch": 0.07244094488188976, "grad_norm": 56.6866569519043, "learning_rate": 9.99964409448819e-06, "loss": 0.5982, "step": 230 }, { "epoch": 0.07275590551181102, "grad_norm": 34.4522705078125, "learning_rate": 9.99964251968504e-06, "loss": 0.5496, "step": 231 }, { "epoch": 0.07307086614173228, "grad_norm": 36.2459831237793, "learning_rate": 9.999640944881891e-06, "loss": 0.4069, "step": 232 }, { "epoch": 0.07338582677165355, "grad_norm": 57.894195556640625, "learning_rate": 9.99963937007874e-06, "loss": 0.4789, "step": 233 }, { "epoch": 0.0737007874015748, "grad_norm": 100.86152648925781, "learning_rate": 9.999637795275591e-06, "loss": 0.3101, "step": 234 }, { "epoch": 0.07401574803149606, "grad_norm": 49.66980743408203, "learning_rate": 9.999636220472442e-06, "loss": 0.516, "step": 235 }, { "epoch": 0.07433070866141732, "grad_norm": 52.82820510864258, "learning_rate": 9.999634645669292e-06, "loss": 0.3732, "step": 236 }, { "epoch": 0.07464566929133858, "grad_norm": 56.593467712402344, "learning_rate": 9.999633070866143e-06, "loss": 0.3873, "step": 237 }, { "epoch": 0.07496062992125985, "grad_norm": 20.434045791625977, "learning_rate": 9.999631496062992e-06, "loss": 0.2495, "step": 238 }, { "epoch": 0.0752755905511811, "grad_norm": 65.34156799316406, "learning_rate": 9.999629921259843e-06, "loss": 0.6973, "step": 239 }, { "epoch": 0.07559055118110236, "grad_norm": 32.1629638671875, "learning_rate": 9.999628346456694e-06, "loss": 0.3504, "step": 240 }, { "epoch": 0.07559055118110236, "eval_loss": 0.7387034296989441, "eval_runtime": 307.2747, "eval_samples_per_second": 0.381, "eval_steps_per_second": 0.381, "step": 240 }, { "epoch": 0.07590551181102362, "grad_norm": 78.75101470947266, "learning_rate": 9.999626771653545e-06, "loss": 0.5959, "step": 241 }, { "epoch": 0.07622047244094488, "grad_norm": 52.39651870727539, "learning_rate": 9.999625196850394e-06, "loss": 0.3459, "step": 242 }, { "epoch": 0.07653543307086615, "grad_norm": 100.4014663696289, "learning_rate": 9.999623622047245e-06, "loss": 1.4185, "step": 243 }, { "epoch": 0.0768503937007874, "grad_norm": 63.76593780517578, "learning_rate": 9.999622047244095e-06, "loss": 0.2557, "step": 244 }, { "epoch": 0.07716535433070866, "grad_norm": 141.5568084716797, "learning_rate": 9.999620472440946e-06, "loss": 0.4333, "step": 245 }, { "epoch": 0.07748031496062992, "grad_norm": 104.44050598144531, "learning_rate": 9.999618897637797e-06, "loss": 0.5989, "step": 246 }, { "epoch": 0.07779527559055117, "grad_norm": 43.31072998046875, "learning_rate": 9.999617322834646e-06, "loss": 0.579, "step": 247 }, { "epoch": 0.07811023622047245, "grad_norm": 47.9522819519043, "learning_rate": 9.999615748031497e-06, "loss": 0.2994, "step": 248 }, { "epoch": 0.0784251968503937, "grad_norm": 40.461368560791016, "learning_rate": 9.999614173228346e-06, "loss": 0.4243, "step": 249 }, { "epoch": 0.07874015748031496, "grad_norm": 26.12702751159668, "learning_rate": 9.999612598425197e-06, "loss": 0.3175, "step": 250 }, { "epoch": 0.07905511811023622, "grad_norm": 76.74534606933594, "learning_rate": 9.999611023622048e-06, "loss": 0.6386, "step": 251 }, { "epoch": 0.07937007874015747, "grad_norm": 61.15847396850586, "learning_rate": 9.999609448818899e-06, "loss": 0.2725, "step": 252 }, { "epoch": 0.07968503937007874, "grad_norm": 119.91480255126953, "learning_rate": 9.999607874015748e-06, "loss": 0.2381, "step": 253 }, { "epoch": 0.08, "grad_norm": 138.52313232421875, "learning_rate": 9.9996062992126e-06, "loss": 1.0807, "step": 254 }, { "epoch": 0.08031496062992126, "grad_norm": 99.42451477050781, "learning_rate": 9.99960472440945e-06, "loss": 0.4007, "step": 255 }, { "epoch": 0.08062992125984252, "grad_norm": 51.6858024597168, "learning_rate": 9.9996031496063e-06, "loss": 0.2963, "step": 256 }, { "epoch": 0.08094488188976377, "grad_norm": 52.566734313964844, "learning_rate": 9.99960157480315e-06, "loss": 0.2889, "step": 257 }, { "epoch": 0.08125984251968504, "grad_norm": 94.96017456054688, "learning_rate": 9.9996e-06, "loss": 1.354, "step": 258 }, { "epoch": 0.0815748031496063, "grad_norm": 54.514915466308594, "learning_rate": 9.999598425196851e-06, "loss": 0.1296, "step": 259 }, { "epoch": 0.08188976377952756, "grad_norm": 165.0517578125, "learning_rate": 9.999596850393702e-06, "loss": 0.7929, "step": 260 }, { "epoch": 0.08188976377952756, "eval_loss": 0.767807126045227, "eval_runtime": 307.1459, "eval_samples_per_second": 0.381, "eval_steps_per_second": 0.381, "step": 260 }, { "epoch": 0.08220472440944881, "grad_norm": 62.38746643066406, "learning_rate": 9.999595275590553e-06, "loss": 0.2907, "step": 261 }, { "epoch": 0.08251968503937007, "grad_norm": 107.24059295654297, "learning_rate": 9.999593700787402e-06, "loss": 0.9582, "step": 262 }, { "epoch": 0.08283464566929134, "grad_norm": 182.7991943359375, "learning_rate": 9.999592125984253e-06, "loss": 1.1002, "step": 263 }, { "epoch": 0.0831496062992126, "grad_norm": 39.42921829223633, "learning_rate": 9.999590551181102e-06, "loss": 0.3591, "step": 264 }, { "epoch": 0.08346456692913386, "grad_norm": 78.57293701171875, "learning_rate": 9.999588976377953e-06, "loss": 0.5395, "step": 265 }, { "epoch": 0.08377952755905511, "grad_norm": 46.55572509765625, "learning_rate": 9.999587401574804e-06, "loss": 0.4244, "step": 266 }, { "epoch": 0.08409448818897637, "grad_norm": 72.65052795410156, "learning_rate": 9.999585826771654e-06, "loss": 0.5564, "step": 267 }, { "epoch": 0.08440944881889764, "grad_norm": 40.577850341796875, "learning_rate": 9.999584251968505e-06, "loss": 0.319, "step": 268 }, { "epoch": 0.0847244094488189, "grad_norm": 60.15317916870117, "learning_rate": 9.999582677165354e-06, "loss": 0.2762, "step": 269 }, { "epoch": 0.08503937007874016, "grad_norm": 64.16828918457031, "learning_rate": 9.999581102362205e-06, "loss": 0.5016, "step": 270 }, { "epoch": 0.08535433070866141, "grad_norm": 62.69063949584961, "learning_rate": 9.999579527559056e-06, "loss": 0.4175, "step": 271 }, { "epoch": 0.08566929133858267, "grad_norm": 49.130157470703125, "learning_rate": 9.999577952755907e-06, "loss": 0.3619, "step": 272 }, { "epoch": 0.08598425196850394, "grad_norm": 71.22623443603516, "learning_rate": 9.999576377952756e-06, "loss": 0.3924, "step": 273 }, { "epoch": 0.0862992125984252, "grad_norm": 77.7140884399414, "learning_rate": 9.999574803149607e-06, "loss": 0.4717, "step": 274 }, { "epoch": 0.08661417322834646, "grad_norm": 78.33636474609375, "learning_rate": 9.999573228346458e-06, "loss": 0.5174, "step": 275 }, { "epoch": 0.08692913385826771, "grad_norm": 48.37542724609375, "learning_rate": 9.999571653543308e-06, "loss": 0.2733, "step": 276 }, { "epoch": 0.08724409448818897, "grad_norm": 57.93960189819336, "learning_rate": 9.999570078740159e-06, "loss": 0.6626, "step": 277 }, { "epoch": 0.08755905511811024, "grad_norm": 58.80123519897461, "learning_rate": 9.999568503937008e-06, "loss": 0.4598, "step": 278 }, { "epoch": 0.0878740157480315, "grad_norm": 49.037818908691406, "learning_rate": 9.999566929133859e-06, "loss": 0.4452, "step": 279 }, { "epoch": 0.08818897637795275, "grad_norm": 62.81136703491211, "learning_rate": 9.99956535433071e-06, "loss": 0.348, "step": 280 }, { "epoch": 0.08818897637795275, "eval_loss": 0.618428111076355, "eval_runtime": 306.6914, "eval_samples_per_second": 0.381, "eval_steps_per_second": 0.381, "step": 280 }, { "epoch": 0.08850393700787401, "grad_norm": 61.92344284057617, "learning_rate": 9.999563779527561e-06, "loss": 0.2919, "step": 281 }, { "epoch": 0.08881889763779527, "grad_norm": 75.2997817993164, "learning_rate": 9.99956220472441e-06, "loss": 0.4883, "step": 282 }, { "epoch": 0.08913385826771654, "grad_norm": 34.474639892578125, "learning_rate": 9.999560629921261e-06, "loss": 0.1744, "step": 283 }, { "epoch": 0.0894488188976378, "grad_norm": 79.71351623535156, "learning_rate": 9.99955905511811e-06, "loss": 0.4965, "step": 284 }, { "epoch": 0.08976377952755905, "grad_norm": 127.44145202636719, "learning_rate": 9.999557480314961e-06, "loss": 0.7171, "step": 285 }, { "epoch": 0.09007874015748031, "grad_norm": 57.06454086303711, "learning_rate": 9.999555905511812e-06, "loss": 0.5155, "step": 286 }, { "epoch": 0.09039370078740157, "grad_norm": 86.23944854736328, "learning_rate": 9.999554330708662e-06, "loss": 0.613, "step": 287 }, { "epoch": 0.09070866141732284, "grad_norm": 188.0777587890625, "learning_rate": 9.999552755905513e-06, "loss": 0.3885, "step": 288 }, { "epoch": 0.0910236220472441, "grad_norm": 49.08794403076172, "learning_rate": 9.999551181102362e-06, "loss": 0.3903, "step": 289 }, { "epoch": 0.09133858267716535, "grad_norm": 77.4734115600586, "learning_rate": 9.999549606299213e-06, "loss": 0.629, "step": 290 }, { "epoch": 0.09165354330708661, "grad_norm": 31.26721954345703, "learning_rate": 9.999548031496064e-06, "loss": 0.2508, "step": 291 }, { "epoch": 0.09196850393700787, "grad_norm": 59.19281768798828, "learning_rate": 9.999546456692915e-06, "loss": 0.1862, "step": 292 }, { "epoch": 0.09228346456692914, "grad_norm": 97.84364318847656, "learning_rate": 9.999544881889764e-06, "loss": 0.8973, "step": 293 }, { "epoch": 0.0925984251968504, "grad_norm": 94.82438659667969, "learning_rate": 9.999543307086615e-06, "loss": 0.8992, "step": 294 }, { "epoch": 0.09291338582677165, "grad_norm": 57.59076690673828, "learning_rate": 9.999541732283465e-06, "loss": 0.6437, "step": 295 }, { "epoch": 0.09322834645669291, "grad_norm": 37.8861198425293, "learning_rate": 9.999540157480316e-06, "loss": 0.4185, "step": 296 }, { "epoch": 0.09354330708661417, "grad_norm": 28.94227409362793, "learning_rate": 9.999538582677167e-06, "loss": 0.285, "step": 297 }, { "epoch": 0.09385826771653544, "grad_norm": 50.66032409667969, "learning_rate": 9.999537007874016e-06, "loss": 0.6087, "step": 298 }, { "epoch": 0.0941732283464567, "grad_norm": 24.23774528503418, "learning_rate": 9.999535433070867e-06, "loss": 0.2489, "step": 299 }, { "epoch": 0.09448818897637795, "grad_norm": 50.08018493652344, "learning_rate": 9.999533858267718e-06, "loss": 0.6251, "step": 300 }, { "epoch": 0.09448818897637795, "eval_loss": 0.5396940112113953, "eval_runtime": 294.9781, "eval_samples_per_second": 0.397, "eval_steps_per_second": 0.397, "step": 300 }, { "epoch": 0.09480314960629921, "grad_norm": 87.98992919921875, "learning_rate": 9.999532283464569e-06, "loss": 0.6764, "step": 301 }, { "epoch": 0.09511811023622047, "grad_norm": 47.72505187988281, "learning_rate": 9.999530708661418e-06, "loss": 0.754, "step": 302 }, { "epoch": 0.09543307086614174, "grad_norm": 29.56645393371582, "learning_rate": 9.999529133858269e-06, "loss": 0.2754, "step": 303 }, { "epoch": 0.095748031496063, "grad_norm": 66.290283203125, "learning_rate": 9.999527559055118e-06, "loss": 0.5663, "step": 304 }, { "epoch": 0.09606299212598425, "grad_norm": 38.929725646972656, "learning_rate": 9.99952598425197e-06, "loss": 0.3187, "step": 305 }, { "epoch": 0.09637795275590551, "grad_norm": 55.97653579711914, "learning_rate": 9.99952440944882e-06, "loss": 0.4682, "step": 306 }, { "epoch": 0.09669291338582676, "grad_norm": 41.88676071166992, "learning_rate": 9.99952283464567e-06, "loss": 0.4863, "step": 307 }, { "epoch": 0.09700787401574804, "grad_norm": 39.72370529174805, "learning_rate": 9.99952125984252e-06, "loss": 0.3445, "step": 308 }, { "epoch": 0.09732283464566929, "grad_norm": 48.20722579956055, "learning_rate": 9.99951968503937e-06, "loss": 0.4213, "step": 309 }, { "epoch": 0.09763779527559055, "grad_norm": 54.454715728759766, "learning_rate": 9.999518110236221e-06, "loss": 0.4397, "step": 310 }, { "epoch": 0.0979527559055118, "grad_norm": 64.91082000732422, "learning_rate": 9.999516535433072e-06, "loss": 0.2707, "step": 311 }, { "epoch": 0.09826771653543306, "grad_norm": 94.55459594726562, "learning_rate": 9.999514960629923e-06, "loss": 0.3693, "step": 312 }, { "epoch": 0.09858267716535433, "grad_norm": 96.01959991455078, "learning_rate": 9.999513385826772e-06, "loss": 0.5796, "step": 313 }, { "epoch": 0.09889763779527559, "grad_norm": 44.687355041503906, "learning_rate": 9.999511811023623e-06, "loss": 0.4414, "step": 314 }, { "epoch": 0.09921259842519685, "grad_norm": 108.15480041503906, "learning_rate": 9.999510236220473e-06, "loss": 0.4173, "step": 315 }, { "epoch": 0.0995275590551181, "grad_norm": 42.95850372314453, "learning_rate": 9.999508661417324e-06, "loss": 0.1891, "step": 316 }, { "epoch": 0.09984251968503936, "grad_norm": 66.48217010498047, "learning_rate": 9.999507086614174e-06, "loss": 0.2298, "step": 317 }, { "epoch": 0.10015748031496063, "grad_norm": 152.71719360351562, "learning_rate": 9.999505511811024e-06, "loss": 0.4982, "step": 318 }, { "epoch": 0.10047244094488189, "grad_norm": 68.98497772216797, "learning_rate": 9.999503937007875e-06, "loss": 0.6307, "step": 319 }, { "epoch": 0.10078740157480315, "grad_norm": 68.25215911865234, "learning_rate": 9.999502362204724e-06, "loss": 0.2943, "step": 320 }, { "epoch": 0.10078740157480315, "eval_loss": 0.6371558904647827, "eval_runtime": 297.2013, "eval_samples_per_second": 0.394, "eval_steps_per_second": 0.394, "step": 320 }, { "epoch": 0.1011023622047244, "grad_norm": 64.41675567626953, "learning_rate": 9.999500787401577e-06, "loss": 0.3071, "step": 321 }, { "epoch": 0.10141732283464568, "grad_norm": 146.19937133789062, "learning_rate": 9.999499212598426e-06, "loss": 0.1765, "step": 322 }, { "epoch": 0.10173228346456693, "grad_norm": 81.2210464477539, "learning_rate": 9.999497637795277e-06, "loss": 0.854, "step": 323 }, { "epoch": 0.10204724409448819, "grad_norm": 142.67649841308594, "learning_rate": 9.999496062992126e-06, "loss": 0.431, "step": 324 }, { "epoch": 0.10236220472440945, "grad_norm": 23.75156593322754, "learning_rate": 9.999494488188977e-06, "loss": 0.0743, "step": 325 }, { "epoch": 0.1026771653543307, "grad_norm": 154.65882873535156, "learning_rate": 9.999492913385828e-06, "loss": 0.7567, "step": 326 }, { "epoch": 0.10299212598425198, "grad_norm": 284.99822998046875, "learning_rate": 9.999491338582678e-06, "loss": 1.2282, "step": 327 }, { "epoch": 0.10330708661417323, "grad_norm": 87.0584945678711, "learning_rate": 9.999489763779529e-06, "loss": 1.0155, "step": 328 }, { "epoch": 0.10362204724409449, "grad_norm": 54.77091598510742, "learning_rate": 9.999488188976378e-06, "loss": 0.2331, "step": 329 }, { "epoch": 0.10393700787401575, "grad_norm": 93.18637084960938, "learning_rate": 9.999486614173229e-06, "loss": 0.5502, "step": 330 }, { "epoch": 0.104251968503937, "grad_norm": 70.45845031738281, "learning_rate": 9.99948503937008e-06, "loss": 0.5354, "step": 331 }, { "epoch": 0.10456692913385827, "grad_norm": 55.51031494140625, "learning_rate": 9.999483464566931e-06, "loss": 0.5554, "step": 332 }, { "epoch": 0.10488188976377953, "grad_norm": 93.07254028320312, "learning_rate": 9.99948188976378e-06, "loss": 0.521, "step": 333 }, { "epoch": 0.10519685039370079, "grad_norm": 31.241605758666992, "learning_rate": 9.999480314960631e-06, "loss": 0.2664, "step": 334 }, { "epoch": 0.10551181102362205, "grad_norm": 54.83103942871094, "learning_rate": 9.99947874015748e-06, "loss": 0.4913, "step": 335 }, { "epoch": 0.1058267716535433, "grad_norm": 55.371360778808594, "learning_rate": 9.999477165354331e-06, "loss": 0.5836, "step": 336 }, { "epoch": 0.10614173228346457, "grad_norm": 77.42748260498047, "learning_rate": 9.999475590551182e-06, "loss": 0.4551, "step": 337 }, { "epoch": 0.10645669291338583, "grad_norm": 52.34659194946289, "learning_rate": 9.999474015748032e-06, "loss": 0.3793, "step": 338 }, { "epoch": 0.10677165354330709, "grad_norm": 69.15506744384766, "learning_rate": 9.999472440944883e-06, "loss": 0.4266, "step": 339 }, { "epoch": 0.10708661417322834, "grad_norm": 59.99565887451172, "learning_rate": 9.999470866141732e-06, "loss": 0.3804, "step": 340 }, { "epoch": 0.10708661417322834, "eval_loss": 0.5206155776977539, "eval_runtime": 293.7622, "eval_samples_per_second": 0.398, "eval_steps_per_second": 0.398, "step": 340 }, { "epoch": 0.1074015748031496, "grad_norm": 62.315391540527344, "learning_rate": 9.999469291338583e-06, "loss": 0.5939, "step": 341 }, { "epoch": 0.10771653543307087, "grad_norm": 45.72246551513672, "learning_rate": 9.999467716535434e-06, "loss": 0.2962, "step": 342 }, { "epoch": 0.10803149606299213, "grad_norm": 72.52598571777344, "learning_rate": 9.999466141732285e-06, "loss": 0.773, "step": 343 }, { "epoch": 0.10834645669291339, "grad_norm": 50.78411102294922, "learning_rate": 9.999464566929134e-06, "loss": 0.4392, "step": 344 }, { "epoch": 0.10866141732283464, "grad_norm": 70.54865264892578, "learning_rate": 9.999462992125985e-06, "loss": 0.5495, "step": 345 }, { "epoch": 0.1089763779527559, "grad_norm": 42.886756896972656, "learning_rate": 9.999461417322836e-06, "loss": 0.3558, "step": 346 }, { "epoch": 0.10929133858267717, "grad_norm": 61.35227584838867, "learning_rate": 9.999459842519686e-06, "loss": 0.583, "step": 347 }, { "epoch": 0.10960629921259843, "grad_norm": 75.92544555664062, "learning_rate": 9.999458267716537e-06, "loss": 0.6637, "step": 348 }, { "epoch": 0.10992125984251969, "grad_norm": 47.078548431396484, "learning_rate": 9.999456692913386e-06, "loss": 0.5437, "step": 349 }, { "epoch": 0.11023622047244094, "grad_norm": 46.12405014038086, "learning_rate": 9.999455118110237e-06, "loss": 0.463, "step": 350 }, { "epoch": 0.1105511811023622, "grad_norm": 49.5578727722168, "learning_rate": 9.999453543307088e-06, "loss": 0.2595, "step": 351 }, { "epoch": 0.11086614173228347, "grad_norm": 42.03670883178711, "learning_rate": 9.999451968503939e-06, "loss": 0.3669, "step": 352 }, { "epoch": 0.11118110236220473, "grad_norm": 55.1522102355957, "learning_rate": 9.999450393700788e-06, "loss": 0.5018, "step": 353 }, { "epoch": 0.11149606299212599, "grad_norm": 125.36481475830078, "learning_rate": 9.999448818897639e-06, "loss": 0.4888, "step": 354 }, { "epoch": 0.11181102362204724, "grad_norm": 81.62045288085938, "learning_rate": 9.999447244094488e-06, "loss": 0.4195, "step": 355 }, { "epoch": 0.1121259842519685, "grad_norm": 85.13298797607422, "learning_rate": 9.99944566929134e-06, "loss": 0.4851, "step": 356 }, { "epoch": 0.11244094488188977, "grad_norm": 245.27197265625, "learning_rate": 9.99944409448819e-06, "loss": 0.5387, "step": 357 }, { "epoch": 0.11275590551181103, "grad_norm": 71.68444061279297, "learning_rate": 9.99944251968504e-06, "loss": 0.2865, "step": 358 }, { "epoch": 0.11307086614173228, "grad_norm": 44.38494110107422, "learning_rate": 9.99944094488189e-06, "loss": 0.4339, "step": 359 }, { "epoch": 0.11338582677165354, "grad_norm": 62.24411392211914, "learning_rate": 9.99943937007874e-06, "loss": 0.4798, "step": 360 }, { "epoch": 0.11338582677165354, "eval_loss": 0.5544171929359436, "eval_runtime": 293.2522, "eval_samples_per_second": 0.399, "eval_steps_per_second": 0.399, "step": 360 }, { "epoch": 0.1137007874015748, "grad_norm": 97.97586059570312, "learning_rate": 9.999437795275591e-06, "loss": 0.8155, "step": 361 }, { "epoch": 0.11401574803149607, "grad_norm": 101.96649169921875, "learning_rate": 9.999436220472442e-06, "loss": 0.3751, "step": 362 }, { "epoch": 0.11433070866141733, "grad_norm": 109.22547912597656, "learning_rate": 9.999434645669293e-06, "loss": 0.9483, "step": 363 }, { "epoch": 0.11464566929133858, "grad_norm": 101.05289459228516, "learning_rate": 9.999433070866142e-06, "loss": 0.4469, "step": 364 }, { "epoch": 0.11496062992125984, "grad_norm": 97.29914855957031, "learning_rate": 9.999431496062993e-06, "loss": 0.4986, "step": 365 }, { "epoch": 0.1152755905511811, "grad_norm": 52.88810729980469, "learning_rate": 9.999429921259843e-06, "loss": 0.239, "step": 366 }, { "epoch": 0.11559055118110237, "grad_norm": 220.28375244140625, "learning_rate": 9.999428346456694e-06, "loss": 0.9343, "step": 367 }, { "epoch": 0.11590551181102363, "grad_norm": 128.3485565185547, "learning_rate": 9.999426771653545e-06, "loss": 0.6493, "step": 368 }, { "epoch": 0.11622047244094488, "grad_norm": 68.30548095703125, "learning_rate": 9.999425196850394e-06, "loss": 0.6118, "step": 369 }, { "epoch": 0.11653543307086614, "grad_norm": 85.01860809326172, "learning_rate": 9.999423622047245e-06, "loss": 0.9292, "step": 370 }, { "epoch": 0.1168503937007874, "grad_norm": 42.967952728271484, "learning_rate": 9.999422047244096e-06, "loss": 0.3607, "step": 371 }, { "epoch": 0.11716535433070867, "grad_norm": 31.546159744262695, "learning_rate": 9.999420472440947e-06, "loss": 0.1924, "step": 372 }, { "epoch": 0.11748031496062993, "grad_norm": 49.90913391113281, "learning_rate": 9.999418897637796e-06, "loss": 0.3681, "step": 373 }, { "epoch": 0.11779527559055118, "grad_norm": 43.57588195800781, "learning_rate": 9.999417322834647e-06, "loss": 0.2731, "step": 374 }, { "epoch": 0.11811023622047244, "grad_norm": 53.5254020690918, "learning_rate": 9.999415748031496e-06, "loss": 0.5743, "step": 375 }, { "epoch": 0.1184251968503937, "grad_norm": 57.816184997558594, "learning_rate": 9.999414173228347e-06, "loss": 0.6925, "step": 376 }, { "epoch": 0.11874015748031497, "grad_norm": 50.099021911621094, "learning_rate": 9.999412598425198e-06, "loss": 0.4436, "step": 377 }, { "epoch": 0.11905511811023622, "grad_norm": 38.81980514526367, "learning_rate": 9.999411023622048e-06, "loss": 0.3597, "step": 378 }, { "epoch": 0.11937007874015748, "grad_norm": 60.50627517700195, "learning_rate": 9.999409448818899e-06, "loss": 0.3588, "step": 379 }, { "epoch": 0.11968503937007874, "grad_norm": 88.20054626464844, "learning_rate": 9.999407874015748e-06, "loss": 0.3192, "step": 380 }, { "epoch": 0.11968503937007874, "eval_loss": 0.5467623472213745, "eval_runtime": 292.898, "eval_samples_per_second": 0.399, "eval_steps_per_second": 0.399, "step": 380 }, { "epoch": 0.12, "grad_norm": 39.78269577026367, "learning_rate": 9.999406299212599e-06, "loss": 0.4447, "step": 381 }, { "epoch": 0.12031496062992127, "grad_norm": 61.480377197265625, "learning_rate": 9.99940472440945e-06, "loss": 0.3143, "step": 382 }, { "epoch": 0.12062992125984252, "grad_norm": 200.373779296875, "learning_rate": 9.999403149606301e-06, "loss": 1.3858, "step": 383 }, { "epoch": 0.12094488188976378, "grad_norm": 162.78187561035156, "learning_rate": 9.99940157480315e-06, "loss": 0.6117, "step": 384 }, { "epoch": 0.12125984251968504, "grad_norm": 42.38726043701172, "learning_rate": 9.999400000000001e-06, "loss": 0.3125, "step": 385 }, { "epoch": 0.1215748031496063, "grad_norm": 47.82020568847656, "learning_rate": 9.99939842519685e-06, "loss": 0.3962, "step": 386 }, { "epoch": 0.12188976377952757, "grad_norm": 106.18301391601562, "learning_rate": 9.999396850393701e-06, "loss": 0.6107, "step": 387 }, { "epoch": 0.12220472440944882, "grad_norm": 57.39361572265625, "learning_rate": 9.999395275590552e-06, "loss": 0.2136, "step": 388 }, { "epoch": 0.12251968503937008, "grad_norm": 32.14236068725586, "learning_rate": 9.999393700787402e-06, "loss": 0.0976, "step": 389 }, { "epoch": 0.12283464566929134, "grad_norm": 13.914974212646484, "learning_rate": 9.999392125984253e-06, "loss": 0.0404, "step": 390 }, { "epoch": 0.1231496062992126, "grad_norm": 67.5541000366211, "learning_rate": 9.999390551181104e-06, "loss": 0.5028, "step": 391 }, { "epoch": 0.12346456692913386, "grad_norm": 108.35496520996094, "learning_rate": 9.999388976377955e-06, "loss": 0.686, "step": 392 }, { "epoch": 0.12377952755905512, "grad_norm": 69.92194366455078, "learning_rate": 9.999387401574804e-06, "loss": 0.2082, "step": 393 }, { "epoch": 0.12409448818897638, "grad_norm": 23.27518081665039, "learning_rate": 9.999385826771655e-06, "loss": 0.0539, "step": 394 }, { "epoch": 0.12440944881889764, "grad_norm": 48.74710464477539, "learning_rate": 9.999384251968504e-06, "loss": 0.3422, "step": 395 }, { "epoch": 0.1247244094488189, "grad_norm": 138.21983337402344, "learning_rate": 9.999382677165355e-06, "loss": 0.8565, "step": 396 }, { "epoch": 0.12503937007874016, "grad_norm": 112.90072631835938, "learning_rate": 9.999381102362206e-06, "loss": 0.2024, "step": 397 }, { "epoch": 0.12535433070866142, "grad_norm": 256.4333190917969, "learning_rate": 9.999379527559056e-06, "loss": 1.8303, "step": 398 }, { "epoch": 0.12566929133858268, "grad_norm": 112.24787902832031, "learning_rate": 9.999377952755907e-06, "loss": 0.3992, "step": 399 }, { "epoch": 0.12598425196850394, "grad_norm": 131.27490234375, "learning_rate": 9.999376377952756e-06, "loss": 0.5066, "step": 400 }, { "epoch": 0.12598425196850394, "eval_loss": 0.6591930985450745, "eval_runtime": 304.46, "eval_samples_per_second": 0.384, "eval_steps_per_second": 0.384, "step": 400 }, { "epoch": 0.1262992125984252, "grad_norm": 194.88157653808594, "learning_rate": 9.999374803149607e-06, "loss": 1.5517, "step": 401 }, { "epoch": 0.12661417322834645, "grad_norm": 57.20365524291992, "learning_rate": 9.999373228346458e-06, "loss": 0.4244, "step": 402 }, { "epoch": 0.1269291338582677, "grad_norm": 36.50507736206055, "learning_rate": 9.999371653543309e-06, "loss": 0.1584, "step": 403 }, { "epoch": 0.12724409448818896, "grad_norm": 133.4579620361328, "learning_rate": 9.999370078740158e-06, "loss": 1.333, "step": 404 }, { "epoch": 0.12755905511811025, "grad_norm": 128.0843505859375, "learning_rate": 9.999368503937009e-06, "loss": 0.4923, "step": 405 }, { "epoch": 0.1278740157480315, "grad_norm": 100.4643325805664, "learning_rate": 9.999366929133858e-06, "loss": 0.4229, "step": 406 }, { "epoch": 0.12818897637795276, "grad_norm": 51.65610885620117, "learning_rate": 9.99936535433071e-06, "loss": 0.3144, "step": 407 }, { "epoch": 0.12850393700787402, "grad_norm": 86.17060852050781, "learning_rate": 9.99936377952756e-06, "loss": 0.5492, "step": 408 }, { "epoch": 0.12881889763779528, "grad_norm": 11.016338348388672, "learning_rate": 9.99936220472441e-06, "loss": 0.0502, "step": 409 }, { "epoch": 0.12913385826771653, "grad_norm": 75.9980697631836, "learning_rate": 9.99936062992126e-06, "loss": 0.47, "step": 410 }, { "epoch": 0.1294488188976378, "grad_norm": 64.87591552734375, "learning_rate": 9.99935905511811e-06, "loss": 0.2592, "step": 411 }, { "epoch": 0.12976377952755905, "grad_norm": 109.087646484375, "learning_rate": 9.999357480314961e-06, "loss": 0.7454, "step": 412 }, { "epoch": 0.1300787401574803, "grad_norm": 46.490875244140625, "learning_rate": 9.999355905511812e-06, "loss": 0.2048, "step": 413 }, { "epoch": 0.13039370078740156, "grad_norm": 156.15184020996094, "learning_rate": 9.999354330708663e-06, "loss": 0.4924, "step": 414 }, { "epoch": 0.13070866141732285, "grad_norm": 120.5556640625, "learning_rate": 9.999352755905512e-06, "loss": 0.8374, "step": 415 }, { "epoch": 0.1310236220472441, "grad_norm": 151.4144744873047, "learning_rate": 9.999351181102363e-06, "loss": 0.6144, "step": 416 }, { "epoch": 0.13133858267716536, "grad_norm": 90.7903060913086, "learning_rate": 9.999349606299214e-06, "loss": 0.2709, "step": 417 }, { "epoch": 0.13165354330708662, "grad_norm": 27.181320190429688, "learning_rate": 9.999348031496064e-06, "loss": 0.3197, "step": 418 }, { "epoch": 0.13196850393700787, "grad_norm": 98.53707885742188, "learning_rate": 9.999346456692915e-06, "loss": 0.8255, "step": 419 }, { "epoch": 0.13228346456692913, "grad_norm": 34.432132720947266, "learning_rate": 9.999344881889764e-06, "loss": 0.1117, "step": 420 }, { "epoch": 0.13228346456692913, "eval_loss": 0.535347580909729, "eval_runtime": 306.3308, "eval_samples_per_second": 0.382, "eval_steps_per_second": 0.382, "step": 420 }, { "epoch": 0.1325984251968504, "grad_norm": 42.67380142211914, "learning_rate": 9.999343307086615e-06, "loss": 0.3283, "step": 421 }, { "epoch": 0.13291338582677165, "grad_norm": 71.41397857666016, "learning_rate": 9.999341732283466e-06, "loss": 0.295, "step": 422 }, { "epoch": 0.1332283464566929, "grad_norm": 28.39377784729004, "learning_rate": 9.999340157480317e-06, "loss": 0.1143, "step": 423 }, { "epoch": 0.13354330708661416, "grad_norm": 61.82292556762695, "learning_rate": 9.999338582677166e-06, "loss": 0.5727, "step": 424 }, { "epoch": 0.13385826771653545, "grad_norm": 69.15302276611328, "learning_rate": 9.999337007874017e-06, "loss": 0.3849, "step": 425 }, { "epoch": 0.1341732283464567, "grad_norm": 38.20407485961914, "learning_rate": 9.999335433070866e-06, "loss": 0.334, "step": 426 }, { "epoch": 0.13448818897637796, "grad_norm": 25.310161590576172, "learning_rate": 9.999333858267717e-06, "loss": 0.1519, "step": 427 }, { "epoch": 0.13480314960629922, "grad_norm": 28.452911376953125, "learning_rate": 9.999332283464568e-06, "loss": 0.2238, "step": 428 }, { "epoch": 0.13511811023622047, "grad_norm": 127.5716552734375, "learning_rate": 9.999330708661418e-06, "loss": 1.3774, "step": 429 }, { "epoch": 0.13543307086614173, "grad_norm": 60.44778060913086, "learning_rate": 9.999329133858269e-06, "loss": 0.7446, "step": 430 }, { "epoch": 0.135748031496063, "grad_norm": 54.13089370727539, "learning_rate": 9.999327559055118e-06, "loss": 0.7198, "step": 431 }, { "epoch": 0.13606299212598424, "grad_norm": 44.978485107421875, "learning_rate": 9.999325984251969e-06, "loss": 0.457, "step": 432 }, { "epoch": 0.1363779527559055, "grad_norm": 78.32373809814453, "learning_rate": 9.99932440944882e-06, "loss": 0.5859, "step": 433 }, { "epoch": 0.13669291338582676, "grad_norm": 132.4387969970703, "learning_rate": 9.999322834645671e-06, "loss": 0.4658, "step": 434 }, { "epoch": 0.13700787401574804, "grad_norm": 14.64592170715332, "learning_rate": 9.99932125984252e-06, "loss": 0.0912, "step": 435 }, { "epoch": 0.1373228346456693, "grad_norm": 57.1812858581543, "learning_rate": 9.999319685039371e-06, "loss": 0.4572, "step": 436 }, { "epoch": 0.13763779527559056, "grad_norm": 46.73292541503906, "learning_rate": 9.99931811023622e-06, "loss": 0.4464, "step": 437 }, { "epoch": 0.13795275590551181, "grad_norm": 21.240659713745117, "learning_rate": 9.999316535433072e-06, "loss": 0.1034, "step": 438 }, { "epoch": 0.13826771653543307, "grad_norm": 46.250614166259766, "learning_rate": 9.999314960629922e-06, "loss": 0.2897, "step": 439 }, { "epoch": 0.13858267716535433, "grad_norm": 73.32185363769531, "learning_rate": 9.999313385826772e-06, "loss": 0.5788, "step": 440 }, { "epoch": 0.13858267716535433, "eval_loss": 0.5959511399269104, "eval_runtime": 304.5113, "eval_samples_per_second": 0.384, "eval_steps_per_second": 0.384, "step": 440 }, { "epoch": 0.13889763779527559, "grad_norm": 7.073561191558838, "learning_rate": 9.999311811023623e-06, "loss": 0.0235, "step": 441 }, { "epoch": 0.13921259842519684, "grad_norm": 130.6144561767578, "learning_rate": 9.999310236220474e-06, "loss": 1.088, "step": 442 }, { "epoch": 0.1395275590551181, "grad_norm": 105.4767837524414, "learning_rate": 9.999308661417325e-06, "loss": 1.3411, "step": 443 }, { "epoch": 0.13984251968503936, "grad_norm": 45.18183898925781, "learning_rate": 9.999307086614174e-06, "loss": 0.2972, "step": 444 }, { "epoch": 0.14015748031496064, "grad_norm": 84.5200424194336, "learning_rate": 9.999305511811025e-06, "loss": 0.3241, "step": 445 }, { "epoch": 0.1404724409448819, "grad_norm": 33.606468200683594, "learning_rate": 9.999303937007874e-06, "loss": 0.3225, "step": 446 }, { "epoch": 0.14078740157480316, "grad_norm": 156.7371063232422, "learning_rate": 9.999302362204725e-06, "loss": 0.8385, "step": 447 }, { "epoch": 0.1411023622047244, "grad_norm": 51.86471176147461, "learning_rate": 9.999300787401576e-06, "loss": 0.3886, "step": 448 }, { "epoch": 0.14141732283464567, "grad_norm": 98.36966705322266, "learning_rate": 9.999299212598426e-06, "loss": 0.3667, "step": 449 }, { "epoch": 0.14173228346456693, "grad_norm": 163.55325317382812, "learning_rate": 9.999297637795277e-06, "loss": 0.9837, "step": 450 }, { "epoch": 0.14204724409448818, "grad_norm": 52.89970779418945, "learning_rate": 9.999296062992126e-06, "loss": 0.3358, "step": 451 }, { "epoch": 0.14236220472440944, "grad_norm": 71.64704132080078, "learning_rate": 9.999294488188977e-06, "loss": 0.7998, "step": 452 }, { "epoch": 0.1426771653543307, "grad_norm": 19.4017276763916, "learning_rate": 9.999292913385828e-06, "loss": 0.0742, "step": 453 }, { "epoch": 0.14299212598425196, "grad_norm": 89.31649017333984, "learning_rate": 9.999291338582679e-06, "loss": 0.557, "step": 454 }, { "epoch": 0.14330708661417324, "grad_norm": 47.06895446777344, "learning_rate": 9.999289763779528e-06, "loss": 0.5652, "step": 455 }, { "epoch": 0.1436220472440945, "grad_norm": 60.04147720336914, "learning_rate": 9.99928818897638e-06, "loss": 0.4523, "step": 456 }, { "epoch": 0.14393700787401575, "grad_norm": 33.64701843261719, "learning_rate": 9.999286614173228e-06, "loss": 0.3136, "step": 457 }, { "epoch": 0.144251968503937, "grad_norm": 24.181163787841797, "learning_rate": 9.99928503937008e-06, "loss": 0.1441, "step": 458 }, { "epoch": 0.14456692913385827, "grad_norm": 121.68936157226562, "learning_rate": 9.99928346456693e-06, "loss": 0.9653, "step": 459 }, { "epoch": 0.14488188976377953, "grad_norm": 48.23858642578125, "learning_rate": 9.99928188976378e-06, "loss": 0.4576, "step": 460 }, { "epoch": 0.14488188976377953, "eval_loss": 0.6207642555236816, "eval_runtime": 295.7863, "eval_samples_per_second": 0.396, "eval_steps_per_second": 0.396, "step": 460 }, { "epoch": 0.14519685039370078, "grad_norm": 59.632354736328125, "learning_rate": 9.99928031496063e-06, "loss": 0.4868, "step": 461 }, { "epoch": 0.14551181102362204, "grad_norm": 103.69953918457031, "learning_rate": 9.999278740157482e-06, "loss": 0.5991, "step": 462 }, { "epoch": 0.1458267716535433, "grad_norm": 36.89337158203125, "learning_rate": 9.999277165354333e-06, "loss": 0.2331, "step": 463 }, { "epoch": 0.14614173228346455, "grad_norm": 42.319114685058594, "learning_rate": 9.999275590551182e-06, "loss": 0.3162, "step": 464 }, { "epoch": 0.14645669291338584, "grad_norm": 36.906063079833984, "learning_rate": 9.999274015748033e-06, "loss": 0.3396, "step": 465 }, { "epoch": 0.1467716535433071, "grad_norm": 25.45098304748535, "learning_rate": 9.999272440944882e-06, "loss": 0.1806, "step": 466 }, { "epoch": 0.14708661417322835, "grad_norm": 56.87234878540039, "learning_rate": 9.999270866141733e-06, "loss": 0.2911, "step": 467 }, { "epoch": 0.1474015748031496, "grad_norm": 20.72125244140625, "learning_rate": 9.999269291338584e-06, "loss": 0.0734, "step": 468 }, { "epoch": 0.14771653543307087, "grad_norm": 69.94297790527344, "learning_rate": 9.999267716535434e-06, "loss": 0.3137, "step": 469 }, { "epoch": 0.14803149606299212, "grad_norm": 55.783897399902344, "learning_rate": 9.999266141732285e-06, "loss": 0.3248, "step": 470 }, { "epoch": 0.14834645669291338, "grad_norm": 119.89505004882812, "learning_rate": 9.999264566929134e-06, "loss": 0.7729, "step": 471 }, { "epoch": 0.14866141732283464, "grad_norm": 45.10006332397461, "learning_rate": 9.999262992125985e-06, "loss": 0.287, "step": 472 }, { "epoch": 0.1489763779527559, "grad_norm": 146.70803833007812, "learning_rate": 9.999261417322836e-06, "loss": 1.0979, "step": 473 }, { "epoch": 0.14929133858267715, "grad_norm": 156.35951232910156, "learning_rate": 9.999259842519687e-06, "loss": 0.8397, "step": 474 }, { "epoch": 0.14960629921259844, "grad_norm": 99.72000122070312, "learning_rate": 9.999258267716536e-06, "loss": 0.3907, "step": 475 }, { "epoch": 0.1499212598425197, "grad_norm": 119.9110336303711, "learning_rate": 9.999256692913387e-06, "loss": 0.8118, "step": 476 }, { "epoch": 0.15023622047244095, "grad_norm": 160.22637939453125, "learning_rate": 9.999255118110236e-06, "loss": 0.6305, "step": 477 }, { "epoch": 0.1505511811023622, "grad_norm": 53.594276428222656, "learning_rate": 9.999253543307087e-06, "loss": 0.2044, "step": 478 }, { "epoch": 0.15086614173228347, "grad_norm": 76.94689178466797, "learning_rate": 9.999251968503938e-06, "loss": 1.0085, "step": 479 }, { "epoch": 0.15118110236220472, "grad_norm": 133.71861267089844, "learning_rate": 9.999250393700788e-06, "loss": 0.715, "step": 480 }, { "epoch": 0.15118110236220472, "eval_loss": 0.5550094246864319, "eval_runtime": 306.9821, "eval_samples_per_second": 0.381, "eval_steps_per_second": 0.381, "step": 480 }, { "epoch": 0.15149606299212598, "grad_norm": 65.10472106933594, "learning_rate": 9.999248818897639e-06, "loss": 0.3845, "step": 481 }, { "epoch": 0.15181102362204724, "grad_norm": 64.05907440185547, "learning_rate": 9.999247244094488e-06, "loss": 0.4831, "step": 482 }, { "epoch": 0.1521259842519685, "grad_norm": 29.167266845703125, "learning_rate": 9.999245669291339e-06, "loss": 0.1163, "step": 483 }, { "epoch": 0.15244094488188975, "grad_norm": 81.4149398803711, "learning_rate": 9.99924409448819e-06, "loss": 0.5063, "step": 484 }, { "epoch": 0.15275590551181104, "grad_norm": 29.339014053344727, "learning_rate": 9.999242519685041e-06, "loss": 0.3148, "step": 485 }, { "epoch": 0.1530708661417323, "grad_norm": 56.6673698425293, "learning_rate": 9.99924094488189e-06, "loss": 0.5286, "step": 486 }, { "epoch": 0.15338582677165355, "grad_norm": 52.65031051635742, "learning_rate": 9.999239370078741e-06, "loss": 0.2759, "step": 487 }, { "epoch": 0.1537007874015748, "grad_norm": 68.89445495605469, "learning_rate": 9.999237795275592e-06, "loss": 0.4206, "step": 488 }, { "epoch": 0.15401574803149606, "grad_norm": 57.06834030151367, "learning_rate": 9.999236220472442e-06, "loss": 0.3788, "step": 489 }, { "epoch": 0.15433070866141732, "grad_norm": 30.39971923828125, "learning_rate": 9.999234645669293e-06, "loss": 0.4785, "step": 490 }, { "epoch": 0.15464566929133858, "grad_norm": 58.35342025756836, "learning_rate": 9.999233070866142e-06, "loss": 0.5897, "step": 491 }, { "epoch": 0.15496062992125983, "grad_norm": 58.403533935546875, "learning_rate": 9.999231496062993e-06, "loss": 0.8812, "step": 492 }, { "epoch": 0.1552755905511811, "grad_norm": 71.63230895996094, "learning_rate": 9.999229921259844e-06, "loss": 0.1902, "step": 493 }, { "epoch": 0.15559055118110235, "grad_norm": 31.834192276000977, "learning_rate": 9.999228346456695e-06, "loss": 0.7004, "step": 494 }, { "epoch": 0.15590551181102363, "grad_norm": 35.32748794555664, "learning_rate": 9.999226771653544e-06, "loss": 0.5292, "step": 495 }, { "epoch": 0.1562204724409449, "grad_norm": 63.80234909057617, "learning_rate": 9.999225196850395e-06, "loss": 0.69, "step": 496 }, { "epoch": 0.15653543307086615, "grad_norm": 47.65721893310547, "learning_rate": 9.999223622047244e-06, "loss": 0.4307, "step": 497 }, { "epoch": 0.1568503937007874, "grad_norm": 23.857507705688477, "learning_rate": 9.999222047244095e-06, "loss": 0.4278, "step": 498 }, { "epoch": 0.15716535433070866, "grad_norm": 36.79646301269531, "learning_rate": 9.999220472440946e-06, "loss": 0.3957, "step": 499 }, { "epoch": 0.15748031496062992, "grad_norm": 37.079471588134766, "learning_rate": 9.999218897637796e-06, "loss": 0.5078, "step": 500 }, { "epoch": 0.15748031496062992, "eval_loss": 0.5599373579025269, "eval_runtime": 312.5913, "eval_samples_per_second": 0.374, "eval_steps_per_second": 0.374, "step": 500 }, { "epoch": 0.15779527559055118, "grad_norm": 31.857145309448242, "learning_rate": 9.999217322834647e-06, "loss": 0.3308, "step": 501 }, { "epoch": 0.15811023622047243, "grad_norm": 31.417692184448242, "learning_rate": 9.999215748031496e-06, "loss": 0.4725, "step": 502 }, { "epoch": 0.1584251968503937, "grad_norm": 100.06877899169922, "learning_rate": 9.999214173228347e-06, "loss": 0.8381, "step": 503 }, { "epoch": 0.15874015748031495, "grad_norm": 29.754446029663086, "learning_rate": 9.999212598425198e-06, "loss": 0.422, "step": 504 }, { "epoch": 0.15905511811023623, "grad_norm": 36.77560043334961, "learning_rate": 9.999211023622049e-06, "loss": 0.2553, "step": 505 }, { "epoch": 0.1593700787401575, "grad_norm": 45.064300537109375, "learning_rate": 9.999209448818898e-06, "loss": 0.4116, "step": 506 }, { "epoch": 0.15968503937007875, "grad_norm": 82.82674407958984, "learning_rate": 9.999207874015747e-06, "loss": 1.244, "step": 507 }, { "epoch": 0.16, "grad_norm": 30.511226654052734, "learning_rate": 9.999206299212598e-06, "loss": 0.2208, "step": 508 }, { "epoch": 0.16031496062992126, "grad_norm": 26.348651885986328, "learning_rate": 9.99920472440945e-06, "loss": 0.2132, "step": 509 }, { "epoch": 0.16062992125984252, "grad_norm": 88.1176986694336, "learning_rate": 9.9992031496063e-06, "loss": 0.5883, "step": 510 }, { "epoch": 0.16094488188976377, "grad_norm": 36.90149688720703, "learning_rate": 9.99920157480315e-06, "loss": 0.3298, "step": 511 }, { "epoch": 0.16125984251968503, "grad_norm": 30.67510414123535, "learning_rate": 9.9992e-06, "loss": 0.2843, "step": 512 }, { "epoch": 0.1615748031496063, "grad_norm": 89.83963012695312, "learning_rate": 9.999198425196852e-06, "loss": 0.4022, "step": 513 }, { "epoch": 0.16188976377952755, "grad_norm": 46.62565612792969, "learning_rate": 9.999196850393703e-06, "loss": 0.491, "step": 514 }, { "epoch": 0.16220472440944883, "grad_norm": 27.551525115966797, "learning_rate": 9.999195275590552e-06, "loss": 0.1321, "step": 515 }, { "epoch": 0.1625196850393701, "grad_norm": 15.754891395568848, "learning_rate": 9.999193700787403e-06, "loss": 0.0748, "step": 516 }, { "epoch": 0.16283464566929134, "grad_norm": 55.909244537353516, "learning_rate": 9.999192125984252e-06, "loss": 0.2661, "step": 517 }, { "epoch": 0.1631496062992126, "grad_norm": 37.772239685058594, "learning_rate": 9.999190551181103e-06, "loss": 0.1733, "step": 518 }, { "epoch": 0.16346456692913386, "grad_norm": 66.4226303100586, "learning_rate": 9.999188976377954e-06, "loss": 0.4846, "step": 519 }, { "epoch": 0.16377952755905512, "grad_norm": 37.98040771484375, "learning_rate": 9.999187401574804e-06, "loss": 0.1224, "step": 520 }, { "epoch": 0.16377952755905512, "eval_loss": 0.6723836064338684, "eval_runtime": 307.1583, "eval_samples_per_second": 0.381, "eval_steps_per_second": 0.381, "step": 520 }, { "epoch": 0.16409448818897637, "grad_norm": 53.96063995361328, "learning_rate": 9.999185826771655e-06, "loss": 0.4602, "step": 521 }, { "epoch": 0.16440944881889763, "grad_norm": 32.147621154785156, "learning_rate": 9.999184251968504e-06, "loss": 0.0799, "step": 522 }, { "epoch": 0.1647244094488189, "grad_norm": 104.01305389404297, "learning_rate": 9.999182677165355e-06, "loss": 0.2182, "step": 523 }, { "epoch": 0.16503937007874014, "grad_norm": 106.27403259277344, "learning_rate": 9.999181102362206e-06, "loss": 1.2292, "step": 524 }, { "epoch": 0.16535433070866143, "grad_norm": 33.72996139526367, "learning_rate": 9.999179527559057e-06, "loss": 0.2182, "step": 525 }, { "epoch": 0.16566929133858269, "grad_norm": 83.75643920898438, "learning_rate": 9.999177952755906e-06, "loss": 1.4189, "step": 526 }, { "epoch": 0.16598425196850394, "grad_norm": 81.67135620117188, "learning_rate": 9.999176377952755e-06, "loss": 0.3019, "step": 527 }, { "epoch": 0.1662992125984252, "grad_norm": 160.24029541015625, "learning_rate": 9.999174803149606e-06, "loss": 0.2741, "step": 528 }, { "epoch": 0.16661417322834646, "grad_norm": 73.85230255126953, "learning_rate": 9.999173228346457e-06, "loss": 1.037, "step": 529 }, { "epoch": 0.16692913385826771, "grad_norm": 87.96981811523438, "learning_rate": 9.999171653543308e-06, "loss": 0.3799, "step": 530 }, { "epoch": 0.16724409448818897, "grad_norm": 62.46072769165039, "learning_rate": 9.999170078740158e-06, "loss": 0.5531, "step": 531 }, { "epoch": 0.16755905511811023, "grad_norm": 151.0035858154297, "learning_rate": 9.999168503937009e-06, "loss": 1.4022, "step": 532 }, { "epoch": 0.16787401574803149, "grad_norm": 60.152671813964844, "learning_rate": 9.99916692913386e-06, "loss": 0.6236, "step": 533 }, { "epoch": 0.16818897637795274, "grad_norm": 102.74165344238281, "learning_rate": 9.99916535433071e-06, "loss": 1.2304, "step": 534 }, { "epoch": 0.16850393700787403, "grad_norm": 20.071189880371094, "learning_rate": 9.99916377952756e-06, "loss": 0.1347, "step": 535 }, { "epoch": 0.16881889763779528, "grad_norm": 43.08108139038086, "learning_rate": 9.999162204724411e-06, "loss": 0.4371, "step": 536 }, { "epoch": 0.16913385826771654, "grad_norm": 64.68496704101562, "learning_rate": 9.99916062992126e-06, "loss": 0.6083, "step": 537 }, { "epoch": 0.1694488188976378, "grad_norm": 39.30328369140625, "learning_rate": 9.999159055118111e-06, "loss": 0.5559, "step": 538 }, { "epoch": 0.16976377952755906, "grad_norm": 56.48727035522461, "learning_rate": 9.999157480314962e-06, "loss": 0.7363, "step": 539 }, { "epoch": 0.1700787401574803, "grad_norm": 26.32372283935547, "learning_rate": 9.999155905511812e-06, "loss": 0.3275, "step": 540 }, { "epoch": 0.1700787401574803, "eval_loss": 0.6582661867141724, "eval_runtime": 302.7589, "eval_samples_per_second": 0.386, "eval_steps_per_second": 0.386, "step": 540 }, { "epoch": 0.17039370078740157, "grad_norm": 69.46321868896484, "learning_rate": 9.999154330708663e-06, "loss": 0.3794, "step": 541 }, { "epoch": 0.17070866141732283, "grad_norm": 36.254520416259766, "learning_rate": 9.999152755905512e-06, "loss": 0.6954, "step": 542 }, { "epoch": 0.17102362204724408, "grad_norm": 55.22049331665039, "learning_rate": 9.999151181102363e-06, "loss": 0.4438, "step": 543 }, { "epoch": 0.17133858267716534, "grad_norm": 24.40268325805664, "learning_rate": 9.999149606299214e-06, "loss": 0.3439, "step": 544 }, { "epoch": 0.17165354330708663, "grad_norm": 29.554643630981445, "learning_rate": 9.999148031496065e-06, "loss": 0.4619, "step": 545 }, { "epoch": 0.17196850393700788, "grad_norm": 34.69717788696289, "learning_rate": 9.999146456692914e-06, "loss": 0.3959, "step": 546 }, { "epoch": 0.17228346456692914, "grad_norm": 49.8066520690918, "learning_rate": 9.999144881889763e-06, "loss": 0.6021, "step": 547 }, { "epoch": 0.1725984251968504, "grad_norm": 29.715484619140625, "learning_rate": 9.999143307086614e-06, "loss": 0.5588, "step": 548 }, { "epoch": 0.17291338582677165, "grad_norm": 24.99655532836914, "learning_rate": 9.999141732283465e-06, "loss": 0.5414, "step": 549 }, { "epoch": 0.1732283464566929, "grad_norm": 21.271596908569336, "learning_rate": 9.999140157480316e-06, "loss": 0.4187, "step": 550 }, { "epoch": 0.17354330708661417, "grad_norm": 51.398712158203125, "learning_rate": 9.999138582677166e-06, "loss": 0.6258, "step": 551 }, { "epoch": 0.17385826771653543, "grad_norm": 51.94355010986328, "learning_rate": 9.999137007874017e-06, "loss": 0.3318, "step": 552 }, { "epoch": 0.17417322834645668, "grad_norm": 30.253637313842773, "learning_rate": 9.999135433070866e-06, "loss": 0.2938, "step": 553 }, { "epoch": 0.17448818897637794, "grad_norm": 44.27308654785156, "learning_rate": 9.999133858267717e-06, "loss": 0.5734, "step": 554 }, { "epoch": 0.17480314960629922, "grad_norm": 33.18519973754883, "learning_rate": 9.999132283464568e-06, "loss": 0.566, "step": 555 }, { "epoch": 0.17511811023622048, "grad_norm": 48.19463348388672, "learning_rate": 9.999130708661419e-06, "loss": 0.6241, "step": 556 }, { "epoch": 0.17543307086614174, "grad_norm": 34.98441696166992, "learning_rate": 9.999129133858268e-06, "loss": 0.4501, "step": 557 }, { "epoch": 0.175748031496063, "grad_norm": 30.51637077331543, "learning_rate": 9.99912755905512e-06, "loss": 0.2408, "step": 558 }, { "epoch": 0.17606299212598425, "grad_norm": 65.26117706298828, "learning_rate": 9.99912598425197e-06, "loss": 0.6767, "step": 559 }, { "epoch": 0.1763779527559055, "grad_norm": 37.751888275146484, "learning_rate": 9.99912440944882e-06, "loss": 0.3929, "step": 560 }, { "epoch": 0.1763779527559055, "eval_loss": 0.5896673798561096, "eval_runtime": 296.0217, "eval_samples_per_second": 0.395, "eval_steps_per_second": 0.395, "step": 560 }, { "epoch": 0.17669291338582677, "grad_norm": 47.340492248535156, "learning_rate": 9.99912283464567e-06, "loss": 0.3066, "step": 561 }, { "epoch": 0.17700787401574802, "grad_norm": 86.48099517822266, "learning_rate": 9.99912125984252e-06, "loss": 0.4933, "step": 562 }, { "epoch": 0.17732283464566928, "grad_norm": 33.50886917114258, "learning_rate": 9.99911968503937e-06, "loss": 0.1638, "step": 563 }, { "epoch": 0.17763779527559054, "grad_norm": 48.44023895263672, "learning_rate": 9.999118110236222e-06, "loss": 0.4082, "step": 564 }, { "epoch": 0.17795275590551182, "grad_norm": 54.50333786010742, "learning_rate": 9.999116535433073e-06, "loss": 0.3993, "step": 565 }, { "epoch": 0.17826771653543308, "grad_norm": 82.37716674804688, "learning_rate": 9.999114960629922e-06, "loss": 0.9308, "step": 566 }, { "epoch": 0.17858267716535434, "grad_norm": 53.5648307800293, "learning_rate": 9.999113385826771e-06, "loss": 0.757, "step": 567 }, { "epoch": 0.1788976377952756, "grad_norm": 51.68220901489258, "learning_rate": 9.999111811023622e-06, "loss": 0.408, "step": 568 }, { "epoch": 0.17921259842519685, "grad_norm": 32.468666076660156, "learning_rate": 9.999110236220473e-06, "loss": 0.1988, "step": 569 }, { "epoch": 0.1795275590551181, "grad_norm": 92.7658462524414, "learning_rate": 9.999108661417324e-06, "loss": 0.737, "step": 570 }, { "epoch": 0.17984251968503936, "grad_norm": 94.98796844482422, "learning_rate": 9.999107086614174e-06, "loss": 0.2773, "step": 571 }, { "epoch": 0.18015748031496062, "grad_norm": 45.95973587036133, "learning_rate": 9.999105511811025e-06, "loss": 0.417, "step": 572 }, { "epoch": 0.18047244094488188, "grad_norm": 42.59912872314453, "learning_rate": 9.999103937007874e-06, "loss": 0.437, "step": 573 }, { "epoch": 0.18078740157480314, "grad_norm": 73.75167083740234, "learning_rate": 9.999102362204725e-06, "loss": 0.2676, "step": 574 }, { "epoch": 0.18110236220472442, "grad_norm": 46.682533264160156, "learning_rate": 9.999100787401576e-06, "loss": 0.1902, "step": 575 }, { "epoch": 0.18141732283464568, "grad_norm": 35.30620574951172, "learning_rate": 9.999099212598427e-06, "loss": 0.2953, "step": 576 }, { "epoch": 0.18173228346456693, "grad_norm": 113.94246673583984, "learning_rate": 9.999097637795276e-06, "loss": 0.7046, "step": 577 }, { "epoch": 0.1820472440944882, "grad_norm": 207.06141662597656, "learning_rate": 9.999096062992125e-06, "loss": 0.5952, "step": 578 }, { "epoch": 0.18236220472440945, "grad_norm": 34.89611053466797, "learning_rate": 9.999094488188978e-06, "loss": 0.2947, "step": 579 }, { "epoch": 0.1826771653543307, "grad_norm": 108.61929321289062, "learning_rate": 9.999092913385827e-06, "loss": 0.2417, "step": 580 }, { "epoch": 0.1826771653543307, "eval_loss": 0.5409280061721802, "eval_runtime": 307.2823, "eval_samples_per_second": 0.381, "eval_steps_per_second": 0.381, "step": 580 }, { "epoch": 0.18299212598425196, "grad_norm": 15.958565711975098, "learning_rate": 9.999091338582678e-06, "loss": 0.1013, "step": 581 }, { "epoch": 0.18330708661417322, "grad_norm": 68.7283706665039, "learning_rate": 9.999089763779528e-06, "loss": 0.8978, "step": 582 }, { "epoch": 0.18362204724409448, "grad_norm": 80.29981994628906, "learning_rate": 9.999088188976379e-06, "loss": 0.5132, "step": 583 }, { "epoch": 0.18393700787401573, "grad_norm": 25.062374114990234, "learning_rate": 9.99908661417323e-06, "loss": 0.1291, "step": 584 }, { "epoch": 0.18425196850393702, "grad_norm": 58.20054244995117, "learning_rate": 9.99908503937008e-06, "loss": 0.7141, "step": 585 }, { "epoch": 0.18456692913385828, "grad_norm": 73.24835968017578, "learning_rate": 9.99908346456693e-06, "loss": 0.5306, "step": 586 }, { "epoch": 0.18488188976377953, "grad_norm": 145.0093994140625, "learning_rate": 9.99908188976378e-06, "loss": 0.6498, "step": 587 }, { "epoch": 0.1851968503937008, "grad_norm": 62.22865295410156, "learning_rate": 9.99908031496063e-06, "loss": 0.6765, "step": 588 }, { "epoch": 0.18551181102362205, "grad_norm": 11.53496265411377, "learning_rate": 9.999078740157481e-06, "loss": 0.0789, "step": 589 }, { "epoch": 0.1858267716535433, "grad_norm": 104.29961395263672, "learning_rate": 9.999077165354332e-06, "loss": 0.444, "step": 590 }, { "epoch": 0.18614173228346456, "grad_norm": 26.030893325805664, "learning_rate": 9.999075590551182e-06, "loss": 0.0719, "step": 591 }, { "epoch": 0.18645669291338582, "grad_norm": 59.57289123535156, "learning_rate": 9.999074015748033e-06, "loss": 0.4676, "step": 592 }, { "epoch": 0.18677165354330708, "grad_norm": 47.10686111450195, "learning_rate": 9.999072440944882e-06, "loss": 0.6408, "step": 593 }, { "epoch": 0.18708661417322833, "grad_norm": 97.91781616210938, "learning_rate": 9.999070866141733e-06, "loss": 0.8239, "step": 594 }, { "epoch": 0.18740157480314962, "grad_norm": 82.05168151855469, "learning_rate": 9.999069291338584e-06, "loss": 0.2311, "step": 595 }, { "epoch": 0.18771653543307087, "grad_norm": 73.30006408691406, "learning_rate": 9.999067716535435e-06, "loss": 0.6648, "step": 596 }, { "epoch": 0.18803149606299213, "grad_norm": 36.88441467285156, "learning_rate": 9.999066141732284e-06, "loss": 0.411, "step": 597 }, { "epoch": 0.1883464566929134, "grad_norm": 21.77279281616211, "learning_rate": 9.999064566929133e-06, "loss": 0.0762, "step": 598 }, { "epoch": 0.18866141732283465, "grad_norm": 41.469337463378906, "learning_rate": 9.999062992125984e-06, "loss": 0.2768, "step": 599 }, { "epoch": 0.1889763779527559, "grad_norm": 23.42574119567871, "learning_rate": 9.999061417322835e-06, "loss": 0.3597, "step": 600 }, { "epoch": 0.1889763779527559, "eval_loss": 0.5386444330215454, "eval_runtime": 309.0194, "eval_samples_per_second": 0.379, "eval_steps_per_second": 0.379, "step": 600 }, { "epoch": 0.18929133858267716, "grad_norm": 49.9908332824707, "learning_rate": 9.999059842519686e-06, "loss": 0.3259, "step": 601 }, { "epoch": 0.18960629921259842, "grad_norm": 111.2554702758789, "learning_rate": 9.999058267716536e-06, "loss": 0.451, "step": 602 }, { "epoch": 0.18992125984251967, "grad_norm": 87.17556762695312, "learning_rate": 9.999056692913387e-06, "loss": 0.2611, "step": 603 }, { "epoch": 0.19023622047244093, "grad_norm": 24.059478759765625, "learning_rate": 9.999055118110238e-06, "loss": 0.3407, "step": 604 }, { "epoch": 0.19055118110236222, "grad_norm": 121.00428009033203, "learning_rate": 9.999053543307089e-06, "loss": 1.3334, "step": 605 }, { "epoch": 0.19086614173228347, "grad_norm": 26.074909210205078, "learning_rate": 9.999051968503938e-06, "loss": 0.2243, "step": 606 }, { "epoch": 0.19118110236220473, "grad_norm": 60.16206359863281, "learning_rate": 9.999050393700787e-06, "loss": 0.1694, "step": 607 }, { "epoch": 0.191496062992126, "grad_norm": 94.56045532226562, "learning_rate": 9.999048818897638e-06, "loss": 0.453, "step": 608 }, { "epoch": 0.19181102362204724, "grad_norm": 106.34618377685547, "learning_rate": 9.99904724409449e-06, "loss": 0.5918, "step": 609 }, { "epoch": 0.1921259842519685, "grad_norm": 33.97660827636719, "learning_rate": 9.99904566929134e-06, "loss": 0.6824, "step": 610 }, { "epoch": 0.19244094488188976, "grad_norm": 25.390705108642578, "learning_rate": 9.99904409448819e-06, "loss": 0.4475, "step": 611 }, { "epoch": 0.19275590551181102, "grad_norm": 77.29833984375, "learning_rate": 9.99904251968504e-06, "loss": 0.5393, "step": 612 }, { "epoch": 0.19307086614173227, "grad_norm": 56.867801666259766, "learning_rate": 9.99904094488189e-06, "loss": 0.4419, "step": 613 }, { "epoch": 0.19338582677165353, "grad_norm": 62.77841567993164, "learning_rate": 9.99903937007874e-06, "loss": 0.7734, "step": 614 }, { "epoch": 0.19370078740157481, "grad_norm": 76.52714538574219, "learning_rate": 9.999037795275592e-06, "loss": 0.7938, "step": 615 }, { "epoch": 0.19401574803149607, "grad_norm": 26.540964126586914, "learning_rate": 9.999036220472443e-06, "loss": 0.2187, "step": 616 }, { "epoch": 0.19433070866141733, "grad_norm": 20.736865997314453, "learning_rate": 9.999034645669292e-06, "loss": 0.1165, "step": 617 }, { "epoch": 0.19464566929133859, "grad_norm": 86.58599853515625, "learning_rate": 9.999033070866141e-06, "loss": 0.4113, "step": 618 }, { "epoch": 0.19496062992125984, "grad_norm": 17.32464599609375, "learning_rate": 9.999031496062992e-06, "loss": 0.2206, "step": 619 }, { "epoch": 0.1952755905511811, "grad_norm": 54.347904205322266, "learning_rate": 9.999029921259843e-06, "loss": 0.6016, "step": 620 }, { "epoch": 0.1952755905511811, "eval_loss": 0.5922896265983582, "eval_runtime": 309.8531, "eval_samples_per_second": 0.378, "eval_steps_per_second": 0.378, "step": 620 }, { "epoch": 0.19559055118110236, "grad_norm": 62.3914794921875, "learning_rate": 9.999028346456694e-06, "loss": 0.5295, "step": 621 }, { "epoch": 0.1959055118110236, "grad_norm": 29.178783416748047, "learning_rate": 9.999026771653544e-06, "loss": 0.2871, "step": 622 }, { "epoch": 0.19622047244094487, "grad_norm": 28.286312103271484, "learning_rate": 9.999025196850395e-06, "loss": 0.3118, "step": 623 }, { "epoch": 0.19653543307086613, "grad_norm": 25.735105514526367, "learning_rate": 9.999023622047244e-06, "loss": 0.2462, "step": 624 }, { "epoch": 0.1968503937007874, "grad_norm": 173.3253936767578, "learning_rate": 9.999022047244095e-06, "loss": 0.9632, "step": 625 }, { "epoch": 0.19716535433070867, "grad_norm": 85.50897979736328, "learning_rate": 9.999020472440946e-06, "loss": 0.643, "step": 626 }, { "epoch": 0.19748031496062993, "grad_norm": 29.653247833251953, "learning_rate": 9.999018897637795e-06, "loss": 0.4286, "step": 627 }, { "epoch": 0.19779527559055118, "grad_norm": 37.991817474365234, "learning_rate": 9.999017322834646e-06, "loss": 0.3509, "step": 628 }, { "epoch": 0.19811023622047244, "grad_norm": 97.16632843017578, "learning_rate": 9.999015748031497e-06, "loss": 0.3752, "step": 629 }, { "epoch": 0.1984251968503937, "grad_norm": 50.35764694213867, "learning_rate": 9.999014173228348e-06, "loss": 0.5741, "step": 630 }, { "epoch": 0.19874015748031496, "grad_norm": 28.755027770996094, "learning_rate": 9.999012598425197e-06, "loss": 0.278, "step": 631 }, { "epoch": 0.1990551181102362, "grad_norm": 74.70952606201172, "learning_rate": 9.999011023622048e-06, "loss": 0.6146, "step": 632 }, { "epoch": 0.19937007874015747, "grad_norm": 78.7210693359375, "learning_rate": 9.999009448818898e-06, "loss": 0.2827, "step": 633 }, { "epoch": 0.19968503937007873, "grad_norm": 17.010251998901367, "learning_rate": 9.999007874015749e-06, "loss": 0.239, "step": 634 }, { "epoch": 0.2, "grad_norm": 81.048095703125, "learning_rate": 9.9990062992126e-06, "loss": 1.0794, "step": 635 }, { "epoch": 0.20031496062992127, "grad_norm": 30.376026153564453, "learning_rate": 9.99900472440945e-06, "loss": 0.2238, "step": 636 }, { "epoch": 0.20062992125984253, "grad_norm": 45.60966873168945, "learning_rate": 9.9990031496063e-06, "loss": 0.7123, "step": 637 }, { "epoch": 0.20094488188976378, "grad_norm": 47.69063186645508, "learning_rate": 9.99900157480315e-06, "loss": 0.5227, "step": 638 }, { "epoch": 0.20125984251968504, "grad_norm": 25.39598274230957, "learning_rate": 9.999e-06, "loss": 0.0982, "step": 639 }, { "epoch": 0.2015748031496063, "grad_norm": 36.95790481567383, "learning_rate": 9.998998425196851e-06, "loss": 0.4809, "step": 640 }, { "epoch": 0.2015748031496063, "eval_loss": 0.5429102182388306, "eval_runtime": 295.5911, "eval_samples_per_second": 0.396, "eval_steps_per_second": 0.396, "step": 640 }, { "epoch": 0.20188976377952755, "grad_norm": 47.62435531616211, "learning_rate": 9.998996850393702e-06, "loss": 0.6058, "step": 641 }, { "epoch": 0.2022047244094488, "grad_norm": 31.27633285522461, "learning_rate": 9.998995275590552e-06, "loss": 0.2615, "step": 642 }, { "epoch": 0.20251968503937007, "grad_norm": 47.4228630065918, "learning_rate": 9.998993700787403e-06, "loss": 0.6306, "step": 643 }, { "epoch": 0.20283464566929135, "grad_norm": 67.43081665039062, "learning_rate": 9.998992125984252e-06, "loss": 0.654, "step": 644 }, { "epoch": 0.2031496062992126, "grad_norm": 57.79238510131836, "learning_rate": 9.998990551181103e-06, "loss": 0.3423, "step": 645 }, { "epoch": 0.20346456692913387, "grad_norm": 38.009735107421875, "learning_rate": 9.998988976377954e-06, "loss": 0.5406, "step": 646 }, { "epoch": 0.20377952755905512, "grad_norm": 36.380531311035156, "learning_rate": 9.998987401574805e-06, "loss": 0.3381, "step": 647 }, { "epoch": 0.20409448818897638, "grad_norm": 33.59734344482422, "learning_rate": 9.998985826771654e-06, "loss": 0.2425, "step": 648 }, { "epoch": 0.20440944881889764, "grad_norm": 40.288902282714844, "learning_rate": 9.998984251968505e-06, "loss": 0.4959, "step": 649 }, { "epoch": 0.2047244094488189, "grad_norm": 51.55220413208008, "learning_rate": 9.998982677165356e-06, "loss": 0.4607, "step": 650 }, { "epoch": 0.20503937007874015, "grad_norm": 51.30996322631836, "learning_rate": 9.998981102362205e-06, "loss": 0.4445, "step": 651 }, { "epoch": 0.2053543307086614, "grad_norm": 139.98907470703125, "learning_rate": 9.998979527559056e-06, "loss": 1.0222, "step": 652 }, { "epoch": 0.20566929133858267, "grad_norm": 42.455169677734375, "learning_rate": 9.998977952755906e-06, "loss": 0.3876, "step": 653 }, { "epoch": 0.20598425196850395, "grad_norm": 22.130889892578125, "learning_rate": 9.998976377952757e-06, "loss": 0.3178, "step": 654 }, { "epoch": 0.2062992125984252, "grad_norm": 27.28899574279785, "learning_rate": 9.998974803149608e-06, "loss": 0.236, "step": 655 }, { "epoch": 0.20661417322834646, "grad_norm": 19.113969802856445, "learning_rate": 9.998973228346459e-06, "loss": 0.2199, "step": 656 }, { "epoch": 0.20692913385826772, "grad_norm": 70.42593383789062, "learning_rate": 9.998971653543308e-06, "loss": 0.5523, "step": 657 }, { "epoch": 0.20724409448818898, "grad_norm": 51.796207427978516, "learning_rate": 9.998970078740157e-06, "loss": 0.5754, "step": 658 }, { "epoch": 0.20755905511811024, "grad_norm": 95.93733215332031, "learning_rate": 9.998968503937008e-06, "loss": 0.7566, "step": 659 }, { "epoch": 0.2078740157480315, "grad_norm": 27.715049743652344, "learning_rate": 9.99896692913386e-06, "loss": 0.2745, "step": 660 }, { "epoch": 0.2078740157480315, "eval_loss": 0.4993188977241516, "eval_runtime": 309.1118, "eval_samples_per_second": 0.379, "eval_steps_per_second": 0.379, "step": 660 }, { "epoch": 0.20818897637795275, "grad_norm": 43.9770393371582, "learning_rate": 9.99896535433071e-06, "loss": 0.5356, "step": 661 }, { "epoch": 0.208503937007874, "grad_norm": 28.040441513061523, "learning_rate": 9.99896377952756e-06, "loss": 0.315, "step": 662 }, { "epoch": 0.20881889763779526, "grad_norm": 45.72781753540039, "learning_rate": 9.99896220472441e-06, "loss": 0.4161, "step": 663 }, { "epoch": 0.20913385826771655, "grad_norm": 27.964317321777344, "learning_rate": 9.99896062992126e-06, "loss": 0.2185, "step": 664 }, { "epoch": 0.2094488188976378, "grad_norm": 32.59132766723633, "learning_rate": 9.99895905511811e-06, "loss": 0.3825, "step": 665 }, { "epoch": 0.20976377952755906, "grad_norm": 38.677207946777344, "learning_rate": 9.998957480314962e-06, "loss": 0.2456, "step": 666 }, { "epoch": 0.21007874015748032, "grad_norm": 14.392455101013184, "learning_rate": 9.998955905511813e-06, "loss": 0.0907, "step": 667 }, { "epoch": 0.21039370078740158, "grad_norm": 50.628910064697266, "learning_rate": 9.998954330708662e-06, "loss": 0.7256, "step": 668 }, { "epoch": 0.21070866141732283, "grad_norm": 60.51618194580078, "learning_rate": 9.998952755905511e-06, "loss": 0.4601, "step": 669 }, { "epoch": 0.2110236220472441, "grad_norm": 40.12845230102539, "learning_rate": 9.998951181102362e-06, "loss": 0.5639, "step": 670 }, { "epoch": 0.21133858267716535, "grad_norm": 45.10561752319336, "learning_rate": 9.998949606299213e-06, "loss": 0.4125, "step": 671 }, { "epoch": 0.2116535433070866, "grad_norm": 79.58858489990234, "learning_rate": 9.998948031496064e-06, "loss": 0.8415, "step": 672 }, { "epoch": 0.21196850393700786, "grad_norm": 35.8792724609375, "learning_rate": 9.998946456692914e-06, "loss": 0.3386, "step": 673 }, { "epoch": 0.21228346456692915, "grad_norm": 13.449446678161621, "learning_rate": 9.998944881889765e-06, "loss": 0.1095, "step": 674 }, { "epoch": 0.2125984251968504, "grad_norm": 82.7651596069336, "learning_rate": 9.998943307086616e-06, "loss": 0.331, "step": 675 }, { "epoch": 0.21291338582677166, "grad_norm": 80.85045623779297, "learning_rate": 9.998941732283467e-06, "loss": 0.9251, "step": 676 }, { "epoch": 0.21322834645669292, "grad_norm": 34.32810974121094, "learning_rate": 9.998940157480316e-06, "loss": 0.2547, "step": 677 }, { "epoch": 0.21354330708661418, "grad_norm": 39.6412239074707, "learning_rate": 9.998938582677165e-06, "loss": 0.3042, "step": 678 }, { "epoch": 0.21385826771653543, "grad_norm": 29.662391662597656, "learning_rate": 9.998937007874016e-06, "loss": 0.353, "step": 679 }, { "epoch": 0.2141732283464567, "grad_norm": 15.640954971313477, "learning_rate": 9.998935433070867e-06, "loss": 0.0869, "step": 680 }, { "epoch": 0.2141732283464567, "eval_loss": 0.5842440128326416, "eval_runtime": 303.7569, "eval_samples_per_second": 0.385, "eval_steps_per_second": 0.385, "step": 680 }, { "epoch": 0.21448818897637795, "grad_norm": 47.02978515625, "learning_rate": 9.998933858267718e-06, "loss": 0.6137, "step": 681 }, { "epoch": 0.2148031496062992, "grad_norm": 23.671756744384766, "learning_rate": 9.998932283464568e-06, "loss": 0.2595, "step": 682 }, { "epoch": 0.21511811023622046, "grad_norm": 35.9589729309082, "learning_rate": 9.998930708661418e-06, "loss": 0.3242, "step": 683 }, { "epoch": 0.21543307086614175, "grad_norm": 42.95161819458008, "learning_rate": 9.998929133858268e-06, "loss": 0.2721, "step": 684 }, { "epoch": 0.215748031496063, "grad_norm": 58.640968322753906, "learning_rate": 9.998927559055119e-06, "loss": 0.3166, "step": 685 }, { "epoch": 0.21606299212598426, "grad_norm": 35.683467864990234, "learning_rate": 9.99892598425197e-06, "loss": 0.3474, "step": 686 }, { "epoch": 0.21637795275590552, "grad_norm": 9.069628715515137, "learning_rate": 9.99892440944882e-06, "loss": 0.0522, "step": 687 }, { "epoch": 0.21669291338582677, "grad_norm": 29.538806915283203, "learning_rate": 9.99892283464567e-06, "loss": 0.3609, "step": 688 }, { "epoch": 0.21700787401574803, "grad_norm": 40.99892044067383, "learning_rate": 9.99892125984252e-06, "loss": 0.5832, "step": 689 }, { "epoch": 0.2173228346456693, "grad_norm": 48.680538177490234, "learning_rate": 9.99891968503937e-06, "loss": 0.5598, "step": 690 }, { "epoch": 0.21763779527559055, "grad_norm": 37.50501251220703, "learning_rate": 9.998918110236221e-06, "loss": 0.211, "step": 691 }, { "epoch": 0.2179527559055118, "grad_norm": 145.6537322998047, "learning_rate": 9.998916535433072e-06, "loss": 1.0281, "step": 692 }, { "epoch": 0.21826771653543306, "grad_norm": 82.14835357666016, "learning_rate": 9.998914960629922e-06, "loss": 0.4626, "step": 693 }, { "epoch": 0.21858267716535434, "grad_norm": 114.24347686767578, "learning_rate": 9.998913385826773e-06, "loss": 0.4935, "step": 694 }, { "epoch": 0.2188976377952756, "grad_norm": 44.75710678100586, "learning_rate": 9.998911811023622e-06, "loss": 0.2071, "step": 695 }, { "epoch": 0.21921259842519686, "grad_norm": 35.995880126953125, "learning_rate": 9.998910236220473e-06, "loss": 0.1709, "step": 696 }, { "epoch": 0.21952755905511812, "grad_norm": 74.69808197021484, "learning_rate": 9.998908661417324e-06, "loss": 0.7127, "step": 697 }, { "epoch": 0.21984251968503937, "grad_norm": 46.23347854614258, "learning_rate": 9.998907086614173e-06, "loss": 0.2524, "step": 698 }, { "epoch": 0.22015748031496063, "grad_norm": 23.196392059326172, "learning_rate": 9.998905511811024e-06, "loss": 0.0743, "step": 699 }, { "epoch": 0.2204724409448819, "grad_norm": 68.11265563964844, "learning_rate": 9.998903937007875e-06, "loss": 0.5743, "step": 700 }, { "epoch": 0.2204724409448819, "eval_loss": 0.6227251291275024, "eval_runtime": 306.9107, "eval_samples_per_second": 0.381, "eval_steps_per_second": 0.381, "step": 700 }, { "epoch": 0.22078740157480314, "grad_norm": 49.117034912109375, "learning_rate": 9.998902362204726e-06, "loss": 0.5395, "step": 701 }, { "epoch": 0.2211023622047244, "grad_norm": 72.21612548828125, "learning_rate": 9.998900787401575e-06, "loss": 0.1053, "step": 702 }, { "epoch": 0.22141732283464566, "grad_norm": 106.6301498413086, "learning_rate": 9.998899212598426e-06, "loss": 0.4932, "step": 703 }, { "epoch": 0.22173228346456694, "grad_norm": 171.39988708496094, "learning_rate": 9.998897637795276e-06, "loss": 0.9959, "step": 704 }, { "epoch": 0.2220472440944882, "grad_norm": 33.84798812866211, "learning_rate": 9.998896062992127e-06, "loss": 0.3472, "step": 705 }, { "epoch": 0.22236220472440946, "grad_norm": 35.71989059448242, "learning_rate": 9.998894488188978e-06, "loss": 0.1522, "step": 706 }, { "epoch": 0.22267716535433071, "grad_norm": 65.45905303955078, "learning_rate": 9.998892913385829e-06, "loss": 0.9923, "step": 707 }, { "epoch": 0.22299212598425197, "grad_norm": 88.24420166015625, "learning_rate": 9.998891338582678e-06, "loss": 0.7105, "step": 708 }, { "epoch": 0.22330708661417323, "grad_norm": 71.33065032958984, "learning_rate": 9.998889763779527e-06, "loss": 1.2078, "step": 709 }, { "epoch": 0.22362204724409449, "grad_norm": 25.207542419433594, "learning_rate": 9.998888188976378e-06, "loss": 0.1815, "step": 710 }, { "epoch": 0.22393700787401574, "grad_norm": 23.143169403076172, "learning_rate": 9.99888661417323e-06, "loss": 0.1984, "step": 711 }, { "epoch": 0.224251968503937, "grad_norm": 32.797752380371094, "learning_rate": 9.99888503937008e-06, "loss": 0.3693, "step": 712 }, { "epoch": 0.22456692913385826, "grad_norm": 35.14696502685547, "learning_rate": 9.99888346456693e-06, "loss": 0.4894, "step": 713 }, { "epoch": 0.22488188976377954, "grad_norm": 24.498680114746094, "learning_rate": 9.99888188976378e-06, "loss": 0.3493, "step": 714 }, { "epoch": 0.2251968503937008, "grad_norm": 57.850730895996094, "learning_rate": 9.99888031496063e-06, "loss": 0.483, "step": 715 }, { "epoch": 0.22551181102362206, "grad_norm": 47.62955093383789, "learning_rate": 9.998878740157481e-06, "loss": 0.3677, "step": 716 }, { "epoch": 0.2258267716535433, "grad_norm": 25.785797119140625, "learning_rate": 9.998877165354332e-06, "loss": 0.324, "step": 717 }, { "epoch": 0.22614173228346457, "grad_norm": 44.619041442871094, "learning_rate": 9.998875590551181e-06, "loss": 0.3471, "step": 718 }, { "epoch": 0.22645669291338583, "grad_norm": 10.855001449584961, "learning_rate": 9.998874015748032e-06, "loss": 0.205, "step": 719 }, { "epoch": 0.22677165354330708, "grad_norm": 53.096466064453125, "learning_rate": 9.998872440944883e-06, "loss": 0.4501, "step": 720 }, { "epoch": 0.22677165354330708, "eval_loss": 0.5146042704582214, "eval_runtime": 316.9221, "eval_samples_per_second": 0.369, "eval_steps_per_second": 0.369, "step": 720 }, { "epoch": 0.22708661417322834, "grad_norm": 44.9933967590332, "learning_rate": 9.998870866141734e-06, "loss": 0.4339, "step": 721 }, { "epoch": 0.2274015748031496, "grad_norm": 46.176700592041016, "learning_rate": 9.998869291338583e-06, "loss": 0.4345, "step": 722 }, { "epoch": 0.22771653543307085, "grad_norm": 21.4019832611084, "learning_rate": 9.998867716535434e-06, "loss": 0.5765, "step": 723 }, { "epoch": 0.22803149606299214, "grad_norm": 17.84157943725586, "learning_rate": 9.998866141732284e-06, "loss": 0.1562, "step": 724 }, { "epoch": 0.2283464566929134, "grad_norm": 18.121217727661133, "learning_rate": 9.998864566929135e-06, "loss": 0.1687, "step": 725 }, { "epoch": 0.22866141732283465, "grad_norm": 32.00659942626953, "learning_rate": 9.998862992125986e-06, "loss": 0.1804, "step": 726 }, { "epoch": 0.2289763779527559, "grad_norm": 32.684757232666016, "learning_rate": 9.998861417322837e-06, "loss": 0.2921, "step": 727 }, { "epoch": 0.22929133858267717, "grad_norm": 65.45207214355469, "learning_rate": 9.998859842519686e-06, "loss": 0.237, "step": 728 }, { "epoch": 0.22960629921259842, "grad_norm": 24.639812469482422, "learning_rate": 9.998858267716535e-06, "loss": 0.1602, "step": 729 }, { "epoch": 0.22992125984251968, "grad_norm": 105.74681854248047, "learning_rate": 9.998856692913386e-06, "loss": 0.6624, "step": 730 }, { "epoch": 0.23023622047244094, "grad_norm": 123.29035949707031, "learning_rate": 9.998855118110237e-06, "loss": 0.1592, "step": 731 }, { "epoch": 0.2305511811023622, "grad_norm": 75.88599395751953, "learning_rate": 9.998853543307088e-06, "loss": 0.8749, "step": 732 }, { "epoch": 0.23086614173228345, "grad_norm": 93.44769287109375, "learning_rate": 9.998851968503938e-06, "loss": 0.2961, "step": 733 }, { "epoch": 0.23118110236220474, "grad_norm": 149.9324188232422, "learning_rate": 9.998850393700789e-06, "loss": 1.0564, "step": 734 }, { "epoch": 0.231496062992126, "grad_norm": 158.12449645996094, "learning_rate": 9.998848818897638e-06, "loss": 1.6293, "step": 735 }, { "epoch": 0.23181102362204725, "grad_norm": 149.76597595214844, "learning_rate": 9.998847244094489e-06, "loss": 1.6253, "step": 736 }, { "epoch": 0.2321259842519685, "grad_norm": 40.33038330078125, "learning_rate": 9.99884566929134e-06, "loss": 0.5073, "step": 737 }, { "epoch": 0.23244094488188977, "grad_norm": 67.39472198486328, "learning_rate": 9.998844094488189e-06, "loss": 0.303, "step": 738 }, { "epoch": 0.23275590551181102, "grad_norm": 48.84297561645508, "learning_rate": 9.99884251968504e-06, "loss": 0.3308, "step": 739 }, { "epoch": 0.23307086614173228, "grad_norm": 50.47072219848633, "learning_rate": 9.99884094488189e-06, "loss": 0.5499, "step": 740 }, { "epoch": 0.23307086614173228, "eval_loss": 0.563422441482544, "eval_runtime": 300.8217, "eval_samples_per_second": 0.389, "eval_steps_per_second": 0.389, "step": 740 }, { "epoch": 0.23338582677165354, "grad_norm": 40.9046745300293, "learning_rate": 9.99883937007874e-06, "loss": 0.275, "step": 741 }, { "epoch": 0.2337007874015748, "grad_norm": 68.57015991210938, "learning_rate": 9.998837795275591e-06, "loss": 0.5954, "step": 742 }, { "epoch": 0.23401574803149605, "grad_norm": 32.3016357421875, "learning_rate": 9.998836220472442e-06, "loss": 0.1714, "step": 743 }, { "epoch": 0.23433070866141734, "grad_norm": 35.811279296875, "learning_rate": 9.998834645669292e-06, "loss": 0.3879, "step": 744 }, { "epoch": 0.2346456692913386, "grad_norm": 25.600717544555664, "learning_rate": 9.998833070866143e-06, "loss": 0.0832, "step": 745 }, { "epoch": 0.23496062992125985, "grad_norm": 38.07957458496094, "learning_rate": 9.998831496062994e-06, "loss": 0.6504, "step": 746 }, { "epoch": 0.2352755905511811, "grad_norm": 62.875614166259766, "learning_rate": 9.998829921259845e-06, "loss": 0.6823, "step": 747 }, { "epoch": 0.23559055118110236, "grad_norm": 95.92493438720703, "learning_rate": 9.998828346456694e-06, "loss": 1.5914, "step": 748 }, { "epoch": 0.23590551181102362, "grad_norm": 40.44503402709961, "learning_rate": 9.998826771653543e-06, "loss": 0.6454, "step": 749 }, { "epoch": 0.23622047244094488, "grad_norm": 16.39349365234375, "learning_rate": 9.998825196850394e-06, "loss": 0.2515, "step": 750 }, { "epoch": 0.23653543307086614, "grad_norm": 54.119468688964844, "learning_rate": 9.998823622047245e-06, "loss": 0.497, "step": 751 }, { "epoch": 0.2368503937007874, "grad_norm": 55.42496871948242, "learning_rate": 9.998822047244096e-06, "loss": 0.2491, "step": 752 }, { "epoch": 0.23716535433070865, "grad_norm": 55.21287155151367, "learning_rate": 9.998820472440945e-06, "loss": 0.9017, "step": 753 }, { "epoch": 0.23748031496062993, "grad_norm": 26.37238311767578, "learning_rate": 9.998818897637796e-06, "loss": 0.4861, "step": 754 }, { "epoch": 0.2377952755905512, "grad_norm": 23.099788665771484, "learning_rate": 9.998817322834646e-06, "loss": 0.4609, "step": 755 }, { "epoch": 0.23811023622047245, "grad_norm": 35.12017822265625, "learning_rate": 9.998815748031497e-06, "loss": 0.4897, "step": 756 }, { "epoch": 0.2384251968503937, "grad_norm": 13.813502311706543, "learning_rate": 9.998814173228348e-06, "loss": 0.1271, "step": 757 }, { "epoch": 0.23874015748031496, "grad_norm": 29.46474266052246, "learning_rate": 9.998812598425197e-06, "loss": 0.505, "step": 758 }, { "epoch": 0.23905511811023622, "grad_norm": 26.503684997558594, "learning_rate": 9.998811023622048e-06, "loss": 0.424, "step": 759 }, { "epoch": 0.23937007874015748, "grad_norm": 36.68299102783203, "learning_rate": 9.998809448818897e-06, "loss": 0.5153, "step": 760 }, { "epoch": 0.23937007874015748, "eval_loss": 0.467477947473526, "eval_runtime": 299.9771, "eval_samples_per_second": 0.39, "eval_steps_per_second": 0.39, "step": 760 }, { "epoch": 0.23968503937007873, "grad_norm": 27.404645919799805, "learning_rate": 9.998807874015748e-06, "loss": 0.4891, "step": 761 }, { "epoch": 0.24, "grad_norm": 35.5542106628418, "learning_rate": 9.9988062992126e-06, "loss": 0.6146, "step": 762 }, { "epoch": 0.24031496062992125, "grad_norm": 64.38270568847656, "learning_rate": 9.99880472440945e-06, "loss": 0.7001, "step": 763 }, { "epoch": 0.24062992125984253, "grad_norm": 11.979158401489258, "learning_rate": 9.9988031496063e-06, "loss": 0.1519, "step": 764 }, { "epoch": 0.2409448818897638, "grad_norm": 45.081756591796875, "learning_rate": 9.99880157480315e-06, "loss": 0.5738, "step": 765 }, { "epoch": 0.24125984251968505, "grad_norm": 16.82332992553711, "learning_rate": 9.998800000000002e-06, "loss": 0.2367, "step": 766 }, { "epoch": 0.2415748031496063, "grad_norm": 58.26005554199219, "learning_rate": 9.998798425196853e-06, "loss": 0.6255, "step": 767 }, { "epoch": 0.24188976377952756, "grad_norm": 33.454200744628906, "learning_rate": 9.998796850393702e-06, "loss": 0.282, "step": 768 }, { "epoch": 0.24220472440944882, "grad_norm": 26.192066192626953, "learning_rate": 9.998795275590551e-06, "loss": 0.2512, "step": 769 }, { "epoch": 0.24251968503937008, "grad_norm": 24.451866149902344, "learning_rate": 9.998793700787402e-06, "loss": 0.1288, "step": 770 }, { "epoch": 0.24283464566929133, "grad_norm": 17.908388137817383, "learning_rate": 9.998792125984253e-06, "loss": 0.1228, "step": 771 }, { "epoch": 0.2431496062992126, "grad_norm": 30.55709457397461, "learning_rate": 9.998790551181104e-06, "loss": 0.4937, "step": 772 }, { "epoch": 0.24346456692913385, "grad_norm": 47.04617691040039, "learning_rate": 9.998788976377953e-06, "loss": 0.7825, "step": 773 }, { "epoch": 0.24377952755905513, "grad_norm": 66.30603790283203, "learning_rate": 9.998787401574804e-06, "loss": 0.3853, "step": 774 }, { "epoch": 0.2440944881889764, "grad_norm": 28.97539520263672, "learning_rate": 9.998785826771654e-06, "loss": 0.2909, "step": 775 }, { "epoch": 0.24440944881889765, "grad_norm": 71.36101531982422, "learning_rate": 9.998784251968505e-06, "loss": 0.6988, "step": 776 }, { "epoch": 0.2447244094488189, "grad_norm": 35.699249267578125, "learning_rate": 9.998782677165356e-06, "loss": 0.4579, "step": 777 }, { "epoch": 0.24503937007874016, "grad_norm": 4.760364532470703, "learning_rate": 9.998781102362205e-06, "loss": 0.0287, "step": 778 }, { "epoch": 0.24535433070866142, "grad_norm": 30.33928108215332, "learning_rate": 9.998779527559056e-06, "loss": 0.5643, "step": 779 }, { "epoch": 0.24566929133858267, "grad_norm": 43.770694732666016, "learning_rate": 9.998777952755905e-06, "loss": 0.4347, "step": 780 }, { "epoch": 0.24566929133858267, "eval_loss": 0.5863191485404968, "eval_runtime": 301.5247, "eval_samples_per_second": 0.388, "eval_steps_per_second": 0.388, "step": 780 }, { "epoch": 0.24598425196850393, "grad_norm": 38.59556198120117, "learning_rate": 9.998776377952756e-06, "loss": 0.2601, "step": 781 }, { "epoch": 0.2462992125984252, "grad_norm": 42.7963981628418, "learning_rate": 9.998774803149607e-06, "loss": 0.329, "step": 782 }, { "epoch": 0.24661417322834644, "grad_norm": 48.59577941894531, "learning_rate": 9.998773228346458e-06, "loss": 0.5232, "step": 783 }, { "epoch": 0.24692913385826773, "grad_norm": 51.12166976928711, "learning_rate": 9.998771653543308e-06, "loss": 0.955, "step": 784 }, { "epoch": 0.247244094488189, "grad_norm": 29.71060562133789, "learning_rate": 9.998770078740159e-06, "loss": 0.1399, "step": 785 }, { "epoch": 0.24755905511811024, "grad_norm": 73.006103515625, "learning_rate": 9.998768503937008e-06, "loss": 0.8551, "step": 786 }, { "epoch": 0.2478740157480315, "grad_norm": 24.96092414855957, "learning_rate": 9.998766929133859e-06, "loss": 0.379, "step": 787 }, { "epoch": 0.24818897637795276, "grad_norm": 23.48893165588379, "learning_rate": 9.99876535433071e-06, "loss": 0.1811, "step": 788 }, { "epoch": 0.24850393700787402, "grad_norm": 38.28635025024414, "learning_rate": 9.998763779527559e-06, "loss": 0.4841, "step": 789 }, { "epoch": 0.24881889763779527, "grad_norm": 53.374549865722656, "learning_rate": 9.99876220472441e-06, "loss": 0.5892, "step": 790 }, { "epoch": 0.24913385826771653, "grad_norm": 32.456485748291016, "learning_rate": 9.998760629921261e-06, "loss": 0.4001, "step": 791 }, { "epoch": 0.2494488188976378, "grad_norm": 75.20841217041016, "learning_rate": 9.998759055118112e-06, "loss": 0.5947, "step": 792 }, { "epoch": 0.24976377952755904, "grad_norm": 43.24211502075195, "learning_rate": 9.998757480314961e-06, "loss": 0.4858, "step": 793 }, { "epoch": 0.25007874015748033, "grad_norm": 50.684852600097656, "learning_rate": 9.998755905511812e-06, "loss": 0.3602, "step": 794 }, { "epoch": 0.2503937007874016, "grad_norm": 24.13330078125, "learning_rate": 9.998754330708662e-06, "loss": 0.3898, "step": 795 }, { "epoch": 0.25070866141732284, "grad_norm": 32.628719329833984, "learning_rate": 9.998752755905513e-06, "loss": 0.547, "step": 796 }, { "epoch": 0.2510236220472441, "grad_norm": 24.97728157043457, "learning_rate": 9.998751181102364e-06, "loss": 0.536, "step": 797 }, { "epoch": 0.25133858267716536, "grad_norm": 21.705215454101562, "learning_rate": 9.998749606299213e-06, "loss": 0.315, "step": 798 }, { "epoch": 0.2516535433070866, "grad_norm": 27.06643295288086, "learning_rate": 9.998748031496064e-06, "loss": 0.4722, "step": 799 }, { "epoch": 0.25196850393700787, "grad_norm": 23.71688461303711, "learning_rate": 9.998746456692913e-06, "loss": 0.5445, "step": 800 }, { "epoch": 0.25196850393700787, "eval_loss": 0.51589435338974, "eval_runtime": 316.4111, "eval_samples_per_second": 0.37, "eval_steps_per_second": 0.37, "step": 800 }, { "epoch": 0.2522834645669291, "grad_norm": 35.45967483520508, "learning_rate": 9.998744881889764e-06, "loss": 0.5146, "step": 801 }, { "epoch": 0.2525984251968504, "grad_norm": 24.311609268188477, "learning_rate": 9.998743307086615e-06, "loss": 0.596, "step": 802 }, { "epoch": 0.25291338582677164, "grad_norm": 37.100257873535156, "learning_rate": 9.998741732283466e-06, "loss": 0.4307, "step": 803 }, { "epoch": 0.2532283464566929, "grad_norm": 21.675411224365234, "learning_rate": 9.998740157480315e-06, "loss": 0.3512, "step": 804 }, { "epoch": 0.25354330708661416, "grad_norm": 21.883447647094727, "learning_rate": 9.998738582677166e-06, "loss": 0.4789, "step": 805 }, { "epoch": 0.2538582677165354, "grad_norm": 26.242074966430664, "learning_rate": 9.998737007874016e-06, "loss": 0.6633, "step": 806 }, { "epoch": 0.25417322834645667, "grad_norm": 36.36134338378906, "learning_rate": 9.998735433070867e-06, "loss": 0.2947, "step": 807 }, { "epoch": 0.2544881889763779, "grad_norm": 58.036354064941406, "learning_rate": 9.998733858267718e-06, "loss": 0.5375, "step": 808 }, { "epoch": 0.25480314960629924, "grad_norm": 77.08882141113281, "learning_rate": 9.998732283464567e-06, "loss": 0.2146, "step": 809 }, { "epoch": 0.2551181102362205, "grad_norm": 25.003931045532227, "learning_rate": 9.998730708661418e-06, "loss": 0.3593, "step": 810 }, { "epoch": 0.25543307086614175, "grad_norm": 35.63140869140625, "learning_rate": 9.998729133858267e-06, "loss": 0.3855, "step": 811 }, { "epoch": 0.255748031496063, "grad_norm": 13.581232070922852, "learning_rate": 9.998727559055118e-06, "loss": 0.1482, "step": 812 }, { "epoch": 0.25606299212598427, "grad_norm": 46.440670013427734, "learning_rate": 9.99872598425197e-06, "loss": 0.5954, "step": 813 }, { "epoch": 0.2563779527559055, "grad_norm": 24.534271240234375, "learning_rate": 9.99872440944882e-06, "loss": 0.2221, "step": 814 }, { "epoch": 0.2566929133858268, "grad_norm": 28.46855926513672, "learning_rate": 9.99872283464567e-06, "loss": 0.1972, "step": 815 }, { "epoch": 0.25700787401574804, "grad_norm": 27.41106605529785, "learning_rate": 9.99872125984252e-06, "loss": 0.2933, "step": 816 }, { "epoch": 0.2573228346456693, "grad_norm": 77.73954772949219, "learning_rate": 9.998719685039372e-06, "loss": 0.5044, "step": 817 }, { "epoch": 0.25763779527559055, "grad_norm": 48.21875, "learning_rate": 9.998718110236221e-06, "loss": 0.5613, "step": 818 }, { "epoch": 0.2579527559055118, "grad_norm": 33.06459045410156, "learning_rate": 9.998716535433072e-06, "loss": 0.4511, "step": 819 }, { "epoch": 0.25826771653543307, "grad_norm": 91.82710266113281, "learning_rate": 9.998714960629921e-06, "loss": 1.3236, "step": 820 }, { "epoch": 0.25826771653543307, "eval_loss": 0.4647013545036316, "eval_runtime": 297.9126, "eval_samples_per_second": 0.393, "eval_steps_per_second": 0.393, "step": 820 }, { "epoch": 0.2585826771653543, "grad_norm": 80.5277328491211, "learning_rate": 9.998713385826772e-06, "loss": 0.7502, "step": 821 }, { "epoch": 0.2588976377952756, "grad_norm": 200.29324340820312, "learning_rate": 9.998711811023623e-06, "loss": 1.2537, "step": 822 }, { "epoch": 0.25921259842519684, "grad_norm": 22.48586654663086, "learning_rate": 9.998710236220474e-06, "loss": 0.1774, "step": 823 }, { "epoch": 0.2595275590551181, "grad_norm": 43.362831115722656, "learning_rate": 9.998708661417323e-06, "loss": 0.336, "step": 824 }, { "epoch": 0.25984251968503935, "grad_norm": 27.631332397460938, "learning_rate": 9.998707086614174e-06, "loss": 0.2185, "step": 825 }, { "epoch": 0.2601574803149606, "grad_norm": 44.31364440917969, "learning_rate": 9.998705511811024e-06, "loss": 0.3823, "step": 826 }, { "epoch": 0.26047244094488187, "grad_norm": 78.68717193603516, "learning_rate": 9.998703937007875e-06, "loss": 0.5907, "step": 827 }, { "epoch": 0.2607874015748031, "grad_norm": 86.36324310302734, "learning_rate": 9.998702362204726e-06, "loss": 0.7028, "step": 828 }, { "epoch": 0.26110236220472444, "grad_norm": 108.75439453125, "learning_rate": 9.998700787401575e-06, "loss": 0.9105, "step": 829 }, { "epoch": 0.2614173228346457, "grad_norm": 33.248592376708984, "learning_rate": 9.998699212598426e-06, "loss": 0.2809, "step": 830 }, { "epoch": 0.26173228346456695, "grad_norm": 61.382259368896484, "learning_rate": 9.998697637795275e-06, "loss": 0.5749, "step": 831 }, { "epoch": 0.2620472440944882, "grad_norm": 34.68425750732422, "learning_rate": 9.998696062992126e-06, "loss": 0.5513, "step": 832 }, { "epoch": 0.26236220472440946, "grad_norm": 25.614322662353516, "learning_rate": 9.998694488188977e-06, "loss": 0.0668, "step": 833 }, { "epoch": 0.2626771653543307, "grad_norm": 42.22956848144531, "learning_rate": 9.998692913385828e-06, "loss": 0.3998, "step": 834 }, { "epoch": 0.262992125984252, "grad_norm": 34.24924087524414, "learning_rate": 9.998691338582678e-06, "loss": 0.2085, "step": 835 }, { "epoch": 0.26330708661417324, "grad_norm": 72.84844970703125, "learning_rate": 9.998689763779529e-06, "loss": 0.7099, "step": 836 }, { "epoch": 0.2636220472440945, "grad_norm": 16.073625564575195, "learning_rate": 9.99868818897638e-06, "loss": 0.1643, "step": 837 }, { "epoch": 0.26393700787401575, "grad_norm": 12.518115997314453, "learning_rate": 9.998686614173229e-06, "loss": 0.1489, "step": 838 }, { "epoch": 0.264251968503937, "grad_norm": 19.30543327331543, "learning_rate": 9.99868503937008e-06, "loss": 0.2271, "step": 839 }, { "epoch": 0.26456692913385826, "grad_norm": 80.19608306884766, "learning_rate": 9.998683464566929e-06, "loss": 0.3996, "step": 840 }, { "epoch": 0.26456692913385826, "eval_loss": 0.46735909581184387, "eval_runtime": 308.0881, "eval_samples_per_second": 0.38, "eval_steps_per_second": 0.38, "step": 840 }, { "epoch": 0.2648818897637795, "grad_norm": 47.69651412963867, "learning_rate": 9.99868188976378e-06, "loss": 0.4229, "step": 841 }, { "epoch": 0.2651968503937008, "grad_norm": 40.88669204711914, "learning_rate": 9.998680314960631e-06, "loss": 0.594, "step": 842 }, { "epoch": 0.26551181102362204, "grad_norm": 44.67588806152344, "learning_rate": 9.998678740157482e-06, "loss": 0.1496, "step": 843 }, { "epoch": 0.2658267716535433, "grad_norm": 24.644241333007812, "learning_rate": 9.998677165354331e-06, "loss": 0.3401, "step": 844 }, { "epoch": 0.26614173228346455, "grad_norm": 18.624984741210938, "learning_rate": 9.998675590551182e-06, "loss": 0.2221, "step": 845 }, { "epoch": 0.2664566929133858, "grad_norm": 17.709203720092773, "learning_rate": 9.998674015748032e-06, "loss": 0.1178, "step": 846 }, { "epoch": 0.26677165354330706, "grad_norm": 24.67478370666504, "learning_rate": 9.998672440944883e-06, "loss": 0.1389, "step": 847 }, { "epoch": 0.2670866141732283, "grad_norm": 31.281604766845703, "learning_rate": 9.998670866141734e-06, "loss": 0.2201, "step": 848 }, { "epoch": 0.26740157480314963, "grad_norm": 37.749542236328125, "learning_rate": 9.998669291338583e-06, "loss": 0.2498, "step": 849 }, { "epoch": 0.2677165354330709, "grad_norm": 77.09476470947266, "learning_rate": 9.998667716535434e-06, "loss": 1.0283, "step": 850 }, { "epoch": 0.26803149606299215, "grad_norm": 63.68134307861328, "learning_rate": 9.998666141732283e-06, "loss": 0.4823, "step": 851 }, { "epoch": 0.2683464566929134, "grad_norm": 78.47075653076172, "learning_rate": 9.998664566929134e-06, "loss": 0.42, "step": 852 }, { "epoch": 0.26866141732283466, "grad_norm": 116.93570709228516, "learning_rate": 9.998662992125985e-06, "loss": 0.2312, "step": 853 }, { "epoch": 0.2689763779527559, "grad_norm": 20.844566345214844, "learning_rate": 9.998661417322836e-06, "loss": 0.1281, "step": 854 }, { "epoch": 0.2692913385826772, "grad_norm": 26.171772003173828, "learning_rate": 9.998659842519686e-06, "loss": 0.2235, "step": 855 }, { "epoch": 0.26960629921259843, "grad_norm": 63.36984634399414, "learning_rate": 9.998658267716537e-06, "loss": 0.6844, "step": 856 }, { "epoch": 0.2699212598425197, "grad_norm": 76.0230941772461, "learning_rate": 9.998656692913386e-06, "loss": 0.5884, "step": 857 }, { "epoch": 0.27023622047244095, "grad_norm": 57.21022033691406, "learning_rate": 9.998655118110237e-06, "loss": 0.6529, "step": 858 }, { "epoch": 0.2705511811023622, "grad_norm": 17.419769287109375, "learning_rate": 9.998653543307088e-06, "loss": 0.0747, "step": 859 }, { "epoch": 0.27086614173228346, "grad_norm": 87.31539154052734, "learning_rate": 9.998651968503937e-06, "loss": 0.6404, "step": 860 }, { "epoch": 0.27086614173228346, "eval_loss": 0.5078207850456238, "eval_runtime": 304.5599, "eval_samples_per_second": 0.384, "eval_steps_per_second": 0.384, "step": 860 }, { "epoch": 0.2711811023622047, "grad_norm": 39.05491638183594, "learning_rate": 9.998650393700788e-06, "loss": 0.3182, "step": 861 }, { "epoch": 0.271496062992126, "grad_norm": 43.85835647583008, "learning_rate": 9.998648818897639e-06, "loss": 0.2367, "step": 862 }, { "epoch": 0.27181102362204723, "grad_norm": 61.60994338989258, "learning_rate": 9.99864724409449e-06, "loss": 0.7357, "step": 863 }, { "epoch": 0.2721259842519685, "grad_norm": 61.263484954833984, "learning_rate": 9.99864566929134e-06, "loss": 0.1954, "step": 864 }, { "epoch": 0.27244094488188975, "grad_norm": 41.59515380859375, "learning_rate": 9.99864409448819e-06, "loss": 0.4512, "step": 865 }, { "epoch": 0.272755905511811, "grad_norm": 24.524341583251953, "learning_rate": 9.99864251968504e-06, "loss": 0.1307, "step": 866 }, { "epoch": 0.27307086614173226, "grad_norm": 58.72618865966797, "learning_rate": 9.99864094488189e-06, "loss": 0.707, "step": 867 }, { "epoch": 0.2733858267716535, "grad_norm": 46.32933044433594, "learning_rate": 9.998639370078742e-06, "loss": 0.218, "step": 868 }, { "epoch": 0.27370078740157483, "grad_norm": 36.803565979003906, "learning_rate": 9.998637795275591e-06, "loss": 0.1957, "step": 869 }, { "epoch": 0.2740157480314961, "grad_norm": 29.207927703857422, "learning_rate": 9.998636220472442e-06, "loss": 0.3488, "step": 870 }, { "epoch": 0.27433070866141734, "grad_norm": 26.461669921875, "learning_rate": 9.998634645669291e-06, "loss": 0.1342, "step": 871 }, { "epoch": 0.2746456692913386, "grad_norm": 98.34436798095703, "learning_rate": 9.998633070866142e-06, "loss": 0.2521, "step": 872 }, { "epoch": 0.27496062992125986, "grad_norm": 40.949153900146484, "learning_rate": 9.998631496062993e-06, "loss": 0.3699, "step": 873 }, { "epoch": 0.2752755905511811, "grad_norm": 26.012012481689453, "learning_rate": 9.998629921259844e-06, "loss": 0.2719, "step": 874 }, { "epoch": 0.2755905511811024, "grad_norm": 28.45779800415039, "learning_rate": 9.998628346456693e-06, "loss": 0.1321, "step": 875 }, { "epoch": 0.27590551181102363, "grad_norm": 63.54460144042969, "learning_rate": 9.998626771653544e-06, "loss": 0.5929, "step": 876 }, { "epoch": 0.2762204724409449, "grad_norm": 101.21806335449219, "learning_rate": 9.998625196850394e-06, "loss": 0.6633, "step": 877 }, { "epoch": 0.27653543307086614, "grad_norm": 8.460577964782715, "learning_rate": 9.998623622047245e-06, "loss": 0.0187, "step": 878 }, { "epoch": 0.2768503937007874, "grad_norm": 101.49215698242188, "learning_rate": 9.998622047244096e-06, "loss": 0.7255, "step": 879 }, { "epoch": 0.27716535433070866, "grad_norm": 110.3086929321289, "learning_rate": 9.998620472440945e-06, "loss": 0.9578, "step": 880 }, { "epoch": 0.27716535433070866, "eval_loss": 0.49943187832832336, "eval_runtime": 306.9283, "eval_samples_per_second": 0.381, "eval_steps_per_second": 0.381, "step": 880 }, { "epoch": 0.2774803149606299, "grad_norm": 6.745794773101807, "learning_rate": 9.998618897637796e-06, "loss": 0.0379, "step": 881 }, { "epoch": 0.27779527559055117, "grad_norm": 51.96881866455078, "learning_rate": 9.998617322834645e-06, "loss": 0.7563, "step": 882 }, { "epoch": 0.27811023622047243, "grad_norm": 52.106834411621094, "learning_rate": 9.998615748031496e-06, "loss": 0.4385, "step": 883 }, { "epoch": 0.2784251968503937, "grad_norm": 69.72996520996094, "learning_rate": 9.998614173228347e-06, "loss": 0.8885, "step": 884 }, { "epoch": 0.27874015748031494, "grad_norm": 67.1977310180664, "learning_rate": 9.998612598425198e-06, "loss": 0.814, "step": 885 }, { "epoch": 0.2790551181102362, "grad_norm": 111.88276672363281, "learning_rate": 9.998611023622048e-06, "loss": 0.6336, "step": 886 }, { "epoch": 0.27937007874015746, "grad_norm": 40.74708557128906, "learning_rate": 9.998609448818899e-06, "loss": 0.2043, "step": 887 }, { "epoch": 0.2796850393700787, "grad_norm": 42.77908706665039, "learning_rate": 9.99860787401575e-06, "loss": 0.2083, "step": 888 }, { "epoch": 0.28, "grad_norm": 57.296024322509766, "learning_rate": 9.998606299212599e-06, "loss": 0.1104, "step": 889 }, { "epoch": 0.2803149606299213, "grad_norm": 32.91524124145508, "learning_rate": 9.99860472440945e-06, "loss": 0.498, "step": 890 }, { "epoch": 0.28062992125984254, "grad_norm": 71.90412139892578, "learning_rate": 9.9986031496063e-06, "loss": 0.8651, "step": 891 }, { "epoch": 0.2809448818897638, "grad_norm": 11.306217193603516, "learning_rate": 9.99860157480315e-06, "loss": 0.0642, "step": 892 }, { "epoch": 0.28125984251968505, "grad_norm": 15.339298248291016, "learning_rate": 9.998600000000001e-06, "loss": 0.0829, "step": 893 }, { "epoch": 0.2815748031496063, "grad_norm": 59.414466857910156, "learning_rate": 9.998598425196852e-06, "loss": 0.5543, "step": 894 }, { "epoch": 0.28188976377952757, "grad_norm": 66.70774841308594, "learning_rate": 9.998596850393701e-06, "loss": 0.3054, "step": 895 }, { "epoch": 0.2822047244094488, "grad_norm": 20.609098434448242, "learning_rate": 9.998595275590552e-06, "loss": 0.1729, "step": 896 }, { "epoch": 0.2825196850393701, "grad_norm": 7.707085609436035, "learning_rate": 9.998593700787402e-06, "loss": 0.0473, "step": 897 }, { "epoch": 0.28283464566929134, "grad_norm": 41.48631286621094, "learning_rate": 9.998592125984253e-06, "loss": 0.4578, "step": 898 }, { "epoch": 0.2831496062992126, "grad_norm": 23.614355087280273, "learning_rate": 9.998590551181104e-06, "loss": 0.193, "step": 899 }, { "epoch": 0.28346456692913385, "grad_norm": 15.626468658447266, "learning_rate": 9.998588976377953e-06, "loss": 0.0845, "step": 900 }, { "epoch": 0.28346456692913385, "eval_loss": 0.5890966653823853, "eval_runtime": 302.1316, "eval_samples_per_second": 0.387, "eval_steps_per_second": 0.387, "step": 900 }, { "epoch": 0.2837795275590551, "grad_norm": 60.57646942138672, "learning_rate": 9.998587401574804e-06, "loss": 0.3963, "step": 901 }, { "epoch": 0.28409448818897637, "grad_norm": 47.83597946166992, "learning_rate": 9.998585826771653e-06, "loss": 0.6909, "step": 902 }, { "epoch": 0.2844094488188976, "grad_norm": 26.267818450927734, "learning_rate": 9.998584251968504e-06, "loss": 0.0956, "step": 903 }, { "epoch": 0.2847244094488189, "grad_norm": 63.959110260009766, "learning_rate": 9.998582677165355e-06, "loss": 0.4272, "step": 904 }, { "epoch": 0.28503937007874014, "grad_norm": 59.13768005371094, "learning_rate": 9.998581102362206e-06, "loss": 0.3178, "step": 905 }, { "epoch": 0.2853543307086614, "grad_norm": 39.657814025878906, "learning_rate": 9.998579527559056e-06, "loss": 0.6433, "step": 906 }, { "epoch": 0.28566929133858265, "grad_norm": 90.1864013671875, "learning_rate": 9.998577952755907e-06, "loss": 0.6963, "step": 907 }, { "epoch": 0.2859842519685039, "grad_norm": 121.48163604736328, "learning_rate": 9.998576377952758e-06, "loss": 1.8743, "step": 908 }, { "epoch": 0.2862992125984252, "grad_norm": 37.84361267089844, "learning_rate": 9.998574803149607e-06, "loss": 0.3433, "step": 909 }, { "epoch": 0.2866141732283465, "grad_norm": 42.82717514038086, "learning_rate": 9.998573228346458e-06, "loss": 0.3181, "step": 910 }, { "epoch": 0.28692913385826774, "grad_norm": 55.58892822265625, "learning_rate": 9.998571653543307e-06, "loss": 1.2413, "step": 911 }, { "epoch": 0.287244094488189, "grad_norm": 56.79817581176758, "learning_rate": 9.998570078740158e-06, "loss": 0.3707, "step": 912 }, { "epoch": 0.28755905511811025, "grad_norm": 54.66569900512695, "learning_rate": 9.998568503937009e-06, "loss": 0.7417, "step": 913 }, { "epoch": 0.2878740157480315, "grad_norm": 22.454544067382812, "learning_rate": 9.99856692913386e-06, "loss": 0.1303, "step": 914 }, { "epoch": 0.28818897637795277, "grad_norm": 72.16681671142578, "learning_rate": 9.99856535433071e-06, "loss": 0.4007, "step": 915 }, { "epoch": 0.288503937007874, "grad_norm": 52.15703201293945, "learning_rate": 9.99856377952756e-06, "loss": 0.7637, "step": 916 }, { "epoch": 0.2888188976377953, "grad_norm": 8.107488632202148, "learning_rate": 9.99856220472441e-06, "loss": 0.0643, "step": 917 }, { "epoch": 0.28913385826771654, "grad_norm": 47.568267822265625, "learning_rate": 9.99856062992126e-06, "loss": 0.4449, "step": 918 }, { "epoch": 0.2894488188976378, "grad_norm": 21.596525192260742, "learning_rate": 9.998559055118112e-06, "loss": 0.1228, "step": 919 }, { "epoch": 0.28976377952755905, "grad_norm": 40.52389144897461, "learning_rate": 9.998557480314961e-06, "loss": 0.446, "step": 920 }, { "epoch": 0.28976377952755905, "eval_loss": 0.5098508596420288, "eval_runtime": 307.5399, "eval_samples_per_second": 0.38, "eval_steps_per_second": 0.38, "step": 920 }, { "epoch": 0.2900787401574803, "grad_norm": 43.565303802490234, "learning_rate": 9.998555905511812e-06, "loss": 0.5184, "step": 921 }, { "epoch": 0.29039370078740157, "grad_norm": 54.92490768432617, "learning_rate": 9.998554330708661e-06, "loss": 0.7516, "step": 922 }, { "epoch": 0.2907086614173228, "grad_norm": 47.38011169433594, "learning_rate": 9.998552755905512e-06, "loss": 0.4989, "step": 923 }, { "epoch": 0.2910236220472441, "grad_norm": 63.52509689331055, "learning_rate": 9.998551181102363e-06, "loss": 0.4689, "step": 924 }, { "epoch": 0.29133858267716534, "grad_norm": 38.14700698852539, "learning_rate": 9.998549606299214e-06, "loss": 0.5137, "step": 925 }, { "epoch": 0.2916535433070866, "grad_norm": 63.95713806152344, "learning_rate": 9.998548031496063e-06, "loss": 0.3511, "step": 926 }, { "epoch": 0.29196850393700785, "grad_norm": 38.79820251464844, "learning_rate": 9.998546456692914e-06, "loss": 0.3497, "step": 927 }, { "epoch": 0.2922834645669291, "grad_norm": 76.02424621582031, "learning_rate": 9.998544881889764e-06, "loss": 0.416, "step": 928 }, { "epoch": 0.2925984251968504, "grad_norm": 45.44684982299805, "learning_rate": 9.998543307086615e-06, "loss": 0.4091, "step": 929 }, { "epoch": 0.2929133858267717, "grad_norm": 32.81157684326172, "learning_rate": 9.998541732283466e-06, "loss": 0.4614, "step": 930 }, { "epoch": 0.29322834645669293, "grad_norm": 45.81043243408203, "learning_rate": 9.998540157480315e-06, "loss": 0.4832, "step": 931 }, { "epoch": 0.2935433070866142, "grad_norm": 42.5070915222168, "learning_rate": 9.998538582677166e-06, "loss": 0.352, "step": 932 }, { "epoch": 0.29385826771653545, "grad_norm": 40.78940200805664, "learning_rate": 9.998537007874017e-06, "loss": 0.9412, "step": 933 }, { "epoch": 0.2941732283464567, "grad_norm": 44.96437454223633, "learning_rate": 9.998535433070868e-06, "loss": 0.9758, "step": 934 }, { "epoch": 0.29448818897637796, "grad_norm": 8.247536659240723, "learning_rate": 9.998533858267717e-06, "loss": 0.0863, "step": 935 }, { "epoch": 0.2948031496062992, "grad_norm": 72.65038299560547, "learning_rate": 9.998532283464568e-06, "loss": 0.6317, "step": 936 }, { "epoch": 0.2951181102362205, "grad_norm": 28.256349563598633, "learning_rate": 9.998530708661418e-06, "loss": 0.3451, "step": 937 }, { "epoch": 0.29543307086614173, "grad_norm": 17.122854232788086, "learning_rate": 9.998529133858269e-06, "loss": 0.2172, "step": 938 }, { "epoch": 0.295748031496063, "grad_norm": 25.158649444580078, "learning_rate": 9.99852755905512e-06, "loss": 0.3431, "step": 939 }, { "epoch": 0.29606299212598425, "grad_norm": 48.18063735961914, "learning_rate": 9.998525984251969e-06, "loss": 0.6472, "step": 940 }, { "epoch": 0.29606299212598425, "eval_loss": 0.4980691373348236, "eval_runtime": 301.7364, "eval_samples_per_second": 0.388, "eval_steps_per_second": 0.388, "step": 940 }, { "epoch": 0.2963779527559055, "grad_norm": 93.8964614868164, "learning_rate": 9.99852440944882e-06, "loss": 0.8136, "step": 941 }, { "epoch": 0.29669291338582676, "grad_norm": 19.963516235351562, "learning_rate": 9.99852283464567e-06, "loss": 0.194, "step": 942 }, { "epoch": 0.297007874015748, "grad_norm": 20.99871826171875, "learning_rate": 9.99852125984252e-06, "loss": 0.294, "step": 943 }, { "epoch": 0.2973228346456693, "grad_norm": 43.991703033447266, "learning_rate": 9.998519685039371e-06, "loss": 0.4679, "step": 944 }, { "epoch": 0.29763779527559053, "grad_norm": 48.08595275878906, "learning_rate": 9.998518110236222e-06, "loss": 0.6872, "step": 945 }, { "epoch": 0.2979527559055118, "grad_norm": 14.398991584777832, "learning_rate": 9.998516535433071e-06, "loss": 0.1213, "step": 946 }, { "epoch": 0.29826771653543305, "grad_norm": 60.385154724121094, "learning_rate": 9.998514960629922e-06, "loss": 0.9587, "step": 947 }, { "epoch": 0.2985826771653543, "grad_norm": 23.951004028320312, "learning_rate": 9.998513385826772e-06, "loss": 0.1824, "step": 948 }, { "epoch": 0.2988976377952756, "grad_norm": 24.64979362487793, "learning_rate": 9.998511811023623e-06, "loss": 0.2494, "step": 949 }, { "epoch": 0.2992125984251969, "grad_norm": 33.05177307128906, "learning_rate": 9.998510236220474e-06, "loss": 0.8105, "step": 950 }, { "epoch": 0.29952755905511813, "grad_norm": 59.83413314819336, "learning_rate": 9.998508661417323e-06, "loss": 0.6087, "step": 951 }, { "epoch": 0.2998425196850394, "grad_norm": 79.51978302001953, "learning_rate": 9.998507086614174e-06, "loss": 0.8573, "step": 952 }, { "epoch": 0.30015748031496065, "grad_norm": 59.22591018676758, "learning_rate": 9.998505511811023e-06, "loss": 0.7464, "step": 953 }, { "epoch": 0.3004724409448819, "grad_norm": 53.090614318847656, "learning_rate": 9.998503937007876e-06, "loss": 0.6472, "step": 954 }, { "epoch": 0.30078740157480316, "grad_norm": 35.13800048828125, "learning_rate": 9.998502362204725e-06, "loss": 0.2764, "step": 955 }, { "epoch": 0.3011023622047244, "grad_norm": 18.853023529052734, "learning_rate": 9.998500787401576e-06, "loss": 0.1602, "step": 956 }, { "epoch": 0.3014173228346457, "grad_norm": 13.263671875, "learning_rate": 9.998499212598426e-06, "loss": 0.1145, "step": 957 }, { "epoch": 0.30173228346456693, "grad_norm": 38.00738525390625, "learning_rate": 9.998497637795277e-06, "loss": 0.1779, "step": 958 }, { "epoch": 0.3020472440944882, "grad_norm": 29.51807403564453, "learning_rate": 9.998496062992128e-06, "loss": 0.3998, "step": 959 }, { "epoch": 0.30236220472440944, "grad_norm": 49.63967514038086, "learning_rate": 9.998494488188977e-06, "loss": 1.2491, "step": 960 }, { "epoch": 0.30236220472440944, "eval_loss": 0.505746066570282, "eval_runtime": 303.3875, "eval_samples_per_second": 0.386, "eval_steps_per_second": 0.386, "step": 960 }, { "epoch": 0.3026771653543307, "grad_norm": 22.166236877441406, "learning_rate": 9.998492913385828e-06, "loss": 0.2884, "step": 961 }, { "epoch": 0.30299212598425196, "grad_norm": 9.977280616760254, "learning_rate": 9.998491338582677e-06, "loss": 0.0985, "step": 962 }, { "epoch": 0.3033070866141732, "grad_norm": 51.85643005371094, "learning_rate": 9.998489763779528e-06, "loss": 0.4792, "step": 963 }, { "epoch": 0.3036220472440945, "grad_norm": 11.42037582397461, "learning_rate": 9.998488188976379e-06, "loss": 0.0609, "step": 964 }, { "epoch": 0.30393700787401573, "grad_norm": 29.179784774780273, "learning_rate": 9.99848661417323e-06, "loss": 0.2065, "step": 965 }, { "epoch": 0.304251968503937, "grad_norm": 99.96176147460938, "learning_rate": 9.99848503937008e-06, "loss": 0.1855, "step": 966 }, { "epoch": 0.30456692913385824, "grad_norm": 67.8636245727539, "learning_rate": 9.99848346456693e-06, "loss": 0.4098, "step": 967 }, { "epoch": 0.3048818897637795, "grad_norm": 67.14287567138672, "learning_rate": 9.99848188976378e-06, "loss": 0.8962, "step": 968 }, { "epoch": 0.3051968503937008, "grad_norm": 52.694175720214844, "learning_rate": 9.99848031496063e-06, "loss": 0.3033, "step": 969 }, { "epoch": 0.30551181102362207, "grad_norm": 124.31986236572266, "learning_rate": 9.998478740157482e-06, "loss": 0.6216, "step": 970 }, { "epoch": 0.30582677165354333, "grad_norm": 59.26445770263672, "learning_rate": 9.998477165354331e-06, "loss": 0.4527, "step": 971 }, { "epoch": 0.3061417322834646, "grad_norm": 28.591516494750977, "learning_rate": 9.998475590551182e-06, "loss": 0.1072, "step": 972 }, { "epoch": 0.30645669291338584, "grad_norm": 117.83760833740234, "learning_rate": 9.998474015748031e-06, "loss": 0.622, "step": 973 }, { "epoch": 0.3067716535433071, "grad_norm": 34.50752639770508, "learning_rate": 9.998472440944882e-06, "loss": 0.4129, "step": 974 }, { "epoch": 0.30708661417322836, "grad_norm": 8.890520095825195, "learning_rate": 9.998470866141733e-06, "loss": 0.0409, "step": 975 }, { "epoch": 0.3074015748031496, "grad_norm": 52.015785217285156, "learning_rate": 9.998469291338584e-06, "loss": 0.4855, "step": 976 }, { "epoch": 0.30771653543307087, "grad_norm": 23.1546688079834, "learning_rate": 9.998467716535434e-06, "loss": 0.1683, "step": 977 }, { "epoch": 0.3080314960629921, "grad_norm": 26.71653938293457, "learning_rate": 9.998466141732285e-06, "loss": 0.1988, "step": 978 }, { "epoch": 0.3083464566929134, "grad_norm": 55.756309509277344, "learning_rate": 9.998464566929135e-06, "loss": 0.6107, "step": 979 }, { "epoch": 0.30866141732283464, "grad_norm": 38.31888198852539, "learning_rate": 9.998462992125985e-06, "loss": 0.3716, "step": 980 }, { "epoch": 0.30866141732283464, "eval_loss": 0.4945707321166992, "eval_runtime": 303.1038, "eval_samples_per_second": 0.386, "eval_steps_per_second": 0.386, "step": 980 }, { "epoch": 0.3089763779527559, "grad_norm": 136.65208435058594, "learning_rate": 9.998461417322836e-06, "loss": 1.04, "step": 981 }, { "epoch": 0.30929133858267716, "grad_norm": 70.81889343261719, "learning_rate": 9.998459842519685e-06, "loss": 0.7749, "step": 982 }, { "epoch": 0.3096062992125984, "grad_norm": 81.47750854492188, "learning_rate": 9.998458267716536e-06, "loss": 0.4565, "step": 983 }, { "epoch": 0.30992125984251967, "grad_norm": 84.3187255859375, "learning_rate": 9.998456692913387e-06, "loss": 1.1115, "step": 984 }, { "epoch": 0.3102362204724409, "grad_norm": 121.83627319335938, "learning_rate": 9.998455118110238e-06, "loss": 0.4089, "step": 985 }, { "epoch": 0.3105511811023622, "grad_norm": 46.44225311279297, "learning_rate": 9.998453543307087e-06, "loss": 0.6732, "step": 986 }, { "epoch": 0.31086614173228344, "grad_norm": 60.36497497558594, "learning_rate": 9.998451968503938e-06, "loss": 0.3421, "step": 987 }, { "epoch": 0.3111811023622047, "grad_norm": 34.300819396972656, "learning_rate": 9.998450393700788e-06, "loss": 0.1633, "step": 988 }, { "epoch": 0.311496062992126, "grad_norm": 29.893108367919922, "learning_rate": 9.998448818897639e-06, "loss": 0.2435, "step": 989 }, { "epoch": 0.31181102362204727, "grad_norm": 56.01438522338867, "learning_rate": 9.99844724409449e-06, "loss": 0.2676, "step": 990 }, { "epoch": 0.3121259842519685, "grad_norm": 58.90812301635742, "learning_rate": 9.998445669291339e-06, "loss": 0.362, "step": 991 }, { "epoch": 0.3124409448818898, "grad_norm": 54.015743255615234, "learning_rate": 9.99844409448819e-06, "loss": 0.3922, "step": 992 }, { "epoch": 0.31275590551181104, "grad_norm": 43.66928482055664, "learning_rate": 9.99844251968504e-06, "loss": 0.3873, "step": 993 }, { "epoch": 0.3130708661417323, "grad_norm": 37.3400764465332, "learning_rate": 9.99844094488189e-06, "loss": 0.1752, "step": 994 }, { "epoch": 0.31338582677165355, "grad_norm": 22.93543243408203, "learning_rate": 9.998439370078741e-06, "loss": 0.2166, "step": 995 }, { "epoch": 0.3137007874015748, "grad_norm": 54.434444427490234, "learning_rate": 9.998437795275592e-06, "loss": 0.6486, "step": 996 }, { "epoch": 0.31401574803149607, "grad_norm": 29.223251342773438, "learning_rate": 9.998436220472441e-06, "loss": 0.2945, "step": 997 }, { "epoch": 0.3143307086614173, "grad_norm": 40.362152099609375, "learning_rate": 9.998434645669292e-06, "loss": 0.6047, "step": 998 }, { "epoch": 0.3146456692913386, "grad_norm": 17.493072509765625, "learning_rate": 9.998433070866142e-06, "loss": 0.3557, "step": 999 }, { "epoch": 0.31496062992125984, "grad_norm": 32.33248519897461, "learning_rate": 9.998431496062993e-06, "loss": 0.1763, "step": 1000 }, { "epoch": 0.31496062992125984, "eval_loss": 0.4581963121891022, "eval_runtime": 306.4597, "eval_samples_per_second": 0.382, "eval_steps_per_second": 0.382, "step": 1000 }, { "epoch": 0.3152755905511811, "grad_norm": 41.288719177246094, "learning_rate": 9.998429921259844e-06, "loss": 0.3175, "step": 1001 }, { "epoch": 0.31559055118110235, "grad_norm": 21.213180541992188, "learning_rate": 9.998428346456693e-06, "loss": 0.1175, "step": 1002 }, { "epoch": 0.3159055118110236, "grad_norm": 35.53864669799805, "learning_rate": 9.998426771653544e-06, "loss": 0.2783, "step": 1003 }, { "epoch": 0.31622047244094487, "grad_norm": 46.64375305175781, "learning_rate": 9.998425196850395e-06, "loss": 0.4337, "step": 1004 }, { "epoch": 0.3165354330708661, "grad_norm": 25.14255142211914, "learning_rate": 9.998423622047246e-06, "loss": 0.3534, "step": 1005 }, { "epoch": 0.3168503937007874, "grad_norm": 64.11309051513672, "learning_rate": 9.998422047244095e-06, "loss": 0.1756, "step": 1006 }, { "epoch": 0.31716535433070864, "grad_norm": 54.7104377746582, "learning_rate": 9.998420472440946e-06, "loss": 0.9319, "step": 1007 }, { "epoch": 0.3174803149606299, "grad_norm": 84.07182312011719, "learning_rate": 9.998418897637796e-06, "loss": 0.9422, "step": 1008 }, { "epoch": 0.3177952755905512, "grad_norm": 46.56437683105469, "learning_rate": 9.998417322834647e-06, "loss": 0.2973, "step": 1009 }, { "epoch": 0.31811023622047246, "grad_norm": 39.493682861328125, "learning_rate": 9.998415748031498e-06, "loss": 0.4972, "step": 1010 }, { "epoch": 0.3184251968503937, "grad_norm": 51.46127700805664, "learning_rate": 9.998414173228347e-06, "loss": 0.6015, "step": 1011 }, { "epoch": 0.318740157480315, "grad_norm": 27.754695892333984, "learning_rate": 9.998412598425198e-06, "loss": 0.2744, "step": 1012 }, { "epoch": 0.31905511811023624, "grad_norm": 26.403295516967773, "learning_rate": 9.998411023622047e-06, "loss": 0.2464, "step": 1013 }, { "epoch": 0.3193700787401575, "grad_norm": 66.5213623046875, "learning_rate": 9.998409448818898e-06, "loss": 0.3692, "step": 1014 }, { "epoch": 0.31968503937007875, "grad_norm": 35.4731559753418, "learning_rate": 9.998407874015749e-06, "loss": 0.3412, "step": 1015 }, { "epoch": 0.32, "grad_norm": 23.23394203186035, "learning_rate": 9.9984062992126e-06, "loss": 0.4104, "step": 1016 }, { "epoch": 0.32031496062992126, "grad_norm": 44.30449676513672, "learning_rate": 9.99840472440945e-06, "loss": 0.4564, "step": 1017 }, { "epoch": 0.3206299212598425, "grad_norm": 18.734786987304688, "learning_rate": 9.9984031496063e-06, "loss": 0.2802, "step": 1018 }, { "epoch": 0.3209448818897638, "grad_norm": 40.950653076171875, "learning_rate": 9.99840157480315e-06, "loss": 0.3802, "step": 1019 }, { "epoch": 0.32125984251968503, "grad_norm": 97.77069091796875, "learning_rate": 9.9984e-06, "loss": 0.467, "step": 1020 }, { "epoch": 0.32125984251968503, "eval_loss": 0.4887102544307709, "eval_runtime": 304.7183, "eval_samples_per_second": 0.384, "eval_steps_per_second": 0.384, "step": 1020 }, { "epoch": 0.3215748031496063, "grad_norm": 131.0202178955078, "learning_rate": 9.998398425196852e-06, "loss": 0.9191, "step": 1021 }, { "epoch": 0.32188976377952755, "grad_norm": 72.55668640136719, "learning_rate": 9.998396850393701e-06, "loss": 0.9462, "step": 1022 }, { "epoch": 0.3222047244094488, "grad_norm": 47.58323669433594, "learning_rate": 9.998395275590552e-06, "loss": 0.7648, "step": 1023 }, { "epoch": 0.32251968503937006, "grad_norm": 15.815386772155762, "learning_rate": 9.998393700787403e-06, "loss": 0.1747, "step": 1024 }, { "epoch": 0.3228346456692913, "grad_norm": 23.02329444885254, "learning_rate": 9.998392125984254e-06, "loss": 0.2737, "step": 1025 }, { "epoch": 0.3231496062992126, "grad_norm": 69.93936157226562, "learning_rate": 9.998390551181103e-06, "loss": 0.6328, "step": 1026 }, { "epoch": 0.32346456692913383, "grad_norm": 44.680335998535156, "learning_rate": 9.998388976377954e-06, "loss": 0.2808, "step": 1027 }, { "epoch": 0.3237795275590551, "grad_norm": 61.480506896972656, "learning_rate": 9.998387401574804e-06, "loss": 0.3987, "step": 1028 }, { "epoch": 0.3240944881889764, "grad_norm": 49.2515754699707, "learning_rate": 9.998385826771655e-06, "loss": 0.2713, "step": 1029 }, { "epoch": 0.32440944881889766, "grad_norm": 39.322608947753906, "learning_rate": 9.998384251968506e-06, "loss": 0.5236, "step": 1030 }, { "epoch": 0.3247244094488189, "grad_norm": 50.91118621826172, "learning_rate": 9.998382677165355e-06, "loss": 0.3365, "step": 1031 }, { "epoch": 0.3250393700787402, "grad_norm": 57.56412887573242, "learning_rate": 9.998381102362206e-06, "loss": 0.5609, "step": 1032 }, { "epoch": 0.32535433070866143, "grad_norm": 106.12908935546875, "learning_rate": 9.998379527559055e-06, "loss": 0.6946, "step": 1033 }, { "epoch": 0.3256692913385827, "grad_norm": 28.783655166625977, "learning_rate": 9.998377952755906e-06, "loss": 0.3571, "step": 1034 }, { "epoch": 0.32598425196850395, "grad_norm": 35.49094772338867, "learning_rate": 9.998376377952757e-06, "loss": 0.2201, "step": 1035 }, { "epoch": 0.3262992125984252, "grad_norm": 41.67929458618164, "learning_rate": 9.998374803149608e-06, "loss": 0.2486, "step": 1036 }, { "epoch": 0.32661417322834646, "grad_norm": 55.53398132324219, "learning_rate": 9.998373228346457e-06, "loss": 0.6653, "step": 1037 }, { "epoch": 0.3269291338582677, "grad_norm": 63.74702835083008, "learning_rate": 9.998371653543308e-06, "loss": 0.3952, "step": 1038 }, { "epoch": 0.327244094488189, "grad_norm": 43.730770111083984, "learning_rate": 9.998370078740158e-06, "loss": 0.2713, "step": 1039 }, { "epoch": 0.32755905511811023, "grad_norm": 20.90167808532715, "learning_rate": 9.998368503937009e-06, "loss": 0.3298, "step": 1040 }, { "epoch": 0.32755905511811023, "eval_loss": 0.553530216217041, "eval_runtime": 338.2916, "eval_samples_per_second": 0.346, "eval_steps_per_second": 0.346, "step": 1040 }, { "epoch": 0.3278740157480315, "grad_norm": 53.132686614990234, "learning_rate": 9.99836692913386e-06, "loss": 0.2454, "step": 1041 }, { "epoch": 0.32818897637795275, "grad_norm": 62.30380630493164, "learning_rate": 9.998365354330709e-06, "loss": 0.6391, "step": 1042 }, { "epoch": 0.328503937007874, "grad_norm": 34.83534622192383, "learning_rate": 9.99836377952756e-06, "loss": 0.2441, "step": 1043 }, { "epoch": 0.32881889763779526, "grad_norm": 37.7114372253418, "learning_rate": 9.99836220472441e-06, "loss": 0.1986, "step": 1044 }, { "epoch": 0.3291338582677165, "grad_norm": 26.971843719482422, "learning_rate": 9.99836062992126e-06, "loss": 0.1753, "step": 1045 }, { "epoch": 0.3294488188976378, "grad_norm": 64.88057708740234, "learning_rate": 9.998359055118111e-06, "loss": 0.3672, "step": 1046 }, { "epoch": 0.32976377952755903, "grad_norm": 16.45389747619629, "learning_rate": 9.998357480314962e-06, "loss": 0.0662, "step": 1047 }, { "epoch": 0.3300787401574803, "grad_norm": 53.416934967041016, "learning_rate": 9.998355905511811e-06, "loss": 0.6081, "step": 1048 }, { "epoch": 0.3303937007874016, "grad_norm": 44.44949722290039, "learning_rate": 9.998354330708662e-06, "loss": 0.2946, "step": 1049 }, { "epoch": 0.33070866141732286, "grad_norm": 85.5646743774414, "learning_rate": 9.998352755905513e-06, "loss": 0.7022, "step": 1050 }, { "epoch": 0.3310236220472441, "grad_norm": 158.97926330566406, "learning_rate": 9.998351181102363e-06, "loss": 0.3364, "step": 1051 }, { "epoch": 0.33133858267716537, "grad_norm": 53.93890380859375, "learning_rate": 9.998349606299214e-06, "loss": 0.0856, "step": 1052 }, { "epoch": 0.33165354330708663, "grad_norm": 58.113338470458984, "learning_rate": 9.998348031496063e-06, "loss": 0.4051, "step": 1053 }, { "epoch": 0.3319685039370079, "grad_norm": 113.30328369140625, "learning_rate": 9.998346456692914e-06, "loss": 1.3079, "step": 1054 }, { "epoch": 0.33228346456692914, "grad_norm": 52.5933952331543, "learning_rate": 9.998344881889765e-06, "loss": 0.2001, "step": 1055 }, { "epoch": 0.3325984251968504, "grad_norm": 111.05465698242188, "learning_rate": 9.998343307086616e-06, "loss": 0.8016, "step": 1056 }, { "epoch": 0.33291338582677166, "grad_norm": 80.2440185546875, "learning_rate": 9.998341732283465e-06, "loss": 0.6905, "step": 1057 }, { "epoch": 0.3332283464566929, "grad_norm": 46.98655319213867, "learning_rate": 9.998340157480316e-06, "loss": 0.645, "step": 1058 }, { "epoch": 0.33354330708661417, "grad_norm": 88.55839538574219, "learning_rate": 9.998338582677166e-06, "loss": 0.632, "step": 1059 }, { "epoch": 0.33385826771653543, "grad_norm": 45.15827178955078, "learning_rate": 9.998337007874017e-06, "loss": 0.3994, "step": 1060 }, { "epoch": 0.33385826771653543, "eval_loss": 0.49371591210365295, "eval_runtime": 338.6729, "eval_samples_per_second": 0.345, "eval_steps_per_second": 0.345, "step": 1060 }, { "epoch": 0.3341732283464567, "grad_norm": 89.61175537109375, "learning_rate": 9.998335433070868e-06, "loss": 0.7211, "step": 1061 }, { "epoch": 0.33448818897637794, "grad_norm": 43.96299362182617, "learning_rate": 9.998333858267717e-06, "loss": 0.2502, "step": 1062 }, { "epoch": 0.3348031496062992, "grad_norm": 48.51971435546875, "learning_rate": 9.998332283464568e-06, "loss": 0.2617, "step": 1063 }, { "epoch": 0.33511811023622046, "grad_norm": 56.80341720581055, "learning_rate": 9.998330708661417e-06, "loss": 0.315, "step": 1064 }, { "epoch": 0.3354330708661417, "grad_norm": 24.417354583740234, "learning_rate": 9.998329133858268e-06, "loss": 0.2498, "step": 1065 }, { "epoch": 0.33574803149606297, "grad_norm": 21.887855529785156, "learning_rate": 9.99832755905512e-06, "loss": 0.1483, "step": 1066 }, { "epoch": 0.33606299212598423, "grad_norm": 52.2086181640625, "learning_rate": 9.99832598425197e-06, "loss": 0.3647, "step": 1067 }, { "epoch": 0.3363779527559055, "grad_norm": 50.21038818359375, "learning_rate": 9.99832440944882e-06, "loss": 0.5818, "step": 1068 }, { "epoch": 0.3366929133858268, "grad_norm": 25.27126693725586, "learning_rate": 9.99832283464567e-06, "loss": 0.1273, "step": 1069 }, { "epoch": 0.33700787401574805, "grad_norm": 36.34380340576172, "learning_rate": 9.99832125984252e-06, "loss": 0.3509, "step": 1070 }, { "epoch": 0.3373228346456693, "grad_norm": 35.567527770996094, "learning_rate": 9.99831968503937e-06, "loss": 0.2472, "step": 1071 }, { "epoch": 0.33763779527559057, "grad_norm": 39.77647399902344, "learning_rate": 9.998318110236222e-06, "loss": 0.5746, "step": 1072 }, { "epoch": 0.3379527559055118, "grad_norm": 10.02043628692627, "learning_rate": 9.998316535433071e-06, "loss": 0.0749, "step": 1073 }, { "epoch": 0.3382677165354331, "grad_norm": 31.038677215576172, "learning_rate": 9.998314960629922e-06, "loss": 0.2673, "step": 1074 }, { "epoch": 0.33858267716535434, "grad_norm": 34.27322769165039, "learning_rate": 9.998313385826773e-06, "loss": 0.1064, "step": 1075 }, { "epoch": 0.3388976377952756, "grad_norm": 91.75303649902344, "learning_rate": 9.998311811023624e-06, "loss": 0.7232, "step": 1076 }, { "epoch": 0.33921259842519685, "grad_norm": 68.7430419921875, "learning_rate": 9.998310236220473e-06, "loss": 0.5586, "step": 1077 }, { "epoch": 0.3395275590551181, "grad_norm": 94.78008270263672, "learning_rate": 9.998308661417324e-06, "loss": 0.5504, "step": 1078 }, { "epoch": 0.33984251968503937, "grad_norm": 54.0759162902832, "learning_rate": 9.998307086614174e-06, "loss": 0.1873, "step": 1079 }, { "epoch": 0.3401574803149606, "grad_norm": 65.1077651977539, "learning_rate": 9.998305511811025e-06, "loss": 1.0425, "step": 1080 }, { "epoch": 0.3401574803149606, "eval_loss": 0.6948055028915405, "eval_runtime": 327.3153, "eval_samples_per_second": 0.357, "eval_steps_per_second": 0.357, "step": 1080 }, { "epoch": 0.3404724409448819, "grad_norm": 79.78382873535156, "learning_rate": 9.998303937007876e-06, "loss": 0.6999, "step": 1081 }, { "epoch": 0.34078740157480314, "grad_norm": 97.7957534790039, "learning_rate": 9.998302362204725e-06, "loss": 0.4083, "step": 1082 }, { "epoch": 0.3411023622047244, "grad_norm": 90.36141967773438, "learning_rate": 9.998300787401576e-06, "loss": 0.7225, "step": 1083 }, { "epoch": 0.34141732283464565, "grad_norm": 14.357733726501465, "learning_rate": 9.998299212598425e-06, "loss": 0.0595, "step": 1084 }, { "epoch": 0.3417322834645669, "grad_norm": 47.068233489990234, "learning_rate": 9.998297637795276e-06, "loss": 0.7842, "step": 1085 }, { "epoch": 0.34204724409448817, "grad_norm": 22.276060104370117, "learning_rate": 9.998296062992127e-06, "loss": 0.1631, "step": 1086 }, { "epoch": 0.3423622047244094, "grad_norm": 38.76866149902344, "learning_rate": 9.998294488188978e-06, "loss": 0.2738, "step": 1087 }, { "epoch": 0.3426771653543307, "grad_norm": 45.607505798339844, "learning_rate": 9.998292913385827e-06, "loss": 0.4995, "step": 1088 }, { "epoch": 0.342992125984252, "grad_norm": 34.9421272277832, "learning_rate": 9.998291338582678e-06, "loss": 0.614, "step": 1089 }, { "epoch": 0.34330708661417325, "grad_norm": 36.95371627807617, "learning_rate": 9.998289763779528e-06, "loss": 0.6949, "step": 1090 }, { "epoch": 0.3436220472440945, "grad_norm": 32.992279052734375, "learning_rate": 9.998288188976379e-06, "loss": 0.2076, "step": 1091 }, { "epoch": 0.34393700787401577, "grad_norm": 15.995903015136719, "learning_rate": 9.99828661417323e-06, "loss": 0.1538, "step": 1092 }, { "epoch": 0.344251968503937, "grad_norm": 38.209495544433594, "learning_rate": 9.998285039370079e-06, "loss": 0.5808, "step": 1093 }, { "epoch": 0.3445669291338583, "grad_norm": 52.266441345214844, "learning_rate": 9.99828346456693e-06, "loss": 0.3555, "step": 1094 }, { "epoch": 0.34488188976377954, "grad_norm": 11.709747314453125, "learning_rate": 9.998281889763781e-06, "loss": 0.0999, "step": 1095 }, { "epoch": 0.3451968503937008, "grad_norm": 37.43159484863281, "learning_rate": 9.998280314960632e-06, "loss": 0.5477, "step": 1096 }, { "epoch": 0.34551181102362205, "grad_norm": 31.790834426879883, "learning_rate": 9.998278740157481e-06, "loss": 0.4586, "step": 1097 }, { "epoch": 0.3458267716535433, "grad_norm": 15.95163631439209, "learning_rate": 9.998277165354332e-06, "loss": 0.1477, "step": 1098 }, { "epoch": 0.34614173228346456, "grad_norm": 57.9958610534668, "learning_rate": 9.998275590551182e-06, "loss": 0.2226, "step": 1099 }, { "epoch": 0.3464566929133858, "grad_norm": 31.550888061523438, "learning_rate": 9.998274015748033e-06, "loss": 0.3224, "step": 1100 }, { "epoch": 0.3464566929133858, "eval_loss": 0.5266521573066711, "eval_runtime": 337.1873, "eval_samples_per_second": 0.347, "eval_steps_per_second": 0.347, "step": 1100 }, { "epoch": 0.3467716535433071, "grad_norm": 27.716707229614258, "learning_rate": 9.998272440944883e-06, "loss": 0.2458, "step": 1101 }, { "epoch": 0.34708661417322834, "grad_norm": 36.678863525390625, "learning_rate": 9.998270866141733e-06, "loss": 0.315, "step": 1102 }, { "epoch": 0.3474015748031496, "grad_norm": 21.499208450317383, "learning_rate": 9.998269291338584e-06, "loss": 0.2061, "step": 1103 }, { "epoch": 0.34771653543307085, "grad_norm": 63.6679801940918, "learning_rate": 9.998267716535433e-06, "loss": 1.0103, "step": 1104 }, { "epoch": 0.3480314960629921, "grad_norm": 34.54896545410156, "learning_rate": 9.998266141732284e-06, "loss": 0.5376, "step": 1105 }, { "epoch": 0.34834645669291336, "grad_norm": 16.613237380981445, "learning_rate": 9.998264566929135e-06, "loss": 0.0988, "step": 1106 }, { "epoch": 0.3486614173228346, "grad_norm": 73.5743179321289, "learning_rate": 9.998262992125986e-06, "loss": 0.8173, "step": 1107 }, { "epoch": 0.3489763779527559, "grad_norm": 6.472119331359863, "learning_rate": 9.998261417322835e-06, "loss": 0.0331, "step": 1108 }, { "epoch": 0.3492913385826772, "grad_norm": 24.923892974853516, "learning_rate": 9.998259842519686e-06, "loss": 0.1636, "step": 1109 }, { "epoch": 0.34960629921259845, "grad_norm": 76.55589294433594, "learning_rate": 9.998258267716536e-06, "loss": 0.5014, "step": 1110 }, { "epoch": 0.3499212598425197, "grad_norm": 15.523338317871094, "learning_rate": 9.998256692913387e-06, "loss": 0.0701, "step": 1111 }, { "epoch": 0.35023622047244096, "grad_norm": 72.51299285888672, "learning_rate": 9.998255118110238e-06, "loss": 0.5609, "step": 1112 }, { "epoch": 0.3505511811023622, "grad_norm": 41.60987854003906, "learning_rate": 9.998253543307087e-06, "loss": 0.2, "step": 1113 }, { "epoch": 0.3508661417322835, "grad_norm": 4.41101598739624, "learning_rate": 9.998251968503938e-06, "loss": 0.0192, "step": 1114 }, { "epoch": 0.35118110236220473, "grad_norm": 69.58025360107422, "learning_rate": 9.998250393700787e-06, "loss": 0.8551, "step": 1115 }, { "epoch": 0.351496062992126, "grad_norm": 34.081336975097656, "learning_rate": 9.998248818897638e-06, "loss": 0.6141, "step": 1116 }, { "epoch": 0.35181102362204725, "grad_norm": 46.96076965332031, "learning_rate": 9.99824724409449e-06, "loss": 0.8069, "step": 1117 }, { "epoch": 0.3521259842519685, "grad_norm": 88.97361755371094, "learning_rate": 9.99824566929134e-06, "loss": 0.3247, "step": 1118 }, { "epoch": 0.35244094488188976, "grad_norm": 16.12812042236328, "learning_rate": 9.99824409448819e-06, "loss": 0.0924, "step": 1119 }, { "epoch": 0.352755905511811, "grad_norm": 3.950244188308716, "learning_rate": 9.99824251968504e-06, "loss": 0.0298, "step": 1120 }, { "epoch": 0.352755905511811, "eval_loss": 0.5020915865898132, "eval_runtime": 340.8492, "eval_samples_per_second": 0.343, "eval_steps_per_second": 0.343, "step": 1120 }, { "epoch": 0.3530708661417323, "grad_norm": 86.11933898925781, "learning_rate": 9.998240944881891e-06, "loss": 0.9179, "step": 1121 }, { "epoch": 0.35338582677165353, "grad_norm": 41.664955139160156, "learning_rate": 9.99823937007874e-06, "loss": 0.1796, "step": 1122 }, { "epoch": 0.3537007874015748, "grad_norm": 22.196773529052734, "learning_rate": 9.998237795275592e-06, "loss": 0.143, "step": 1123 }, { "epoch": 0.35401574803149605, "grad_norm": 57.799415588378906, "learning_rate": 9.998236220472441e-06, "loss": 0.441, "step": 1124 }, { "epoch": 0.3543307086614173, "grad_norm": 23.602643966674805, "learning_rate": 9.998234645669292e-06, "loss": 0.325, "step": 1125 }, { "epoch": 0.35464566929133856, "grad_norm": 49.98581314086914, "learning_rate": 9.998233070866143e-06, "loss": 0.5312, "step": 1126 }, { "epoch": 0.3549606299212598, "grad_norm": 32.001861572265625, "learning_rate": 9.998231496062994e-06, "loss": 0.2678, "step": 1127 }, { "epoch": 0.3552755905511811, "grad_norm": 22.768354415893555, "learning_rate": 9.998229921259843e-06, "loss": 0.2655, "step": 1128 }, { "epoch": 0.3555905511811024, "grad_norm": 10.575422286987305, "learning_rate": 9.998228346456694e-06, "loss": 0.067, "step": 1129 }, { "epoch": 0.35590551181102364, "grad_norm": 23.114152908325195, "learning_rate": 9.998226771653544e-06, "loss": 0.1502, "step": 1130 }, { "epoch": 0.3562204724409449, "grad_norm": 55.854827880859375, "learning_rate": 9.998225196850395e-06, "loss": 0.7007, "step": 1131 }, { "epoch": 0.35653543307086616, "grad_norm": 43.69165802001953, "learning_rate": 9.998223622047246e-06, "loss": 0.2763, "step": 1132 }, { "epoch": 0.3568503937007874, "grad_norm": 10.879396438598633, "learning_rate": 9.998222047244095e-06, "loss": 0.0361, "step": 1133 }, { "epoch": 0.3571653543307087, "grad_norm": 78.35888671875, "learning_rate": 9.998220472440946e-06, "loss": 0.089, "step": 1134 }, { "epoch": 0.35748031496062993, "grad_norm": 24.78093147277832, "learning_rate": 9.998218897637795e-06, "loss": 0.1069, "step": 1135 }, { "epoch": 0.3577952755905512, "grad_norm": 46.83030700683594, "learning_rate": 9.998217322834646e-06, "loss": 0.472, "step": 1136 }, { "epoch": 0.35811023622047244, "grad_norm": 15.252365112304688, "learning_rate": 9.998215748031497e-06, "loss": 0.0778, "step": 1137 }, { "epoch": 0.3584251968503937, "grad_norm": 15.580936431884766, "learning_rate": 9.998214173228348e-06, "loss": 0.1253, "step": 1138 }, { "epoch": 0.35874015748031496, "grad_norm": 2.145813226699829, "learning_rate": 9.998212598425197e-06, "loss": 0.0093, "step": 1139 }, { "epoch": 0.3590551181102362, "grad_norm": 204.69932556152344, "learning_rate": 9.998211023622047e-06, "loss": 1.3589, "step": 1140 }, { "epoch": 0.3590551181102362, "eval_loss": 0.5789304971694946, "eval_runtime": 339.4483, "eval_samples_per_second": 0.345, "eval_steps_per_second": 0.345, "step": 1140 }, { "epoch": 0.3593700787401575, "grad_norm": 134.10121154785156, "learning_rate": 9.998209448818898e-06, "loss": 0.3836, "step": 1141 }, { "epoch": 0.35968503937007873, "grad_norm": 49.37288284301758, "learning_rate": 9.998207874015749e-06, "loss": 0.5611, "step": 1142 }, { "epoch": 0.36, "grad_norm": 42.33864212036133, "learning_rate": 9.9982062992126e-06, "loss": 0.1906, "step": 1143 }, { "epoch": 0.36031496062992124, "grad_norm": 75.55062866210938, "learning_rate": 9.998204724409449e-06, "loss": 0.1209, "step": 1144 }, { "epoch": 0.3606299212598425, "grad_norm": 30.669452667236328, "learning_rate": 9.9982031496063e-06, "loss": 0.097, "step": 1145 }, { "epoch": 0.36094488188976376, "grad_norm": 96.1650390625, "learning_rate": 9.998201574803151e-06, "loss": 0.9096, "step": 1146 }, { "epoch": 0.361259842519685, "grad_norm": 108.84801483154297, "learning_rate": 9.998200000000002e-06, "loss": 0.4399, "step": 1147 }, { "epoch": 0.36157480314960627, "grad_norm": 101.94995880126953, "learning_rate": 9.998198425196851e-06, "loss": 0.5331, "step": 1148 }, { "epoch": 0.3618897637795276, "grad_norm": 42.10065841674805, "learning_rate": 9.998196850393702e-06, "loss": 0.2717, "step": 1149 }, { "epoch": 0.36220472440944884, "grad_norm": 34.27854919433594, "learning_rate": 9.998195275590552e-06, "loss": 0.0578, "step": 1150 }, { "epoch": 0.3625196850393701, "grad_norm": 80.90559387207031, "learning_rate": 9.998193700787403e-06, "loss": 0.2224, "step": 1151 }, { "epoch": 0.36283464566929136, "grad_norm": 89.94649505615234, "learning_rate": 9.998192125984254e-06, "loss": 0.2526, "step": 1152 }, { "epoch": 0.3631496062992126, "grad_norm": 96.36505126953125, "learning_rate": 9.998190551181103e-06, "loss": 0.521, "step": 1153 }, { "epoch": 0.36346456692913387, "grad_norm": 177.08819580078125, "learning_rate": 9.998188976377954e-06, "loss": 0.6533, "step": 1154 }, { "epoch": 0.3637795275590551, "grad_norm": 60.227481842041016, "learning_rate": 9.998187401574803e-06, "loss": 0.8652, "step": 1155 }, { "epoch": 0.3640944881889764, "grad_norm": 69.29286193847656, "learning_rate": 9.998185826771654e-06, "loss": 1.2497, "step": 1156 }, { "epoch": 0.36440944881889764, "grad_norm": 59.95965576171875, "learning_rate": 9.998184251968505e-06, "loss": 0.4519, "step": 1157 }, { "epoch": 0.3647244094488189, "grad_norm": 61.35934066772461, "learning_rate": 9.998182677165356e-06, "loss": 0.2029, "step": 1158 }, { "epoch": 0.36503937007874016, "grad_norm": 32.42390060424805, "learning_rate": 9.998181102362205e-06, "loss": 0.5222, "step": 1159 }, { "epoch": 0.3653543307086614, "grad_norm": 58.62075424194336, "learning_rate": 9.998179527559055e-06, "loss": 0.6079, "step": 1160 }, { "epoch": 0.3653543307086614, "eval_loss": 0.5354205965995789, "eval_runtime": 339.1978, "eval_samples_per_second": 0.345, "eval_steps_per_second": 0.345, "step": 1160 }, { "epoch": 0.36566929133858267, "grad_norm": 61.64841079711914, "learning_rate": 9.998177952755906e-06, "loss": 0.603, "step": 1161 }, { "epoch": 0.3659842519685039, "grad_norm": 34.8085823059082, "learning_rate": 9.998176377952757e-06, "loss": 0.3863, "step": 1162 }, { "epoch": 0.3662992125984252, "grad_norm": 25.715442657470703, "learning_rate": 9.998174803149608e-06, "loss": 0.2857, "step": 1163 }, { "epoch": 0.36661417322834644, "grad_norm": 33.884483337402344, "learning_rate": 9.998173228346457e-06, "loss": 0.549, "step": 1164 }, { "epoch": 0.3669291338582677, "grad_norm": 24.125484466552734, "learning_rate": 9.998171653543308e-06, "loss": 0.2794, "step": 1165 }, { "epoch": 0.36724409448818895, "grad_norm": 67.11617279052734, "learning_rate": 9.998170078740159e-06, "loss": 0.6092, "step": 1166 }, { "epoch": 0.3675590551181102, "grad_norm": 23.704805374145508, "learning_rate": 9.99816850393701e-06, "loss": 0.4884, "step": 1167 }, { "epoch": 0.36787401574803147, "grad_norm": 25.822975158691406, "learning_rate": 9.99816692913386e-06, "loss": 0.5046, "step": 1168 }, { "epoch": 0.3681889763779528, "grad_norm": 31.311058044433594, "learning_rate": 9.99816535433071e-06, "loss": 0.3914, "step": 1169 }, { "epoch": 0.36850393700787404, "grad_norm": 18.38756561279297, "learning_rate": 9.99816377952756e-06, "loss": 0.4273, "step": 1170 }, { "epoch": 0.3688188976377953, "grad_norm": 20.738372802734375, "learning_rate": 9.99816220472441e-06, "loss": 0.2612, "step": 1171 }, { "epoch": 0.36913385826771655, "grad_norm": 23.450641632080078, "learning_rate": 9.998160629921261e-06, "loss": 0.3474, "step": 1172 }, { "epoch": 0.3694488188976378, "grad_norm": 38.92578125, "learning_rate": 9.99815905511811e-06, "loss": 0.5643, "step": 1173 }, { "epoch": 0.36976377952755907, "grad_norm": 59.437496185302734, "learning_rate": 9.998157480314962e-06, "loss": 0.7971, "step": 1174 }, { "epoch": 0.3700787401574803, "grad_norm": 22.928701400756836, "learning_rate": 9.998155905511811e-06, "loss": 0.2843, "step": 1175 }, { "epoch": 0.3703937007874016, "grad_norm": 38.7431526184082, "learning_rate": 9.998154330708662e-06, "loss": 0.3979, "step": 1176 }, { "epoch": 0.37070866141732284, "grad_norm": 5.290953636169434, "learning_rate": 9.998152755905513e-06, "loss": 0.0298, "step": 1177 }, { "epoch": 0.3710236220472441, "grad_norm": 23.754051208496094, "learning_rate": 9.998151181102364e-06, "loss": 0.4325, "step": 1178 }, { "epoch": 0.37133858267716535, "grad_norm": 27.173952102661133, "learning_rate": 9.998149606299213e-06, "loss": 0.5674, "step": 1179 }, { "epoch": 0.3716535433070866, "grad_norm": 42.20391845703125, "learning_rate": 9.998148031496063e-06, "loss": 0.5973, "step": 1180 }, { "epoch": 0.3716535433070866, "eval_loss": 0.5375993251800537, "eval_runtime": 339.4527, "eval_samples_per_second": 0.345, "eval_steps_per_second": 0.345, "step": 1180 }, { "epoch": 0.37196850393700787, "grad_norm": 11.690238952636719, "learning_rate": 9.998146456692914e-06, "loss": 0.0917, "step": 1181 }, { "epoch": 0.3722834645669291, "grad_norm": 10.012290000915527, "learning_rate": 9.998144881889765e-06, "loss": 0.0842, "step": 1182 }, { "epoch": 0.3725984251968504, "grad_norm": 21.645814895629883, "learning_rate": 9.998143307086616e-06, "loss": 0.152, "step": 1183 }, { "epoch": 0.37291338582677164, "grad_norm": 31.5441837310791, "learning_rate": 9.998141732283465e-06, "loss": 0.5621, "step": 1184 }, { "epoch": 0.3732283464566929, "grad_norm": 41.74612808227539, "learning_rate": 9.998140157480316e-06, "loss": 0.7019, "step": 1185 }, { "epoch": 0.37354330708661415, "grad_norm": 35.047794342041016, "learning_rate": 9.998138582677165e-06, "loss": 0.1264, "step": 1186 }, { "epoch": 0.3738582677165354, "grad_norm": 77.93087005615234, "learning_rate": 9.998137007874016e-06, "loss": 0.8623, "step": 1187 }, { "epoch": 0.37417322834645667, "grad_norm": 32.53571701049805, "learning_rate": 9.998135433070867e-06, "loss": 0.2572, "step": 1188 }, { "epoch": 0.374488188976378, "grad_norm": 57.9747428894043, "learning_rate": 9.998133858267718e-06, "loss": 0.5365, "step": 1189 }, { "epoch": 0.37480314960629924, "grad_norm": 46.87409210205078, "learning_rate": 9.998132283464567e-06, "loss": 1.2091, "step": 1190 }, { "epoch": 0.3751181102362205, "grad_norm": 29.17926788330078, "learning_rate": 9.998130708661418e-06, "loss": 0.3407, "step": 1191 }, { "epoch": 0.37543307086614175, "grad_norm": 43.51498794555664, "learning_rate": 9.99812913385827e-06, "loss": 0.747, "step": 1192 }, { "epoch": 0.375748031496063, "grad_norm": 63.52394104003906, "learning_rate": 9.998127559055119e-06, "loss": 0.5045, "step": 1193 }, { "epoch": 0.37606299212598426, "grad_norm": 82.31373596191406, "learning_rate": 9.99812598425197e-06, "loss": 0.9014, "step": 1194 }, { "epoch": 0.3763779527559055, "grad_norm": 70.40677642822266, "learning_rate": 9.998124409448819e-06, "loss": 0.5663, "step": 1195 }, { "epoch": 0.3766929133858268, "grad_norm": 16.95841407775879, "learning_rate": 9.99812283464567e-06, "loss": 0.3155, "step": 1196 }, { "epoch": 0.37700787401574803, "grad_norm": 46.58156967163086, "learning_rate": 9.998121259842521e-06, "loss": 0.6919, "step": 1197 }, { "epoch": 0.3773228346456693, "grad_norm": 37.353492736816406, "learning_rate": 9.998119685039372e-06, "loss": 0.669, "step": 1198 }, { "epoch": 0.37763779527559055, "grad_norm": 23.72784996032715, "learning_rate": 9.998118110236221e-06, "loss": 0.3445, "step": 1199 }, { "epoch": 0.3779527559055118, "grad_norm": 39.98880386352539, "learning_rate": 9.99811653543307e-06, "loss": 0.4265, "step": 1200 }, { "epoch": 0.3779527559055118, "eval_loss": 0.5579959154129028, "eval_runtime": 337.5471, "eval_samples_per_second": 0.347, "eval_steps_per_second": 0.347, "step": 1200 }, { "epoch": 0.37826771653543306, "grad_norm": 13.343225479125977, "learning_rate": 9.998114960629922e-06, "loss": 0.346, "step": 1201 }, { "epoch": 0.3785826771653543, "grad_norm": 31.93812370300293, "learning_rate": 9.998113385826773e-06, "loss": 0.7538, "step": 1202 }, { "epoch": 0.3788976377952756, "grad_norm": 22.874921798706055, "learning_rate": 9.998111811023624e-06, "loss": 0.2865, "step": 1203 }, { "epoch": 0.37921259842519683, "grad_norm": 50.595577239990234, "learning_rate": 9.998110236220473e-06, "loss": 0.7486, "step": 1204 }, { "epoch": 0.3795275590551181, "grad_norm": 20.3195858001709, "learning_rate": 9.998108661417324e-06, "loss": 0.2986, "step": 1205 }, { "epoch": 0.37984251968503935, "grad_norm": 18.59178352355957, "learning_rate": 9.998107086614173e-06, "loss": 0.3259, "step": 1206 }, { "epoch": 0.3801574803149606, "grad_norm": 23.734966278076172, "learning_rate": 9.998105511811024e-06, "loss": 0.4835, "step": 1207 }, { "epoch": 0.38047244094488186, "grad_norm": 47.375789642333984, "learning_rate": 9.998103937007875e-06, "loss": 0.4708, "step": 1208 }, { "epoch": 0.3807874015748032, "grad_norm": 12.326082229614258, "learning_rate": 9.998102362204726e-06, "loss": 0.3698, "step": 1209 }, { "epoch": 0.38110236220472443, "grad_norm": 30.175519943237305, "learning_rate": 9.998100787401575e-06, "loss": 0.4141, "step": 1210 }, { "epoch": 0.3814173228346457, "grad_norm": 27.1715087890625, "learning_rate": 9.998099212598425e-06, "loss": 0.4674, "step": 1211 }, { "epoch": 0.38173228346456695, "grad_norm": 31.19744300842285, "learning_rate": 9.998097637795277e-06, "loss": 0.3239, "step": 1212 }, { "epoch": 0.3820472440944882, "grad_norm": 44.07891845703125, "learning_rate": 9.998096062992127e-06, "loss": 0.7672, "step": 1213 }, { "epoch": 0.38236220472440946, "grad_norm": 15.165576934814453, "learning_rate": 9.998094488188978e-06, "loss": 0.2029, "step": 1214 }, { "epoch": 0.3826771653543307, "grad_norm": 41.70854187011719, "learning_rate": 9.998092913385827e-06, "loss": 0.3442, "step": 1215 }, { "epoch": 0.382992125984252, "grad_norm": 25.64063835144043, "learning_rate": 9.998091338582678e-06, "loss": 0.3051, "step": 1216 }, { "epoch": 0.38330708661417323, "grad_norm": 19.23823356628418, "learning_rate": 9.998089763779529e-06, "loss": 0.1983, "step": 1217 }, { "epoch": 0.3836220472440945, "grad_norm": 65.09869384765625, "learning_rate": 9.99808818897638e-06, "loss": 0.4903, "step": 1218 }, { "epoch": 0.38393700787401575, "grad_norm": 14.515801429748535, "learning_rate": 9.99808661417323e-06, "loss": 0.1511, "step": 1219 }, { "epoch": 0.384251968503937, "grad_norm": 34.14856719970703, "learning_rate": 9.998085039370079e-06, "loss": 0.4351, "step": 1220 }, { "epoch": 0.384251968503937, "eval_loss": 0.49825048446655273, "eval_runtime": 340.3122, "eval_samples_per_second": 0.344, "eval_steps_per_second": 0.344, "step": 1220 }, { "epoch": 0.38456692913385826, "grad_norm": 42.81884765625, "learning_rate": 9.99808346456693e-06, "loss": 0.4317, "step": 1221 }, { "epoch": 0.3848818897637795, "grad_norm": 40.3559455871582, "learning_rate": 9.99808188976378e-06, "loss": 0.5042, "step": 1222 }, { "epoch": 0.3851968503937008, "grad_norm": 28.505815505981445, "learning_rate": 9.998080314960631e-06, "loss": 0.316, "step": 1223 }, { "epoch": 0.38551181102362203, "grad_norm": 20.508024215698242, "learning_rate": 9.99807874015748e-06, "loss": 0.15, "step": 1224 }, { "epoch": 0.3858267716535433, "grad_norm": 73.8309555053711, "learning_rate": 9.998077165354332e-06, "loss": 0.4734, "step": 1225 }, { "epoch": 0.38614173228346454, "grad_norm": 24.401338577270508, "learning_rate": 9.998075590551181e-06, "loss": 0.2553, "step": 1226 }, { "epoch": 0.3864566929133858, "grad_norm": 28.52256202697754, "learning_rate": 9.998074015748032e-06, "loss": 0.0865, "step": 1227 }, { "epoch": 0.38677165354330706, "grad_norm": 45.155696868896484, "learning_rate": 9.998072440944883e-06, "loss": 0.413, "step": 1228 }, { "epoch": 0.38708661417322837, "grad_norm": 23.677753448486328, "learning_rate": 9.998070866141734e-06, "loss": 0.1919, "step": 1229 }, { "epoch": 0.38740157480314963, "grad_norm": 27.503589630126953, "learning_rate": 9.998069291338583e-06, "loss": 0.6631, "step": 1230 }, { "epoch": 0.3877165354330709, "grad_norm": 33.99694061279297, "learning_rate": 9.998067716535433e-06, "loss": 0.5537, "step": 1231 }, { "epoch": 0.38803149606299214, "grad_norm": 73.67473602294922, "learning_rate": 9.998066141732284e-06, "loss": 0.2761, "step": 1232 }, { "epoch": 0.3883464566929134, "grad_norm": 47.667327880859375, "learning_rate": 9.998064566929135e-06, "loss": 0.1336, "step": 1233 }, { "epoch": 0.38866141732283466, "grad_norm": 32.623802185058594, "learning_rate": 9.998062992125986e-06, "loss": 0.1881, "step": 1234 }, { "epoch": 0.3889763779527559, "grad_norm": 92.79457092285156, "learning_rate": 9.998061417322835e-06, "loss": 0.9026, "step": 1235 }, { "epoch": 0.38929133858267717, "grad_norm": 47.63346862792969, "learning_rate": 9.998059842519686e-06, "loss": 0.2402, "step": 1236 }, { "epoch": 0.38960629921259843, "grad_norm": 113.52471923828125, "learning_rate": 9.998058267716537e-06, "loss": 0.7758, "step": 1237 }, { "epoch": 0.3899212598425197, "grad_norm": 44.212303161621094, "learning_rate": 9.998056692913388e-06, "loss": 0.1154, "step": 1238 }, { "epoch": 0.39023622047244094, "grad_norm": 113.48955535888672, "learning_rate": 9.998055118110237e-06, "loss": 0.7735, "step": 1239 }, { "epoch": 0.3905511811023622, "grad_norm": 95.3028564453125, "learning_rate": 9.998053543307086e-06, "loss": 0.5174, "step": 1240 }, { "epoch": 0.3905511811023622, "eval_loss": 0.680210292339325, "eval_runtime": 337.0826, "eval_samples_per_second": 0.347, "eval_steps_per_second": 0.347, "step": 1240 }, { "epoch": 0.39086614173228346, "grad_norm": 75.17251586914062, "learning_rate": 9.998051968503937e-06, "loss": 0.6919, "step": 1241 }, { "epoch": 0.3911811023622047, "grad_norm": 94.41243743896484, "learning_rate": 9.998050393700788e-06, "loss": 1.1939, "step": 1242 }, { "epoch": 0.39149606299212597, "grad_norm": 17.574474334716797, "learning_rate": 9.99804881889764e-06, "loss": 0.0841, "step": 1243 }, { "epoch": 0.3918110236220472, "grad_norm": 22.532442092895508, "learning_rate": 9.998047244094489e-06, "loss": 0.0756, "step": 1244 }, { "epoch": 0.3921259842519685, "grad_norm": 48.4405632019043, "learning_rate": 9.99804566929134e-06, "loss": 0.225, "step": 1245 }, { "epoch": 0.39244094488188974, "grad_norm": 76.005859375, "learning_rate": 9.998044094488189e-06, "loss": 0.1798, "step": 1246 }, { "epoch": 0.392755905511811, "grad_norm": 90.26568603515625, "learning_rate": 9.99804251968504e-06, "loss": 1.1972, "step": 1247 }, { "epoch": 0.39307086614173226, "grad_norm": 69.64557647705078, "learning_rate": 9.998040944881891e-06, "loss": 0.1896, "step": 1248 }, { "epoch": 0.39338582677165357, "grad_norm": 68.00067901611328, "learning_rate": 9.998039370078742e-06, "loss": 0.7109, "step": 1249 }, { "epoch": 0.3937007874015748, "grad_norm": 57.22831344604492, "learning_rate": 9.998037795275591e-06, "loss": 0.8236, "step": 1250 }, { "epoch": 0.3940157480314961, "grad_norm": 47.59756851196289, "learning_rate": 9.99803622047244e-06, "loss": 0.2775, "step": 1251 }, { "epoch": 0.39433070866141734, "grad_norm": 23.361492156982422, "learning_rate": 9.998034645669292e-06, "loss": 0.0931, "step": 1252 }, { "epoch": 0.3946456692913386, "grad_norm": 40.62185287475586, "learning_rate": 9.998033070866143e-06, "loss": 0.3612, "step": 1253 }, { "epoch": 0.39496062992125985, "grad_norm": 31.514081954956055, "learning_rate": 9.998031496062994e-06, "loss": 0.0952, "step": 1254 }, { "epoch": 0.3952755905511811, "grad_norm": 39.17756652832031, "learning_rate": 9.998029921259843e-06, "loss": 0.2162, "step": 1255 }, { "epoch": 0.39559055118110237, "grad_norm": 19.34990882873535, "learning_rate": 9.998028346456694e-06, "loss": 0.1018, "step": 1256 }, { "epoch": 0.3959055118110236, "grad_norm": 40.027671813964844, "learning_rate": 9.998026771653543e-06, "loss": 0.217, "step": 1257 }, { "epoch": 0.3962204724409449, "grad_norm": 97.05489349365234, "learning_rate": 9.998025196850394e-06, "loss": 0.3828, "step": 1258 }, { "epoch": 0.39653543307086614, "grad_norm": 20.452852249145508, "learning_rate": 9.998023622047245e-06, "loss": 0.1665, "step": 1259 }, { "epoch": 0.3968503937007874, "grad_norm": 76.07334899902344, "learning_rate": 9.998022047244094e-06, "loss": 0.9711, "step": 1260 }, { "epoch": 0.3968503937007874, "eval_loss": 0.508359432220459, "eval_runtime": 337.8333, "eval_samples_per_second": 0.346, "eval_steps_per_second": 0.346, "step": 1260 }, { "epoch": 0.39716535433070865, "grad_norm": 44.56214904785156, "learning_rate": 9.998020472440945e-06, "loss": 0.1862, "step": 1261 }, { "epoch": 0.3974803149606299, "grad_norm": 105.54949188232422, "learning_rate": 9.998018897637796e-06, "loss": 1.0023, "step": 1262 }, { "epoch": 0.39779527559055117, "grad_norm": 64.92313385009766, "learning_rate": 9.998017322834647e-06, "loss": 0.9656, "step": 1263 }, { "epoch": 0.3981102362204724, "grad_norm": 60.504730224609375, "learning_rate": 9.998015748031497e-06, "loss": 0.3303, "step": 1264 }, { "epoch": 0.3984251968503937, "grad_norm": 57.36290740966797, "learning_rate": 9.998014173228348e-06, "loss": 0.2705, "step": 1265 }, { "epoch": 0.39874015748031494, "grad_norm": 49.48656463623047, "learning_rate": 9.998012598425197e-06, "loss": 0.5267, "step": 1266 }, { "epoch": 0.3990551181102362, "grad_norm": 73.73528289794922, "learning_rate": 9.998011023622048e-06, "loss": 0.2752, "step": 1267 }, { "epoch": 0.39937007874015745, "grad_norm": 77.58939361572266, "learning_rate": 9.998009448818899e-06, "loss": 0.782, "step": 1268 }, { "epoch": 0.39968503937007877, "grad_norm": 13.655567169189453, "learning_rate": 9.99800787401575e-06, "loss": 0.0997, "step": 1269 }, { "epoch": 0.4, "grad_norm": 31.09360122680664, "learning_rate": 9.9980062992126e-06, "loss": 0.275, "step": 1270 }, { "epoch": 0.4003149606299213, "grad_norm": 57.30973815917969, "learning_rate": 9.998004724409449e-06, "loss": 0.5763, "step": 1271 }, { "epoch": 0.40062992125984254, "grad_norm": 55.74612808227539, "learning_rate": 9.9980031496063e-06, "loss": 0.8855, "step": 1272 }, { "epoch": 0.4009448818897638, "grad_norm": 45.396583557128906, "learning_rate": 9.99800157480315e-06, "loss": 0.4024, "step": 1273 }, { "epoch": 0.40125984251968505, "grad_norm": 17.555898666381836, "learning_rate": 9.998000000000002e-06, "loss": 0.128, "step": 1274 }, { "epoch": 0.4015748031496063, "grad_norm": 54.747310638427734, "learning_rate": 9.99799842519685e-06, "loss": 0.8548, "step": 1275 }, { "epoch": 0.40188976377952756, "grad_norm": 36.224857330322266, "learning_rate": 9.997996850393702e-06, "loss": 0.2832, "step": 1276 }, { "epoch": 0.4022047244094488, "grad_norm": 20.606124877929688, "learning_rate": 9.997995275590551e-06, "loss": 0.2242, "step": 1277 }, { "epoch": 0.4025196850393701, "grad_norm": 37.51258087158203, "learning_rate": 9.997993700787402e-06, "loss": 0.3867, "step": 1278 }, { "epoch": 0.40283464566929134, "grad_norm": 14.421310424804688, "learning_rate": 9.997992125984253e-06, "loss": 0.1174, "step": 1279 }, { "epoch": 0.4031496062992126, "grad_norm": 41.61854553222656, "learning_rate": 9.997990551181104e-06, "loss": 0.3572, "step": 1280 }, { "epoch": 0.4031496062992126, "eval_loss": 0.48692360520362854, "eval_runtime": 338.2236, "eval_samples_per_second": 0.346, "eval_steps_per_second": 0.346, "step": 1280 }, { "epoch": 0.40346456692913385, "grad_norm": 23.846805572509766, "learning_rate": 9.997988976377953e-06, "loss": 0.419, "step": 1281 }, { "epoch": 0.4037795275590551, "grad_norm": 48.06615447998047, "learning_rate": 9.997987401574804e-06, "loss": 0.8735, "step": 1282 }, { "epoch": 0.40409448818897636, "grad_norm": 15.782271385192871, "learning_rate": 9.997985826771655e-06, "loss": 0.0723, "step": 1283 }, { "epoch": 0.4044094488188976, "grad_norm": 19.87238883972168, "learning_rate": 9.997984251968505e-06, "loss": 0.1359, "step": 1284 }, { "epoch": 0.4047244094488189, "grad_norm": 38.17216110229492, "learning_rate": 9.997982677165356e-06, "loss": 0.3926, "step": 1285 }, { "epoch": 0.40503937007874014, "grad_norm": 45.54020309448242, "learning_rate": 9.997981102362205e-06, "loss": 0.3831, "step": 1286 }, { "epoch": 0.4053543307086614, "grad_norm": 26.464305877685547, "learning_rate": 9.997979527559056e-06, "loss": 0.4216, "step": 1287 }, { "epoch": 0.4056692913385827, "grad_norm": 64.78831481933594, "learning_rate": 9.997977952755907e-06, "loss": 0.861, "step": 1288 }, { "epoch": 0.40598425196850396, "grad_norm": 38.6380615234375, "learning_rate": 9.997976377952758e-06, "loss": 0.3388, "step": 1289 }, { "epoch": 0.4062992125984252, "grad_norm": 44.91901779174805, "learning_rate": 9.997974803149607e-06, "loss": 0.2208, "step": 1290 }, { "epoch": 0.4066141732283465, "grad_norm": 59.079185485839844, "learning_rate": 9.997973228346456e-06, "loss": 0.6019, "step": 1291 }, { "epoch": 0.40692913385826773, "grad_norm": 65.11566925048828, "learning_rate": 9.997971653543307e-06, "loss": 0.64, "step": 1292 }, { "epoch": 0.407244094488189, "grad_norm": 45.68169021606445, "learning_rate": 9.997970078740158e-06, "loss": 0.6369, "step": 1293 }, { "epoch": 0.40755905511811025, "grad_norm": 6.096194744110107, "learning_rate": 9.99796850393701e-06, "loss": 0.0484, "step": 1294 }, { "epoch": 0.4078740157480315, "grad_norm": 31.807212829589844, "learning_rate": 9.997966929133859e-06, "loss": 0.3238, "step": 1295 }, { "epoch": 0.40818897637795276, "grad_norm": 66.85626983642578, "learning_rate": 9.99796535433071e-06, "loss": 0.6099, "step": 1296 }, { "epoch": 0.408503937007874, "grad_norm": 15.900289535522461, "learning_rate": 9.997963779527559e-06, "loss": 0.3017, "step": 1297 }, { "epoch": 0.4088188976377953, "grad_norm": 58.991703033447266, "learning_rate": 9.99796220472441e-06, "loss": 0.441, "step": 1298 }, { "epoch": 0.40913385826771653, "grad_norm": 64.84235382080078, "learning_rate": 9.997960629921261e-06, "loss": 0.5636, "step": 1299 }, { "epoch": 0.4094488188976378, "grad_norm": 48.843505859375, "learning_rate": 9.997959055118112e-06, "loss": 0.5989, "step": 1300 }, { "epoch": 0.4094488188976378, "eval_loss": 0.5216355323791504, "eval_runtime": 339.9169, "eval_samples_per_second": 0.344, "eval_steps_per_second": 0.344, "step": 1300 }, { "epoch": 0.40976377952755905, "grad_norm": 22.199951171875, "learning_rate": 9.997957480314961e-06, "loss": 0.331, "step": 1301 }, { "epoch": 0.4100787401574803, "grad_norm": 56.680816650390625, "learning_rate": 9.99795590551181e-06, "loss": 0.4215, "step": 1302 }, { "epoch": 0.41039370078740156, "grad_norm": 33.23557662963867, "learning_rate": 9.997954330708662e-06, "loss": 0.444, "step": 1303 }, { "epoch": 0.4107086614173228, "grad_norm": 31.659833908081055, "learning_rate": 9.997952755905513e-06, "loss": 0.4401, "step": 1304 }, { "epoch": 0.4110236220472441, "grad_norm": 14.66598129272461, "learning_rate": 9.997951181102364e-06, "loss": 0.1447, "step": 1305 }, { "epoch": 0.41133858267716533, "grad_norm": 20.910083770751953, "learning_rate": 9.997949606299213e-06, "loss": 0.2887, "step": 1306 }, { "epoch": 0.4116535433070866, "grad_norm": 39.282596588134766, "learning_rate": 9.997948031496064e-06, "loss": 0.5081, "step": 1307 }, { "epoch": 0.4119685039370079, "grad_norm": 37.22988510131836, "learning_rate": 9.997946456692915e-06, "loss": 0.6354, "step": 1308 }, { "epoch": 0.41228346456692916, "grad_norm": 51.220314025878906, "learning_rate": 9.997944881889766e-06, "loss": 0.4205, "step": 1309 }, { "epoch": 0.4125984251968504, "grad_norm": 31.497386932373047, "learning_rate": 9.997943307086615e-06, "loss": 0.4645, "step": 1310 }, { "epoch": 0.4129133858267717, "grad_norm": 48.64023971557617, "learning_rate": 9.997941732283464e-06, "loss": 0.5369, "step": 1311 }, { "epoch": 0.41322834645669293, "grad_norm": 44.41730499267578, "learning_rate": 9.997940157480315e-06, "loss": 0.4616, "step": 1312 }, { "epoch": 0.4135433070866142, "grad_norm": 32.24418640136719, "learning_rate": 9.997938582677166e-06, "loss": 0.4085, "step": 1313 }, { "epoch": 0.41385826771653544, "grad_norm": 25.68863296508789, "learning_rate": 9.997937007874017e-06, "loss": 0.1576, "step": 1314 }, { "epoch": 0.4141732283464567, "grad_norm": 57.67222595214844, "learning_rate": 9.997935433070867e-06, "loss": 0.6927, "step": 1315 }, { "epoch": 0.41448818897637796, "grad_norm": 46.1195182800293, "learning_rate": 9.997933858267718e-06, "loss": 0.3948, "step": 1316 }, { "epoch": 0.4148031496062992, "grad_norm": 16.534828186035156, "learning_rate": 9.997932283464567e-06, "loss": 0.1382, "step": 1317 }, { "epoch": 0.41511811023622047, "grad_norm": 40.07267379760742, "learning_rate": 9.997930708661418e-06, "loss": 0.3056, "step": 1318 }, { "epoch": 0.41543307086614173, "grad_norm": 29.960514068603516, "learning_rate": 9.997929133858269e-06, "loss": 0.1982, "step": 1319 }, { "epoch": 0.415748031496063, "grad_norm": 26.702951431274414, "learning_rate": 9.99792755905512e-06, "loss": 0.4233, "step": 1320 }, { "epoch": 0.415748031496063, "eval_loss": 0.4609772861003876, "eval_runtime": 338.1687, "eval_samples_per_second": 0.346, "eval_steps_per_second": 0.346, "step": 1320 }, { "epoch": 0.41606299212598424, "grad_norm": 31.864824295043945, "learning_rate": 9.99792598425197e-06, "loss": 0.5666, "step": 1321 }, { "epoch": 0.4163779527559055, "grad_norm": 32.256771087646484, "learning_rate": 9.997924409448819e-06, "loss": 0.2308, "step": 1322 }, { "epoch": 0.41669291338582676, "grad_norm": 20.752614974975586, "learning_rate": 9.99792283464567e-06, "loss": 0.3462, "step": 1323 }, { "epoch": 0.417007874015748, "grad_norm": 86.646484375, "learning_rate": 9.99792125984252e-06, "loss": 0.7229, "step": 1324 }, { "epoch": 0.41732283464566927, "grad_norm": 36.791202545166016, "learning_rate": 9.997919685039372e-06, "loss": 0.6036, "step": 1325 }, { "epoch": 0.41763779527559053, "grad_norm": 60.902095794677734, "learning_rate": 9.99791811023622e-06, "loss": 0.7423, "step": 1326 }, { "epoch": 0.4179527559055118, "grad_norm": 15.44100284576416, "learning_rate": 9.997916535433072e-06, "loss": 0.0577, "step": 1327 }, { "epoch": 0.4182677165354331, "grad_norm": 39.82502365112305, "learning_rate": 9.997914960629921e-06, "loss": 0.351, "step": 1328 }, { "epoch": 0.41858267716535436, "grad_norm": 19.225820541381836, "learning_rate": 9.997913385826772e-06, "loss": 0.1259, "step": 1329 }, { "epoch": 0.4188976377952756, "grad_norm": 20.358154296875, "learning_rate": 9.997911811023623e-06, "loss": 0.1947, "step": 1330 }, { "epoch": 0.41921259842519687, "grad_norm": 15.202303886413574, "learning_rate": 9.997910236220472e-06, "loss": 0.0907, "step": 1331 }, { "epoch": 0.4195275590551181, "grad_norm": 16.259374618530273, "learning_rate": 9.997908661417323e-06, "loss": 0.0561, "step": 1332 }, { "epoch": 0.4198425196850394, "grad_norm": 40.858028411865234, "learning_rate": 9.997907086614174e-06, "loss": 0.3551, "step": 1333 }, { "epoch": 0.42015748031496064, "grad_norm": 21.958782196044922, "learning_rate": 9.997905511811025e-06, "loss": 0.1549, "step": 1334 }, { "epoch": 0.4204724409448819, "grad_norm": 60.448734283447266, "learning_rate": 9.997903937007875e-06, "loss": 0.4945, "step": 1335 }, { "epoch": 0.42078740157480315, "grad_norm": 9.495431900024414, "learning_rate": 9.997902362204726e-06, "loss": 0.0391, "step": 1336 }, { "epoch": 0.4211023622047244, "grad_norm": 96.30876922607422, "learning_rate": 9.997900787401575e-06, "loss": 0.5236, "step": 1337 }, { "epoch": 0.42141732283464567, "grad_norm": 100.25637817382812, "learning_rate": 9.997899212598426e-06, "loss": 0.5044, "step": 1338 }, { "epoch": 0.4217322834645669, "grad_norm": 10.620447158813477, "learning_rate": 9.997897637795277e-06, "loss": 0.0333, "step": 1339 }, { "epoch": 0.4220472440944882, "grad_norm": 99.30320739746094, "learning_rate": 9.997896062992128e-06, "loss": 0.4552, "step": 1340 }, { "epoch": 0.4220472440944882, "eval_loss": 0.6429303884506226, "eval_runtime": 339.7084, "eval_samples_per_second": 0.344, "eval_steps_per_second": 0.344, "step": 1340 }, { "epoch": 0.42236220472440944, "grad_norm": 90.87741088867188, "learning_rate": 9.997894488188977e-06, "loss": 0.8964, "step": 1341 }, { "epoch": 0.4226771653543307, "grad_norm": 53.61817169189453, "learning_rate": 9.997892913385827e-06, "loss": 0.2336, "step": 1342 }, { "epoch": 0.42299212598425195, "grad_norm": 69.91616821289062, "learning_rate": 9.997891338582678e-06, "loss": 0.3807, "step": 1343 }, { "epoch": 0.4233070866141732, "grad_norm": 64.52349853515625, "learning_rate": 9.997889763779528e-06, "loss": 0.3968, "step": 1344 }, { "epoch": 0.42362204724409447, "grad_norm": 101.8323745727539, "learning_rate": 9.99788818897638e-06, "loss": 0.6942, "step": 1345 }, { "epoch": 0.4239370078740157, "grad_norm": 137.70166015625, "learning_rate": 9.997886614173229e-06, "loss": 1.0158, "step": 1346 }, { "epoch": 0.424251968503937, "grad_norm": 95.1364974975586, "learning_rate": 9.99788503937008e-06, "loss": 0.4745, "step": 1347 }, { "epoch": 0.4245669291338583, "grad_norm": 42.36204147338867, "learning_rate": 9.997883464566929e-06, "loss": 0.265, "step": 1348 }, { "epoch": 0.42488188976377955, "grad_norm": 37.66621017456055, "learning_rate": 9.99788188976378e-06, "loss": 0.5518, "step": 1349 }, { "epoch": 0.4251968503937008, "grad_norm": 54.68177032470703, "learning_rate": 9.997880314960631e-06, "loss": 0.2369, "step": 1350 }, { "epoch": 0.42551181102362207, "grad_norm": 88.8768310546875, "learning_rate": 9.99787874015748e-06, "loss": 1.004, "step": 1351 }, { "epoch": 0.4258267716535433, "grad_norm": 9.298436164855957, "learning_rate": 9.997877165354331e-06, "loss": 0.0451, "step": 1352 }, { "epoch": 0.4261417322834646, "grad_norm": 68.624755859375, "learning_rate": 9.997875590551182e-06, "loss": 0.9096, "step": 1353 }, { "epoch": 0.42645669291338584, "grad_norm": 42.33887481689453, "learning_rate": 9.997874015748033e-06, "loss": 0.5008, "step": 1354 }, { "epoch": 0.4267716535433071, "grad_norm": 84.9489517211914, "learning_rate": 9.997872440944883e-06, "loss": 0.4279, "step": 1355 }, { "epoch": 0.42708661417322835, "grad_norm": 26.046403884887695, "learning_rate": 9.997870866141734e-06, "loss": 0.2294, "step": 1356 }, { "epoch": 0.4274015748031496, "grad_norm": 24.110301971435547, "learning_rate": 9.997869291338583e-06, "loss": 0.1951, "step": 1357 }, { "epoch": 0.42771653543307087, "grad_norm": 31.500816345214844, "learning_rate": 9.997867716535434e-06, "loss": 0.1093, "step": 1358 }, { "epoch": 0.4280314960629921, "grad_norm": 37.86301040649414, "learning_rate": 9.997866141732285e-06, "loss": 0.5599, "step": 1359 }, { "epoch": 0.4283464566929134, "grad_norm": 68.08171081542969, "learning_rate": 9.997864566929136e-06, "loss": 1.32, "step": 1360 }, { "epoch": 0.4283464566929134, "eval_loss": 0.47423291206359863, "eval_runtime": 339.2488, "eval_samples_per_second": 0.345, "eval_steps_per_second": 0.345, "step": 1360 }, { "epoch": 0.42866141732283464, "grad_norm": 32.4645881652832, "learning_rate": 9.997862992125985e-06, "loss": 0.2244, "step": 1361 }, { "epoch": 0.4289763779527559, "grad_norm": 21.44221305847168, "learning_rate": 9.997861417322834e-06, "loss": 0.1674, "step": 1362 }, { "epoch": 0.42929133858267715, "grad_norm": 67.8936538696289, "learning_rate": 9.997859842519685e-06, "loss": 1.0609, "step": 1363 }, { "epoch": 0.4296062992125984, "grad_norm": 44.46934127807617, "learning_rate": 9.997858267716536e-06, "loss": 0.4285, "step": 1364 }, { "epoch": 0.42992125984251967, "grad_norm": 25.02652359008789, "learning_rate": 9.997856692913387e-06, "loss": 0.5787, "step": 1365 }, { "epoch": 0.4302362204724409, "grad_norm": 53.81482696533203, "learning_rate": 9.997855118110237e-06, "loss": 0.3341, "step": 1366 }, { "epoch": 0.4305511811023622, "grad_norm": 47.348201751708984, "learning_rate": 9.997853543307088e-06, "loss": 0.6322, "step": 1367 }, { "epoch": 0.4308661417322835, "grad_norm": 22.753206253051758, "learning_rate": 9.997851968503937e-06, "loss": 0.4667, "step": 1368 }, { "epoch": 0.43118110236220475, "grad_norm": 13.805994033813477, "learning_rate": 9.997850393700788e-06, "loss": 0.2587, "step": 1369 }, { "epoch": 0.431496062992126, "grad_norm": 39.555076599121094, "learning_rate": 9.997848818897639e-06, "loss": 0.4924, "step": 1370 }, { "epoch": 0.43181102362204726, "grad_norm": 34.64474105834961, "learning_rate": 9.997847244094488e-06, "loss": 0.5824, "step": 1371 }, { "epoch": 0.4321259842519685, "grad_norm": 19.8635311126709, "learning_rate": 9.99784566929134e-06, "loss": 0.3249, "step": 1372 }, { "epoch": 0.4324409448818898, "grad_norm": 45.133663177490234, "learning_rate": 9.997844094488189e-06, "loss": 0.4831, "step": 1373 }, { "epoch": 0.43275590551181103, "grad_norm": 39.303409576416016, "learning_rate": 9.99784251968504e-06, "loss": 0.3898, "step": 1374 }, { "epoch": 0.4330708661417323, "grad_norm": 34.55439758300781, "learning_rate": 9.99784094488189e-06, "loss": 0.4709, "step": 1375 }, { "epoch": 0.43338582677165355, "grad_norm": 23.586498260498047, "learning_rate": 9.997839370078742e-06, "loss": 0.2142, "step": 1376 }, { "epoch": 0.4337007874015748, "grad_norm": 25.976821899414062, "learning_rate": 9.997837795275591e-06, "loss": 0.4925, "step": 1377 }, { "epoch": 0.43401574803149606, "grad_norm": 42.66455841064453, "learning_rate": 9.997836220472442e-06, "loss": 0.7777, "step": 1378 }, { "epoch": 0.4343307086614173, "grad_norm": 24.16051483154297, "learning_rate": 9.997834645669293e-06, "loss": 0.3197, "step": 1379 }, { "epoch": 0.4346456692913386, "grad_norm": 62.70100021362305, "learning_rate": 9.997833070866144e-06, "loss": 0.2733, "step": 1380 }, { "epoch": 0.4346456692913386, "eval_loss": 0.4844158887863159, "eval_runtime": 338.7196, "eval_samples_per_second": 0.345, "eval_steps_per_second": 0.345, "step": 1380 }, { "epoch": 0.43496062992125983, "grad_norm": 43.02194595336914, "learning_rate": 9.997831496062993e-06, "loss": 0.3504, "step": 1381 }, { "epoch": 0.4352755905511811, "grad_norm": 38.13003158569336, "learning_rate": 9.997829921259842e-06, "loss": 0.39, "step": 1382 }, { "epoch": 0.43559055118110235, "grad_norm": 65.61643981933594, "learning_rate": 9.997828346456693e-06, "loss": 0.7555, "step": 1383 }, { "epoch": 0.4359055118110236, "grad_norm": 31.994688034057617, "learning_rate": 9.997826771653544e-06, "loss": 0.4433, "step": 1384 }, { "epoch": 0.43622047244094486, "grad_norm": 39.22994613647461, "learning_rate": 9.997825196850395e-06, "loss": 0.235, "step": 1385 }, { "epoch": 0.4365354330708661, "grad_norm": 18.356678009033203, "learning_rate": 9.997823622047245e-06, "loss": 0.1587, "step": 1386 }, { "epoch": 0.4368503937007874, "grad_norm": 16.878463745117188, "learning_rate": 9.997822047244096e-06, "loss": 0.3613, "step": 1387 }, { "epoch": 0.4371653543307087, "grad_norm": 27.259780883789062, "learning_rate": 9.997820472440945e-06, "loss": 0.3243, "step": 1388 }, { "epoch": 0.43748031496062995, "grad_norm": 25.487789154052734, "learning_rate": 9.997818897637796e-06, "loss": 0.4375, "step": 1389 }, { "epoch": 0.4377952755905512, "grad_norm": 15.02742862701416, "learning_rate": 9.997817322834647e-06, "loss": 0.2624, "step": 1390 }, { "epoch": 0.43811023622047246, "grad_norm": 26.566652297973633, "learning_rate": 9.997815748031496e-06, "loss": 0.4211, "step": 1391 }, { "epoch": 0.4384251968503937, "grad_norm": 67.95926666259766, "learning_rate": 9.997814173228347e-06, "loss": 0.9252, "step": 1392 }, { "epoch": 0.438740157480315, "grad_norm": 29.216794967651367, "learning_rate": 9.997812598425197e-06, "loss": 0.4028, "step": 1393 }, { "epoch": 0.43905511811023623, "grad_norm": 50.51660919189453, "learning_rate": 9.997811023622048e-06, "loss": 0.4433, "step": 1394 }, { "epoch": 0.4393700787401575, "grad_norm": 13.636687278747559, "learning_rate": 9.997809448818899e-06, "loss": 0.2541, "step": 1395 }, { "epoch": 0.43968503937007875, "grad_norm": 26.405738830566406, "learning_rate": 9.99780787401575e-06, "loss": 0.3662, "step": 1396 }, { "epoch": 0.44, "grad_norm": 31.268871307373047, "learning_rate": 9.997806299212599e-06, "loss": 0.2522, "step": 1397 }, { "epoch": 0.44031496062992126, "grad_norm": 52.73464584350586, "learning_rate": 9.99780472440945e-06, "loss": 0.5821, "step": 1398 }, { "epoch": 0.4406299212598425, "grad_norm": 7.454155921936035, "learning_rate": 9.997803149606299e-06, "loss": 0.0584, "step": 1399 }, { "epoch": 0.4409448818897638, "grad_norm": 51.72380828857422, "learning_rate": 9.997801574803152e-06, "loss": 0.4622, "step": 1400 }, { "epoch": 0.4409448818897638, "eval_loss": 0.47116619348526, "eval_runtime": 337.7576, "eval_samples_per_second": 0.346, "eval_steps_per_second": 0.346, "step": 1400 }, { "epoch": 0.44125984251968503, "grad_norm": 10.400198936462402, "learning_rate": 9.997800000000001e-06, "loss": 0.0597, "step": 1401 }, { "epoch": 0.4415748031496063, "grad_norm": 48.547332763671875, "learning_rate": 9.99779842519685e-06, "loss": 0.3959, "step": 1402 }, { "epoch": 0.44188976377952754, "grad_norm": 23.85326385498047, "learning_rate": 9.997796850393701e-06, "loss": 0.2755, "step": 1403 }, { "epoch": 0.4422047244094488, "grad_norm": 25.128524780273438, "learning_rate": 9.997795275590552e-06, "loss": 0.0733, "step": 1404 }, { "epoch": 0.44251968503937006, "grad_norm": 51.35587692260742, "learning_rate": 9.997793700787403e-06, "loss": 0.6701, "step": 1405 }, { "epoch": 0.4428346456692913, "grad_norm": 14.554354667663574, "learning_rate": 9.997792125984253e-06, "loss": 0.064, "step": 1406 }, { "epoch": 0.4431496062992126, "grad_norm": 67.32514953613281, "learning_rate": 9.997790551181104e-06, "loss": 0.4264, "step": 1407 }, { "epoch": 0.4434645669291339, "grad_norm": 13.576302528381348, "learning_rate": 9.997788976377953e-06, "loss": 0.0385, "step": 1408 }, { "epoch": 0.44377952755905514, "grad_norm": 126.57298278808594, "learning_rate": 9.997787401574804e-06, "loss": 0.7388, "step": 1409 }, { "epoch": 0.4440944881889764, "grad_norm": 9.836527824401855, "learning_rate": 9.997785826771655e-06, "loss": 0.0383, "step": 1410 }, { "epoch": 0.44440944881889766, "grad_norm": 76.40158081054688, "learning_rate": 9.997784251968504e-06, "loss": 0.3938, "step": 1411 }, { "epoch": 0.4447244094488189, "grad_norm": 60.20785140991211, "learning_rate": 9.997782677165355e-06, "loss": 0.2053, "step": 1412 }, { "epoch": 0.44503937007874017, "grad_norm": 14.930899620056152, "learning_rate": 9.997781102362204e-06, "loss": 0.0554, "step": 1413 }, { "epoch": 0.44535433070866143, "grad_norm": 101.52273559570312, "learning_rate": 9.997779527559055e-06, "loss": 0.6495, "step": 1414 }, { "epoch": 0.4456692913385827, "grad_norm": 6.1531267166137695, "learning_rate": 9.997777952755906e-06, "loss": 0.0153, "step": 1415 }, { "epoch": 0.44598425196850394, "grad_norm": 55.18434143066406, "learning_rate": 9.997776377952757e-06, "loss": 0.5381, "step": 1416 }, { "epoch": 0.4462992125984252, "grad_norm": 5.298098564147949, "learning_rate": 9.997774803149607e-06, "loss": 0.0163, "step": 1417 }, { "epoch": 0.44661417322834646, "grad_norm": 33.11286163330078, "learning_rate": 9.997773228346458e-06, "loss": 0.051, "step": 1418 }, { "epoch": 0.4469291338582677, "grad_norm": 86.3932876586914, "learning_rate": 9.997771653543307e-06, "loss": 0.7798, "step": 1419 }, { "epoch": 0.44724409448818897, "grad_norm": 86.36060333251953, "learning_rate": 9.997770078740158e-06, "loss": 0.812, "step": 1420 }, { "epoch": 0.44724409448818897, "eval_loss": 0.7094002366065979, "eval_runtime": 338.8984, "eval_samples_per_second": 0.345, "eval_steps_per_second": 0.345, "step": 1420 }, { "epoch": 0.4475590551181102, "grad_norm": 85.274169921875, "learning_rate": 9.997768503937009e-06, "loss": 1.2007, "step": 1421 }, { "epoch": 0.4478740157480315, "grad_norm": 37.372406005859375, "learning_rate": 9.997766929133858e-06, "loss": 0.6024, "step": 1422 }, { "epoch": 0.44818897637795274, "grad_norm": 72.87973022460938, "learning_rate": 9.99776535433071e-06, "loss": 0.9482, "step": 1423 }, { "epoch": 0.448503937007874, "grad_norm": 19.026866912841797, "learning_rate": 9.99776377952756e-06, "loss": 0.0342, "step": 1424 }, { "epoch": 0.44881889763779526, "grad_norm": 124.86366271972656, "learning_rate": 9.997762204724411e-06, "loss": 1.1195, "step": 1425 }, { "epoch": 0.4491338582677165, "grad_norm": 91.7364273071289, "learning_rate": 9.99776062992126e-06, "loss": 0.6728, "step": 1426 }, { "epoch": 0.44944881889763777, "grad_norm": 199.09945678710938, "learning_rate": 9.997759055118112e-06, "loss": 0.1931, "step": 1427 }, { "epoch": 0.4497637795275591, "grad_norm": 88.02843475341797, "learning_rate": 9.997757480314961e-06, "loss": 0.8221, "step": 1428 }, { "epoch": 0.45007874015748034, "grad_norm": 50.00212860107422, "learning_rate": 9.997755905511812e-06, "loss": 0.4988, "step": 1429 }, { "epoch": 0.4503937007874016, "grad_norm": 46.531864166259766, "learning_rate": 9.997754330708663e-06, "loss": 0.166, "step": 1430 }, { "epoch": 0.45070866141732285, "grad_norm": 62.54853439331055, "learning_rate": 9.997752755905512e-06, "loss": 0.8291, "step": 1431 }, { "epoch": 0.4510236220472441, "grad_norm": 50.34440994262695, "learning_rate": 9.997751181102363e-06, "loss": 0.9383, "step": 1432 }, { "epoch": 0.45133858267716537, "grad_norm": 93.70585632324219, "learning_rate": 9.997749606299212e-06, "loss": 0.7288, "step": 1433 }, { "epoch": 0.4516535433070866, "grad_norm": 16.99148941040039, "learning_rate": 9.997748031496063e-06, "loss": 0.193, "step": 1434 }, { "epoch": 0.4519685039370079, "grad_norm": 51.41777420043945, "learning_rate": 9.997746456692914e-06, "loss": 0.2534, "step": 1435 }, { "epoch": 0.45228346456692914, "grad_norm": 60.7148323059082, "learning_rate": 9.997744881889765e-06, "loss": 0.824, "step": 1436 }, { "epoch": 0.4525984251968504, "grad_norm": 42.44105529785156, "learning_rate": 9.997743307086615e-06, "loss": 0.5518, "step": 1437 }, { "epoch": 0.45291338582677165, "grad_norm": 68.7177963256836, "learning_rate": 9.997741732283466e-06, "loss": 0.1653, "step": 1438 }, { "epoch": 0.4532283464566929, "grad_norm": 34.2302131652832, "learning_rate": 9.997740157480315e-06, "loss": 0.3412, "step": 1439 }, { "epoch": 0.45354330708661417, "grad_norm": 18.738046646118164, "learning_rate": 9.997738582677166e-06, "loss": 0.3157, "step": 1440 }, { "epoch": 0.45354330708661417, "eval_loss": 0.5591472387313843, "eval_runtime": 340.3548, "eval_samples_per_second": 0.344, "eval_steps_per_second": 0.344, "step": 1440 }, { "epoch": 0.4538582677165354, "grad_norm": 113.84444427490234, "learning_rate": 9.997737007874017e-06, "loss": 1.0174, "step": 1441 }, { "epoch": 0.4541732283464567, "grad_norm": 38.3621940612793, "learning_rate": 9.997735433070866e-06, "loss": 0.6811, "step": 1442 }, { "epoch": 0.45448818897637794, "grad_norm": 34.84929656982422, "learning_rate": 9.997733858267717e-06, "loss": 0.2798, "step": 1443 }, { "epoch": 0.4548031496062992, "grad_norm": 10.604235649108887, "learning_rate": 9.997732283464567e-06, "loss": 0.0995, "step": 1444 }, { "epoch": 0.45511811023622045, "grad_norm": 48.81785202026367, "learning_rate": 9.997730708661418e-06, "loss": 0.713, "step": 1445 }, { "epoch": 0.4554330708661417, "grad_norm": 22.666885375976562, "learning_rate": 9.997729133858269e-06, "loss": 0.5139, "step": 1446 }, { "epoch": 0.45574803149606297, "grad_norm": 12.92003059387207, "learning_rate": 9.99772755905512e-06, "loss": 0.1314, "step": 1447 }, { "epoch": 0.4560629921259843, "grad_norm": 29.922138214111328, "learning_rate": 9.997725984251969e-06, "loss": 0.4134, "step": 1448 }, { "epoch": 0.45637795275590554, "grad_norm": 47.216609954833984, "learning_rate": 9.99772440944882e-06, "loss": 1.0959, "step": 1449 }, { "epoch": 0.4566929133858268, "grad_norm": 20.693220138549805, "learning_rate": 9.99772283464567e-06, "loss": 0.1627, "step": 1450 }, { "epoch": 0.45700787401574805, "grad_norm": 27.217304229736328, "learning_rate": 9.99772125984252e-06, "loss": 0.2322, "step": 1451 }, { "epoch": 0.4573228346456693, "grad_norm": 49.21244430541992, "learning_rate": 9.997719685039371e-06, "loss": 0.2963, "step": 1452 }, { "epoch": 0.45763779527559056, "grad_norm": 29.46310043334961, "learning_rate": 9.99771811023622e-06, "loss": 0.3798, "step": 1453 }, { "epoch": 0.4579527559055118, "grad_norm": 30.430831909179688, "learning_rate": 9.997716535433071e-06, "loss": 0.3264, "step": 1454 }, { "epoch": 0.4582677165354331, "grad_norm": 61.625083923339844, "learning_rate": 9.997714960629922e-06, "loss": 1.0914, "step": 1455 }, { "epoch": 0.45858267716535434, "grad_norm": 28.40131187438965, "learning_rate": 9.997713385826773e-06, "loss": 0.1856, "step": 1456 }, { "epoch": 0.4588976377952756, "grad_norm": 30.0435733795166, "learning_rate": 9.997711811023623e-06, "loss": 0.5765, "step": 1457 }, { "epoch": 0.45921259842519685, "grad_norm": 15.2453031539917, "learning_rate": 9.997710236220474e-06, "loss": 0.287, "step": 1458 }, { "epoch": 0.4595275590551181, "grad_norm": 36.65151596069336, "learning_rate": 9.997708661417323e-06, "loss": 0.2213, "step": 1459 }, { "epoch": 0.45984251968503936, "grad_norm": 37.6453971862793, "learning_rate": 9.997707086614174e-06, "loss": 0.5465, "step": 1460 }, { "epoch": 0.45984251968503936, "eval_loss": 0.49336767196655273, "eval_runtime": 340.9663, "eval_samples_per_second": 0.343, "eval_steps_per_second": 0.343, "step": 1460 }, { "epoch": 0.4601574803149606, "grad_norm": 21.586259841918945, "learning_rate": 9.997705511811025e-06, "loss": 0.1797, "step": 1461 }, { "epoch": 0.4604724409448819, "grad_norm": 15.331527709960938, "learning_rate": 9.997703937007874e-06, "loss": 0.2192, "step": 1462 }, { "epoch": 0.46078740157480313, "grad_norm": 27.347505569458008, "learning_rate": 9.997702362204725e-06, "loss": 0.6414, "step": 1463 }, { "epoch": 0.4611023622047244, "grad_norm": 49.80681610107422, "learning_rate": 9.997700787401575e-06, "loss": 0.2621, "step": 1464 }, { "epoch": 0.46141732283464565, "grad_norm": 20.706789016723633, "learning_rate": 9.997699212598426e-06, "loss": 0.1548, "step": 1465 }, { "epoch": 0.4617322834645669, "grad_norm": 59.39773941040039, "learning_rate": 9.997697637795276e-06, "loss": 0.514, "step": 1466 }, { "epoch": 0.46204724409448816, "grad_norm": 29.0017147064209, "learning_rate": 9.997696062992127e-06, "loss": 0.4804, "step": 1467 }, { "epoch": 0.4623622047244095, "grad_norm": 59.08629608154297, "learning_rate": 9.997694488188977e-06, "loss": 0.6536, "step": 1468 }, { "epoch": 0.46267716535433073, "grad_norm": 39.09469223022461, "learning_rate": 9.997692913385828e-06, "loss": 0.3326, "step": 1469 }, { "epoch": 0.462992125984252, "grad_norm": 14.819294929504395, "learning_rate": 9.997691338582679e-06, "loss": 0.164, "step": 1470 }, { "epoch": 0.46330708661417325, "grad_norm": 23.74578094482422, "learning_rate": 9.997689763779528e-06, "loss": 0.3426, "step": 1471 }, { "epoch": 0.4636220472440945, "grad_norm": 31.888647079467773, "learning_rate": 9.997688188976379e-06, "loss": 0.4113, "step": 1472 }, { "epoch": 0.46393700787401576, "grad_norm": 42.581398010253906, "learning_rate": 9.997686614173228e-06, "loss": 0.3324, "step": 1473 }, { "epoch": 0.464251968503937, "grad_norm": 45.32304382324219, "learning_rate": 9.99768503937008e-06, "loss": 0.5497, "step": 1474 }, { "epoch": 0.4645669291338583, "grad_norm": 31.52800750732422, "learning_rate": 9.99768346456693e-06, "loss": 0.6702, "step": 1475 }, { "epoch": 0.46488188976377953, "grad_norm": 41.72283172607422, "learning_rate": 9.997681889763781e-06, "loss": 0.2619, "step": 1476 }, { "epoch": 0.4651968503937008, "grad_norm": 97.3839340209961, "learning_rate": 9.99768031496063e-06, "loss": 0.4892, "step": 1477 }, { "epoch": 0.46551181102362205, "grad_norm": 39.59928512573242, "learning_rate": 9.997678740157482e-06, "loss": 0.5622, "step": 1478 }, { "epoch": 0.4658267716535433, "grad_norm": 23.758737564086914, "learning_rate": 9.997677165354331e-06, "loss": 0.3074, "step": 1479 }, { "epoch": 0.46614173228346456, "grad_norm": 39.63565444946289, "learning_rate": 9.997675590551182e-06, "loss": 0.3154, "step": 1480 }, { "epoch": 0.46614173228346456, "eval_loss": 0.4589572548866272, "eval_runtime": 341.5074, "eval_samples_per_second": 0.343, "eval_steps_per_second": 0.343, "step": 1480 }, { "epoch": 0.4664566929133858, "grad_norm": 19.873315811157227, "learning_rate": 9.997674015748033e-06, "loss": 0.2133, "step": 1481 }, { "epoch": 0.4667716535433071, "grad_norm": 32.47504806518555, "learning_rate": 9.997672440944882e-06, "loss": 0.3143, "step": 1482 }, { "epoch": 0.46708661417322833, "grad_norm": 27.075977325439453, "learning_rate": 9.997670866141733e-06, "loss": 0.315, "step": 1483 }, { "epoch": 0.4674015748031496, "grad_norm": 18.573650360107422, "learning_rate": 9.997669291338582e-06, "loss": 0.1998, "step": 1484 }, { "epoch": 0.46771653543307085, "grad_norm": 19.674373626708984, "learning_rate": 9.997667716535433e-06, "loss": 0.2505, "step": 1485 }, { "epoch": 0.4680314960629921, "grad_norm": 40.249202728271484, "learning_rate": 9.997666141732284e-06, "loss": 0.4321, "step": 1486 }, { "epoch": 0.46834645669291336, "grad_norm": 12.53464412689209, "learning_rate": 9.997664566929135e-06, "loss": 0.069, "step": 1487 }, { "epoch": 0.4686614173228347, "grad_norm": 38.792728424072266, "learning_rate": 9.997662992125985e-06, "loss": 0.417, "step": 1488 }, { "epoch": 0.46897637795275593, "grad_norm": 56.2455940246582, "learning_rate": 9.997661417322836e-06, "loss": 0.7037, "step": 1489 }, { "epoch": 0.4692913385826772, "grad_norm": 47.736263275146484, "learning_rate": 9.997659842519685e-06, "loss": 0.2714, "step": 1490 }, { "epoch": 0.46960629921259844, "grad_norm": 75.22129821777344, "learning_rate": 9.997658267716536e-06, "loss": 0.5145, "step": 1491 }, { "epoch": 0.4699212598425197, "grad_norm": 21.2304744720459, "learning_rate": 9.997656692913387e-06, "loss": 0.2883, "step": 1492 }, { "epoch": 0.47023622047244096, "grad_norm": 62.4765625, "learning_rate": 9.997655118110236e-06, "loss": 0.4325, "step": 1493 }, { "epoch": 0.4705511811023622, "grad_norm": 49.23244094848633, "learning_rate": 9.997653543307087e-06, "loss": 0.6759, "step": 1494 }, { "epoch": 0.47086614173228347, "grad_norm": 21.292394638061523, "learning_rate": 9.997651968503938e-06, "loss": 0.1809, "step": 1495 }, { "epoch": 0.47118110236220473, "grad_norm": 57.66270446777344, "learning_rate": 9.99765039370079e-06, "loss": 0.7726, "step": 1496 }, { "epoch": 0.471496062992126, "grad_norm": 61.524253845214844, "learning_rate": 9.997648818897639e-06, "loss": 0.3385, "step": 1497 }, { "epoch": 0.47181102362204724, "grad_norm": 51.74467468261719, "learning_rate": 9.99764724409449e-06, "loss": 0.1262, "step": 1498 }, { "epoch": 0.4721259842519685, "grad_norm": 64.45372009277344, "learning_rate": 9.997645669291339e-06, "loss": 0.5856, "step": 1499 }, { "epoch": 0.47244094488188976, "grad_norm": 33.162811279296875, "learning_rate": 9.99764409448819e-06, "loss": 0.8524, "step": 1500 }, { "epoch": 0.47244094488188976, "eval_loss": 0.4843520522117615, "eval_runtime": 337.6077, "eval_samples_per_second": 0.347, "eval_steps_per_second": 0.347, "step": 1500 }, { "epoch": 0.472755905511811, "grad_norm": 58.66725158691406, "learning_rate": 9.99764251968504e-06, "loss": 0.7346, "step": 1501 }, { "epoch": 0.47307086614173227, "grad_norm": 113.04570770263672, "learning_rate": 9.99764094488189e-06, "loss": 0.4166, "step": 1502 }, { "epoch": 0.47338582677165353, "grad_norm": 49.267024993896484, "learning_rate": 9.997639370078741e-06, "loss": 0.8303, "step": 1503 }, { "epoch": 0.4737007874015748, "grad_norm": 18.804046630859375, "learning_rate": 9.99763779527559e-06, "loss": 0.2172, "step": 1504 }, { "epoch": 0.47401574803149604, "grad_norm": 23.38251495361328, "learning_rate": 9.997636220472441e-06, "loss": 0.1723, "step": 1505 }, { "epoch": 0.4743307086614173, "grad_norm": 32.252750396728516, "learning_rate": 9.997634645669292e-06, "loss": 0.1733, "step": 1506 }, { "epoch": 0.47464566929133856, "grad_norm": 58.86079025268555, "learning_rate": 9.997633070866143e-06, "loss": 0.8325, "step": 1507 }, { "epoch": 0.47496062992125987, "grad_norm": 18.70965003967285, "learning_rate": 9.997631496062993e-06, "loss": 0.1669, "step": 1508 }, { "epoch": 0.4752755905511811, "grad_norm": 31.96597671508789, "learning_rate": 9.997629921259844e-06, "loss": 0.3987, "step": 1509 }, { "epoch": 0.4755905511811024, "grad_norm": 21.83759880065918, "learning_rate": 9.997628346456693e-06, "loss": 0.1909, "step": 1510 }, { "epoch": 0.47590551181102364, "grad_norm": 27.163360595703125, "learning_rate": 9.997626771653544e-06, "loss": 0.3432, "step": 1511 }, { "epoch": 0.4762204724409449, "grad_norm": 81.91073608398438, "learning_rate": 9.997625196850395e-06, "loss": 0.4286, "step": 1512 }, { "epoch": 0.47653543307086615, "grad_norm": 23.85365867614746, "learning_rate": 9.997623622047244e-06, "loss": 0.2956, "step": 1513 }, { "epoch": 0.4768503937007874, "grad_norm": 20.397815704345703, "learning_rate": 9.997622047244095e-06, "loss": 0.4461, "step": 1514 }, { "epoch": 0.47716535433070867, "grad_norm": 30.885562896728516, "learning_rate": 9.997620472440945e-06, "loss": 0.477, "step": 1515 }, { "epoch": 0.4774803149606299, "grad_norm": 33.92880630493164, "learning_rate": 9.997618897637796e-06, "loss": 0.3881, "step": 1516 }, { "epoch": 0.4777952755905512, "grad_norm": 24.906478881835938, "learning_rate": 9.997617322834647e-06, "loss": 0.3944, "step": 1517 }, { "epoch": 0.47811023622047244, "grad_norm": 17.711050033569336, "learning_rate": 9.997615748031498e-06, "loss": 0.3327, "step": 1518 }, { "epoch": 0.4784251968503937, "grad_norm": 28.30522918701172, "learning_rate": 9.997614173228347e-06, "loss": 0.4556, "step": 1519 }, { "epoch": 0.47874015748031495, "grad_norm": 22.290746688842773, "learning_rate": 9.997612598425198e-06, "loss": 0.177, "step": 1520 }, { "epoch": 0.47874015748031495, "eval_loss": 0.45914146304130554, "eval_runtime": 337.9882, "eval_samples_per_second": 0.346, "eval_steps_per_second": 0.346, "step": 1520 }, { "epoch": 0.4790551181102362, "grad_norm": 51.145206451416016, "learning_rate": 9.997611023622049e-06, "loss": 0.2974, "step": 1521 }, { "epoch": 0.47937007874015747, "grad_norm": 53.45519256591797, "learning_rate": 9.997609448818898e-06, "loss": 0.2501, "step": 1522 }, { "epoch": 0.4796850393700787, "grad_norm": 41.410377502441406, "learning_rate": 9.997607874015749e-06, "loss": 0.5987, "step": 1523 }, { "epoch": 0.48, "grad_norm": 29.04597282409668, "learning_rate": 9.997606299212598e-06, "loss": 0.3196, "step": 1524 }, { "epoch": 0.48031496062992124, "grad_norm": 37.950008392333984, "learning_rate": 9.99760472440945e-06, "loss": 0.444, "step": 1525 }, { "epoch": 0.4806299212598425, "grad_norm": 47.16240692138672, "learning_rate": 9.9976031496063e-06, "loss": 0.6532, "step": 1526 }, { "epoch": 0.48094488188976375, "grad_norm": 33.97493362426758, "learning_rate": 9.997601574803151e-06, "loss": 0.4194, "step": 1527 }, { "epoch": 0.48125984251968507, "grad_norm": 36.6172981262207, "learning_rate": 9.9976e-06, "loss": 0.4505, "step": 1528 }, { "epoch": 0.4815748031496063, "grad_norm": 26.39623260498047, "learning_rate": 9.997598425196852e-06, "loss": 0.2039, "step": 1529 }, { "epoch": 0.4818897637795276, "grad_norm": 24.622024536132812, "learning_rate": 9.997596850393701e-06, "loss": 0.2808, "step": 1530 }, { "epoch": 0.48220472440944884, "grad_norm": 58.132904052734375, "learning_rate": 9.997595275590552e-06, "loss": 0.4782, "step": 1531 }, { "epoch": 0.4825196850393701, "grad_norm": 53.385154724121094, "learning_rate": 9.997593700787403e-06, "loss": 0.4064, "step": 1532 }, { "epoch": 0.48283464566929135, "grad_norm": 18.42597007751465, "learning_rate": 9.997592125984252e-06, "loss": 0.0884, "step": 1533 }, { "epoch": 0.4831496062992126, "grad_norm": 68.61125183105469, "learning_rate": 9.997590551181103e-06, "loss": 0.52, "step": 1534 }, { "epoch": 0.48346456692913387, "grad_norm": 19.761259078979492, "learning_rate": 9.997588976377952e-06, "loss": 0.1054, "step": 1535 }, { "epoch": 0.4837795275590551, "grad_norm": 20.919004440307617, "learning_rate": 9.997587401574803e-06, "loss": 0.2819, "step": 1536 }, { "epoch": 0.4840944881889764, "grad_norm": 24.11286735534668, "learning_rate": 9.997585826771654e-06, "loss": 0.2413, "step": 1537 }, { "epoch": 0.48440944881889764, "grad_norm": 34.09194564819336, "learning_rate": 9.997584251968505e-06, "loss": 0.2417, "step": 1538 }, { "epoch": 0.4847244094488189, "grad_norm": 53.1319580078125, "learning_rate": 9.997582677165355e-06, "loss": 0.5983, "step": 1539 }, { "epoch": 0.48503937007874015, "grad_norm": 23.620807647705078, "learning_rate": 9.997581102362206e-06, "loss": 0.1747, "step": 1540 }, { "epoch": 0.48503937007874015, "eval_loss": 0.453780859708786, "eval_runtime": 336.017, "eval_samples_per_second": 0.348, "eval_steps_per_second": 0.348, "step": 1540 }, { "epoch": 0.4853543307086614, "grad_norm": 45.581031799316406, "learning_rate": 9.997579527559057e-06, "loss": 0.4102, "step": 1541 }, { "epoch": 0.48566929133858266, "grad_norm": 49.25587463378906, "learning_rate": 9.997577952755906e-06, "loss": 0.71, "step": 1542 }, { "epoch": 0.4859842519685039, "grad_norm": 4.6258649826049805, "learning_rate": 9.997576377952757e-06, "loss": 0.0368, "step": 1543 }, { "epoch": 0.4862992125984252, "grad_norm": 17.629261016845703, "learning_rate": 9.997574803149606e-06, "loss": 0.2871, "step": 1544 }, { "epoch": 0.48661417322834644, "grad_norm": 53.029666900634766, "learning_rate": 9.997573228346457e-06, "loss": 0.4007, "step": 1545 }, { "epoch": 0.4869291338582677, "grad_norm": 25.097759246826172, "learning_rate": 9.997571653543308e-06, "loss": 0.2665, "step": 1546 }, { "epoch": 0.48724409448818895, "grad_norm": 59.07927703857422, "learning_rate": 9.99757007874016e-06, "loss": 0.8133, "step": 1547 }, { "epoch": 0.48755905511811026, "grad_norm": 39.4317741394043, "learning_rate": 9.997568503937009e-06, "loss": 0.1684, "step": 1548 }, { "epoch": 0.4878740157480315, "grad_norm": 61.61114501953125, "learning_rate": 9.99756692913386e-06, "loss": 0.5758, "step": 1549 }, { "epoch": 0.4881889763779528, "grad_norm": 57.046142578125, "learning_rate": 9.997565354330709e-06, "loss": 0.4657, "step": 1550 }, { "epoch": 0.48850393700787403, "grad_norm": 28.860565185546875, "learning_rate": 9.99756377952756e-06, "loss": 0.4151, "step": 1551 }, { "epoch": 0.4888188976377953, "grad_norm": 92.83087158203125, "learning_rate": 9.997562204724411e-06, "loss": 0.7444, "step": 1552 }, { "epoch": 0.48913385826771655, "grad_norm": 41.72923278808594, "learning_rate": 9.99756062992126e-06, "loss": 0.7972, "step": 1553 }, { "epoch": 0.4894488188976378, "grad_norm": 43.05347442626953, "learning_rate": 9.997559055118111e-06, "loss": 0.1623, "step": 1554 }, { "epoch": 0.48976377952755906, "grad_norm": 37.73609161376953, "learning_rate": 9.99755748031496e-06, "loss": 0.4308, "step": 1555 }, { "epoch": 0.4900787401574803, "grad_norm": 22.235315322875977, "learning_rate": 9.997555905511811e-06, "loss": 0.2682, "step": 1556 }, { "epoch": 0.4903937007874016, "grad_norm": 4.401269435882568, "learning_rate": 9.997554330708662e-06, "loss": 0.0164, "step": 1557 }, { "epoch": 0.49070866141732283, "grad_norm": 44.11044692993164, "learning_rate": 9.997552755905513e-06, "loss": 0.3743, "step": 1558 }, { "epoch": 0.4910236220472441, "grad_norm": 38.67685317993164, "learning_rate": 9.997551181102363e-06, "loss": 0.3507, "step": 1559 }, { "epoch": 0.49133858267716535, "grad_norm": 20.494264602661133, "learning_rate": 9.997549606299214e-06, "loss": 0.3759, "step": 1560 }, { "epoch": 0.49133858267716535, "eval_loss": 0.43347296118736267, "eval_runtime": 339.139, "eval_samples_per_second": 0.345, "eval_steps_per_second": 0.345, "step": 1560 }, { "epoch": 0.4916535433070866, "grad_norm": 21.03292465209961, "learning_rate": 9.997548031496063e-06, "loss": 0.2048, "step": 1561 }, { "epoch": 0.49196850393700786, "grad_norm": 40.08341979980469, "learning_rate": 9.997546456692914e-06, "loss": 0.1279, "step": 1562 }, { "epoch": 0.4922834645669291, "grad_norm": 31.835474014282227, "learning_rate": 9.997544881889765e-06, "loss": 0.26, "step": 1563 }, { "epoch": 0.4925984251968504, "grad_norm": 26.725603103637695, "learning_rate": 9.997543307086614e-06, "loss": 0.2091, "step": 1564 }, { "epoch": 0.49291338582677163, "grad_norm": 33.68645095825195, "learning_rate": 9.997541732283465e-06, "loss": 0.2875, "step": 1565 }, { "epoch": 0.4932283464566929, "grad_norm": 56.92742156982422, "learning_rate": 9.997540157480316e-06, "loss": 1.1372, "step": 1566 }, { "epoch": 0.49354330708661415, "grad_norm": 10.900672912597656, "learning_rate": 9.997538582677167e-06, "loss": 0.0537, "step": 1567 }, { "epoch": 0.49385826771653546, "grad_norm": 20.076473236083984, "learning_rate": 9.997537007874017e-06, "loss": 0.1373, "step": 1568 }, { "epoch": 0.4941732283464567, "grad_norm": 30.94587516784668, "learning_rate": 9.997535433070868e-06, "loss": 0.348, "step": 1569 }, { "epoch": 0.494488188976378, "grad_norm": 79.50629425048828, "learning_rate": 9.997533858267717e-06, "loss": 0.9603, "step": 1570 }, { "epoch": 0.49480314960629923, "grad_norm": 51.20693588256836, "learning_rate": 9.997532283464568e-06, "loss": 1.1457, "step": 1571 }, { "epoch": 0.4951181102362205, "grad_norm": 34.06742477416992, "learning_rate": 9.997530708661419e-06, "loss": 0.4457, "step": 1572 }, { "epoch": 0.49543307086614174, "grad_norm": 40.76358413696289, "learning_rate": 9.997529133858268e-06, "loss": 0.2998, "step": 1573 }, { "epoch": 0.495748031496063, "grad_norm": 23.580713272094727, "learning_rate": 9.997527559055119e-06, "loss": 0.1263, "step": 1574 }, { "epoch": 0.49606299212598426, "grad_norm": 6.127594470977783, "learning_rate": 9.997525984251968e-06, "loss": 0.0331, "step": 1575 }, { "epoch": 0.4963779527559055, "grad_norm": 28.260459899902344, "learning_rate": 9.99752440944882e-06, "loss": 0.3719, "step": 1576 }, { "epoch": 0.4966929133858268, "grad_norm": 68.38239288330078, "learning_rate": 9.99752283464567e-06, "loss": 0.6032, "step": 1577 }, { "epoch": 0.49700787401574803, "grad_norm": 60.19215393066406, "learning_rate": 9.997521259842521e-06, "loss": 0.2563, "step": 1578 }, { "epoch": 0.4973228346456693, "grad_norm": 40.08391571044922, "learning_rate": 9.99751968503937e-06, "loss": 0.2775, "step": 1579 }, { "epoch": 0.49763779527559054, "grad_norm": 37.87879180908203, "learning_rate": 9.997518110236222e-06, "loss": 0.117, "step": 1580 }, { "epoch": 0.49763779527559054, "eval_loss": 0.5100625157356262, "eval_runtime": 338.9567, "eval_samples_per_second": 0.345, "eval_steps_per_second": 0.345, "step": 1580 }, { "epoch": 0.4979527559055118, "grad_norm": 43.88140869140625, "learning_rate": 9.997516535433071e-06, "loss": 0.7505, "step": 1581 }, { "epoch": 0.49826771653543306, "grad_norm": 10.613367080688477, "learning_rate": 9.997514960629922e-06, "loss": 0.0331, "step": 1582 }, { "epoch": 0.4985826771653543, "grad_norm": 115.95816040039062, "learning_rate": 9.997513385826773e-06, "loss": 0.6034, "step": 1583 }, { "epoch": 0.4988976377952756, "grad_norm": 52.247474670410156, "learning_rate": 9.997511811023622e-06, "loss": 0.7637, "step": 1584 }, { "epoch": 0.49921259842519683, "grad_norm": 49.17156982421875, "learning_rate": 9.997510236220473e-06, "loss": 0.2622, "step": 1585 }, { "epoch": 0.4995275590551181, "grad_norm": 52.29065704345703, "learning_rate": 9.997508661417323e-06, "loss": 0.2516, "step": 1586 }, { "epoch": 0.49984251968503934, "grad_norm": 4.530742168426514, "learning_rate": 9.997507086614175e-06, "loss": 0.039, "step": 1587 }, { "epoch": 0.5001574803149607, "grad_norm": 54.83158874511719, "learning_rate": 9.997505511811024e-06, "loss": 1.2184, "step": 1588 }, { "epoch": 0.5004724409448819, "grad_norm": 80.9505615234375, "learning_rate": 9.997503937007875e-06, "loss": 0.6721, "step": 1589 }, { "epoch": 0.5007874015748032, "grad_norm": 88.2104263305664, "learning_rate": 9.997502362204725e-06, "loss": 0.8541, "step": 1590 }, { "epoch": 0.5011023622047244, "grad_norm": 54.88927459716797, "learning_rate": 9.997500787401576e-06, "loss": 0.7247, "step": 1591 }, { "epoch": 0.5014173228346457, "grad_norm": 58.55228805541992, "learning_rate": 9.997499212598427e-06, "loss": 0.6949, "step": 1592 }, { "epoch": 0.5017322834645669, "grad_norm": 70.87881469726562, "learning_rate": 9.997497637795276e-06, "loss": 0.5249, "step": 1593 }, { "epoch": 0.5020472440944882, "grad_norm": 49.13249206542969, "learning_rate": 9.997496062992127e-06, "loss": 0.3263, "step": 1594 }, { "epoch": 0.5023622047244094, "grad_norm": 44.89517593383789, "learning_rate": 9.997494488188976e-06, "loss": 0.2266, "step": 1595 }, { "epoch": 0.5026771653543307, "grad_norm": 34.858665466308594, "learning_rate": 9.997492913385827e-06, "loss": 0.3243, "step": 1596 }, { "epoch": 0.5029921259842519, "grad_norm": 48.156105041503906, "learning_rate": 9.997491338582678e-06, "loss": 0.6439, "step": 1597 }, { "epoch": 0.5033070866141732, "grad_norm": 23.67864418029785, "learning_rate": 9.99748976377953e-06, "loss": 0.4013, "step": 1598 }, { "epoch": 0.5036220472440945, "grad_norm": 33.883583068847656, "learning_rate": 9.997488188976379e-06, "loss": 0.3123, "step": 1599 }, { "epoch": 0.5039370078740157, "grad_norm": 14.358415603637695, "learning_rate": 9.99748661417323e-06, "loss": 0.1317, "step": 1600 }, { "epoch": 0.5039370078740157, "eval_loss": 0.46397241950035095, "eval_runtime": 338.8895, "eval_samples_per_second": 0.345, "eval_steps_per_second": 0.345, "step": 1600 }, { "epoch": 0.504251968503937, "grad_norm": 29.868349075317383, "learning_rate": 9.997485039370079e-06, "loss": 0.2516, "step": 1601 }, { "epoch": 0.5045669291338583, "grad_norm": 21.991357803344727, "learning_rate": 9.99748346456693e-06, "loss": 0.2228, "step": 1602 }, { "epoch": 0.5048818897637796, "grad_norm": 15.654339790344238, "learning_rate": 9.997481889763781e-06, "loss": 0.1345, "step": 1603 }, { "epoch": 0.5051968503937008, "grad_norm": 15.310891151428223, "learning_rate": 9.99748031496063e-06, "loss": 0.1679, "step": 1604 }, { "epoch": 0.5055118110236221, "grad_norm": 38.296146392822266, "learning_rate": 9.997478740157481e-06, "loss": 0.3338, "step": 1605 }, { "epoch": 0.5058267716535433, "grad_norm": 43.396427154541016, "learning_rate": 9.99747716535433e-06, "loss": 0.4996, "step": 1606 }, { "epoch": 0.5061417322834646, "grad_norm": 94.86878204345703, "learning_rate": 9.997475590551181e-06, "loss": 0.8225, "step": 1607 }, { "epoch": 0.5064566929133858, "grad_norm": 52.50116729736328, "learning_rate": 9.997474015748032e-06, "loss": 0.3828, "step": 1608 }, { "epoch": 0.5067716535433071, "grad_norm": 38.4481086730957, "learning_rate": 9.997472440944883e-06, "loss": 0.1611, "step": 1609 }, { "epoch": 0.5070866141732283, "grad_norm": 49.7927131652832, "learning_rate": 9.997470866141733e-06, "loss": 1.0682, "step": 1610 }, { "epoch": 0.5074015748031496, "grad_norm": 8.284367561340332, "learning_rate": 9.997469291338584e-06, "loss": 0.0571, "step": 1611 }, { "epoch": 0.5077165354330708, "grad_norm": 24.673439025878906, "learning_rate": 9.997467716535435e-06, "loss": 0.1879, "step": 1612 }, { "epoch": 0.5080314960629921, "grad_norm": 33.2513542175293, "learning_rate": 9.997466141732284e-06, "loss": 0.636, "step": 1613 }, { "epoch": 0.5083464566929133, "grad_norm": 62.058738708496094, "learning_rate": 9.997464566929135e-06, "loss": 0.878, "step": 1614 }, { "epoch": 0.5086614173228347, "grad_norm": 79.89865112304688, "learning_rate": 9.997462992125984e-06, "loss": 0.2207, "step": 1615 }, { "epoch": 0.5089763779527559, "grad_norm": 37.931819915771484, "learning_rate": 9.997461417322835e-06, "loss": 0.1365, "step": 1616 }, { "epoch": 0.5092913385826772, "grad_norm": 30.21257209777832, "learning_rate": 9.997459842519686e-06, "loss": 0.4815, "step": 1617 }, { "epoch": 0.5096062992125985, "grad_norm": 35.088890075683594, "learning_rate": 9.997458267716537e-06, "loss": 0.2641, "step": 1618 }, { "epoch": 0.5099212598425197, "grad_norm": 31.518869400024414, "learning_rate": 9.997456692913387e-06, "loss": 0.3096, "step": 1619 }, { "epoch": 0.510236220472441, "grad_norm": 43.287391662597656, "learning_rate": 9.997455118110238e-06, "loss": 0.4844, "step": 1620 }, { "epoch": 0.510236220472441, "eval_loss": 0.5515217781066895, "eval_runtime": 339.8588, "eval_samples_per_second": 0.344, "eval_steps_per_second": 0.344, "step": 1620 }, { "epoch": 0.5105511811023622, "grad_norm": 20.773025512695312, "learning_rate": 9.997453543307087e-06, "loss": 0.0673, "step": 1621 }, { "epoch": 0.5108661417322835, "grad_norm": 22.52477264404297, "learning_rate": 9.997451968503938e-06, "loss": 0.2089, "step": 1622 }, { "epoch": 0.5111811023622047, "grad_norm": 21.807056427001953, "learning_rate": 9.997450393700789e-06, "loss": 0.1542, "step": 1623 }, { "epoch": 0.511496062992126, "grad_norm": 60.92094802856445, "learning_rate": 9.997448818897638e-06, "loss": 0.9874, "step": 1624 }, { "epoch": 0.5118110236220472, "grad_norm": 100.76142120361328, "learning_rate": 9.997447244094489e-06, "loss": 0.3835, "step": 1625 }, { "epoch": 0.5121259842519685, "grad_norm": 36.08515167236328, "learning_rate": 9.997445669291338e-06, "loss": 0.3638, "step": 1626 }, { "epoch": 0.5124409448818897, "grad_norm": 36.939170837402344, "learning_rate": 9.99744409448819e-06, "loss": 0.7436, "step": 1627 }, { "epoch": 0.512755905511811, "grad_norm": 48.17253494262695, "learning_rate": 9.99744251968504e-06, "loss": 0.4945, "step": 1628 }, { "epoch": 0.5130708661417323, "grad_norm": 79.20890808105469, "learning_rate": 9.997440944881891e-06, "loss": 0.8642, "step": 1629 }, { "epoch": 0.5133858267716536, "grad_norm": 19.723230361938477, "learning_rate": 9.99743937007874e-06, "loss": 0.1813, "step": 1630 }, { "epoch": 0.5137007874015748, "grad_norm": 39.59589385986328, "learning_rate": 9.997437795275592e-06, "loss": 0.3752, "step": 1631 }, { "epoch": 0.5140157480314961, "grad_norm": 77.0748062133789, "learning_rate": 9.997436220472441e-06, "loss": 0.6325, "step": 1632 }, { "epoch": 0.5143307086614173, "grad_norm": 20.3349552154541, "learning_rate": 9.997434645669292e-06, "loss": 0.2886, "step": 1633 }, { "epoch": 0.5146456692913386, "grad_norm": 46.71043014526367, "learning_rate": 9.997433070866143e-06, "loss": 0.9402, "step": 1634 }, { "epoch": 0.5149606299212598, "grad_norm": 34.89626693725586, "learning_rate": 9.997431496062992e-06, "loss": 0.4485, "step": 1635 }, { "epoch": 0.5152755905511811, "grad_norm": 28.397123336791992, "learning_rate": 9.997429921259843e-06, "loss": 0.3992, "step": 1636 }, { "epoch": 0.5155905511811023, "grad_norm": 50.874427795410156, "learning_rate": 9.997428346456694e-06, "loss": 0.9479, "step": 1637 }, { "epoch": 0.5159055118110236, "grad_norm": 38.96868896484375, "learning_rate": 9.997426771653545e-06, "loss": 0.5054, "step": 1638 }, { "epoch": 0.5162204724409449, "grad_norm": 7.598008155822754, "learning_rate": 9.997425196850395e-06, "loss": 0.0787, "step": 1639 }, { "epoch": 0.5165354330708661, "grad_norm": 24.391128540039062, "learning_rate": 9.997423622047246e-06, "loss": 0.1297, "step": 1640 }, { "epoch": 0.5165354330708661, "eval_loss": 0.4850241541862488, "eval_runtime": 337.9062, "eval_samples_per_second": 0.346, "eval_steps_per_second": 0.346, "step": 1640 }, { "epoch": 0.5168503937007874, "grad_norm": 21.02195167541504, "learning_rate": 9.997422047244095e-06, "loss": 0.3174, "step": 1641 }, { "epoch": 0.5171653543307086, "grad_norm": 19.174169540405273, "learning_rate": 9.997420472440946e-06, "loss": 0.4078, "step": 1642 }, { "epoch": 0.51748031496063, "grad_norm": 35.037322998046875, "learning_rate": 9.997418897637797e-06, "loss": 0.3795, "step": 1643 }, { "epoch": 0.5177952755905512, "grad_norm": 19.232954025268555, "learning_rate": 9.997417322834646e-06, "loss": 0.4354, "step": 1644 }, { "epoch": 0.5181102362204725, "grad_norm": 41.107784271240234, "learning_rate": 9.997415748031497e-06, "loss": 0.1806, "step": 1645 }, { "epoch": 0.5184251968503937, "grad_norm": 31.30357551574707, "learning_rate": 9.997414173228346e-06, "loss": 0.5762, "step": 1646 }, { "epoch": 0.518740157480315, "grad_norm": 27.336881637573242, "learning_rate": 9.997412598425197e-06, "loss": 0.5801, "step": 1647 }, { "epoch": 0.5190551181102362, "grad_norm": 43.55338668823242, "learning_rate": 9.997411023622048e-06, "loss": 0.9297, "step": 1648 }, { "epoch": 0.5193700787401575, "grad_norm": 33.737937927246094, "learning_rate": 9.9974094488189e-06, "loss": 0.3568, "step": 1649 }, { "epoch": 0.5196850393700787, "grad_norm": 24.022850036621094, "learning_rate": 9.997407874015749e-06, "loss": 0.1614, "step": 1650 }, { "epoch": 0.52, "grad_norm": 27.58238410949707, "learning_rate": 9.9974062992126e-06, "loss": 0.3589, "step": 1651 }, { "epoch": 0.5203149606299212, "grad_norm": 41.2049446105957, "learning_rate": 9.997404724409449e-06, "loss": 0.2244, "step": 1652 }, { "epoch": 0.5206299212598425, "grad_norm": 19.164548873901367, "learning_rate": 9.9974031496063e-06, "loss": 0.3534, "step": 1653 }, { "epoch": 0.5209448818897637, "grad_norm": 24.569271087646484, "learning_rate": 9.997401574803151e-06, "loss": 0.3207, "step": 1654 }, { "epoch": 0.521259842519685, "grad_norm": 24.17620849609375, "learning_rate": 9.9974e-06, "loss": 0.2762, "step": 1655 }, { "epoch": 0.5215748031496062, "grad_norm": 45.885250091552734, "learning_rate": 9.997398425196851e-06, "loss": 0.1903, "step": 1656 }, { "epoch": 0.5218897637795276, "grad_norm": 10.778152465820312, "learning_rate": 9.997396850393702e-06, "loss": 0.0588, "step": 1657 }, { "epoch": 0.5222047244094489, "grad_norm": 36.7594108581543, "learning_rate": 9.997395275590553e-06, "loss": 0.2227, "step": 1658 }, { "epoch": 0.5225196850393701, "grad_norm": 13.830334663391113, "learning_rate": 9.997393700787402e-06, "loss": 0.2058, "step": 1659 }, { "epoch": 0.5228346456692914, "grad_norm": 70.07479095458984, "learning_rate": 9.997392125984253e-06, "loss": 0.3415, "step": 1660 }, { "epoch": 0.5228346456692914, "eval_loss": 0.4270039498806, "eval_runtime": 338.5531, "eval_samples_per_second": 0.346, "eval_steps_per_second": 0.346, "step": 1660 }, { "epoch": 0.5231496062992126, "grad_norm": 54.377689361572266, "learning_rate": 9.997390551181103e-06, "loss": 0.699, "step": 1661 }, { "epoch": 0.5234645669291339, "grad_norm": 20.332332611083984, "learning_rate": 9.997388976377954e-06, "loss": 0.3042, "step": 1662 }, { "epoch": 0.5237795275590551, "grad_norm": 29.605314254760742, "learning_rate": 9.997387401574805e-06, "loss": 0.616, "step": 1663 }, { "epoch": 0.5240944881889764, "grad_norm": 26.555742263793945, "learning_rate": 9.997385826771654e-06, "loss": 0.2776, "step": 1664 }, { "epoch": 0.5244094488188976, "grad_norm": 24.46210479736328, "learning_rate": 9.997384251968505e-06, "loss": 0.3006, "step": 1665 }, { "epoch": 0.5247244094488189, "grad_norm": 135.4957275390625, "learning_rate": 9.997382677165354e-06, "loss": 0.4272, "step": 1666 }, { "epoch": 0.5250393700787401, "grad_norm": 8.375167846679688, "learning_rate": 9.997381102362205e-06, "loss": 0.0964, "step": 1667 }, { "epoch": 0.5253543307086614, "grad_norm": 17.8012752532959, "learning_rate": 9.997379527559056e-06, "loss": 0.1773, "step": 1668 }, { "epoch": 0.5256692913385826, "grad_norm": 43.08966064453125, "learning_rate": 9.997377952755907e-06, "loss": 0.3974, "step": 1669 }, { "epoch": 0.525984251968504, "grad_norm": 34.40397644042969, "learning_rate": 9.997376377952757e-06, "loss": 0.3196, "step": 1670 }, { "epoch": 0.5262992125984252, "grad_norm": 16.06707000732422, "learning_rate": 9.997374803149608e-06, "loss": 0.1063, "step": 1671 }, { "epoch": 0.5266141732283465, "grad_norm": 12.523601531982422, "learning_rate": 9.997373228346457e-06, "loss": 0.1203, "step": 1672 }, { "epoch": 0.5269291338582677, "grad_norm": 43.67927169799805, "learning_rate": 9.997371653543308e-06, "loss": 0.285, "step": 1673 }, { "epoch": 0.527244094488189, "grad_norm": 58.05824279785156, "learning_rate": 9.997370078740159e-06, "loss": 0.2903, "step": 1674 }, { "epoch": 0.5275590551181102, "grad_norm": 52.33161163330078, "learning_rate": 9.997368503937008e-06, "loss": 0.3971, "step": 1675 }, { "epoch": 0.5278740157480315, "grad_norm": 43.415462493896484, "learning_rate": 9.997366929133859e-06, "loss": 0.2529, "step": 1676 }, { "epoch": 0.5281889763779527, "grad_norm": 37.16794967651367, "learning_rate": 9.997365354330708e-06, "loss": 0.6718, "step": 1677 }, { "epoch": 0.528503937007874, "grad_norm": 19.802600860595703, "learning_rate": 9.99736377952756e-06, "loss": 0.1329, "step": 1678 }, { "epoch": 0.5288188976377953, "grad_norm": 18.722341537475586, "learning_rate": 9.99736220472441e-06, "loss": 0.0449, "step": 1679 }, { "epoch": 0.5291338582677165, "grad_norm": 30.56394386291504, "learning_rate": 9.997360629921261e-06, "loss": 0.0851, "step": 1680 }, { "epoch": 0.5291338582677165, "eval_loss": 0.5506137013435364, "eval_runtime": 340.4385, "eval_samples_per_second": 0.344, "eval_steps_per_second": 0.344, "step": 1680 }, { "epoch": 0.5294488188976378, "grad_norm": 15.041778564453125, "learning_rate": 9.99735905511811e-06, "loss": 0.0869, "step": 1681 }, { "epoch": 0.529763779527559, "grad_norm": 81.58389282226562, "learning_rate": 9.997357480314962e-06, "loss": 0.6074, "step": 1682 }, { "epoch": 0.5300787401574804, "grad_norm": 99.90658569335938, "learning_rate": 9.997355905511813e-06, "loss": 0.2411, "step": 1683 }, { "epoch": 0.5303937007874016, "grad_norm": 25.98404312133789, "learning_rate": 9.997354330708662e-06, "loss": 0.1055, "step": 1684 }, { "epoch": 0.5307086614173229, "grad_norm": 34.40901184082031, "learning_rate": 9.997352755905513e-06, "loss": 0.4056, "step": 1685 }, { "epoch": 0.5310236220472441, "grad_norm": 12.796487808227539, "learning_rate": 9.997351181102362e-06, "loss": 0.0303, "step": 1686 }, { "epoch": 0.5313385826771654, "grad_norm": 14.561634063720703, "learning_rate": 9.997349606299213e-06, "loss": 0.0729, "step": 1687 }, { "epoch": 0.5316535433070866, "grad_norm": 38.87438201904297, "learning_rate": 9.997348031496064e-06, "loss": 0.1667, "step": 1688 }, { "epoch": 0.5319685039370079, "grad_norm": 39.482303619384766, "learning_rate": 9.997346456692915e-06, "loss": 0.0675, "step": 1689 }, { "epoch": 0.5322834645669291, "grad_norm": 162.6251220703125, "learning_rate": 9.997344881889765e-06, "loss": 0.5522, "step": 1690 }, { "epoch": 0.5325984251968504, "grad_norm": 10.552274703979492, "learning_rate": 9.997343307086616e-06, "loss": 0.036, "step": 1691 }, { "epoch": 0.5329133858267716, "grad_norm": 118.2337646484375, "learning_rate": 9.997341732283465e-06, "loss": 0.1975, "step": 1692 }, { "epoch": 0.5332283464566929, "grad_norm": 46.612266540527344, "learning_rate": 9.997340157480316e-06, "loss": 0.1823, "step": 1693 }, { "epoch": 0.5335433070866141, "grad_norm": 93.66093444824219, "learning_rate": 9.997338582677167e-06, "loss": 1.0047, "step": 1694 }, { "epoch": 0.5338582677165354, "grad_norm": 42.73204040527344, "learning_rate": 9.997337007874016e-06, "loss": 0.1202, "step": 1695 }, { "epoch": 0.5341732283464566, "grad_norm": 116.17491912841797, "learning_rate": 9.997335433070867e-06, "loss": 0.5415, "step": 1696 }, { "epoch": 0.534488188976378, "grad_norm": 38.49724197387695, "learning_rate": 9.997333858267716e-06, "loss": 0.7654, "step": 1697 }, { "epoch": 0.5348031496062993, "grad_norm": 76.99092864990234, "learning_rate": 9.997332283464567e-06, "loss": 0.3025, "step": 1698 }, { "epoch": 0.5351181102362205, "grad_norm": 102.11122131347656, "learning_rate": 9.997330708661418e-06, "loss": 0.6456, "step": 1699 }, { "epoch": 0.5354330708661418, "grad_norm": 140.7232666015625, "learning_rate": 9.99732913385827e-06, "loss": 0.4218, "step": 1700 }, { "epoch": 0.5354330708661418, "eval_loss": 0.5890030860900879, "eval_runtime": 338.8108, "eval_samples_per_second": 0.345, "eval_steps_per_second": 0.345, "step": 1700 }, { "epoch": 0.535748031496063, "grad_norm": 111.65464782714844, "learning_rate": 9.997327559055119e-06, "loss": 1.0181, "step": 1701 }, { "epoch": 0.5360629921259843, "grad_norm": 37.92324447631836, "learning_rate": 9.99732598425197e-06, "loss": 0.7354, "step": 1702 }, { "epoch": 0.5363779527559055, "grad_norm": 42.922149658203125, "learning_rate": 9.997324409448819e-06, "loss": 0.2447, "step": 1703 }, { "epoch": 0.5366929133858268, "grad_norm": 45.183082580566406, "learning_rate": 9.99732283464567e-06, "loss": 0.337, "step": 1704 }, { "epoch": 0.537007874015748, "grad_norm": 39.247962951660156, "learning_rate": 9.997321259842521e-06, "loss": 0.3811, "step": 1705 }, { "epoch": 0.5373228346456693, "grad_norm": 128.93247985839844, "learning_rate": 9.99731968503937e-06, "loss": 0.5403, "step": 1706 }, { "epoch": 0.5376377952755905, "grad_norm": 107.43035888671875, "learning_rate": 9.997318110236221e-06, "loss": 0.3911, "step": 1707 }, { "epoch": 0.5379527559055118, "grad_norm": 43.16225051879883, "learning_rate": 9.997316535433072e-06, "loss": 0.2933, "step": 1708 }, { "epoch": 0.538267716535433, "grad_norm": 87.62896728515625, "learning_rate": 9.997314960629923e-06, "loss": 0.4866, "step": 1709 }, { "epoch": 0.5385826771653544, "grad_norm": 40.5244255065918, "learning_rate": 9.997313385826772e-06, "loss": 0.382, "step": 1710 }, { "epoch": 0.5388976377952756, "grad_norm": 34.54233932495117, "learning_rate": 9.997311811023623e-06, "loss": 0.2301, "step": 1711 }, { "epoch": 0.5392125984251969, "grad_norm": 95.98748016357422, "learning_rate": 9.997310236220473e-06, "loss": 0.8653, "step": 1712 }, { "epoch": 0.5395275590551181, "grad_norm": 86.79911041259766, "learning_rate": 9.997308661417324e-06, "loss": 0.4673, "step": 1713 }, { "epoch": 0.5398425196850394, "grad_norm": 13.791953086853027, "learning_rate": 9.997307086614175e-06, "loss": 0.0652, "step": 1714 }, { "epoch": 0.5401574803149606, "grad_norm": 16.453601837158203, "learning_rate": 9.997305511811024e-06, "loss": 0.103, "step": 1715 }, { "epoch": 0.5404724409448819, "grad_norm": 40.83580780029297, "learning_rate": 9.997303937007875e-06, "loss": 0.4618, "step": 1716 }, { "epoch": 0.5407874015748031, "grad_norm": 57.49140548706055, "learning_rate": 9.997302362204724e-06, "loss": 0.2785, "step": 1717 }, { "epoch": 0.5411023622047244, "grad_norm": 56.28849792480469, "learning_rate": 9.997300787401575e-06, "loss": 0.6193, "step": 1718 }, { "epoch": 0.5414173228346457, "grad_norm": 37.70257568359375, "learning_rate": 9.997299212598426e-06, "loss": 0.2423, "step": 1719 }, { "epoch": 0.5417322834645669, "grad_norm": 60.47585678100586, "learning_rate": 9.997297637795277e-06, "loss": 0.3203, "step": 1720 }, { "epoch": 0.5417322834645669, "eval_loss": 0.4833095669746399, "eval_runtime": 339.1777, "eval_samples_per_second": 0.345, "eval_steps_per_second": 0.345, "step": 1720 }, { "epoch": 0.5420472440944882, "grad_norm": 45.25422286987305, "learning_rate": 9.997296062992127e-06, "loss": 0.4233, "step": 1721 }, { "epoch": 0.5423622047244094, "grad_norm": 21.047258377075195, "learning_rate": 9.997294488188978e-06, "loss": 0.1222, "step": 1722 }, { "epoch": 0.5426771653543307, "grad_norm": 41.04436492919922, "learning_rate": 9.997292913385827e-06, "loss": 0.1144, "step": 1723 }, { "epoch": 0.542992125984252, "grad_norm": 9.362834930419922, "learning_rate": 9.997291338582678e-06, "loss": 0.0498, "step": 1724 }, { "epoch": 0.5433070866141733, "grad_norm": 50.78684997558594, "learning_rate": 9.997289763779529e-06, "loss": 0.6966, "step": 1725 }, { "epoch": 0.5436220472440945, "grad_norm": 33.23752212524414, "learning_rate": 9.997288188976378e-06, "loss": 0.1696, "step": 1726 }, { "epoch": 0.5439370078740158, "grad_norm": 29.13566780090332, "learning_rate": 9.99728661417323e-06, "loss": 0.511, "step": 1727 }, { "epoch": 0.544251968503937, "grad_norm": 36.36374282836914, "learning_rate": 9.99728503937008e-06, "loss": 0.3232, "step": 1728 }, { "epoch": 0.5445669291338583, "grad_norm": 9.268793106079102, "learning_rate": 9.997283464566931e-06, "loss": 0.0484, "step": 1729 }, { "epoch": 0.5448818897637795, "grad_norm": 10.723762512207031, "learning_rate": 9.99728188976378e-06, "loss": 0.0748, "step": 1730 }, { "epoch": 0.5451968503937008, "grad_norm": 42.956424713134766, "learning_rate": 9.997280314960631e-06, "loss": 0.2678, "step": 1731 }, { "epoch": 0.545511811023622, "grad_norm": 33.43272018432617, "learning_rate": 9.99727874015748e-06, "loss": 0.4772, "step": 1732 }, { "epoch": 0.5458267716535433, "grad_norm": 59.14657974243164, "learning_rate": 9.997277165354332e-06, "loss": 0.2235, "step": 1733 }, { "epoch": 0.5461417322834645, "grad_norm": 29.12813377380371, "learning_rate": 9.997275590551183e-06, "loss": 0.0931, "step": 1734 }, { "epoch": 0.5464566929133858, "grad_norm": 17.903165817260742, "learning_rate": 9.997274015748032e-06, "loss": 0.4091, "step": 1735 }, { "epoch": 0.546771653543307, "grad_norm": 72.72879028320312, "learning_rate": 9.997272440944883e-06, "loss": 0.3515, "step": 1736 }, { "epoch": 0.5470866141732283, "grad_norm": 42.820655822753906, "learning_rate": 9.997270866141732e-06, "loss": 0.8093, "step": 1737 }, { "epoch": 0.5474015748031497, "grad_norm": 34.83147048950195, "learning_rate": 9.997269291338583e-06, "loss": 0.0899, "step": 1738 }, { "epoch": 0.5477165354330709, "grad_norm": 165.32769775390625, "learning_rate": 9.997267716535434e-06, "loss": 1.2147, "step": 1739 }, { "epoch": 0.5480314960629922, "grad_norm": 54.89258575439453, "learning_rate": 9.997266141732285e-06, "loss": 0.3279, "step": 1740 }, { "epoch": 0.5480314960629922, "eval_loss": 0.5086050033569336, "eval_runtime": 340.7446, "eval_samples_per_second": 0.343, "eval_steps_per_second": 0.343, "step": 1740 }, { "epoch": 0.5483464566929134, "grad_norm": 19.018238067626953, "learning_rate": 9.997264566929135e-06, "loss": 0.0628, "step": 1741 }, { "epoch": 0.5486614173228347, "grad_norm": 44.377803802490234, "learning_rate": 9.997262992125986e-06, "loss": 0.1875, "step": 1742 }, { "epoch": 0.5489763779527559, "grad_norm": 51.6882209777832, "learning_rate": 9.997261417322835e-06, "loss": 0.5064, "step": 1743 }, { "epoch": 0.5492913385826772, "grad_norm": 6.151045799255371, "learning_rate": 9.997259842519686e-06, "loss": 0.0298, "step": 1744 }, { "epoch": 0.5496062992125984, "grad_norm": 76.562255859375, "learning_rate": 9.997258267716537e-06, "loss": 0.6654, "step": 1745 }, { "epoch": 0.5499212598425197, "grad_norm": 48.19951629638672, "learning_rate": 9.997256692913386e-06, "loss": 0.4798, "step": 1746 }, { "epoch": 0.5502362204724409, "grad_norm": 68.81988525390625, "learning_rate": 9.997255118110237e-06, "loss": 0.4747, "step": 1747 }, { "epoch": 0.5505511811023622, "grad_norm": 24.68579864501953, "learning_rate": 9.997253543307086e-06, "loss": 0.1881, "step": 1748 }, { "epoch": 0.5508661417322834, "grad_norm": 21.575008392333984, "learning_rate": 9.997251968503937e-06, "loss": 0.3208, "step": 1749 }, { "epoch": 0.5511811023622047, "grad_norm": 45.560203552246094, "learning_rate": 9.997250393700788e-06, "loss": 0.7353, "step": 1750 }, { "epoch": 0.551496062992126, "grad_norm": 11.304056167602539, "learning_rate": 9.99724881889764e-06, "loss": 0.0611, "step": 1751 }, { "epoch": 0.5518110236220473, "grad_norm": 53.74604797363281, "learning_rate": 9.997247244094489e-06, "loss": 0.3013, "step": 1752 }, { "epoch": 0.5521259842519685, "grad_norm": 40.34061050415039, "learning_rate": 9.99724566929134e-06, "loss": 0.9412, "step": 1753 }, { "epoch": 0.5524409448818898, "grad_norm": 38.01073455810547, "learning_rate": 9.99724409448819e-06, "loss": 0.1551, "step": 1754 }, { "epoch": 0.552755905511811, "grad_norm": 64.29601287841797, "learning_rate": 9.99724251968504e-06, "loss": 1.8098, "step": 1755 }, { "epoch": 0.5530708661417323, "grad_norm": 50.7859001159668, "learning_rate": 9.997240944881891e-06, "loss": 0.3809, "step": 1756 }, { "epoch": 0.5533858267716535, "grad_norm": 23.059926986694336, "learning_rate": 9.99723937007874e-06, "loss": 0.1878, "step": 1757 }, { "epoch": 0.5537007874015748, "grad_norm": 15.40706729888916, "learning_rate": 9.997237795275591e-06, "loss": 0.2192, "step": 1758 }, { "epoch": 0.5540157480314961, "grad_norm": 50.74030685424805, "learning_rate": 9.997236220472442e-06, "loss": 0.8022, "step": 1759 }, { "epoch": 0.5543307086614173, "grad_norm": 49.88266372680664, "learning_rate": 9.997234645669293e-06, "loss": 0.3615, "step": 1760 }, { "epoch": 0.5543307086614173, "eval_loss": 0.4197517931461334, "eval_runtime": 338.0619, "eval_samples_per_second": 0.346, "eval_steps_per_second": 0.346, "step": 1760 }, { "epoch": 0.5546456692913386, "grad_norm": 33.29900360107422, "learning_rate": 9.997233070866143e-06, "loss": 0.4475, "step": 1761 }, { "epoch": 0.5549606299212598, "grad_norm": 57.282649993896484, "learning_rate": 9.997231496062994e-06, "loss": 0.5688, "step": 1762 }, { "epoch": 0.5552755905511811, "grad_norm": 16.435340881347656, "learning_rate": 9.997229921259843e-06, "loss": 0.2603, "step": 1763 }, { "epoch": 0.5555905511811023, "grad_norm": 45.144737243652344, "learning_rate": 9.997228346456694e-06, "loss": 0.6854, "step": 1764 }, { "epoch": 0.5559055118110237, "grad_norm": 42.861515045166016, "learning_rate": 9.997226771653545e-06, "loss": 0.6296, "step": 1765 }, { "epoch": 0.5562204724409449, "grad_norm": 48.16584777832031, "learning_rate": 9.997225196850394e-06, "loss": 0.4873, "step": 1766 }, { "epoch": 0.5565354330708662, "grad_norm": 61.42652893066406, "learning_rate": 9.997223622047245e-06, "loss": 0.7604, "step": 1767 }, { "epoch": 0.5568503937007874, "grad_norm": 24.190889358520508, "learning_rate": 9.997222047244094e-06, "loss": 0.3546, "step": 1768 }, { "epoch": 0.5571653543307087, "grad_norm": 19.951372146606445, "learning_rate": 9.997220472440945e-06, "loss": 0.4902, "step": 1769 }, { "epoch": 0.5574803149606299, "grad_norm": 24.672101974487305, "learning_rate": 9.997218897637796e-06, "loss": 0.3603, "step": 1770 }, { "epoch": 0.5577952755905512, "grad_norm": 16.915287017822266, "learning_rate": 9.997217322834647e-06, "loss": 0.1767, "step": 1771 }, { "epoch": 0.5581102362204724, "grad_norm": 28.743501663208008, "learning_rate": 9.997215748031497e-06, "loss": 0.3357, "step": 1772 }, { "epoch": 0.5584251968503937, "grad_norm": 48.64558029174805, "learning_rate": 9.997214173228346e-06, "loss": 0.3531, "step": 1773 }, { "epoch": 0.5587401574803149, "grad_norm": 18.73015022277832, "learning_rate": 9.997212598425197e-06, "loss": 0.2041, "step": 1774 }, { "epoch": 0.5590551181102362, "grad_norm": 27.524059295654297, "learning_rate": 9.997211023622048e-06, "loss": 0.2056, "step": 1775 }, { "epoch": 0.5593700787401574, "grad_norm": 15.509268760681152, "learning_rate": 9.997209448818899e-06, "loss": 0.1131, "step": 1776 }, { "epoch": 0.5596850393700787, "grad_norm": 39.2099609375, "learning_rate": 9.997207874015748e-06, "loss": 0.2837, "step": 1777 }, { "epoch": 0.56, "grad_norm": 41.994815826416016, "learning_rate": 9.9972062992126e-06, "loss": 0.3212, "step": 1778 }, { "epoch": 0.5603149606299213, "grad_norm": 50.88380432128906, "learning_rate": 9.99720472440945e-06, "loss": 0.476, "step": 1779 }, { "epoch": 0.5606299212598426, "grad_norm": 23.539987564086914, "learning_rate": 9.997203149606301e-06, "loss": 0.1563, "step": 1780 }, { "epoch": 0.5606299212598426, "eval_loss": 0.5017465353012085, "eval_runtime": 339.1116, "eval_samples_per_second": 0.345, "eval_steps_per_second": 0.345, "step": 1780 }, { "epoch": 0.5609448818897638, "grad_norm": 48.950191497802734, "learning_rate": 9.99720157480315e-06, "loss": 0.7028, "step": 1781 }, { "epoch": 0.5612598425196851, "grad_norm": 15.981232643127441, "learning_rate": 9.997200000000001e-06, "loss": 0.0955, "step": 1782 }, { "epoch": 0.5615748031496063, "grad_norm": 53.745574951171875, "learning_rate": 9.99719842519685e-06, "loss": 0.5677, "step": 1783 }, { "epoch": 0.5618897637795276, "grad_norm": 57.129024505615234, "learning_rate": 9.997196850393702e-06, "loss": 0.5015, "step": 1784 }, { "epoch": 0.5622047244094488, "grad_norm": 25.303375244140625, "learning_rate": 9.997195275590553e-06, "loss": 0.3141, "step": 1785 }, { "epoch": 0.5625196850393701, "grad_norm": 29.012252807617188, "learning_rate": 9.997193700787402e-06, "loss": 0.3536, "step": 1786 }, { "epoch": 0.5628346456692913, "grad_norm": 10.699189186096191, "learning_rate": 9.997192125984253e-06, "loss": 0.0627, "step": 1787 }, { "epoch": 0.5631496062992126, "grad_norm": 8.34593391418457, "learning_rate": 9.997190551181102e-06, "loss": 0.0391, "step": 1788 }, { "epoch": 0.5634645669291338, "grad_norm": 42.121437072753906, "learning_rate": 9.997188976377953e-06, "loss": 0.8368, "step": 1789 }, { "epoch": 0.5637795275590551, "grad_norm": 35.72675704956055, "learning_rate": 9.997187401574804e-06, "loss": 0.5532, "step": 1790 }, { "epoch": 0.5640944881889763, "grad_norm": 45.88935470581055, "learning_rate": 9.997185826771655e-06, "loss": 0.1764, "step": 1791 }, { "epoch": 0.5644094488188977, "grad_norm": 32.38536071777344, "learning_rate": 9.997184251968505e-06, "loss": 0.1779, "step": 1792 }, { "epoch": 0.5647244094488189, "grad_norm": 72.07355499267578, "learning_rate": 9.997182677165354e-06, "loss": 0.3102, "step": 1793 }, { "epoch": 0.5650393700787402, "grad_norm": 82.99500274658203, "learning_rate": 9.997181102362205e-06, "loss": 0.6726, "step": 1794 }, { "epoch": 0.5653543307086614, "grad_norm": 80.94422149658203, "learning_rate": 9.997179527559056e-06, "loss": 0.7582, "step": 1795 }, { "epoch": 0.5656692913385827, "grad_norm": 4.303206443786621, "learning_rate": 9.997177952755907e-06, "loss": 0.0169, "step": 1796 }, { "epoch": 0.5659842519685039, "grad_norm": 38.44367980957031, "learning_rate": 9.997176377952756e-06, "loss": 0.5267, "step": 1797 }, { "epoch": 0.5662992125984252, "grad_norm": 48.196807861328125, "learning_rate": 9.997174803149607e-06, "loss": 0.6009, "step": 1798 }, { "epoch": 0.5666141732283465, "grad_norm": 33.66616439819336, "learning_rate": 9.997173228346458e-06, "loss": 0.5003, "step": 1799 }, { "epoch": 0.5669291338582677, "grad_norm": 34.213218688964844, "learning_rate": 9.997171653543309e-06, "loss": 0.2974, "step": 1800 }, { "epoch": 0.5669291338582677, "eval_loss": 0.477321982383728, "eval_runtime": 338.9113, "eval_samples_per_second": 0.345, "eval_steps_per_second": 0.345, "step": 1800 }, { "epoch": 0.567244094488189, "grad_norm": 63.08566665649414, "learning_rate": 9.997170078740158e-06, "loss": 0.1859, "step": 1801 }, { "epoch": 0.5675590551181102, "grad_norm": 61.78851318359375, "learning_rate": 9.99716850393701e-06, "loss": 0.9895, "step": 1802 }, { "epoch": 0.5678740157480315, "grad_norm": 47.29201126098633, "learning_rate": 9.997166929133859e-06, "loss": 0.5153, "step": 1803 }, { "epoch": 0.5681889763779527, "grad_norm": 33.21207046508789, "learning_rate": 9.99716535433071e-06, "loss": 0.3221, "step": 1804 }, { "epoch": 0.568503937007874, "grad_norm": 20.20362663269043, "learning_rate": 9.99716377952756e-06, "loss": 0.0697, "step": 1805 }, { "epoch": 0.5688188976377953, "grad_norm": 23.14605712890625, "learning_rate": 9.99716220472441e-06, "loss": 0.3096, "step": 1806 }, { "epoch": 0.5691338582677166, "grad_norm": 58.990699768066406, "learning_rate": 9.997160629921261e-06, "loss": 0.6046, "step": 1807 }, { "epoch": 0.5694488188976378, "grad_norm": 36.48517608642578, "learning_rate": 9.99715905511811e-06, "loss": 0.2886, "step": 1808 }, { "epoch": 0.5697637795275591, "grad_norm": 56.9811897277832, "learning_rate": 9.997157480314961e-06, "loss": 0.4658, "step": 1809 }, { "epoch": 0.5700787401574803, "grad_norm": 40.79648208618164, "learning_rate": 9.997155905511812e-06, "loss": 0.3177, "step": 1810 }, { "epoch": 0.5703937007874016, "grad_norm": 31.444948196411133, "learning_rate": 9.997154330708663e-06, "loss": 0.3236, "step": 1811 }, { "epoch": 0.5707086614173228, "grad_norm": 21.288089752197266, "learning_rate": 9.997152755905513e-06, "loss": 0.4159, "step": 1812 }, { "epoch": 0.5710236220472441, "grad_norm": 58.0438117980957, "learning_rate": 9.997151181102362e-06, "loss": 0.6176, "step": 1813 }, { "epoch": 0.5713385826771653, "grad_norm": 19.920330047607422, "learning_rate": 9.997149606299213e-06, "loss": 0.305, "step": 1814 }, { "epoch": 0.5716535433070866, "grad_norm": 35.30345916748047, "learning_rate": 9.997148031496064e-06, "loss": 0.4661, "step": 1815 }, { "epoch": 0.5719685039370078, "grad_norm": 16.50104522705078, "learning_rate": 9.997146456692915e-06, "loss": 0.2696, "step": 1816 }, { "epoch": 0.5722834645669291, "grad_norm": 40.40153121948242, "learning_rate": 9.997144881889764e-06, "loss": 0.5205, "step": 1817 }, { "epoch": 0.5725984251968504, "grad_norm": 12.835118293762207, "learning_rate": 9.997143307086615e-06, "loss": 0.1722, "step": 1818 }, { "epoch": 0.5729133858267716, "grad_norm": 40.450950622558594, "learning_rate": 9.997141732283464e-06, "loss": 0.3829, "step": 1819 }, { "epoch": 0.573228346456693, "grad_norm": 20.867347717285156, "learning_rate": 9.997140157480315e-06, "loss": 0.4121, "step": 1820 }, { "epoch": 0.573228346456693, "eval_loss": 0.40022820234298706, "eval_runtime": 337.3982, "eval_samples_per_second": 0.347, "eval_steps_per_second": 0.347, "step": 1820 }, { "epoch": 0.5735433070866142, "grad_norm": 27.65648651123047, "learning_rate": 9.997138582677166e-06, "loss": 0.1822, "step": 1821 }, { "epoch": 0.5738582677165355, "grad_norm": 57.512733459472656, "learning_rate": 9.997137007874017e-06, "loss": 0.8702, "step": 1822 }, { "epoch": 0.5741732283464567, "grad_norm": 13.143925666809082, "learning_rate": 9.997135433070867e-06, "loss": 0.183, "step": 1823 }, { "epoch": 0.574488188976378, "grad_norm": 49.15605926513672, "learning_rate": 9.997133858267718e-06, "loss": 0.3416, "step": 1824 }, { "epoch": 0.5748031496062992, "grad_norm": 12.571438789367676, "learning_rate": 9.997132283464569e-06, "loss": 0.1638, "step": 1825 }, { "epoch": 0.5751181102362205, "grad_norm": 66.83301544189453, "learning_rate": 9.997130708661418e-06, "loss": 0.6822, "step": 1826 }, { "epoch": 0.5754330708661417, "grad_norm": 22.997888565063477, "learning_rate": 9.997129133858269e-06, "loss": 0.2096, "step": 1827 }, { "epoch": 0.575748031496063, "grad_norm": 13.75336742401123, "learning_rate": 9.997127559055118e-06, "loss": 0.0664, "step": 1828 }, { "epoch": 0.5760629921259842, "grad_norm": 60.55500411987305, "learning_rate": 9.99712598425197e-06, "loss": 0.8159, "step": 1829 }, { "epoch": 0.5763779527559055, "grad_norm": 35.01063537597656, "learning_rate": 9.99712440944882e-06, "loss": 0.3943, "step": 1830 }, { "epoch": 0.5766929133858267, "grad_norm": 31.06571388244629, "learning_rate": 9.997122834645671e-06, "loss": 0.4897, "step": 1831 }, { "epoch": 0.577007874015748, "grad_norm": 78.39657592773438, "learning_rate": 9.99712125984252e-06, "loss": 0.5343, "step": 1832 }, { "epoch": 0.5773228346456692, "grad_norm": 46.089942932128906, "learning_rate": 9.99711968503937e-06, "loss": 0.5686, "step": 1833 }, { "epoch": 0.5776377952755906, "grad_norm": 37.687374114990234, "learning_rate": 9.99711811023622e-06, "loss": 0.6468, "step": 1834 }, { "epoch": 0.5779527559055118, "grad_norm": 22.699844360351562, "learning_rate": 9.997116535433072e-06, "loss": 0.3065, "step": 1835 }, { "epoch": 0.5782677165354331, "grad_norm": 89.35218048095703, "learning_rate": 9.997114960629923e-06, "loss": 0.4874, "step": 1836 }, { "epoch": 0.5785826771653543, "grad_norm": 95.06536102294922, "learning_rate": 9.997113385826772e-06, "loss": 0.4179, "step": 1837 }, { "epoch": 0.5788976377952756, "grad_norm": 44.81085205078125, "learning_rate": 9.997111811023623e-06, "loss": 0.3927, "step": 1838 }, { "epoch": 0.5792125984251969, "grad_norm": 49.45285415649414, "learning_rate": 9.997110236220472e-06, "loss": 0.5742, "step": 1839 }, { "epoch": 0.5795275590551181, "grad_norm": 35.893402099609375, "learning_rate": 9.997108661417323e-06, "loss": 0.4282, "step": 1840 }, { "epoch": 0.5795275590551181, "eval_loss": 0.43516087532043457, "eval_runtime": 337.9832, "eval_samples_per_second": 0.346, "eval_steps_per_second": 0.346, "step": 1840 }, { "epoch": 0.5798425196850394, "grad_norm": 42.341896057128906, "learning_rate": 9.997107086614174e-06, "loss": 0.2899, "step": 1841 }, { "epoch": 0.5801574803149606, "grad_norm": 28.99087142944336, "learning_rate": 9.997105511811025e-06, "loss": 0.1677, "step": 1842 }, { "epoch": 0.5804724409448819, "grad_norm": 13.114380836486816, "learning_rate": 9.997103937007875e-06, "loss": 0.063, "step": 1843 }, { "epoch": 0.5807874015748031, "grad_norm": 28.455913543701172, "learning_rate": 9.997102362204724e-06, "loss": 0.3202, "step": 1844 }, { "epoch": 0.5811023622047244, "grad_norm": 65.52840423583984, "learning_rate": 9.997100787401577e-06, "loss": 0.5528, "step": 1845 }, { "epoch": 0.5814173228346456, "grad_norm": 14.398350715637207, "learning_rate": 9.997099212598426e-06, "loss": 0.0874, "step": 1846 }, { "epoch": 0.581732283464567, "grad_norm": 46.19868850708008, "learning_rate": 9.997097637795277e-06, "loss": 0.2662, "step": 1847 }, { "epoch": 0.5820472440944882, "grad_norm": 13.150195121765137, "learning_rate": 9.997096062992126e-06, "loss": 0.1102, "step": 1848 }, { "epoch": 0.5823622047244095, "grad_norm": 19.446687698364258, "learning_rate": 9.997094488188977e-06, "loss": 0.1399, "step": 1849 }, { "epoch": 0.5826771653543307, "grad_norm": 112.38624572753906, "learning_rate": 9.997092913385828e-06, "loss": 0.3521, "step": 1850 }, { "epoch": 0.582992125984252, "grad_norm": 14.854012489318848, "learning_rate": 9.997091338582679e-06, "loss": 0.1212, "step": 1851 }, { "epoch": 0.5833070866141732, "grad_norm": 58.18556213378906, "learning_rate": 9.997089763779528e-06, "loss": 0.4609, "step": 1852 }, { "epoch": 0.5836220472440945, "grad_norm": 8.105414390563965, "learning_rate": 9.997088188976378e-06, "loss": 0.047, "step": 1853 }, { "epoch": 0.5839370078740157, "grad_norm": 78.19012451171875, "learning_rate": 9.997086614173229e-06, "loss": 0.3938, "step": 1854 }, { "epoch": 0.584251968503937, "grad_norm": 60.095645904541016, "learning_rate": 9.99708503937008e-06, "loss": 1.0973, "step": 1855 }, { "epoch": 0.5845669291338582, "grad_norm": 27.53265953063965, "learning_rate": 9.99708346456693e-06, "loss": 0.3794, "step": 1856 }, { "epoch": 0.5848818897637795, "grad_norm": 90.6170654296875, "learning_rate": 9.99708188976378e-06, "loss": 0.7531, "step": 1857 }, { "epoch": 0.5851968503937008, "grad_norm": 82.42227935791016, "learning_rate": 9.997080314960631e-06, "loss": 0.8938, "step": 1858 }, { "epoch": 0.585511811023622, "grad_norm": 37.9282112121582, "learning_rate": 9.99707874015748e-06, "loss": 0.2363, "step": 1859 }, { "epoch": 0.5858267716535434, "grad_norm": 17.56612777709961, "learning_rate": 9.997077165354331e-06, "loss": 0.3155, "step": 1860 }, { "epoch": 0.5858267716535434, "eval_loss": 0.5003868937492371, "eval_runtime": 368.9458, "eval_samples_per_second": 0.317, "eval_steps_per_second": 0.317, "step": 1860 }, { "epoch": 0.5861417322834646, "grad_norm": 40.288368225097656, "learning_rate": 9.997075590551182e-06, "loss": 0.3474, "step": 1861 }, { "epoch": 0.5864566929133859, "grad_norm": 29.836639404296875, "learning_rate": 9.997074015748033e-06, "loss": 0.3913, "step": 1862 }, { "epoch": 0.5867716535433071, "grad_norm": 32.93700408935547, "learning_rate": 9.997072440944883e-06, "loss": 0.2077, "step": 1863 }, { "epoch": 0.5870866141732284, "grad_norm": 44.143409729003906, "learning_rate": 9.997070866141732e-06, "loss": 0.359, "step": 1864 }, { "epoch": 0.5874015748031496, "grad_norm": 18.514257431030273, "learning_rate": 9.997069291338583e-06, "loss": 0.2217, "step": 1865 }, { "epoch": 0.5877165354330709, "grad_norm": 72.15349578857422, "learning_rate": 9.997067716535434e-06, "loss": 0.6858, "step": 1866 }, { "epoch": 0.5880314960629921, "grad_norm": 53.466983795166016, "learning_rate": 9.997066141732285e-06, "loss": 0.4746, "step": 1867 }, { "epoch": 0.5883464566929134, "grad_norm": 22.331130981445312, "learning_rate": 9.997064566929134e-06, "loss": 0.2087, "step": 1868 }, { "epoch": 0.5886614173228346, "grad_norm": 49.12759017944336, "learning_rate": 9.997062992125985e-06, "loss": 0.8626, "step": 1869 }, { "epoch": 0.5889763779527559, "grad_norm": 68.28533935546875, "learning_rate": 9.997061417322836e-06, "loss": 0.9698, "step": 1870 }, { "epoch": 0.5892913385826771, "grad_norm": 61.11606979370117, "learning_rate": 9.997059842519687e-06, "loss": 0.6654, "step": 1871 }, { "epoch": 0.5896062992125984, "grad_norm": 39.162593841552734, "learning_rate": 9.997058267716536e-06, "loss": 0.3398, "step": 1872 }, { "epoch": 0.5899212598425196, "grad_norm": 34.68936538696289, "learning_rate": 9.997056692913386e-06, "loss": 0.4498, "step": 1873 }, { "epoch": 0.590236220472441, "grad_norm": 31.508249282836914, "learning_rate": 9.997055118110237e-06, "loss": 0.5166, "step": 1874 }, { "epoch": 0.5905511811023622, "grad_norm": 26.828365325927734, "learning_rate": 9.997053543307088e-06, "loss": 0.3993, "step": 1875 }, { "epoch": 0.5908661417322835, "grad_norm": 31.714258193969727, "learning_rate": 9.997051968503939e-06, "loss": 0.4829, "step": 1876 }, { "epoch": 0.5911811023622047, "grad_norm": 13.222908020019531, "learning_rate": 9.997050393700788e-06, "loss": 0.2407, "step": 1877 }, { "epoch": 0.591496062992126, "grad_norm": 18.35155487060547, "learning_rate": 9.997048818897639e-06, "loss": 0.4279, "step": 1878 }, { "epoch": 0.5918110236220473, "grad_norm": 11.358118057250977, "learning_rate": 9.997047244094488e-06, "loss": 0.208, "step": 1879 }, { "epoch": 0.5921259842519685, "grad_norm": 26.622526168823242, "learning_rate": 9.99704566929134e-06, "loss": 0.4925, "step": 1880 }, { "epoch": 0.5921259842519685, "eval_loss": 0.5159415602684021, "eval_runtime": 351.9211, "eval_samples_per_second": 0.332, "eval_steps_per_second": 0.332, "step": 1880 }, { "epoch": 0.5924409448818898, "grad_norm": 32.826812744140625, "learning_rate": 9.99704409448819e-06, "loss": 0.6125, "step": 1881 }, { "epoch": 0.592755905511811, "grad_norm": 55.511314392089844, "learning_rate": 9.997042519685041e-06, "loss": 0.3093, "step": 1882 }, { "epoch": 0.5930708661417323, "grad_norm": 35.73579788208008, "learning_rate": 9.99704094488189e-06, "loss": 0.7267, "step": 1883 }, { "epoch": 0.5933858267716535, "grad_norm": 62.44449234008789, "learning_rate": 9.99703937007874e-06, "loss": 0.6222, "step": 1884 }, { "epoch": 0.5937007874015748, "grad_norm": 33.10036087036133, "learning_rate": 9.99703779527559e-06, "loss": 0.9203, "step": 1885 }, { "epoch": 0.594015748031496, "grad_norm": 12.248406410217285, "learning_rate": 9.997036220472442e-06, "loss": 0.1355, "step": 1886 }, { "epoch": 0.5943307086614174, "grad_norm": 30.380659103393555, "learning_rate": 9.997034645669293e-06, "loss": 0.4853, "step": 1887 }, { "epoch": 0.5946456692913386, "grad_norm": 32.66392135620117, "learning_rate": 9.997033070866142e-06, "loss": 0.6233, "step": 1888 }, { "epoch": 0.5949606299212599, "grad_norm": 50.096702575683594, "learning_rate": 9.997031496062993e-06, "loss": 0.4143, "step": 1889 }, { "epoch": 0.5952755905511811, "grad_norm": 10.48125171661377, "learning_rate": 9.997029921259842e-06, "loss": 0.1208, "step": 1890 }, { "epoch": 0.5955905511811024, "grad_norm": 39.09800720214844, "learning_rate": 9.997028346456693e-06, "loss": 0.2144, "step": 1891 }, { "epoch": 0.5959055118110236, "grad_norm": 18.518939971923828, "learning_rate": 9.997026771653544e-06, "loss": 0.3954, "step": 1892 }, { "epoch": 0.5962204724409449, "grad_norm": 28.824628829956055, "learning_rate": 9.997025196850394e-06, "loss": 0.5161, "step": 1893 }, { "epoch": 0.5965354330708661, "grad_norm": 36.205902099609375, "learning_rate": 9.997023622047245e-06, "loss": 0.3268, "step": 1894 }, { "epoch": 0.5968503937007874, "grad_norm": 44.05073165893555, "learning_rate": 9.997022047244096e-06, "loss": 0.6844, "step": 1895 }, { "epoch": 0.5971653543307086, "grad_norm": 32.883384704589844, "learning_rate": 9.997020472440947e-06, "loss": 0.2967, "step": 1896 }, { "epoch": 0.5974803149606299, "grad_norm": 41.35356140136719, "learning_rate": 9.997018897637796e-06, "loss": 0.8327, "step": 1897 }, { "epoch": 0.5977952755905512, "grad_norm": 30.476848602294922, "learning_rate": 9.997017322834647e-06, "loss": 0.384, "step": 1898 }, { "epoch": 0.5981102362204724, "grad_norm": 20.232952117919922, "learning_rate": 9.997015748031496e-06, "loss": 0.4447, "step": 1899 }, { "epoch": 0.5984251968503937, "grad_norm": 5.822659492492676, "learning_rate": 9.997014173228347e-06, "loss": 0.0435, "step": 1900 }, { "epoch": 0.5984251968503937, "eval_loss": 0.49179500341415405, "eval_runtime": 348.9961, "eval_samples_per_second": 0.335, "eval_steps_per_second": 0.335, "step": 1900 }, { "epoch": 0.598740157480315, "grad_norm": 18.793500900268555, "learning_rate": 9.997012598425198e-06, "loss": 0.2095, "step": 1901 }, { "epoch": 0.5990551181102363, "grad_norm": 17.714536666870117, "learning_rate": 9.99701102362205e-06, "loss": 0.222, "step": 1902 }, { "epoch": 0.5993700787401575, "grad_norm": 35.780208587646484, "learning_rate": 9.997009448818898e-06, "loss": 0.2701, "step": 1903 }, { "epoch": 0.5996850393700788, "grad_norm": 53.45842361450195, "learning_rate": 9.997007874015748e-06, "loss": 0.8297, "step": 1904 }, { "epoch": 0.6, "grad_norm": 40.31769561767578, "learning_rate": 9.997006299212599e-06, "loss": 0.3894, "step": 1905 }, { "epoch": 0.6003149606299213, "grad_norm": 24.99709129333496, "learning_rate": 9.99700472440945e-06, "loss": 0.3682, "step": 1906 }, { "epoch": 0.6006299212598425, "grad_norm": 52.774383544921875, "learning_rate": 9.9970031496063e-06, "loss": 0.3242, "step": 1907 }, { "epoch": 0.6009448818897638, "grad_norm": 19.92904281616211, "learning_rate": 9.99700157480315e-06, "loss": 0.2621, "step": 1908 }, { "epoch": 0.601259842519685, "grad_norm": 38.63066864013672, "learning_rate": 9.997000000000001e-06, "loss": 0.2645, "step": 1909 }, { "epoch": 0.6015748031496063, "grad_norm": 20.248315811157227, "learning_rate": 9.99699842519685e-06, "loss": 0.0956, "step": 1910 }, { "epoch": 0.6018897637795275, "grad_norm": 162.4951934814453, "learning_rate": 9.996996850393701e-06, "loss": 0.4668, "step": 1911 }, { "epoch": 0.6022047244094488, "grad_norm": 11.444823265075684, "learning_rate": 9.996995275590552e-06, "loss": 0.0933, "step": 1912 }, { "epoch": 0.60251968503937, "grad_norm": 16.599746704101562, "learning_rate": 9.996993700787403e-06, "loss": 0.1043, "step": 1913 }, { "epoch": 0.6028346456692913, "grad_norm": 66.4779281616211, "learning_rate": 9.996992125984253e-06, "loss": 0.818, "step": 1914 }, { "epoch": 0.6031496062992125, "grad_norm": 26.693084716796875, "learning_rate": 9.996990551181104e-06, "loss": 0.572, "step": 1915 }, { "epoch": 0.6034645669291339, "grad_norm": 71.67481994628906, "learning_rate": 9.996988976377955e-06, "loss": 0.1498, "step": 1916 }, { "epoch": 0.6037795275590552, "grad_norm": 32.80756378173828, "learning_rate": 9.996987401574804e-06, "loss": 0.2654, "step": 1917 }, { "epoch": 0.6040944881889764, "grad_norm": 73.54534149169922, "learning_rate": 9.996985826771655e-06, "loss": 0.8395, "step": 1918 }, { "epoch": 0.6044094488188977, "grad_norm": 124.98379516601562, "learning_rate": 9.996984251968504e-06, "loss": 0.4122, "step": 1919 }, { "epoch": 0.6047244094488189, "grad_norm": 58.07841873168945, "learning_rate": 9.996982677165355e-06, "loss": 0.4811, "step": 1920 }, { "epoch": 0.6047244094488189, "eval_loss": 0.490536630153656, "eval_runtime": 351.4081, "eval_samples_per_second": 0.333, "eval_steps_per_second": 0.333, "step": 1920 }, { "epoch": 0.6050393700787402, "grad_norm": 47.089378356933594, "learning_rate": 9.996981102362206e-06, "loss": 0.5479, "step": 1921 }, { "epoch": 0.6053543307086614, "grad_norm": 40.61234664916992, "learning_rate": 9.996979527559057e-06, "loss": 0.1702, "step": 1922 }, { "epoch": 0.6056692913385827, "grad_norm": 76.10828399658203, "learning_rate": 9.996977952755906e-06, "loss": 1.0415, "step": 1923 }, { "epoch": 0.6059842519685039, "grad_norm": 8.746352195739746, "learning_rate": 9.996976377952756e-06, "loss": 0.0595, "step": 1924 }, { "epoch": 0.6062992125984252, "grad_norm": 49.07436752319336, "learning_rate": 9.996974803149607e-06, "loss": 0.8017, "step": 1925 }, { "epoch": 0.6066141732283464, "grad_norm": 37.18422317504883, "learning_rate": 9.996973228346458e-06, "loss": 0.3798, "step": 1926 }, { "epoch": 0.6069291338582677, "grad_norm": 41.602073669433594, "learning_rate": 9.996971653543309e-06, "loss": 0.4223, "step": 1927 }, { "epoch": 0.607244094488189, "grad_norm": 46.16876220703125, "learning_rate": 9.996970078740158e-06, "loss": 0.5156, "step": 1928 }, { "epoch": 0.6075590551181103, "grad_norm": 38.204036712646484, "learning_rate": 9.996968503937009e-06, "loss": 0.3032, "step": 1929 }, { "epoch": 0.6078740157480315, "grad_norm": 7.078056335449219, "learning_rate": 9.996966929133858e-06, "loss": 0.0285, "step": 1930 }, { "epoch": 0.6081889763779528, "grad_norm": 76.96788787841797, "learning_rate": 9.99696535433071e-06, "loss": 0.4164, "step": 1931 }, { "epoch": 0.608503937007874, "grad_norm": 33.73112487792969, "learning_rate": 9.99696377952756e-06, "loss": 0.1971, "step": 1932 }, { "epoch": 0.6088188976377953, "grad_norm": 50.70700454711914, "learning_rate": 9.996962204724411e-06, "loss": 0.6604, "step": 1933 }, { "epoch": 0.6091338582677165, "grad_norm": 16.3695011138916, "learning_rate": 9.99696062992126e-06, "loss": 0.2403, "step": 1934 }, { "epoch": 0.6094488188976378, "grad_norm": 32.59841537475586, "learning_rate": 9.99695905511811e-06, "loss": 0.3409, "step": 1935 }, { "epoch": 0.609763779527559, "grad_norm": 33.59854507446289, "learning_rate": 9.99695748031496e-06, "loss": 0.2444, "step": 1936 }, { "epoch": 0.6100787401574803, "grad_norm": 53.62559509277344, "learning_rate": 9.996955905511812e-06, "loss": 0.1804, "step": 1937 }, { "epoch": 0.6103937007874016, "grad_norm": 41.21846389770508, "learning_rate": 9.996954330708663e-06, "loss": 0.8659, "step": 1938 }, { "epoch": 0.6107086614173228, "grad_norm": 65.04766845703125, "learning_rate": 9.996952755905512e-06, "loss": 0.4908, "step": 1939 }, { "epoch": 0.6110236220472441, "grad_norm": 63.45843505859375, "learning_rate": 9.996951181102363e-06, "loss": 0.8725, "step": 1940 }, { "epoch": 0.6110236220472441, "eval_loss": 0.5175274014472961, "eval_runtime": 349.2545, "eval_samples_per_second": 0.335, "eval_steps_per_second": 0.335, "step": 1940 }, { "epoch": 0.6113385826771653, "grad_norm": 59.46821212768555, "learning_rate": 9.996949606299214e-06, "loss": 0.5103, "step": 1941 }, { "epoch": 0.6116535433070867, "grad_norm": 18.964391708374023, "learning_rate": 9.996948031496065e-06, "loss": 0.3203, "step": 1942 }, { "epoch": 0.6119685039370079, "grad_norm": 101.10069274902344, "learning_rate": 9.996946456692914e-06, "loss": 0.7592, "step": 1943 }, { "epoch": 0.6122834645669292, "grad_norm": 36.87664794921875, "learning_rate": 9.996944881889764e-06, "loss": 0.5184, "step": 1944 }, { "epoch": 0.6125984251968504, "grad_norm": 31.650327682495117, "learning_rate": 9.996943307086615e-06, "loss": 0.3825, "step": 1945 }, { "epoch": 0.6129133858267717, "grad_norm": 23.641586303710938, "learning_rate": 9.996941732283466e-06, "loss": 0.1946, "step": 1946 }, { "epoch": 0.6132283464566929, "grad_norm": 19.757226943969727, "learning_rate": 9.996940157480317e-06, "loss": 0.0762, "step": 1947 }, { "epoch": 0.6135433070866142, "grad_norm": 10.93740177154541, "learning_rate": 9.996938582677166e-06, "loss": 0.0708, "step": 1948 }, { "epoch": 0.6138582677165354, "grad_norm": 21.542688369750977, "learning_rate": 9.996937007874017e-06, "loss": 0.2724, "step": 1949 }, { "epoch": 0.6141732283464567, "grad_norm": 27.340009689331055, "learning_rate": 9.996935433070866e-06, "loss": 0.0854, "step": 1950 }, { "epoch": 0.6144881889763779, "grad_norm": 47.907955169677734, "learning_rate": 9.996933858267717e-06, "loss": 0.3816, "step": 1951 }, { "epoch": 0.6148031496062992, "grad_norm": 65.70764923095703, "learning_rate": 9.996932283464568e-06, "loss": 0.2682, "step": 1952 }, { "epoch": 0.6151181102362204, "grad_norm": 39.29658889770508, "learning_rate": 9.99693070866142e-06, "loss": 0.3791, "step": 1953 }, { "epoch": 0.6154330708661417, "grad_norm": 7.722301006317139, "learning_rate": 9.996929133858268e-06, "loss": 0.0215, "step": 1954 }, { "epoch": 0.6157480314960629, "grad_norm": 47.819461822509766, "learning_rate": 9.996927559055118e-06, "loss": 0.6154, "step": 1955 }, { "epoch": 0.6160629921259843, "grad_norm": 32.92623519897461, "learning_rate": 9.996925984251969e-06, "loss": 0.1975, "step": 1956 }, { "epoch": 0.6163779527559056, "grad_norm": 73.50929260253906, "learning_rate": 9.99692440944882e-06, "loss": 0.2781, "step": 1957 }, { "epoch": 0.6166929133858268, "grad_norm": 26.083810806274414, "learning_rate": 9.99692283464567e-06, "loss": 0.3916, "step": 1958 }, { "epoch": 0.6170078740157481, "grad_norm": 3.3253297805786133, "learning_rate": 9.99692125984252e-06, "loss": 0.0313, "step": 1959 }, { "epoch": 0.6173228346456693, "grad_norm": 0.6226401329040527, "learning_rate": 9.996919685039371e-06, "loss": 0.0049, "step": 1960 }, { "epoch": 0.6173228346456693, "eval_loss": 0.4793933033943176, "eval_runtime": 341.1297, "eval_samples_per_second": 0.343, "eval_steps_per_second": 0.343, "step": 1960 }, { "epoch": 0.6176377952755906, "grad_norm": 38.08949279785156, "learning_rate": 9.99691811023622e-06, "loss": 0.2089, "step": 1961 }, { "epoch": 0.6179527559055118, "grad_norm": 96.51960754394531, "learning_rate": 9.996916535433071e-06, "loss": 0.6145, "step": 1962 }, { "epoch": 0.6182677165354331, "grad_norm": 36.37595748901367, "learning_rate": 9.996914960629922e-06, "loss": 0.1043, "step": 1963 }, { "epoch": 0.6185826771653543, "grad_norm": 112.60614013671875, "learning_rate": 9.996913385826772e-06, "loss": 1.0763, "step": 1964 }, { "epoch": 0.6188976377952756, "grad_norm": 49.78593444824219, "learning_rate": 9.996911811023623e-06, "loss": 0.1965, "step": 1965 }, { "epoch": 0.6192125984251968, "grad_norm": 58.12906265258789, "learning_rate": 9.996910236220474e-06, "loss": 0.3536, "step": 1966 }, { "epoch": 0.6195275590551181, "grad_norm": 71.21797180175781, "learning_rate": 9.996908661417325e-06, "loss": 0.6605, "step": 1967 }, { "epoch": 0.6198425196850393, "grad_norm": 104.464111328125, "learning_rate": 9.996907086614174e-06, "loss": 0.3808, "step": 1968 }, { "epoch": 0.6201574803149607, "grad_norm": 79.76933288574219, "learning_rate": 9.996905511811025e-06, "loss": 0.5754, "step": 1969 }, { "epoch": 0.6204724409448819, "grad_norm": 21.67342758178711, "learning_rate": 9.996903937007874e-06, "loss": 0.0917, "step": 1970 }, { "epoch": 0.6207874015748032, "grad_norm": 34.137447357177734, "learning_rate": 9.996902362204725e-06, "loss": 0.5807, "step": 1971 }, { "epoch": 0.6211023622047244, "grad_norm": 46.111595153808594, "learning_rate": 9.996900787401576e-06, "loss": 0.3738, "step": 1972 }, { "epoch": 0.6214173228346457, "grad_norm": 45.28417205810547, "learning_rate": 9.996899212598427e-06, "loss": 0.7565, "step": 1973 }, { "epoch": 0.6217322834645669, "grad_norm": 36.73102569580078, "learning_rate": 9.996897637795276e-06, "loss": 0.7887, "step": 1974 }, { "epoch": 0.6220472440944882, "grad_norm": 69.32728576660156, "learning_rate": 9.996896062992126e-06, "loss": 0.4563, "step": 1975 }, { "epoch": 0.6223622047244094, "grad_norm": 43.308555603027344, "learning_rate": 9.996894488188977e-06, "loss": 0.2911, "step": 1976 }, { "epoch": 0.6226771653543307, "grad_norm": 40.517086029052734, "learning_rate": 9.996892913385828e-06, "loss": 0.3979, "step": 1977 }, { "epoch": 0.622992125984252, "grad_norm": 68.62828826904297, "learning_rate": 9.996891338582679e-06, "loss": 0.3365, "step": 1978 }, { "epoch": 0.6233070866141732, "grad_norm": 28.892871856689453, "learning_rate": 9.996889763779528e-06, "loss": 0.5079, "step": 1979 }, { "epoch": 0.6236220472440945, "grad_norm": 22.088882446289062, "learning_rate": 9.996888188976379e-06, "loss": 0.3113, "step": 1980 }, { "epoch": 0.6236220472440945, "eval_loss": 0.4940292537212372, "eval_runtime": 352.4729, "eval_samples_per_second": 0.332, "eval_steps_per_second": 0.332, "step": 1980 }, { "epoch": 0.6239370078740157, "grad_norm": 37.80302047729492, "learning_rate": 9.996886614173228e-06, "loss": 0.6656, "step": 1981 }, { "epoch": 0.624251968503937, "grad_norm": 19.02584457397461, "learning_rate": 9.99688503937008e-06, "loss": 0.1548, "step": 1982 }, { "epoch": 0.6245669291338583, "grad_norm": 10.896634101867676, "learning_rate": 9.99688346456693e-06, "loss": 0.1322, "step": 1983 }, { "epoch": 0.6248818897637796, "grad_norm": 33.37139892578125, "learning_rate": 9.99688188976378e-06, "loss": 0.3109, "step": 1984 }, { "epoch": 0.6251968503937008, "grad_norm": 20.222166061401367, "learning_rate": 9.99688031496063e-06, "loss": 0.3568, "step": 1985 }, { "epoch": 0.6255118110236221, "grad_norm": 26.97902488708496, "learning_rate": 9.996878740157482e-06, "loss": 0.4469, "step": 1986 }, { "epoch": 0.6258267716535433, "grad_norm": 3.6876025199890137, "learning_rate": 9.996877165354333e-06, "loss": 0.017, "step": 1987 }, { "epoch": 0.6261417322834646, "grad_norm": 3.7643909454345703, "learning_rate": 9.996875590551182e-06, "loss": 0.0312, "step": 1988 }, { "epoch": 0.6264566929133858, "grad_norm": 58.000083923339844, "learning_rate": 9.996874015748033e-06, "loss": 0.5242, "step": 1989 }, { "epoch": 0.6267716535433071, "grad_norm": 57.707000732421875, "learning_rate": 9.996872440944882e-06, "loss": 0.537, "step": 1990 }, { "epoch": 0.6270866141732283, "grad_norm": 23.221139907836914, "learning_rate": 9.996870866141733e-06, "loss": 0.1436, "step": 1991 }, { "epoch": 0.6274015748031496, "grad_norm": 17.156478881835938, "learning_rate": 9.996869291338584e-06, "loss": 0.2777, "step": 1992 }, { "epoch": 0.6277165354330708, "grad_norm": 81.11396789550781, "learning_rate": 9.996867716535435e-06, "loss": 0.6183, "step": 1993 }, { "epoch": 0.6280314960629921, "grad_norm": 17.866575241088867, "learning_rate": 9.996866141732284e-06, "loss": 0.2052, "step": 1994 }, { "epoch": 0.6283464566929133, "grad_norm": 49.33943557739258, "learning_rate": 9.996864566929134e-06, "loss": 0.2541, "step": 1995 }, { "epoch": 0.6286614173228346, "grad_norm": 24.039709091186523, "learning_rate": 9.996862992125985e-06, "loss": 0.1116, "step": 1996 }, { "epoch": 0.628976377952756, "grad_norm": 17.275371551513672, "learning_rate": 9.996861417322836e-06, "loss": 0.054, "step": 1997 }, { "epoch": 0.6292913385826772, "grad_norm": 65.09687805175781, "learning_rate": 9.996859842519687e-06, "loss": 0.842, "step": 1998 }, { "epoch": 0.6296062992125985, "grad_norm": 63.773319244384766, "learning_rate": 9.996858267716536e-06, "loss": 0.2476, "step": 1999 }, { "epoch": 0.6299212598425197, "grad_norm": 101.19293212890625, "learning_rate": 9.996856692913387e-06, "loss": 1.0292, "step": 2000 }, { "epoch": 0.6299212598425197, "eval_loss": 0.49186941981315613, "eval_runtime": 611.4185, "eval_samples_per_second": 0.191, "eval_steps_per_second": 0.191, "step": 2000 } ], "logging_steps": 1, "max_steps": 6350000, "num_input_tokens_seen": 0, "num_train_epochs": 2000, "save_steps": 20, "total_flos": 1.252556321381376e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }