{ "best_metric": null, "best_model_checkpoint": null, "epoch": 4.995079950799508, "eval_steps": 102, "global_step": 1015, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.004920049200492005, "grad_norm": 2.296875, "learning_rate": 1.0000000000000002e-06, "loss": 0.2408, "step": 1 }, { "epoch": 0.00984009840098401, "grad_norm": 2.8125, "learning_rate": 2.0000000000000003e-06, "loss": 0.2626, "step": 2 }, { "epoch": 0.014760147601476014, "grad_norm": 2.484375, "learning_rate": 3e-06, "loss": 0.2438, "step": 3 }, { "epoch": 0.01968019680196802, "grad_norm": 2.375, "learning_rate": 4.000000000000001e-06, "loss": 0.2335, "step": 4 }, { "epoch": 0.024600246002460024, "grad_norm": 2.453125, "learning_rate": 5e-06, "loss": 0.229, "step": 5 }, { "epoch": 0.02952029520295203, "grad_norm": 1.78125, "learning_rate": 6e-06, "loss": 0.2125, "step": 6 }, { "epoch": 0.03444034440344403, "grad_norm": 1.4765625, "learning_rate": 7e-06, "loss": 0.1421, "step": 7 }, { "epoch": 0.03936039360393604, "grad_norm": 1.53125, "learning_rate": 8.000000000000001e-06, "loss": 0.1727, "step": 8 }, { "epoch": 0.04428044280442804, "grad_norm": 1.4296875, "learning_rate": 9e-06, "loss": 0.1599, "step": 9 }, { "epoch": 0.04920049200492005, "grad_norm": 1.7265625, "learning_rate": 1e-05, "loss": 0.2048, "step": 10 }, { "epoch": 0.05412054120541206, "grad_norm": 2.125, "learning_rate": 9.99997557091071e-06, "loss": 0.1458, "step": 11 }, { "epoch": 0.05904059040590406, "grad_norm": 1.6015625, "learning_rate": 9.99990228388155e-06, "loss": 0.1435, "step": 12 }, { "epoch": 0.06396063960639606, "grad_norm": 1.2578125, "learning_rate": 9.999780139628657e-06, "loss": 0.1791, "step": 13 }, { "epoch": 0.06888068880688807, "grad_norm": 1.375, "learning_rate": 9.999609139345578e-06, "loss": 0.1412, "step": 14 }, { "epoch": 0.07380073800738007, "grad_norm": 1.0859375, "learning_rate": 9.999389284703265e-06, "loss": 0.1416, "step": 15 }, { "epoch": 0.07872078720787208, "grad_norm": 1.21875, "learning_rate": 9.99912057785006e-06, "loss": 0.1255, "step": 16 }, { "epoch": 0.08364083640836409, "grad_norm": 0.9140625, "learning_rate": 9.998803021411668e-06, "loss": 0.1201, "step": 17 }, { "epoch": 0.08856088560885608, "grad_norm": 1.6875, "learning_rate": 9.99843661849113e-06, "loss": 0.2039, "step": 18 }, { "epoch": 0.09348093480934809, "grad_norm": 0.96875, "learning_rate": 9.998021372668809e-06, "loss": 0.1352, "step": 19 }, { "epoch": 0.0984009840098401, "grad_norm": 0.9296875, "learning_rate": 9.99755728800233e-06, "loss": 0.1448, "step": 20 }, { "epoch": 0.1033210332103321, "grad_norm": 1.109375, "learning_rate": 9.997044369026563e-06, "loss": 0.1537, "step": 21 }, { "epoch": 0.10824108241082411, "grad_norm": 1.0234375, "learning_rate": 9.996482620753565e-06, "loss": 0.1506, "step": 22 }, { "epoch": 0.11316113161131611, "grad_norm": 1.109375, "learning_rate": 9.995872048672535e-06, "loss": 0.1474, "step": 23 }, { "epoch": 0.11808118081180811, "grad_norm": 0.8671875, "learning_rate": 9.99521265874976e-06, "loss": 0.1215, "step": 24 }, { "epoch": 0.12300123001230012, "grad_norm": 1.0859375, "learning_rate": 9.994504457428557e-06, "loss": 0.126, "step": 25 }, { "epoch": 0.12792127921279212, "grad_norm": 1.015625, "learning_rate": 9.993747451629217e-06, "loss": 0.1417, "step": 26 }, { "epoch": 0.13284132841328414, "grad_norm": 1.15625, "learning_rate": 9.992941648748917e-06, "loss": 0.1163, "step": 27 }, { "epoch": 0.13776137761377613, "grad_norm": 1.046875, "learning_rate": 9.992087056661676e-06, "loss": 0.1332, "step": 28 }, { "epoch": 0.14268142681426815, "grad_norm": 0.8671875, "learning_rate": 9.991183683718253e-06, "loss": 0.1269, "step": 29 }, { "epoch": 0.14760147601476015, "grad_norm": 1.0, "learning_rate": 9.99023153874608e-06, "loss": 0.1071, "step": 30 }, { "epoch": 0.15252152521525214, "grad_norm": 1.078125, "learning_rate": 9.989230631049171e-06, "loss": 0.1415, "step": 31 }, { "epoch": 0.15744157441574416, "grad_norm": 1.1015625, "learning_rate": 9.988180970408032e-06, "loss": 0.1439, "step": 32 }, { "epoch": 0.16236162361623616, "grad_norm": 1.046875, "learning_rate": 9.987082567079563e-06, "loss": 0.1333, "step": 33 }, { "epoch": 0.16728167281672818, "grad_norm": 0.9296875, "learning_rate": 9.985935431796962e-06, "loss": 0.134, "step": 34 }, { "epoch": 0.17220172201722017, "grad_norm": 1.1015625, "learning_rate": 9.984739575769619e-06, "loss": 0.1315, "step": 35 }, { "epoch": 0.17712177121771217, "grad_norm": 1.0625, "learning_rate": 9.983495010683e-06, "loss": 0.1258, "step": 36 }, { "epoch": 0.1820418204182042, "grad_norm": 0.87890625, "learning_rate": 9.982201748698543e-06, "loss": 0.1141, "step": 37 }, { "epoch": 0.18696186961869618, "grad_norm": 1.0703125, "learning_rate": 9.980859802453533e-06, "loss": 0.1437, "step": 38 }, { "epoch": 0.1918819188191882, "grad_norm": 0.95703125, "learning_rate": 9.97946918506098e-06, "loss": 0.1326, "step": 39 }, { "epoch": 0.1968019680196802, "grad_norm": 0.87109375, "learning_rate": 9.978029910109491e-06, "loss": 0.0979, "step": 40 }, { "epoch": 0.2017220172201722, "grad_norm": 0.8203125, "learning_rate": 9.976541991663136e-06, "loss": 0.1086, "step": 41 }, { "epoch": 0.2066420664206642, "grad_norm": 1.2265625, "learning_rate": 9.975005444261312e-06, "loss": 0.1504, "step": 42 }, { "epoch": 0.2115621156211562, "grad_norm": 0.88671875, "learning_rate": 9.973420282918603e-06, "loss": 0.1282, "step": 43 }, { "epoch": 0.21648216482164823, "grad_norm": 0.75, "learning_rate": 9.971786523124623e-06, "loss": 0.0807, "step": 44 }, { "epoch": 0.22140221402214022, "grad_norm": 0.77734375, "learning_rate": 9.970104180843881e-06, "loss": 0.1084, "step": 45 }, { "epoch": 0.22632226322263221, "grad_norm": 1.046875, "learning_rate": 9.968373272515612e-06, "loss": 0.1385, "step": 46 }, { "epoch": 0.23124231242312424, "grad_norm": 1.0078125, "learning_rate": 9.966593815053623e-06, "loss": 0.1538, "step": 47 }, { "epoch": 0.23616236162361623, "grad_norm": 0.82421875, "learning_rate": 9.964765825846123e-06, "loss": 0.1035, "step": 48 }, { "epoch": 0.24108241082410825, "grad_norm": 0.96484375, "learning_rate": 9.962889322755556e-06, "loss": 0.1055, "step": 49 }, { "epoch": 0.24600246002460024, "grad_norm": 0.7734375, "learning_rate": 9.960964324118428e-06, "loss": 0.108, "step": 50 }, { "epoch": 0.25092250922509224, "grad_norm": 0.94921875, "learning_rate": 9.958990848745123e-06, "loss": 0.1407, "step": 51 }, { "epoch": 0.25584255842558423, "grad_norm": 0.890625, "learning_rate": 9.956968915919725e-06, "loss": 0.1281, "step": 52 }, { "epoch": 0.2607626076260763, "grad_norm": 1.2109375, "learning_rate": 9.954898545399822e-06, "loss": 0.1094, "step": 53 }, { "epoch": 0.2656826568265683, "grad_norm": 0.8984375, "learning_rate": 9.952779757416326e-06, "loss": 0.1341, "step": 54 }, { "epoch": 0.27060270602706027, "grad_norm": 1.1484375, "learning_rate": 9.950612572673255e-06, "loss": 0.1497, "step": 55 }, { "epoch": 0.27552275522755226, "grad_norm": 1.0859375, "learning_rate": 9.948397012347553e-06, "loss": 0.146, "step": 56 }, { "epoch": 0.28044280442804426, "grad_norm": 0.94140625, "learning_rate": 9.946133098088868e-06, "loss": 0.1294, "step": 57 }, { "epoch": 0.2853628536285363, "grad_norm": 1.046875, "learning_rate": 9.943820852019345e-06, "loss": 0.1228, "step": 58 }, { "epoch": 0.2902829028290283, "grad_norm": 0.890625, "learning_rate": 9.941460296733409e-06, "loss": 0.0948, "step": 59 }, { "epoch": 0.2952029520295203, "grad_norm": 1.09375, "learning_rate": 9.939051455297548e-06, "loss": 0.1337, "step": 60 }, { "epoch": 0.3001230012300123, "grad_norm": 0.90234375, "learning_rate": 9.936594351250082e-06, "loss": 0.1267, "step": 61 }, { "epoch": 0.3050430504305043, "grad_norm": 0.6640625, "learning_rate": 9.934089008600937e-06, "loss": 0.0852, "step": 62 }, { "epoch": 0.30996309963099633, "grad_norm": 0.89453125, "learning_rate": 9.931535451831408e-06, "loss": 0.0853, "step": 63 }, { "epoch": 0.3148831488314883, "grad_norm": 1.2578125, "learning_rate": 9.928933705893924e-06, "loss": 0.104, "step": 64 }, { "epoch": 0.3198031980319803, "grad_norm": 0.96484375, "learning_rate": 9.926283796211796e-06, "loss": 0.0999, "step": 65 }, { "epoch": 0.3247232472324723, "grad_norm": 0.83203125, "learning_rate": 9.923585748678976e-06, "loss": 0.1169, "step": 66 }, { "epoch": 0.3296432964329643, "grad_norm": 0.984375, "learning_rate": 9.920839589659803e-06, "loss": 0.1413, "step": 67 }, { "epoch": 0.33456334563345635, "grad_norm": 0.90625, "learning_rate": 9.918045345988743e-06, "loss": 0.0995, "step": 68 }, { "epoch": 0.33948339483394835, "grad_norm": 0.95703125, "learning_rate": 9.915203044970125e-06, "loss": 0.1242, "step": 69 }, { "epoch": 0.34440344403444034, "grad_norm": 1.078125, "learning_rate": 9.91231271437788e-06, "loss": 0.0816, "step": 70 }, { "epoch": 0.34932349323493234, "grad_norm": 0.96875, "learning_rate": 9.909374382455268e-06, "loss": 0.1365, "step": 71 }, { "epoch": 0.35424354243542433, "grad_norm": 0.91796875, "learning_rate": 9.906388077914598e-06, "loss": 0.1388, "step": 72 }, { "epoch": 0.3591635916359164, "grad_norm": 1.0546875, "learning_rate": 9.903353829936943e-06, "loss": 0.1291, "step": 73 }, { "epoch": 0.3640836408364084, "grad_norm": 0.8828125, "learning_rate": 9.900271668171879e-06, "loss": 0.1014, "step": 74 }, { "epoch": 0.36900369003690037, "grad_norm": 1.0625, "learning_rate": 9.89714162273716e-06, "loss": 0.1653, "step": 75 }, { "epoch": 0.37392373923739236, "grad_norm": 0.76953125, "learning_rate": 9.893963724218456e-06, "loss": 0.1097, "step": 76 }, { "epoch": 0.37884378843788435, "grad_norm": 1.09375, "learning_rate": 9.890738003669029e-06, "loss": 0.1509, "step": 77 }, { "epoch": 0.3837638376383764, "grad_norm": 1.15625, "learning_rate": 9.887464492609447e-06, "loss": 0.119, "step": 78 }, { "epoch": 0.3886838868388684, "grad_norm": 0.90234375, "learning_rate": 9.884143223027267e-06, "loss": 0.1354, "step": 79 }, { "epoch": 0.3936039360393604, "grad_norm": 0.84375, "learning_rate": 9.880774227376727e-06, "loss": 0.0938, "step": 80 }, { "epoch": 0.3985239852398524, "grad_norm": 0.890625, "learning_rate": 9.877357538578426e-06, "loss": 0.1072, "step": 81 }, { "epoch": 0.4034440344403444, "grad_norm": 0.8125, "learning_rate": 9.873893190018995e-06, "loss": 0.099, "step": 82 }, { "epoch": 0.40836408364083643, "grad_norm": 1.0703125, "learning_rate": 9.870381215550796e-06, "loss": 0.1036, "step": 83 }, { "epoch": 0.4132841328413284, "grad_norm": 0.7734375, "learning_rate": 9.866821649491558e-06, "loss": 0.0779, "step": 84 }, { "epoch": 0.4182041820418204, "grad_norm": 1.140625, "learning_rate": 9.863214526624065e-06, "loss": 0.1597, "step": 85 }, { "epoch": 0.4231242312423124, "grad_norm": 0.96875, "learning_rate": 9.85955988219581e-06, "loss": 0.1283, "step": 86 }, { "epoch": 0.4280442804428044, "grad_norm": 1.1171875, "learning_rate": 9.855857751918644e-06, "loss": 0.1327, "step": 87 }, { "epoch": 0.43296432964329645, "grad_norm": 0.80859375, "learning_rate": 9.852108171968438e-06, "loss": 0.1, "step": 88 }, { "epoch": 0.43788437884378845, "grad_norm": 1.0625, "learning_rate": 9.848311178984718e-06, "loss": 0.1653, "step": 89 }, { "epoch": 0.44280442804428044, "grad_norm": 0.69921875, "learning_rate": 9.844466810070319e-06, "loss": 0.0841, "step": 90 }, { "epoch": 0.44772447724477243, "grad_norm": 0.8828125, "learning_rate": 9.840575102791013e-06, "loss": 0.1193, "step": 91 }, { "epoch": 0.45264452644526443, "grad_norm": 0.9921875, "learning_rate": 9.836636095175148e-06, "loss": 0.1187, "step": 92 }, { "epoch": 0.4575645756457565, "grad_norm": 0.95703125, "learning_rate": 9.832649825713266e-06, "loss": 0.1198, "step": 93 }, { "epoch": 0.46248462484624847, "grad_norm": 0.96875, "learning_rate": 9.828616333357743e-06, "loss": 0.1148, "step": 94 }, { "epoch": 0.46740467404674046, "grad_norm": 0.79296875, "learning_rate": 9.8245356575224e-06, "loss": 0.1074, "step": 95 }, { "epoch": 0.47232472324723246, "grad_norm": 0.7109375, "learning_rate": 9.820407838082109e-06, "loss": 0.1193, "step": 96 }, { "epoch": 0.47724477244772445, "grad_norm": 0.66796875, "learning_rate": 9.816232915372423e-06, "loss": 0.0722, "step": 97 }, { "epoch": 0.4821648216482165, "grad_norm": 0.9765625, "learning_rate": 9.812010930189163e-06, "loss": 0.0929, "step": 98 }, { "epoch": 0.4870848708487085, "grad_norm": 0.89453125, "learning_rate": 9.80774192378803e-06, "loss": 0.0973, "step": 99 }, { "epoch": 0.4920049200492005, "grad_norm": 1.1328125, "learning_rate": 9.803425937884202e-06, "loss": 0.1335, "step": 100 }, { "epoch": 0.4969249692496925, "grad_norm": 0.77734375, "learning_rate": 9.799063014651918e-06, "loss": 0.1169, "step": 101 }, { "epoch": 0.5018450184501845, "grad_norm": 0.93359375, "learning_rate": 9.794653196724077e-06, "loss": 0.1159, "step": 102 }, { "epoch": 0.5018450184501845, "eval_loss": 0.12004748731851578, "eval_runtime": 203.4298, "eval_samples_per_second": 8.042, "eval_steps_per_second": 1.342, "step": 102 }, { "epoch": 0.5067650676506765, "grad_norm": 0.9375, "learning_rate": 9.790196527191811e-06, "loss": 0.1142, "step": 103 }, { "epoch": 0.5116851168511685, "grad_norm": 0.8203125, "learning_rate": 9.785693049604073e-06, "loss": 0.1113, "step": 104 }, { "epoch": 0.5166051660516605, "grad_norm": 0.80859375, "learning_rate": 9.781142807967205e-06, "loss": 0.1035, "step": 105 }, { "epoch": 0.5215252152521526, "grad_norm": 0.91796875, "learning_rate": 9.776545846744509e-06, "loss": 0.1238, "step": 106 }, { "epoch": 0.5264452644526445, "grad_norm": 1.4765625, "learning_rate": 9.771902210855819e-06, "loss": 0.0921, "step": 107 }, { "epoch": 0.5313653136531366, "grad_norm": 1.1015625, "learning_rate": 9.767211945677051e-06, "loss": 0.1459, "step": 108 }, { "epoch": 0.5362853628536285, "grad_norm": 1.0, "learning_rate": 9.762475097039767e-06, "loss": 0.101, "step": 109 }, { "epoch": 0.5412054120541205, "grad_norm": 1.078125, "learning_rate": 9.757691711230728e-06, "loss": 0.1259, "step": 110 }, { "epoch": 0.5461254612546126, "grad_norm": 0.8828125, "learning_rate": 9.752861834991436e-06, "loss": 0.1146, "step": 111 }, { "epoch": 0.5510455104551045, "grad_norm": 1.0078125, "learning_rate": 9.747985515517683e-06, "loss": 0.1609, "step": 112 }, { "epoch": 0.5559655596555966, "grad_norm": 0.94921875, "learning_rate": 9.743062800459089e-06, "loss": 0.1242, "step": 113 }, { "epoch": 0.5608856088560885, "grad_norm": 0.9765625, "learning_rate": 9.738093737918627e-06, "loss": 0.1095, "step": 114 }, { "epoch": 0.5658056580565806, "grad_norm": 1.0, "learning_rate": 9.733078376452172e-06, "loss": 0.1083, "step": 115 }, { "epoch": 0.5707257072570726, "grad_norm": 1.0390625, "learning_rate": 9.728016765068004e-06, "loss": 0.1428, "step": 116 }, { "epoch": 0.5756457564575646, "grad_norm": 0.859375, "learning_rate": 9.722908953226349e-06, "loss": 0.1073, "step": 117 }, { "epoch": 0.5805658056580566, "grad_norm": 0.8125, "learning_rate": 9.717754990838882e-06, "loss": 0.1049, "step": 118 }, { "epoch": 0.5854858548585485, "grad_norm": 1.140625, "learning_rate": 9.712554928268245e-06, "loss": 0.1479, "step": 119 }, { "epoch": 0.5904059040590406, "grad_norm": 0.90234375, "learning_rate": 9.707308816327557e-06, "loss": 0.1195, "step": 120 }, { "epoch": 0.5953259532595326, "grad_norm": 0.82421875, "learning_rate": 9.702016706279914e-06, "loss": 0.1048, "step": 121 }, { "epoch": 0.6002460024600246, "grad_norm": 0.984375, "learning_rate": 9.696678649837883e-06, "loss": 0.1073, "step": 122 }, { "epoch": 0.6051660516605166, "grad_norm": 1.0859375, "learning_rate": 9.691294699163013e-06, "loss": 0.1308, "step": 123 }, { "epoch": 0.6100861008610086, "grad_norm": 0.86328125, "learning_rate": 9.685864906865303e-06, "loss": 0.1248, "step": 124 }, { "epoch": 0.6150061500615006, "grad_norm": 0.8515625, "learning_rate": 9.680389326002708e-06, "loss": 0.1312, "step": 125 }, { "epoch": 0.6199261992619927, "grad_norm": 0.71875, "learning_rate": 9.67486801008061e-06, "loss": 0.0868, "step": 126 }, { "epoch": 0.6248462484624846, "grad_norm": 0.76953125, "learning_rate": 9.669301013051297e-06, "loss": 0.0859, "step": 127 }, { "epoch": 0.6297662976629766, "grad_norm": 0.734375, "learning_rate": 9.663688389313435e-06, "loss": 0.0878, "step": 128 }, { "epoch": 0.6346863468634686, "grad_norm": 1.0234375, "learning_rate": 9.658030193711538e-06, "loss": 0.1229, "step": 129 }, { "epoch": 0.6396063960639606, "grad_norm": 0.97265625, "learning_rate": 9.652326481535434e-06, "loss": 0.1439, "step": 130 }, { "epoch": 0.6445264452644527, "grad_norm": 0.703125, "learning_rate": 9.646577308519719e-06, "loss": 0.1063, "step": 131 }, { "epoch": 0.6494464944649446, "grad_norm": 0.9609375, "learning_rate": 9.640782730843219e-06, "loss": 0.1023, "step": 132 }, { "epoch": 0.6543665436654367, "grad_norm": 0.984375, "learning_rate": 9.634942805128433e-06, "loss": 0.1002, "step": 133 }, { "epoch": 0.6592865928659286, "grad_norm": 0.76953125, "learning_rate": 9.629057588440993e-06, "loss": 0.0978, "step": 134 }, { "epoch": 0.6642066420664207, "grad_norm": 0.8671875, "learning_rate": 9.623127138289087e-06, "loss": 0.1042, "step": 135 }, { "epoch": 0.6691266912669127, "grad_norm": 0.8515625, "learning_rate": 9.617151512622918e-06, "loss": 0.0873, "step": 136 }, { "epoch": 0.6740467404674046, "grad_norm": 0.91796875, "learning_rate": 9.611130769834121e-06, "loss": 0.1118, "step": 137 }, { "epoch": 0.6789667896678967, "grad_norm": 0.9375, "learning_rate": 9.6050649687552e-06, "loss": 0.13, "step": 138 }, { "epoch": 0.6838868388683886, "grad_norm": 0.828125, "learning_rate": 9.598954168658956e-06, "loss": 0.1096, "step": 139 }, { "epoch": 0.6888068880688807, "grad_norm": 0.86328125, "learning_rate": 9.592798429257899e-06, "loss": 0.1219, "step": 140 }, { "epoch": 0.6937269372693727, "grad_norm": 0.95703125, "learning_rate": 9.586597810703674e-06, "loss": 0.1002, "step": 141 }, { "epoch": 0.6986469864698647, "grad_norm": 0.765625, "learning_rate": 9.580352373586468e-06, "loss": 0.0765, "step": 142 }, { "epoch": 0.7035670356703567, "grad_norm": 0.90234375, "learning_rate": 9.574062178934414e-06, "loss": 0.1227, "step": 143 }, { "epoch": 0.7084870848708487, "grad_norm": 0.82421875, "learning_rate": 9.567727288213005e-06, "loss": 0.1052, "step": 144 }, { "epoch": 0.7134071340713407, "grad_norm": 0.87109375, "learning_rate": 9.561347763324484e-06, "loss": 0.0909, "step": 145 }, { "epoch": 0.7183271832718328, "grad_norm": 1.0546875, "learning_rate": 9.554923666607245e-06, "loss": 0.1219, "step": 146 }, { "epoch": 0.7232472324723247, "grad_norm": 0.98046875, "learning_rate": 9.548455060835223e-06, "loss": 0.1077, "step": 147 }, { "epoch": 0.7281672816728167, "grad_norm": 0.94921875, "learning_rate": 9.541942009217273e-06, "loss": 0.132, "step": 148 }, { "epoch": 0.7330873308733087, "grad_norm": 1.03125, "learning_rate": 9.535384575396565e-06, "loss": 0.1462, "step": 149 }, { "epoch": 0.7380073800738007, "grad_norm": 0.79296875, "learning_rate": 9.528782823449954e-06, "loss": 0.0951, "step": 150 }, { "epoch": 0.7429274292742928, "grad_norm": 0.99609375, "learning_rate": 9.522136817887353e-06, "loss": 0.1451, "step": 151 }, { "epoch": 0.7478474784747847, "grad_norm": 0.81640625, "learning_rate": 9.51544662365111e-06, "loss": 0.072, "step": 152 }, { "epoch": 0.7527675276752768, "grad_norm": 0.80078125, "learning_rate": 9.508712306115366e-06, "loss": 0.1202, "step": 153 }, { "epoch": 0.7576875768757687, "grad_norm": 0.890625, "learning_rate": 9.501933931085416e-06, "loss": 0.1068, "step": 154 }, { "epoch": 0.7626076260762608, "grad_norm": 0.8984375, "learning_rate": 9.495111564797073e-06, "loss": 0.1091, "step": 155 }, { "epoch": 0.7675276752767528, "grad_norm": 0.84375, "learning_rate": 9.488245273916016e-06, "loss": 0.0982, "step": 156 }, { "epoch": 0.7724477244772447, "grad_norm": 0.828125, "learning_rate": 9.481335125537138e-06, "loss": 0.1064, "step": 157 }, { "epoch": 0.7773677736777368, "grad_norm": 0.8828125, "learning_rate": 9.47438118718389e-06, "loss": 0.1027, "step": 158 }, { "epoch": 0.7822878228782287, "grad_norm": 0.765625, "learning_rate": 9.467383526807626e-06, "loss": 0.0917, "step": 159 }, { "epoch": 0.7872078720787208, "grad_norm": 0.875, "learning_rate": 9.460342212786933e-06, "loss": 0.0975, "step": 160 }, { "epoch": 0.7921279212792128, "grad_norm": 1.03125, "learning_rate": 9.453257313926969e-06, "loss": 0.0837, "step": 161 }, { "epoch": 0.7970479704797048, "grad_norm": 0.68359375, "learning_rate": 9.44612889945878e-06, "loss": 0.093, "step": 162 }, { "epoch": 0.8019680196801968, "grad_norm": 0.7578125, "learning_rate": 9.43895703903864e-06, "loss": 0.1094, "step": 163 }, { "epoch": 0.8068880688806888, "grad_norm": 1.015625, "learning_rate": 9.431741802747355e-06, "loss": 0.1081, "step": 164 }, { "epoch": 0.8118081180811808, "grad_norm": 0.96875, "learning_rate": 9.424483261089584e-06, "loss": 0.1207, "step": 165 }, { "epoch": 0.8167281672816729, "grad_norm": 0.67578125, "learning_rate": 9.417181484993154e-06, "loss": 0.0717, "step": 166 }, { "epoch": 0.8216482164821648, "grad_norm": 0.99609375, "learning_rate": 9.40983654580836e-06, "loss": 0.1013, "step": 167 }, { "epoch": 0.8265682656826568, "grad_norm": 0.87890625, "learning_rate": 9.402448515307274e-06, "loss": 0.1122, "step": 168 }, { "epoch": 0.8314883148831488, "grad_norm": 0.99609375, "learning_rate": 9.395017465683036e-06, "loss": 0.1234, "step": 169 }, { "epoch": 0.8364083640836408, "grad_norm": 0.85546875, "learning_rate": 9.387543469549156e-06, "loss": 0.1259, "step": 170 }, { "epoch": 0.8413284132841329, "grad_norm": 0.78515625, "learning_rate": 9.380026599938804e-06, "loss": 0.1007, "step": 171 }, { "epoch": 0.8462484624846248, "grad_norm": 0.95703125, "learning_rate": 9.372466930304091e-06, "loss": 0.1148, "step": 172 }, { "epoch": 0.8511685116851169, "grad_norm": 0.765625, "learning_rate": 9.364864534515353e-06, "loss": 0.11, "step": 173 }, { "epoch": 0.8560885608856088, "grad_norm": 1.0078125, "learning_rate": 9.357219486860434e-06, "loss": 0.1255, "step": 174 }, { "epoch": 0.8610086100861009, "grad_norm": 0.8671875, "learning_rate": 9.349531862043952e-06, "loss": 0.0973, "step": 175 }, { "epoch": 0.8659286592865929, "grad_norm": 0.8671875, "learning_rate": 9.34180173518658e-06, "loss": 0.0872, "step": 176 }, { "epoch": 0.8708487084870848, "grad_norm": 0.796875, "learning_rate": 9.3340291818243e-06, "loss": 0.089, "step": 177 }, { "epoch": 0.8757687576875769, "grad_norm": 1.0078125, "learning_rate": 9.32621427790767e-06, "loss": 0.1148, "step": 178 }, { "epoch": 0.8806888068880688, "grad_norm": 0.78515625, "learning_rate": 9.318357099801087e-06, "loss": 0.0896, "step": 179 }, { "epoch": 0.8856088560885609, "grad_norm": 0.9296875, "learning_rate": 9.310457724282034e-06, "loss": 0.1104, "step": 180 }, { "epoch": 0.8905289052890529, "grad_norm": 0.859375, "learning_rate": 9.302516228540328e-06, "loss": 0.1115, "step": 181 }, { "epoch": 0.8954489544895449, "grad_norm": 0.921875, "learning_rate": 9.294532690177373e-06, "loss": 0.12, "step": 182 }, { "epoch": 0.9003690036900369, "grad_norm": 1.0625, "learning_rate": 9.286507187205399e-06, "loss": 0.137, "step": 183 }, { "epoch": 0.9052890528905289, "grad_norm": 0.80859375, "learning_rate": 9.278439798046697e-06, "loss": 0.1273, "step": 184 }, { "epoch": 0.9102091020910209, "grad_norm": 0.96484375, "learning_rate": 9.270330601532855e-06, "loss": 0.1339, "step": 185 }, { "epoch": 0.915129151291513, "grad_norm": 0.90234375, "learning_rate": 9.262179676903986e-06, "loss": 0.1011, "step": 186 }, { "epoch": 0.9200492004920049, "grad_norm": 0.90234375, "learning_rate": 9.25398710380796e-06, "loss": 0.1096, "step": 187 }, { "epoch": 0.9249692496924969, "grad_norm": 0.8125, "learning_rate": 9.245752962299612e-06, "loss": 0.1316, "step": 188 }, { "epoch": 0.9298892988929889, "grad_norm": 0.93359375, "learning_rate": 9.237477332839975e-06, "loss": 0.0908, "step": 189 }, { "epoch": 0.9348093480934809, "grad_norm": 0.97265625, "learning_rate": 9.229160296295488e-06, "loss": 0.138, "step": 190 }, { "epoch": 0.939729397293973, "grad_norm": 0.81640625, "learning_rate": 9.220801933937199e-06, "loss": 0.1259, "step": 191 }, { "epoch": 0.9446494464944649, "grad_norm": 0.92578125, "learning_rate": 9.212402327439982e-06, "loss": 0.117, "step": 192 }, { "epoch": 0.949569495694957, "grad_norm": 1.0859375, "learning_rate": 9.20396155888173e-06, "loss": 0.1405, "step": 193 }, { "epoch": 0.9544895448954489, "grad_norm": 1.015625, "learning_rate": 9.19547971074256e-06, "loss": 0.1154, "step": 194 }, { "epoch": 0.959409594095941, "grad_norm": 0.8828125, "learning_rate": 9.186956865904004e-06, "loss": 0.1203, "step": 195 }, { "epoch": 0.964329643296433, "grad_norm": 0.71875, "learning_rate": 9.178393107648193e-06, "loss": 0.1009, "step": 196 }, { "epoch": 0.9692496924969249, "grad_norm": 0.859375, "learning_rate": 9.169788519657056e-06, "loss": 0.1023, "step": 197 }, { "epoch": 0.974169741697417, "grad_norm": 1.015625, "learning_rate": 9.161143186011492e-06, "loss": 0.1402, "step": 198 }, { "epoch": 0.9790897908979089, "grad_norm": 0.828125, "learning_rate": 9.15245719119055e-06, "loss": 0.1065, "step": 199 }, { "epoch": 0.984009840098401, "grad_norm": 0.8515625, "learning_rate": 9.143730620070609e-06, "loss": 0.097, "step": 200 }, { "epoch": 0.988929889298893, "grad_norm": 0.87109375, "learning_rate": 9.134963557924543e-06, "loss": 0.0979, "step": 201 }, { "epoch": 0.993849938499385, "grad_norm": 0.87109375, "learning_rate": 9.12615609042089e-06, "loss": 0.1012, "step": 202 }, { "epoch": 0.998769987699877, "grad_norm": 0.7578125, "learning_rate": 9.11730830362301e-06, "loss": 0.0764, "step": 203 }, { "epoch": 1.0030750307503076, "grad_norm": 1.9296875, "learning_rate": 9.10842028398826e-06, "loss": 0.1883, "step": 204 }, { "epoch": 1.0030750307503076, "eval_loss": 0.11158129572868347, "eval_runtime": 204.933, "eval_samples_per_second": 7.983, "eval_steps_per_second": 1.332, "step": 204 }, { "epoch": 1.0079950799507995, "grad_norm": 1.09375, "learning_rate": 9.099492118367123e-06, "loss": 0.1227, "step": 205 }, { "epoch": 1.0129151291512914, "grad_norm": 0.80078125, "learning_rate": 9.090523894002386e-06, "loss": 0.0759, "step": 206 }, { "epoch": 1.0178351783517836, "grad_norm": 1.234375, "learning_rate": 9.081515698528267e-06, "loss": 0.1026, "step": 207 }, { "epoch": 1.0227552275522755, "grad_norm": 1.296875, "learning_rate": 9.072467619969574e-06, "loss": 0.1144, "step": 208 }, { "epoch": 1.0276752767527675, "grad_norm": 1.2578125, "learning_rate": 9.06337974674083e-06, "loss": 0.1129, "step": 209 }, { "epoch": 1.0325953259532596, "grad_norm": 1.5078125, "learning_rate": 9.054252167645426e-06, "loss": 0.081, "step": 210 }, { "epoch": 1.0375153751537516, "grad_norm": 1.1640625, "learning_rate": 9.045084971874738e-06, "loss": 0.1211, "step": 211 }, { "epoch": 1.0424354243542435, "grad_norm": 1.0546875, "learning_rate": 9.035878249007264e-06, "loss": 0.0857, "step": 212 }, { "epoch": 1.0473554735547355, "grad_norm": 1.0703125, "learning_rate": 9.026632089007745e-06, "loss": 0.1407, "step": 213 }, { "epoch": 1.0522755227552276, "grad_norm": 1.1484375, "learning_rate": 9.017346582226289e-06, "loss": 0.0835, "step": 214 }, { "epoch": 1.0571955719557196, "grad_norm": 1.21875, "learning_rate": 9.008021819397488e-06, "loss": 0.1263, "step": 215 }, { "epoch": 1.0621156211562115, "grad_norm": 1.2109375, "learning_rate": 8.998657891639523e-06, "loss": 0.0747, "step": 216 }, { "epoch": 1.0670356703567037, "grad_norm": 0.89453125, "learning_rate": 8.989254890453289e-06, "loss": 0.09, "step": 217 }, { "epoch": 1.0719557195571956, "grad_norm": 1.0, "learning_rate": 8.979812907721485e-06, "loss": 0.0924, "step": 218 }, { "epoch": 1.0768757687576875, "grad_norm": 1.1171875, "learning_rate": 8.97033203570773e-06, "loss": 0.1017, "step": 219 }, { "epoch": 1.0817958179581795, "grad_norm": 0.9921875, "learning_rate": 8.960812367055646e-06, "loss": 0.1033, "step": 220 }, { "epoch": 1.0867158671586716, "grad_norm": 1.3515625, "learning_rate": 8.951253994787975e-06, "loss": 0.1425, "step": 221 }, { "epoch": 1.0916359163591636, "grad_norm": 1.140625, "learning_rate": 8.941657012305644e-06, "loss": 0.1076, "step": 222 }, { "epoch": 1.0965559655596555, "grad_norm": 0.9296875, "learning_rate": 8.93202151338687e-06, "loss": 0.081, "step": 223 }, { "epoch": 1.1014760147601477, "grad_norm": 0.90234375, "learning_rate": 8.922347592186237e-06, "loss": 0.096, "step": 224 }, { "epoch": 1.1063960639606396, "grad_norm": 0.79296875, "learning_rate": 8.912635343233784e-06, "loss": 0.0681, "step": 225 }, { "epoch": 1.1113161131611315, "grad_norm": 1.28125, "learning_rate": 8.902884861434066e-06, "loss": 0.117, "step": 226 }, { "epoch": 1.1162361623616237, "grad_norm": 1.2734375, "learning_rate": 8.89309624206524e-06, "loss": 0.1133, "step": 227 }, { "epoch": 1.1211562115621156, "grad_norm": 0.96484375, "learning_rate": 8.883269580778127e-06, "loss": 0.0933, "step": 228 }, { "epoch": 1.1260762607626076, "grad_norm": 1.2265625, "learning_rate": 8.873404973595284e-06, "loss": 0.1243, "step": 229 }, { "epoch": 1.1309963099630997, "grad_norm": 1.6796875, "learning_rate": 8.863502516910058e-06, "loss": 0.0959, "step": 230 }, { "epoch": 1.1359163591635917, "grad_norm": 1.0390625, "learning_rate": 8.853562307485649e-06, "loss": 0.1095, "step": 231 }, { "epoch": 1.1408364083640836, "grad_norm": 1.3984375, "learning_rate": 8.84358444245416e-06, "loss": 0.1128, "step": 232 }, { "epoch": 1.1457564575645756, "grad_norm": 1.046875, "learning_rate": 8.833569019315654e-06, "loss": 0.098, "step": 233 }, { "epoch": 1.1506765067650677, "grad_norm": 1.6328125, "learning_rate": 8.8235161359372e-06, "loss": 0.1001, "step": 234 }, { "epoch": 1.1555965559655597, "grad_norm": 1.1796875, "learning_rate": 8.81342589055191e-06, "loss": 0.1213, "step": 235 }, { "epoch": 1.1605166051660516, "grad_norm": 1.078125, "learning_rate": 8.803298381757987e-06, "loss": 0.0891, "step": 236 }, { "epoch": 1.1654366543665438, "grad_norm": 0.9375, "learning_rate": 8.793133708517759e-06, "loss": 0.0878, "step": 237 }, { "epoch": 1.1703567035670357, "grad_norm": 1.0625, "learning_rate": 8.782931970156708e-06, "loss": 0.0967, "step": 238 }, { "epoch": 1.1752767527675276, "grad_norm": 1.171875, "learning_rate": 8.772693266362504e-06, "loss": 0.0976, "step": 239 }, { "epoch": 1.1801968019680196, "grad_norm": 0.875, "learning_rate": 8.762417697184034e-06, "loss": 0.0737, "step": 240 }, { "epoch": 1.1851168511685117, "grad_norm": 1.109375, "learning_rate": 8.752105363030414e-06, "loss": 0.1093, "step": 241 }, { "epoch": 1.1900369003690037, "grad_norm": 1.0703125, "learning_rate": 8.741756364670018e-06, "loss": 0.0853, "step": 242 }, { "epoch": 1.1949569495694956, "grad_norm": 1.1640625, "learning_rate": 8.731370803229488e-06, "loss": 0.0941, "step": 243 }, { "epoch": 1.1998769987699878, "grad_norm": 1.125, "learning_rate": 8.720948780192747e-06, "loss": 0.0935, "step": 244 }, { "epoch": 1.2047970479704797, "grad_norm": 1.109375, "learning_rate": 8.710490397400007e-06, "loss": 0.1011, "step": 245 }, { "epoch": 1.2097170971709716, "grad_norm": 1.171875, "learning_rate": 8.699995757046773e-06, "loss": 0.1141, "step": 246 }, { "epoch": 1.2146371463714638, "grad_norm": 1.421875, "learning_rate": 8.689464961682853e-06, "loss": 0.1281, "step": 247 }, { "epoch": 1.2195571955719557, "grad_norm": 1.1953125, "learning_rate": 8.678898114211338e-06, "loss": 0.1005, "step": 248 }, { "epoch": 1.2244772447724477, "grad_norm": 1.1171875, "learning_rate": 8.668295317887615e-06, "loss": 0.1019, "step": 249 }, { "epoch": 1.2293972939729398, "grad_norm": 1.0703125, "learning_rate": 8.657656676318346e-06, "loss": 0.1101, "step": 250 }, { "epoch": 1.2343173431734318, "grad_norm": 1.2265625, "learning_rate": 8.646982293460461e-06, "loss": 0.1366, "step": 251 }, { "epoch": 1.2392373923739237, "grad_norm": 1.21875, "learning_rate": 8.63627227362014e-06, "loss": 0.0937, "step": 252 }, { "epoch": 1.2441574415744157, "grad_norm": 1.2265625, "learning_rate": 8.625526721451799e-06, "loss": 0.1273, "step": 253 }, { "epoch": 1.2490774907749078, "grad_norm": 1.125, "learning_rate": 8.614745741957054e-06, "loss": 0.097, "step": 254 }, { "epoch": 1.2539975399753998, "grad_norm": 1.4453125, "learning_rate": 8.603929440483714e-06, "loss": 0.1064, "step": 255 }, { "epoch": 1.2589175891758917, "grad_norm": 1.1953125, "learning_rate": 8.593077922724733e-06, "loss": 0.118, "step": 256 }, { "epoch": 1.2638376383763839, "grad_norm": 1.0625, "learning_rate": 8.582191294717192e-06, "loss": 0.0993, "step": 257 }, { "epoch": 1.2687576875768758, "grad_norm": 1.34375, "learning_rate": 8.571269662841253e-06, "loss": 0.0659, "step": 258 }, { "epoch": 1.2736777367773677, "grad_norm": 1.3828125, "learning_rate": 8.560313133819124e-06, "loss": 0.12, "step": 259 }, { "epoch": 1.2785977859778597, "grad_norm": 1.390625, "learning_rate": 8.549321814714018e-06, "loss": 0.0868, "step": 260 }, { "epoch": 1.2835178351783518, "grad_norm": 1.2734375, "learning_rate": 8.538295812929096e-06, "loss": 0.0902, "step": 261 }, { "epoch": 1.2884378843788438, "grad_norm": 1.09375, "learning_rate": 8.527235236206438e-06, "loss": 0.0966, "step": 262 }, { "epoch": 1.293357933579336, "grad_norm": 0.984375, "learning_rate": 8.516140192625962e-06, "loss": 0.0993, "step": 263 }, { "epoch": 1.2982779827798279, "grad_norm": 0.97265625, "learning_rate": 8.5050107906044e-06, "loss": 0.0865, "step": 264 }, { "epoch": 1.3031980319803198, "grad_norm": 1.1640625, "learning_rate": 8.49384713889421e-06, "loss": 0.089, "step": 265 }, { "epoch": 1.3081180811808117, "grad_norm": 1.046875, "learning_rate": 8.482649346582529e-06, "loss": 0.0956, "step": 266 }, { "epoch": 1.3130381303813037, "grad_norm": 1.1953125, "learning_rate": 8.471417523090109e-06, "loss": 0.0985, "step": 267 }, { "epoch": 1.3179581795817958, "grad_norm": 0.875, "learning_rate": 8.460151778170231e-06, "loss": 0.0789, "step": 268 }, { "epoch": 1.3228782287822878, "grad_norm": 1.3515625, "learning_rate": 8.448852221907657e-06, "loss": 0.0995, "step": 269 }, { "epoch": 1.32779827798278, "grad_norm": 1.1328125, "learning_rate": 8.43751896471753e-06, "loss": 0.1006, "step": 270 }, { "epoch": 1.3327183271832719, "grad_norm": 1.1328125, "learning_rate": 8.426152117344314e-06, "loss": 0.1119, "step": 271 }, { "epoch": 1.3376383763837638, "grad_norm": 1.3359375, "learning_rate": 8.414751790860697e-06, "loss": 0.1329, "step": 272 }, { "epoch": 1.3425584255842558, "grad_norm": 1.0234375, "learning_rate": 8.40331809666652e-06, "loss": 0.0863, "step": 273 }, { "epoch": 1.347478474784748, "grad_norm": 1.265625, "learning_rate": 8.391851146487675e-06, "loss": 0.1255, "step": 274 }, { "epoch": 1.3523985239852399, "grad_norm": 1.453125, "learning_rate": 8.380351052375023e-06, "loss": 0.1031, "step": 275 }, { "epoch": 1.3573185731857318, "grad_norm": 1.03125, "learning_rate": 8.368817926703296e-06, "loss": 0.0798, "step": 276 }, { "epoch": 1.362238622386224, "grad_norm": 1.28125, "learning_rate": 8.357251882169993e-06, "loss": 0.0923, "step": 277 }, { "epoch": 1.367158671586716, "grad_norm": 1.0859375, "learning_rate": 8.345653031794292e-06, "loss": 0.0831, "step": 278 }, { "epoch": 1.3720787207872078, "grad_norm": 1.09375, "learning_rate": 8.33402148891593e-06, "loss": 0.071, "step": 279 }, { "epoch": 1.3769987699876998, "grad_norm": 1.1015625, "learning_rate": 8.32235736719411e-06, "loss": 0.0697, "step": 280 }, { "epoch": 1.381918819188192, "grad_norm": 0.97265625, "learning_rate": 8.310660780606376e-06, "loss": 0.0704, "step": 281 }, { "epoch": 1.3868388683886839, "grad_norm": 1.34375, "learning_rate": 8.298931843447517e-06, "loss": 0.1128, "step": 282 }, { "epoch": 1.391758917589176, "grad_norm": 1.2421875, "learning_rate": 8.28717067032843e-06, "loss": 0.1112, "step": 283 }, { "epoch": 1.396678966789668, "grad_norm": 0.90625, "learning_rate": 8.275377376175015e-06, "loss": 0.0798, "step": 284 }, { "epoch": 1.40159901599016, "grad_norm": 1.1484375, "learning_rate": 8.263552076227048e-06, "loss": 0.1072, "step": 285 }, { "epoch": 1.4065190651906518, "grad_norm": 1.0390625, "learning_rate": 8.251694886037052e-06, "loss": 0.0953, "step": 286 }, { "epoch": 1.4114391143911438, "grad_norm": 1.21875, "learning_rate": 8.239805921469168e-06, "loss": 0.1145, "step": 287 }, { "epoch": 1.416359163591636, "grad_norm": 1.4453125, "learning_rate": 8.227885298698029e-06, "loss": 0.1047, "step": 288 }, { "epoch": 1.4212792127921279, "grad_norm": 1.046875, "learning_rate": 8.21593313420762e-06, "loss": 0.0905, "step": 289 }, { "epoch": 1.42619926199262, "grad_norm": 0.9609375, "learning_rate": 8.203949544790131e-06, "loss": 0.1013, "step": 290 }, { "epoch": 1.431119311193112, "grad_norm": 1.2109375, "learning_rate": 8.19193464754484e-06, "loss": 0.085, "step": 291 }, { "epoch": 1.436039360393604, "grad_norm": 1.2421875, "learning_rate": 8.179888559876943e-06, "loss": 0.1029, "step": 292 }, { "epoch": 1.4409594095940959, "grad_norm": 1.1796875, "learning_rate": 8.16781139949642e-06, "loss": 0.1262, "step": 293 }, { "epoch": 1.445879458794588, "grad_norm": 1.421875, "learning_rate": 8.155703284416884e-06, "loss": 0.125, "step": 294 }, { "epoch": 1.45079950799508, "grad_norm": 0.8984375, "learning_rate": 8.143564332954426e-06, "loss": 0.0796, "step": 295 }, { "epoch": 1.455719557195572, "grad_norm": 1.4453125, "learning_rate": 8.131394663726452e-06, "loss": 0.1024, "step": 296 }, { "epoch": 1.460639606396064, "grad_norm": 1.109375, "learning_rate": 8.119194395650545e-06, "loss": 0.0994, "step": 297 }, { "epoch": 1.465559655596556, "grad_norm": 1.15625, "learning_rate": 8.106963647943273e-06, "loss": 0.1104, "step": 298 }, { "epoch": 1.470479704797048, "grad_norm": 1.359375, "learning_rate": 8.09470254011905e-06, "loss": 0.1446, "step": 299 }, { "epoch": 1.4753997539975399, "grad_norm": 1.0078125, "learning_rate": 8.082411191988956e-06, "loss": 0.1061, "step": 300 }, { "epoch": 1.480319803198032, "grad_norm": 0.953125, "learning_rate": 8.070089723659567e-06, "loss": 0.074, "step": 301 }, { "epoch": 1.485239852398524, "grad_norm": 1.296875, "learning_rate": 8.057738255531781e-06, "loss": 0.0852, "step": 302 }, { "epoch": 1.4901599015990161, "grad_norm": 1.015625, "learning_rate": 8.045356908299647e-06, "loss": 0.091, "step": 303 }, { "epoch": 1.495079950799508, "grad_norm": 1.0, "learning_rate": 8.032945802949179e-06, "loss": 0.0849, "step": 304 }, { "epoch": 1.5, "grad_norm": 0.98828125, "learning_rate": 8.02050506075718e-06, "loss": 0.1125, "step": 305 }, { "epoch": 1.504920049200492, "grad_norm": 0.8984375, "learning_rate": 8.008034803290045e-06, "loss": 0.086, "step": 306 }, { "epoch": 1.504920049200492, "eval_loss": 0.10929346084594727, "eval_runtime": 203.7138, "eval_samples_per_second": 8.031, "eval_steps_per_second": 1.34, "step": 306 }, { "epoch": 1.5098400984009839, "grad_norm": 1.0390625, "learning_rate": 7.995535152402592e-06, "loss": 0.1118, "step": 307 }, { "epoch": 1.514760147601476, "grad_norm": 1.203125, "learning_rate": 7.983006230236855e-06, "loss": 0.0959, "step": 308 }, { "epoch": 1.519680196801968, "grad_norm": 1.234375, "learning_rate": 7.970448159220897e-06, "loss": 0.0916, "step": 309 }, { "epoch": 1.5246002460024601, "grad_norm": 1.21875, "learning_rate": 7.957861062067614e-06, "loss": 0.1117, "step": 310 }, { "epoch": 1.529520295202952, "grad_norm": 0.87109375, "learning_rate": 7.945245061773531e-06, "loss": 0.073, "step": 311 }, { "epoch": 1.534440344403444, "grad_norm": 0.96875, "learning_rate": 7.93260028161761e-06, "loss": 0.072, "step": 312 }, { "epoch": 1.539360393603936, "grad_norm": 1.046875, "learning_rate": 7.919926845160037e-06, "loss": 0.1082, "step": 313 }, { "epoch": 1.5442804428044279, "grad_norm": 0.9375, "learning_rate": 7.907224876241015e-06, "loss": 0.0828, "step": 314 }, { "epoch": 1.54920049200492, "grad_norm": 0.9296875, "learning_rate": 7.894494498979558e-06, "loss": 0.0924, "step": 315 }, { "epoch": 1.5541205412054122, "grad_norm": 1.203125, "learning_rate": 7.881735837772274e-06, "loss": 0.1099, "step": 316 }, { "epoch": 1.5590405904059041, "grad_norm": 0.984375, "learning_rate": 7.868949017292153e-06, "loss": 0.0909, "step": 317 }, { "epoch": 1.563960639606396, "grad_norm": 1.09375, "learning_rate": 7.856134162487346e-06, "loss": 0.0986, "step": 318 }, { "epoch": 1.568880688806888, "grad_norm": 0.94140625, "learning_rate": 7.843291398579946e-06, "loss": 0.1045, "step": 319 }, { "epoch": 1.57380073800738, "grad_norm": 1.0234375, "learning_rate": 7.830420851064767e-06, "loss": 0.0971, "step": 320 }, { "epoch": 1.5787207872078721, "grad_norm": 1.03125, "learning_rate": 7.817522645708106e-06, "loss": 0.0974, "step": 321 }, { "epoch": 1.583640836408364, "grad_norm": 1.015625, "learning_rate": 7.804596908546529e-06, "loss": 0.1178, "step": 322 }, { "epoch": 1.5885608856088562, "grad_norm": 1.1015625, "learning_rate": 7.791643765885632e-06, "loss": 0.093, "step": 323 }, { "epoch": 1.5934809348093482, "grad_norm": 1.015625, "learning_rate": 7.778663344298804e-06, "loss": 0.0798, "step": 324 }, { "epoch": 1.59840098400984, "grad_norm": 0.8828125, "learning_rate": 7.765655770625997e-06, "loss": 0.0991, "step": 325 }, { "epoch": 1.603321033210332, "grad_norm": 0.93359375, "learning_rate": 7.752621171972486e-06, "loss": 0.0673, "step": 326 }, { "epoch": 1.608241082410824, "grad_norm": 0.94140625, "learning_rate": 7.739559675707615e-06, "loss": 0.094, "step": 327 }, { "epoch": 1.6131611316113161, "grad_norm": 1.15625, "learning_rate": 7.726471409463572e-06, "loss": 0.1286, "step": 328 }, { "epoch": 1.618081180811808, "grad_norm": 1.3125, "learning_rate": 7.713356501134124e-06, "loss": 0.0938, "step": 329 }, { "epoch": 1.6230012300123002, "grad_norm": 1.203125, "learning_rate": 7.70021507887338e-06, "loss": 0.1141, "step": 330 }, { "epoch": 1.6279212792127922, "grad_norm": 0.71875, "learning_rate": 7.687047271094528e-06, "loss": 0.0708, "step": 331 }, { "epoch": 1.632841328413284, "grad_norm": 0.96875, "learning_rate": 7.673853206468593e-06, "loss": 0.0816, "step": 332 }, { "epoch": 1.637761377613776, "grad_norm": 1.3515625, "learning_rate": 7.660633013923164e-06, "loss": 0.1224, "step": 333 }, { "epoch": 1.642681426814268, "grad_norm": 1.265625, "learning_rate": 7.64738682264115e-06, "loss": 0.1012, "step": 334 }, { "epoch": 1.6476014760147601, "grad_norm": 0.8984375, "learning_rate": 7.634114762059504e-06, "loss": 0.103, "step": 335 }, { "epoch": 1.6525215252152523, "grad_norm": 0.9375, "learning_rate": 7.62081696186797e-06, "loss": 0.0773, "step": 336 }, { "epoch": 1.6574415744157442, "grad_norm": 0.98828125, "learning_rate": 7.607493552007805e-06, "loss": 0.1176, "step": 337 }, { "epoch": 1.6623616236162362, "grad_norm": 0.9765625, "learning_rate": 7.5941446626705175e-06, "loss": 0.0996, "step": 338 }, { "epoch": 1.6672816728167281, "grad_norm": 0.96484375, "learning_rate": 7.580770424296591e-06, "loss": 0.1001, "step": 339 }, { "epoch": 1.67220172201722, "grad_norm": 0.8359375, "learning_rate": 7.56737096757421e-06, "loss": 0.1052, "step": 340 }, { "epoch": 1.6771217712177122, "grad_norm": 1.28125, "learning_rate": 7.553946423437988e-06, "loss": 0.1104, "step": 341 }, { "epoch": 1.6820418204182042, "grad_norm": 1.078125, "learning_rate": 7.540496923067675e-06, "loss": 0.1092, "step": 342 }, { "epoch": 1.6869618696186963, "grad_norm": 1.1640625, "learning_rate": 7.527022597886895e-06, "loss": 0.0994, "step": 343 }, { "epoch": 1.6918819188191883, "grad_norm": 1.484375, "learning_rate": 7.513523579561839e-06, "loss": 0.0787, "step": 344 }, { "epoch": 1.6968019680196802, "grad_norm": 1.125, "learning_rate": 7.500000000000001e-06, "loss": 0.0986, "step": 345 }, { "epoch": 1.7017220172201721, "grad_norm": 1.0546875, "learning_rate": 7.486451991348872e-06, "loss": 0.0806, "step": 346 }, { "epoch": 1.706642066420664, "grad_norm": 1.2890625, "learning_rate": 7.472879685994658e-06, "loss": 0.1059, "step": 347 }, { "epoch": 1.7115621156211562, "grad_norm": 1.0234375, "learning_rate": 7.459283216560982e-06, "loss": 0.0947, "step": 348 }, { "epoch": 1.7164821648216482, "grad_norm": 1.015625, "learning_rate": 7.445662715907591e-06, "loss": 0.0897, "step": 349 }, { "epoch": 1.7214022140221403, "grad_norm": 1.046875, "learning_rate": 7.432018317129056e-06, "loss": 0.1083, "step": 350 }, { "epoch": 1.7263222632226323, "grad_norm": 0.8984375, "learning_rate": 7.418350153553471e-06, "loss": 0.07, "step": 351 }, { "epoch": 1.7312423124231242, "grad_norm": 1.0703125, "learning_rate": 7.40465835874115e-06, "loss": 0.0942, "step": 352 }, { "epoch": 1.7361623616236161, "grad_norm": 0.9765625, "learning_rate": 7.390943066483327e-06, "loss": 0.0881, "step": 353 }, { "epoch": 1.741082410824108, "grad_norm": 1.046875, "learning_rate": 7.377204410800839e-06, "loss": 0.1029, "step": 354 }, { "epoch": 1.7460024600246002, "grad_norm": 0.96484375, "learning_rate": 7.363442525942827e-06, "loss": 0.093, "step": 355 }, { "epoch": 1.7509225092250924, "grad_norm": 1.1328125, "learning_rate": 7.349657546385414e-06, "loss": 0.0857, "step": 356 }, { "epoch": 1.7558425584255843, "grad_norm": 0.9375, "learning_rate": 7.335849606830402e-06, "loss": 0.1016, "step": 357 }, { "epoch": 1.7607626076260763, "grad_norm": 1.015625, "learning_rate": 7.322018842203942e-06, "loss": 0.1107, "step": 358 }, { "epoch": 1.7656826568265682, "grad_norm": 1.15625, "learning_rate": 7.308165387655231e-06, "loss": 0.1411, "step": 359 }, { "epoch": 1.7706027060270602, "grad_norm": 0.89453125, "learning_rate": 7.294289378555179e-06, "loss": 0.1037, "step": 360 }, { "epoch": 1.7755227552275523, "grad_norm": 0.87109375, "learning_rate": 7.2803909504950935e-06, "loss": 0.0838, "step": 361 }, { "epoch": 1.7804428044280443, "grad_norm": 0.99609375, "learning_rate": 7.266470239285347e-06, "loss": 0.0947, "step": 362 }, { "epoch": 1.7853628536285364, "grad_norm": 1.0625, "learning_rate": 7.252527380954062e-06, "loss": 0.0904, "step": 363 }, { "epoch": 1.7902829028290284, "grad_norm": 1.1640625, "learning_rate": 7.238562511745768e-06, "loss": 0.0877, "step": 364 }, { "epoch": 1.7952029520295203, "grad_norm": 1.015625, "learning_rate": 7.2245757681200835e-06, "loss": 0.0871, "step": 365 }, { "epoch": 1.8001230012300122, "grad_norm": 1.09375, "learning_rate": 7.210567286750368e-06, "loss": 0.0898, "step": 366 }, { "epoch": 1.8050430504305042, "grad_norm": 1.140625, "learning_rate": 7.196537204522401e-06, "loss": 0.0826, "step": 367 }, { "epoch": 1.8099630996309963, "grad_norm": 1.109375, "learning_rate": 7.182485658533036e-06, "loss": 0.0991, "step": 368 }, { "epoch": 1.8148831488314883, "grad_norm": 1.0078125, "learning_rate": 7.168412786088857e-06, "loss": 0.0775, "step": 369 }, { "epoch": 1.8198031980319804, "grad_norm": 0.8984375, "learning_rate": 7.1543187247048525e-06, "loss": 0.0915, "step": 370 }, { "epoch": 1.8247232472324724, "grad_norm": 1.109375, "learning_rate": 7.140203612103052e-06, "loss": 0.0983, "step": 371 }, { "epoch": 1.8296432964329643, "grad_norm": 0.89453125, "learning_rate": 7.1260675862111986e-06, "loss": 0.086, "step": 372 }, { "epoch": 1.8345633456334562, "grad_norm": 1.34375, "learning_rate": 7.111910785161381e-06, "loss": 0.1281, "step": 373 }, { "epoch": 1.8394833948339482, "grad_norm": 1.3203125, "learning_rate": 7.0977333472887076e-06, "loss": 0.0902, "step": 374 }, { "epoch": 1.8444034440344403, "grad_norm": 1.234375, "learning_rate": 7.083535411129934e-06, "loss": 0.0996, "step": 375 }, { "epoch": 1.8493234932349325, "grad_norm": 1.09375, "learning_rate": 7.069317115422121e-06, "loss": 0.0887, "step": 376 }, { "epoch": 1.8542435424354244, "grad_norm": 1.15625, "learning_rate": 7.055078599101275e-06, "loss": 0.1046, "step": 377 }, { "epoch": 1.8591635916359164, "grad_norm": 1.09375, "learning_rate": 7.040820001300992e-06, "loss": 0.1243, "step": 378 }, { "epoch": 1.8640836408364083, "grad_norm": 1.1328125, "learning_rate": 7.026541461351092e-06, "loss": 0.0664, "step": 379 }, { "epoch": 1.8690036900369003, "grad_norm": 1.0859375, "learning_rate": 7.01224311877627e-06, "loss": 0.1314, "step": 380 }, { "epoch": 1.8739237392373924, "grad_norm": 0.82421875, "learning_rate": 6.997925113294717e-06, "loss": 0.0811, "step": 381 }, { "epoch": 1.8788437884378844, "grad_norm": 0.9921875, "learning_rate": 6.983587584816769e-06, "loss": 0.0776, "step": 382 }, { "epoch": 1.8837638376383765, "grad_norm": 1.359375, "learning_rate": 6.969230673443531e-06, "loss": 0.116, "step": 383 }, { "epoch": 1.8886838868388685, "grad_norm": 0.9765625, "learning_rate": 6.9548545194655115e-06, "loss": 0.0983, "step": 384 }, { "epoch": 1.8936039360393604, "grad_norm": 1.28125, "learning_rate": 6.9404592633612486e-06, "loss": 0.1323, "step": 385 }, { "epoch": 1.8985239852398523, "grad_norm": 1.328125, "learning_rate": 6.926045045795944e-06, "loss": 0.1106, "step": 386 }, { "epoch": 1.9034440344403443, "grad_norm": 1.3046875, "learning_rate": 6.911612007620077e-06, "loss": 0.0884, "step": 387 }, { "epoch": 1.9083640836408364, "grad_norm": 0.953125, "learning_rate": 6.897160289868042e-06, "loss": 0.101, "step": 388 }, { "epoch": 1.9132841328413284, "grad_norm": 0.81640625, "learning_rate": 6.8826900337567595e-06, "loss": 0.0907, "step": 389 }, { "epoch": 1.9182041820418205, "grad_norm": 1.0078125, "learning_rate": 6.8682013806842985e-06, "loss": 0.083, "step": 390 }, { "epoch": 1.9231242312423125, "grad_norm": 0.9765625, "learning_rate": 6.853694472228504e-06, "loss": 0.0821, "step": 391 }, { "epoch": 1.9280442804428044, "grad_norm": 1.046875, "learning_rate": 6.839169450145595e-06, "loss": 0.0605, "step": 392 }, { "epoch": 1.9329643296432963, "grad_norm": 1.0234375, "learning_rate": 6.824626456368802e-06, "loss": 0.1001, "step": 393 }, { "epoch": 1.9378843788437883, "grad_norm": 1.203125, "learning_rate": 6.810065633006956e-06, "loss": 0.0936, "step": 394 }, { "epoch": 1.9428044280442804, "grad_norm": 1.046875, "learning_rate": 6.795487122343124e-06, "loss": 0.0866, "step": 395 }, { "epoch": 1.9477244772447726, "grad_norm": 1.3125, "learning_rate": 6.7808910668331985e-06, "loss": 0.1078, "step": 396 }, { "epoch": 1.9526445264452645, "grad_norm": 1.109375, "learning_rate": 6.766277609104518e-06, "loss": 0.0854, "step": 397 }, { "epoch": 1.9575645756457565, "grad_norm": 0.9453125, "learning_rate": 6.751646891954466e-06, "loss": 0.0848, "step": 398 }, { "epoch": 1.9624846248462484, "grad_norm": 1.0234375, "learning_rate": 6.736999058349085e-06, "loss": 0.0861, "step": 399 }, { "epoch": 1.9674046740467404, "grad_norm": 1.4296875, "learning_rate": 6.722334251421665e-06, "loss": 0.108, "step": 400 }, { "epoch": 1.9723247232472325, "grad_norm": 0.76171875, "learning_rate": 6.707652614471359e-06, "loss": 0.0691, "step": 401 }, { "epoch": 1.9772447724477245, "grad_norm": 1.4765625, "learning_rate": 6.692954290961774e-06, "loss": 0.1234, "step": 402 }, { "epoch": 1.9821648216482166, "grad_norm": 0.9140625, "learning_rate": 6.678239424519575e-06, "loss": 0.0785, "step": 403 }, { "epoch": 1.9870848708487086, "grad_norm": 1.328125, "learning_rate": 6.6635081589330745e-06, "loss": 0.096, "step": 404 }, { "epoch": 1.9920049200492005, "grad_norm": 1.25, "learning_rate": 6.648760638150833e-06, "loss": 0.1109, "step": 405 }, { "epoch": 1.9969249692496924, "grad_norm": 1.125, "learning_rate": 6.6339970062802526e-06, "loss": 0.0892, "step": 406 }, { "epoch": 2.0030750307503076, "grad_norm": 0.83984375, "learning_rate": 6.619217407586167e-06, "loss": 0.065, "step": 407 }, { "epoch": 2.0079950799507995, "grad_norm": 1.2578125, "learning_rate": 6.604421986489428e-06, "loss": 0.1044, "step": 408 }, { "epoch": 2.0079950799507995, "eval_loss": 0.10599377751350403, "eval_runtime": 203.102, "eval_samples_per_second": 8.055, "eval_steps_per_second": 1.344, "step": 408 }, { "epoch": 2.0129151291512914, "grad_norm": 0.8828125, "learning_rate": 6.589610887565503e-06, "loss": 0.0784, "step": 409 }, { "epoch": 2.0178351783517834, "grad_norm": 0.9609375, "learning_rate": 6.574784255543052e-06, "loss": 0.0788, "step": 410 }, { "epoch": 2.0227552275522753, "grad_norm": 0.81640625, "learning_rate": 6.559942235302527e-06, "loss": 0.0876, "step": 411 }, { "epoch": 2.0276752767527677, "grad_norm": 1.125, "learning_rate": 6.545084971874738e-06, "loss": 0.0702, "step": 412 }, { "epoch": 2.0325953259532596, "grad_norm": 1.703125, "learning_rate": 6.530212610439455e-06, "loss": 0.1014, "step": 413 }, { "epoch": 2.0375153751537516, "grad_norm": 0.87890625, "learning_rate": 6.5153252963239736e-06, "loss": 0.0846, "step": 414 }, { "epoch": 2.0424354243542435, "grad_norm": 1.34375, "learning_rate": 6.500423175001705e-06, "loss": 0.1422, "step": 415 }, { "epoch": 2.0473554735547355, "grad_norm": 1.1796875, "learning_rate": 6.485506392090749e-06, "loss": 0.1003, "step": 416 }, { "epoch": 2.0522755227552274, "grad_norm": 1.1328125, "learning_rate": 6.470575093352477e-06, "loss": 0.1012, "step": 417 }, { "epoch": 2.0571955719557193, "grad_norm": 1.1484375, "learning_rate": 6.4556294246900994e-06, "loss": 0.0778, "step": 418 }, { "epoch": 2.0621156211562117, "grad_norm": 1.109375, "learning_rate": 6.440669532147246e-06, "loss": 0.0975, "step": 419 }, { "epoch": 2.0670356703567037, "grad_norm": 0.91015625, "learning_rate": 6.4256955619065375e-06, "loss": 0.0772, "step": 420 }, { "epoch": 2.0719557195571956, "grad_norm": 1.078125, "learning_rate": 6.410707660288154e-06, "loss": 0.1044, "step": 421 }, { "epoch": 2.0768757687576875, "grad_norm": 1.1796875, "learning_rate": 6.395705973748414e-06, "loss": 0.0868, "step": 422 }, { "epoch": 2.0817958179581795, "grad_norm": 1.234375, "learning_rate": 6.3806906488783305e-06, "loss": 0.1051, "step": 423 }, { "epoch": 2.0867158671586714, "grad_norm": 1.1171875, "learning_rate": 6.36566183240219e-06, "loss": 0.0572, "step": 424 }, { "epoch": 2.091635916359164, "grad_norm": 1.078125, "learning_rate": 6.350619671176111e-06, "loss": 0.1055, "step": 425 }, { "epoch": 2.0965559655596557, "grad_norm": 0.9140625, "learning_rate": 6.335564312186615e-06, "loss": 0.0956, "step": 426 }, { "epoch": 2.1014760147601477, "grad_norm": 1.75, "learning_rate": 6.3204959025491844e-06, "loss": 0.0727, "step": 427 }, { "epoch": 2.1063960639606396, "grad_norm": 1.1015625, "learning_rate": 6.30541458950683e-06, "loss": 0.1014, "step": 428 }, { "epoch": 2.1113161131611315, "grad_norm": 1.1953125, "learning_rate": 6.2903205204286474e-06, "loss": 0.0968, "step": 429 }, { "epoch": 2.1162361623616235, "grad_norm": 1.21875, "learning_rate": 6.275213842808383e-06, "loss": 0.0856, "step": 430 }, { "epoch": 2.1211562115621154, "grad_norm": 1.359375, "learning_rate": 6.260094704262986e-06, "loss": 0.0773, "step": 431 }, { "epoch": 2.126076260762608, "grad_norm": 1.125, "learning_rate": 6.244963252531171e-06, "loss": 0.0802, "step": 432 }, { "epoch": 2.1309963099630997, "grad_norm": 1.21875, "learning_rate": 6.229819635471972e-06, "loss": 0.0809, "step": 433 }, { "epoch": 2.1359163591635917, "grad_norm": 1.1171875, "learning_rate": 6.2146640010633e-06, "loss": 0.0833, "step": 434 }, { "epoch": 2.1408364083640836, "grad_norm": 0.984375, "learning_rate": 6.19949649740049e-06, "loss": 0.0784, "step": 435 }, { "epoch": 2.1457564575645756, "grad_norm": 0.9453125, "learning_rate": 6.184317272694866e-06, "loss": 0.0885, "step": 436 }, { "epoch": 2.1506765067650675, "grad_norm": 1.140625, "learning_rate": 6.16912647527228e-06, "loss": 0.0763, "step": 437 }, { "epoch": 2.15559655596556, "grad_norm": 1.1484375, "learning_rate": 6.1539242535716704e-06, "loss": 0.0805, "step": 438 }, { "epoch": 2.160516605166052, "grad_norm": 1.2421875, "learning_rate": 6.138710756143613e-06, "loss": 0.1051, "step": 439 }, { "epoch": 2.1654366543665438, "grad_norm": 0.96875, "learning_rate": 6.123486131648859e-06, "loss": 0.0741, "step": 440 }, { "epoch": 2.1703567035670357, "grad_norm": 1.171875, "learning_rate": 6.108250528856895e-06, "loss": 0.0725, "step": 441 }, { "epoch": 2.1752767527675276, "grad_norm": 1.0859375, "learning_rate": 6.0930040966444815e-06, "loss": 0.0957, "step": 442 }, { "epoch": 2.1801968019680196, "grad_norm": 1.421875, "learning_rate": 6.077746983994198e-06, "loss": 0.1151, "step": 443 }, { "epoch": 2.1851168511685115, "grad_norm": 1.1328125, "learning_rate": 6.062479339992993e-06, "loss": 0.0939, "step": 444 }, { "epoch": 2.190036900369004, "grad_norm": 1.0234375, "learning_rate": 6.047201313830724e-06, "loss": 0.0944, "step": 445 }, { "epoch": 2.194956949569496, "grad_norm": 1.0859375, "learning_rate": 6.031913054798692e-06, "loss": 0.0825, "step": 446 }, { "epoch": 2.1998769987699878, "grad_norm": 0.94921875, "learning_rate": 6.016614712288198e-06, "loss": 0.0732, "step": 447 }, { "epoch": 2.2047970479704797, "grad_norm": 1.2421875, "learning_rate": 6.0013064357890715e-06, "loss": 0.0749, "step": 448 }, { "epoch": 2.2097170971709716, "grad_norm": 1.3828125, "learning_rate": 5.985988374888216e-06, "loss": 0.0935, "step": 449 }, { "epoch": 2.2146371463714636, "grad_norm": 1.0625, "learning_rate": 5.970660679268139e-06, "loss": 0.0762, "step": 450 }, { "epoch": 2.2195571955719555, "grad_norm": 1.3671875, "learning_rate": 5.955323498705501e-06, "loss": 0.0929, "step": 451 }, { "epoch": 2.224477244772448, "grad_norm": 1.3046875, "learning_rate": 5.9399769830696404e-06, "loss": 0.1067, "step": 452 }, { "epoch": 2.22939729397294, "grad_norm": 0.9609375, "learning_rate": 5.924621282321123e-06, "loss": 0.0736, "step": 453 }, { "epoch": 2.234317343173432, "grad_norm": 1.0078125, "learning_rate": 5.909256546510257e-06, "loss": 0.0699, "step": 454 }, { "epoch": 2.2392373923739237, "grad_norm": 1.0, "learning_rate": 5.893882925775648e-06, "loss": 0.0792, "step": 455 }, { "epoch": 2.2441574415744157, "grad_norm": 1.15625, "learning_rate": 5.878500570342714e-06, "loss": 0.0874, "step": 456 }, { "epoch": 2.2490774907749076, "grad_norm": 1.296875, "learning_rate": 5.86310963052223e-06, "loss": 0.1197, "step": 457 }, { "epoch": 2.2539975399753995, "grad_norm": 1.21875, "learning_rate": 5.847710256708854e-06, "loss": 0.1147, "step": 458 }, { "epoch": 2.258917589175892, "grad_norm": 0.9453125, "learning_rate": 5.832302599379657e-06, "loss": 0.0647, "step": 459 }, { "epoch": 2.263837638376384, "grad_norm": 0.86328125, "learning_rate": 5.816886809092651e-06, "loss": 0.0794, "step": 460 }, { "epoch": 2.268757687576876, "grad_norm": 1.1171875, "learning_rate": 5.8014630364853275e-06, "loss": 0.0787, "step": 461 }, { "epoch": 2.2736777367773677, "grad_norm": 1.375, "learning_rate": 5.7860314322731705e-06, "loss": 0.1395, "step": 462 }, { "epoch": 2.2785977859778597, "grad_norm": 1.6640625, "learning_rate": 5.770592147248197e-06, "loss": 0.083, "step": 463 }, { "epoch": 2.2835178351783516, "grad_norm": 1.390625, "learning_rate": 5.755145332277472e-06, "loss": 0.1035, "step": 464 }, { "epoch": 2.288437884378844, "grad_norm": 1.2265625, "learning_rate": 5.73969113830165e-06, "loss": 0.1229, "step": 465 }, { "epoch": 2.293357933579336, "grad_norm": 0.96875, "learning_rate": 5.72422971633348e-06, "loss": 0.0847, "step": 466 }, { "epoch": 2.298277982779828, "grad_norm": 1.2734375, "learning_rate": 5.70876121745635e-06, "loss": 0.0944, "step": 467 }, { "epoch": 2.30319803198032, "grad_norm": 0.8828125, "learning_rate": 5.69328579282279e-06, "loss": 0.0729, "step": 468 }, { "epoch": 2.3081180811808117, "grad_norm": 1.2890625, "learning_rate": 5.677803593653018e-06, "loss": 0.076, "step": 469 }, { "epoch": 2.3130381303813037, "grad_norm": 1.3515625, "learning_rate": 5.66231477123344e-06, "loss": 0.0779, "step": 470 }, { "epoch": 2.3179581795817956, "grad_norm": 1.09375, "learning_rate": 5.646819476915189e-06, "loss": 0.0596, "step": 471 }, { "epoch": 2.322878228782288, "grad_norm": 0.98046875, "learning_rate": 5.631317862112636e-06, "loss": 0.1077, "step": 472 }, { "epoch": 2.32779827798278, "grad_norm": 1.3125, "learning_rate": 5.615810078301912e-06, "loss": 0.1024, "step": 473 }, { "epoch": 2.332718327183272, "grad_norm": 1.3984375, "learning_rate": 5.600296277019434e-06, "loss": 0.1357, "step": 474 }, { "epoch": 2.337638376383764, "grad_norm": 1.09375, "learning_rate": 5.584776609860414e-06, "loss": 0.0891, "step": 475 }, { "epoch": 2.3425584255842558, "grad_norm": 1.1328125, "learning_rate": 5.569251228477386e-06, "loss": 0.0744, "step": 476 }, { "epoch": 2.3474784747847477, "grad_norm": 0.984375, "learning_rate": 5.553720284578723e-06, "loss": 0.0669, "step": 477 }, { "epoch": 2.35239852398524, "grad_norm": 1.015625, "learning_rate": 5.538183929927152e-06, "loss": 0.0819, "step": 478 }, { "epoch": 2.357318573185732, "grad_norm": 1.140625, "learning_rate": 5.522642316338268e-06, "loss": 0.0794, "step": 479 }, { "epoch": 2.362238622386224, "grad_norm": 1.03125, "learning_rate": 5.507095595679059e-06, "loss": 0.0887, "step": 480 }, { "epoch": 2.367158671586716, "grad_norm": 1.4296875, "learning_rate": 5.491543919866417e-06, "loss": 0.0863, "step": 481 }, { "epoch": 2.372078720787208, "grad_norm": 0.9140625, "learning_rate": 5.47598744086565e-06, "loss": 0.0844, "step": 482 }, { "epoch": 2.3769987699876998, "grad_norm": 1.0625, "learning_rate": 5.460426310689006e-06, "loss": 0.0712, "step": 483 }, { "epoch": 2.3819188191881917, "grad_norm": 1.359375, "learning_rate": 5.4448606813941805e-06, "loss": 0.0768, "step": 484 }, { "epoch": 2.3868388683886836, "grad_norm": 1.1015625, "learning_rate": 5.42929070508283e-06, "loss": 0.079, "step": 485 }, { "epoch": 2.391758917589176, "grad_norm": 0.96875, "learning_rate": 5.413716533899096e-06, "loss": 0.0853, "step": 486 }, { "epoch": 2.396678966789668, "grad_norm": 1.1171875, "learning_rate": 5.3981383200281004e-06, "loss": 0.1153, "step": 487 }, { "epoch": 2.40159901599016, "grad_norm": 1.1171875, "learning_rate": 5.382556215694478e-06, "loss": 0.0533, "step": 488 }, { "epoch": 2.406519065190652, "grad_norm": 1.234375, "learning_rate": 5.366970373160873e-06, "loss": 0.0905, "step": 489 }, { "epoch": 2.411439114391144, "grad_norm": 1.03125, "learning_rate": 5.351380944726465e-06, "loss": 0.0787, "step": 490 }, { "epoch": 2.416359163591636, "grad_norm": 1.21875, "learning_rate": 5.335788082725467e-06, "loss": 0.0781, "step": 491 }, { "epoch": 2.421279212792128, "grad_norm": 0.90234375, "learning_rate": 5.3201919395256475e-06, "loss": 0.0689, "step": 492 }, { "epoch": 2.42619926199262, "grad_norm": 1.2265625, "learning_rate": 5.304592667526835e-06, "loss": 0.0849, "step": 493 }, { "epoch": 2.431119311193112, "grad_norm": 1.109375, "learning_rate": 5.288990419159433e-06, "loss": 0.1141, "step": 494 }, { "epoch": 2.436039360393604, "grad_norm": 0.85546875, "learning_rate": 5.2733853468829295e-06, "loss": 0.0842, "step": 495 }, { "epoch": 2.440959409594096, "grad_norm": 1.1796875, "learning_rate": 5.257777603184408e-06, "loss": 0.0599, "step": 496 }, { "epoch": 2.445879458794588, "grad_norm": 0.859375, "learning_rate": 5.24216734057705e-06, "loss": 0.0906, "step": 497 }, { "epoch": 2.4507995079950797, "grad_norm": 1.84375, "learning_rate": 5.226554711598659e-06, "loss": 0.0934, "step": 498 }, { "epoch": 2.455719557195572, "grad_norm": 0.75390625, "learning_rate": 5.210939868810156e-06, "loss": 0.0783, "step": 499 }, { "epoch": 2.460639606396064, "grad_norm": 1.078125, "learning_rate": 5.195322964794098e-06, "loss": 0.0944, "step": 500 }, { "epoch": 2.465559655596556, "grad_norm": 1.0234375, "learning_rate": 5.1797041521531795e-06, "loss": 0.0831, "step": 501 }, { "epoch": 2.470479704797048, "grad_norm": 1.3828125, "learning_rate": 5.16408358350875e-06, "loss": 0.1166, "step": 502 }, { "epoch": 2.47539975399754, "grad_norm": 1.0625, "learning_rate": 5.1484614114993156e-06, "loss": 0.089, "step": 503 }, { "epoch": 2.480319803198032, "grad_norm": 1.546875, "learning_rate": 5.132837788779049e-06, "loss": 0.1219, "step": 504 }, { "epoch": 2.485239852398524, "grad_norm": 0.95703125, "learning_rate": 5.117212868016303e-06, "loss": 0.0802, "step": 505 }, { "epoch": 2.490159901599016, "grad_norm": 0.90625, "learning_rate": 5.101586801892109e-06, "loss": 0.0862, "step": 506 }, { "epoch": 2.495079950799508, "grad_norm": 1.46875, "learning_rate": 5.085959743098693e-06, "loss": 0.0834, "step": 507 }, { "epoch": 2.5, "grad_norm": 1.125, "learning_rate": 5.07033184433798e-06, "loss": 0.0764, "step": 508 }, { "epoch": 2.504920049200492, "grad_norm": 1.21875, "learning_rate": 5.054703258320107e-06, "loss": 0.0789, "step": 509 }, { "epoch": 2.509840098400984, "grad_norm": 1.0390625, "learning_rate": 5.03907413776192e-06, "loss": 0.0909, "step": 510 }, { "epoch": 2.509840098400984, "eval_loss": 0.10507026314735413, "eval_runtime": 204.2755, "eval_samples_per_second": 8.009, "eval_steps_per_second": 1.336, "step": 510 }, { "epoch": 2.514760147601476, "grad_norm": 0.88671875, "learning_rate": 5.0234446353854934e-06, "loss": 0.0751, "step": 511 }, { "epoch": 2.5196801968019678, "grad_norm": 1.1796875, "learning_rate": 5.00781490391663e-06, "loss": 0.1011, "step": 512 }, { "epoch": 2.52460024600246, "grad_norm": 1.2890625, "learning_rate": 4.992185096083372e-06, "loss": 0.1007, "step": 513 }, { "epoch": 2.529520295202952, "grad_norm": 1.0625, "learning_rate": 4.976555364614509e-06, "loss": 0.0879, "step": 514 }, { "epoch": 2.534440344403444, "grad_norm": 1.3203125, "learning_rate": 4.96092586223808e-06, "loss": 0.0806, "step": 515 }, { "epoch": 2.539360393603936, "grad_norm": 1.1875, "learning_rate": 4.9452967416798945e-06, "loss": 0.0916, "step": 516 }, { "epoch": 2.544280442804428, "grad_norm": 1.4296875, "learning_rate": 4.929668155662021e-06, "loss": 0.0825, "step": 517 }, { "epoch": 2.5492004920049203, "grad_norm": 1.1953125, "learning_rate": 4.914040256901309e-06, "loss": 0.0761, "step": 518 }, { "epoch": 2.554120541205412, "grad_norm": 1.1015625, "learning_rate": 4.898413198107892e-06, "loss": 0.0657, "step": 519 }, { "epoch": 2.559040590405904, "grad_norm": 1.625, "learning_rate": 4.882787131983698e-06, "loss": 0.1164, "step": 520 }, { "epoch": 2.563960639606396, "grad_norm": 1.0078125, "learning_rate": 4.867162211220952e-06, "loss": 0.0794, "step": 521 }, { "epoch": 2.568880688806888, "grad_norm": 1.0859375, "learning_rate": 4.851538588500687e-06, "loss": 0.0943, "step": 522 }, { "epoch": 2.57380073800738, "grad_norm": 1.1953125, "learning_rate": 4.835916416491251e-06, "loss": 0.0911, "step": 523 }, { "epoch": 2.578720787207872, "grad_norm": 0.9140625, "learning_rate": 4.820295847846822e-06, "loss": 0.0579, "step": 524 }, { "epoch": 2.583640836408364, "grad_norm": 1.125, "learning_rate": 4.804677035205903e-06, "loss": 0.0955, "step": 525 }, { "epoch": 2.588560885608856, "grad_norm": 1.484375, "learning_rate": 4.789060131189845e-06, "loss": 0.0879, "step": 526 }, { "epoch": 2.593480934809348, "grad_norm": 1.140625, "learning_rate": 4.773445288401343e-06, "loss": 0.089, "step": 527 }, { "epoch": 2.59840098400984, "grad_norm": 0.9375, "learning_rate": 4.75783265942295e-06, "loss": 0.0871, "step": 528 }, { "epoch": 2.603321033210332, "grad_norm": 1.1640625, "learning_rate": 4.742222396815593e-06, "loss": 0.1021, "step": 529 }, { "epoch": 2.608241082410824, "grad_norm": 1.0859375, "learning_rate": 4.726614653117071e-06, "loss": 0.0909, "step": 530 }, { "epoch": 2.6131611316113164, "grad_norm": 1.234375, "learning_rate": 4.711009580840569e-06, "loss": 0.1082, "step": 531 }, { "epoch": 2.6180811808118083, "grad_norm": 1.203125, "learning_rate": 4.695407332473166e-06, "loss": 0.083, "step": 532 }, { "epoch": 2.6230012300123002, "grad_norm": 1.0859375, "learning_rate": 4.679808060474354e-06, "loss": 0.0745, "step": 533 }, { "epoch": 2.627921279212792, "grad_norm": 1.0703125, "learning_rate": 4.6642119172745345e-06, "loss": 0.0776, "step": 534 }, { "epoch": 2.632841328413284, "grad_norm": 0.9453125, "learning_rate": 4.6486190552735375e-06, "loss": 0.0676, "step": 535 }, { "epoch": 2.637761377613776, "grad_norm": 1.0703125, "learning_rate": 4.633029626839128e-06, "loss": 0.0794, "step": 536 }, { "epoch": 2.642681426814268, "grad_norm": 1.0859375, "learning_rate": 4.617443784305524e-06, "loss": 0.1136, "step": 537 }, { "epoch": 2.64760147601476, "grad_norm": 1.1875, "learning_rate": 4.601861679971901e-06, "loss": 0.0747, "step": 538 }, { "epoch": 2.6525215252152523, "grad_norm": 1.625, "learning_rate": 4.5862834661009074e-06, "loss": 0.0982, "step": 539 }, { "epoch": 2.6574415744157442, "grad_norm": 1.046875, "learning_rate": 4.57070929491717e-06, "loss": 0.0788, "step": 540 }, { "epoch": 2.662361623616236, "grad_norm": 1.4765625, "learning_rate": 4.555139318605821e-06, "loss": 0.0857, "step": 541 }, { "epoch": 2.667281672816728, "grad_norm": 1.1796875, "learning_rate": 4.539573689310995e-06, "loss": 0.09, "step": 542 }, { "epoch": 2.67220172201722, "grad_norm": 1.4609375, "learning_rate": 4.524012559134352e-06, "loss": 0.0934, "step": 543 }, { "epoch": 2.6771217712177124, "grad_norm": 1.0234375, "learning_rate": 4.508456080133584e-06, "loss": 0.0742, "step": 544 }, { "epoch": 2.6820418204182044, "grad_norm": 1.046875, "learning_rate": 4.492904404320942e-06, "loss": 0.0677, "step": 545 }, { "epoch": 2.6869618696186963, "grad_norm": 1.078125, "learning_rate": 4.477357683661734e-06, "loss": 0.072, "step": 546 }, { "epoch": 2.6918819188191883, "grad_norm": 1.2578125, "learning_rate": 4.461816070072851e-06, "loss": 0.0948, "step": 547 }, { "epoch": 2.69680196801968, "grad_norm": 1.0625, "learning_rate": 4.446279715421277e-06, "loss": 0.0846, "step": 548 }, { "epoch": 2.701722017220172, "grad_norm": 1.3828125, "learning_rate": 4.430748771522615e-06, "loss": 0.0952, "step": 549 }, { "epoch": 2.706642066420664, "grad_norm": 1.09375, "learning_rate": 4.415223390139588e-06, "loss": 0.0797, "step": 550 }, { "epoch": 2.711562115621156, "grad_norm": 1.0078125, "learning_rate": 4.399703722980569e-06, "loss": 0.0822, "step": 551 }, { "epoch": 2.716482164821648, "grad_norm": 1.328125, "learning_rate": 4.3841899216980895e-06, "loss": 0.1041, "step": 552 }, { "epoch": 2.7214022140221403, "grad_norm": 1.21875, "learning_rate": 4.368682137887365e-06, "loss": 0.0884, "step": 553 }, { "epoch": 2.7263222632226323, "grad_norm": 1.0546875, "learning_rate": 4.3531805230848116e-06, "loss": 0.0799, "step": 554 }, { "epoch": 2.731242312423124, "grad_norm": 1.375, "learning_rate": 4.337685228766561e-06, "loss": 0.1055, "step": 555 }, { "epoch": 2.736162361623616, "grad_norm": 1.15625, "learning_rate": 4.322196406346984e-06, "loss": 0.1206, "step": 556 }, { "epoch": 2.741082410824108, "grad_norm": 1.2265625, "learning_rate": 4.30671420717721e-06, "loss": 0.085, "step": 557 }, { "epoch": 2.7460024600246005, "grad_norm": 1.4765625, "learning_rate": 4.291238782543652e-06, "loss": 0.1146, "step": 558 }, { "epoch": 2.7509225092250924, "grad_norm": 1.125, "learning_rate": 4.275770283666521e-06, "loss": 0.0742, "step": 559 }, { "epoch": 2.7558425584255843, "grad_norm": 1.2109375, "learning_rate": 4.260308861698351e-06, "loss": 0.091, "step": 560 }, { "epoch": 2.7607626076260763, "grad_norm": 1.171875, "learning_rate": 4.244854667722527e-06, "loss": 0.0741, "step": 561 }, { "epoch": 2.765682656826568, "grad_norm": 1.1328125, "learning_rate": 4.229407852751806e-06, "loss": 0.0738, "step": 562 }, { "epoch": 2.77060270602706, "grad_norm": 1.3359375, "learning_rate": 4.21396856772683e-06, "loss": 0.0999, "step": 563 }, { "epoch": 2.775522755227552, "grad_norm": 1.0703125, "learning_rate": 4.198536963514674e-06, "loss": 0.0619, "step": 564 }, { "epoch": 2.780442804428044, "grad_norm": 1.0390625, "learning_rate": 4.183113190907349e-06, "loss": 0.0841, "step": 565 }, { "epoch": 2.7853628536285364, "grad_norm": 0.96484375, "learning_rate": 4.1676974006203456e-06, "loss": 0.0738, "step": 566 }, { "epoch": 2.7902829028290284, "grad_norm": 1.21875, "learning_rate": 4.152289743291148e-06, "loss": 0.1018, "step": 567 }, { "epoch": 2.7952029520295203, "grad_norm": 1.0234375, "learning_rate": 4.136890369477773e-06, "loss": 0.0664, "step": 568 }, { "epoch": 2.8001230012300122, "grad_norm": 1.09375, "learning_rate": 4.121499429657287e-06, "loss": 0.0769, "step": 569 }, { "epoch": 2.805043050430504, "grad_norm": 1.0234375, "learning_rate": 4.106117074224354e-06, "loss": 0.0779, "step": 570 }, { "epoch": 2.8099630996309966, "grad_norm": 1.203125, "learning_rate": 4.090743453489744e-06, "loss": 0.0958, "step": 571 }, { "epoch": 2.8148831488314885, "grad_norm": 1.4609375, "learning_rate": 4.07537871767888e-06, "loss": 0.1121, "step": 572 }, { "epoch": 2.8198031980319804, "grad_norm": 1.4375, "learning_rate": 4.060023016930359e-06, "loss": 0.1435, "step": 573 }, { "epoch": 2.8247232472324724, "grad_norm": 1.2265625, "learning_rate": 4.044676501294501e-06, "loss": 0.1366, "step": 574 }, { "epoch": 2.8296432964329643, "grad_norm": 1.171875, "learning_rate": 4.029339320731862e-06, "loss": 0.0755, "step": 575 }, { "epoch": 2.8345633456334562, "grad_norm": 1.0625, "learning_rate": 4.0140116251117865e-06, "loss": 0.0837, "step": 576 }, { "epoch": 2.839483394833948, "grad_norm": 1.25, "learning_rate": 3.998693564210929e-06, "loss": 0.1097, "step": 577 }, { "epoch": 2.84440344403444, "grad_norm": 1.0078125, "learning_rate": 3.983385287711803e-06, "loss": 0.0652, "step": 578 }, { "epoch": 2.8493234932349325, "grad_norm": 0.94921875, "learning_rate": 3.96808694520131e-06, "loss": 0.0785, "step": 579 }, { "epoch": 2.8542435424354244, "grad_norm": 1.3125, "learning_rate": 3.952798686169279e-06, "loss": 0.0947, "step": 580 }, { "epoch": 2.8591635916359164, "grad_norm": 1.1796875, "learning_rate": 3.937520660007008e-06, "loss": 0.0803, "step": 581 }, { "epoch": 2.8640836408364083, "grad_norm": 0.8203125, "learning_rate": 3.9222530160058025e-06, "loss": 0.0664, "step": 582 }, { "epoch": 2.8690036900369003, "grad_norm": 1.0703125, "learning_rate": 3.90699590335552e-06, "loss": 0.0737, "step": 583 }, { "epoch": 2.8739237392373926, "grad_norm": 1.6875, "learning_rate": 3.891749471143106e-06, "loss": 0.1152, "step": 584 }, { "epoch": 2.8788437884378846, "grad_norm": 1.0, "learning_rate": 3.876513868351142e-06, "loss": 0.0932, "step": 585 }, { "epoch": 2.8837638376383765, "grad_norm": 1.6015625, "learning_rate": 3.861289243856388e-06, "loss": 0.1052, "step": 586 }, { "epoch": 2.8886838868388685, "grad_norm": 1.5234375, "learning_rate": 3.84607574642833e-06, "loss": 0.1152, "step": 587 }, { "epoch": 2.8936039360393604, "grad_norm": 1.1796875, "learning_rate": 3.830873524727722e-06, "loss": 0.083, "step": 588 }, { "epoch": 2.8985239852398523, "grad_norm": 1.2109375, "learning_rate": 3.815682727305136e-06, "loss": 0.0824, "step": 589 }, { "epoch": 2.9034440344403443, "grad_norm": 1.171875, "learning_rate": 3.800503502599511e-06, "loss": 0.0933, "step": 590 }, { "epoch": 2.908364083640836, "grad_norm": 0.91015625, "learning_rate": 3.7853359989367023e-06, "loss": 0.0831, "step": 591 }, { "epoch": 2.913284132841328, "grad_norm": 1.1484375, "learning_rate": 3.7701803645280296e-06, "loss": 0.0954, "step": 592 }, { "epoch": 2.9182041820418205, "grad_norm": 1.046875, "learning_rate": 3.7550367474688315e-06, "loss": 0.0829, "step": 593 }, { "epoch": 2.9231242312423125, "grad_norm": 1.703125, "learning_rate": 3.739905295737015e-06, "loss": 0.0905, "step": 594 }, { "epoch": 2.9280442804428044, "grad_norm": 1.3125, "learning_rate": 3.7247861571916183e-06, "loss": 0.0838, "step": 595 }, { "epoch": 2.9329643296432963, "grad_norm": 1.2109375, "learning_rate": 3.7096794795713542e-06, "loss": 0.1098, "step": 596 }, { "epoch": 2.9378843788437883, "grad_norm": 1.140625, "learning_rate": 3.6945854104931726e-06, "loss": 0.0776, "step": 597 }, { "epoch": 2.9428044280442807, "grad_norm": 1.203125, "learning_rate": 3.6795040974508164e-06, "loss": 0.0694, "step": 598 }, { "epoch": 2.9477244772447726, "grad_norm": 1.2890625, "learning_rate": 3.6644356878133862e-06, "loss": 0.0996, "step": 599 }, { "epoch": 2.9526445264452645, "grad_norm": 1.1953125, "learning_rate": 3.6493803288238894e-06, "loss": 0.0801, "step": 600 }, { "epoch": 2.9575645756457565, "grad_norm": 1.265625, "learning_rate": 3.634338167597812e-06, "loss": 0.0904, "step": 601 }, { "epoch": 2.9624846248462484, "grad_norm": 1.3125, "learning_rate": 3.61930935112167e-06, "loss": 0.0676, "step": 602 }, { "epoch": 2.9674046740467404, "grad_norm": 1.265625, "learning_rate": 3.6042940262515867e-06, "loss": 0.0985, "step": 603 }, { "epoch": 2.9723247232472323, "grad_norm": 1.109375, "learning_rate": 3.5892923397118473e-06, "loss": 0.0816, "step": 604 }, { "epoch": 2.9772447724477242, "grad_norm": 1.1328125, "learning_rate": 3.5743044380934655e-06, "loss": 0.076, "step": 605 }, { "epoch": 2.9821648216482166, "grad_norm": 0.97265625, "learning_rate": 3.5593304678527547e-06, "loss": 0.0932, "step": 606 }, { "epoch": 2.9870848708487086, "grad_norm": 1.140625, "learning_rate": 3.544370575309902e-06, "loss": 0.0982, "step": 607 }, { "epoch": 2.9920049200492005, "grad_norm": 1.15625, "learning_rate": 3.5294249066475245e-06, "loss": 0.0997, "step": 608 }, { "epoch": 2.9969249692496924, "grad_norm": 1.1953125, "learning_rate": 3.5144936079092528e-06, "loss": 0.0907, "step": 609 }, { "epoch": 3.002460024600246, "grad_norm": 7.90625, "learning_rate": 3.4995768249982975e-06, "loss": 0.1999, "step": 610 }, { "epoch": 3.007380073800738, "grad_norm": 1.0546875, "learning_rate": 3.4846747036760285e-06, "loss": 0.0815, "step": 611 }, { "epoch": 3.01230012300123, "grad_norm": 1.0234375, "learning_rate": 3.4697873895605466e-06, "loss": 0.0754, "step": 612 }, { "epoch": 3.01230012300123, "eval_loss": 0.1035788357257843, "eval_runtime": 204.6775, "eval_samples_per_second": 7.993, "eval_steps_per_second": 1.334, "step": 612 }, { "epoch": 3.0172201722017222, "grad_norm": 1.1796875, "learning_rate": 3.4549150281252635e-06, "loss": 0.1022, "step": 613 }, { "epoch": 3.022140221402214, "grad_norm": 0.953125, "learning_rate": 3.4400577646974766e-06, "loss": 0.074, "step": 614 }, { "epoch": 3.027060270602706, "grad_norm": 0.98046875, "learning_rate": 3.4252157444569478e-06, "loss": 0.0619, "step": 615 }, { "epoch": 3.031980319803198, "grad_norm": 1.2265625, "learning_rate": 3.410389112434499e-06, "loss": 0.0887, "step": 616 }, { "epoch": 3.03690036900369, "grad_norm": 0.87109375, "learning_rate": 3.3955780135105736e-06, "loss": 0.0831, "step": 617 }, { "epoch": 3.041820418204182, "grad_norm": 1.0625, "learning_rate": 3.3807825924138356e-06, "loss": 0.1026, "step": 618 }, { "epoch": 3.046740467404674, "grad_norm": 1.1953125, "learning_rate": 3.366002993719747e-06, "loss": 0.0814, "step": 619 }, { "epoch": 3.0516605166051662, "grad_norm": 1.1015625, "learning_rate": 3.351239361849168e-06, "loss": 0.1059, "step": 620 }, { "epoch": 3.056580565805658, "grad_norm": 0.82421875, "learning_rate": 3.336491841066928e-06, "loss": 0.083, "step": 621 }, { "epoch": 3.06150061500615, "grad_norm": 1.109375, "learning_rate": 3.3217605754804273e-06, "loss": 0.0612, "step": 622 }, { "epoch": 3.066420664206642, "grad_norm": 1.1015625, "learning_rate": 3.307045709038226e-06, "loss": 0.0833, "step": 623 }, { "epoch": 3.071340713407134, "grad_norm": 1.0078125, "learning_rate": 3.2923473855286426e-06, "loss": 0.0717, "step": 624 }, { "epoch": 3.076260762607626, "grad_norm": 0.8671875, "learning_rate": 3.2776657485783357e-06, "loss": 0.0752, "step": 625 }, { "epoch": 3.081180811808118, "grad_norm": 1.09375, "learning_rate": 3.2630009416509167e-06, "loss": 0.0891, "step": 626 }, { "epoch": 3.0861008610086103, "grad_norm": 1.2109375, "learning_rate": 3.2483531080455334e-06, "loss": 0.1059, "step": 627 }, { "epoch": 3.091020910209102, "grad_norm": 1.015625, "learning_rate": 3.2337223908954834e-06, "loss": 0.0874, "step": 628 }, { "epoch": 3.095940959409594, "grad_norm": 1.0078125, "learning_rate": 3.2191089331668036e-06, "loss": 0.0814, "step": 629 }, { "epoch": 3.100861008610086, "grad_norm": 1.1875, "learning_rate": 3.2045128776568783e-06, "loss": 0.0982, "step": 630 }, { "epoch": 3.105781057810578, "grad_norm": 0.7265625, "learning_rate": 3.1899343669930446e-06, "loss": 0.0647, "step": 631 }, { "epoch": 3.11070110701107, "grad_norm": 1.2734375, "learning_rate": 3.1753735436312005e-06, "loss": 0.1091, "step": 632 }, { "epoch": 3.1156211562115623, "grad_norm": 1.046875, "learning_rate": 3.1608305498544056e-06, "loss": 0.0939, "step": 633 }, { "epoch": 3.1205412054120543, "grad_norm": 1.0390625, "learning_rate": 3.146305527771499e-06, "loss": 0.074, "step": 634 }, { "epoch": 3.125461254612546, "grad_norm": 1.03125, "learning_rate": 3.1317986193157023e-06, "loss": 0.0734, "step": 635 }, { "epoch": 3.130381303813038, "grad_norm": 1.3203125, "learning_rate": 3.1173099662432426e-06, "loss": 0.1068, "step": 636 }, { "epoch": 3.13530135301353, "grad_norm": 1.484375, "learning_rate": 3.1028397101319584e-06, "loss": 0.0917, "step": 637 }, { "epoch": 3.140221402214022, "grad_norm": 1.2265625, "learning_rate": 3.0883879923799244e-06, "loss": 0.0743, "step": 638 }, { "epoch": 3.145141451414514, "grad_norm": 1.2734375, "learning_rate": 3.0739549542040583e-06, "loss": 0.0971, "step": 639 }, { "epoch": 3.1500615006150063, "grad_norm": 1.0625, "learning_rate": 3.059540736638751e-06, "loss": 0.1033, "step": 640 }, { "epoch": 3.1549815498154983, "grad_norm": 0.94921875, "learning_rate": 3.0451454805344893e-06, "loss": 0.0899, "step": 641 }, { "epoch": 3.15990159901599, "grad_norm": 1.0859375, "learning_rate": 3.0307693265564708e-06, "loss": 0.0845, "step": 642 }, { "epoch": 3.164821648216482, "grad_norm": 1.1484375, "learning_rate": 3.016412415183233e-06, "loss": 0.0914, "step": 643 }, { "epoch": 3.169741697416974, "grad_norm": 1.1796875, "learning_rate": 3.002074886705284e-06, "loss": 0.0802, "step": 644 }, { "epoch": 3.174661746617466, "grad_norm": 1.3125, "learning_rate": 2.9877568812237325e-06, "loss": 0.0946, "step": 645 }, { "epoch": 3.179581795817958, "grad_norm": 1.2265625, "learning_rate": 2.9734585386489095e-06, "loss": 0.0782, "step": 646 }, { "epoch": 3.1845018450184504, "grad_norm": 1.0078125, "learning_rate": 2.9591799986990098e-06, "loss": 0.0594, "step": 647 }, { "epoch": 3.1894218942189423, "grad_norm": 1.3515625, "learning_rate": 2.9449214008987253e-06, "loss": 0.0794, "step": 648 }, { "epoch": 3.1943419434194342, "grad_norm": 1.1640625, "learning_rate": 2.93068288457788e-06, "loss": 0.0818, "step": 649 }, { "epoch": 3.199261992619926, "grad_norm": 1.1796875, "learning_rate": 2.916464588870067e-06, "loss": 0.1144, "step": 650 }, { "epoch": 3.204182041820418, "grad_norm": 1.265625, "learning_rate": 2.9022666527112954e-06, "loss": 0.093, "step": 651 }, { "epoch": 3.20910209102091, "grad_norm": 0.98828125, "learning_rate": 2.8880892148386198e-06, "loss": 0.0725, "step": 652 }, { "epoch": 3.2140221402214024, "grad_norm": 1.1796875, "learning_rate": 2.873932413788805e-06, "loss": 0.068, "step": 653 }, { "epoch": 3.2189421894218944, "grad_norm": 0.96484375, "learning_rate": 2.859796387896949e-06, "loss": 0.0583, "step": 654 }, { "epoch": 3.2238622386223863, "grad_norm": 1.078125, "learning_rate": 2.8456812752951483e-06, "loss": 0.0748, "step": 655 }, { "epoch": 3.2287822878228782, "grad_norm": 1.3984375, "learning_rate": 2.831587213911142e-06, "loss": 0.1128, "step": 656 }, { "epoch": 3.23370233702337, "grad_norm": 1.328125, "learning_rate": 2.817514341466965e-06, "loss": 0.0812, "step": 657 }, { "epoch": 3.238622386223862, "grad_norm": 1.1875, "learning_rate": 2.8034627954775993e-06, "loss": 0.0995, "step": 658 }, { "epoch": 3.243542435424354, "grad_norm": 1.0546875, "learning_rate": 2.7894327132496324e-06, "loss": 0.0865, "step": 659 }, { "epoch": 3.2484624846248464, "grad_norm": 1.171875, "learning_rate": 2.7754242318799174e-06, "loss": 0.0807, "step": 660 }, { "epoch": 3.2533825338253384, "grad_norm": 1.015625, "learning_rate": 2.761437488254232e-06, "loss": 0.1041, "step": 661 }, { "epoch": 3.2583025830258303, "grad_norm": 1.0703125, "learning_rate": 2.74747261904594e-06, "loss": 0.074, "step": 662 }, { "epoch": 3.2632226322263223, "grad_norm": 1.203125, "learning_rate": 2.733529760714655e-06, "loss": 0.0852, "step": 663 }, { "epoch": 3.268142681426814, "grad_norm": 1.3515625, "learning_rate": 2.719609049504911e-06, "loss": 0.1049, "step": 664 }, { "epoch": 3.273062730627306, "grad_norm": 1.2265625, "learning_rate": 2.7057106214448216e-06, "loss": 0.0682, "step": 665 }, { "epoch": 3.2779827798277985, "grad_norm": 1.3828125, "learning_rate": 2.6918346123447708e-06, "loss": 0.1297, "step": 666 }, { "epoch": 3.2829028290282904, "grad_norm": 1.1328125, "learning_rate": 2.677981157796059e-06, "loss": 0.0748, "step": 667 }, { "epoch": 3.2878228782287824, "grad_norm": 1.125, "learning_rate": 2.6641503931696e-06, "loss": 0.1069, "step": 668 }, { "epoch": 3.2927429274292743, "grad_norm": 1.0703125, "learning_rate": 2.650342453614586e-06, "loss": 0.0802, "step": 669 }, { "epoch": 3.2976629766297663, "grad_norm": 1.1796875, "learning_rate": 2.636557474057173e-06, "loss": 0.0908, "step": 670 }, { "epoch": 3.302583025830258, "grad_norm": 0.91015625, "learning_rate": 2.6227955891991617e-06, "loss": 0.0663, "step": 671 }, { "epoch": 3.30750307503075, "grad_norm": 1.390625, "learning_rate": 2.609056933516675e-06, "loss": 0.0686, "step": 672 }, { "epoch": 3.312423124231242, "grad_norm": 1.0390625, "learning_rate": 2.5953416412588504e-06, "loss": 0.0833, "step": 673 }, { "epoch": 3.3173431734317345, "grad_norm": 1.1328125, "learning_rate": 2.58164984644653e-06, "loss": 0.0934, "step": 674 }, { "epoch": 3.3222632226322264, "grad_norm": 1.1796875, "learning_rate": 2.567981682870946e-06, "loss": 0.0953, "step": 675 }, { "epoch": 3.3271832718327183, "grad_norm": 1.1328125, "learning_rate": 2.554337284092411e-06, "loss": 0.1022, "step": 676 }, { "epoch": 3.3321033210332103, "grad_norm": 1.1953125, "learning_rate": 2.540716783439019e-06, "loss": 0.0674, "step": 677 }, { "epoch": 3.337023370233702, "grad_norm": 1.28125, "learning_rate": 2.5271203140053436e-06, "loss": 0.1064, "step": 678 }, { "epoch": 3.341943419434194, "grad_norm": 1.1796875, "learning_rate": 2.5135480086511306e-06, "loss": 0.0989, "step": 679 }, { "epoch": 3.3468634686346865, "grad_norm": 1.25, "learning_rate": 2.5000000000000015e-06, "loss": 0.0908, "step": 680 }, { "epoch": 3.3517835178351785, "grad_norm": 1.2421875, "learning_rate": 2.4864764204381624e-06, "loss": 0.0834, "step": 681 }, { "epoch": 3.3567035670356704, "grad_norm": 1.4921875, "learning_rate": 2.472977402113107e-06, "loss": 0.0902, "step": 682 }, { "epoch": 3.3616236162361623, "grad_norm": 1.1875, "learning_rate": 2.4595030769323246e-06, "loss": 0.0718, "step": 683 }, { "epoch": 3.3665436654366543, "grad_norm": 1.609375, "learning_rate": 2.4460535765620147e-06, "loss": 0.1142, "step": 684 }, { "epoch": 3.3714637146371462, "grad_norm": 1.1953125, "learning_rate": 2.4326290324257896e-06, "loss": 0.0951, "step": 685 }, { "epoch": 3.376383763837638, "grad_norm": 1.328125, "learning_rate": 2.419229575703411e-06, "loss": 0.0985, "step": 686 }, { "epoch": 3.3813038130381305, "grad_norm": 1.109375, "learning_rate": 2.4058553373294846e-06, "loss": 0.0713, "step": 687 }, { "epoch": 3.3862238622386225, "grad_norm": 1.3359375, "learning_rate": 2.3925064479921985e-06, "loss": 0.0882, "step": 688 }, { "epoch": 3.3911439114391144, "grad_norm": 1.2109375, "learning_rate": 2.379183038132031e-06, "loss": 0.1156, "step": 689 }, { "epoch": 3.3960639606396064, "grad_norm": 1.171875, "learning_rate": 2.3658852379404973e-06, "loss": 0.0914, "step": 690 }, { "epoch": 3.4009840098400983, "grad_norm": 1.1953125, "learning_rate": 2.352613177358852e-06, "loss": 0.0794, "step": 691 }, { "epoch": 3.4059040590405902, "grad_norm": 1.0859375, "learning_rate": 2.3393669860768364e-06, "loss": 0.0941, "step": 692 }, { "epoch": 3.4108241082410826, "grad_norm": 1.0078125, "learning_rate": 2.3261467935314097e-06, "loss": 0.0648, "step": 693 }, { "epoch": 3.4157441574415746, "grad_norm": 1.25, "learning_rate": 2.3129527289054716e-06, "loss": 0.0912, "step": 694 }, { "epoch": 3.4206642066420665, "grad_norm": 1.0390625, "learning_rate": 2.299784921126622e-06, "loss": 0.0801, "step": 695 }, { "epoch": 3.4255842558425584, "grad_norm": 1.078125, "learning_rate": 2.286643498865877e-06, "loss": 0.072, "step": 696 }, { "epoch": 3.4305043050430504, "grad_norm": 0.96875, "learning_rate": 2.2735285905364307e-06, "loss": 0.0599, "step": 697 }, { "epoch": 3.4354243542435423, "grad_norm": 1.015625, "learning_rate": 2.260440324292385e-06, "loss": 0.0748, "step": 698 }, { "epoch": 3.4403444034440342, "grad_norm": 0.95703125, "learning_rate": 2.2473788280275164e-06, "loss": 0.0613, "step": 699 }, { "epoch": 3.4452644526445266, "grad_norm": 1.390625, "learning_rate": 2.234344229374003e-06, "loss": 0.107, "step": 700 }, { "epoch": 3.4501845018450186, "grad_norm": 1.109375, "learning_rate": 2.2213366557011988e-06, "loss": 0.083, "step": 701 }, { "epoch": 3.4551045510455105, "grad_norm": 0.953125, "learning_rate": 2.2083562341143695e-06, "loss": 0.0823, "step": 702 }, { "epoch": 3.4600246002460024, "grad_norm": 1.4609375, "learning_rate": 2.195403091453473e-06, "loss": 0.1011, "step": 703 }, { "epoch": 3.4649446494464944, "grad_norm": 1.171875, "learning_rate": 2.1824773542918957e-06, "loss": 0.0832, "step": 704 }, { "epoch": 3.4698646986469863, "grad_norm": 1.5859375, "learning_rate": 2.1695791489352346e-06, "loss": 0.1078, "step": 705 }, { "epoch": 3.4747847478474787, "grad_norm": 0.8984375, "learning_rate": 2.156708601420053e-06, "loss": 0.0529, "step": 706 }, { "epoch": 3.4797047970479706, "grad_norm": 1.0546875, "learning_rate": 2.1438658375126544e-06, "loss": 0.0788, "step": 707 }, { "epoch": 3.4846248462484626, "grad_norm": 1.15625, "learning_rate": 2.131050982707849e-06, "loss": 0.0749, "step": 708 }, { "epoch": 3.4895448954489545, "grad_norm": 1.296875, "learning_rate": 2.1182641622277273e-06, "loss": 0.0666, "step": 709 }, { "epoch": 3.4944649446494465, "grad_norm": 1.0, "learning_rate": 2.1055055010204427e-06, "loss": 0.0892, "step": 710 }, { "epoch": 3.4993849938499384, "grad_norm": 1.203125, "learning_rate": 2.092775123758985e-06, "loss": 0.1143, "step": 711 }, { "epoch": 3.5043050430504303, "grad_norm": 0.9765625, "learning_rate": 2.080073154839964e-06, "loss": 0.0703, "step": 712 }, { "epoch": 3.5092250922509223, "grad_norm": 1.0703125, "learning_rate": 2.06739971838239e-06, "loss": 0.083, "step": 713 }, { "epoch": 3.5141451414514147, "grad_norm": 1.09375, "learning_rate": 2.05475493822647e-06, "loss": 0.0697, "step": 714 }, { "epoch": 3.5141451414514147, "eval_loss": 0.10337568074464798, "eval_runtime": 203.8015, "eval_samples_per_second": 8.027, "eval_steps_per_second": 1.34, "step": 714 }, { "epoch": 3.5190651906519066, "grad_norm": 1.375, "learning_rate": 2.042138937932388e-06, "loss": 0.0886, "step": 715 }, { "epoch": 3.5239852398523985, "grad_norm": 1.21875, "learning_rate": 2.0295518407791054e-06, "loss": 0.1, "step": 716 }, { "epoch": 3.5289052890528905, "grad_norm": 1.1484375, "learning_rate": 2.016993769763147e-06, "loss": 0.0802, "step": 717 }, { "epoch": 3.5338253382533824, "grad_norm": 0.98828125, "learning_rate": 2.0044648475974095e-06, "loss": 0.058, "step": 718 }, { "epoch": 3.538745387453875, "grad_norm": 1.0234375, "learning_rate": 1.9919651967099556e-06, "loss": 0.0875, "step": 719 }, { "epoch": 3.5436654366543667, "grad_norm": 1.25, "learning_rate": 1.979494939242822e-06, "loss": 0.0656, "step": 720 }, { "epoch": 3.5485854858548587, "grad_norm": 1.09375, "learning_rate": 1.9670541970508224e-06, "loss": 0.0676, "step": 721 }, { "epoch": 3.5535055350553506, "grad_norm": 1.0859375, "learning_rate": 1.954643091700354e-06, "loss": 0.0672, "step": 722 }, { "epoch": 3.5584255842558425, "grad_norm": 1.3984375, "learning_rate": 1.9422617444682196e-06, "loss": 0.0816, "step": 723 }, { "epoch": 3.5633456334563345, "grad_norm": 1.0234375, "learning_rate": 1.9299102763404335e-06, "loss": 0.0769, "step": 724 }, { "epoch": 3.5682656826568264, "grad_norm": 0.81640625, "learning_rate": 1.917588808011045e-06, "loss": 0.069, "step": 725 }, { "epoch": 3.5731857318573184, "grad_norm": 1.4765625, "learning_rate": 1.9052974598809505e-06, "loss": 0.075, "step": 726 }, { "epoch": 3.5781057810578107, "grad_norm": 1.34375, "learning_rate": 1.893036352056728e-06, "loss": 0.1036, "step": 727 }, { "epoch": 3.5830258302583027, "grad_norm": 1.2734375, "learning_rate": 1.880805604349456e-06, "loss": 0.0658, "step": 728 }, { "epoch": 3.5879458794587946, "grad_norm": 0.8046875, "learning_rate": 1.8686053362735468e-06, "loss": 0.052, "step": 729 }, { "epoch": 3.5928659286592866, "grad_norm": 0.828125, "learning_rate": 1.856435667045577e-06, "loss": 0.0592, "step": 730 }, { "epoch": 3.5977859778597785, "grad_norm": 1.2265625, "learning_rate": 1.844296715583117e-06, "loss": 0.0679, "step": 731 }, { "epoch": 3.602706027060271, "grad_norm": 1.0390625, "learning_rate": 1.8321886005035812e-06, "loss": 0.0883, "step": 732 }, { "epoch": 3.607626076260763, "grad_norm": 1.25, "learning_rate": 1.8201114401230586e-06, "loss": 0.0875, "step": 733 }, { "epoch": 3.6125461254612548, "grad_norm": 1.1953125, "learning_rate": 1.8080653524551623e-06, "loss": 0.0662, "step": 734 }, { "epoch": 3.6174661746617467, "grad_norm": 1.859375, "learning_rate": 1.796050455209869e-06, "loss": 0.1162, "step": 735 }, { "epoch": 3.6223862238622386, "grad_norm": 1.21875, "learning_rate": 1.7840668657923838e-06, "loss": 0.0753, "step": 736 }, { "epoch": 3.6273062730627306, "grad_norm": 1.1953125, "learning_rate": 1.772114701301972e-06, "loss": 0.0767, "step": 737 }, { "epoch": 3.6322263222632225, "grad_norm": 1.1640625, "learning_rate": 1.7601940785308337e-06, "loss": 0.1026, "step": 738 }, { "epoch": 3.6371463714637144, "grad_norm": 1.3203125, "learning_rate": 1.7483051139629482e-06, "loss": 0.0724, "step": 739 }, { "epoch": 3.6420664206642064, "grad_norm": 1.015625, "learning_rate": 1.7364479237729526e-06, "loss": 0.0915, "step": 740 }, { "epoch": 3.6469864698646988, "grad_norm": 1.375, "learning_rate": 1.7246226238249853e-06, "loss": 0.057, "step": 741 }, { "epoch": 3.6519065190651907, "grad_norm": 1.3515625, "learning_rate": 1.7128293296715704e-06, "loss": 0.0772, "step": 742 }, { "epoch": 3.6568265682656826, "grad_norm": 1.15625, "learning_rate": 1.7010681565524834e-06, "loss": 0.0597, "step": 743 }, { "epoch": 3.6617466174661746, "grad_norm": 1.2578125, "learning_rate": 1.6893392193936231e-06, "loss": 0.1057, "step": 744 }, { "epoch": 3.6666666666666665, "grad_norm": 1.0390625, "learning_rate": 1.677642632805892e-06, "loss": 0.0869, "step": 745 }, { "epoch": 3.671586715867159, "grad_norm": 1.1015625, "learning_rate": 1.6659785110840704e-06, "loss": 0.0846, "step": 746 }, { "epoch": 3.676506765067651, "grad_norm": 1.3125, "learning_rate": 1.6543469682057105e-06, "loss": 0.0953, "step": 747 }, { "epoch": 3.6814268142681428, "grad_norm": 1.5234375, "learning_rate": 1.6427481178300064e-06, "loss": 0.0984, "step": 748 }, { "epoch": 3.6863468634686347, "grad_norm": 1.2265625, "learning_rate": 1.631182073296706e-06, "loss": 0.099, "step": 749 }, { "epoch": 3.6912669126691267, "grad_norm": 1.3828125, "learning_rate": 1.6196489476249777e-06, "loss": 0.1055, "step": 750 }, { "epoch": 3.6961869618696186, "grad_norm": 1.265625, "learning_rate": 1.6081488535123274e-06, "loss": 0.1011, "step": 751 }, { "epoch": 3.7011070110701105, "grad_norm": 0.91015625, "learning_rate": 1.5966819033334807e-06, "loss": 0.0697, "step": 752 }, { "epoch": 3.7060270602706025, "grad_norm": 1.203125, "learning_rate": 1.5852482091393045e-06, "loss": 0.0792, "step": 753 }, { "epoch": 3.710947109471095, "grad_norm": 0.8671875, "learning_rate": 1.5738478826556885e-06, "loss": 0.0589, "step": 754 }, { "epoch": 3.715867158671587, "grad_norm": 1.328125, "learning_rate": 1.5624810352824709e-06, "loss": 0.1095, "step": 755 }, { "epoch": 3.7207872078720787, "grad_norm": 1.03125, "learning_rate": 1.551147778092344e-06, "loss": 0.0697, "step": 756 }, { "epoch": 3.7257072570725707, "grad_norm": 0.7890625, "learning_rate": 1.539848221829769e-06, "loss": 0.0562, "step": 757 }, { "epoch": 3.7306273062730626, "grad_norm": 1.125, "learning_rate": 1.5285824769098938e-06, "loss": 0.064, "step": 758 }, { "epoch": 3.735547355473555, "grad_norm": 1.234375, "learning_rate": 1.517350653417472e-06, "loss": 0.0805, "step": 759 }, { "epoch": 3.740467404674047, "grad_norm": 1.34375, "learning_rate": 1.5061528611057917e-06, "loss": 0.0751, "step": 760 }, { "epoch": 3.745387453874539, "grad_norm": 1.234375, "learning_rate": 1.4949892093956015e-06, "loss": 0.0958, "step": 761 }, { "epoch": 3.750307503075031, "grad_norm": 1.1875, "learning_rate": 1.4838598073740395e-06, "loss": 0.1045, "step": 762 }, { "epoch": 3.7552275522755227, "grad_norm": 0.984375, "learning_rate": 1.472764763793565e-06, "loss": 0.0532, "step": 763 }, { "epoch": 3.7601476014760147, "grad_norm": 1.3046875, "learning_rate": 1.4617041870709042e-06, "loss": 0.0961, "step": 764 }, { "epoch": 3.7650676506765066, "grad_norm": 1.2265625, "learning_rate": 1.4506781852859836e-06, "loss": 0.0696, "step": 765 }, { "epoch": 3.7699876998769986, "grad_norm": 1.015625, "learning_rate": 1.4396868661808777e-06, "loss": 0.076, "step": 766 }, { "epoch": 3.774907749077491, "grad_norm": 1.0078125, "learning_rate": 1.428730337158749e-06, "loss": 0.0949, "step": 767 }, { "epoch": 3.779827798277983, "grad_norm": 1.3203125, "learning_rate": 1.4178087052828098e-06, "loss": 0.0981, "step": 768 }, { "epoch": 3.784747847478475, "grad_norm": 1.25, "learning_rate": 1.4069220772752685e-06, "loss": 0.1072, "step": 769 }, { "epoch": 3.7896678966789668, "grad_norm": 0.9921875, "learning_rate": 1.3960705595162876e-06, "loss": 0.063, "step": 770 }, { "epoch": 3.7945879458794587, "grad_norm": 0.86328125, "learning_rate": 1.385254258042948e-06, "loss": 0.0707, "step": 771 }, { "epoch": 3.799507995079951, "grad_norm": 1.1484375, "learning_rate": 1.3744732785482035e-06, "loss": 0.0804, "step": 772 }, { "epoch": 3.804428044280443, "grad_norm": 1.2109375, "learning_rate": 1.3637277263798603e-06, "loss": 0.1245, "step": 773 }, { "epoch": 3.809348093480935, "grad_norm": 1.046875, "learning_rate": 1.35301770653954e-06, "loss": 0.0788, "step": 774 }, { "epoch": 3.814268142681427, "grad_norm": 0.85546875, "learning_rate": 1.3423433236816563e-06, "loss": 0.0559, "step": 775 }, { "epoch": 3.819188191881919, "grad_norm": 1.265625, "learning_rate": 1.3317046821123868e-06, "loss": 0.0958, "step": 776 }, { "epoch": 3.8241082410824108, "grad_norm": 1.3671875, "learning_rate": 1.3211018857886632e-06, "loss": 0.0748, "step": 777 }, { "epoch": 3.8290282902829027, "grad_norm": 0.80078125, "learning_rate": 1.3105350383171484e-06, "loss": 0.0673, "step": 778 }, { "epoch": 3.8339483394833946, "grad_norm": 1.234375, "learning_rate": 1.3000042429532267e-06, "loss": 0.0911, "step": 779 }, { "epoch": 3.8388683886838866, "grad_norm": 0.90625, "learning_rate": 1.289509602599996e-06, "loss": 0.0776, "step": 780 }, { "epoch": 3.843788437884379, "grad_norm": 0.9453125, "learning_rate": 1.2790512198072558e-06, "loss": 0.0996, "step": 781 }, { "epoch": 3.848708487084871, "grad_norm": 1.1328125, "learning_rate": 1.268629196770514e-06, "loss": 0.0685, "step": 782 }, { "epoch": 3.853628536285363, "grad_norm": 1.203125, "learning_rate": 1.2582436353299832e-06, "loss": 0.0924, "step": 783 }, { "epoch": 3.8585485854858548, "grad_norm": 1.28125, "learning_rate": 1.2478946369695882e-06, "loss": 0.0901, "step": 784 }, { "epoch": 3.8634686346863467, "grad_norm": 1.03125, "learning_rate": 1.2375823028159667e-06, "loss": 0.0971, "step": 785 }, { "epoch": 3.868388683886839, "grad_norm": 1.28125, "learning_rate": 1.2273067336374972e-06, "loss": 0.087, "step": 786 }, { "epoch": 3.873308733087331, "grad_norm": 1.4609375, "learning_rate": 1.2170680298432934e-06, "loss": 0.106, "step": 787 }, { "epoch": 3.878228782287823, "grad_norm": 1.15625, "learning_rate": 1.2068662914822432e-06, "loss": 0.0634, "step": 788 }, { "epoch": 3.883148831488315, "grad_norm": 1.734375, "learning_rate": 1.1967016182420122e-06, "loss": 0.0983, "step": 789 }, { "epoch": 3.888068880688807, "grad_norm": 1.1484375, "learning_rate": 1.186574109448091e-06, "loss": 0.0899, "step": 790 }, { "epoch": 3.892988929889299, "grad_norm": 1.046875, "learning_rate": 1.1764838640628011e-06, "loss": 0.0799, "step": 791 }, { "epoch": 3.8979089790897907, "grad_norm": 1.390625, "learning_rate": 1.1664309806843466e-06, "loss": 0.0942, "step": 792 }, { "epoch": 3.9028290282902827, "grad_norm": 1.4921875, "learning_rate": 1.1564155575458414e-06, "loss": 0.0895, "step": 793 }, { "epoch": 3.907749077490775, "grad_norm": 1.3359375, "learning_rate": 1.1464376925143528e-06, "loss": 0.0802, "step": 794 }, { "epoch": 3.912669126691267, "grad_norm": 1.2578125, "learning_rate": 1.1364974830899438e-06, "loss": 0.0689, "step": 795 }, { "epoch": 3.917589175891759, "grad_norm": 0.96875, "learning_rate": 1.1265950264047171e-06, "loss": 0.0677, "step": 796 }, { "epoch": 3.922509225092251, "grad_norm": 1.1328125, "learning_rate": 1.1167304192218737e-06, "loss": 0.1054, "step": 797 }, { "epoch": 3.927429274292743, "grad_norm": 1.203125, "learning_rate": 1.1069037579347613e-06, "loss": 0.1051, "step": 798 }, { "epoch": 3.932349323493235, "grad_norm": 1.0, "learning_rate": 1.0971151385659357e-06, "loss": 0.0892, "step": 799 }, { "epoch": 3.937269372693727, "grad_norm": 0.8671875, "learning_rate": 1.0873646567662165e-06, "loss": 0.0544, "step": 800 }, { "epoch": 3.942189421894219, "grad_norm": 1.328125, "learning_rate": 1.077652407813764e-06, "loss": 0.0906, "step": 801 }, { "epoch": 3.947109471094711, "grad_norm": 0.8984375, "learning_rate": 1.067978486613131e-06, "loss": 0.056, "step": 802 }, { "epoch": 3.952029520295203, "grad_norm": 1.1328125, "learning_rate": 1.0583429876943585e-06, "loss": 0.0857, "step": 803 }, { "epoch": 3.956949569495695, "grad_norm": 1.0078125, "learning_rate": 1.0487460052120262e-06, "loss": 0.0857, "step": 804 }, { "epoch": 3.961869618696187, "grad_norm": 1.5859375, "learning_rate": 1.0391876329443534e-06, "loss": 0.0975, "step": 805 }, { "epoch": 3.9667896678966788, "grad_norm": 1.0546875, "learning_rate": 1.0296679642922718e-06, "loss": 0.0768, "step": 806 }, { "epoch": 3.971709717097171, "grad_norm": 1.484375, "learning_rate": 1.0201870922785156e-06, "loss": 0.1113, "step": 807 }, { "epoch": 3.976629766297663, "grad_norm": 1.171875, "learning_rate": 1.010745109546713e-06, "loss": 0.0812, "step": 808 }, { "epoch": 3.981549815498155, "grad_norm": 0.82421875, "learning_rate": 1.0013421083604779e-06, "loss": 0.062, "step": 809 }, { "epoch": 3.986469864698647, "grad_norm": 0.96875, "learning_rate": 9.919781806025136e-07, "loss": 0.0666, "step": 810 }, { "epoch": 3.991389913899139, "grad_norm": 1.0859375, "learning_rate": 9.826534177737106e-07, "loss": 0.0616, "step": 811 }, { "epoch": 3.9963099630996313, "grad_norm": 1.140625, "learning_rate": 9.733679109922567e-07, "loss": 0.0663, "step": 812 }, { "epoch": 4.001230012300123, "grad_norm": 2.390625, "learning_rate": 9.641217509927376e-07, "loss": 0.0926, "step": 813 }, { "epoch": 4.006150061500615, "grad_norm": 1.1875, "learning_rate": 9.549150281252633e-07, "loss": 0.0742, "step": 814 }, { "epoch": 4.011070110701107, "grad_norm": 1.0078125, "learning_rate": 9.457478323545749e-07, "loss": 0.1039, "step": 815 }, { "epoch": 4.015990159901599, "grad_norm": 1.0859375, "learning_rate": 9.366202532591717e-07, "loss": 0.0794, "step": 816 }, { "epoch": 4.015990159901599, "eval_loss": 0.10322786867618561, "eval_runtime": 204.3584, "eval_samples_per_second": 8.006, "eval_steps_per_second": 1.336, "step": 816 }, { "epoch": 4.020910209102091, "grad_norm": 0.78515625, "learning_rate": 9.275323800304287e-07, "loss": 0.0628, "step": 817 }, { "epoch": 4.025830258302583, "grad_norm": 0.9296875, "learning_rate": 9.184843014717337e-07, "loss": 0.0652, "step": 818 }, { "epoch": 4.030750307503075, "grad_norm": 1.0625, "learning_rate": 9.094761059976154e-07, "loss": 0.0884, "step": 819 }, { "epoch": 4.035670356703567, "grad_norm": 1.0234375, "learning_rate": 9.005078816328772e-07, "loss": 0.0858, "step": 820 }, { "epoch": 4.040590405904059, "grad_norm": 1.125, "learning_rate": 8.915797160117423e-07, "loss": 0.1029, "step": 821 }, { "epoch": 4.045510455104551, "grad_norm": 1.140625, "learning_rate": 8.826916963769888e-07, "loss": 0.0981, "step": 822 }, { "epoch": 4.0504305043050435, "grad_norm": 1.0078125, "learning_rate": 8.738439095791123e-07, "loss": 0.0688, "step": 823 }, { "epoch": 4.055350553505535, "grad_norm": 0.98828125, "learning_rate": 8.650364420754581e-07, "loss": 0.0715, "step": 824 }, { "epoch": 4.060270602706027, "grad_norm": 0.94140625, "learning_rate": 8.562693799293931e-07, "loss": 0.0624, "step": 825 }, { "epoch": 4.065190651906519, "grad_norm": 1.328125, "learning_rate": 8.475428088094517e-07, "loss": 0.0977, "step": 826 }, { "epoch": 4.070110701107011, "grad_norm": 1.140625, "learning_rate": 8.388568139885101e-07, "loss": 0.0827, "step": 827 }, { "epoch": 4.075030750307503, "grad_norm": 1.0078125, "learning_rate": 8.30211480342945e-07, "loss": 0.0824, "step": 828 }, { "epoch": 4.079950799507995, "grad_norm": 1.0390625, "learning_rate": 8.216068923518072e-07, "loss": 0.0713, "step": 829 }, { "epoch": 4.084870848708487, "grad_norm": 1.171875, "learning_rate": 8.130431340959982e-07, "loss": 0.0905, "step": 830 }, { "epoch": 4.089790897908979, "grad_norm": 1.0390625, "learning_rate": 8.045202892574395e-07, "loss": 0.0692, "step": 831 }, { "epoch": 4.094710947109471, "grad_norm": 0.9765625, "learning_rate": 7.960384411182709e-07, "loss": 0.0713, "step": 832 }, { "epoch": 4.099630996309963, "grad_norm": 1.03125, "learning_rate": 7.875976725600193e-07, "loss": 0.0824, "step": 833 }, { "epoch": 4.104551045510455, "grad_norm": 0.76953125, "learning_rate": 7.791980660628029e-07, "loss": 0.0579, "step": 834 }, { "epoch": 4.109471094710947, "grad_norm": 0.98828125, "learning_rate": 7.708397037045129e-07, "loss": 0.081, "step": 835 }, { "epoch": 4.114391143911439, "grad_norm": 1.0078125, "learning_rate": 7.625226671600256e-07, "loss": 0.0884, "step": 836 }, { "epoch": 4.1193111931119315, "grad_norm": 1.0546875, "learning_rate": 7.542470377003897e-07, "loss": 0.0809, "step": 837 }, { "epoch": 4.124231242312423, "grad_norm": 1.0546875, "learning_rate": 7.460128961920432e-07, "loss": 0.0768, "step": 838 }, { "epoch": 4.129151291512915, "grad_norm": 0.8203125, "learning_rate": 7.37820323096014e-07, "loss": 0.0809, "step": 839 }, { "epoch": 4.134071340713407, "grad_norm": 0.8359375, "learning_rate": 7.296693984671465e-07, "loss": 0.0748, "step": 840 }, { "epoch": 4.138991389913899, "grad_norm": 1.046875, "learning_rate": 7.215602019533041e-07, "loss": 0.1046, "step": 841 }, { "epoch": 4.143911439114391, "grad_norm": 1.1015625, "learning_rate": 7.134928127946017e-07, "loss": 0.0804, "step": 842 }, { "epoch": 4.148831488314883, "grad_norm": 0.9375, "learning_rate": 7.054673098226278e-07, "loss": 0.0735, "step": 843 }, { "epoch": 4.153751537515375, "grad_norm": 1.1328125, "learning_rate": 6.974837714596732e-07, "loss": 0.0918, "step": 844 }, { "epoch": 4.158671586715867, "grad_norm": 1.1328125, "learning_rate": 6.895422757179682e-07, "loss": 0.0926, "step": 845 }, { "epoch": 4.163591635916359, "grad_norm": 1.03125, "learning_rate": 6.816429001989133e-07, "loss": 0.0887, "step": 846 }, { "epoch": 4.168511685116851, "grad_norm": 1.078125, "learning_rate": 6.737857220923305e-07, "loss": 0.0546, "step": 847 }, { "epoch": 4.173431734317343, "grad_norm": 1.015625, "learning_rate": 6.659708181757013e-07, "loss": 0.0709, "step": 848 }, { "epoch": 4.178351783517835, "grad_norm": 0.84375, "learning_rate": 6.581982648134217e-07, "loss": 0.0709, "step": 849 }, { "epoch": 4.183271832718328, "grad_norm": 1.015625, "learning_rate": 6.50468137956049e-07, "loss": 0.0639, "step": 850 }, { "epoch": 4.1881918819188195, "grad_norm": 0.90234375, "learning_rate": 6.427805131395681e-07, "loss": 0.0931, "step": 851 }, { "epoch": 4.1931119311193115, "grad_norm": 0.87890625, "learning_rate": 6.351354654846481e-07, "loss": 0.0811, "step": 852 }, { "epoch": 4.198031980319803, "grad_norm": 1.171875, "learning_rate": 6.275330696959109e-07, "loss": 0.0872, "step": 853 }, { "epoch": 4.202952029520295, "grad_norm": 0.9921875, "learning_rate": 6.199734000611968e-07, "loss": 0.0826, "step": 854 }, { "epoch": 4.207872078720787, "grad_norm": 0.81640625, "learning_rate": 6.12456530450844e-07, "loss": 0.0732, "step": 855 }, { "epoch": 4.212792127921279, "grad_norm": 0.91796875, "learning_rate": 6.049825343169652e-07, "loss": 0.0697, "step": 856 }, { "epoch": 4.217712177121771, "grad_norm": 0.99609375, "learning_rate": 5.975514846927271e-07, "loss": 0.0767, "step": 857 }, { "epoch": 4.222632226322263, "grad_norm": 0.921875, "learning_rate": 5.901634541916406e-07, "loss": 0.0632, "step": 858 }, { "epoch": 4.227552275522755, "grad_norm": 0.984375, "learning_rate": 5.828185150068472e-07, "loss": 0.0798, "step": 859 }, { "epoch": 4.232472324723247, "grad_norm": 0.796875, "learning_rate": 5.755167389104166e-07, "loss": 0.0685, "step": 860 }, { "epoch": 4.237392373923739, "grad_norm": 0.87109375, "learning_rate": 5.682581972526463e-07, "loss": 0.0645, "step": 861 }, { "epoch": 4.242312423124231, "grad_norm": 0.84375, "learning_rate": 5.610429609613615e-07, "loss": 0.0588, "step": 862 }, { "epoch": 4.247232472324724, "grad_norm": 0.7890625, "learning_rate": 5.538711005412212e-07, "loss": 0.0581, "step": 863 }, { "epoch": 4.252152521525216, "grad_norm": 0.984375, "learning_rate": 5.467426860730334e-07, "loss": 0.082, "step": 864 }, { "epoch": 4.2570725707257075, "grad_norm": 0.86328125, "learning_rate": 5.396577872130676e-07, "loss": 0.0584, "step": 865 }, { "epoch": 4.2619926199261995, "grad_norm": 1.0234375, "learning_rate": 5.32616473192375e-07, "loss": 0.0822, "step": 866 }, { "epoch": 4.266912669126691, "grad_norm": 1.1328125, "learning_rate": 5.256188128161116e-07, "loss": 0.0834, "step": 867 }, { "epoch": 4.271832718327183, "grad_norm": 1.078125, "learning_rate": 5.186648744628637e-07, "loss": 0.0703, "step": 868 }, { "epoch": 4.276752767527675, "grad_norm": 1.2734375, "learning_rate": 5.117547260839845e-07, "loss": 0.1001, "step": 869 }, { "epoch": 4.281672816728167, "grad_norm": 1.3046875, "learning_rate": 5.048884352029271e-07, "loss": 0.1088, "step": 870 }, { "epoch": 4.286592865928659, "grad_norm": 0.80859375, "learning_rate": 4.980660689145855e-07, "loss": 0.0635, "step": 871 }, { "epoch": 4.291512915129151, "grad_norm": 1.0546875, "learning_rate": 4.912876938846345e-07, "loss": 0.0876, "step": 872 }, { "epoch": 4.296432964329643, "grad_norm": 1.203125, "learning_rate": 4.845533763488902e-07, "loss": 0.0893, "step": 873 }, { "epoch": 4.301353013530135, "grad_norm": 1.2421875, "learning_rate": 4.778631821126473e-07, "loss": 0.0946, "step": 874 }, { "epoch": 4.306273062730627, "grad_norm": 1.1875, "learning_rate": 4.712171765500484e-07, "loss": 0.0911, "step": 875 }, { "epoch": 4.31119311193112, "grad_norm": 1.1171875, "learning_rate": 4.6461542460343565e-07, "loss": 0.1092, "step": 876 }, { "epoch": 4.316113161131612, "grad_norm": 1.0546875, "learning_rate": 4.580579907827287e-07, "loss": 0.0921, "step": 877 }, { "epoch": 4.321033210332104, "grad_norm": 1.0546875, "learning_rate": 4.515449391647786e-07, "loss": 0.0759, "step": 878 }, { "epoch": 4.325953259532596, "grad_norm": 1.109375, "learning_rate": 4.4507633339275494e-07, "loss": 0.0808, "step": 879 }, { "epoch": 4.3308733087330875, "grad_norm": 0.98046875, "learning_rate": 4.386522366755169e-07, "loss": 0.0842, "step": 880 }, { "epoch": 4.3357933579335795, "grad_norm": 1.078125, "learning_rate": 4.322727117869951e-07, "loss": 0.1048, "step": 881 }, { "epoch": 4.340713407134071, "grad_norm": 1.15625, "learning_rate": 4.2593782106558676e-07, "loss": 0.0922, "step": 882 }, { "epoch": 4.345633456334563, "grad_norm": 0.9765625, "learning_rate": 4.1964762641353297e-07, "loss": 0.0989, "step": 883 }, { "epoch": 4.350553505535055, "grad_norm": 1.171875, "learning_rate": 4.1340218929632636e-07, "loss": 0.0922, "step": 884 }, { "epoch": 4.355473554735547, "grad_norm": 1.078125, "learning_rate": 4.072015707421006e-07, "loss": 0.0831, "step": 885 }, { "epoch": 4.360393603936039, "grad_norm": 1.2109375, "learning_rate": 4.0104583134104593e-07, "loss": 0.092, "step": 886 }, { "epoch": 4.365313653136531, "grad_norm": 0.9921875, "learning_rate": 3.9493503124480135e-07, "loss": 0.0722, "step": 887 }, { "epoch": 4.370233702337023, "grad_norm": 1.1796875, "learning_rate": 3.8886923016588195e-07, "loss": 0.0792, "step": 888 }, { "epoch": 4.375153751537516, "grad_norm": 1.21875, "learning_rate": 3.828484873770832e-07, "loss": 0.0541, "step": 889 }, { "epoch": 4.380073800738008, "grad_norm": 1.140625, "learning_rate": 3.7687286171091355e-07, "loss": 0.1026, "step": 890 }, { "epoch": 4.3849938499385, "grad_norm": 1.0, "learning_rate": 3.709424115590088e-07, "loss": 0.0849, "step": 891 }, { "epoch": 4.389913899138992, "grad_norm": 1.109375, "learning_rate": 3.65057194871567e-07, "loss": 0.098, "step": 892 }, { "epoch": 4.394833948339484, "grad_norm": 0.8359375, "learning_rate": 3.5921726915678247e-07, "loss": 0.0659, "step": 893 }, { "epoch": 4.3997539975399755, "grad_norm": 1.09375, "learning_rate": 3.534226914802813e-07, "loss": 0.0839, "step": 894 }, { "epoch": 4.4046740467404675, "grad_norm": 0.94921875, "learning_rate": 3.4767351846456744e-07, "loss": 0.1014, "step": 895 }, { "epoch": 4.409594095940959, "grad_norm": 1.296875, "learning_rate": 3.4196980628846297e-07, "loss": 0.0843, "step": 896 }, { "epoch": 4.414514145141451, "grad_norm": 0.9609375, "learning_rate": 3.3631161068656604e-07, "loss": 0.0727, "step": 897 }, { "epoch": 4.419434194341943, "grad_norm": 1.03125, "learning_rate": 3.3069898694870373e-07, "loss": 0.0771, "step": 898 }, { "epoch": 4.424354243542435, "grad_norm": 1.21875, "learning_rate": 3.2513198991939054e-07, "loss": 0.0975, "step": 899 }, { "epoch": 4.429274292742927, "grad_norm": 0.8984375, "learning_rate": 3.196106739972926e-07, "loss": 0.0649, "step": 900 }, { "epoch": 4.434194341943419, "grad_norm": 1.015625, "learning_rate": 3.1413509313469816e-07, "loss": 0.0885, "step": 901 }, { "epoch": 4.439114391143911, "grad_norm": 0.88671875, "learning_rate": 3.087053008369889e-07, "loss": 0.069, "step": 902 }, { "epoch": 4.444034440344403, "grad_norm": 0.859375, "learning_rate": 3.0332135016211794e-07, "loss": 0.0539, "step": 903 }, { "epoch": 4.448954489544896, "grad_norm": 0.97265625, "learning_rate": 2.979832937200883e-07, "loss": 0.0869, "step": 904 }, { "epoch": 4.453874538745388, "grad_norm": 1.0, "learning_rate": 2.9269118367244385e-07, "loss": 0.0936, "step": 905 }, { "epoch": 4.45879458794588, "grad_norm": 1.109375, "learning_rate": 2.8744507173175564e-07, "loss": 0.07, "step": 906 }, { "epoch": 4.463714637146372, "grad_norm": 1.15625, "learning_rate": 2.822450091611195e-07, "loss": 0.0865, "step": 907 }, { "epoch": 4.468634686346864, "grad_norm": 1.0078125, "learning_rate": 2.770910467736532e-07, "loss": 0.0797, "step": 908 }, { "epoch": 4.4735547355473555, "grad_norm": 0.87890625, "learning_rate": 2.719832349319973e-07, "loss": 0.0591, "step": 909 }, { "epoch": 4.478474784747847, "grad_norm": 0.9765625, "learning_rate": 2.669216235478295e-07, "loss": 0.0716, "step": 910 }, { "epoch": 4.483394833948339, "grad_norm": 1.390625, "learning_rate": 2.619062620813728e-07, "loss": 0.0919, "step": 911 }, { "epoch": 4.488314883148831, "grad_norm": 1.171875, "learning_rate": 2.5693719954091257e-07, "loss": 0.073, "step": 912 }, { "epoch": 4.493234932349323, "grad_norm": 1.0, "learning_rate": 2.520144844823169e-07, "loss": 0.0683, "step": 913 }, { "epoch": 4.498154981549815, "grad_norm": 1.1328125, "learning_rate": 2.471381650085647e-07, "loss": 0.0885, "step": 914 }, { "epoch": 4.503075030750307, "grad_norm": 1.1640625, "learning_rate": 2.4230828876927293e-07, "loss": 0.0676, "step": 915 }, { "epoch": 4.507995079950799, "grad_norm": 1.078125, "learning_rate": 2.375249029602339e-07, "loss": 0.0829, "step": 916 }, { "epoch": 4.512915129151292, "grad_norm": 1.171875, "learning_rate": 2.327880543229505e-07, "loss": 0.0787, "step": 917 }, { "epoch": 4.517835178351784, "grad_norm": 1.1171875, "learning_rate": 2.2809778914418156e-07, "loss": 0.0839, "step": 918 }, { "epoch": 4.517835178351784, "eval_loss": 0.10323299467563629, "eval_runtime": 203.6707, "eval_samples_per_second": 8.033, "eval_steps_per_second": 1.34, "step": 918 }, { "epoch": 4.522755227552276, "grad_norm": 1.1875, "learning_rate": 2.2345415325549125e-07, "loss": 0.1202, "step": 919 }, { "epoch": 4.527675276752768, "grad_norm": 0.90234375, "learning_rate": 2.1885719203279587e-07, "loss": 0.0706, "step": 920 }, { "epoch": 4.53259532595326, "grad_norm": 0.953125, "learning_rate": 2.143069503959283e-07, "loss": 0.0835, "step": 921 }, { "epoch": 4.537515375153752, "grad_norm": 1.140625, "learning_rate": 2.0980347280818935e-07, "loss": 0.0678, "step": 922 }, { "epoch": 4.5424354243542435, "grad_norm": 0.98046875, "learning_rate": 2.0534680327592426e-07, "loss": 0.0662, "step": 923 }, { "epoch": 4.5473554735547355, "grad_norm": 0.93359375, "learning_rate": 2.009369853480825e-07, "loss": 0.0582, "step": 924 }, { "epoch": 4.552275522755227, "grad_norm": 1.09375, "learning_rate": 1.9657406211579966e-07, "loss": 0.1065, "step": 925 }, { "epoch": 4.557195571955719, "grad_norm": 0.97265625, "learning_rate": 1.922580762119697e-07, "loss": 0.0513, "step": 926 }, { "epoch": 4.562115621156211, "grad_norm": 1.1796875, "learning_rate": 1.8798906981083832e-07, "loss": 0.0734, "step": 927 }, { "epoch": 4.567035670356703, "grad_norm": 1.03125, "learning_rate": 1.8376708462757798e-07, "loss": 0.0901, "step": 928 }, { "epoch": 4.571955719557195, "grad_norm": 1.0703125, "learning_rate": 1.7959216191789142e-07, "loss": 0.0806, "step": 929 }, { "epoch": 4.576875768757688, "grad_norm": 1.234375, "learning_rate": 1.7546434247760147e-07, "loss": 0.0858, "step": 930 }, { "epoch": 4.58179581795818, "grad_norm": 1.0625, "learning_rate": 1.713836666422569e-07, "loss": 0.0808, "step": 931 }, { "epoch": 4.586715867158672, "grad_norm": 0.984375, "learning_rate": 1.673501742867356e-07, "loss": 0.0951, "step": 932 }, { "epoch": 4.591635916359164, "grad_norm": 1.0078125, "learning_rate": 1.633639048248542e-07, "loss": 0.0835, "step": 933 }, { "epoch": 4.596555965559656, "grad_norm": 0.9140625, "learning_rate": 1.594248972089879e-07, "loss": 0.0686, "step": 934 }, { "epoch": 4.601476014760148, "grad_norm": 1.109375, "learning_rate": 1.555331899296808e-07, "loss": 0.1139, "step": 935 }, { "epoch": 4.60639606396064, "grad_norm": 1.15625, "learning_rate": 1.5168882101528282e-07, "loss": 0.0884, "step": 936 }, { "epoch": 4.6113161131611315, "grad_norm": 1.1171875, "learning_rate": 1.4789182803156333e-07, "loss": 0.0855, "step": 937 }, { "epoch": 4.6162361623616235, "grad_norm": 1.2109375, "learning_rate": 1.44142248081357e-07, "loss": 0.0854, "step": 938 }, { "epoch": 4.621156211562115, "grad_norm": 1.3125, "learning_rate": 1.4044011780419032e-07, "loss": 0.0947, "step": 939 }, { "epoch": 4.626076260762607, "grad_norm": 1.046875, "learning_rate": 1.3678547337593494e-07, "loss": 0.0637, "step": 940 }, { "epoch": 4.630996309963099, "grad_norm": 1.140625, "learning_rate": 1.3317835050844275e-07, "loss": 0.0743, "step": 941 }, { "epoch": 4.635916359163591, "grad_norm": 1.2265625, "learning_rate": 1.296187844492053e-07, "loss": 0.0928, "step": 942 }, { "epoch": 4.640836408364084, "grad_norm": 1.140625, "learning_rate": 1.2610680998100476e-07, "loss": 0.0978, "step": 943 }, { "epoch": 4.645756457564576, "grad_norm": 0.94921875, "learning_rate": 1.2264246142157656e-07, "loss": 0.0776, "step": 944 }, { "epoch": 4.650676506765068, "grad_norm": 0.91796875, "learning_rate": 1.1922577262327374e-07, "loss": 0.0896, "step": 945 }, { "epoch": 4.65559655596556, "grad_norm": 0.95703125, "learning_rate": 1.1585677697273312e-07, "loss": 0.0723, "step": 946 }, { "epoch": 4.660516605166052, "grad_norm": 1.1796875, "learning_rate": 1.1253550739055374e-07, "loss": 0.0738, "step": 947 }, { "epoch": 4.665436654366544, "grad_norm": 1.0390625, "learning_rate": 1.0926199633097156e-07, "loss": 0.0792, "step": 948 }, { "epoch": 4.670356703567036, "grad_norm": 1.265625, "learning_rate": 1.060362757815453e-07, "loss": 0.0992, "step": 949 }, { "epoch": 4.675276752767528, "grad_norm": 1.078125, "learning_rate": 1.0285837726283999e-07, "loss": 0.0638, "step": 950 }, { "epoch": 4.68019680196802, "grad_norm": 1.0, "learning_rate": 9.972833182812225e-08, "loss": 0.0741, "step": 951 }, { "epoch": 4.6851168511685115, "grad_norm": 1.4140625, "learning_rate": 9.664617006305665e-08, "loss": 0.1135, "step": 952 }, { "epoch": 4.6900369003690034, "grad_norm": 0.90625, "learning_rate": 9.361192208540427e-08, "loss": 0.0603, "step": 953 }, { "epoch": 4.694956949569495, "grad_norm": 1.0, "learning_rate": 9.062561754473231e-08, "loss": 0.0811, "step": 954 }, { "epoch": 4.699876998769987, "grad_norm": 1.8984375, "learning_rate": 8.768728562211948e-08, "loss": 0.1155, "step": 955 }, { "epoch": 4.70479704797048, "grad_norm": 0.9140625, "learning_rate": 8.479695502987551e-08, "loss": 0.0756, "step": 956 }, { "epoch": 4.709717097170972, "grad_norm": 1.265625, "learning_rate": 8.195465401125812e-08, "loss": 0.1058, "step": 957 }, { "epoch": 4.714637146371464, "grad_norm": 0.8359375, "learning_rate": 7.916041034019773e-08, "loss": 0.0538, "step": 958 }, { "epoch": 4.719557195571956, "grad_norm": 1.0859375, "learning_rate": 7.64142513210242e-08, "loss": 0.0974, "step": 959 }, { "epoch": 4.724477244772448, "grad_norm": 1.1484375, "learning_rate": 7.371620378820555e-08, "loss": 0.0975, "step": 960 }, { "epoch": 4.72939729397294, "grad_norm": 0.953125, "learning_rate": 7.10662941060769e-08, "loss": 0.0675, "step": 961 }, { "epoch": 4.734317343173432, "grad_norm": 1.1796875, "learning_rate": 6.84645481685925e-08, "loss": 0.081, "step": 962 }, { "epoch": 4.739237392373924, "grad_norm": 0.9453125, "learning_rate": 6.59109913990641e-08, "loss": 0.0836, "step": 963 }, { "epoch": 4.744157441574416, "grad_norm": 1.1640625, "learning_rate": 6.340564874991906e-08, "loss": 0.0772, "step": 964 }, { "epoch": 4.749077490774908, "grad_norm": 1.3515625, "learning_rate": 6.094854470245326e-08, "loss": 0.1065, "step": 965 }, { "epoch": 4.7539975399753995, "grad_norm": 1.046875, "learning_rate": 5.853970326659186e-08, "loss": 0.0749, "step": 966 }, { "epoch": 4.7589175891758915, "grad_norm": 0.97265625, "learning_rate": 5.6179147980656154e-08, "loss": 0.0792, "step": 967 }, { "epoch": 4.763837638376383, "grad_norm": 1.0390625, "learning_rate": 5.3866901911132086e-08, "loss": 0.0924, "step": 968 }, { "epoch": 4.768757687576876, "grad_norm": 0.9453125, "learning_rate": 5.160298765244709e-08, "loss": 0.0824, "step": 969 }, { "epoch": 4.773677736777367, "grad_norm": 0.96484375, "learning_rate": 4.9387427326745287e-08, "loss": 0.0622, "step": 970 }, { "epoch": 4.77859778597786, "grad_norm": 0.85546875, "learning_rate": 4.722024258367597e-08, "loss": 0.0668, "step": 971 }, { "epoch": 4.783517835178352, "grad_norm": 1.359375, "learning_rate": 4.5101454600177676e-08, "loss": 0.1176, "step": 972 }, { "epoch": 4.788437884378844, "grad_norm": 1.2109375, "learning_rate": 4.303108408027668e-08, "loss": 0.1058, "step": 973 }, { "epoch": 4.793357933579336, "grad_norm": 0.9921875, "learning_rate": 4.1009151254878254e-08, "loss": 0.091, "step": 974 }, { "epoch": 4.798277982779828, "grad_norm": 0.88671875, "learning_rate": 3.903567588157353e-08, "loss": 0.0783, "step": 975 }, { "epoch": 4.80319803198032, "grad_norm": 1.015625, "learning_rate": 3.711067724444517e-08, "loss": 0.0883, "step": 976 }, { "epoch": 4.808118081180812, "grad_norm": 1.1640625, "learning_rate": 3.523417415387864e-08, "loss": 0.0801, "step": 977 }, { "epoch": 4.813038130381304, "grad_norm": 1.0859375, "learning_rate": 3.340618494637793e-08, "loss": 0.0695, "step": 978 }, { "epoch": 4.817958179581796, "grad_norm": 1.0234375, "learning_rate": 3.162672748438844e-08, "loss": 0.0959, "step": 979 }, { "epoch": 4.822878228782288, "grad_norm": 1.21875, "learning_rate": 2.989581915611994e-08, "loss": 0.0944, "step": 980 }, { "epoch": 4.8277982779827795, "grad_norm": 0.890625, "learning_rate": 2.821347687537834e-08, "loss": 0.0634, "step": 981 }, { "epoch": 4.832718327183272, "grad_norm": 0.921875, "learning_rate": 2.657971708139917e-08, "loss": 0.0954, "step": 982 }, { "epoch": 4.837638376383763, "grad_norm": 1.2265625, "learning_rate": 2.4994555738688252e-08, "loss": 0.0935, "step": 983 }, { "epoch": 4.842558425584256, "grad_norm": 1.0859375, "learning_rate": 2.3458008336864623e-08, "loss": 0.0965, "step": 984 }, { "epoch": 4.847478474784748, "grad_norm": 1.078125, "learning_rate": 2.1970089890509527e-08, "loss": 0.0944, "step": 985 }, { "epoch": 4.85239852398524, "grad_norm": 1.21875, "learning_rate": 2.0530814939020428e-08, "loss": 0.0842, "step": 986 }, { "epoch": 4.857318573185732, "grad_norm": 1.140625, "learning_rate": 1.9140197546467787e-08, "loss": 0.0857, "step": 987 }, { "epoch": 4.862238622386224, "grad_norm": 1.046875, "learning_rate": 1.7798251301458512e-08, "loss": 0.0715, "step": 988 }, { "epoch": 4.867158671586716, "grad_norm": 0.953125, "learning_rate": 1.6504989317001618e-08, "loss": 0.0868, "step": 989 }, { "epoch": 4.872078720787208, "grad_norm": 1.2578125, "learning_rate": 1.5260424230382763e-08, "loss": 0.0842, "step": 990 }, { "epoch": 4.8769987699877, "grad_norm": 1.1171875, "learning_rate": 1.4064568203037699e-08, "loss": 0.0964, "step": 991 }, { "epoch": 4.881918819188192, "grad_norm": 1.203125, "learning_rate": 1.2917432920437345e-08, "loss": 0.089, "step": 992 }, { "epoch": 4.886838868388684, "grad_norm": 0.875, "learning_rate": 1.1819029591968456e-08, "loss": 0.0584, "step": 993 }, { "epoch": 4.891758917589176, "grad_norm": 1.09375, "learning_rate": 1.0769368950829252e-08, "loss": 0.0723, "step": 994 }, { "epoch": 4.8966789667896675, "grad_norm": 1.0390625, "learning_rate": 9.768461253920614e-09, "loss": 0.1013, "step": 995 }, { "epoch": 4.9015990159901595, "grad_norm": 0.9921875, "learning_rate": 8.816316281747839e-09, "loss": 0.0699, "step": 996 }, { "epoch": 4.906519065190652, "grad_norm": 0.921875, "learning_rate": 7.912943338324598e-09, "loss": 0.0806, "step": 997 }, { "epoch": 4.911439114391144, "grad_norm": 0.8359375, "learning_rate": 7.058351251083007e-09, "loss": 0.0705, "step": 998 }, { "epoch": 4.916359163591636, "grad_norm": 1.0625, "learning_rate": 6.252548370784817e-09, "loss": 0.072, "step": 999 }, { "epoch": 4.921279212792128, "grad_norm": 0.921875, "learning_rate": 5.495542571443135e-09, "loss": 0.0715, "step": 1000 }, { "epoch": 4.92619926199262, "grad_norm": 0.9453125, "learning_rate": 4.787341250241384e-09, "loss": 0.0838, "step": 1001 }, { "epoch": 4.931119311193112, "grad_norm": 0.890625, "learning_rate": 4.127951327466684e-09, "loss": 0.0672, "step": 1002 }, { "epoch": 4.936039360393604, "grad_norm": 1.0703125, "learning_rate": 3.5173792464360256e-09, "loss": 0.0753, "step": 1003 }, { "epoch": 4.940959409594096, "grad_norm": 0.94140625, "learning_rate": 2.955630973437429e-09, "loss": 0.0858, "step": 1004 }, { "epoch": 4.945879458794588, "grad_norm": 0.83984375, "learning_rate": 2.442711997670544e-09, "loss": 0.0674, "step": 1005 }, { "epoch": 4.95079950799508, "grad_norm": 1.0, "learning_rate": 1.978627331192806e-09, "loss": 0.0892, "step": 1006 }, { "epoch": 4.955719557195572, "grad_norm": 1.0625, "learning_rate": 1.5633815088705872e-09, "loss": 0.074, "step": 1007 }, { "epoch": 4.960639606396064, "grad_norm": 1.109375, "learning_rate": 1.196978588334785e-09, "loss": 0.0951, "step": 1008 }, { "epoch": 4.9655596555965555, "grad_norm": 1.1796875, "learning_rate": 8.794221499408562e-10, "loss": 0.0977, "step": 1009 }, { "epoch": 4.970479704797048, "grad_norm": 1.2578125, "learning_rate": 6.107152967349539e-10, "loss": 0.1068, "step": 1010 }, { "epoch": 4.97539975399754, "grad_norm": 1.0078125, "learning_rate": 3.908606544228422e-10, "loss": 0.0803, "step": 1011 }, { "epoch": 4.980319803198032, "grad_norm": 0.88671875, "learning_rate": 2.1986037134325012e-10, "loss": 0.0667, "step": 1012 }, { "epoch": 4.985239852398524, "grad_norm": 0.86328125, "learning_rate": 9.771611844955343e-11, "loss": 0.0616, "step": 1013 }, { "epoch": 4.990159901599016, "grad_norm": 1.0546875, "learning_rate": 2.4429089290345375e-11, "loss": 0.088, "step": 1014 }, { "epoch": 4.995079950799508, "grad_norm": 1.03125, "learning_rate": 0.0, "loss": 0.1053, "step": 1015 } ], "logging_steps": 1, "max_steps": 1015, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 51, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.1980182581321662e+19, "train_batch_size": 2, "trial_name": null, "trial_params": null }