{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.5781153767193594, "eval_steps": 88, "global_step": 704, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.000821186614658181, "grad_norm": 0.480191707611084, "learning_rate": 2e-05, "loss": 1.332, "step": 1 }, { "epoch": 0.001642373229316362, "grad_norm": 0.43048733472824097, "learning_rate": 4e-05, "loss": 1.2784, "step": 2 }, { "epoch": 0.002463559843974543, "grad_norm": 0.33739173412323, "learning_rate": 6e-05, "loss": 1.3286, "step": 3 }, { "epoch": 0.003284746458632724, "grad_norm": 0.432579904794693, "learning_rate": 8e-05, "loss": 1.3079, "step": 4 }, { "epoch": 0.0041059330732909054, "grad_norm": 0.3490436375141144, "learning_rate": 0.0001, "loss": 1.2182, "step": 5 }, { "epoch": 0.004927119687949086, "grad_norm": 0.20206260681152344, "learning_rate": 9.997257268239166e-05, "loss": 1.2828, "step": 6 }, { "epoch": 0.005748306302607267, "grad_norm": 0.15355628728866577, "learning_rate": 9.994514536478333e-05, "loss": 1.1616, "step": 7 }, { "epoch": 0.006569492917265448, "grad_norm": 0.16756780445575714, "learning_rate": 9.9917718047175e-05, "loss": 1.0935, "step": 8 }, { "epoch": 0.00739067953192363, "grad_norm": 0.17429664731025696, "learning_rate": 9.989029072956665e-05, "loss": 0.9694, "step": 9 }, { "epoch": 0.008211866146581811, "grad_norm": 0.22355175018310547, "learning_rate": 9.986286341195832e-05, "loss": 0.9407, "step": 10 }, { "epoch": 0.009033052761239991, "grad_norm": 0.33407703042030334, "learning_rate": 9.983543609434997e-05, "loss": 0.7717, "step": 11 }, { "epoch": 0.009854239375898173, "grad_norm": 0.47473278641700745, "learning_rate": 9.980800877674164e-05, "loss": 0.7763, "step": 12 }, { "epoch": 0.010675425990556354, "grad_norm": 0.2812059819698334, "learning_rate": 9.978058145913331e-05, "loss": 0.6258, "step": 13 }, { "epoch": 0.011496612605214535, "grad_norm": 0.23547925055027008, "learning_rate": 9.975315414152496e-05, "loss": 0.5968, "step": 14 }, { "epoch": 0.012317799219872716, "grad_norm": 0.18453630805015564, "learning_rate": 9.972572682391662e-05, "loss": 0.5368, "step": 15 }, { "epoch": 0.013138985834530896, "grad_norm": 0.16103577613830566, "learning_rate": 9.969829950630828e-05, "loss": 0.4909, "step": 16 }, { "epoch": 0.013960172449189078, "grad_norm": 0.18651455640792847, "learning_rate": 9.967087218869995e-05, "loss": 0.5135, "step": 17 }, { "epoch": 0.01478135906384726, "grad_norm": 0.11300642043352127, "learning_rate": 9.96434448710916e-05, "loss": 0.5069, "step": 18 }, { "epoch": 0.01560254567850544, "grad_norm": 0.10415703803300858, "learning_rate": 9.961601755348327e-05, "loss": 0.4851, "step": 19 }, { "epoch": 0.016423732293163622, "grad_norm": 0.11693017929792404, "learning_rate": 9.958859023587493e-05, "loss": 0.4625, "step": 20 }, { "epoch": 0.017244918907821802, "grad_norm": 0.10035043954849243, "learning_rate": 9.95611629182666e-05, "loss": 0.4822, "step": 21 }, { "epoch": 0.018066105522479982, "grad_norm": 0.10483390837907791, "learning_rate": 9.953373560065826e-05, "loss": 0.4342, "step": 22 }, { "epoch": 0.018887292137138165, "grad_norm": 2.8405802249908447, "learning_rate": 9.950630828304992e-05, "loss": 0.4664, "step": 23 }, { "epoch": 0.019708478751796345, "grad_norm": 0.13821998238563538, "learning_rate": 9.947888096544159e-05, "loss": 0.4468, "step": 24 }, { "epoch": 0.020529665366454525, "grad_norm": 0.1991378366947174, "learning_rate": 9.945145364783325e-05, "loss": 0.4605, "step": 25 }, { "epoch": 0.02135085198111271, "grad_norm": 0.07619134336709976, "learning_rate": 9.942402633022491e-05, "loss": 0.4597, "step": 26 }, { "epoch": 0.02217203859577089, "grad_norm": 0.13373583555221558, "learning_rate": 9.939659901261658e-05, "loss": 0.4626, "step": 27 }, { "epoch": 0.02299322521042907, "grad_norm": 0.09962721168994904, "learning_rate": 9.936917169500823e-05, "loss": 0.4638, "step": 28 }, { "epoch": 0.023814411825087253, "grad_norm": 0.09395964443683624, "learning_rate": 9.93417443773999e-05, "loss": 0.4569, "step": 29 }, { "epoch": 0.024635598439745433, "grad_norm": 0.09109952300786972, "learning_rate": 9.931431705979157e-05, "loss": 0.4439, "step": 30 }, { "epoch": 0.025456785054403613, "grad_norm": 0.10370515286922455, "learning_rate": 9.928688974218322e-05, "loss": 0.4425, "step": 31 }, { "epoch": 0.026277971669061793, "grad_norm": 0.2153477966785431, "learning_rate": 9.925946242457488e-05, "loss": 0.4503, "step": 32 }, { "epoch": 0.027099158283719976, "grad_norm": 0.08772841095924377, "learning_rate": 9.923203510696654e-05, "loss": 0.419, "step": 33 }, { "epoch": 0.027920344898378156, "grad_norm": 0.10951374471187592, "learning_rate": 9.920460778935821e-05, "loss": 0.4353, "step": 34 }, { "epoch": 0.028741531513036336, "grad_norm": 0.09190870076417923, "learning_rate": 9.917718047174987e-05, "loss": 0.5196, "step": 35 }, { "epoch": 0.02956271812769452, "grad_norm": 0.07667124271392822, "learning_rate": 9.914975315414153e-05, "loss": 0.4358, "step": 36 }, { "epoch": 0.0303839047423527, "grad_norm": 0.1514267474412918, "learning_rate": 9.912232583653319e-05, "loss": 0.411, "step": 37 }, { "epoch": 0.03120509135701088, "grad_norm": 0.09086549282073975, "learning_rate": 9.909489851892486e-05, "loss": 0.4003, "step": 38 }, { "epoch": 0.032026277971669063, "grad_norm": 0.2616782486438751, "learning_rate": 9.906747120131652e-05, "loss": 0.4842, "step": 39 }, { "epoch": 0.032847464586327244, "grad_norm": 0.11908283084630966, "learning_rate": 9.904004388370818e-05, "loss": 0.4143, "step": 40 }, { "epoch": 0.033668651200985424, "grad_norm": 0.07770542800426483, "learning_rate": 9.901261656609983e-05, "loss": 0.3873, "step": 41 }, { "epoch": 0.034489837815643604, "grad_norm": 0.08934606611728668, "learning_rate": 9.89851892484915e-05, "loss": 0.4235, "step": 42 }, { "epoch": 0.035311024430301784, "grad_norm": 0.09303563088178635, "learning_rate": 9.895776193088317e-05, "loss": 0.4103, "step": 43 }, { "epoch": 0.036132211044959964, "grad_norm": 0.08622181415557861, "learning_rate": 9.893033461327482e-05, "loss": 0.448, "step": 44 }, { "epoch": 0.03695339765961815, "grad_norm": 0.08822862058877945, "learning_rate": 9.890290729566649e-05, "loss": 0.3855, "step": 45 }, { "epoch": 0.03777458427427633, "grad_norm": 0.08557698875665665, "learning_rate": 9.887547997805814e-05, "loss": 0.3945, "step": 46 }, { "epoch": 0.03859577088893451, "grad_norm": 0.07540106773376465, "learning_rate": 9.884805266044981e-05, "loss": 0.4192, "step": 47 }, { "epoch": 0.03941695750359269, "grad_norm": 0.1023702397942543, "learning_rate": 9.882062534284148e-05, "loss": 0.4126, "step": 48 }, { "epoch": 0.04023814411825087, "grad_norm": 0.07779772579669952, "learning_rate": 9.879319802523313e-05, "loss": 0.4244, "step": 49 }, { "epoch": 0.04105933073290905, "grad_norm": 0.08826564252376556, "learning_rate": 9.876577070762479e-05, "loss": 0.415, "step": 50 }, { "epoch": 0.04188051734756724, "grad_norm": 0.08254576474428177, "learning_rate": 9.873834339001646e-05, "loss": 0.4346, "step": 51 }, { "epoch": 0.04270170396222542, "grad_norm": 0.08287151902914047, "learning_rate": 9.871091607240812e-05, "loss": 0.4142, "step": 52 }, { "epoch": 0.0435228905768836, "grad_norm": 0.08196476101875305, "learning_rate": 9.868348875479978e-05, "loss": 0.3822, "step": 53 }, { "epoch": 0.04434407719154178, "grad_norm": 0.08654092252254486, "learning_rate": 9.865606143719145e-05, "loss": 0.3937, "step": 54 }, { "epoch": 0.04516526380619996, "grad_norm": 0.1102684736251831, "learning_rate": 9.86286341195831e-05, "loss": 0.4508, "step": 55 }, { "epoch": 0.04598645042085814, "grad_norm": 0.08240954577922821, "learning_rate": 9.860120680197478e-05, "loss": 0.4069, "step": 56 }, { "epoch": 0.04680763703551632, "grad_norm": 0.08542217314243317, "learning_rate": 9.857377948436644e-05, "loss": 0.5002, "step": 57 }, { "epoch": 0.047628823650174505, "grad_norm": 0.08390172570943832, "learning_rate": 9.854635216675809e-05, "loss": 0.3851, "step": 58 }, { "epoch": 0.048450010264832685, "grad_norm": 0.10694168508052826, "learning_rate": 9.851892484914976e-05, "loss": 0.4026, "step": 59 }, { "epoch": 0.049271196879490865, "grad_norm": 0.0852806493639946, "learning_rate": 9.849149753154143e-05, "loss": 0.424, "step": 60 }, { "epoch": 0.050092383494149045, "grad_norm": 0.15425831079483032, "learning_rate": 9.846407021393308e-05, "loss": 0.4307, "step": 61 }, { "epoch": 0.050913570108807225, "grad_norm": 0.08546218276023865, "learning_rate": 9.843664289632475e-05, "loss": 0.4386, "step": 62 }, { "epoch": 0.051734756723465405, "grad_norm": 0.08588795363903046, "learning_rate": 9.84092155787164e-05, "loss": 0.4433, "step": 63 }, { "epoch": 0.052555943338123585, "grad_norm": 0.08901514858007431, "learning_rate": 9.838178826110807e-05, "loss": 0.4263, "step": 64 }, { "epoch": 0.05337712995278177, "grad_norm": 0.17668095231056213, "learning_rate": 9.835436094349974e-05, "loss": 0.407, "step": 65 }, { "epoch": 0.05419831656743995, "grad_norm": 0.081763856112957, "learning_rate": 9.83269336258914e-05, "loss": 0.374, "step": 66 }, { "epoch": 0.05501950318209813, "grad_norm": 0.09063572436571121, "learning_rate": 9.829950630828305e-05, "loss": 0.3839, "step": 67 }, { "epoch": 0.05584068979675631, "grad_norm": 0.08264392614364624, "learning_rate": 9.827207899067472e-05, "loss": 0.4676, "step": 68 }, { "epoch": 0.05666187641141449, "grad_norm": 0.0827123150229454, "learning_rate": 9.824465167306638e-05, "loss": 0.3801, "step": 69 }, { "epoch": 0.05748306302607267, "grad_norm": 0.07972189038991928, "learning_rate": 9.821722435545804e-05, "loss": 0.3758, "step": 70 }, { "epoch": 0.05830424964073085, "grad_norm": 0.08738942444324493, "learning_rate": 9.81897970378497e-05, "loss": 0.3634, "step": 71 }, { "epoch": 0.05912543625538904, "grad_norm": 0.08442792296409607, "learning_rate": 9.816236972024136e-05, "loss": 0.3803, "step": 72 }, { "epoch": 0.05994662287004722, "grad_norm": 0.2751137316226959, "learning_rate": 9.813494240263303e-05, "loss": 0.3869, "step": 73 }, { "epoch": 0.0607678094847054, "grad_norm": 0.19429296255111694, "learning_rate": 9.81075150850247e-05, "loss": 0.3992, "step": 74 }, { "epoch": 0.06158899609936358, "grad_norm": 0.08392605930566788, "learning_rate": 9.808008776741635e-05, "loss": 0.361, "step": 75 }, { "epoch": 0.06241018271402176, "grad_norm": 2.01043963432312, "learning_rate": 9.8052660449808e-05, "loss": 0.3787, "step": 76 }, { "epoch": 0.06323136932867994, "grad_norm": 0.0808538943529129, "learning_rate": 9.802523313219967e-05, "loss": 0.4013, "step": 77 }, { "epoch": 0.06405255594333813, "grad_norm": 0.07846518605947495, "learning_rate": 9.799780581459134e-05, "loss": 0.3808, "step": 78 }, { "epoch": 0.0648737425579963, "grad_norm": 0.08152970671653748, "learning_rate": 9.7970378496983e-05, "loss": 0.4076, "step": 79 }, { "epoch": 0.06569492917265449, "grad_norm": 0.07745791226625443, "learning_rate": 9.794295117937466e-05, "loss": 0.3981, "step": 80 }, { "epoch": 0.06651611578731266, "grad_norm": 0.0842173770070076, "learning_rate": 9.791552386176632e-05, "loss": 0.4469, "step": 81 }, { "epoch": 0.06733730240197085, "grad_norm": 0.07920663058757782, "learning_rate": 9.788809654415799e-05, "loss": 0.3988, "step": 82 }, { "epoch": 0.06815848901662903, "grad_norm": 0.07899456471204758, "learning_rate": 9.786066922654965e-05, "loss": 0.4418, "step": 83 }, { "epoch": 0.06897967563128721, "grad_norm": 0.07229727506637573, "learning_rate": 9.783324190894131e-05, "loss": 0.4016, "step": 84 }, { "epoch": 0.0698008622459454, "grad_norm": 0.08062436431646347, "learning_rate": 9.780581459133296e-05, "loss": 0.3762, "step": 85 }, { "epoch": 0.07062204886060357, "grad_norm": 0.08440960198640823, "learning_rate": 9.777838727372464e-05, "loss": 0.4362, "step": 86 }, { "epoch": 0.07144323547526175, "grad_norm": 0.07544733583927155, "learning_rate": 9.77509599561163e-05, "loss": 0.4135, "step": 87 }, { "epoch": 0.07226442208991993, "grad_norm": 0.1760656237602234, "learning_rate": 9.772353263850797e-05, "loss": 0.3887, "step": 88 }, { "epoch": 0.07226442208991993, "eval_runtime": 494.3052, "eval_samples_per_second": 0.399, "eval_steps_per_second": 0.2, "step": 88 }, { "epoch": 0.07308560870457811, "grad_norm": 0.08049104362726212, "learning_rate": 9.769610532089962e-05, "loss": 0.4272, "step": 89 }, { "epoch": 0.0739067953192363, "grad_norm": 0.07617965340614319, "learning_rate": 9.766867800329129e-05, "loss": 0.3873, "step": 90 }, { "epoch": 0.07472798193389447, "grad_norm": 0.07975462824106216, "learning_rate": 9.764125068568296e-05, "loss": 0.3903, "step": 91 }, { "epoch": 0.07554916854855266, "grad_norm": 0.08299189805984497, "learning_rate": 9.761382336807461e-05, "loss": 0.404, "step": 92 }, { "epoch": 0.07637035516321083, "grad_norm": 0.08578819036483765, "learning_rate": 9.758639605046626e-05, "loss": 0.3983, "step": 93 }, { "epoch": 0.07719154177786902, "grad_norm": 0.10872071981430054, "learning_rate": 9.755896873285793e-05, "loss": 0.3553, "step": 94 }, { "epoch": 0.07801272839252721, "grad_norm": 0.08121436834335327, "learning_rate": 9.75315414152496e-05, "loss": 0.408, "step": 95 }, { "epoch": 0.07883391500718538, "grad_norm": 0.09175996482372284, "learning_rate": 9.750411409764125e-05, "loss": 0.4008, "step": 96 }, { "epoch": 0.07965510162184357, "grad_norm": 0.08122789114713669, "learning_rate": 9.747668678003292e-05, "loss": 0.4045, "step": 97 }, { "epoch": 0.08047628823650174, "grad_norm": 0.08402436226606369, "learning_rate": 9.744925946242458e-05, "loss": 0.3814, "step": 98 }, { "epoch": 0.08129747485115993, "grad_norm": 0.08454861491918564, "learning_rate": 9.742183214481624e-05, "loss": 0.3904, "step": 99 }, { "epoch": 0.0821186614658181, "grad_norm": 0.08125888556241989, "learning_rate": 9.739440482720791e-05, "loss": 0.3681, "step": 100 }, { "epoch": 0.08293984808047629, "grad_norm": 0.08544078469276428, "learning_rate": 9.736697750959957e-05, "loss": 0.367, "step": 101 }, { "epoch": 0.08376103469513448, "grad_norm": 0.07539089769124985, "learning_rate": 9.733955019199122e-05, "loss": 0.3821, "step": 102 }, { "epoch": 0.08458222130979265, "grad_norm": 0.07529085874557495, "learning_rate": 9.731212287438289e-05, "loss": 0.4169, "step": 103 }, { "epoch": 0.08540340792445084, "grad_norm": 0.07588034868240356, "learning_rate": 9.728469555677456e-05, "loss": 0.4211, "step": 104 }, { "epoch": 0.08622459453910901, "grad_norm": 0.08019097149372101, "learning_rate": 9.725726823916621e-05, "loss": 0.4033, "step": 105 }, { "epoch": 0.0870457811537672, "grad_norm": 0.07878712564706802, "learning_rate": 9.722984092155788e-05, "loss": 0.3842, "step": 106 }, { "epoch": 0.08786696776842537, "grad_norm": 0.08083963394165039, "learning_rate": 9.720241360394953e-05, "loss": 0.3558, "step": 107 }, { "epoch": 0.08868815438308356, "grad_norm": 0.08340411633253098, "learning_rate": 9.71749862863412e-05, "loss": 0.4037, "step": 108 }, { "epoch": 0.08950934099774174, "grad_norm": 0.09317754209041595, "learning_rate": 9.714755896873287e-05, "loss": 0.3874, "step": 109 }, { "epoch": 0.09033052761239992, "grad_norm": 0.07143125683069229, "learning_rate": 9.712013165112452e-05, "loss": 0.3441, "step": 110 }, { "epoch": 0.0911517142270581, "grad_norm": 0.0909111350774765, "learning_rate": 9.709270433351618e-05, "loss": 0.3552, "step": 111 }, { "epoch": 0.09197290084171628, "grad_norm": 0.08041603863239288, "learning_rate": 9.706527701590785e-05, "loss": 0.394, "step": 112 }, { "epoch": 0.09279408745637446, "grad_norm": 0.08147992193698883, "learning_rate": 9.703784969829951e-05, "loss": 0.3916, "step": 113 }, { "epoch": 0.09361527407103264, "grad_norm": 0.08365318924188614, "learning_rate": 9.701042238069117e-05, "loss": 0.4075, "step": 114 }, { "epoch": 0.09443646068569082, "grad_norm": 0.083246149122715, "learning_rate": 9.698299506308284e-05, "loss": 0.3566, "step": 115 }, { "epoch": 0.09525764730034901, "grad_norm": 0.0942649245262146, "learning_rate": 9.695556774547449e-05, "loss": 0.3516, "step": 116 }, { "epoch": 0.09607883391500718, "grad_norm": 0.08110091835260391, "learning_rate": 9.692814042786616e-05, "loss": 0.3902, "step": 117 }, { "epoch": 0.09690002052966537, "grad_norm": 0.43124014139175415, "learning_rate": 9.690071311025783e-05, "loss": 0.3956, "step": 118 }, { "epoch": 0.09772120714432354, "grad_norm": 0.08204706013202667, "learning_rate": 9.687328579264948e-05, "loss": 0.3865, "step": 119 }, { "epoch": 0.09854239375898173, "grad_norm": 0.08262762427330017, "learning_rate": 9.684585847504113e-05, "loss": 0.3777, "step": 120 }, { "epoch": 0.0993635803736399, "grad_norm": 0.08233962953090668, "learning_rate": 9.681843115743282e-05, "loss": 0.3976, "step": 121 }, { "epoch": 0.10018476698829809, "grad_norm": 0.7428120970726013, "learning_rate": 9.679100383982447e-05, "loss": 0.377, "step": 122 }, { "epoch": 0.10100595360295628, "grad_norm": 0.07909400761127472, "learning_rate": 9.676357652221614e-05, "loss": 0.3654, "step": 123 }, { "epoch": 0.10182714021761445, "grad_norm": 0.08714035898447037, "learning_rate": 9.673614920460779e-05, "loss": 0.4181, "step": 124 }, { "epoch": 0.10264832683227264, "grad_norm": 0.08017311245203018, "learning_rate": 9.670872188699946e-05, "loss": 0.4052, "step": 125 }, { "epoch": 0.10346951344693081, "grad_norm": 0.13821078836917877, "learning_rate": 9.668129456939113e-05, "loss": 0.3733, "step": 126 }, { "epoch": 0.104290700061589, "grad_norm": 0.1609969437122345, "learning_rate": 9.665386725178278e-05, "loss": 0.3686, "step": 127 }, { "epoch": 0.10511188667624717, "grad_norm": 0.1280309557914734, "learning_rate": 9.662643993417444e-05, "loss": 0.4036, "step": 128 }, { "epoch": 0.10593307329090536, "grad_norm": 0.08747898787260056, "learning_rate": 9.65990126165661e-05, "loss": 0.3587, "step": 129 }, { "epoch": 0.10675425990556354, "grad_norm": 0.1252209097146988, "learning_rate": 9.657158529895777e-05, "loss": 0.3626, "step": 130 }, { "epoch": 0.10757544652022172, "grad_norm": 0.09374388307332993, "learning_rate": 9.654415798134943e-05, "loss": 0.3824, "step": 131 }, { "epoch": 0.1083966331348799, "grad_norm": 0.08730709552764893, "learning_rate": 9.65167306637411e-05, "loss": 0.3827, "step": 132 }, { "epoch": 0.10921781974953808, "grad_norm": 0.07719024270772934, "learning_rate": 9.648930334613275e-05, "loss": 0.3762, "step": 133 }, { "epoch": 0.11003900636419627, "grad_norm": 0.08843278139829636, "learning_rate": 9.646187602852442e-05, "loss": 0.3843, "step": 134 }, { "epoch": 0.11086019297885444, "grad_norm": 0.08568207919597626, "learning_rate": 9.643444871091608e-05, "loss": 0.3594, "step": 135 }, { "epoch": 0.11168137959351263, "grad_norm": 0.08556952327489853, "learning_rate": 9.640702139330774e-05, "loss": 0.3886, "step": 136 }, { "epoch": 0.11250256620817081, "grad_norm": 0.09016801416873932, "learning_rate": 9.63795940756994e-05, "loss": 0.4023, "step": 137 }, { "epoch": 0.11332375282282899, "grad_norm": 0.08133590966463089, "learning_rate": 9.635216675809106e-05, "loss": 0.3634, "step": 138 }, { "epoch": 0.11414493943748717, "grad_norm": 0.2211730182170868, "learning_rate": 9.632473944048273e-05, "loss": 0.3518, "step": 139 }, { "epoch": 0.11496612605214535, "grad_norm": 0.08816584199666977, "learning_rate": 9.629731212287438e-05, "loss": 0.3727, "step": 140 }, { "epoch": 0.11578731266680353, "grad_norm": 0.11618969589471817, "learning_rate": 9.626988480526605e-05, "loss": 0.4247, "step": 141 }, { "epoch": 0.1166084992814617, "grad_norm": 0.09243030101060867, "learning_rate": 9.62424574876577e-05, "loss": 0.396, "step": 142 }, { "epoch": 0.11742968589611989, "grad_norm": 0.08566376566886902, "learning_rate": 9.621503017004937e-05, "loss": 0.3852, "step": 143 }, { "epoch": 0.11825087251077808, "grad_norm": 0.08220973610877991, "learning_rate": 9.618760285244104e-05, "loss": 0.3961, "step": 144 }, { "epoch": 0.11907205912543625, "grad_norm": 0.08240345865488052, "learning_rate": 9.61601755348327e-05, "loss": 0.3518, "step": 145 }, { "epoch": 0.11989324574009444, "grad_norm": 0.08472532778978348, "learning_rate": 9.613274821722435e-05, "loss": 0.3586, "step": 146 }, { "epoch": 0.12071443235475261, "grad_norm": 0.08407485485076904, "learning_rate": 9.610532089961602e-05, "loss": 0.3797, "step": 147 }, { "epoch": 0.1215356189694108, "grad_norm": 0.09284385293722153, "learning_rate": 9.607789358200769e-05, "loss": 0.3499, "step": 148 }, { "epoch": 0.12235680558406897, "grad_norm": 0.08499818295240402, "learning_rate": 9.605046626439934e-05, "loss": 0.3722, "step": 149 }, { "epoch": 0.12317799219872716, "grad_norm": 0.080271415412426, "learning_rate": 9.602303894679101e-05, "loss": 0.3871, "step": 150 }, { "epoch": 0.12399917881338535, "grad_norm": 0.07850060611963272, "learning_rate": 9.599561162918266e-05, "loss": 0.3679, "step": 151 }, { "epoch": 0.12482036542804352, "grad_norm": 0.07685016840696335, "learning_rate": 9.596818431157433e-05, "loss": 0.4078, "step": 152 }, { "epoch": 0.1256415520427017, "grad_norm": 0.09402357786893845, "learning_rate": 9.5940756993966e-05, "loss": 0.3996, "step": 153 }, { "epoch": 0.12646273865735988, "grad_norm": 0.08445476740598679, "learning_rate": 9.591332967635765e-05, "loss": 0.3841, "step": 154 }, { "epoch": 0.12728392527201807, "grad_norm": 0.08233911544084549, "learning_rate": 9.588590235874932e-05, "loss": 0.3529, "step": 155 }, { "epoch": 0.12810511188667625, "grad_norm": 0.07896068692207336, "learning_rate": 9.585847504114099e-05, "loss": 0.3585, "step": 156 }, { "epoch": 0.12892629850133444, "grad_norm": 0.0822276696562767, "learning_rate": 9.583104772353264e-05, "loss": 0.3962, "step": 157 }, { "epoch": 0.1297474851159926, "grad_norm": 0.07977598905563354, "learning_rate": 9.580362040592431e-05, "loss": 0.454, "step": 158 }, { "epoch": 0.1305686717306508, "grad_norm": 0.0857616737484932, "learning_rate": 9.577619308831597e-05, "loss": 0.3954, "step": 159 }, { "epoch": 0.13138985834530897, "grad_norm": 0.0874355211853981, "learning_rate": 9.574876577070763e-05, "loss": 0.3591, "step": 160 }, { "epoch": 0.13221104495996716, "grad_norm": 0.07877468317747116, "learning_rate": 9.57213384530993e-05, "loss": 0.348, "step": 161 }, { "epoch": 0.13303223157462532, "grad_norm": 0.08618593961000443, "learning_rate": 9.569391113549096e-05, "loss": 0.3723, "step": 162 }, { "epoch": 0.1338534181892835, "grad_norm": 0.08144336938858032, "learning_rate": 9.566648381788261e-05, "loss": 0.4146, "step": 163 }, { "epoch": 0.1346746048039417, "grad_norm": 0.07322760671377182, "learning_rate": 9.563905650027428e-05, "loss": 0.3107, "step": 164 }, { "epoch": 0.13549579141859988, "grad_norm": 0.08007095754146576, "learning_rate": 9.561162918266595e-05, "loss": 0.4128, "step": 165 }, { "epoch": 0.13631697803325807, "grad_norm": 0.09636646509170532, "learning_rate": 9.55842018650576e-05, "loss": 0.4089, "step": 166 }, { "epoch": 0.13713816464791623, "grad_norm": 0.08381053060293198, "learning_rate": 9.555677454744927e-05, "loss": 0.3624, "step": 167 }, { "epoch": 0.13795935126257441, "grad_norm": 0.07476504147052765, "learning_rate": 9.552934722984092e-05, "loss": 0.3906, "step": 168 }, { "epoch": 0.1387805378772326, "grad_norm": 0.0901239663362503, "learning_rate": 9.550191991223259e-05, "loss": 0.3378, "step": 169 }, { "epoch": 0.1396017244918908, "grad_norm": 0.0813356265425682, "learning_rate": 9.547449259462426e-05, "loss": 0.3627, "step": 170 }, { "epoch": 0.14042291110654898, "grad_norm": 0.14319093525409698, "learning_rate": 9.544706527701591e-05, "loss": 0.3512, "step": 171 }, { "epoch": 0.14124409772120713, "grad_norm": 0.13329866528511047, "learning_rate": 9.541963795940757e-05, "loss": 0.3809, "step": 172 }, { "epoch": 0.14206528433586532, "grad_norm": 0.0815596953034401, "learning_rate": 9.539221064179923e-05, "loss": 0.3444, "step": 173 }, { "epoch": 0.1428864709505235, "grad_norm": 0.08646956831216812, "learning_rate": 9.53647833241909e-05, "loss": 0.3699, "step": 174 }, { "epoch": 0.1437076575651817, "grad_norm": 0.09374339133501053, "learning_rate": 9.533735600658256e-05, "loss": 0.3569, "step": 175 }, { "epoch": 0.14452884417983985, "grad_norm": 0.0834718644618988, "learning_rate": 9.530992868897422e-05, "loss": 0.3708, "step": 176 }, { "epoch": 0.14452884417983985, "eval_runtime": 493.9539, "eval_samples_per_second": 0.399, "eval_steps_per_second": 0.2, "step": 176 }, { "epoch": 0.14535003079449804, "grad_norm": 0.08648290485143661, "learning_rate": 9.528250137136588e-05, "loss": 0.379, "step": 177 }, { "epoch": 0.14617121740915623, "grad_norm": 0.08577203750610352, "learning_rate": 9.525507405375755e-05, "loss": 0.4368, "step": 178 }, { "epoch": 0.14699240402381442, "grad_norm": 0.1023576483130455, "learning_rate": 9.522764673614921e-05, "loss": 0.3553, "step": 179 }, { "epoch": 0.1478135906384726, "grad_norm": 0.08062634617090225, "learning_rate": 9.520021941854087e-05, "loss": 0.3616, "step": 180 }, { "epoch": 0.14863477725313076, "grad_norm": 0.07487751543521881, "learning_rate": 9.517279210093252e-05, "loss": 0.4413, "step": 181 }, { "epoch": 0.14945596386778895, "grad_norm": 0.07332492619752884, "learning_rate": 9.514536478332419e-05, "loss": 0.3166, "step": 182 }, { "epoch": 0.15027715048244714, "grad_norm": 0.08658608049154282, "learning_rate": 9.511793746571586e-05, "loss": 0.332, "step": 183 }, { "epoch": 0.15109833709710532, "grad_norm": 0.10461894422769547, "learning_rate": 9.509051014810751e-05, "loss": 0.3653, "step": 184 }, { "epoch": 0.1519195237117635, "grad_norm": 0.19619494676589966, "learning_rate": 9.506308283049918e-05, "loss": 0.4224, "step": 185 }, { "epoch": 0.15274071032642167, "grad_norm": 0.08313202857971191, "learning_rate": 9.503565551289084e-05, "loss": 0.3512, "step": 186 }, { "epoch": 0.15356189694107986, "grad_norm": 0.08686342090368271, "learning_rate": 9.50082281952825e-05, "loss": 0.3352, "step": 187 }, { "epoch": 0.15438308355573804, "grad_norm": 0.09020522236824036, "learning_rate": 9.498080087767417e-05, "loss": 0.4138, "step": 188 }, { "epoch": 0.15520427017039623, "grad_norm": 0.0799839124083519, "learning_rate": 9.495337356006583e-05, "loss": 0.3436, "step": 189 }, { "epoch": 0.15602545678505442, "grad_norm": 0.08851379156112671, "learning_rate": 9.49259462424575e-05, "loss": 0.3933, "step": 190 }, { "epoch": 0.15684664339971258, "grad_norm": 0.08521082252264023, "learning_rate": 9.489851892484916e-05, "loss": 0.3667, "step": 191 }, { "epoch": 0.15766783001437076, "grad_norm": 0.09019312262535095, "learning_rate": 9.487109160724082e-05, "loss": 0.3556, "step": 192 }, { "epoch": 0.15848901662902895, "grad_norm": 0.08162654936313629, "learning_rate": 9.484366428963248e-05, "loss": 0.3864, "step": 193 }, { "epoch": 0.15931020324368714, "grad_norm": 0.08963490277528763, "learning_rate": 9.481623697202414e-05, "loss": 0.391, "step": 194 }, { "epoch": 0.1601313898583453, "grad_norm": 0.10057719051837921, "learning_rate": 9.47888096544158e-05, "loss": 0.3599, "step": 195 }, { "epoch": 0.16095257647300348, "grad_norm": 0.1694149523973465, "learning_rate": 9.476138233680747e-05, "loss": 0.3612, "step": 196 }, { "epoch": 0.16177376308766167, "grad_norm": 0.08255323767662048, "learning_rate": 9.473395501919913e-05, "loss": 0.3441, "step": 197 }, { "epoch": 0.16259494970231986, "grad_norm": 0.09228333085775375, "learning_rate": 9.470652770159078e-05, "loss": 0.3432, "step": 198 }, { "epoch": 0.16341613631697804, "grad_norm": 0.11736617982387543, "learning_rate": 9.467910038398245e-05, "loss": 0.3408, "step": 199 }, { "epoch": 0.1642373229316362, "grad_norm": 0.16636626422405243, "learning_rate": 9.465167306637412e-05, "loss": 0.3639, "step": 200 }, { "epoch": 0.1650585095462944, "grad_norm": 0.0912085771560669, "learning_rate": 9.462424574876577e-05, "loss": 0.4055, "step": 201 }, { "epoch": 0.16587969616095258, "grad_norm": 0.08452475070953369, "learning_rate": 9.459681843115744e-05, "loss": 0.3497, "step": 202 }, { "epoch": 0.16670088277561076, "grad_norm": 0.09225429594516754, "learning_rate": 9.45693911135491e-05, "loss": 0.3765, "step": 203 }, { "epoch": 0.16752206939026895, "grad_norm": 0.08500406891107559, "learning_rate": 9.454196379594076e-05, "loss": 0.3622, "step": 204 }, { "epoch": 0.1683432560049271, "grad_norm": 0.08277002722024918, "learning_rate": 9.451453647833243e-05, "loss": 0.3596, "step": 205 }, { "epoch": 0.1691644426195853, "grad_norm": 0.0855122059583664, "learning_rate": 9.448710916072408e-05, "loss": 0.3495, "step": 206 }, { "epoch": 0.16998562923424348, "grad_norm": 0.11409081518650055, "learning_rate": 9.445968184311574e-05, "loss": 0.3123, "step": 207 }, { "epoch": 0.17080681584890167, "grad_norm": 0.08898866921663284, "learning_rate": 9.443225452550741e-05, "loss": 0.3383, "step": 208 }, { "epoch": 0.17162800246355983, "grad_norm": 0.07967101782560349, "learning_rate": 9.440482720789908e-05, "loss": 0.3927, "step": 209 }, { "epoch": 0.17244918907821802, "grad_norm": 0.08267655968666077, "learning_rate": 9.437739989029073e-05, "loss": 0.3503, "step": 210 }, { "epoch": 0.1732703756928762, "grad_norm": 0.1024966835975647, "learning_rate": 9.43499725726824e-05, "loss": 0.343, "step": 211 }, { "epoch": 0.1740915623075344, "grad_norm": 0.11825034767389297, "learning_rate": 9.432254525507405e-05, "loss": 0.3694, "step": 212 }, { "epoch": 0.17491274892219258, "grad_norm": 0.08225111663341522, "learning_rate": 9.429511793746572e-05, "loss": 0.3625, "step": 213 }, { "epoch": 0.17573393553685074, "grad_norm": 0.082975834608078, "learning_rate": 9.426769061985739e-05, "loss": 0.3692, "step": 214 }, { "epoch": 0.17655512215150893, "grad_norm": 0.09659875184297562, "learning_rate": 9.424026330224904e-05, "loss": 0.3857, "step": 215 }, { "epoch": 0.1773763087661671, "grad_norm": 0.08381886035203934, "learning_rate": 9.42128359846407e-05, "loss": 0.3658, "step": 216 }, { "epoch": 0.1781974953808253, "grad_norm": 0.08097488433122635, "learning_rate": 9.418540866703238e-05, "loss": 0.3556, "step": 217 }, { "epoch": 0.1790186819954835, "grad_norm": 0.08661879599094391, "learning_rate": 9.415798134942403e-05, "loss": 0.3495, "step": 218 }, { "epoch": 0.17983986861014165, "grad_norm": 0.09621778875589371, "learning_rate": 9.413055403181569e-05, "loss": 0.3488, "step": 219 }, { "epoch": 0.18066105522479983, "grad_norm": 0.0867924615740776, "learning_rate": 9.410312671420735e-05, "loss": 0.3496, "step": 220 }, { "epoch": 0.18148224183945802, "grad_norm": 0.09928230196237564, "learning_rate": 9.407569939659901e-05, "loss": 0.411, "step": 221 }, { "epoch": 0.1823034284541162, "grad_norm": 0.08545473217964172, "learning_rate": 9.404827207899069e-05, "loss": 0.3707, "step": 222 }, { "epoch": 0.18312461506877437, "grad_norm": 0.1317296177148819, "learning_rate": 9.402084476138234e-05, "loss": 0.4021, "step": 223 }, { "epoch": 0.18394580168343255, "grad_norm": 0.09009065479040146, "learning_rate": 9.3993417443774e-05, "loss": 0.3451, "step": 224 }, { "epoch": 0.18476698829809074, "grad_norm": 0.08332253247499466, "learning_rate": 9.396599012616567e-05, "loss": 0.3457, "step": 225 }, { "epoch": 0.18558817491274893, "grad_norm": 0.08737312257289886, "learning_rate": 9.393856280855733e-05, "loss": 0.3721, "step": 226 }, { "epoch": 0.1864093615274071, "grad_norm": 0.07472239434719086, "learning_rate": 9.391113549094899e-05, "loss": 0.3542, "step": 227 }, { "epoch": 0.18723054814206527, "grad_norm": 0.08319877088069916, "learning_rate": 9.388370817334066e-05, "loss": 0.361, "step": 228 }, { "epoch": 0.18805173475672346, "grad_norm": 0.08192326873540878, "learning_rate": 9.385628085573231e-05, "loss": 0.344, "step": 229 }, { "epoch": 0.18887292137138165, "grad_norm": 0.11502642929553986, "learning_rate": 9.382885353812398e-05, "loss": 0.363, "step": 230 }, { "epoch": 0.18969410798603983, "grad_norm": 0.08567750453948975, "learning_rate": 9.380142622051565e-05, "loss": 0.3565, "step": 231 }, { "epoch": 0.19051529460069802, "grad_norm": 0.08059141039848328, "learning_rate": 9.37739989029073e-05, "loss": 0.3722, "step": 232 }, { "epoch": 0.19133648121535618, "grad_norm": 0.07670270651578903, "learning_rate": 9.374657158529896e-05, "loss": 0.3493, "step": 233 }, { "epoch": 0.19215766783001437, "grad_norm": 0.08022642135620117, "learning_rate": 9.371914426769062e-05, "loss": 0.4142, "step": 234 }, { "epoch": 0.19297885444467255, "grad_norm": 0.08015397936105728, "learning_rate": 9.369171695008229e-05, "loss": 0.3442, "step": 235 }, { "epoch": 0.19380004105933074, "grad_norm": 0.08329442143440247, "learning_rate": 9.366428963247395e-05, "loss": 0.3587, "step": 236 }, { "epoch": 0.1946212276739889, "grad_norm": 0.08311276882886887, "learning_rate": 9.363686231486561e-05, "loss": 0.3792, "step": 237 }, { "epoch": 0.1954424142886471, "grad_norm": 0.1332862675189972, "learning_rate": 9.360943499725727e-05, "loss": 0.3451, "step": 238 }, { "epoch": 0.19626360090330527, "grad_norm": 0.081804558634758, "learning_rate": 9.358200767964894e-05, "loss": 0.4212, "step": 239 }, { "epoch": 0.19708478751796346, "grad_norm": 0.08397019654512405, "learning_rate": 9.35545803620406e-05, "loss": 0.3482, "step": 240 }, { "epoch": 0.19790597413262165, "grad_norm": 0.09495637565851212, "learning_rate": 9.352715304443226e-05, "loss": 0.3432, "step": 241 }, { "epoch": 0.1987271607472798, "grad_norm": 0.09187504649162292, "learning_rate": 9.349972572682391e-05, "loss": 0.3938, "step": 242 }, { "epoch": 0.199548347361938, "grad_norm": 0.100834921002388, "learning_rate": 9.347229840921558e-05, "loss": 0.3567, "step": 243 }, { "epoch": 0.20036953397659618, "grad_norm": 0.09292273223400116, "learning_rate": 9.344487109160725e-05, "loss": 0.3802, "step": 244 }, { "epoch": 0.20119072059125437, "grad_norm": 0.07785986363887787, "learning_rate": 9.34174437739989e-05, "loss": 0.3832, "step": 245 }, { "epoch": 0.20201190720591256, "grad_norm": 0.081189826130867, "learning_rate": 9.339001645639057e-05, "loss": 0.3448, "step": 246 }, { "epoch": 0.20283309382057071, "grad_norm": 0.08209879696369171, "learning_rate": 9.336258913878222e-05, "loss": 0.3602, "step": 247 }, { "epoch": 0.2036542804352289, "grad_norm": 0.0813421905040741, "learning_rate": 9.333516182117389e-05, "loss": 0.3477, "step": 248 }, { "epoch": 0.2044754670498871, "grad_norm": 0.08188773691654205, "learning_rate": 9.330773450356556e-05, "loss": 0.3466, "step": 249 }, { "epoch": 0.20529665366454528, "grad_norm": 0.09966633468866348, "learning_rate": 9.328030718595721e-05, "loss": 0.3776, "step": 250 }, { "epoch": 0.20611784027920346, "grad_norm": 0.08154954016208649, "learning_rate": 9.325287986834887e-05, "loss": 0.3727, "step": 251 }, { "epoch": 0.20693902689386162, "grad_norm": 0.07942931354045868, "learning_rate": 9.322545255074055e-05, "loss": 0.3468, "step": 252 }, { "epoch": 0.2077602135085198, "grad_norm": 0.07944433391094208, "learning_rate": 9.31980252331322e-05, "loss": 0.3512, "step": 253 }, { "epoch": 0.208581400123178, "grad_norm": 0.08637971431016922, "learning_rate": 9.317059791552387e-05, "loss": 0.3397, "step": 254 }, { "epoch": 0.20940258673783618, "grad_norm": 0.08189195394515991, "learning_rate": 9.314317059791553e-05, "loss": 0.4105, "step": 255 }, { "epoch": 0.21022377335249434, "grad_norm": 0.08190836012363434, "learning_rate": 9.31157432803072e-05, "loss": 0.346, "step": 256 }, { "epoch": 0.21104495996715253, "grad_norm": 0.08331865072250366, "learning_rate": 9.308831596269886e-05, "loss": 0.3655, "step": 257 }, { "epoch": 0.21186614658181072, "grad_norm": 0.0822620540857315, "learning_rate": 9.306088864509052e-05, "loss": 0.372, "step": 258 }, { "epoch": 0.2126873331964689, "grad_norm": 0.08011777698993683, "learning_rate": 9.303346132748217e-05, "loss": 0.3304, "step": 259 }, { "epoch": 0.2135085198111271, "grad_norm": 0.08395062386989594, "learning_rate": 9.300603400987384e-05, "loss": 0.3903, "step": 260 }, { "epoch": 0.21432970642578525, "grad_norm": 0.08101452887058258, "learning_rate": 9.297860669226551e-05, "loss": 0.3579, "step": 261 }, { "epoch": 0.21515089304044344, "grad_norm": 0.11504925042390823, "learning_rate": 9.295117937465716e-05, "loss": 0.3723, "step": 262 }, { "epoch": 0.21597207965510162, "grad_norm": 0.0704338401556015, "learning_rate": 9.292375205704883e-05, "loss": 0.3341, "step": 263 }, { "epoch": 0.2167932662697598, "grad_norm": 0.07944470643997192, "learning_rate": 9.289632473944048e-05, "loss": 0.3273, "step": 264 }, { "epoch": 0.2167932662697598, "eval_runtime": 507.3366, "eval_samples_per_second": 0.388, "eval_steps_per_second": 0.195, "step": 264 }, { "epoch": 0.217614452884418, "grad_norm": 0.0778859481215477, "learning_rate": 9.286889742183215e-05, "loss": 0.3707, "step": 265 }, { "epoch": 0.21843563949907616, "grad_norm": 0.07564377039670944, "learning_rate": 9.284147010422382e-05, "loss": 0.334, "step": 266 }, { "epoch": 0.21925682611373434, "grad_norm": 0.07744503021240234, "learning_rate": 9.281404278661547e-05, "loss": 0.3521, "step": 267 }, { "epoch": 0.22007801272839253, "grad_norm": 0.08139525353908539, "learning_rate": 9.278661546900713e-05, "loss": 0.4069, "step": 268 }, { "epoch": 0.22089919934305072, "grad_norm": 0.08162204176187515, "learning_rate": 9.27591881513988e-05, "loss": 0.3292, "step": 269 }, { "epoch": 0.22172038595770888, "grad_norm": 0.07798879593610764, "learning_rate": 9.273176083379046e-05, "loss": 0.3441, "step": 270 }, { "epoch": 0.22254157257236706, "grad_norm": 0.07629597932100296, "learning_rate": 9.270433351618212e-05, "loss": 0.3989, "step": 271 }, { "epoch": 0.22336275918702525, "grad_norm": 0.07695835083723068, "learning_rate": 9.267690619857379e-05, "loss": 0.3506, "step": 272 }, { "epoch": 0.22418394580168344, "grad_norm": 0.07938527315855026, "learning_rate": 9.264947888096544e-05, "loss": 0.327, "step": 273 }, { "epoch": 0.22500513241634162, "grad_norm": 0.08125253766775131, "learning_rate": 9.262205156335711e-05, "loss": 0.3749, "step": 274 }, { "epoch": 0.22582631903099978, "grad_norm": 0.07831385731697083, "learning_rate": 9.259462424574878e-05, "loss": 0.3477, "step": 275 }, { "epoch": 0.22664750564565797, "grad_norm": 0.08408527821302414, "learning_rate": 9.256719692814043e-05, "loss": 0.3488, "step": 276 }, { "epoch": 0.22746869226031616, "grad_norm": 0.0800737589597702, "learning_rate": 9.253976961053208e-05, "loss": 0.3607, "step": 277 }, { "epoch": 0.22828987887497434, "grad_norm": 0.07969113439321518, "learning_rate": 9.251234229292375e-05, "loss": 0.38, "step": 278 }, { "epoch": 0.22911106548963253, "grad_norm": 0.08066916465759277, "learning_rate": 9.248491497531542e-05, "loss": 0.3469, "step": 279 }, { "epoch": 0.2299322521042907, "grad_norm": 0.07686860859394073, "learning_rate": 9.245748765770708e-05, "loss": 0.3311, "step": 280 }, { "epoch": 0.23075343871894888, "grad_norm": 0.13063722848892212, "learning_rate": 9.243006034009874e-05, "loss": 0.3091, "step": 281 }, { "epoch": 0.23157462533360706, "grad_norm": 0.08465570211410522, "learning_rate": 9.24026330224904e-05, "loss": 0.3548, "step": 282 }, { "epoch": 0.23239581194826525, "grad_norm": 0.07824988663196564, "learning_rate": 9.237520570488207e-05, "loss": 0.3485, "step": 283 }, { "epoch": 0.2332169985629234, "grad_norm": 0.087236687541008, "learning_rate": 9.234777838727373e-05, "loss": 0.3564, "step": 284 }, { "epoch": 0.2340381851775816, "grad_norm": 0.09531310200691223, "learning_rate": 9.232035106966539e-05, "loss": 0.3763, "step": 285 }, { "epoch": 0.23485937179223978, "grad_norm": 0.0783536285161972, "learning_rate": 9.229292375205704e-05, "loss": 0.3479, "step": 286 }, { "epoch": 0.23568055840689797, "grad_norm": 0.08209957927465439, "learning_rate": 9.226549643444872e-05, "loss": 0.3563, "step": 287 }, { "epoch": 0.23650174502155616, "grad_norm": 0.08195500820875168, "learning_rate": 9.223806911684038e-05, "loss": 0.3879, "step": 288 }, { "epoch": 0.23732293163621432, "grad_norm": 0.0796385332942009, "learning_rate": 9.221064179923205e-05, "loss": 0.3787, "step": 289 }, { "epoch": 0.2381441182508725, "grad_norm": 0.08523210138082504, "learning_rate": 9.21832144816237e-05, "loss": 0.3608, "step": 290 }, { "epoch": 0.2389653048655307, "grad_norm": 0.08573302626609802, "learning_rate": 9.215578716401537e-05, "loss": 0.3366, "step": 291 }, { "epoch": 0.23978649148018888, "grad_norm": 0.08111650496721268, "learning_rate": 9.212835984640704e-05, "loss": 0.3331, "step": 292 }, { "epoch": 0.24060767809484707, "grad_norm": 0.07877853512763977, "learning_rate": 9.210093252879869e-05, "loss": 0.3346, "step": 293 }, { "epoch": 0.24142886470950523, "grad_norm": 0.0814940333366394, "learning_rate": 9.207350521119034e-05, "loss": 0.3701, "step": 294 }, { "epoch": 0.2422500513241634, "grad_norm": 0.0843539908528328, "learning_rate": 9.204607789358201e-05, "loss": 0.3339, "step": 295 }, { "epoch": 0.2430712379388216, "grad_norm": 0.0875789001584053, "learning_rate": 9.201865057597368e-05, "loss": 0.362, "step": 296 }, { "epoch": 0.2438924245534798, "grad_norm": 0.08261071145534515, "learning_rate": 9.199122325836533e-05, "loss": 0.3578, "step": 297 }, { "epoch": 0.24471361116813795, "grad_norm": 0.09160041064023972, "learning_rate": 9.1963795940757e-05, "loss": 0.387, "step": 298 }, { "epoch": 0.24553479778279613, "grad_norm": 0.08152728527784348, "learning_rate": 9.193636862314866e-05, "loss": 0.3282, "step": 299 }, { "epoch": 0.24635598439745432, "grad_norm": 0.079656682908535, "learning_rate": 9.190894130554032e-05, "loss": 0.3238, "step": 300 }, { "epoch": 0.2471771710121125, "grad_norm": 0.0797642394900322, "learning_rate": 9.188151398793199e-05, "loss": 0.3506, "step": 301 }, { "epoch": 0.2479983576267707, "grad_norm": 0.08688945323228836, "learning_rate": 9.185408667032365e-05, "loss": 0.4177, "step": 302 }, { "epoch": 0.24881954424142885, "grad_norm": 0.106851726770401, "learning_rate": 9.18266593527153e-05, "loss": 0.3527, "step": 303 }, { "epoch": 0.24964073085608704, "grad_norm": 0.08586329221725464, "learning_rate": 9.179923203510697e-05, "loss": 0.3763, "step": 304 }, { "epoch": 0.25046191747074525, "grad_norm": 0.1141560897231102, "learning_rate": 9.177180471749864e-05, "loss": 0.3451, "step": 305 }, { "epoch": 0.2512831040854034, "grad_norm": 0.08728976547718048, "learning_rate": 9.174437739989029e-05, "loss": 0.3444, "step": 306 }, { "epoch": 0.2521042907000616, "grad_norm": 0.07848083972930908, "learning_rate": 9.171695008228196e-05, "loss": 0.344, "step": 307 }, { "epoch": 0.25292547731471976, "grad_norm": 0.08140011131763458, "learning_rate": 9.168952276467361e-05, "loss": 0.3338, "step": 308 }, { "epoch": 0.25374666392937795, "grad_norm": 0.07934946566820145, "learning_rate": 9.166209544706528e-05, "loss": 0.3869, "step": 309 }, { "epoch": 0.25456785054403613, "grad_norm": 0.08028525859117508, "learning_rate": 9.163466812945695e-05, "loss": 0.3141, "step": 310 }, { "epoch": 0.2553890371586943, "grad_norm": 0.08121496438980103, "learning_rate": 9.16072408118486e-05, "loss": 0.4098, "step": 311 }, { "epoch": 0.2562102237733525, "grad_norm": 0.09907019138336182, "learning_rate": 9.157981349424026e-05, "loss": 0.3594, "step": 312 }, { "epoch": 0.2570314103880107, "grad_norm": 0.10270131379365921, "learning_rate": 9.155238617663193e-05, "loss": 0.3713, "step": 313 }, { "epoch": 0.2578525970026689, "grad_norm": 0.08354981243610382, "learning_rate": 9.15249588590236e-05, "loss": 0.3943, "step": 314 }, { "epoch": 0.258673783617327, "grad_norm": 0.09153895825147629, "learning_rate": 9.149753154141525e-05, "loss": 0.3499, "step": 315 }, { "epoch": 0.2594949702319852, "grad_norm": 0.07772937417030334, "learning_rate": 9.147010422380692e-05, "loss": 0.3323, "step": 316 }, { "epoch": 0.2603161568466434, "grad_norm": 0.08953258395195007, "learning_rate": 9.144267690619857e-05, "loss": 0.3684, "step": 317 }, { "epoch": 0.2611373434613016, "grad_norm": 0.08204226195812225, "learning_rate": 9.141524958859024e-05, "loss": 0.3387, "step": 318 }, { "epoch": 0.26195853007595976, "grad_norm": 0.07604125887155533, "learning_rate": 9.13878222709819e-05, "loss": 0.3104, "step": 319 }, { "epoch": 0.26277971669061795, "grad_norm": 0.07614021003246307, "learning_rate": 9.136039495337356e-05, "loss": 0.329, "step": 320 }, { "epoch": 0.26360090330527614, "grad_norm": 0.19391578435897827, "learning_rate": 9.133296763576523e-05, "loss": 0.349, "step": 321 }, { "epoch": 0.2644220899199343, "grad_norm": 0.07927168160676956, "learning_rate": 9.13055403181569e-05, "loss": 0.3495, "step": 322 }, { "epoch": 0.2652432765345925, "grad_norm": 0.08091975003480911, "learning_rate": 9.127811300054855e-05, "loss": 0.3563, "step": 323 }, { "epoch": 0.26606446314925064, "grad_norm": 0.07865800708532333, "learning_rate": 9.125068568294022e-05, "loss": 0.36, "step": 324 }, { "epoch": 0.2668856497639088, "grad_norm": 0.08946891874074936, "learning_rate": 9.122325836533187e-05, "loss": 0.3434, "step": 325 }, { "epoch": 0.267706836378567, "grad_norm": 0.07987753301858902, "learning_rate": 9.119583104772354e-05, "loss": 0.3593, "step": 326 }, { "epoch": 0.2685280229932252, "grad_norm": 0.07894265651702881, "learning_rate": 9.116840373011521e-05, "loss": 0.3746, "step": 327 }, { "epoch": 0.2693492096078834, "grad_norm": 0.08121046423912048, "learning_rate": 9.114097641250686e-05, "loss": 0.357, "step": 328 }, { "epoch": 0.2701703962225416, "grad_norm": 0.08320184051990509, "learning_rate": 9.111354909489852e-05, "loss": 0.3783, "step": 329 }, { "epoch": 0.27099158283719976, "grad_norm": 0.10911049693822861, "learning_rate": 9.108612177729018e-05, "loss": 0.3699, "step": 330 }, { "epoch": 0.27181276945185795, "grad_norm": 0.09739838540554047, "learning_rate": 9.105869445968185e-05, "loss": 0.3524, "step": 331 }, { "epoch": 0.27263395606651614, "grad_norm": 0.0874953344464302, "learning_rate": 9.103126714207351e-05, "loss": 0.3807, "step": 332 }, { "epoch": 0.2734551426811743, "grad_norm": 0.09758396446704865, "learning_rate": 9.100383982446518e-05, "loss": 0.3876, "step": 333 }, { "epoch": 0.27427632929583246, "grad_norm": 0.08628170937299728, "learning_rate": 9.097641250685683e-05, "loss": 0.3632, "step": 334 }, { "epoch": 0.27509751591049064, "grad_norm": 0.08430638164281845, "learning_rate": 9.09489851892485e-05, "loss": 0.3448, "step": 335 }, { "epoch": 0.27591870252514883, "grad_norm": 0.0725574865937233, "learning_rate": 9.092155787164017e-05, "loss": 0.2914, "step": 336 }, { "epoch": 0.276739889139807, "grad_norm": 0.08421092480421066, "learning_rate": 9.089413055403182e-05, "loss": 0.3676, "step": 337 }, { "epoch": 0.2775610757544652, "grad_norm": 0.08702068030834198, "learning_rate": 9.086670323642347e-05, "loss": 0.345, "step": 338 }, { "epoch": 0.2783822623691234, "grad_norm": 0.08053667843341827, "learning_rate": 9.083927591881514e-05, "loss": 0.324, "step": 339 }, { "epoch": 0.2792034489837816, "grad_norm": 0.08294267952442169, "learning_rate": 9.081184860120681e-05, "loss": 0.348, "step": 340 }, { "epoch": 0.28002463559843976, "grad_norm": 0.18557026982307434, "learning_rate": 9.078442128359846e-05, "loss": 0.3483, "step": 341 }, { "epoch": 0.28084582221309795, "grad_norm": 0.08436847478151321, "learning_rate": 9.075699396599013e-05, "loss": 0.3641, "step": 342 }, { "epoch": 0.2816670088277561, "grad_norm": 0.08285505324602127, "learning_rate": 9.072956664838179e-05, "loss": 0.3641, "step": 343 }, { "epoch": 0.28248819544241427, "grad_norm": 0.08681569993495941, "learning_rate": 9.070213933077345e-05, "loss": 0.342, "step": 344 }, { "epoch": 0.28330938205707246, "grad_norm": 0.08668892085552216, "learning_rate": 9.067471201316512e-05, "loss": 0.3514, "step": 345 }, { "epoch": 0.28413056867173064, "grad_norm": 0.08393285423517227, "learning_rate": 9.064728469555678e-05, "loss": 0.3531, "step": 346 }, { "epoch": 0.28495175528638883, "grad_norm": 0.07783684879541397, "learning_rate": 9.061985737794843e-05, "loss": 0.3792, "step": 347 }, { "epoch": 0.285772941901047, "grad_norm": 0.08259117603302002, "learning_rate": 9.059243006034011e-05, "loss": 0.375, "step": 348 }, { "epoch": 0.2865941285157052, "grad_norm": 0.0969453975558281, "learning_rate": 9.056500274273177e-05, "loss": 0.3776, "step": 349 }, { "epoch": 0.2874153151303634, "grad_norm": 0.08154502511024475, "learning_rate": 9.053757542512342e-05, "loss": 0.3577, "step": 350 }, { "epoch": 0.2882365017450216, "grad_norm": 0.09322196245193481, "learning_rate": 9.051014810751509e-05, "loss": 0.3774, "step": 351 }, { "epoch": 0.2890576883596797, "grad_norm": 0.0804852545261383, "learning_rate": 9.048272078990674e-05, "loss": 0.333, "step": 352 }, { "epoch": 0.2890576883596797, "eval_runtime": 493.668, "eval_samples_per_second": 0.399, "eval_steps_per_second": 0.201, "step": 352 }, { "epoch": 0.2898788749743379, "grad_norm": 0.08082845062017441, "learning_rate": 9.045529347229841e-05, "loss": 0.3111, "step": 353 }, { "epoch": 0.2907000615889961, "grad_norm": 0.08396881073713303, "learning_rate": 9.042786615469008e-05, "loss": 0.3429, "step": 354 }, { "epoch": 0.29152124820365427, "grad_norm": 0.08037351071834564, "learning_rate": 9.040043883708173e-05, "loss": 0.349, "step": 355 }, { "epoch": 0.29234243481831246, "grad_norm": 0.10421521961688995, "learning_rate": 9.03730115194734e-05, "loss": 0.3228, "step": 356 }, { "epoch": 0.29316362143297064, "grad_norm": 0.08268436044454575, "learning_rate": 9.034558420186507e-05, "loss": 0.3414, "step": 357 }, { "epoch": 0.29398480804762883, "grad_norm": 0.08203061670064926, "learning_rate": 9.031815688425672e-05, "loss": 0.3596, "step": 358 }, { "epoch": 0.294805994662287, "grad_norm": 0.07778481394052505, "learning_rate": 9.029072956664839e-05, "loss": 0.3337, "step": 359 }, { "epoch": 0.2956271812769452, "grad_norm": 0.14194349944591522, "learning_rate": 9.026330224904005e-05, "loss": 0.3613, "step": 360 }, { "epoch": 0.2964483678916034, "grad_norm": 0.09854190051555634, "learning_rate": 9.023587493143171e-05, "loss": 0.3862, "step": 361 }, { "epoch": 0.2972695545062615, "grad_norm": 0.07878783345222473, "learning_rate": 9.020844761382338e-05, "loss": 0.3294, "step": 362 }, { "epoch": 0.2980907411209197, "grad_norm": 0.07692375034093857, "learning_rate": 9.018102029621504e-05, "loss": 0.3083, "step": 363 }, { "epoch": 0.2989119277355779, "grad_norm": 0.0772586315870285, "learning_rate": 9.015359297860669e-05, "loss": 0.3086, "step": 364 }, { "epoch": 0.2997331143502361, "grad_norm": 0.0739215686917305, "learning_rate": 9.012616566099836e-05, "loss": 0.3103, "step": 365 }, { "epoch": 0.30055430096489427, "grad_norm": 0.08434160053730011, "learning_rate": 9.009873834339003e-05, "loss": 0.3485, "step": 366 }, { "epoch": 0.30137548757955246, "grad_norm": 0.08038529008626938, "learning_rate": 9.007131102578168e-05, "loss": 0.3421, "step": 367 }, { "epoch": 0.30219667419421065, "grad_norm": 0.08303964138031006, "learning_rate": 9.004388370817335e-05, "loss": 0.3349, "step": 368 }, { "epoch": 0.30301786080886883, "grad_norm": 0.08672379702329636, "learning_rate": 9.0016456390565e-05, "loss": 0.3228, "step": 369 }, { "epoch": 0.303839047423527, "grad_norm": 0.08417553454637527, "learning_rate": 8.998902907295667e-05, "loss": 0.3312, "step": 370 }, { "epoch": 0.30466023403818515, "grad_norm": 0.0842244029045105, "learning_rate": 8.996160175534834e-05, "loss": 0.3545, "step": 371 }, { "epoch": 0.30548142065284334, "grad_norm": 0.08479683846235275, "learning_rate": 8.993417443773999e-05, "loss": 0.3446, "step": 372 }, { "epoch": 0.3063026072675015, "grad_norm": 0.07726123929023743, "learning_rate": 8.990674712013165e-05, "loss": 0.3116, "step": 373 }, { "epoch": 0.3071237938821597, "grad_norm": 0.08586291968822479, "learning_rate": 8.987931980252331e-05, "loss": 0.3617, "step": 374 }, { "epoch": 0.3079449804968179, "grad_norm": 0.07475403696298599, "learning_rate": 8.985189248491498e-05, "loss": 0.3481, "step": 375 }, { "epoch": 0.3087661671114761, "grad_norm": 0.08379320055246353, "learning_rate": 8.982446516730664e-05, "loss": 0.3434, "step": 376 }, { "epoch": 0.3095873537261343, "grad_norm": 0.07624776661396027, "learning_rate": 8.97970378496983e-05, "loss": 0.3262, "step": 377 }, { "epoch": 0.31040854034079246, "grad_norm": 0.07878948748111725, "learning_rate": 8.976961053208996e-05, "loss": 0.3366, "step": 378 }, { "epoch": 0.31122972695545065, "grad_norm": 0.08924887329339981, "learning_rate": 8.974218321448163e-05, "loss": 0.3175, "step": 379 }, { "epoch": 0.31205091357010883, "grad_norm": 0.08927389234304428, "learning_rate": 8.97147558968733e-05, "loss": 0.3338, "step": 380 }, { "epoch": 0.31287210018476697, "grad_norm": 0.08250463008880615, "learning_rate": 8.968732857926495e-05, "loss": 0.3314, "step": 381 }, { "epoch": 0.31369328679942515, "grad_norm": 0.0977792888879776, "learning_rate": 8.96599012616566e-05, "loss": 0.3222, "step": 382 }, { "epoch": 0.31451447341408334, "grad_norm": 0.08197218924760818, "learning_rate": 8.963247394404828e-05, "loss": 0.3288, "step": 383 }, { "epoch": 0.3153356600287415, "grad_norm": 0.12502075731754303, "learning_rate": 8.960504662643994e-05, "loss": 0.3053, "step": 384 }, { "epoch": 0.3161568466433997, "grad_norm": 0.08058564364910126, "learning_rate": 8.95776193088316e-05, "loss": 0.3292, "step": 385 }, { "epoch": 0.3169780332580579, "grad_norm": 0.08365239202976227, "learning_rate": 8.955019199122326e-05, "loss": 0.3747, "step": 386 }, { "epoch": 0.3177992198727161, "grad_norm": 0.07539816945791245, "learning_rate": 8.952276467361493e-05, "loss": 0.3179, "step": 387 }, { "epoch": 0.3186204064873743, "grad_norm": 0.08136752992868423, "learning_rate": 8.94953373560066e-05, "loss": 0.3647, "step": 388 }, { "epoch": 0.31944159310203246, "grad_norm": 0.09546865522861481, "learning_rate": 8.946791003839825e-05, "loss": 0.4056, "step": 389 }, { "epoch": 0.3202627797166906, "grad_norm": 0.08654627203941345, "learning_rate": 8.94404827207899e-05, "loss": 0.3853, "step": 390 }, { "epoch": 0.3210839663313488, "grad_norm": 0.0822821855545044, "learning_rate": 8.941305540318157e-05, "loss": 0.327, "step": 391 }, { "epoch": 0.32190515294600697, "grad_norm": 0.08066459000110626, "learning_rate": 8.938562808557324e-05, "loss": 0.3585, "step": 392 }, { "epoch": 0.32272633956066515, "grad_norm": 0.08845409005880356, "learning_rate": 8.93582007679649e-05, "loss": 0.3419, "step": 393 }, { "epoch": 0.32354752617532334, "grad_norm": 0.08482253551483154, "learning_rate": 8.933077345035656e-05, "loss": 0.3263, "step": 394 }, { "epoch": 0.32436871278998153, "grad_norm": 0.0801006481051445, "learning_rate": 8.930334613274822e-05, "loss": 0.3713, "step": 395 }, { "epoch": 0.3251898994046397, "grad_norm": 0.08470006287097931, "learning_rate": 8.927591881513989e-05, "loss": 0.3631, "step": 396 }, { "epoch": 0.3260110860192979, "grad_norm": 0.08457473665475845, "learning_rate": 8.924849149753155e-05, "loss": 0.339, "step": 397 }, { "epoch": 0.3268322726339561, "grad_norm": 0.14836351573467255, "learning_rate": 8.922106417992321e-05, "loss": 0.3039, "step": 398 }, { "epoch": 0.3276534592486142, "grad_norm": 0.08391403406858444, "learning_rate": 8.919363686231486e-05, "loss": 0.3415, "step": 399 }, { "epoch": 0.3284746458632724, "grad_norm": 0.07998157292604446, "learning_rate": 8.916620954470653e-05, "loss": 0.2943, "step": 400 }, { "epoch": 0.3292958324779306, "grad_norm": 0.07682473957538605, "learning_rate": 8.91387822270982e-05, "loss": 0.3413, "step": 401 }, { "epoch": 0.3301170190925888, "grad_norm": 0.09960923343896866, "learning_rate": 8.911135490948985e-05, "loss": 0.356, "step": 402 }, { "epoch": 0.33093820570724697, "grad_norm": 0.08010439574718475, "learning_rate": 8.908392759188152e-05, "loss": 0.3408, "step": 403 }, { "epoch": 0.33175939232190516, "grad_norm": 0.0811014398932457, "learning_rate": 8.905650027427318e-05, "loss": 0.3159, "step": 404 }, { "epoch": 0.33258057893656334, "grad_norm": 0.08086768537759781, "learning_rate": 8.902907295666484e-05, "loss": 0.3145, "step": 405 }, { "epoch": 0.33340176555122153, "grad_norm": 0.0778844803571701, "learning_rate": 8.900164563905651e-05, "loss": 0.3271, "step": 406 }, { "epoch": 0.3342229521658797, "grad_norm": 0.08636100590229034, "learning_rate": 8.897421832144817e-05, "loss": 0.395, "step": 407 }, { "epoch": 0.3350441387805379, "grad_norm": 0.08080980181694031, "learning_rate": 8.894679100383982e-05, "loss": 0.3819, "step": 408 }, { "epoch": 0.33586532539519603, "grad_norm": 0.08312460780143738, "learning_rate": 8.891936368623149e-05, "loss": 0.3413, "step": 409 }, { "epoch": 0.3366865120098542, "grad_norm": 0.07980356365442276, "learning_rate": 8.889193636862316e-05, "loss": 0.3546, "step": 410 }, { "epoch": 0.3375076986245124, "grad_norm": 0.0868874043226242, "learning_rate": 8.886450905101481e-05, "loss": 0.3354, "step": 411 }, { "epoch": 0.3383288852391706, "grad_norm": 0.08026042580604553, "learning_rate": 8.883708173340648e-05, "loss": 0.355, "step": 412 }, { "epoch": 0.3391500718538288, "grad_norm": 0.0777151882648468, "learning_rate": 8.880965441579813e-05, "loss": 0.2835, "step": 413 }, { "epoch": 0.33997125846848697, "grad_norm": 0.08007453382015228, "learning_rate": 8.87822270981898e-05, "loss": 0.3042, "step": 414 }, { "epoch": 0.34079244508314516, "grad_norm": 0.08239760994911194, "learning_rate": 8.875479978058147e-05, "loss": 0.4139, "step": 415 }, { "epoch": 0.34161363169780334, "grad_norm": 0.08022072911262512, "learning_rate": 8.872737246297312e-05, "loss": 0.319, "step": 416 }, { "epoch": 0.34243481831246153, "grad_norm": 0.09205158054828644, "learning_rate": 8.869994514536478e-05, "loss": 0.3283, "step": 417 }, { "epoch": 0.34325600492711966, "grad_norm": 0.07974108308553696, "learning_rate": 8.867251782775646e-05, "loss": 0.3331, "step": 418 }, { "epoch": 0.34407719154177785, "grad_norm": 0.0810483992099762, "learning_rate": 8.864509051014811e-05, "loss": 0.4005, "step": 419 }, { "epoch": 0.34489837815643604, "grad_norm": 0.08448805660009384, "learning_rate": 8.861766319253978e-05, "loss": 0.3327, "step": 420 }, { "epoch": 0.3457195647710942, "grad_norm": 0.0865749716758728, "learning_rate": 8.859023587493143e-05, "loss": 0.4459, "step": 421 }, { "epoch": 0.3465407513857524, "grad_norm": 0.08855796605348587, "learning_rate": 8.85628085573231e-05, "loss": 0.3566, "step": 422 }, { "epoch": 0.3473619380004106, "grad_norm": 0.08421412855386734, "learning_rate": 8.853538123971477e-05, "loss": 0.3248, "step": 423 }, { "epoch": 0.3481831246150688, "grad_norm": 0.07903390377759933, "learning_rate": 8.850795392210642e-05, "loss": 0.3608, "step": 424 }, { "epoch": 0.34900431122972697, "grad_norm": 0.09979040920734406, "learning_rate": 8.848052660449808e-05, "loss": 0.3079, "step": 425 }, { "epoch": 0.34982549784438516, "grad_norm": 0.08713112026453018, "learning_rate": 8.845309928688975e-05, "loss": 0.3462, "step": 426 }, { "epoch": 0.35064668445904335, "grad_norm": 0.07836464792490005, "learning_rate": 8.842567196928141e-05, "loss": 0.3467, "step": 427 }, { "epoch": 0.3514678710737015, "grad_norm": 0.07553427666425705, "learning_rate": 8.839824465167307e-05, "loss": 0.3325, "step": 428 }, { "epoch": 0.35228905768835966, "grad_norm": 0.08221621066331863, "learning_rate": 8.837081733406474e-05, "loss": 0.3366, "step": 429 }, { "epoch": 0.35311024430301785, "grad_norm": 0.09283772855997086, "learning_rate": 8.834339001645639e-05, "loss": 0.3518, "step": 430 }, { "epoch": 0.35393143091767604, "grad_norm": 0.08067970722913742, "learning_rate": 8.831596269884806e-05, "loss": 0.3521, "step": 431 }, { "epoch": 0.3547526175323342, "grad_norm": 0.08192427456378937, "learning_rate": 8.828853538123973e-05, "loss": 0.3794, "step": 432 }, { "epoch": 0.3555738041469924, "grad_norm": 0.08332204073667526, "learning_rate": 8.826110806363138e-05, "loss": 0.3243, "step": 433 }, { "epoch": 0.3563949907616506, "grad_norm": 0.08099368214607239, "learning_rate": 8.823368074602304e-05, "loss": 0.3409, "step": 434 }, { "epoch": 0.3572161773763088, "grad_norm": 0.0801381841301918, "learning_rate": 8.82062534284147e-05, "loss": 0.3489, "step": 435 }, { "epoch": 0.358037363990967, "grad_norm": 0.0828125849366188, "learning_rate": 8.817882611080637e-05, "loss": 0.3332, "step": 436 }, { "epoch": 0.3588585506056251, "grad_norm": 0.08463995903730392, "learning_rate": 8.815139879319803e-05, "loss": 0.3532, "step": 437 }, { "epoch": 0.3596797372202833, "grad_norm": 0.08150382339954376, "learning_rate": 8.81239714755897e-05, "loss": 0.3271, "step": 438 }, { "epoch": 0.3605009238349415, "grad_norm": 0.08224772661924362, "learning_rate": 8.809654415798135e-05, "loss": 0.3847, "step": 439 }, { "epoch": 0.36132211044959966, "grad_norm": 0.07509076595306396, "learning_rate": 8.806911684037302e-05, "loss": 0.346, "step": 440 }, { "epoch": 0.36132211044959966, "eval_runtime": 493.5837, "eval_samples_per_second": 0.399, "eval_steps_per_second": 0.201, "step": 440 }, { "epoch": 0.36214329706425785, "grad_norm": 0.07801686972379684, "learning_rate": 8.804168952276468e-05, "loss": 0.3352, "step": 441 }, { "epoch": 0.36296448367891604, "grad_norm": 0.08195506781339645, "learning_rate": 8.801426220515634e-05, "loss": 0.3076, "step": 442 }, { "epoch": 0.3637856702935742, "grad_norm": 0.08178213983774185, "learning_rate": 8.798683488754799e-05, "loss": 0.3565, "step": 443 }, { "epoch": 0.3646068569082324, "grad_norm": 0.07905025035142899, "learning_rate": 8.795940756993966e-05, "loss": 0.3513, "step": 444 }, { "epoch": 0.3654280435228906, "grad_norm": 0.1251908391714096, "learning_rate": 8.793198025233133e-05, "loss": 0.3656, "step": 445 }, { "epoch": 0.36624923013754873, "grad_norm": 0.08813164383172989, "learning_rate": 8.790455293472298e-05, "loss": 0.3352, "step": 446 }, { "epoch": 0.3670704167522069, "grad_norm": 0.0773930549621582, "learning_rate": 8.787712561711465e-05, "loss": 0.3241, "step": 447 }, { "epoch": 0.3678916033668651, "grad_norm": 0.08977091312408447, "learning_rate": 8.78496982995063e-05, "loss": 0.3308, "step": 448 }, { "epoch": 0.3687127899815233, "grad_norm": 0.0813109427690506, "learning_rate": 8.782227098189797e-05, "loss": 0.344, "step": 449 }, { "epoch": 0.3695339765961815, "grad_norm": 0.08452528715133667, "learning_rate": 8.779484366428964e-05, "loss": 0.2985, "step": 450 }, { "epoch": 0.37035516321083967, "grad_norm": 0.08880957961082458, "learning_rate": 8.77674163466813e-05, "loss": 0.3674, "step": 451 }, { "epoch": 0.37117634982549785, "grad_norm": 0.08020122349262238, "learning_rate": 8.773998902907295e-05, "loss": 0.3487, "step": 452 }, { "epoch": 0.37199753644015604, "grad_norm": 0.09180288016796112, "learning_rate": 8.771256171146463e-05, "loss": 0.364, "step": 453 }, { "epoch": 0.3728187230548142, "grad_norm": 0.08132430911064148, "learning_rate": 8.768513439385628e-05, "loss": 0.3428, "step": 454 }, { "epoch": 0.3736399096694724, "grad_norm": 0.08801679313182831, "learning_rate": 8.765770707624795e-05, "loss": 0.3287, "step": 455 }, { "epoch": 0.37446109628413055, "grad_norm": 0.088813915848732, "learning_rate": 8.763027975863961e-05, "loss": 0.3508, "step": 456 }, { "epoch": 0.37528228289878873, "grad_norm": 0.07454930245876312, "learning_rate": 8.760285244103128e-05, "loss": 0.3267, "step": 457 }, { "epoch": 0.3761034695134469, "grad_norm": 0.0816444680094719, "learning_rate": 8.757542512342294e-05, "loss": 0.3237, "step": 458 }, { "epoch": 0.3769246561281051, "grad_norm": 0.08916571736335754, "learning_rate": 8.75479978058146e-05, "loss": 0.3599, "step": 459 }, { "epoch": 0.3777458427427633, "grad_norm": 0.08580256998538971, "learning_rate": 8.752057048820625e-05, "loss": 0.351, "step": 460 }, { "epoch": 0.3785670293574215, "grad_norm": 0.0849752202630043, "learning_rate": 8.749314317059792e-05, "loss": 0.3616, "step": 461 }, { "epoch": 0.37938821597207967, "grad_norm": 0.11230983585119247, "learning_rate": 8.746571585298959e-05, "loss": 0.3401, "step": 462 }, { "epoch": 0.38020940258673785, "grad_norm": 0.08614380657672882, "learning_rate": 8.743828853538124e-05, "loss": 0.3603, "step": 463 }, { "epoch": 0.38103058920139604, "grad_norm": 0.08322657644748688, "learning_rate": 8.741086121777291e-05, "loss": 0.3087, "step": 464 }, { "epoch": 0.3818517758160542, "grad_norm": 0.08578125387430191, "learning_rate": 8.738343390016456e-05, "loss": 0.3483, "step": 465 }, { "epoch": 0.38267296243071236, "grad_norm": 0.08472792059183121, "learning_rate": 8.735600658255623e-05, "loss": 0.3445, "step": 466 }, { "epoch": 0.38349414904537055, "grad_norm": 0.08300035446882248, "learning_rate": 8.73285792649479e-05, "loss": 0.3384, "step": 467 }, { "epoch": 0.38431533566002873, "grad_norm": 0.08582165092229843, "learning_rate": 8.730115194733955e-05, "loss": 0.3497, "step": 468 }, { "epoch": 0.3851365222746869, "grad_norm": 0.28932541608810425, "learning_rate": 8.727372462973121e-05, "loss": 0.3091, "step": 469 }, { "epoch": 0.3859577088893451, "grad_norm": 0.08396289497613907, "learning_rate": 8.724629731212288e-05, "loss": 0.3472, "step": 470 }, { "epoch": 0.3867788955040033, "grad_norm": 0.08243526518344879, "learning_rate": 8.721886999451454e-05, "loss": 0.3641, "step": 471 }, { "epoch": 0.3876000821186615, "grad_norm": 0.08014799654483795, "learning_rate": 8.71914426769062e-05, "loss": 0.3389, "step": 472 }, { "epoch": 0.38842126873331967, "grad_norm": 0.08125128597021103, "learning_rate": 8.716401535929787e-05, "loss": 0.3293, "step": 473 }, { "epoch": 0.3892424553479778, "grad_norm": 0.08730563521385193, "learning_rate": 8.713658804168952e-05, "loss": 0.3102, "step": 474 }, { "epoch": 0.390063641962636, "grad_norm": 0.08945506811141968, "learning_rate": 8.710916072408119e-05, "loss": 0.3496, "step": 475 }, { "epoch": 0.3908848285772942, "grad_norm": 0.09083441644906998, "learning_rate": 8.708173340647286e-05, "loss": 0.3411, "step": 476 }, { "epoch": 0.39170601519195236, "grad_norm": 0.08667677640914917, "learning_rate": 8.705430608886451e-05, "loss": 0.3289, "step": 477 }, { "epoch": 0.39252720180661055, "grad_norm": 0.08793242275714874, "learning_rate": 8.702687877125617e-05, "loss": 0.3542, "step": 478 }, { "epoch": 0.39334838842126874, "grad_norm": 0.08769497275352478, "learning_rate": 8.699945145364785e-05, "loss": 0.3207, "step": 479 }, { "epoch": 0.3941695750359269, "grad_norm": 0.08303678780794144, "learning_rate": 8.69720241360395e-05, "loss": 0.343, "step": 480 }, { "epoch": 0.3949907616505851, "grad_norm": 0.07943850755691528, "learning_rate": 8.694459681843116e-05, "loss": 0.3273, "step": 481 }, { "epoch": 0.3958119482652433, "grad_norm": 0.08450903743505478, "learning_rate": 8.691716950082282e-05, "loss": 0.3895, "step": 482 }, { "epoch": 0.3966331348799015, "grad_norm": 0.07759539037942886, "learning_rate": 8.688974218321448e-05, "loss": 0.3426, "step": 483 }, { "epoch": 0.3974543214945596, "grad_norm": 0.09013447165489197, "learning_rate": 8.686231486560615e-05, "loss": 0.336, "step": 484 }, { "epoch": 0.3982755081092178, "grad_norm": 0.08396545797586441, "learning_rate": 8.683488754799781e-05, "loss": 0.3647, "step": 485 }, { "epoch": 0.399096694723876, "grad_norm": 0.08143208920955658, "learning_rate": 8.680746023038947e-05, "loss": 0.3239, "step": 486 }, { "epoch": 0.3999178813385342, "grad_norm": 0.07628796994686127, "learning_rate": 8.678003291278114e-05, "loss": 0.3389, "step": 487 }, { "epoch": 0.40073906795319236, "grad_norm": 0.08144954591989517, "learning_rate": 8.67526055951728e-05, "loss": 0.3258, "step": 488 }, { "epoch": 0.40156025456785055, "grad_norm": 0.08240176737308502, "learning_rate": 8.672517827756446e-05, "loss": 0.3316, "step": 489 }, { "epoch": 0.40238144118250874, "grad_norm": 0.09237710386514664, "learning_rate": 8.669775095995613e-05, "loss": 0.3435, "step": 490 }, { "epoch": 0.4032026277971669, "grad_norm": 0.07773042470216751, "learning_rate": 8.667032364234778e-05, "loss": 0.3479, "step": 491 }, { "epoch": 0.4040238144118251, "grad_norm": 0.08410782366991043, "learning_rate": 8.664289632473945e-05, "loss": 0.3792, "step": 492 }, { "epoch": 0.40484500102648324, "grad_norm": 0.08251044154167175, "learning_rate": 8.661546900713112e-05, "loss": 0.3388, "step": 493 }, { "epoch": 0.40566618764114143, "grad_norm": 0.08804575353860855, "learning_rate": 8.658804168952277e-05, "loss": 0.3388, "step": 494 }, { "epoch": 0.4064873742557996, "grad_norm": 0.10321494936943054, "learning_rate": 8.656061437191442e-05, "loss": 0.3348, "step": 495 }, { "epoch": 0.4073085608704578, "grad_norm": 0.07642900943756104, "learning_rate": 8.653318705430609e-05, "loss": 0.3155, "step": 496 }, { "epoch": 0.408129747485116, "grad_norm": 0.08629249781370163, "learning_rate": 8.650575973669776e-05, "loss": 0.3303, "step": 497 }, { "epoch": 0.4089509340997742, "grad_norm": 0.07933122664690018, "learning_rate": 8.647833241908941e-05, "loss": 0.3399, "step": 498 }, { "epoch": 0.40977212071443236, "grad_norm": 0.0802338495850563, "learning_rate": 8.645090510148108e-05, "loss": 0.3623, "step": 499 }, { "epoch": 0.41059330732909055, "grad_norm": 0.07566798478364944, "learning_rate": 8.642347778387274e-05, "loss": 0.3195, "step": 500 }, { "epoch": 0.41141449394374874, "grad_norm": 0.0790770947933197, "learning_rate": 8.63960504662644e-05, "loss": 0.3478, "step": 501 }, { "epoch": 0.4122356805584069, "grad_norm": 0.08219192922115326, "learning_rate": 8.636862314865607e-05, "loss": 0.4037, "step": 502 }, { "epoch": 0.41305686717306506, "grad_norm": 0.09459681063890457, "learning_rate": 8.634119583104773e-05, "loss": 0.3604, "step": 503 }, { "epoch": 0.41387805378772324, "grad_norm": 0.08201157301664352, "learning_rate": 8.631376851343938e-05, "loss": 0.3561, "step": 504 }, { "epoch": 0.41469924040238143, "grad_norm": 0.08181356638669968, "learning_rate": 8.628634119583105e-05, "loss": 0.3359, "step": 505 }, { "epoch": 0.4155204270170396, "grad_norm": 0.08242693543434143, "learning_rate": 8.625891387822272e-05, "loss": 0.3501, "step": 506 }, { "epoch": 0.4163416136316978, "grad_norm": 0.08045560866594315, "learning_rate": 8.623148656061437e-05, "loss": 0.3188, "step": 507 }, { "epoch": 0.417162800246356, "grad_norm": 0.08392325043678284, "learning_rate": 8.620405924300604e-05, "loss": 0.3607, "step": 508 }, { "epoch": 0.4179839868610142, "grad_norm": 0.0920289009809494, "learning_rate": 8.61766319253977e-05, "loss": 0.3425, "step": 509 }, { "epoch": 0.41880517347567237, "grad_norm": 0.08921059966087341, "learning_rate": 8.614920460778936e-05, "loss": 0.3386, "step": 510 }, { "epoch": 0.41962636009033055, "grad_norm": 0.08129517734050751, "learning_rate": 8.612177729018103e-05, "loss": 0.3631, "step": 511 }, { "epoch": 0.4204475467049887, "grad_norm": 0.19061192870140076, "learning_rate": 8.609434997257268e-05, "loss": 0.3285, "step": 512 }, { "epoch": 0.42126873331964687, "grad_norm": 0.08446541428565979, "learning_rate": 8.606692265496434e-05, "loss": 0.4133, "step": 513 }, { "epoch": 0.42208991993430506, "grad_norm": 0.07793641835451126, "learning_rate": 8.603949533735602e-05, "loss": 0.3166, "step": 514 }, { "epoch": 0.42291110654896324, "grad_norm": 0.07862336188554764, "learning_rate": 8.601206801974767e-05, "loss": 0.3448, "step": 515 }, { "epoch": 0.42373229316362143, "grad_norm": 0.07776626199483871, "learning_rate": 8.598464070213933e-05, "loss": 0.3717, "step": 516 }, { "epoch": 0.4245534797782796, "grad_norm": 0.08171868324279785, "learning_rate": 8.5957213384531e-05, "loss": 0.3384, "step": 517 }, { "epoch": 0.4253746663929378, "grad_norm": 0.08469153940677643, "learning_rate": 8.592978606692266e-05, "loss": 0.3238, "step": 518 }, { "epoch": 0.426195853007596, "grad_norm": 0.0739326924085617, "learning_rate": 8.590235874931433e-05, "loss": 0.3033, "step": 519 }, { "epoch": 0.4270170396222542, "grad_norm": 0.08084508031606674, "learning_rate": 8.587493143170599e-05, "loss": 0.3557, "step": 520 }, { "epoch": 0.4278382262369123, "grad_norm": 0.079572893679142, "learning_rate": 8.584750411409764e-05, "loss": 0.3369, "step": 521 }, { "epoch": 0.4286594128515705, "grad_norm": 0.09056146442890167, "learning_rate": 8.582007679648931e-05, "loss": 0.3362, "step": 522 }, { "epoch": 0.4294805994662287, "grad_norm": 0.09055227041244507, "learning_rate": 8.579264947888098e-05, "loss": 0.3457, "step": 523 }, { "epoch": 0.43030178608088687, "grad_norm": 0.11645710468292236, "learning_rate": 8.576522216127263e-05, "loss": 0.3378, "step": 524 }, { "epoch": 0.43112297269554506, "grad_norm": 0.08169250190258026, "learning_rate": 8.57377948436643e-05, "loss": 0.3406, "step": 525 }, { "epoch": 0.43194415931020325, "grad_norm": 0.08108235895633698, "learning_rate": 8.571036752605595e-05, "loss": 0.3253, "step": 526 }, { "epoch": 0.43276534592486143, "grad_norm": 0.07971398532390594, "learning_rate": 8.568294020844762e-05, "loss": 0.3389, "step": 527 }, { "epoch": 0.4335865325395196, "grad_norm": 0.08281555771827698, "learning_rate": 8.565551289083929e-05, "loss": 0.3289, "step": 528 }, { "epoch": 0.4335865325395196, "eval_runtime": 493.1791, "eval_samples_per_second": 0.399, "eval_steps_per_second": 0.201, "step": 528 }, { "epoch": 0.4344077191541778, "grad_norm": 0.08380738645792007, "learning_rate": 8.562808557323094e-05, "loss": 0.3445, "step": 529 }, { "epoch": 0.435228905768836, "grad_norm": 0.07987751066684723, "learning_rate": 8.56006582556226e-05, "loss": 0.3149, "step": 530 }, { "epoch": 0.4360500923834941, "grad_norm": 0.08102700114250183, "learning_rate": 8.557323093801427e-05, "loss": 0.3309, "step": 531 }, { "epoch": 0.4368712789981523, "grad_norm": 0.08056002855300903, "learning_rate": 8.554580362040593e-05, "loss": 0.3579, "step": 532 }, { "epoch": 0.4376924656128105, "grad_norm": 0.08129940181970596, "learning_rate": 8.551837630279759e-05, "loss": 0.3414, "step": 533 }, { "epoch": 0.4385136522274687, "grad_norm": 0.07720111310482025, "learning_rate": 8.549094898518926e-05, "loss": 0.3066, "step": 534 }, { "epoch": 0.4393348388421269, "grad_norm": 0.08578766137361526, "learning_rate": 8.546352166758091e-05, "loss": 0.4056, "step": 535 }, { "epoch": 0.44015602545678506, "grad_norm": 0.14550408720970154, "learning_rate": 8.543609434997258e-05, "loss": 0.353, "step": 536 }, { "epoch": 0.44097721207144325, "grad_norm": 0.07624532282352448, "learning_rate": 8.540866703236425e-05, "loss": 0.323, "step": 537 }, { "epoch": 0.44179839868610143, "grad_norm": 0.07774993777275085, "learning_rate": 8.53812397147559e-05, "loss": 0.3244, "step": 538 }, { "epoch": 0.4426195853007596, "grad_norm": 0.08061113953590393, "learning_rate": 8.535381239714755e-05, "loss": 0.3209, "step": 539 }, { "epoch": 0.44344077191541775, "grad_norm": 0.07771355658769608, "learning_rate": 8.532638507953922e-05, "loss": 0.3604, "step": 540 }, { "epoch": 0.44426195853007594, "grad_norm": 0.08160841464996338, "learning_rate": 8.529895776193089e-05, "loss": 0.3373, "step": 541 }, { "epoch": 0.4450831451447341, "grad_norm": 0.0819459930062294, "learning_rate": 8.527153044432254e-05, "loss": 0.3097, "step": 542 }, { "epoch": 0.4459043317593923, "grad_norm": 0.08429809659719467, "learning_rate": 8.524410312671421e-05, "loss": 0.3538, "step": 543 }, { "epoch": 0.4467255183740505, "grad_norm": 0.08519409596920013, "learning_rate": 8.521667580910587e-05, "loss": 0.3486, "step": 544 }, { "epoch": 0.4475467049887087, "grad_norm": 0.08900930732488632, "learning_rate": 8.518924849149753e-05, "loss": 0.3339, "step": 545 }, { "epoch": 0.4483678916033669, "grad_norm": 0.08914072066545486, "learning_rate": 8.51618211738892e-05, "loss": 0.2933, "step": 546 }, { "epoch": 0.44918907821802506, "grad_norm": 0.08086919039487839, "learning_rate": 8.513439385628086e-05, "loss": 0.3367, "step": 547 }, { "epoch": 0.45001026483268325, "grad_norm": 0.08358173817396164, "learning_rate": 8.510696653867251e-05, "loss": 0.3443, "step": 548 }, { "epoch": 0.45083145144734144, "grad_norm": 0.0807274803519249, "learning_rate": 8.507953922106419e-05, "loss": 0.3332, "step": 549 }, { "epoch": 0.45165263806199957, "grad_norm": 0.08126116544008255, "learning_rate": 8.505211190345585e-05, "loss": 0.3485, "step": 550 }, { "epoch": 0.45247382467665775, "grad_norm": 0.08835196495056152, "learning_rate": 8.50246845858475e-05, "loss": 0.3739, "step": 551 }, { "epoch": 0.45329501129131594, "grad_norm": 0.07614720612764359, "learning_rate": 8.499725726823917e-05, "loss": 0.3464, "step": 552 }, { "epoch": 0.45411619790597413, "grad_norm": 0.07444674521684647, "learning_rate": 8.496982995063084e-05, "loss": 0.3553, "step": 553 }, { "epoch": 0.4549373845206323, "grad_norm": 0.0741548240184784, "learning_rate": 8.49424026330225e-05, "loss": 0.341, "step": 554 }, { "epoch": 0.4557585711352905, "grad_norm": 0.0820796862244606, "learning_rate": 8.491497531541416e-05, "loss": 0.3359, "step": 555 }, { "epoch": 0.4565797577499487, "grad_norm": 0.08371058106422424, "learning_rate": 8.488754799780581e-05, "loss": 0.3339, "step": 556 }, { "epoch": 0.4574009443646069, "grad_norm": 0.08354955166578293, "learning_rate": 8.486012068019748e-05, "loss": 0.3619, "step": 557 }, { "epoch": 0.45822213097926506, "grad_norm": 0.08976832777261734, "learning_rate": 8.483269336258915e-05, "loss": 0.353, "step": 558 }, { "epoch": 0.4590433175939232, "grad_norm": 0.07933076471090317, "learning_rate": 8.48052660449808e-05, "loss": 0.3764, "step": 559 }, { "epoch": 0.4598645042085814, "grad_norm": 0.08624580502510071, "learning_rate": 8.477783872737247e-05, "loss": 0.3372, "step": 560 }, { "epoch": 0.46068569082323957, "grad_norm": 0.08301227539777756, "learning_rate": 8.475041140976413e-05, "loss": 0.3748, "step": 561 }, { "epoch": 0.46150687743789776, "grad_norm": 0.07408706098794937, "learning_rate": 8.47229840921558e-05, "loss": 0.2622, "step": 562 }, { "epoch": 0.46232806405255594, "grad_norm": 0.07799118012189865, "learning_rate": 8.469555677454746e-05, "loss": 0.3614, "step": 563 }, { "epoch": 0.46314925066721413, "grad_norm": 0.08185340464115143, "learning_rate": 8.466812945693912e-05, "loss": 0.3423, "step": 564 }, { "epoch": 0.4639704372818723, "grad_norm": 0.0817733034491539, "learning_rate": 8.464070213933077e-05, "loss": 0.3519, "step": 565 }, { "epoch": 0.4647916238965305, "grad_norm": 0.08026120066642761, "learning_rate": 8.461327482172244e-05, "loss": 0.3372, "step": 566 }, { "epoch": 0.4656128105111887, "grad_norm": 0.07413862645626068, "learning_rate": 8.45858475041141e-05, "loss": 0.3017, "step": 567 }, { "epoch": 0.4664339971258468, "grad_norm": 0.09256873279809952, "learning_rate": 8.455842018650576e-05, "loss": 0.3495, "step": 568 }, { "epoch": 0.467255183740505, "grad_norm": 0.08319979161024094, "learning_rate": 8.453099286889743e-05, "loss": 0.3339, "step": 569 }, { "epoch": 0.4680763703551632, "grad_norm": 0.08824891597032547, "learning_rate": 8.450356555128908e-05, "loss": 0.3272, "step": 570 }, { "epoch": 0.4688975569698214, "grad_norm": 0.0790686085820198, "learning_rate": 8.447613823368075e-05, "loss": 0.3446, "step": 571 }, { "epoch": 0.46971874358447957, "grad_norm": 0.08730736374855042, "learning_rate": 8.444871091607242e-05, "loss": 0.3731, "step": 572 }, { "epoch": 0.47053993019913776, "grad_norm": 0.08023872971534729, "learning_rate": 8.442128359846407e-05, "loss": 0.3226, "step": 573 }, { "epoch": 0.47136111681379594, "grad_norm": 0.09242340922355652, "learning_rate": 8.439385628085573e-05, "loss": 0.3229, "step": 574 }, { "epoch": 0.47218230342845413, "grad_norm": 0.07871225476264954, "learning_rate": 8.43664289632474e-05, "loss": 0.3112, "step": 575 }, { "epoch": 0.4730034900431123, "grad_norm": 0.08368773013353348, "learning_rate": 8.433900164563906e-05, "loss": 0.3357, "step": 576 }, { "epoch": 0.4738246766577705, "grad_norm": 0.07561603933572769, "learning_rate": 8.431157432803072e-05, "loss": 0.3125, "step": 577 }, { "epoch": 0.47464586327242864, "grad_norm": 0.08441564440727234, "learning_rate": 8.428414701042238e-05, "loss": 0.3211, "step": 578 }, { "epoch": 0.4754670498870868, "grad_norm": 0.09772761911153793, "learning_rate": 8.425671969281404e-05, "loss": 0.3285, "step": 579 }, { "epoch": 0.476288236501745, "grad_norm": 0.07900305837392807, "learning_rate": 8.422929237520571e-05, "loss": 0.344, "step": 580 }, { "epoch": 0.4771094231164032, "grad_norm": 0.13551943004131317, "learning_rate": 8.420186505759737e-05, "loss": 0.3515, "step": 581 }, { "epoch": 0.4779306097310614, "grad_norm": 0.07592803239822388, "learning_rate": 8.417443773998903e-05, "loss": 0.3377, "step": 582 }, { "epoch": 0.47875179634571957, "grad_norm": 0.08301271498203278, "learning_rate": 8.414701042238068e-05, "loss": 0.365, "step": 583 }, { "epoch": 0.47957298296037776, "grad_norm": 0.11750177294015884, "learning_rate": 8.411958310477237e-05, "loss": 0.3159, "step": 584 }, { "epoch": 0.48039416957503595, "grad_norm": 0.08506017178297043, "learning_rate": 8.409215578716402e-05, "loss": 0.3381, "step": 585 }, { "epoch": 0.48121535618969413, "grad_norm": 0.0827595517039299, "learning_rate": 8.406472846955569e-05, "loss": 0.3497, "step": 586 }, { "epoch": 0.48203654280435226, "grad_norm": 0.07703883200883865, "learning_rate": 8.403730115194734e-05, "loss": 0.2989, "step": 587 }, { "epoch": 0.48285772941901045, "grad_norm": 0.07976412773132324, "learning_rate": 8.400987383433901e-05, "loss": 0.3617, "step": 588 }, { "epoch": 0.48367891603366864, "grad_norm": 0.09424680471420288, "learning_rate": 8.398244651673068e-05, "loss": 0.3623, "step": 589 }, { "epoch": 0.4845001026483268, "grad_norm": 0.08831481635570526, "learning_rate": 8.395501919912233e-05, "loss": 0.3526, "step": 590 }, { "epoch": 0.485321289262985, "grad_norm": 0.07811284810304642, "learning_rate": 8.392759188151399e-05, "loss": 0.3721, "step": 591 }, { "epoch": 0.4861424758776432, "grad_norm": 0.08139461278915405, "learning_rate": 8.390016456390565e-05, "loss": 0.33, "step": 592 }, { "epoch": 0.4869636624923014, "grad_norm": 0.0807470753788948, "learning_rate": 8.387273724629732e-05, "loss": 0.372, "step": 593 }, { "epoch": 0.4877848491069596, "grad_norm": 0.08856850117444992, "learning_rate": 8.384530992868898e-05, "loss": 0.3288, "step": 594 }, { "epoch": 0.48860603572161776, "grad_norm": 0.08670219779014587, "learning_rate": 8.381788261108064e-05, "loss": 0.4113, "step": 595 }, { "epoch": 0.4894272223362759, "grad_norm": 0.0814259722828865, "learning_rate": 8.37904552934723e-05, "loss": 0.3735, "step": 596 }, { "epoch": 0.4902484089509341, "grad_norm": 0.08457124978303909, "learning_rate": 8.376302797586397e-05, "loss": 0.3516, "step": 597 }, { "epoch": 0.49106959556559227, "grad_norm": 0.0774846151471138, "learning_rate": 8.373560065825563e-05, "loss": 0.3604, "step": 598 }, { "epoch": 0.49189078218025045, "grad_norm": 0.07951226830482483, "learning_rate": 8.370817334064729e-05, "loss": 0.3016, "step": 599 }, { "epoch": 0.49271196879490864, "grad_norm": 0.08235177397727966, "learning_rate": 8.368074602303894e-05, "loss": 0.3502, "step": 600 }, { "epoch": 0.4935331554095668, "grad_norm": 0.08112053573131561, "learning_rate": 8.365331870543061e-05, "loss": 0.3122, "step": 601 }, { "epoch": 0.494354342024225, "grad_norm": 0.07517421990633011, "learning_rate": 8.362589138782228e-05, "loss": 0.3482, "step": 602 }, { "epoch": 0.4951755286388832, "grad_norm": 0.08243374526500702, "learning_rate": 8.359846407021393e-05, "loss": 0.3455, "step": 603 }, { "epoch": 0.4959967152535414, "grad_norm": 0.08028637617826462, "learning_rate": 8.35710367526056e-05, "loss": 0.3454, "step": 604 }, { "epoch": 0.4968179018681996, "grad_norm": 0.08006370067596436, "learning_rate": 8.354360943499726e-05, "loss": 0.3216, "step": 605 }, { "epoch": 0.4976390884828577, "grad_norm": 0.1940547525882721, "learning_rate": 8.351618211738892e-05, "loss": 0.3017, "step": 606 }, { "epoch": 0.4984602750975159, "grad_norm": 0.07875222712755203, "learning_rate": 8.348875479978059e-05, "loss": 0.3295, "step": 607 }, { "epoch": 0.4992814617121741, "grad_norm": 0.07511032372713089, "learning_rate": 8.346132748217225e-05, "loss": 0.328, "step": 608 }, { "epoch": 0.5001026483268323, "grad_norm": 0.08197803795337677, "learning_rate": 8.34339001645639e-05, "loss": 0.3232, "step": 609 }, { "epoch": 0.5009238349414905, "grad_norm": 0.08152435719966888, "learning_rate": 8.340647284695558e-05, "loss": 0.3546, "step": 610 }, { "epoch": 0.5017450215561486, "grad_norm": 0.0796707347035408, "learning_rate": 8.337904552934724e-05, "loss": 0.3518, "step": 611 }, { "epoch": 0.5025662081708068, "grad_norm": 0.07969588786363602, "learning_rate": 8.335161821173889e-05, "loss": 0.3195, "step": 612 }, { "epoch": 0.503387394785465, "grad_norm": 0.08270443230867386, "learning_rate": 8.332419089413056e-05, "loss": 0.3244, "step": 613 }, { "epoch": 0.5042085814001231, "grad_norm": 0.08303502202033997, "learning_rate": 8.329676357652221e-05, "loss": 0.3386, "step": 614 }, { "epoch": 0.5050297680147814, "grad_norm": 0.07783270627260208, "learning_rate": 8.326933625891388e-05, "loss": 0.3046, "step": 615 }, { "epoch": 0.5058509546294395, "grad_norm": 0.08112157136201859, "learning_rate": 8.324190894130555e-05, "loss": 0.3244, "step": 616 }, { "epoch": 0.5058509546294395, "eval_runtime": 494.4481, "eval_samples_per_second": 0.398, "eval_steps_per_second": 0.2, "step": 616 }, { "epoch": 0.5066721412440978, "grad_norm": 0.08969878405332565, "learning_rate": 8.32144816236972e-05, "loss": 0.3365, "step": 617 }, { "epoch": 0.5074933278587559, "grad_norm": 0.08142217993736267, "learning_rate": 8.318705430608886e-05, "loss": 0.3413, "step": 618 }, { "epoch": 0.5083145144734141, "grad_norm": 0.08319269865751266, "learning_rate": 8.315962698848054e-05, "loss": 0.3458, "step": 619 }, { "epoch": 0.5091357010880723, "grad_norm": 0.08104688674211502, "learning_rate": 8.313219967087219e-05, "loss": 0.3286, "step": 620 }, { "epoch": 0.5099568877027304, "grad_norm": 0.08447278290987015, "learning_rate": 8.310477235326386e-05, "loss": 0.3485, "step": 621 }, { "epoch": 0.5107780743173886, "grad_norm": 0.08879225701093674, "learning_rate": 8.307734503565551e-05, "loss": 0.2986, "step": 622 }, { "epoch": 0.5115992609320468, "grad_norm": 0.07878706604242325, "learning_rate": 8.304991771804718e-05, "loss": 0.3301, "step": 623 }, { "epoch": 0.512420447546705, "grad_norm": 0.08089771121740341, "learning_rate": 8.302249040043885e-05, "loss": 0.3183, "step": 624 }, { "epoch": 0.5132416341613631, "grad_norm": 0.08130493015050888, "learning_rate": 8.29950630828305e-05, "loss": 0.3559, "step": 625 }, { "epoch": 0.5140628207760214, "grad_norm": 0.0880448967218399, "learning_rate": 8.296763576522216e-05, "loss": 0.3567, "step": 626 }, { "epoch": 0.5148840073906795, "grad_norm": 0.087735615670681, "learning_rate": 8.294020844761383e-05, "loss": 0.3123, "step": 627 }, { "epoch": 0.5157051940053378, "grad_norm": 0.08659686893224716, "learning_rate": 8.29127811300055e-05, "loss": 0.3212, "step": 628 }, { "epoch": 0.5165263806199959, "grad_norm": 0.0853281244635582, "learning_rate": 8.288535381239715e-05, "loss": 0.3337, "step": 629 }, { "epoch": 0.517347567234654, "grad_norm": 0.07861575484275818, "learning_rate": 8.285792649478882e-05, "loss": 0.3666, "step": 630 }, { "epoch": 0.5181687538493123, "grad_norm": 0.08911813795566559, "learning_rate": 8.283049917718047e-05, "loss": 0.349, "step": 631 }, { "epoch": 0.5189899404639704, "grad_norm": 0.08234114944934845, "learning_rate": 8.280307185957214e-05, "loss": 0.3354, "step": 632 }, { "epoch": 0.5198111270786286, "grad_norm": 0.08694402128458023, "learning_rate": 8.277564454196381e-05, "loss": 0.3321, "step": 633 }, { "epoch": 0.5206323136932868, "grad_norm": 0.08186987787485123, "learning_rate": 8.274821722435546e-05, "loss": 0.3595, "step": 634 }, { "epoch": 0.521453500307945, "grad_norm": 0.08425264060497284, "learning_rate": 8.272078990674712e-05, "loss": 0.3641, "step": 635 }, { "epoch": 0.5222746869226031, "grad_norm": 0.1728411465883255, "learning_rate": 8.269336258913878e-05, "loss": 0.3646, "step": 636 }, { "epoch": 0.5230958735372614, "grad_norm": 0.09049440920352936, "learning_rate": 8.266593527153045e-05, "loss": 0.3368, "step": 637 }, { "epoch": 0.5239170601519195, "grad_norm": 0.081216000020504, "learning_rate": 8.26385079539221e-05, "loss": 0.3814, "step": 638 }, { "epoch": 0.5247382467665777, "grad_norm": 0.08497685194015503, "learning_rate": 8.261108063631377e-05, "loss": 0.3315, "step": 639 }, { "epoch": 0.5255594333812359, "grad_norm": 0.08426640927791595, "learning_rate": 8.258365331870543e-05, "loss": 0.3366, "step": 640 }, { "epoch": 0.526380619995894, "grad_norm": 0.08354829251766205, "learning_rate": 8.25562260010971e-05, "loss": 0.3439, "step": 641 }, { "epoch": 0.5272018066105523, "grad_norm": 0.08015681803226471, "learning_rate": 8.252879868348876e-05, "loss": 0.3401, "step": 642 }, { "epoch": 0.5280229932252104, "grad_norm": 0.08374017477035522, "learning_rate": 8.250137136588042e-05, "loss": 0.3533, "step": 643 }, { "epoch": 0.5288441798398686, "grad_norm": 0.09182158857584, "learning_rate": 8.247394404827207e-05, "loss": 0.2983, "step": 644 }, { "epoch": 0.5296653664545268, "grad_norm": 0.08268212527036667, "learning_rate": 8.244651673066375e-05, "loss": 0.331, "step": 645 }, { "epoch": 0.530486553069185, "grad_norm": 0.08059857040643692, "learning_rate": 8.241908941305541e-05, "loss": 0.3448, "step": 646 }, { "epoch": 0.5313077396838431, "grad_norm": 0.08811867982149124, "learning_rate": 8.239166209544706e-05, "loss": 0.3678, "step": 647 }, { "epoch": 0.5321289262985013, "grad_norm": 0.07894317060709, "learning_rate": 8.236423477783873e-05, "loss": 0.3456, "step": 648 }, { "epoch": 0.5329501129131595, "grad_norm": 0.07806131988763809, "learning_rate": 8.23368074602304e-05, "loss": 0.3286, "step": 649 }, { "epoch": 0.5337712995278177, "grad_norm": 0.08501176536083221, "learning_rate": 8.230938014262205e-05, "loss": 0.3383, "step": 650 }, { "epoch": 0.5345924861424759, "grad_norm": 0.12988948822021484, "learning_rate": 8.228195282501372e-05, "loss": 0.3384, "step": 651 }, { "epoch": 0.535413672757134, "grad_norm": 0.08005869388580322, "learning_rate": 8.225452550740537e-05, "loss": 0.323, "step": 652 }, { "epoch": 0.5362348593717923, "grad_norm": 0.07902330160140991, "learning_rate": 8.222709818979704e-05, "loss": 0.3248, "step": 653 }, { "epoch": 0.5370560459864504, "grad_norm": 0.08319025486707687, "learning_rate": 8.219967087218871e-05, "loss": 0.3145, "step": 654 }, { "epoch": 0.5378772326011086, "grad_norm": 0.08141138404607773, "learning_rate": 8.217224355458037e-05, "loss": 0.3105, "step": 655 }, { "epoch": 0.5386984192157668, "grad_norm": 0.07592695206403732, "learning_rate": 8.214481623697203e-05, "loss": 0.3184, "step": 656 }, { "epoch": 0.5395196058304249, "grad_norm": 0.16522720456123352, "learning_rate": 8.211738891936369e-05, "loss": 0.3492, "step": 657 }, { "epoch": 0.5403407924450832, "grad_norm": 0.0965307429432869, "learning_rate": 8.208996160175536e-05, "loss": 0.3012, "step": 658 }, { "epoch": 0.5411619790597413, "grad_norm": 0.08132067322731018, "learning_rate": 8.206253428414702e-05, "loss": 0.3256, "step": 659 }, { "epoch": 0.5419831656743995, "grad_norm": 0.08183111250400543, "learning_rate": 8.203510696653868e-05, "loss": 0.3591, "step": 660 }, { "epoch": 0.5428043522890577, "grad_norm": 0.07674401253461838, "learning_rate": 8.200767964893033e-05, "loss": 0.3273, "step": 661 }, { "epoch": 0.5436255389037159, "grad_norm": 0.07937440276145935, "learning_rate": 8.1980252331322e-05, "loss": 0.3028, "step": 662 }, { "epoch": 0.544446725518374, "grad_norm": 0.07886291295289993, "learning_rate": 8.195282501371367e-05, "loss": 0.3452, "step": 663 }, { "epoch": 0.5452679121330323, "grad_norm": 0.078894704580307, "learning_rate": 8.192539769610532e-05, "loss": 0.3447, "step": 664 }, { "epoch": 0.5460890987476904, "grad_norm": 0.0840119943022728, "learning_rate": 8.189797037849699e-05, "loss": 0.3348, "step": 665 }, { "epoch": 0.5469102853623486, "grad_norm": 0.08318603038787842, "learning_rate": 8.187054306088864e-05, "loss": 0.3523, "step": 666 }, { "epoch": 0.5477314719770068, "grad_norm": 0.08060989528894424, "learning_rate": 8.184311574328031e-05, "loss": 0.3211, "step": 667 }, { "epoch": 0.5485526585916649, "grad_norm": 0.07878737151622772, "learning_rate": 8.181568842567198e-05, "loss": 0.3865, "step": 668 }, { "epoch": 0.5493738452063232, "grad_norm": 0.08277317881584167, "learning_rate": 8.178826110806363e-05, "loss": 0.3507, "step": 669 }, { "epoch": 0.5501950318209813, "grad_norm": 0.15883035957813263, "learning_rate": 8.176083379045529e-05, "loss": 0.3428, "step": 670 }, { "epoch": 0.5510162184356395, "grad_norm": 0.08117318898439407, "learning_rate": 8.173340647284696e-05, "loss": 0.3297, "step": 671 }, { "epoch": 0.5518374050502977, "grad_norm": 0.07886844873428345, "learning_rate": 8.170597915523862e-05, "loss": 0.3489, "step": 672 }, { "epoch": 0.5526585916649559, "grad_norm": 0.0774691253900528, "learning_rate": 8.167855183763028e-05, "loss": 0.3438, "step": 673 }, { "epoch": 0.553479778279614, "grad_norm": 0.07944227755069733, "learning_rate": 8.165112452002195e-05, "loss": 0.3183, "step": 674 }, { "epoch": 0.5543009648942723, "grad_norm": 0.08308850973844528, "learning_rate": 8.16236972024136e-05, "loss": 0.3268, "step": 675 }, { "epoch": 0.5551221515089304, "grad_norm": 0.07913679629564285, "learning_rate": 8.159626988480527e-05, "loss": 0.3358, "step": 676 }, { "epoch": 0.5559433381235885, "grad_norm": 0.08472824096679688, "learning_rate": 8.156884256719694e-05, "loss": 0.3619, "step": 677 }, { "epoch": 0.5567645247382468, "grad_norm": 0.07870602607727051, "learning_rate": 8.154141524958859e-05, "loss": 0.3372, "step": 678 }, { "epoch": 0.5575857113529049, "grad_norm": 0.07889708876609802, "learning_rate": 8.151398793198025e-05, "loss": 0.34, "step": 679 }, { "epoch": 0.5584068979675632, "grad_norm": 0.0856415405869484, "learning_rate": 8.148656061437193e-05, "loss": 0.3132, "step": 680 }, { "epoch": 0.5592280845822213, "grad_norm": 0.0840810164809227, "learning_rate": 8.145913329676358e-05, "loss": 0.3451, "step": 681 }, { "epoch": 0.5600492711968795, "grad_norm": 0.07759146392345428, "learning_rate": 8.143170597915524e-05, "loss": 0.3452, "step": 682 }, { "epoch": 0.5608704578115377, "grad_norm": 0.08038268238306046, "learning_rate": 8.14042786615469e-05, "loss": 0.3103, "step": 683 }, { "epoch": 0.5616916444261959, "grad_norm": 0.07695576548576355, "learning_rate": 8.137685134393857e-05, "loss": 0.342, "step": 684 }, { "epoch": 0.562512831040854, "grad_norm": 0.08458397537469864, "learning_rate": 8.134942402633024e-05, "loss": 0.344, "step": 685 }, { "epoch": 0.5633340176555122, "grad_norm": 0.08124341070652008, "learning_rate": 8.13219967087219e-05, "loss": 0.3235, "step": 686 }, { "epoch": 0.5641552042701704, "grad_norm": 0.09097409248352051, "learning_rate": 8.129456939111355e-05, "loss": 0.3142, "step": 687 }, { "epoch": 0.5649763908848285, "grad_norm": 0.07753387093544006, "learning_rate": 8.126714207350522e-05, "loss": 0.3295, "step": 688 }, { "epoch": 0.5657975774994868, "grad_norm": 0.07960804551839828, "learning_rate": 8.123971475589688e-05, "loss": 0.3156, "step": 689 }, { "epoch": 0.5666187641141449, "grad_norm": 0.0779690146446228, "learning_rate": 8.121228743828854e-05, "loss": 0.3224, "step": 690 }, { "epoch": 0.5674399507288032, "grad_norm": 0.07964863628149033, "learning_rate": 8.11848601206802e-05, "loss": 0.3846, "step": 691 }, { "epoch": 0.5682611373434613, "grad_norm": 0.08143079280853271, "learning_rate": 8.115743280307186e-05, "loss": 0.3268, "step": 692 }, { "epoch": 0.5690823239581195, "grad_norm": 0.08909327536821365, "learning_rate": 8.113000548546353e-05, "loss": 0.3354, "step": 693 }, { "epoch": 0.5699035105727777, "grad_norm": 0.07217192649841309, "learning_rate": 8.11025781678552e-05, "loss": 0.3077, "step": 694 }, { "epoch": 0.5707246971874358, "grad_norm": 0.07756021618843079, "learning_rate": 8.107515085024685e-05, "loss": 0.3244, "step": 695 }, { "epoch": 0.571545883802094, "grad_norm": 0.08227197080850601, "learning_rate": 8.10477235326385e-05, "loss": 0.34, "step": 696 }, { "epoch": 0.5723670704167522, "grad_norm": 0.07765581458806992, "learning_rate": 8.102029621503017e-05, "loss": 0.2898, "step": 697 }, { "epoch": 0.5731882570314104, "grad_norm": 0.07710105925798416, "learning_rate": 8.099286889742184e-05, "loss": 0.3056, "step": 698 }, { "epoch": 0.5740094436460685, "grad_norm": 0.07832709699869156, "learning_rate": 8.09654415798135e-05, "loss": 0.3338, "step": 699 }, { "epoch": 0.5748306302607268, "grad_norm": 0.08406300097703934, "learning_rate": 8.093801426220516e-05, "loss": 0.3483, "step": 700 }, { "epoch": 0.5756518168753849, "grad_norm": 0.07969654351472855, "learning_rate": 8.091058694459682e-05, "loss": 0.3645, "step": 701 }, { "epoch": 0.5764730034900432, "grad_norm": 0.07744769752025604, "learning_rate": 8.088315962698848e-05, "loss": 0.3324, "step": 702 }, { "epoch": 0.5772941901047013, "grad_norm": 0.07306936383247375, "learning_rate": 8.085573230938015e-05, "loss": 0.3011, "step": 703 }, { "epoch": 0.5781153767193594, "grad_norm": 0.09406828135251999, "learning_rate": 8.082830499177181e-05, "loss": 0.3518, "step": 704 }, { "epoch": 0.5781153767193594, "eval_runtime": 494.5472, "eval_samples_per_second": 0.398, "eval_steps_per_second": 0.2, "step": 704 } ], "logging_steps": 1, "max_steps": 3651, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 88, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 3.1068114755114435e+19, "train_batch_size": 4, "trial_name": null, "trial_params": null }