{ "best_metric": 3.4992141723632812, "best_model_checkpoint": "en-to-lg-nllb-cc/checkpoint-267034", "epoch": 2.9999943827482967, "eval_steps": 500, "global_step": 267034, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0005617251703431579, "grad_norm": 1.403859257698059, "learning_rate": 1.999775308669715e-05, "loss": 4.8949, "step": 50 }, { "epoch": 0.0011234503406863158, "grad_norm": 1.5137478113174438, "learning_rate": 1.9995596049926417e-05, "loss": 4.7613, "step": 100 }, { "epoch": 0.0016851755110294736, "grad_norm": 0.933453381061554, "learning_rate": 1.9993349136623566e-05, "loss": 4.5466, "step": 150 }, { "epoch": 0.0022469006813726315, "grad_norm": 0.7829797863960266, "learning_rate": 1.9991102223320716e-05, "loss": 4.3399, "step": 200 }, { "epoch": 0.0028086258517157894, "grad_norm": 0.8315831422805786, "learning_rate": 1.9988855310017865e-05, "loss": 4.3509, "step": 250 }, { "epoch": 0.0033703510220589473, "grad_norm": 0.8067842125892639, "learning_rate": 1.9986608396715014e-05, "loss": 4.3661, "step": 300 }, { "epoch": 0.003932076192402105, "grad_norm": 0.5773376822471619, "learning_rate": 1.9984361483412163e-05, "loss": 4.1911, "step": 350 }, { "epoch": 0.004493801362745263, "grad_norm": 0.6907451748847961, "learning_rate": 1.9982114570109313e-05, "loss": 4.2254, "step": 400 }, { "epoch": 0.005055526533088421, "grad_norm": 0.6285920143127441, "learning_rate": 1.9979867656806465e-05, "loss": 4.211, "step": 450 }, { "epoch": 0.005617251703431579, "grad_norm": 0.6995010375976562, "learning_rate": 1.9977620743503615e-05, "loss": 4.1618, "step": 500 }, { "epoch": 0.006178976873774737, "grad_norm": 0.7596259117126465, "learning_rate": 1.9975373830200764e-05, "loss": 4.2133, "step": 550 }, { "epoch": 0.0067407020441178946, "grad_norm": 0.6245077848434448, "learning_rate": 1.9973126916897913e-05, "loss": 4.1891, "step": 600 }, { "epoch": 0.0073024272144610524, "grad_norm": 0.6988118886947632, "learning_rate": 1.9970880003595063e-05, "loss": 4.1388, "step": 650 }, { "epoch": 0.00786415238480421, "grad_norm": 0.6634756922721863, "learning_rate": 1.9968633090292212e-05, "loss": 4.1258, "step": 700 }, { "epoch": 0.008425877555147369, "grad_norm": 0.5972195863723755, "learning_rate": 1.996638617698936e-05, "loss": 4.1774, "step": 750 }, { "epoch": 0.008987602725490526, "grad_norm": 0.7192025184631348, "learning_rate": 1.9964139263686514e-05, "loss": 4.1227, "step": 800 }, { "epoch": 0.009549327895833685, "grad_norm": 0.6875787377357483, "learning_rate": 1.9961892350383663e-05, "loss": 4.076, "step": 850 }, { "epoch": 0.010111053066176842, "grad_norm": 0.7681994438171387, "learning_rate": 1.9959645437080813e-05, "loss": 4.1251, "step": 900 }, { "epoch": 0.01067277823652, "grad_norm": 0.7524916529655457, "learning_rate": 1.9957398523777962e-05, "loss": 4.0552, "step": 950 }, { "epoch": 0.011234503406863158, "grad_norm": 0.7998524904251099, "learning_rate": 1.995515161047511e-05, "loss": 4.1645, "step": 1000 }, { "epoch": 0.011796228577206316, "grad_norm": 0.8324928879737854, "learning_rate": 1.995290469717226e-05, "loss": 4.0986, "step": 1050 }, { "epoch": 0.012357953747549473, "grad_norm": 0.9623536467552185, "learning_rate": 1.995065778386941e-05, "loss": 4.1097, "step": 1100 }, { "epoch": 0.012919678917892632, "grad_norm": 0.6794806122779846, "learning_rate": 1.9948410870566563e-05, "loss": 4.0934, "step": 1150 }, { "epoch": 0.013481404088235789, "grad_norm": 0.8683093786239624, "learning_rate": 1.9946163957263712e-05, "loss": 4.0484, "step": 1200 }, { "epoch": 0.014043129258578948, "grad_norm": 0.8103470802307129, "learning_rate": 1.994391704396086e-05, "loss": 4.1611, "step": 1250 }, { "epoch": 0.014604854428922105, "grad_norm": 0.779623806476593, "learning_rate": 1.994167013065801e-05, "loss": 4.0097, "step": 1300 }, { "epoch": 0.015166579599265264, "grad_norm": 0.7256409525871277, "learning_rate": 1.993942321735516e-05, "loss": 4.0929, "step": 1350 }, { "epoch": 0.01572830476960842, "grad_norm": 0.8892433643341064, "learning_rate": 1.993717630405231e-05, "loss": 4.0195, "step": 1400 }, { "epoch": 0.01629002993995158, "grad_norm": 0.8380836248397827, "learning_rate": 1.993492939074946e-05, "loss": 4.0752, "step": 1450 }, { "epoch": 0.016851755110294738, "grad_norm": 0.803108811378479, "learning_rate": 1.993268247744661e-05, "loss": 4.0432, "step": 1500 }, { "epoch": 0.017413480280637893, "grad_norm": 0.8453735113143921, "learning_rate": 1.993043556414376e-05, "loss": 4.119, "step": 1550 }, { "epoch": 0.017975205450981052, "grad_norm": 0.8968884348869324, "learning_rate": 1.992818865084091e-05, "loss": 4.0632, "step": 1600 }, { "epoch": 0.01853693062132421, "grad_norm": 0.7678434252738953, "learning_rate": 1.992594173753806e-05, "loss": 3.9581, "step": 1650 }, { "epoch": 0.01909865579166737, "grad_norm": 0.7526470422744751, "learning_rate": 1.992369482423521e-05, "loss": 4.0942, "step": 1700 }, { "epoch": 0.01966038096201053, "grad_norm": 0.7269778847694397, "learning_rate": 1.9921447910932358e-05, "loss": 4.0031, "step": 1750 }, { "epoch": 0.020222106132353684, "grad_norm": 0.8589780926704407, "learning_rate": 1.9919200997629507e-05, "loss": 4.0544, "step": 1800 }, { "epoch": 0.020783831302696842, "grad_norm": 0.7709614038467407, "learning_rate": 1.991695408432666e-05, "loss": 4.0253, "step": 1850 }, { "epoch": 0.02134555647304, "grad_norm": 1.058897614479065, "learning_rate": 1.991470717102381e-05, "loss": 4.0436, "step": 1900 }, { "epoch": 0.02190728164338316, "grad_norm": 0.7720006704330444, "learning_rate": 1.991246025772096e-05, "loss": 4.0896, "step": 1950 }, { "epoch": 0.022469006813726315, "grad_norm": 0.8622137904167175, "learning_rate": 1.9910213344418108e-05, "loss": 4.1142, "step": 2000 }, { "epoch": 0.023030731984069474, "grad_norm": 0.7667924761772156, "learning_rate": 1.9907966431115257e-05, "loss": 4.0354, "step": 2050 }, { "epoch": 0.023592457154412633, "grad_norm": 0.7575067281723022, "learning_rate": 1.9905719517812406e-05, "loss": 4.0697, "step": 2100 }, { "epoch": 0.02415418232475579, "grad_norm": 0.6507667899131775, "learning_rate": 1.9903472604509556e-05, "loss": 4.0695, "step": 2150 }, { "epoch": 0.024715907495098947, "grad_norm": 0.7584772109985352, "learning_rate": 1.990122569120671e-05, "loss": 4.0884, "step": 2200 }, { "epoch": 0.025277632665442105, "grad_norm": 0.7901142835617065, "learning_rate": 1.9898978777903858e-05, "loss": 3.9693, "step": 2250 }, { "epoch": 0.025839357835785264, "grad_norm": 0.8463560342788696, "learning_rate": 1.9896731864601007e-05, "loss": 4.0007, "step": 2300 }, { "epoch": 0.026401083006128423, "grad_norm": 0.8583828806877136, "learning_rate": 1.9894484951298156e-05, "loss": 4.1057, "step": 2350 }, { "epoch": 0.026962808176471578, "grad_norm": 0.8160439133644104, "learning_rate": 1.9892238037995306e-05, "loss": 4.0635, "step": 2400 }, { "epoch": 0.027524533346814737, "grad_norm": 0.8757029175758362, "learning_rate": 1.9889991124692455e-05, "loss": 4.1076, "step": 2450 }, { "epoch": 0.028086258517157896, "grad_norm": 0.9236558079719543, "learning_rate": 1.9887744211389604e-05, "loss": 3.9999, "step": 2500 }, { "epoch": 0.028647983687501054, "grad_norm": 0.866594135761261, "learning_rate": 1.9885497298086757e-05, "loss": 3.9782, "step": 2550 }, { "epoch": 0.02920970885784421, "grad_norm": 0.9537388682365417, "learning_rate": 1.9883250384783906e-05, "loss": 4.0354, "step": 2600 }, { "epoch": 0.02977143402818737, "grad_norm": 0.7946441769599915, "learning_rate": 1.9881003471481056e-05, "loss": 4.0499, "step": 2650 }, { "epoch": 0.030333159198530527, "grad_norm": 0.7472932934761047, "learning_rate": 1.9878756558178205e-05, "loss": 3.9899, "step": 2700 }, { "epoch": 0.030894884368873686, "grad_norm": 0.7326238751411438, "learning_rate": 1.9876509644875354e-05, "loss": 4.0243, "step": 2750 }, { "epoch": 0.03145660953921684, "grad_norm": 0.9543971419334412, "learning_rate": 1.9874262731572504e-05, "loss": 4.0569, "step": 2800 }, { "epoch": 0.032018334709560003, "grad_norm": 0.7659034132957458, "learning_rate": 1.9872015818269653e-05, "loss": 3.9717, "step": 2850 }, { "epoch": 0.03258005987990316, "grad_norm": 0.8442956209182739, "learning_rate": 1.9869768904966806e-05, "loss": 4.1692, "step": 2900 }, { "epoch": 0.033141785050246314, "grad_norm": 0.7701689600944519, "learning_rate": 1.9867521991663955e-05, "loss": 3.9567, "step": 2950 }, { "epoch": 0.033703510220589476, "grad_norm": 11.462121963500977, "learning_rate": 1.9865275078361104e-05, "loss": 4.0557, "step": 3000 }, { "epoch": 0.03426523539093263, "grad_norm": 0.9673064947128296, "learning_rate": 1.9863028165058254e-05, "loss": 3.983, "step": 3050 }, { "epoch": 0.03482696056127579, "grad_norm": 0.8910813927650452, "learning_rate": 1.9860781251755403e-05, "loss": 4.055, "step": 3100 }, { "epoch": 0.03538868573161895, "grad_norm": 0.8741032481193542, "learning_rate": 1.9858534338452552e-05, "loss": 3.9814, "step": 3150 }, { "epoch": 0.035950410901962104, "grad_norm": 0.789903461933136, "learning_rate": 1.98562874251497e-05, "loss": 3.9904, "step": 3200 }, { "epoch": 0.03651213607230527, "grad_norm": 0.7773197293281555, "learning_rate": 1.9854040511846854e-05, "loss": 4.0051, "step": 3250 }, { "epoch": 0.03707386124264842, "grad_norm": 0.8257423043251038, "learning_rate": 1.9851793598544004e-05, "loss": 4.0133, "step": 3300 }, { "epoch": 0.03763558641299158, "grad_norm": 0.946387529373169, "learning_rate": 1.9849546685241153e-05, "loss": 4.0593, "step": 3350 }, { "epoch": 0.03819731158333474, "grad_norm": 0.8175517916679382, "learning_rate": 1.9847299771938302e-05, "loss": 4.007, "step": 3400 }, { "epoch": 0.038759036753677895, "grad_norm": 1.033706545829773, "learning_rate": 1.984505285863545e-05, "loss": 4.0251, "step": 3450 }, { "epoch": 0.03932076192402106, "grad_norm": 0.8714016675949097, "learning_rate": 1.98428059453326e-05, "loss": 4.0367, "step": 3500 }, { "epoch": 0.03988248709436421, "grad_norm": 1.0509799718856812, "learning_rate": 1.984055903202975e-05, "loss": 4.0318, "step": 3550 }, { "epoch": 0.04044421226470737, "grad_norm": 0.8878036737442017, "learning_rate": 1.98383121187269e-05, "loss": 4.0543, "step": 3600 }, { "epoch": 0.04100593743505053, "grad_norm": 0.7888545989990234, "learning_rate": 1.9836065205424052e-05, "loss": 3.9551, "step": 3650 }, { "epoch": 0.041567662605393685, "grad_norm": 0.8576290011405945, "learning_rate": 1.98338182921212e-05, "loss": 4.0179, "step": 3700 }, { "epoch": 0.04212938777573684, "grad_norm": 0.8358418345451355, "learning_rate": 1.983157137881835e-05, "loss": 4.0658, "step": 3750 }, { "epoch": 0.04269111294608, "grad_norm": 0.9385461807250977, "learning_rate": 1.98293244655155e-05, "loss": 4.051, "step": 3800 }, { "epoch": 0.04325283811642316, "grad_norm": 0.9508741497993469, "learning_rate": 1.982707755221265e-05, "loss": 4.0268, "step": 3850 }, { "epoch": 0.04381456328676632, "grad_norm": 0.8164873719215393, "learning_rate": 1.98248306389098e-05, "loss": 4.009, "step": 3900 }, { "epoch": 0.044376288457109475, "grad_norm": 0.986486554145813, "learning_rate": 1.9822583725606948e-05, "loss": 4.0036, "step": 3950 }, { "epoch": 0.04493801362745263, "grad_norm": 0.9176609516143799, "learning_rate": 1.98203368123041e-05, "loss": 4.0901, "step": 4000 }, { "epoch": 0.04549973879779579, "grad_norm": 0.9912000298500061, "learning_rate": 1.981808989900125e-05, "loss": 3.9454, "step": 4050 }, { "epoch": 0.04606146396813895, "grad_norm": 1.068574070930481, "learning_rate": 1.98158429856984e-05, "loss": 4.1076, "step": 4100 }, { "epoch": 0.0466231891384821, "grad_norm": 0.7717074751853943, "learning_rate": 1.981359607239555e-05, "loss": 4.0415, "step": 4150 }, { "epoch": 0.047184914308825265, "grad_norm": 0.8814721703529358, "learning_rate": 1.9811349159092698e-05, "loss": 3.9491, "step": 4200 }, { "epoch": 0.04774663947916842, "grad_norm": 0.8565316796302795, "learning_rate": 1.9809102245789847e-05, "loss": 4.0645, "step": 4250 }, { "epoch": 0.04830836464951158, "grad_norm": 0.9432961940765381, "learning_rate": 1.9806855332486997e-05, "loss": 4.0238, "step": 4300 }, { "epoch": 0.04887008981985474, "grad_norm": 1.158420205116272, "learning_rate": 1.9804653357450206e-05, "loss": 3.9297, "step": 4350 }, { "epoch": 0.04943181499019789, "grad_norm": 0.9152107834815979, "learning_rate": 1.9802406444147355e-05, "loss": 3.8737, "step": 4400 }, { "epoch": 0.049993540160541056, "grad_norm": 0.9017531871795654, "learning_rate": 1.9800159530844505e-05, "loss": 4.0031, "step": 4450 }, { "epoch": 0.05055526533088421, "grad_norm": 3.250427007675171, "learning_rate": 1.9797912617541654e-05, "loss": 3.9004, "step": 4500 }, { "epoch": 0.051116990501227366, "grad_norm": 1.2878793478012085, "learning_rate": 1.9795665704238803e-05, "loss": 3.9565, "step": 4550 }, { "epoch": 0.05167871567157053, "grad_norm": 1.0191071033477783, "learning_rate": 1.9793418790935953e-05, "loss": 3.9954, "step": 4600 }, { "epoch": 0.052240440841913684, "grad_norm": 0.8436045050621033, "learning_rate": 1.9791171877633105e-05, "loss": 4.0165, "step": 4650 }, { "epoch": 0.052802166012256846, "grad_norm": 1.0180222988128662, "learning_rate": 1.9788924964330255e-05, "loss": 3.9563, "step": 4700 }, { "epoch": 0.0533638911826, "grad_norm": 0.9640880227088928, "learning_rate": 1.9786678051027404e-05, "loss": 3.9195, "step": 4750 }, { "epoch": 0.053925616352943156, "grad_norm": 1.0035834312438965, "learning_rate": 1.9784431137724553e-05, "loss": 3.9712, "step": 4800 }, { "epoch": 0.05448734152328632, "grad_norm": 0.8614216446876526, "learning_rate": 1.9782184224421703e-05, "loss": 3.969, "step": 4850 }, { "epoch": 0.055049066693629474, "grad_norm": 1.079421877861023, "learning_rate": 1.9779937311118852e-05, "loss": 3.9919, "step": 4900 }, { "epoch": 0.055610791863972636, "grad_norm": 0.8971319198608398, "learning_rate": 1.9777690397816e-05, "loss": 4.1197, "step": 4950 }, { "epoch": 0.05617251703431579, "grad_norm": 0.8105068802833557, "learning_rate": 1.9775443484513154e-05, "loss": 3.9572, "step": 5000 }, { "epoch": 0.05673424220465895, "grad_norm": 1.0254584550857544, "learning_rate": 1.9773196571210303e-05, "loss": 3.9764, "step": 5050 }, { "epoch": 0.05729596737500211, "grad_norm": 0.8980420231819153, "learning_rate": 1.9770949657907453e-05, "loss": 3.9426, "step": 5100 }, { "epoch": 0.057857692545345264, "grad_norm": 1.1586838960647583, "learning_rate": 1.9768702744604602e-05, "loss": 3.9144, "step": 5150 }, { "epoch": 0.05841941771568842, "grad_norm": 0.858310341835022, "learning_rate": 1.976645583130175e-05, "loss": 3.9721, "step": 5200 }, { "epoch": 0.05898114288603158, "grad_norm": 0.8306008577346802, "learning_rate": 1.97642089179989e-05, "loss": 3.9499, "step": 5250 }, { "epoch": 0.05954286805637474, "grad_norm": 0.9015403985977173, "learning_rate": 1.976196200469605e-05, "loss": 3.9362, "step": 5300 }, { "epoch": 0.0601045932267179, "grad_norm": 0.7949343323707581, "learning_rate": 1.9759715091393203e-05, "loss": 3.9575, "step": 5350 }, { "epoch": 0.060666318397061055, "grad_norm": 1.0842740535736084, "learning_rate": 1.9757468178090352e-05, "loss": 3.9215, "step": 5400 }, { "epoch": 0.06122804356740421, "grad_norm": 0.8949707746505737, "learning_rate": 1.97552212647875e-05, "loss": 3.9472, "step": 5450 }, { "epoch": 0.06178976873774737, "grad_norm": 1.0328564643859863, "learning_rate": 1.975297435148465e-05, "loss": 4.0258, "step": 5500 }, { "epoch": 0.06235149390809053, "grad_norm": 0.9843789339065552, "learning_rate": 1.97507274381818e-05, "loss": 3.9164, "step": 5550 }, { "epoch": 0.06291321907843368, "grad_norm": 1.3258886337280273, "learning_rate": 1.974848052487895e-05, "loss": 4.0296, "step": 5600 }, { "epoch": 0.06347494424877684, "grad_norm": 0.9008415341377258, "learning_rate": 1.97462336115761e-05, "loss": 3.9413, "step": 5650 }, { "epoch": 0.06403666941912001, "grad_norm": 0.7957488298416138, "learning_rate": 1.974398669827325e-05, "loss": 4.0003, "step": 5700 }, { "epoch": 0.06459839458946316, "grad_norm": 0.9556325078010559, "learning_rate": 1.97417397849704e-05, "loss": 3.9204, "step": 5750 }, { "epoch": 0.06516011975980632, "grad_norm": 0.8846645951271057, "learning_rate": 1.973949287166755e-05, "loss": 3.9414, "step": 5800 }, { "epoch": 0.06572184493014947, "grad_norm": 0.9160873889923096, "learning_rate": 1.97372459583647e-05, "loss": 3.9641, "step": 5850 }, { "epoch": 0.06628357010049263, "grad_norm": 0.9347891807556152, "learning_rate": 1.973499904506185e-05, "loss": 4.0136, "step": 5900 }, { "epoch": 0.0668452952708358, "grad_norm": 0.8801557421684265, "learning_rate": 1.9732752131758998e-05, "loss": 4.0376, "step": 5950 }, { "epoch": 0.06740702044117895, "grad_norm": 0.849269688129425, "learning_rate": 1.9730505218456147e-05, "loss": 3.9612, "step": 6000 }, { "epoch": 0.06796874561152211, "grad_norm": 1.066200852394104, "learning_rate": 1.97282583051533e-05, "loss": 4.0119, "step": 6050 }, { "epoch": 0.06853047078186526, "grad_norm": 0.9853302240371704, "learning_rate": 1.972601139185045e-05, "loss": 3.913, "step": 6100 }, { "epoch": 0.06909219595220842, "grad_norm": 0.849815309047699, "learning_rate": 1.97237644785476e-05, "loss": 3.9944, "step": 6150 }, { "epoch": 0.06965392112255157, "grad_norm": 0.7928768396377563, "learning_rate": 1.9721517565244748e-05, "loss": 3.925, "step": 6200 }, { "epoch": 0.07021564629289474, "grad_norm": 1.0310895442962646, "learning_rate": 1.9719270651941897e-05, "loss": 3.9194, "step": 6250 }, { "epoch": 0.0707773714632379, "grad_norm": 1.1710147857666016, "learning_rate": 1.9717023738639046e-05, "loss": 3.9733, "step": 6300 }, { "epoch": 0.07133909663358105, "grad_norm": 0.8295989036560059, "learning_rate": 1.9714776825336196e-05, "loss": 3.9791, "step": 6350 }, { "epoch": 0.07190082180392421, "grad_norm": 0.7764708995819092, "learning_rate": 1.971252991203335e-05, "loss": 3.933, "step": 6400 }, { "epoch": 0.07246254697426736, "grad_norm": 0.9420117139816284, "learning_rate": 1.9710282998730498e-05, "loss": 3.9969, "step": 6450 }, { "epoch": 0.07302427214461053, "grad_norm": 0.956198513507843, "learning_rate": 1.9708036085427647e-05, "loss": 3.9355, "step": 6500 }, { "epoch": 0.07358599731495369, "grad_norm": 0.965861439704895, "learning_rate": 1.9705789172124796e-05, "loss": 3.8971, "step": 6550 }, { "epoch": 0.07414772248529684, "grad_norm": 1.030989408493042, "learning_rate": 1.9703542258821946e-05, "loss": 3.994, "step": 6600 }, { "epoch": 0.07470944765564, "grad_norm": 0.9188275933265686, "learning_rate": 1.9701295345519095e-05, "loss": 3.9572, "step": 6650 }, { "epoch": 0.07527117282598315, "grad_norm": 0.9203095436096191, "learning_rate": 1.9699048432216244e-05, "loss": 3.9916, "step": 6700 }, { "epoch": 0.07583289799632632, "grad_norm": 0.9306748509407043, "learning_rate": 1.9696801518913397e-05, "loss": 3.9701, "step": 6750 }, { "epoch": 0.07639462316666948, "grad_norm": 0.7836359143257141, "learning_rate": 1.9694554605610546e-05, "loss": 3.9381, "step": 6800 }, { "epoch": 0.07695634833701263, "grad_norm": 1.0823968648910522, "learning_rate": 1.9692307692307696e-05, "loss": 3.9308, "step": 6850 }, { "epoch": 0.07751807350735579, "grad_norm": 1.1342588663101196, "learning_rate": 1.9690060779004845e-05, "loss": 3.958, "step": 6900 }, { "epoch": 0.07807979867769894, "grad_norm": 0.9671859741210938, "learning_rate": 1.9687813865701994e-05, "loss": 3.9878, "step": 6950 }, { "epoch": 0.07864152384804211, "grad_norm": 0.925604522228241, "learning_rate": 1.9685566952399144e-05, "loss": 3.9283, "step": 7000 }, { "epoch": 0.07920324901838527, "grad_norm": 0.8741235136985779, "learning_rate": 1.9683320039096293e-05, "loss": 3.865, "step": 7050 }, { "epoch": 0.07976497418872842, "grad_norm": 0.8538075089454651, "learning_rate": 1.9681073125793446e-05, "loss": 3.8876, "step": 7100 }, { "epoch": 0.08032669935907158, "grad_norm": 0.948414146900177, "learning_rate": 1.967882621249059e-05, "loss": 3.9869, "step": 7150 }, { "epoch": 0.08088842452941473, "grad_norm": 1.2384617328643799, "learning_rate": 1.967657929918774e-05, "loss": 3.971, "step": 7200 }, { "epoch": 0.08145014969975789, "grad_norm": 0.9849569797515869, "learning_rate": 1.967433238588489e-05, "loss": 3.9783, "step": 7250 }, { "epoch": 0.08201187487010106, "grad_norm": 0.923664927482605, "learning_rate": 1.967208547258204e-05, "loss": 3.9348, "step": 7300 }, { "epoch": 0.08257360004044421, "grad_norm": 0.9964247345924377, "learning_rate": 1.9669838559279192e-05, "loss": 3.954, "step": 7350 }, { "epoch": 0.08313532521078737, "grad_norm": 1.320538878440857, "learning_rate": 1.966759164597634e-05, "loss": 3.9864, "step": 7400 }, { "epoch": 0.08369705038113052, "grad_norm": 0.930242657661438, "learning_rate": 1.966534473267349e-05, "loss": 3.9783, "step": 7450 }, { "epoch": 0.08425877555147368, "grad_norm": 0.9589186906814575, "learning_rate": 1.966309781937064e-05, "loss": 4.0511, "step": 7500 }, { "epoch": 0.08482050072181685, "grad_norm": 0.9792759418487549, "learning_rate": 1.966085090606779e-05, "loss": 3.9537, "step": 7550 }, { "epoch": 0.08538222589216, "grad_norm": 1.03499436378479, "learning_rate": 1.965860399276494e-05, "loss": 3.9217, "step": 7600 }, { "epoch": 0.08594395106250316, "grad_norm": 0.8685311675071716, "learning_rate": 1.9656357079462088e-05, "loss": 3.8727, "step": 7650 }, { "epoch": 0.08650567623284632, "grad_norm": 0.9089112281799316, "learning_rate": 1.9654155104425297e-05, "loss": 3.9332, "step": 7700 }, { "epoch": 0.08706740140318947, "grad_norm": 0.8686912655830383, "learning_rate": 1.9651908191122447e-05, "loss": 4.0604, "step": 7750 }, { "epoch": 0.08762912657353264, "grad_norm": 1.0852463245391846, "learning_rate": 1.9649661277819596e-05, "loss": 3.9402, "step": 7800 }, { "epoch": 0.0881908517438758, "grad_norm": 1.0768681764602661, "learning_rate": 1.9647414364516745e-05, "loss": 3.9484, "step": 7850 }, { "epoch": 0.08875257691421895, "grad_norm": 0.9111924171447754, "learning_rate": 1.9645167451213895e-05, "loss": 3.897, "step": 7900 }, { "epoch": 0.0893143020845621, "grad_norm": 1.0498020648956299, "learning_rate": 1.9642920537911044e-05, "loss": 3.9207, "step": 7950 }, { "epoch": 0.08987602725490526, "grad_norm": 0.9853581190109253, "learning_rate": 1.9640673624608197e-05, "loss": 3.9408, "step": 8000 }, { "epoch": 0.09043775242524842, "grad_norm": 1.053979754447937, "learning_rate": 1.9638426711305346e-05, "loss": 3.9448, "step": 8050 }, { "epoch": 0.09099947759559159, "grad_norm": 0.9768669009208679, "learning_rate": 1.9636179798002495e-05, "loss": 3.9075, "step": 8100 }, { "epoch": 0.09156120276593474, "grad_norm": 0.9274827241897583, "learning_rate": 1.9633932884699645e-05, "loss": 3.9015, "step": 8150 }, { "epoch": 0.0921229279362779, "grad_norm": 0.9641353487968445, "learning_rate": 1.9631685971396794e-05, "loss": 3.9437, "step": 8200 }, { "epoch": 0.09268465310662105, "grad_norm": 0.9485934972763062, "learning_rate": 1.9629439058093943e-05, "loss": 3.8863, "step": 8250 }, { "epoch": 0.0932463782769642, "grad_norm": 0.880131185054779, "learning_rate": 1.9627192144791093e-05, "loss": 3.8816, "step": 8300 }, { "epoch": 0.09380810344730738, "grad_norm": 1.05781888961792, "learning_rate": 1.9624945231488245e-05, "loss": 3.9312, "step": 8350 }, { "epoch": 0.09436982861765053, "grad_norm": 0.8662311434745789, "learning_rate": 1.9622698318185395e-05, "loss": 3.9234, "step": 8400 }, { "epoch": 0.09493155378799369, "grad_norm": 0.8740962147712708, "learning_rate": 1.9620451404882544e-05, "loss": 3.8708, "step": 8450 }, { "epoch": 0.09549327895833684, "grad_norm": 0.8588781952857971, "learning_rate": 1.9618204491579693e-05, "loss": 3.9236, "step": 8500 }, { "epoch": 0.09605500412868, "grad_norm": 1.1292428970336914, "learning_rate": 1.9615957578276843e-05, "loss": 3.938, "step": 8550 }, { "epoch": 0.09661672929902317, "grad_norm": 0.9856300354003906, "learning_rate": 1.9613710664973992e-05, "loss": 3.8818, "step": 8600 }, { "epoch": 0.09717845446936632, "grad_norm": 0.9480489492416382, "learning_rate": 1.961146375167114e-05, "loss": 3.9085, "step": 8650 }, { "epoch": 0.09774017963970948, "grad_norm": 0.9648436903953552, "learning_rate": 1.9609216838368294e-05, "loss": 3.9123, "step": 8700 }, { "epoch": 0.09830190481005263, "grad_norm": 1.042819857597351, "learning_rate": 1.9606969925065443e-05, "loss": 3.8794, "step": 8750 }, { "epoch": 0.09886362998039579, "grad_norm": 0.9484936594963074, "learning_rate": 1.9604723011762592e-05, "loss": 3.9091, "step": 8800 }, { "epoch": 0.09942535515073896, "grad_norm": 1.0150383710861206, "learning_rate": 1.9602476098459742e-05, "loss": 3.9898, "step": 8850 }, { "epoch": 0.09998708032108211, "grad_norm": 1.0116498470306396, "learning_rate": 1.960022918515689e-05, "loss": 3.9227, "step": 8900 }, { "epoch": 0.10054880549142527, "grad_norm": 1.054053783416748, "learning_rate": 1.959798227185404e-05, "loss": 3.9434, "step": 8950 }, { "epoch": 0.10111053066176842, "grad_norm": 1.049973964691162, "learning_rate": 1.959573535855119e-05, "loss": 3.9419, "step": 9000 }, { "epoch": 0.10167225583211158, "grad_norm": 0.8714705109596252, "learning_rate": 1.9593488445248342e-05, "loss": 3.8853, "step": 9050 }, { "epoch": 0.10223398100245473, "grad_norm": 0.9340712428092957, "learning_rate": 1.9591241531945492e-05, "loss": 3.9283, "step": 9100 }, { "epoch": 0.1027957061727979, "grad_norm": 0.8932335376739502, "learning_rate": 1.958899461864264e-05, "loss": 3.9738, "step": 9150 }, { "epoch": 0.10335743134314106, "grad_norm": 0.9076048731803894, "learning_rate": 1.958674770533979e-05, "loss": 3.865, "step": 9200 }, { "epoch": 0.10391915651348421, "grad_norm": 1.030010461807251, "learning_rate": 1.958450079203694e-05, "loss": 4.0057, "step": 9250 }, { "epoch": 0.10448088168382737, "grad_norm": 0.978993833065033, "learning_rate": 1.958225387873409e-05, "loss": 3.891, "step": 9300 }, { "epoch": 0.10504260685417052, "grad_norm": 1.023034691810608, "learning_rate": 1.958000696543124e-05, "loss": 3.9213, "step": 9350 }, { "epoch": 0.10560433202451369, "grad_norm": 1.2275335788726807, "learning_rate": 1.957776005212839e-05, "loss": 3.9326, "step": 9400 }, { "epoch": 0.10616605719485685, "grad_norm": 1.1626818180084229, "learning_rate": 1.957551313882554e-05, "loss": 3.958, "step": 9450 }, { "epoch": 0.1067277823652, "grad_norm": 0.953780472278595, "learning_rate": 1.957326622552269e-05, "loss": 3.9185, "step": 9500 }, { "epoch": 0.10728950753554316, "grad_norm": 1.0907609462738037, "learning_rate": 1.957101931221984e-05, "loss": 3.8956, "step": 9550 }, { "epoch": 0.10785123270588631, "grad_norm": 0.8553085923194885, "learning_rate": 1.956877239891699e-05, "loss": 3.9707, "step": 9600 }, { "epoch": 0.10841295787622948, "grad_norm": 1.1353096961975098, "learning_rate": 1.9566525485614138e-05, "loss": 3.8704, "step": 9650 }, { "epoch": 0.10897468304657264, "grad_norm": 1.0312119722366333, "learning_rate": 1.9564278572311287e-05, "loss": 3.9181, "step": 9700 }, { "epoch": 0.10953640821691579, "grad_norm": 2.1263670921325684, "learning_rate": 1.9562076597274496e-05, "loss": 3.8704, "step": 9750 }, { "epoch": 0.11009813338725895, "grad_norm": 1.009627103805542, "learning_rate": 1.9559829683971646e-05, "loss": 3.8944, "step": 9800 }, { "epoch": 0.1106598585576021, "grad_norm": 1.0444031953811646, "learning_rate": 1.9557582770668795e-05, "loss": 3.9716, "step": 9850 }, { "epoch": 0.11122158372794527, "grad_norm": 1.0081613063812256, "learning_rate": 1.9555335857365944e-05, "loss": 3.9325, "step": 9900 }, { "epoch": 0.11178330889828843, "grad_norm": 1.05756413936615, "learning_rate": 1.9553088944063094e-05, "loss": 3.9864, "step": 9950 }, { "epoch": 0.11234503406863158, "grad_norm": 1.0573816299438477, "learning_rate": 1.9550842030760243e-05, "loss": 3.9455, "step": 10000 }, { "epoch": 0.11290675923897474, "grad_norm": 1.108229637145996, "learning_rate": 1.9548595117457396e-05, "loss": 3.9542, "step": 10050 }, { "epoch": 0.1134684844093179, "grad_norm": 0.9424736499786377, "learning_rate": 1.9546348204154545e-05, "loss": 3.9492, "step": 10100 }, { "epoch": 0.11403020957966105, "grad_norm": 1.1272293329238892, "learning_rate": 1.9544101290851694e-05, "loss": 3.9152, "step": 10150 }, { "epoch": 0.11459193475000422, "grad_norm": 0.9885505437850952, "learning_rate": 1.9541854377548844e-05, "loss": 3.9176, "step": 10200 }, { "epoch": 0.11515365992034737, "grad_norm": 1.211566686630249, "learning_rate": 1.9539607464245993e-05, "loss": 3.9019, "step": 10250 }, { "epoch": 0.11571538509069053, "grad_norm": 0.9763829112052917, "learning_rate": 1.9537360550943142e-05, "loss": 3.8857, "step": 10300 }, { "epoch": 0.11627711026103368, "grad_norm": 1.1474283933639526, "learning_rate": 1.953511363764029e-05, "loss": 3.9197, "step": 10350 }, { "epoch": 0.11683883543137684, "grad_norm": 0.9613032937049866, "learning_rate": 1.9532866724337444e-05, "loss": 3.9486, "step": 10400 }, { "epoch": 0.11740056060172001, "grad_norm": 0.9909958839416504, "learning_rate": 1.9530619811034593e-05, "loss": 3.9196, "step": 10450 }, { "epoch": 0.11796228577206316, "grad_norm": 1.0783482789993286, "learning_rate": 1.9528372897731743e-05, "loss": 3.9109, "step": 10500 }, { "epoch": 0.11852401094240632, "grad_norm": 0.9794992804527283, "learning_rate": 1.9526125984428892e-05, "loss": 3.8683, "step": 10550 }, { "epoch": 0.11908573611274947, "grad_norm": 1.1354711055755615, "learning_rate": 1.952387907112604e-05, "loss": 3.8885, "step": 10600 }, { "epoch": 0.11964746128309263, "grad_norm": 1.0472307205200195, "learning_rate": 1.952163215782319e-05, "loss": 3.9329, "step": 10650 }, { "epoch": 0.1202091864534358, "grad_norm": 0.9388378262519836, "learning_rate": 1.951938524452034e-05, "loss": 3.9027, "step": 10700 }, { "epoch": 0.12077091162377895, "grad_norm": 0.9691028594970703, "learning_rate": 1.9517138331217493e-05, "loss": 3.8493, "step": 10750 }, { "epoch": 0.12133263679412211, "grad_norm": 0.8358737826347351, "learning_rate": 1.9514891417914642e-05, "loss": 3.8618, "step": 10800 }, { "epoch": 0.12189436196446526, "grad_norm": 1.1211429834365845, "learning_rate": 1.951264450461179e-05, "loss": 3.9208, "step": 10850 }, { "epoch": 0.12245608713480842, "grad_norm": 0.9543302059173584, "learning_rate": 1.951039759130894e-05, "loss": 3.8512, "step": 10900 }, { "epoch": 0.12301781230515157, "grad_norm": 1.0548458099365234, "learning_rate": 1.950815067800609e-05, "loss": 3.9096, "step": 10950 }, { "epoch": 0.12357953747549474, "grad_norm": 0.8686736226081848, "learning_rate": 1.950590376470324e-05, "loss": 3.8479, "step": 11000 }, { "epoch": 0.1241412626458379, "grad_norm": 1.0551873445510864, "learning_rate": 1.950365685140039e-05, "loss": 3.9508, "step": 11050 }, { "epoch": 0.12470298781618105, "grad_norm": 1.0495314598083496, "learning_rate": 1.950140993809754e-05, "loss": 3.9477, "step": 11100 }, { "epoch": 0.12526471298652422, "grad_norm": 0.885094165802002, "learning_rate": 1.949916302479469e-05, "loss": 3.8275, "step": 11150 }, { "epoch": 0.12582643815686737, "grad_norm": 1.0710852146148682, "learning_rate": 1.949691611149184e-05, "loss": 3.8888, "step": 11200 }, { "epoch": 0.12638816332721053, "grad_norm": 0.8449720144271851, "learning_rate": 1.949466919818899e-05, "loss": 3.8312, "step": 11250 }, { "epoch": 0.12694988849755368, "grad_norm": 0.9350094795227051, "learning_rate": 1.949242228488614e-05, "loss": 3.8534, "step": 11300 }, { "epoch": 0.12751161366789684, "grad_norm": 0.9171339869499207, "learning_rate": 1.9490175371583288e-05, "loss": 3.8905, "step": 11350 }, { "epoch": 0.12807333883824001, "grad_norm": 0.8574774861335754, "learning_rate": 1.9487928458280437e-05, "loss": 3.9054, "step": 11400 }, { "epoch": 0.12863506400858316, "grad_norm": 0.9054346084594727, "learning_rate": 1.9485681544977587e-05, "loss": 3.8086, "step": 11450 }, { "epoch": 0.12919678917892632, "grad_norm": 0.9606080055236816, "learning_rate": 1.948343463167474e-05, "loss": 3.9055, "step": 11500 }, { "epoch": 0.12975851434926947, "grad_norm": 1.0003029108047485, "learning_rate": 1.948118771837189e-05, "loss": 3.8765, "step": 11550 }, { "epoch": 0.13032023951961264, "grad_norm": 1.1126985549926758, "learning_rate": 1.9478940805069038e-05, "loss": 3.8959, "step": 11600 }, { "epoch": 0.1308819646899558, "grad_norm": 0.8803656101226807, "learning_rate": 1.9476693891766187e-05, "loss": 3.8938, "step": 11650 }, { "epoch": 0.13144368986029895, "grad_norm": 0.9052019119262695, "learning_rate": 1.9474446978463337e-05, "loss": 3.9625, "step": 11700 }, { "epoch": 0.13200541503064211, "grad_norm": 1.0454570055007935, "learning_rate": 1.9472200065160486e-05, "loss": 3.8853, "step": 11750 }, { "epoch": 0.13256714020098526, "grad_norm": 1.0287830829620361, "learning_rate": 1.9469953151857635e-05, "loss": 3.8961, "step": 11800 }, { "epoch": 0.13312886537132843, "grad_norm": 0.985406219959259, "learning_rate": 1.9467706238554788e-05, "loss": 3.8954, "step": 11850 }, { "epoch": 0.1336905905416716, "grad_norm": 1.020377278327942, "learning_rate": 1.9465459325251937e-05, "loss": 3.9196, "step": 11900 }, { "epoch": 0.13425231571201474, "grad_norm": 1.0363904237747192, "learning_rate": 1.9463212411949087e-05, "loss": 3.8959, "step": 11950 }, { "epoch": 0.1348140408823579, "grad_norm": 0.8309126496315002, "learning_rate": 1.9460965498646236e-05, "loss": 3.8518, "step": 12000 }, { "epoch": 0.13537576605270105, "grad_norm": 1.0168477296829224, "learning_rate": 1.9458718585343385e-05, "loss": 3.8139, "step": 12050 }, { "epoch": 0.13593749122304422, "grad_norm": 0.9926967024803162, "learning_rate": 1.9456471672040534e-05, "loss": 3.9504, "step": 12100 }, { "epoch": 0.13649921639338738, "grad_norm": 1.094438910484314, "learning_rate": 1.9454224758737684e-05, "loss": 3.8508, "step": 12150 }, { "epoch": 0.13706094156373053, "grad_norm": 1.0631167888641357, "learning_rate": 1.9451977845434836e-05, "loss": 3.9198, "step": 12200 }, { "epoch": 0.1376226667340737, "grad_norm": 0.9869749546051025, "learning_rate": 1.9449730932131986e-05, "loss": 3.9386, "step": 12250 }, { "epoch": 0.13818439190441684, "grad_norm": 0.9479770660400391, "learning_rate": 1.9447484018829135e-05, "loss": 3.9549, "step": 12300 }, { "epoch": 0.13874611707476, "grad_norm": 1.067071795463562, "learning_rate": 1.9445237105526284e-05, "loss": 3.9053, "step": 12350 }, { "epoch": 0.13930784224510315, "grad_norm": 1.1760696172714233, "learning_rate": 1.9442990192223434e-05, "loss": 3.984, "step": 12400 }, { "epoch": 0.13986956741544632, "grad_norm": 0.9269402623176575, "learning_rate": 1.9440743278920583e-05, "loss": 3.9179, "step": 12450 }, { "epoch": 0.14043129258578949, "grad_norm": 0.839056134223938, "learning_rate": 1.9438496365617732e-05, "loss": 3.936, "step": 12500 }, { "epoch": 0.14099301775613263, "grad_norm": 0.846478283405304, "learning_rate": 1.9436249452314885e-05, "loss": 3.8743, "step": 12550 }, { "epoch": 0.1415547429264758, "grad_norm": 1.2440580129623413, "learning_rate": 1.9434002539012034e-05, "loss": 3.9124, "step": 12600 }, { "epoch": 0.14211646809681894, "grad_norm": 0.9950169920921326, "learning_rate": 1.9431755625709184e-05, "loss": 3.8883, "step": 12650 }, { "epoch": 0.1426781932671621, "grad_norm": 0.9739125967025757, "learning_rate": 1.9429508712406333e-05, "loss": 3.8438, "step": 12700 }, { "epoch": 0.14323991843750528, "grad_norm": 0.9917782545089722, "learning_rate": 1.9427261799103482e-05, "loss": 3.8871, "step": 12750 }, { "epoch": 0.14380164360784842, "grad_norm": 0.9435528516769409, "learning_rate": 1.942501488580063e-05, "loss": 3.8766, "step": 12800 }, { "epoch": 0.1443633687781916, "grad_norm": 1.1224132776260376, "learning_rate": 1.942276797249778e-05, "loss": 3.9131, "step": 12850 }, { "epoch": 0.14492509394853473, "grad_norm": 0.9452431201934814, "learning_rate": 1.9420521059194934e-05, "loss": 3.8974, "step": 12900 }, { "epoch": 0.1454868191188779, "grad_norm": 0.9240500330924988, "learning_rate": 1.9418274145892083e-05, "loss": 3.8462, "step": 12950 }, { "epoch": 0.14604854428922107, "grad_norm": 1.0114437341690063, "learning_rate": 1.9416027232589232e-05, "loss": 3.8693, "step": 13000 }, { "epoch": 0.1466102694595642, "grad_norm": 0.9811971783638, "learning_rate": 1.941378031928638e-05, "loss": 3.9183, "step": 13050 }, { "epoch": 0.14717199462990738, "grad_norm": 0.8882970213890076, "learning_rate": 1.941153340598353e-05, "loss": 3.9643, "step": 13100 }, { "epoch": 0.14773371980025052, "grad_norm": 0.9701553583145142, "learning_rate": 1.940928649268068e-05, "loss": 3.8478, "step": 13150 }, { "epoch": 0.1482954449705937, "grad_norm": 1.0480194091796875, "learning_rate": 1.940703957937783e-05, "loss": 3.8611, "step": 13200 }, { "epoch": 0.14885717014093686, "grad_norm": 1.0310245752334595, "learning_rate": 1.9404792666074982e-05, "loss": 3.9213, "step": 13250 }, { "epoch": 0.14941889531128, "grad_norm": 1.2251856327056885, "learning_rate": 1.940254575277213e-05, "loss": 3.8666, "step": 13300 }, { "epoch": 0.14998062048162317, "grad_norm": 1.121207356452942, "learning_rate": 1.940029883946928e-05, "loss": 3.8986, "step": 13350 }, { "epoch": 0.1505423456519663, "grad_norm": 1.0388009548187256, "learning_rate": 1.939805192616643e-05, "loss": 3.8735, "step": 13400 }, { "epoch": 0.15110407082230948, "grad_norm": 1.0072270631790161, "learning_rate": 1.939580501286358e-05, "loss": 3.8796, "step": 13450 }, { "epoch": 0.15166579599265265, "grad_norm": 1.1749372482299805, "learning_rate": 1.939355809956073e-05, "loss": 3.9541, "step": 13500 }, { "epoch": 0.1522275211629958, "grad_norm": 0.996511697769165, "learning_rate": 1.9391311186257878e-05, "loss": 3.9101, "step": 13550 }, { "epoch": 0.15278924633333896, "grad_norm": 0.8689504265785217, "learning_rate": 1.938906427295503e-05, "loss": 3.8309, "step": 13600 }, { "epoch": 0.1533509715036821, "grad_norm": 0.8855342268943787, "learning_rate": 1.938681735965218e-05, "loss": 3.9087, "step": 13650 }, { "epoch": 0.15391269667402527, "grad_norm": 0.8916892409324646, "learning_rate": 1.938457044634933e-05, "loss": 3.8328, "step": 13700 }, { "epoch": 0.15447442184436844, "grad_norm": 0.9709575772285461, "learning_rate": 1.938232353304648e-05, "loss": 3.9131, "step": 13750 }, { "epoch": 0.15503614701471158, "grad_norm": 1.1655782461166382, "learning_rate": 1.9380121558009685e-05, "loss": 3.9148, "step": 13800 }, { "epoch": 0.15559787218505475, "grad_norm": 1.018314003944397, "learning_rate": 1.9377874644706834e-05, "loss": 3.9141, "step": 13850 }, { "epoch": 0.1561595973553979, "grad_norm": 0.9596139788627625, "learning_rate": 1.9375627731403987e-05, "loss": 3.9603, "step": 13900 }, { "epoch": 0.15672132252574106, "grad_norm": 0.993695080280304, "learning_rate": 1.9373380818101136e-05, "loss": 3.8694, "step": 13950 }, { "epoch": 0.15728304769608423, "grad_norm": 0.9984117746353149, "learning_rate": 1.9371133904798285e-05, "loss": 3.9069, "step": 14000 }, { "epoch": 0.15784477286642737, "grad_norm": 1.0261962413787842, "learning_rate": 1.9368886991495435e-05, "loss": 3.7683, "step": 14050 }, { "epoch": 0.15840649803677054, "grad_norm": 0.886608362197876, "learning_rate": 1.9366640078192584e-05, "loss": 3.8899, "step": 14100 }, { "epoch": 0.15896822320711368, "grad_norm": 0.9339202046394348, "learning_rate": 1.9364393164889733e-05, "loss": 3.7592, "step": 14150 }, { "epoch": 0.15952994837745685, "grad_norm": 0.919339120388031, "learning_rate": 1.9362146251586883e-05, "loss": 3.9459, "step": 14200 }, { "epoch": 0.1600916735478, "grad_norm": 1.0983084440231323, "learning_rate": 1.9359899338284035e-05, "loss": 3.9416, "step": 14250 }, { "epoch": 0.16065339871814316, "grad_norm": 0.9921066164970398, "learning_rate": 1.9357652424981185e-05, "loss": 3.9033, "step": 14300 }, { "epoch": 0.16121512388848633, "grad_norm": 0.9689832329750061, "learning_rate": 1.9355405511678334e-05, "loss": 3.9637, "step": 14350 }, { "epoch": 0.16177684905882947, "grad_norm": 1.1760526895523071, "learning_rate": 1.9353158598375483e-05, "loss": 3.8931, "step": 14400 }, { "epoch": 0.16233857422917264, "grad_norm": 2.066866636276245, "learning_rate": 1.9350911685072633e-05, "loss": 3.9355, "step": 14450 }, { "epoch": 0.16290029939951578, "grad_norm": 1.0954084396362305, "learning_rate": 1.9348664771769782e-05, "loss": 3.9496, "step": 14500 }, { "epoch": 0.16346202456985895, "grad_norm": 1.043190598487854, "learning_rate": 1.934641785846693e-05, "loss": 3.8599, "step": 14550 }, { "epoch": 0.16402374974020212, "grad_norm": 1.1450411081314087, "learning_rate": 1.9344170945164084e-05, "loss": 3.9175, "step": 14600 }, { "epoch": 0.16458547491054526, "grad_norm": 1.3400977849960327, "learning_rate": 1.9341924031861233e-05, "loss": 3.7876, "step": 14650 }, { "epoch": 0.16514720008088843, "grad_norm": 1.030008316040039, "learning_rate": 1.9339677118558383e-05, "loss": 3.8666, "step": 14700 }, { "epoch": 0.16570892525123157, "grad_norm": 0.9812994599342346, "learning_rate": 1.9337430205255532e-05, "loss": 3.8993, "step": 14750 }, { "epoch": 0.16627065042157474, "grad_norm": 0.9178810119628906, "learning_rate": 1.933518329195268e-05, "loss": 3.9666, "step": 14800 }, { "epoch": 0.1668323755919179, "grad_norm": 1.3065404891967773, "learning_rate": 1.933293637864983e-05, "loss": 3.8952, "step": 14850 }, { "epoch": 0.16739410076226105, "grad_norm": 1.1173884868621826, "learning_rate": 1.933068946534698e-05, "loss": 3.8152, "step": 14900 }, { "epoch": 0.16795582593260422, "grad_norm": 1.0459538698196411, "learning_rate": 1.9328442552044133e-05, "loss": 3.9137, "step": 14950 }, { "epoch": 0.16851755110294736, "grad_norm": 1.1197466850280762, "learning_rate": 1.9326195638741282e-05, "loss": 3.8563, "step": 15000 }, { "epoch": 0.16907927627329053, "grad_norm": 1.0688142776489258, "learning_rate": 1.932394872543843e-05, "loss": 3.8626, "step": 15050 }, { "epoch": 0.1696410014436337, "grad_norm": 0.9067109823226929, "learning_rate": 1.932170181213558e-05, "loss": 3.8535, "step": 15100 }, { "epoch": 0.17020272661397684, "grad_norm": 1.0777970552444458, "learning_rate": 1.931945489883273e-05, "loss": 3.9188, "step": 15150 }, { "epoch": 0.17076445178432, "grad_norm": 0.9435402750968933, "learning_rate": 1.931720798552988e-05, "loss": 3.8346, "step": 15200 }, { "epoch": 0.17132617695466315, "grad_norm": 1.010043978691101, "learning_rate": 1.931496107222703e-05, "loss": 3.8905, "step": 15250 }, { "epoch": 0.17188790212500632, "grad_norm": 0.9614844918251038, "learning_rate": 1.931271415892418e-05, "loss": 3.8939, "step": 15300 }, { "epoch": 0.1724496272953495, "grad_norm": 1.1249394416809082, "learning_rate": 1.931046724562133e-05, "loss": 3.9328, "step": 15350 }, { "epoch": 0.17301135246569263, "grad_norm": 1.04349684715271, "learning_rate": 1.930822033231848e-05, "loss": 3.8523, "step": 15400 }, { "epoch": 0.1735730776360358, "grad_norm": 9.898025512695312, "learning_rate": 1.930597341901563e-05, "loss": 3.8503, "step": 15450 }, { "epoch": 0.17413480280637894, "grad_norm": 0.9678813815116882, "learning_rate": 1.930372650571278e-05, "loss": 3.8466, "step": 15500 }, { "epoch": 0.1746965279767221, "grad_norm": 0.9452058672904968, "learning_rate": 1.9301479592409928e-05, "loss": 3.8458, "step": 15550 }, { "epoch": 0.17525825314706528, "grad_norm": 1.0991216897964478, "learning_rate": 1.9299232679107077e-05, "loss": 3.9693, "step": 15600 }, { "epoch": 0.17581997831740842, "grad_norm": 1.1074953079223633, "learning_rate": 1.9296985765804226e-05, "loss": 3.8205, "step": 15650 }, { "epoch": 0.1763817034877516, "grad_norm": 0.8646476864814758, "learning_rate": 1.929473885250138e-05, "loss": 3.8831, "step": 15700 }, { "epoch": 0.17694342865809473, "grad_norm": 1.0361870527267456, "learning_rate": 1.929249193919853e-05, "loss": 3.985, "step": 15750 }, { "epoch": 0.1775051538284379, "grad_norm": 1.405374526977539, "learning_rate": 1.9290245025895678e-05, "loss": 3.8676, "step": 15800 }, { "epoch": 0.17806687899878107, "grad_norm": 1.0663036108016968, "learning_rate": 1.9287998112592827e-05, "loss": 3.9483, "step": 15850 }, { "epoch": 0.1786286041691242, "grad_norm": 1.0984610319137573, "learning_rate": 1.9285751199289976e-05, "loss": 3.7979, "step": 15900 }, { "epoch": 0.17919032933946738, "grad_norm": 1.0702370405197144, "learning_rate": 1.9283504285987126e-05, "loss": 3.8597, "step": 15950 }, { "epoch": 0.17975205450981052, "grad_norm": 1.00674307346344, "learning_rate": 1.9281257372684275e-05, "loss": 3.9925, "step": 16000 }, { "epoch": 0.1803137796801537, "grad_norm": 1.7243638038635254, "learning_rate": 1.9279010459381428e-05, "loss": 3.9243, "step": 16050 }, { "epoch": 0.18087550485049683, "grad_norm": 0.9632601141929626, "learning_rate": 1.9276763546078577e-05, "loss": 3.9168, "step": 16100 }, { "epoch": 0.18143723002084, "grad_norm": 1.0414319038391113, "learning_rate": 1.9274516632775726e-05, "loss": 3.917, "step": 16150 }, { "epoch": 0.18199895519118317, "grad_norm": 1.0618311166763306, "learning_rate": 1.9272314657738932e-05, "loss": 3.8461, "step": 16200 }, { "epoch": 0.1825606803615263, "grad_norm": 0.9754511117935181, "learning_rate": 1.927006774443608e-05, "loss": 3.8397, "step": 16250 }, { "epoch": 0.18312240553186948, "grad_norm": 1.0348036289215088, "learning_rate": 1.9267820831133234e-05, "loss": 3.9151, "step": 16300 }, { "epoch": 0.18368413070221262, "grad_norm": 1.0387177467346191, "learning_rate": 1.9265573917830384e-05, "loss": 3.8342, "step": 16350 }, { "epoch": 0.1842458558725558, "grad_norm": 1.0977225303649902, "learning_rate": 1.9263327004527533e-05, "loss": 3.9074, "step": 16400 }, { "epoch": 0.18480758104289896, "grad_norm": 0.9499973058700562, "learning_rate": 1.9261080091224682e-05, "loss": 3.9021, "step": 16450 }, { "epoch": 0.1853693062132421, "grad_norm": 0.9755128026008606, "learning_rate": 1.925883317792183e-05, "loss": 3.8882, "step": 16500 }, { "epoch": 0.18593103138358527, "grad_norm": 0.958068311214447, "learning_rate": 1.925658626461898e-05, "loss": 3.916, "step": 16550 }, { "epoch": 0.1864927565539284, "grad_norm": 1.119287133216858, "learning_rate": 1.925433935131613e-05, "loss": 3.8832, "step": 16600 }, { "epoch": 0.18705448172427158, "grad_norm": 1.097261667251587, "learning_rate": 1.925209243801328e-05, "loss": 3.8584, "step": 16650 }, { "epoch": 0.18761620689461475, "grad_norm": 1.1982165575027466, "learning_rate": 1.9249845524710432e-05, "loss": 3.7963, "step": 16700 }, { "epoch": 0.1881779320649579, "grad_norm": 1.0916881561279297, "learning_rate": 1.924759861140758e-05, "loss": 3.9541, "step": 16750 }, { "epoch": 0.18873965723530106, "grad_norm": 1.1338039636611938, "learning_rate": 1.924535169810473e-05, "loss": 3.947, "step": 16800 }, { "epoch": 0.1893013824056442, "grad_norm": 1.116104006767273, "learning_rate": 1.924310478480188e-05, "loss": 3.912, "step": 16850 }, { "epoch": 0.18986310757598737, "grad_norm": 1.2745022773742676, "learning_rate": 1.924085787149903e-05, "loss": 3.8085, "step": 16900 }, { "epoch": 0.19042483274633054, "grad_norm": 1.0067696571350098, "learning_rate": 1.923861095819618e-05, "loss": 3.8644, "step": 16950 }, { "epoch": 0.19098655791667368, "grad_norm": 0.9528264403343201, "learning_rate": 1.9236364044893328e-05, "loss": 3.8716, "step": 17000 }, { "epoch": 0.19154828308701685, "grad_norm": 0.9192308783531189, "learning_rate": 1.923411713159048e-05, "loss": 3.8261, "step": 17050 }, { "epoch": 0.19211000825736, "grad_norm": 0.8801952004432678, "learning_rate": 1.923187021828763e-05, "loss": 3.9287, "step": 17100 }, { "epoch": 0.19267173342770316, "grad_norm": 1.1006280183792114, "learning_rate": 1.922962330498478e-05, "loss": 3.8829, "step": 17150 }, { "epoch": 0.19323345859804633, "grad_norm": 0.9863724112510681, "learning_rate": 1.922737639168193e-05, "loss": 3.8876, "step": 17200 }, { "epoch": 0.19379518376838947, "grad_norm": 0.9066800475120544, "learning_rate": 1.9225129478379078e-05, "loss": 3.7709, "step": 17250 }, { "epoch": 0.19435690893873264, "grad_norm": 1.0071481466293335, "learning_rate": 1.9222882565076227e-05, "loss": 3.8305, "step": 17300 }, { "epoch": 0.19491863410907578, "grad_norm": 0.962822437286377, "learning_rate": 1.9220635651773377e-05, "loss": 3.8863, "step": 17350 }, { "epoch": 0.19548035927941895, "grad_norm": 0.9313477873802185, "learning_rate": 1.921838873847053e-05, "loss": 3.8798, "step": 17400 }, { "epoch": 0.19604208444976212, "grad_norm": 0.8907088041305542, "learning_rate": 1.921614182516768e-05, "loss": 3.8138, "step": 17450 }, { "epoch": 0.19660380962010526, "grad_norm": 0.8724405765533447, "learning_rate": 1.9213894911864828e-05, "loss": 3.8547, "step": 17500 }, { "epoch": 0.19716553479044843, "grad_norm": 1.0886657238006592, "learning_rate": 1.9211647998561977e-05, "loss": 3.8818, "step": 17550 }, { "epoch": 0.19772725996079157, "grad_norm": 1.1140875816345215, "learning_rate": 1.9209446023525183e-05, "loss": 3.8971, "step": 17600 }, { "epoch": 0.19828898513113474, "grad_norm": 1.0241341590881348, "learning_rate": 1.9207199110222333e-05, "loss": 3.8416, "step": 17650 }, { "epoch": 0.1988507103014779, "grad_norm": 0.9751248359680176, "learning_rate": 1.9204952196919485e-05, "loss": 3.7553, "step": 17700 }, { "epoch": 0.19941243547182105, "grad_norm": 1.0477747917175293, "learning_rate": 1.9202705283616635e-05, "loss": 3.8363, "step": 17750 }, { "epoch": 0.19997416064216422, "grad_norm": 1.0788570642471313, "learning_rate": 1.9200458370313784e-05, "loss": 3.7763, "step": 17800 }, { "epoch": 0.20053588581250736, "grad_norm": 1.0035017728805542, "learning_rate": 1.9198211457010933e-05, "loss": 3.8727, "step": 17850 }, { "epoch": 0.20109761098285053, "grad_norm": 0.9635750651359558, "learning_rate": 1.9195964543708083e-05, "loss": 3.8129, "step": 17900 }, { "epoch": 0.2016593361531937, "grad_norm": 1.2292413711547852, "learning_rate": 1.9193717630405232e-05, "loss": 3.8897, "step": 17950 }, { "epoch": 0.20222106132353684, "grad_norm": 1.3051143884658813, "learning_rate": 1.919147071710238e-05, "loss": 3.8518, "step": 18000 }, { "epoch": 0.20278278649388, "grad_norm": 1.1229668855667114, "learning_rate": 1.9189223803799534e-05, "loss": 3.8051, "step": 18050 }, { "epoch": 0.20334451166422315, "grad_norm": 0.9775665402412415, "learning_rate": 1.9186976890496683e-05, "loss": 3.879, "step": 18100 }, { "epoch": 0.20390623683456632, "grad_norm": 1.1217602491378784, "learning_rate": 1.9184729977193833e-05, "loss": 3.8303, "step": 18150 }, { "epoch": 0.20446796200490946, "grad_norm": 1.9044723510742188, "learning_rate": 1.9182483063890982e-05, "loss": 3.872, "step": 18200 }, { "epoch": 0.20502968717525263, "grad_norm": 0.9213973879814148, "learning_rate": 1.918023615058813e-05, "loss": 3.8859, "step": 18250 }, { "epoch": 0.2055914123455958, "grad_norm": 1.2292492389678955, "learning_rate": 1.917798923728528e-05, "loss": 3.8682, "step": 18300 }, { "epoch": 0.20615313751593894, "grad_norm": 0.9231402277946472, "learning_rate": 1.917574232398243e-05, "loss": 3.833, "step": 18350 }, { "epoch": 0.2067148626862821, "grad_norm": 1.0987199544906616, "learning_rate": 1.9173495410679583e-05, "loss": 3.8887, "step": 18400 }, { "epoch": 0.20727658785662526, "grad_norm": 1.301283359527588, "learning_rate": 1.9171248497376732e-05, "loss": 3.962, "step": 18450 }, { "epoch": 0.20783831302696842, "grad_norm": 0.8758549690246582, "learning_rate": 1.916900158407388e-05, "loss": 3.8472, "step": 18500 }, { "epoch": 0.2084000381973116, "grad_norm": 1.331650733947754, "learning_rate": 1.916675467077103e-05, "loss": 3.9557, "step": 18550 }, { "epoch": 0.20896176336765473, "grad_norm": 1.001512885093689, "learning_rate": 1.916450775746818e-05, "loss": 3.9078, "step": 18600 }, { "epoch": 0.2095234885379979, "grad_norm": 1.0210082530975342, "learning_rate": 1.916226084416533e-05, "loss": 3.7859, "step": 18650 }, { "epoch": 0.21008521370834105, "grad_norm": 1.0081088542938232, "learning_rate": 1.916001393086248e-05, "loss": 3.9527, "step": 18700 }, { "epoch": 0.21064693887868421, "grad_norm": 1.1623823642730713, "learning_rate": 1.915776701755963e-05, "loss": 3.8055, "step": 18750 }, { "epoch": 0.21120866404902738, "grad_norm": 1.0312272310256958, "learning_rate": 1.915552010425678e-05, "loss": 3.8481, "step": 18800 }, { "epoch": 0.21177038921937053, "grad_norm": 1.018204927444458, "learning_rate": 1.915327319095393e-05, "loss": 3.8106, "step": 18850 }, { "epoch": 0.2123321143897137, "grad_norm": 1.005672574043274, "learning_rate": 1.915102627765108e-05, "loss": 3.839, "step": 18900 }, { "epoch": 0.21289383956005684, "grad_norm": 0.9410429000854492, "learning_rate": 1.914877936434823e-05, "loss": 3.8118, "step": 18950 }, { "epoch": 0.2134555647304, "grad_norm": 1.0433132648468018, "learning_rate": 1.9146532451045378e-05, "loss": 3.8394, "step": 19000 }, { "epoch": 0.21401728990074317, "grad_norm": 1.1158668994903564, "learning_rate": 1.9144285537742527e-05, "loss": 3.8788, "step": 19050 }, { "epoch": 0.21457901507108632, "grad_norm": 0.9531320929527283, "learning_rate": 1.914203862443968e-05, "loss": 3.9105, "step": 19100 }, { "epoch": 0.21514074024142948, "grad_norm": 0.8815164566040039, "learning_rate": 1.913979171113683e-05, "loss": 3.8621, "step": 19150 }, { "epoch": 0.21570246541177263, "grad_norm": 1.0627539157867432, "learning_rate": 1.913754479783398e-05, "loss": 3.8282, "step": 19200 }, { "epoch": 0.2162641905821158, "grad_norm": 0.9545466303825378, "learning_rate": 1.9135297884531128e-05, "loss": 3.8455, "step": 19250 }, { "epoch": 0.21682591575245896, "grad_norm": 1.0652786493301392, "learning_rate": 1.9133050971228277e-05, "loss": 3.8802, "step": 19300 }, { "epoch": 0.2173876409228021, "grad_norm": 1.087463617324829, "learning_rate": 1.9130804057925426e-05, "loss": 3.7916, "step": 19350 }, { "epoch": 0.21794936609314527, "grad_norm": 0.9182972311973572, "learning_rate": 1.9128557144622576e-05, "loss": 3.8506, "step": 19400 }, { "epoch": 0.21851109126348842, "grad_norm": 0.9740565419197083, "learning_rate": 1.912631023131973e-05, "loss": 3.8396, "step": 19450 }, { "epoch": 0.21907281643383159, "grad_norm": 1.2423253059387207, "learning_rate": 1.9124063318016878e-05, "loss": 3.8392, "step": 19500 }, { "epoch": 0.21963454160417475, "grad_norm": 0.9749589562416077, "learning_rate": 1.9121816404714027e-05, "loss": 3.8517, "step": 19550 }, { "epoch": 0.2201962667745179, "grad_norm": 1.0176291465759277, "learning_rate": 1.9119569491411176e-05, "loss": 3.9056, "step": 19600 }, { "epoch": 0.22075799194486107, "grad_norm": 1.1581487655639648, "learning_rate": 1.9117322578108326e-05, "loss": 3.8801, "step": 19650 }, { "epoch": 0.2213197171152042, "grad_norm": 0.9785657525062561, "learning_rate": 1.9115075664805475e-05, "loss": 3.7482, "step": 19700 }, { "epoch": 0.22188144228554738, "grad_norm": 1.1080032587051392, "learning_rate": 1.9112828751502624e-05, "loss": 3.8123, "step": 19750 }, { "epoch": 0.22244316745589054, "grad_norm": 0.9258034825325012, "learning_rate": 1.9110581838199777e-05, "loss": 3.8657, "step": 19800 }, { "epoch": 0.2230048926262337, "grad_norm": 0.9505318999290466, "learning_rate": 1.9108334924896926e-05, "loss": 3.898, "step": 19850 }, { "epoch": 0.22356661779657686, "grad_norm": 1.0489028692245483, "learning_rate": 1.9106088011594076e-05, "loss": 3.7987, "step": 19900 }, { "epoch": 0.22412834296692, "grad_norm": 1.0717949867248535, "learning_rate": 1.9103841098291225e-05, "loss": 3.9085, "step": 19950 }, { "epoch": 0.22469006813726317, "grad_norm": 0.9286379814147949, "learning_rate": 1.9101594184988374e-05, "loss": 3.845, "step": 20000 }, { "epoch": 0.2252517933076063, "grad_norm": 1.0841072797775269, "learning_rate": 1.9099347271685524e-05, "loss": 3.881, "step": 20050 }, { "epoch": 0.22581351847794948, "grad_norm": 0.8188834190368652, "learning_rate": 1.9097100358382673e-05, "loss": 3.9063, "step": 20100 }, { "epoch": 0.22637524364829265, "grad_norm": 0.9557147026062012, "learning_rate": 1.9094853445079826e-05, "loss": 3.8553, "step": 20150 }, { "epoch": 0.2269369688186358, "grad_norm": 1.0040074586868286, "learning_rate": 1.9092606531776975e-05, "loss": 3.8604, "step": 20200 }, { "epoch": 0.22749869398897896, "grad_norm": 0.9978244304656982, "learning_rate": 1.9090359618474124e-05, "loss": 3.9428, "step": 20250 }, { "epoch": 0.2280604191593221, "grad_norm": 0.998877227306366, "learning_rate": 1.9088112705171274e-05, "loss": 3.8784, "step": 20300 }, { "epoch": 0.22862214432966527, "grad_norm": 1.1113442182540894, "learning_rate": 1.9085865791868423e-05, "loss": 3.8211, "step": 20350 }, { "epoch": 0.22918386950000844, "grad_norm": 0.9248872399330139, "learning_rate": 1.9083618878565572e-05, "loss": 3.8687, "step": 20400 }, { "epoch": 0.22974559467035158, "grad_norm": 1.293752670288086, "learning_rate": 1.908137196526272e-05, "loss": 3.8408, "step": 20450 }, { "epoch": 0.23030731984069475, "grad_norm": 1.0347455739974976, "learning_rate": 1.9079125051959874e-05, "loss": 3.854, "step": 20500 }, { "epoch": 0.2308690450110379, "grad_norm": 1.0671968460083008, "learning_rate": 1.9076878138657023e-05, "loss": 3.8493, "step": 20550 }, { "epoch": 0.23143077018138106, "grad_norm": 1.0472993850708008, "learning_rate": 1.9074631225354173e-05, "loss": 3.7963, "step": 20600 }, { "epoch": 0.23199249535172423, "grad_norm": 1.1053742170333862, "learning_rate": 1.9072384312051322e-05, "loss": 3.8703, "step": 20650 }, { "epoch": 0.23255422052206737, "grad_norm": 0.9656935930252075, "learning_rate": 1.907013739874847e-05, "loss": 3.8551, "step": 20700 }, { "epoch": 0.23311594569241054, "grad_norm": 1.0903059244155884, "learning_rate": 1.906789048544562e-05, "loss": 3.8798, "step": 20750 }, { "epoch": 0.23367767086275368, "grad_norm": 1.178858995437622, "learning_rate": 1.906564357214277e-05, "loss": 3.8763, "step": 20800 }, { "epoch": 0.23423939603309685, "grad_norm": 1.2152944803237915, "learning_rate": 1.906339665883992e-05, "loss": 3.9042, "step": 20850 }, { "epoch": 0.23480112120344002, "grad_norm": 1.1343653202056885, "learning_rate": 1.9061149745537072e-05, "loss": 3.8355, "step": 20900 }, { "epoch": 0.23536284637378316, "grad_norm": 1.0135072469711304, "learning_rate": 1.905890283223422e-05, "loss": 3.886, "step": 20950 }, { "epoch": 0.23592457154412633, "grad_norm": 1.7519757747650146, "learning_rate": 1.905665591893137e-05, "loss": 3.7514, "step": 21000 }, { "epoch": 0.23648629671446947, "grad_norm": 1.1740312576293945, "learning_rate": 1.905440900562852e-05, "loss": 3.8605, "step": 21050 }, { "epoch": 0.23704802188481264, "grad_norm": 1.0153892040252686, "learning_rate": 1.905216209232567e-05, "loss": 3.8826, "step": 21100 }, { "epoch": 0.2376097470551558, "grad_norm": 1.056986927986145, "learning_rate": 1.904991517902282e-05, "loss": 3.7912, "step": 21150 }, { "epoch": 0.23817147222549895, "grad_norm": 1.4163366556167603, "learning_rate": 1.9047668265719968e-05, "loss": 3.8459, "step": 21200 }, { "epoch": 0.23873319739584212, "grad_norm": 1.0286545753479004, "learning_rate": 1.904542135241712e-05, "loss": 3.8383, "step": 21250 }, { "epoch": 0.23929492256618526, "grad_norm": 1.12931227684021, "learning_rate": 1.904317443911427e-05, "loss": 3.8532, "step": 21300 }, { "epoch": 0.23985664773652843, "grad_norm": 0.9231151938438416, "learning_rate": 1.904092752581142e-05, "loss": 3.8494, "step": 21350 }, { "epoch": 0.2404183729068716, "grad_norm": 1.382602572441101, "learning_rate": 1.903868061250857e-05, "loss": 3.851, "step": 21400 }, { "epoch": 0.24098009807721474, "grad_norm": 0.9720520973205566, "learning_rate": 1.9036433699205718e-05, "loss": 3.8763, "step": 21450 }, { "epoch": 0.2415418232475579, "grad_norm": 1.3178774118423462, "learning_rate": 1.9034186785902867e-05, "loss": 3.8769, "step": 21500 }, { "epoch": 0.24210354841790105, "grad_norm": 1.0083391666412354, "learning_rate": 1.9031939872600017e-05, "loss": 3.8123, "step": 21550 }, { "epoch": 0.24266527358824422, "grad_norm": 0.8887113928794861, "learning_rate": 1.902969295929717e-05, "loss": 3.8488, "step": 21600 }, { "epoch": 0.2432269987585874, "grad_norm": 1.0424660444259644, "learning_rate": 1.902744604599432e-05, "loss": 3.922, "step": 21650 }, { "epoch": 0.24378872392893053, "grad_norm": 1.0282403230667114, "learning_rate": 1.9025199132691468e-05, "loss": 3.7764, "step": 21700 }, { "epoch": 0.2443504490992737, "grad_norm": 0.9313274621963501, "learning_rate": 1.9022952219388617e-05, "loss": 3.8267, "step": 21750 }, { "epoch": 0.24491217426961684, "grad_norm": 0.9676574468612671, "learning_rate": 1.9020705306085767e-05, "loss": 3.8855, "step": 21800 }, { "epoch": 0.24547389943996, "grad_norm": 0.9758104681968689, "learning_rate": 1.9018503331048972e-05, "loss": 3.8527, "step": 21850 }, { "epoch": 0.24603562461030315, "grad_norm": 1.0635255575180054, "learning_rate": 1.9016256417746122e-05, "loss": 3.8487, "step": 21900 }, { "epoch": 0.24659734978064632, "grad_norm": 1.0568088293075562, "learning_rate": 1.901400950444327e-05, "loss": 3.826, "step": 21950 }, { "epoch": 0.2471590749509895, "grad_norm": 0.9051997065544128, "learning_rate": 1.901176259114042e-05, "loss": 3.8514, "step": 22000 }, { "epoch": 0.24772080012133263, "grad_norm": 1.0521992444992065, "learning_rate": 1.900951567783757e-05, "loss": 3.8023, "step": 22050 }, { "epoch": 0.2482825252916758, "grad_norm": 1.194155216217041, "learning_rate": 1.9007268764534722e-05, "loss": 3.751, "step": 22100 }, { "epoch": 0.24884425046201894, "grad_norm": 1.0270308256149292, "learning_rate": 1.9005021851231872e-05, "loss": 3.8069, "step": 22150 }, { "epoch": 0.2494059756323621, "grad_norm": 1.0364912748336792, "learning_rate": 1.900277493792902e-05, "loss": 3.8077, "step": 22200 }, { "epoch": 0.24996770080270528, "grad_norm": 1.1937390565872192, "learning_rate": 1.900052802462617e-05, "loss": 3.8628, "step": 22250 }, { "epoch": 0.25052942597304845, "grad_norm": 1.036866307258606, "learning_rate": 1.899828111132332e-05, "loss": 3.8209, "step": 22300 }, { "epoch": 0.2510911511433916, "grad_norm": 1.014030933380127, "learning_rate": 1.899603419802047e-05, "loss": 3.8304, "step": 22350 }, { "epoch": 0.25165287631373473, "grad_norm": 1.2910370826721191, "learning_rate": 1.899378728471762e-05, "loss": 3.8726, "step": 22400 }, { "epoch": 0.2522146014840779, "grad_norm": 0.9999923706054688, "learning_rate": 1.899154037141477e-05, "loss": 3.809, "step": 22450 }, { "epoch": 0.25277632665442107, "grad_norm": 1.0789824724197388, "learning_rate": 1.898929345811192e-05, "loss": 3.8186, "step": 22500 }, { "epoch": 0.2533380518247642, "grad_norm": 0.976632833480835, "learning_rate": 1.898704654480907e-05, "loss": 3.8389, "step": 22550 }, { "epoch": 0.25389977699510735, "grad_norm": 1.0078363418579102, "learning_rate": 1.8984844569772276e-05, "loss": 3.7896, "step": 22600 }, { "epoch": 0.25446150216545055, "grad_norm": 1.0185294151306152, "learning_rate": 1.8982597656469425e-05, "loss": 3.7948, "step": 22650 }, { "epoch": 0.2550232273357937, "grad_norm": 1.1852279901504517, "learning_rate": 1.8980350743166574e-05, "loss": 3.8872, "step": 22700 }, { "epoch": 0.25558495250613683, "grad_norm": 1.039040446281433, "learning_rate": 1.8978103829863727e-05, "loss": 3.8026, "step": 22750 }, { "epoch": 0.25614667767648003, "grad_norm": 0.9940977692604065, "learning_rate": 1.8975856916560876e-05, "loss": 3.9207, "step": 22800 }, { "epoch": 0.25670840284682317, "grad_norm": 0.9345098733901978, "learning_rate": 1.8973610003258026e-05, "loss": 3.8501, "step": 22850 }, { "epoch": 0.2572701280171663, "grad_norm": 1.0778815746307373, "learning_rate": 1.8971363089955175e-05, "loss": 3.9176, "step": 22900 }, { "epoch": 0.25783185318750945, "grad_norm": 1.1364803314208984, "learning_rate": 1.8969116176652324e-05, "loss": 3.8351, "step": 22950 }, { "epoch": 0.25839357835785265, "grad_norm": 0.9966996908187866, "learning_rate": 1.8966869263349474e-05, "loss": 3.89, "step": 23000 }, { "epoch": 0.2589553035281958, "grad_norm": 1.0334755182266235, "learning_rate": 1.8964622350046623e-05, "loss": 3.8796, "step": 23050 }, { "epoch": 0.25951702869853893, "grad_norm": 1.1375792026519775, "learning_rate": 1.8962375436743776e-05, "loss": 3.8086, "step": 23100 }, { "epoch": 0.26007875386888213, "grad_norm": 1.1250351667404175, "learning_rate": 1.8960128523440925e-05, "loss": 3.9187, "step": 23150 }, { "epoch": 0.26064047903922527, "grad_norm": 0.9979472756385803, "learning_rate": 1.8957881610138074e-05, "loss": 3.8808, "step": 23200 }, { "epoch": 0.2612022042095684, "grad_norm": 1.0042188167572021, "learning_rate": 1.8955634696835224e-05, "loss": 3.9456, "step": 23250 }, { "epoch": 0.2617639293799116, "grad_norm": 1.0758593082427979, "learning_rate": 1.8953387783532373e-05, "loss": 3.7855, "step": 23300 }, { "epoch": 0.26232565455025475, "grad_norm": 0.9572558999061584, "learning_rate": 1.8951140870229522e-05, "loss": 3.8474, "step": 23350 }, { "epoch": 0.2628873797205979, "grad_norm": 1.0140413045883179, "learning_rate": 1.894889395692667e-05, "loss": 3.8706, "step": 23400 }, { "epoch": 0.26344910489094103, "grad_norm": 1.0036652088165283, "learning_rate": 1.8946647043623824e-05, "loss": 3.8338, "step": 23450 }, { "epoch": 0.26401083006128423, "grad_norm": 1.1257251501083374, "learning_rate": 1.8944400130320973e-05, "loss": 3.9037, "step": 23500 }, { "epoch": 0.26457255523162737, "grad_norm": 1.1244933605194092, "learning_rate": 1.8942153217018123e-05, "loss": 3.8757, "step": 23550 }, { "epoch": 0.2651342804019705, "grad_norm": 1.1385120153427124, "learning_rate": 1.8939906303715272e-05, "loss": 3.9324, "step": 23600 }, { "epoch": 0.2656960055723137, "grad_norm": 1.2949111461639404, "learning_rate": 1.893765939041242e-05, "loss": 3.8725, "step": 23650 }, { "epoch": 0.26625773074265685, "grad_norm": 1.1291465759277344, "learning_rate": 1.893541247710957e-05, "loss": 3.8566, "step": 23700 }, { "epoch": 0.266819455913, "grad_norm": 1.0833088159561157, "learning_rate": 1.893316556380672e-05, "loss": 3.8305, "step": 23750 }, { "epoch": 0.2673811810833432, "grad_norm": 1.0878947973251343, "learning_rate": 1.8930918650503873e-05, "loss": 3.8964, "step": 23800 }, { "epoch": 0.26794290625368633, "grad_norm": 1.03005051612854, "learning_rate": 1.8928671737201022e-05, "loss": 3.9039, "step": 23850 }, { "epoch": 0.26850463142402947, "grad_norm": 1.0353825092315674, "learning_rate": 1.892642482389817e-05, "loss": 3.7323, "step": 23900 }, { "epoch": 0.2690663565943726, "grad_norm": 0.9908705353736877, "learning_rate": 1.892417791059532e-05, "loss": 3.8291, "step": 23950 }, { "epoch": 0.2696280817647158, "grad_norm": 1.1991065740585327, "learning_rate": 1.892193099729247e-05, "loss": 3.8784, "step": 24000 }, { "epoch": 0.27018980693505895, "grad_norm": 1.3117947578430176, "learning_rate": 1.891968408398962e-05, "loss": 3.8274, "step": 24050 }, { "epoch": 0.2707515321054021, "grad_norm": 0.9964234828948975, "learning_rate": 1.891743717068677e-05, "loss": 3.8596, "step": 24100 }, { "epoch": 0.2713132572757453, "grad_norm": 1.0427865982055664, "learning_rate": 1.8915190257383918e-05, "loss": 3.7689, "step": 24150 }, { "epoch": 0.27187498244608843, "grad_norm": 1.1147197484970093, "learning_rate": 1.891294334408107e-05, "loss": 3.8092, "step": 24200 }, { "epoch": 0.2724367076164316, "grad_norm": 1.0301438570022583, "learning_rate": 1.891069643077822e-05, "loss": 3.7719, "step": 24250 }, { "epoch": 0.27299843278677477, "grad_norm": 1.2350659370422363, "learning_rate": 1.890844951747537e-05, "loss": 3.831, "step": 24300 }, { "epoch": 0.2735601579571179, "grad_norm": 0.917402446269989, "learning_rate": 1.890620260417252e-05, "loss": 3.7391, "step": 24350 }, { "epoch": 0.27412188312746105, "grad_norm": 1.1625767946243286, "learning_rate": 1.8903955690869668e-05, "loss": 3.859, "step": 24400 }, { "epoch": 0.2746836082978042, "grad_norm": 1.1152362823486328, "learning_rate": 1.8901708777566817e-05, "loss": 3.8546, "step": 24450 }, { "epoch": 0.2752453334681474, "grad_norm": 0.9273315668106079, "learning_rate": 1.8899461864263967e-05, "loss": 3.8469, "step": 24500 }, { "epoch": 0.27580705863849053, "grad_norm": 1.0813549757003784, "learning_rate": 1.889721495096112e-05, "loss": 3.8394, "step": 24550 }, { "epoch": 0.2763687838088337, "grad_norm": 0.9964554309844971, "learning_rate": 1.889496803765827e-05, "loss": 3.869, "step": 24600 }, { "epoch": 0.27693050897917687, "grad_norm": 0.8538205623626709, "learning_rate": 1.8892721124355418e-05, "loss": 3.8306, "step": 24650 }, { "epoch": 0.27749223414952, "grad_norm": 1.0182757377624512, "learning_rate": 1.8890474211052567e-05, "loss": 3.8614, "step": 24700 }, { "epoch": 0.27805395931986315, "grad_norm": 1.0067983865737915, "learning_rate": 1.8888227297749717e-05, "loss": 3.8798, "step": 24750 }, { "epoch": 0.2786156844902063, "grad_norm": 1.0838966369628906, "learning_rate": 1.8885980384446866e-05, "loss": 3.8004, "step": 24800 }, { "epoch": 0.2791774096605495, "grad_norm": 1.0354911088943481, "learning_rate": 1.8883733471144015e-05, "loss": 3.7978, "step": 24850 }, { "epoch": 0.27973913483089263, "grad_norm": 1.3074768781661987, "learning_rate": 1.8881486557841168e-05, "loss": 3.8454, "step": 24900 }, { "epoch": 0.2803008600012358, "grad_norm": 0.9944579601287842, "learning_rate": 1.8879239644538317e-05, "loss": 3.9164, "step": 24950 }, { "epoch": 0.28086258517157897, "grad_norm": 0.8879290223121643, "learning_rate": 1.8876992731235467e-05, "loss": 3.8754, "step": 25000 }, { "epoch": 0.2814243103419221, "grad_norm": 1.0343937873840332, "learning_rate": 1.8874745817932616e-05, "loss": 3.8178, "step": 25050 }, { "epoch": 0.28198603551226525, "grad_norm": 1.0482374429702759, "learning_rate": 1.8872498904629765e-05, "loss": 3.8456, "step": 25100 }, { "epoch": 0.28254776068260845, "grad_norm": 1.1767076253890991, "learning_rate": 1.8870251991326914e-05, "loss": 3.8205, "step": 25150 }, { "epoch": 0.2831094858529516, "grad_norm": 1.220213770866394, "learning_rate": 1.8868005078024064e-05, "loss": 3.8225, "step": 25200 }, { "epoch": 0.28367121102329473, "grad_norm": 0.9741342067718506, "learning_rate": 1.8865758164721217e-05, "loss": 3.8642, "step": 25250 }, { "epoch": 0.2842329361936379, "grad_norm": 1.030258297920227, "learning_rate": 1.8863511251418366e-05, "loss": 3.8519, "step": 25300 }, { "epoch": 0.28479466136398107, "grad_norm": 1.2043633460998535, "learning_rate": 1.8861264338115515e-05, "loss": 3.8263, "step": 25350 }, { "epoch": 0.2853563865343242, "grad_norm": 1.0785086154937744, "learning_rate": 1.8859017424812664e-05, "loss": 3.7636, "step": 25400 }, { "epoch": 0.28591811170466735, "grad_norm": 0.9269720911979675, "learning_rate": 1.8856770511509814e-05, "loss": 3.8372, "step": 25450 }, { "epoch": 0.28647983687501055, "grad_norm": 1.099853277206421, "learning_rate": 1.8854523598206963e-05, "loss": 3.8628, "step": 25500 }, { "epoch": 0.2870415620453537, "grad_norm": 1.1465600728988647, "learning_rate": 1.8852276684904112e-05, "loss": 3.8568, "step": 25550 }, { "epoch": 0.28760328721569683, "grad_norm": 0.9685181975364685, "learning_rate": 1.8850029771601265e-05, "loss": 3.8208, "step": 25600 }, { "epoch": 0.28816501238604003, "grad_norm": 1.1846619844436646, "learning_rate": 1.8847782858298414e-05, "loss": 3.8492, "step": 25650 }, { "epoch": 0.2887267375563832, "grad_norm": 1.008122205734253, "learning_rate": 1.8845535944995564e-05, "loss": 3.7525, "step": 25700 }, { "epoch": 0.2892884627267263, "grad_norm": 1.1963845491409302, "learning_rate": 1.8843289031692713e-05, "loss": 3.8768, "step": 25750 }, { "epoch": 0.28985018789706946, "grad_norm": 1.0384691953659058, "learning_rate": 1.8841042118389862e-05, "loss": 3.7988, "step": 25800 }, { "epoch": 0.29041191306741265, "grad_norm": 0.917803168296814, "learning_rate": 1.8838795205087012e-05, "loss": 3.8486, "step": 25850 }, { "epoch": 0.2909736382377558, "grad_norm": 1.0689321756362915, "learning_rate": 1.883654829178416e-05, "loss": 3.8321, "step": 25900 }, { "epoch": 0.29153536340809894, "grad_norm": 0.9739123582839966, "learning_rate": 1.8834301378481314e-05, "loss": 3.8251, "step": 25950 }, { "epoch": 0.29209708857844213, "grad_norm": 0.859523594379425, "learning_rate": 1.8832054465178463e-05, "loss": 3.8407, "step": 26000 }, { "epoch": 0.2926588137487853, "grad_norm": 1.5353630781173706, "learning_rate": 1.8829807551875612e-05, "loss": 3.8114, "step": 26050 }, { "epoch": 0.2932205389191284, "grad_norm": 1.1663600206375122, "learning_rate": 1.882756063857276e-05, "loss": 3.8922, "step": 26100 }, { "epoch": 0.2937822640894716, "grad_norm": 1.1306334733963013, "learning_rate": 1.882531372526991e-05, "loss": 3.8349, "step": 26150 }, { "epoch": 0.29434398925981475, "grad_norm": 1.1433279514312744, "learning_rate": 1.882306681196706e-05, "loss": 3.7827, "step": 26200 }, { "epoch": 0.2949057144301579, "grad_norm": 0.9676677584648132, "learning_rate": 1.882081989866421e-05, "loss": 3.7992, "step": 26250 }, { "epoch": 0.29546743960050104, "grad_norm": 1.019875168800354, "learning_rate": 1.8818572985361362e-05, "loss": 3.813, "step": 26300 }, { "epoch": 0.29602916477084423, "grad_norm": 1.047001600265503, "learning_rate": 1.881632607205851e-05, "loss": 3.8601, "step": 26350 }, { "epoch": 0.2965908899411874, "grad_norm": 1.2332990169525146, "learning_rate": 1.881407915875566e-05, "loss": 3.8068, "step": 26400 }, { "epoch": 0.2971526151115305, "grad_norm": 0.885316789150238, "learning_rate": 1.881183224545281e-05, "loss": 3.8525, "step": 26450 }, { "epoch": 0.2977143402818737, "grad_norm": 1.4388920068740845, "learning_rate": 1.880958533214996e-05, "loss": 3.8855, "step": 26500 }, { "epoch": 0.29827606545221685, "grad_norm": 1.2492674589157104, "learning_rate": 1.880733841884711e-05, "loss": 3.8343, "step": 26550 }, { "epoch": 0.29883779062256, "grad_norm": 0.9600359797477722, "learning_rate": 1.8805091505544258e-05, "loss": 3.8629, "step": 26600 }, { "epoch": 0.29939951579290314, "grad_norm": 1.4856162071228027, "learning_rate": 1.880284459224141e-05, "loss": 3.8172, "step": 26650 }, { "epoch": 0.29996124096324633, "grad_norm": 1.16232168674469, "learning_rate": 1.880059767893856e-05, "loss": 3.8039, "step": 26700 }, { "epoch": 0.3005229661335895, "grad_norm": 1.2946114540100098, "learning_rate": 1.879835076563571e-05, "loss": 3.739, "step": 26750 }, { "epoch": 0.3010846913039326, "grad_norm": 1.1929129362106323, "learning_rate": 1.879610385233286e-05, "loss": 3.8478, "step": 26800 }, { "epoch": 0.3016464164742758, "grad_norm": 1.0640249252319336, "learning_rate": 1.8793856939030008e-05, "loss": 3.7763, "step": 26850 }, { "epoch": 0.30220814164461896, "grad_norm": 1.3911046981811523, "learning_rate": 1.8791610025727158e-05, "loss": 3.8613, "step": 26900 }, { "epoch": 0.3027698668149621, "grad_norm": 1.0830105543136597, "learning_rate": 1.8789408050690367e-05, "loss": 3.854, "step": 26950 }, { "epoch": 0.3033315919853053, "grad_norm": 1.2694368362426758, "learning_rate": 1.8787161137387516e-05, "loss": 3.824, "step": 27000 }, { "epoch": 0.30389331715564843, "grad_norm": 1.2115559577941895, "learning_rate": 1.8784914224084665e-05, "loss": 3.8349, "step": 27050 }, { "epoch": 0.3044550423259916, "grad_norm": 1.0922532081604004, "learning_rate": 1.8782667310781815e-05, "loss": 3.8274, "step": 27100 }, { "epoch": 0.3050167674963347, "grad_norm": 1.0742512941360474, "learning_rate": 1.8780420397478964e-05, "loss": 3.772, "step": 27150 }, { "epoch": 0.3055784926666779, "grad_norm": 1.2460834980010986, "learning_rate": 1.877821842244217e-05, "loss": 3.8323, "step": 27200 }, { "epoch": 0.30614021783702106, "grad_norm": 1.0943710803985596, "learning_rate": 1.8775971509139323e-05, "loss": 3.8507, "step": 27250 }, { "epoch": 0.3067019430073642, "grad_norm": 1.0860657691955566, "learning_rate": 1.8773724595836472e-05, "loss": 3.7839, "step": 27300 }, { "epoch": 0.3072636681777074, "grad_norm": 0.9497836828231812, "learning_rate": 1.877147768253362e-05, "loss": 3.8384, "step": 27350 }, { "epoch": 0.30782539334805054, "grad_norm": 0.932269811630249, "learning_rate": 1.876923076923077e-05, "loss": 3.8193, "step": 27400 }, { "epoch": 0.3083871185183937, "grad_norm": 0.9704727530479431, "learning_rate": 1.876698385592792e-05, "loss": 3.7925, "step": 27450 }, { "epoch": 0.3089488436887369, "grad_norm": 0.8907328248023987, "learning_rate": 1.876473694262507e-05, "loss": 3.8849, "step": 27500 }, { "epoch": 0.30951056885908, "grad_norm": 1.0381135940551758, "learning_rate": 1.876249002932222e-05, "loss": 3.845, "step": 27550 }, { "epoch": 0.31007229402942316, "grad_norm": 0.9771307706832886, "learning_rate": 1.876024311601937e-05, "loss": 3.7812, "step": 27600 }, { "epoch": 0.3106340191997663, "grad_norm": 1.2536208629608154, "learning_rate": 1.875799620271652e-05, "loss": 3.8203, "step": 27650 }, { "epoch": 0.3111957443701095, "grad_norm": 1.2103676795959473, "learning_rate": 1.875574928941367e-05, "loss": 3.8483, "step": 27700 }, { "epoch": 0.31175746954045264, "grad_norm": 1.2546497583389282, "learning_rate": 1.875350237611082e-05, "loss": 3.8319, "step": 27750 }, { "epoch": 0.3123191947107958, "grad_norm": 1.0846409797668457, "learning_rate": 1.875125546280797e-05, "loss": 3.8015, "step": 27800 }, { "epoch": 0.312880919881139, "grad_norm": 1.2220613956451416, "learning_rate": 1.8749053487771175e-05, "loss": 3.7921, "step": 27850 }, { "epoch": 0.3134426450514821, "grad_norm": 1.0720977783203125, "learning_rate": 1.8746806574468327e-05, "loss": 3.8055, "step": 27900 }, { "epoch": 0.31400437022182526, "grad_norm": 1.0860999822616577, "learning_rate": 1.8744559661165477e-05, "loss": 3.8301, "step": 27950 }, { "epoch": 0.31456609539216845, "grad_norm": 1.1940875053405762, "learning_rate": 1.8742312747862626e-05, "loss": 3.793, "step": 28000 }, { "epoch": 0.3151278205625116, "grad_norm": 1.1205674409866333, "learning_rate": 1.8740065834559775e-05, "loss": 3.781, "step": 28050 }, { "epoch": 0.31568954573285474, "grad_norm": 1.1552340984344482, "learning_rate": 1.8737818921256924e-05, "loss": 3.8237, "step": 28100 }, { "epoch": 0.3162512709031979, "grad_norm": 1.0858317613601685, "learning_rate": 1.8735572007954074e-05, "loss": 3.8862, "step": 28150 }, { "epoch": 0.3168129960735411, "grad_norm": 1.2606555223464966, "learning_rate": 1.8733325094651223e-05, "loss": 3.8343, "step": 28200 }, { "epoch": 0.3173747212438842, "grad_norm": 1.202958583831787, "learning_rate": 1.8731078181348376e-05, "loss": 3.7782, "step": 28250 }, { "epoch": 0.31793644641422736, "grad_norm": 1.0832439661026, "learning_rate": 1.8728831268045525e-05, "loss": 3.8162, "step": 28300 }, { "epoch": 0.31849817158457056, "grad_norm": 4.111958980560303, "learning_rate": 1.8726584354742674e-05, "loss": 3.8364, "step": 28350 }, { "epoch": 0.3190598967549137, "grad_norm": 0.9650757312774658, "learning_rate": 1.8724337441439824e-05, "loss": 3.8207, "step": 28400 }, { "epoch": 0.31962162192525684, "grad_norm": 1.0663647651672363, "learning_rate": 1.8722090528136973e-05, "loss": 3.8145, "step": 28450 }, { "epoch": 0.3201833470956, "grad_norm": 0.9724984169006348, "learning_rate": 1.8719843614834122e-05, "loss": 3.7738, "step": 28500 }, { "epoch": 0.3207450722659432, "grad_norm": 1.0874778032302856, "learning_rate": 1.8717596701531272e-05, "loss": 3.8575, "step": 28550 }, { "epoch": 0.3213067974362863, "grad_norm": 1.0395227670669556, "learning_rate": 1.8715349788228424e-05, "loss": 3.8257, "step": 28600 }, { "epoch": 0.32186852260662946, "grad_norm": 0.9556145668029785, "learning_rate": 1.8713102874925574e-05, "loss": 3.8482, "step": 28650 }, { "epoch": 0.32243024777697266, "grad_norm": 0.9420982599258423, "learning_rate": 1.8710855961622723e-05, "loss": 3.7761, "step": 28700 }, { "epoch": 0.3229919729473158, "grad_norm": 0.8945040106773376, "learning_rate": 1.8708609048319872e-05, "loss": 3.8195, "step": 28750 }, { "epoch": 0.32355369811765894, "grad_norm": 1.645750641822815, "learning_rate": 1.870636213501702e-05, "loss": 3.7286, "step": 28800 }, { "epoch": 0.32411542328800214, "grad_norm": 0.9189350605010986, "learning_rate": 1.870411522171417e-05, "loss": 3.8004, "step": 28850 }, { "epoch": 0.3246771484583453, "grad_norm": 1.14378023147583, "learning_rate": 1.870186830841132e-05, "loss": 3.7989, "step": 28900 }, { "epoch": 0.3252388736286884, "grad_norm": 1.208552360534668, "learning_rate": 1.8699621395108473e-05, "loss": 3.7862, "step": 28950 }, { "epoch": 0.32580059879903156, "grad_norm": 1.0845381021499634, "learning_rate": 1.8697374481805622e-05, "loss": 3.8527, "step": 29000 }, { "epoch": 0.32636232396937476, "grad_norm": 0.9852644205093384, "learning_rate": 1.869512756850277e-05, "loss": 3.8067, "step": 29050 }, { "epoch": 0.3269240491397179, "grad_norm": 1.0806910991668701, "learning_rate": 1.869288065519992e-05, "loss": 3.7952, "step": 29100 }, { "epoch": 0.32748577431006104, "grad_norm": 1.2093960046768188, "learning_rate": 1.869063374189707e-05, "loss": 3.9072, "step": 29150 }, { "epoch": 0.32804749948040424, "grad_norm": 1.2655662298202515, "learning_rate": 1.868838682859422e-05, "loss": 3.8486, "step": 29200 }, { "epoch": 0.3286092246507474, "grad_norm": 1.1227494478225708, "learning_rate": 1.868613991529137e-05, "loss": 3.8285, "step": 29250 }, { "epoch": 0.3291709498210905, "grad_norm": 1.0291639566421509, "learning_rate": 1.868389300198852e-05, "loss": 3.8551, "step": 29300 }, { "epoch": 0.3297326749914337, "grad_norm": 1.0368773937225342, "learning_rate": 1.868164608868567e-05, "loss": 3.8432, "step": 29350 }, { "epoch": 0.33029440016177686, "grad_norm": 1.1037966012954712, "learning_rate": 1.867939917538282e-05, "loss": 3.8626, "step": 29400 }, { "epoch": 0.33085612533212, "grad_norm": 0.9336369037628174, "learning_rate": 1.867715226207997e-05, "loss": 3.7892, "step": 29450 }, { "epoch": 0.33141785050246314, "grad_norm": 1.087012767791748, "learning_rate": 1.867490534877712e-05, "loss": 3.8418, "step": 29500 }, { "epoch": 0.33197957567280634, "grad_norm": 1.1471550464630127, "learning_rate": 1.8672658435474268e-05, "loss": 3.7594, "step": 29550 }, { "epoch": 0.3325413008431495, "grad_norm": 1.2329821586608887, "learning_rate": 1.8670411522171418e-05, "loss": 3.8126, "step": 29600 }, { "epoch": 0.3331030260134926, "grad_norm": 1.287199854850769, "learning_rate": 1.866816460886857e-05, "loss": 3.8232, "step": 29650 }, { "epoch": 0.3336647511838358, "grad_norm": 1.009225606918335, "learning_rate": 1.866591769556572e-05, "loss": 3.9107, "step": 29700 }, { "epoch": 0.33422647635417896, "grad_norm": 1.088283896446228, "learning_rate": 1.866367078226287e-05, "loss": 3.8091, "step": 29750 }, { "epoch": 0.3347882015245221, "grad_norm": 1.095376968383789, "learning_rate": 1.8661423868960018e-05, "loss": 3.7958, "step": 29800 }, { "epoch": 0.3353499266948653, "grad_norm": 1.1659440994262695, "learning_rate": 1.8659176955657167e-05, "loss": 3.7734, "step": 29850 }, { "epoch": 0.33591165186520844, "grad_norm": 1.0501476526260376, "learning_rate": 1.8656930042354317e-05, "loss": 3.7342, "step": 29900 }, { "epoch": 0.3364733770355516, "grad_norm": 1.17960786819458, "learning_rate": 1.8654683129051466e-05, "loss": 3.7517, "step": 29950 }, { "epoch": 0.3370351022058947, "grad_norm": 1.0510057210922241, "learning_rate": 1.865243621574862e-05, "loss": 3.8809, "step": 30000 }, { "epoch": 0.3375968273762379, "grad_norm": 0.9583242535591125, "learning_rate": 1.8650189302445768e-05, "loss": 3.7781, "step": 30050 }, { "epoch": 0.33815855254658106, "grad_norm": 1.257326364517212, "learning_rate": 1.8647942389142917e-05, "loss": 3.7793, "step": 30100 }, { "epoch": 0.3387202777169242, "grad_norm": 0.9945911169052124, "learning_rate": 1.8645695475840067e-05, "loss": 3.7762, "step": 30150 }, { "epoch": 0.3392820028872674, "grad_norm": 1.1452234983444214, "learning_rate": 1.8643448562537216e-05, "loss": 3.8735, "step": 30200 }, { "epoch": 0.33984372805761054, "grad_norm": 1.0031211376190186, "learning_rate": 1.8641201649234365e-05, "loss": 3.8606, "step": 30250 }, { "epoch": 0.3404054532279537, "grad_norm": 1.2691720724105835, "learning_rate": 1.8638954735931515e-05, "loss": 3.7905, "step": 30300 }, { "epoch": 0.3409671783982968, "grad_norm": 1.214430332183838, "learning_rate": 1.8636707822628664e-05, "loss": 3.8372, "step": 30350 }, { "epoch": 0.34152890356864, "grad_norm": 1.068723201751709, "learning_rate": 1.8634460909325817e-05, "loss": 3.8458, "step": 30400 }, { "epoch": 0.34209062873898316, "grad_norm": 1.2963320016860962, "learning_rate": 1.8632213996022966e-05, "loss": 3.843, "step": 30450 }, { "epoch": 0.3426523539093263, "grad_norm": 0.9694681763648987, "learning_rate": 1.8629967082720115e-05, "loss": 3.784, "step": 30500 }, { "epoch": 0.3432140790796695, "grad_norm": 0.9249500632286072, "learning_rate": 1.8627720169417265e-05, "loss": 3.7642, "step": 30550 }, { "epoch": 0.34377580425001264, "grad_norm": 0.9498476982116699, "learning_rate": 1.8625473256114414e-05, "loss": 3.8571, "step": 30600 }, { "epoch": 0.3443375294203558, "grad_norm": 1.084003210067749, "learning_rate": 1.8623226342811563e-05, "loss": 3.7847, "step": 30650 }, { "epoch": 0.344899254590699, "grad_norm": 1.0181688070297241, "learning_rate": 1.8620979429508713e-05, "loss": 3.7549, "step": 30700 }, { "epoch": 0.3454609797610421, "grad_norm": 1.1446342468261719, "learning_rate": 1.8618732516205865e-05, "loss": 3.8927, "step": 30750 }, { "epoch": 0.34602270493138526, "grad_norm": 1.0579931735992432, "learning_rate": 1.8616485602903015e-05, "loss": 3.8483, "step": 30800 }, { "epoch": 0.3465844301017284, "grad_norm": 1.0482735633850098, "learning_rate": 1.8614238689600164e-05, "loss": 3.909, "step": 30850 }, { "epoch": 0.3471461552720716, "grad_norm": 1.0333802700042725, "learning_rate": 1.8611991776297313e-05, "loss": 3.8737, "step": 30900 }, { "epoch": 0.34770788044241474, "grad_norm": 1.035236120223999, "learning_rate": 1.8609744862994463e-05, "loss": 3.7939, "step": 30950 }, { "epoch": 0.3482696056127579, "grad_norm": 1.209030270576477, "learning_rate": 1.8607497949691612e-05, "loss": 3.9275, "step": 31000 }, { "epoch": 0.3488313307831011, "grad_norm": 1.029096245765686, "learning_rate": 1.860525103638876e-05, "loss": 3.763, "step": 31050 }, { "epoch": 0.3493930559534442, "grad_norm": 0.9945694208145142, "learning_rate": 1.8603004123085914e-05, "loss": 3.8589, "step": 31100 }, { "epoch": 0.34995478112378736, "grad_norm": 1.048553228378296, "learning_rate": 1.8600757209783063e-05, "loss": 3.8801, "step": 31150 }, { "epoch": 0.35051650629413056, "grad_norm": 1.2170648574829102, "learning_rate": 1.8598510296480213e-05, "loss": 3.8499, "step": 31200 }, { "epoch": 0.3510782314644737, "grad_norm": 0.8793922662734985, "learning_rate": 1.8596263383177362e-05, "loss": 3.9009, "step": 31250 }, { "epoch": 0.35163995663481684, "grad_norm": 0.9794766306877136, "learning_rate": 1.859401646987451e-05, "loss": 3.8049, "step": 31300 }, { "epoch": 0.35220168180516, "grad_norm": 1.1491682529449463, "learning_rate": 1.859176955657166e-05, "loss": 3.7784, "step": 31350 }, { "epoch": 0.3527634069755032, "grad_norm": 0.9648553133010864, "learning_rate": 1.858952264326881e-05, "loss": 3.8466, "step": 31400 }, { "epoch": 0.3533251321458463, "grad_norm": 1.122705101966858, "learning_rate": 1.8587275729965963e-05, "loss": 3.8339, "step": 31450 }, { "epoch": 0.35388685731618946, "grad_norm": 1.4351998567581177, "learning_rate": 1.8585028816663112e-05, "loss": 3.823, "step": 31500 }, { "epoch": 0.35444858248653266, "grad_norm": 1.0422778129577637, "learning_rate": 1.858278190336026e-05, "loss": 3.851, "step": 31550 }, { "epoch": 0.3550103076568758, "grad_norm": 1.0606135129928589, "learning_rate": 1.858053499005741e-05, "loss": 3.8408, "step": 31600 }, { "epoch": 0.35557203282721894, "grad_norm": 1.278016448020935, "learning_rate": 1.857828807675456e-05, "loss": 3.8172, "step": 31650 }, { "epoch": 0.35613375799756214, "grad_norm": 0.9278510808944702, "learning_rate": 1.857604116345171e-05, "loss": 3.7797, "step": 31700 }, { "epoch": 0.3566954831679053, "grad_norm": 1.0965937376022339, "learning_rate": 1.857379425014886e-05, "loss": 3.7864, "step": 31750 }, { "epoch": 0.3572572083382484, "grad_norm": 1.060134768486023, "learning_rate": 1.857154733684601e-05, "loss": 3.831, "step": 31800 }, { "epoch": 0.35781893350859156, "grad_norm": 1.3302299976348877, "learning_rate": 1.856930042354316e-05, "loss": 3.8384, "step": 31850 }, { "epoch": 0.35838065867893476, "grad_norm": 1.0352579355239868, "learning_rate": 1.856705351024031e-05, "loss": 3.7721, "step": 31900 }, { "epoch": 0.3589423838492779, "grad_norm": 1.1207737922668457, "learning_rate": 1.856480659693746e-05, "loss": 3.8045, "step": 31950 }, { "epoch": 0.35950410901962104, "grad_norm": 0.9285112619400024, "learning_rate": 1.856255968363461e-05, "loss": 3.751, "step": 32000 }, { "epoch": 0.36006583418996424, "grad_norm": 1.2063809633255005, "learning_rate": 1.8560312770331758e-05, "loss": 3.799, "step": 32050 }, { "epoch": 0.3606275593603074, "grad_norm": 1.1076829433441162, "learning_rate": 1.8558065857028907e-05, "loss": 3.8278, "step": 32100 }, { "epoch": 0.3611892845306505, "grad_norm": 1.0293713808059692, "learning_rate": 1.855581894372606e-05, "loss": 3.8191, "step": 32150 }, { "epoch": 0.36175100970099366, "grad_norm": 1.1451561450958252, "learning_rate": 1.855357203042321e-05, "loss": 3.8873, "step": 32200 }, { "epoch": 0.36231273487133686, "grad_norm": 1.0879944562911987, "learning_rate": 1.855132511712036e-05, "loss": 3.7784, "step": 32250 }, { "epoch": 0.36287446004168, "grad_norm": 1.052733302116394, "learning_rate": 1.8549078203817508e-05, "loss": 3.8414, "step": 32300 }, { "epoch": 0.36343618521202314, "grad_norm": 1.3449853658676147, "learning_rate": 1.8546831290514657e-05, "loss": 3.7888, "step": 32350 }, { "epoch": 0.36399791038236634, "grad_norm": 0.9949585199356079, "learning_rate": 1.8544584377211806e-05, "loss": 3.8237, "step": 32400 }, { "epoch": 0.3645596355527095, "grad_norm": 0.9596227407455444, "learning_rate": 1.8542337463908956e-05, "loss": 3.8026, "step": 32450 }, { "epoch": 0.3651213607230526, "grad_norm": 1.0283395051956177, "learning_rate": 1.854009055060611e-05, "loss": 3.8657, "step": 32500 }, { "epoch": 0.3656830858933958, "grad_norm": 1.1715400218963623, "learning_rate": 1.8537843637303258e-05, "loss": 3.7421, "step": 32550 }, { "epoch": 0.36624481106373896, "grad_norm": 0.9519177675247192, "learning_rate": 1.8535596724000407e-05, "loss": 3.7558, "step": 32600 }, { "epoch": 0.3668065362340821, "grad_norm": 1.0189740657806396, "learning_rate": 1.8533349810697556e-05, "loss": 3.8338, "step": 32650 }, { "epoch": 0.36736826140442524, "grad_norm": 1.0201365947723389, "learning_rate": 1.8531102897394706e-05, "loss": 3.7984, "step": 32700 }, { "epoch": 0.36792998657476844, "grad_norm": 1.0041818618774414, "learning_rate": 1.8528855984091855e-05, "loss": 3.8174, "step": 32750 }, { "epoch": 0.3684917117451116, "grad_norm": 0.9530069231987, "learning_rate": 1.8526609070789004e-05, "loss": 3.8309, "step": 32800 }, { "epoch": 0.3690534369154547, "grad_norm": 1.1176979541778564, "learning_rate": 1.8524362157486157e-05, "loss": 3.8478, "step": 32850 }, { "epoch": 0.3696151620857979, "grad_norm": 1.1033034324645996, "learning_rate": 1.8522115244183306e-05, "loss": 3.8247, "step": 32900 }, { "epoch": 0.37017688725614106, "grad_norm": 1.0508520603179932, "learning_rate": 1.8519868330880456e-05, "loss": 3.8233, "step": 32950 }, { "epoch": 0.3707386124264842, "grad_norm": 1.377721905708313, "learning_rate": 1.8517621417577605e-05, "loss": 3.7988, "step": 33000 }, { "epoch": 0.3713003375968274, "grad_norm": 1.0736579895019531, "learning_rate": 1.8515374504274754e-05, "loss": 3.8424, "step": 33050 }, { "epoch": 0.37186206276717054, "grad_norm": 1.0143682956695557, "learning_rate": 1.8513127590971904e-05, "loss": 3.8026, "step": 33100 }, { "epoch": 0.3724237879375137, "grad_norm": 1.228042483329773, "learning_rate": 1.8510880677669053e-05, "loss": 3.7849, "step": 33150 }, { "epoch": 0.3729855131078568, "grad_norm": 1.1448063850402832, "learning_rate": 1.8508633764366206e-05, "loss": 3.7735, "step": 33200 }, { "epoch": 0.3735472382782, "grad_norm": 1.9350948333740234, "learning_rate": 1.8506386851063355e-05, "loss": 3.7989, "step": 33250 }, { "epoch": 0.37410896344854316, "grad_norm": 0.9634103178977966, "learning_rate": 1.8504139937760504e-05, "loss": 3.8728, "step": 33300 }, { "epoch": 0.3746706886188863, "grad_norm": 1.0275721549987793, "learning_rate": 1.8501893024457654e-05, "loss": 3.841, "step": 33350 }, { "epoch": 0.3752324137892295, "grad_norm": 1.592858076095581, "learning_rate": 1.8499646111154803e-05, "loss": 3.8184, "step": 33400 }, { "epoch": 0.37579413895957264, "grad_norm": 0.8911204934120178, "learning_rate": 1.8497399197851952e-05, "loss": 3.8351, "step": 33450 }, { "epoch": 0.3763558641299158, "grad_norm": 1.016668438911438, "learning_rate": 1.849519722281516e-05, "loss": 3.789, "step": 33500 }, { "epoch": 0.376917589300259, "grad_norm": 1.1443250179290771, "learning_rate": 1.849295030951231e-05, "loss": 3.8201, "step": 33550 }, { "epoch": 0.3774793144706021, "grad_norm": 1.1348165273666382, "learning_rate": 1.849070339620946e-05, "loss": 3.775, "step": 33600 }, { "epoch": 0.37804103964094526, "grad_norm": 0.9641530513763428, "learning_rate": 1.848845648290661e-05, "loss": 3.8072, "step": 33650 }, { "epoch": 0.3786027648112884, "grad_norm": 1.157254934310913, "learning_rate": 1.848620956960376e-05, "loss": 3.8643, "step": 33700 }, { "epoch": 0.3791644899816316, "grad_norm": 1.276639699935913, "learning_rate": 1.8483962656300908e-05, "loss": 3.8395, "step": 33750 }, { "epoch": 0.37972621515197474, "grad_norm": 1.1866374015808105, "learning_rate": 1.8481715742998057e-05, "loss": 3.8799, "step": 33800 }, { "epoch": 0.3802879403223179, "grad_norm": 1.016951322555542, "learning_rate": 1.847946882969521e-05, "loss": 3.8445, "step": 33850 }, { "epoch": 0.3808496654926611, "grad_norm": 1.0595694780349731, "learning_rate": 1.847722191639236e-05, "loss": 3.7482, "step": 33900 }, { "epoch": 0.3814113906630042, "grad_norm": 1.0442867279052734, "learning_rate": 1.847497500308951e-05, "loss": 3.9199, "step": 33950 }, { "epoch": 0.38197311583334737, "grad_norm": 1.3628323078155518, "learning_rate": 1.8472728089786658e-05, "loss": 3.836, "step": 34000 }, { "epoch": 0.3825348410036905, "grad_norm": 1.1050171852111816, "learning_rate": 1.8470481176483807e-05, "loss": 3.7957, "step": 34050 }, { "epoch": 0.3830965661740337, "grad_norm": 1.0539454221725464, "learning_rate": 1.8468234263180957e-05, "loss": 3.7208, "step": 34100 }, { "epoch": 0.38365829134437685, "grad_norm": 1.0863726139068604, "learning_rate": 1.8465987349878106e-05, "loss": 3.7781, "step": 34150 }, { "epoch": 0.38422001651472, "grad_norm": 0.9803478121757507, "learning_rate": 1.846374043657526e-05, "loss": 3.8272, "step": 34200 }, { "epoch": 0.3847817416850632, "grad_norm": 1.2689176797866821, "learning_rate": 1.8461493523272408e-05, "loss": 3.8052, "step": 34250 }, { "epoch": 0.3853434668554063, "grad_norm": 0.9259783029556274, "learning_rate": 1.8459246609969557e-05, "loss": 3.8363, "step": 34300 }, { "epoch": 0.38590519202574947, "grad_norm": 1.2102526426315308, "learning_rate": 1.8456999696666707e-05, "loss": 3.8209, "step": 34350 }, { "epoch": 0.38646691719609266, "grad_norm": 1.1629327535629272, "learning_rate": 1.8454752783363856e-05, "loss": 3.7606, "step": 34400 }, { "epoch": 0.3870286423664358, "grad_norm": 1.1801527738571167, "learning_rate": 1.8452505870061005e-05, "loss": 3.8129, "step": 34450 }, { "epoch": 0.38759036753677895, "grad_norm": 1.0275895595550537, "learning_rate": 1.8450258956758155e-05, "loss": 3.7998, "step": 34500 }, { "epoch": 0.3881520927071221, "grad_norm": 0.9174814820289612, "learning_rate": 1.8448012043455307e-05, "loss": 3.8167, "step": 34550 }, { "epoch": 0.3887138178774653, "grad_norm": 0.9516996741294861, "learning_rate": 1.8445765130152457e-05, "loss": 3.7663, "step": 34600 }, { "epoch": 0.3892755430478084, "grad_norm": 1.0816078186035156, "learning_rate": 1.8443518216849606e-05, "loss": 3.7428, "step": 34650 }, { "epoch": 0.38983726821815157, "grad_norm": 0.9516173005104065, "learning_rate": 1.8441271303546755e-05, "loss": 3.7543, "step": 34700 }, { "epoch": 0.39039899338849476, "grad_norm": 1.052913784980774, "learning_rate": 1.8439024390243905e-05, "loss": 3.8121, "step": 34750 }, { "epoch": 0.3909607185588379, "grad_norm": 1.1837207078933716, "learning_rate": 1.8436777476941054e-05, "loss": 3.795, "step": 34800 }, { "epoch": 0.39152244372918105, "grad_norm": 1.059316873550415, "learning_rate": 1.8434530563638203e-05, "loss": 3.8403, "step": 34850 }, { "epoch": 0.39208416889952424, "grad_norm": 1.0185049772262573, "learning_rate": 1.8432283650335353e-05, "loss": 3.9025, "step": 34900 }, { "epoch": 0.3926458940698674, "grad_norm": 0.9250991344451904, "learning_rate": 1.8430036737032505e-05, "loss": 3.8261, "step": 34950 }, { "epoch": 0.3932076192402105, "grad_norm": 1.0445061922073364, "learning_rate": 1.8427789823729655e-05, "loss": 3.805, "step": 35000 }, { "epoch": 0.39376934441055367, "grad_norm": 1.1061943769454956, "learning_rate": 1.8425542910426804e-05, "loss": 3.8592, "step": 35050 }, { "epoch": 0.39433106958089686, "grad_norm": 1.0674915313720703, "learning_rate": 1.8423295997123953e-05, "loss": 3.8049, "step": 35100 }, { "epoch": 0.39489279475124, "grad_norm": 1.020250678062439, "learning_rate": 1.8421049083821102e-05, "loss": 3.8215, "step": 35150 }, { "epoch": 0.39545451992158315, "grad_norm": 1.0812455415725708, "learning_rate": 1.8418802170518252e-05, "loss": 3.8594, "step": 35200 }, { "epoch": 0.39601624509192634, "grad_norm": 1.0903102159500122, "learning_rate": 1.84165552572154e-05, "loss": 3.7893, "step": 35250 }, { "epoch": 0.3965779702622695, "grad_norm": 0.9832062125205994, "learning_rate": 1.8414308343912554e-05, "loss": 3.7683, "step": 35300 }, { "epoch": 0.3971396954326126, "grad_norm": 1.0267248153686523, "learning_rate": 1.8412061430609703e-05, "loss": 3.8308, "step": 35350 }, { "epoch": 0.3977014206029558, "grad_norm": 0.9348560571670532, "learning_rate": 1.8409814517306852e-05, "loss": 3.7979, "step": 35400 }, { "epoch": 0.39826314577329897, "grad_norm": 1.093872308731079, "learning_rate": 1.8407567604004002e-05, "loss": 3.7905, "step": 35450 }, { "epoch": 0.3988248709436421, "grad_norm": 1.1041386127471924, "learning_rate": 1.840532069070115e-05, "loss": 3.8041, "step": 35500 }, { "epoch": 0.39938659611398525, "grad_norm": 1.947174072265625, "learning_rate": 1.84030737773983e-05, "loss": 3.8064, "step": 35550 }, { "epoch": 0.39994832128432845, "grad_norm": 1.0973021984100342, "learning_rate": 1.840082686409545e-05, "loss": 3.8587, "step": 35600 }, { "epoch": 0.4005100464546716, "grad_norm": 1.0880939960479736, "learning_rate": 1.8398579950792602e-05, "loss": 3.7737, "step": 35650 }, { "epoch": 0.40107177162501473, "grad_norm": 1.2912418842315674, "learning_rate": 1.8396333037489752e-05, "loss": 3.8031, "step": 35700 }, { "epoch": 0.4016334967953579, "grad_norm": 1.2398872375488281, "learning_rate": 1.83940861241869e-05, "loss": 3.7387, "step": 35750 }, { "epoch": 0.40219522196570107, "grad_norm": 1.6796479225158691, "learning_rate": 1.839183921088405e-05, "loss": 3.8944, "step": 35800 }, { "epoch": 0.4027569471360442, "grad_norm": 1.0580264329910278, "learning_rate": 1.83895922975812e-05, "loss": 3.7805, "step": 35850 }, { "epoch": 0.4033186723063874, "grad_norm": 1.1395939588546753, "learning_rate": 1.838734538427835e-05, "loss": 3.7848, "step": 35900 }, { "epoch": 0.40388039747673055, "grad_norm": 1.790340542793274, "learning_rate": 1.83850984709755e-05, "loss": 3.7637, "step": 35950 }, { "epoch": 0.4044421226470737, "grad_norm": 1.0293999910354614, "learning_rate": 1.838285155767265e-05, "loss": 3.7648, "step": 36000 }, { "epoch": 0.40500384781741683, "grad_norm": 1.108468770980835, "learning_rate": 1.83806046443698e-05, "loss": 3.6868, "step": 36050 }, { "epoch": 0.40556557298776, "grad_norm": 0.9420512914657593, "learning_rate": 1.837835773106695e-05, "loss": 3.7962, "step": 36100 }, { "epoch": 0.40612729815810317, "grad_norm": 1.288345456123352, "learning_rate": 1.83761108177641e-05, "loss": 3.8194, "step": 36150 }, { "epoch": 0.4066890233284463, "grad_norm": 0.9674716591835022, "learning_rate": 1.8373863904461248e-05, "loss": 3.7895, "step": 36200 }, { "epoch": 0.4072507484987895, "grad_norm": 1.0295203924179077, "learning_rate": 1.8371616991158398e-05, "loss": 3.758, "step": 36250 }, { "epoch": 0.40781247366913265, "grad_norm": 1.009635329246521, "learning_rate": 1.8369370077855547e-05, "loss": 3.8189, "step": 36300 }, { "epoch": 0.4083741988394758, "grad_norm": 1.264251947402954, "learning_rate": 1.83671231645527e-05, "loss": 3.7445, "step": 36350 }, { "epoch": 0.40893592400981893, "grad_norm": 1.1355700492858887, "learning_rate": 1.836487625124985e-05, "loss": 3.7772, "step": 36400 }, { "epoch": 0.4094976491801621, "grad_norm": 1.634529709815979, "learning_rate": 1.8362629337946998e-05, "loss": 3.8384, "step": 36450 }, { "epoch": 0.41005937435050527, "grad_norm": 1.0426417589187622, "learning_rate": 1.8360382424644148e-05, "loss": 3.7586, "step": 36500 }, { "epoch": 0.4106210995208484, "grad_norm": 1.0055906772613525, "learning_rate": 1.8358135511341297e-05, "loss": 3.7987, "step": 36550 }, { "epoch": 0.4111828246911916, "grad_norm": 1.060428261756897, "learning_rate": 1.8355888598038446e-05, "loss": 3.8098, "step": 36600 }, { "epoch": 0.41174454986153475, "grad_norm": 1.1187878847122192, "learning_rate": 1.8353641684735596e-05, "loss": 3.7981, "step": 36650 }, { "epoch": 0.4123062750318779, "grad_norm": 1.1359686851501465, "learning_rate": 1.8351394771432748e-05, "loss": 3.8966, "step": 36700 }, { "epoch": 0.4128680002022211, "grad_norm": 1.1621538400650024, "learning_rate": 1.834919279639595e-05, "loss": 3.8835, "step": 36750 }, { "epoch": 0.4134297253725642, "grad_norm": 1.0647566318511963, "learning_rate": 1.83469458830931e-05, "loss": 3.8796, "step": 36800 }, { "epoch": 0.41399145054290737, "grad_norm": 1.1750829219818115, "learning_rate": 1.8344698969790253e-05, "loss": 3.8571, "step": 36850 }, { "epoch": 0.4145531757132505, "grad_norm": 1.250691294670105, "learning_rate": 1.8342452056487402e-05, "loss": 3.877, "step": 36900 }, { "epoch": 0.4151149008835937, "grad_norm": 1.0842593908309937, "learning_rate": 1.834020514318455e-05, "loss": 3.738, "step": 36950 }, { "epoch": 0.41567662605393685, "grad_norm": 1.1104592084884644, "learning_rate": 1.83379582298817e-05, "loss": 3.7505, "step": 37000 }, { "epoch": 0.41623835122428, "grad_norm": 1.1149715185165405, "learning_rate": 1.833571131657885e-05, "loss": 3.8392, "step": 37050 }, { "epoch": 0.4168000763946232, "grad_norm": 1.08084237575531, "learning_rate": 1.8333464403276e-05, "loss": 3.7977, "step": 37100 }, { "epoch": 0.41736180156496633, "grad_norm": 1.0322520732879639, "learning_rate": 1.833121748997315e-05, "loss": 3.7871, "step": 37150 }, { "epoch": 0.41792352673530947, "grad_norm": 1.0011364221572876, "learning_rate": 1.8328970576670298e-05, "loss": 3.8695, "step": 37200 }, { "epoch": 0.41848525190565267, "grad_norm": 1.4675151109695435, "learning_rate": 1.832672366336745e-05, "loss": 3.8172, "step": 37250 }, { "epoch": 0.4190469770759958, "grad_norm": 1.182085394859314, "learning_rate": 1.83244767500646e-05, "loss": 3.6947, "step": 37300 }, { "epoch": 0.41960870224633895, "grad_norm": 1.2066044807434082, "learning_rate": 1.832222983676175e-05, "loss": 3.8281, "step": 37350 }, { "epoch": 0.4201704274166821, "grad_norm": 1.1199288368225098, "learning_rate": 1.83199829234589e-05, "loss": 3.7712, "step": 37400 }, { "epoch": 0.4207321525870253, "grad_norm": 1.0046926736831665, "learning_rate": 1.8317736010156048e-05, "loss": 3.8223, "step": 37450 }, { "epoch": 0.42129387775736843, "grad_norm": 1.0446561574935913, "learning_rate": 1.8315489096853197e-05, "loss": 3.8042, "step": 37500 }, { "epoch": 0.42185560292771157, "grad_norm": 1.0896804332733154, "learning_rate": 1.8313242183550347e-05, "loss": 3.7873, "step": 37550 }, { "epoch": 0.42241732809805477, "grad_norm": 1.162729024887085, "learning_rate": 1.83109952702475e-05, "loss": 3.9176, "step": 37600 }, { "epoch": 0.4229790532683979, "grad_norm": 1.2584238052368164, "learning_rate": 1.830874835694465e-05, "loss": 3.719, "step": 37650 }, { "epoch": 0.42354077843874105, "grad_norm": 0.991865873336792, "learning_rate": 1.8306501443641798e-05, "loss": 3.786, "step": 37700 }, { "epoch": 0.42410250360908425, "grad_norm": 1.1252045631408691, "learning_rate": 1.8304254530338947e-05, "loss": 3.8155, "step": 37750 }, { "epoch": 0.4246642287794274, "grad_norm": 1.1655724048614502, "learning_rate": 1.8302007617036097e-05, "loss": 3.7323, "step": 37800 }, { "epoch": 0.42522595394977053, "grad_norm": 1.1222635507583618, "learning_rate": 1.8299760703733246e-05, "loss": 3.8007, "step": 37850 }, { "epoch": 0.42578767912011367, "grad_norm": 1.05940842628479, "learning_rate": 1.8297513790430395e-05, "loss": 3.7313, "step": 37900 }, { "epoch": 0.42634940429045687, "grad_norm": 1.0890491008758545, "learning_rate": 1.8295266877127548e-05, "loss": 3.8056, "step": 37950 }, { "epoch": 0.4269111294608, "grad_norm": 1.1040856838226318, "learning_rate": 1.8293019963824697e-05, "loss": 3.7458, "step": 38000 }, { "epoch": 0.42747285463114315, "grad_norm": 1.5603495836257935, "learning_rate": 1.8290773050521847e-05, "loss": 3.7967, "step": 38050 }, { "epoch": 0.42803457980148635, "grad_norm": 0.9499339461326599, "learning_rate": 1.8288526137218996e-05, "loss": 3.7156, "step": 38100 }, { "epoch": 0.4285963049718295, "grad_norm": 0.9521780610084534, "learning_rate": 1.8286279223916145e-05, "loss": 3.7458, "step": 38150 }, { "epoch": 0.42915803014217263, "grad_norm": 1.0303791761398315, "learning_rate": 1.8284032310613295e-05, "loss": 3.7431, "step": 38200 }, { "epoch": 0.42971975531251577, "grad_norm": 1.2307860851287842, "learning_rate": 1.8281785397310444e-05, "loss": 3.7695, "step": 38250 }, { "epoch": 0.43028148048285897, "grad_norm": 1.1721601486206055, "learning_rate": 1.8279538484007597e-05, "loss": 3.9299, "step": 38300 }, { "epoch": 0.4308432056532021, "grad_norm": 1.0966248512268066, "learning_rate": 1.8277291570704746e-05, "loss": 3.7414, "step": 38350 }, { "epoch": 0.43140493082354525, "grad_norm": 1.1516059637069702, "learning_rate": 1.8275044657401895e-05, "loss": 3.7757, "step": 38400 }, { "epoch": 0.43196665599388845, "grad_norm": 1.0876235961914062, "learning_rate": 1.8272797744099044e-05, "loss": 3.7672, "step": 38450 }, { "epoch": 0.4325283811642316, "grad_norm": 0.9670385718345642, "learning_rate": 1.8270550830796194e-05, "loss": 3.7899, "step": 38500 }, { "epoch": 0.43309010633457473, "grad_norm": 1.2033318281173706, "learning_rate": 1.8268303917493343e-05, "loss": 3.8375, "step": 38550 }, { "epoch": 0.43365183150491793, "grad_norm": 1.1501555442810059, "learning_rate": 1.8266101942456552e-05, "loss": 3.8231, "step": 38600 }, { "epoch": 0.43421355667526107, "grad_norm": 1.027575969696045, "learning_rate": 1.8263855029153702e-05, "loss": 3.7934, "step": 38650 }, { "epoch": 0.4347752818456042, "grad_norm": 1.1309833526611328, "learning_rate": 1.826160811585085e-05, "loss": 3.7897, "step": 38700 }, { "epoch": 0.43533700701594735, "grad_norm": 1.0121960639953613, "learning_rate": 1.8259361202548e-05, "loss": 3.7848, "step": 38750 }, { "epoch": 0.43589873218629055, "grad_norm": 1.1874136924743652, "learning_rate": 1.825711428924515e-05, "loss": 3.8107, "step": 38800 }, { "epoch": 0.4364604573566337, "grad_norm": 1.5305979251861572, "learning_rate": 1.82548673759423e-05, "loss": 3.8421, "step": 38850 }, { "epoch": 0.43702218252697683, "grad_norm": 1.0823510885238647, "learning_rate": 1.825262046263945e-05, "loss": 3.8185, "step": 38900 }, { "epoch": 0.43758390769732003, "grad_norm": 0.98363196849823, "learning_rate": 1.82503735493366e-05, "loss": 3.7812, "step": 38950 }, { "epoch": 0.43814563286766317, "grad_norm": 1.1372807025909424, "learning_rate": 1.824812663603375e-05, "loss": 3.8593, "step": 39000 }, { "epoch": 0.4387073580380063, "grad_norm": 1.0772751569747925, "learning_rate": 1.82458797227309e-05, "loss": 3.8541, "step": 39050 }, { "epoch": 0.4392690832083495, "grad_norm": 1.145782470703125, "learning_rate": 1.824363280942805e-05, "loss": 3.855, "step": 39100 }, { "epoch": 0.43983080837869265, "grad_norm": 0.8724556565284729, "learning_rate": 1.8241385896125198e-05, "loss": 3.7936, "step": 39150 }, { "epoch": 0.4403925335490358, "grad_norm": 2.4490976333618164, "learning_rate": 1.8239138982822348e-05, "loss": 3.7432, "step": 39200 }, { "epoch": 0.44095425871937893, "grad_norm": 1.1330267190933228, "learning_rate": 1.8236892069519497e-05, "loss": 3.8154, "step": 39250 }, { "epoch": 0.44151598388972213, "grad_norm": 1.1379812955856323, "learning_rate": 1.823464515621665e-05, "loss": 3.78, "step": 39300 }, { "epoch": 0.44207770906006527, "grad_norm": 1.0772302150726318, "learning_rate": 1.82323982429138e-05, "loss": 3.7834, "step": 39350 }, { "epoch": 0.4426394342304084, "grad_norm": 1.1633509397506714, "learning_rate": 1.8230151329610948e-05, "loss": 3.8166, "step": 39400 }, { "epoch": 0.4432011594007516, "grad_norm": 1.1506922245025635, "learning_rate": 1.8227904416308098e-05, "loss": 3.7748, "step": 39450 }, { "epoch": 0.44376288457109475, "grad_norm": 1.0494705438613892, "learning_rate": 1.8225657503005247e-05, "loss": 3.8452, "step": 39500 }, { "epoch": 0.4443246097414379, "grad_norm": 1.2255929708480835, "learning_rate": 1.8223410589702396e-05, "loss": 3.7949, "step": 39550 }, { "epoch": 0.4448863349117811, "grad_norm": 1.0744147300720215, "learning_rate": 1.8221163676399546e-05, "loss": 3.7633, "step": 39600 }, { "epoch": 0.44544806008212423, "grad_norm": 1.0726063251495361, "learning_rate": 1.8218916763096698e-05, "loss": 3.7455, "step": 39650 }, { "epoch": 0.4460097852524674, "grad_norm": 1.1668347120285034, "learning_rate": 1.8216669849793848e-05, "loss": 3.7894, "step": 39700 }, { "epoch": 0.4465715104228105, "grad_norm": 1.0200697183609009, "learning_rate": 1.8214422936490997e-05, "loss": 3.7932, "step": 39750 }, { "epoch": 0.4471332355931537, "grad_norm": 1.0065279006958008, "learning_rate": 1.8212176023188146e-05, "loss": 3.8165, "step": 39800 }, { "epoch": 0.44769496076349685, "grad_norm": 0.9998236298561096, "learning_rate": 1.8209929109885295e-05, "loss": 3.7712, "step": 39850 }, { "epoch": 0.44825668593384, "grad_norm": 1.0551574230194092, "learning_rate": 1.8207682196582445e-05, "loss": 3.8372, "step": 39900 }, { "epoch": 0.4488184111041832, "grad_norm": 1.0093590021133423, "learning_rate": 1.8205435283279594e-05, "loss": 3.8506, "step": 39950 }, { "epoch": 0.44938013627452633, "grad_norm": 1.2401299476623535, "learning_rate": 1.8203188369976747e-05, "loss": 3.7613, "step": 40000 }, { "epoch": 0.4499418614448695, "grad_norm": 1.0292561054229736, "learning_rate": 1.8200941456673896e-05, "loss": 3.734, "step": 40050 }, { "epoch": 0.4505035866152126, "grad_norm": 1.2993911504745483, "learning_rate": 1.8198694543371045e-05, "loss": 3.7656, "step": 40100 }, { "epoch": 0.4510653117855558, "grad_norm": 1.1888697147369385, "learning_rate": 1.8196447630068195e-05, "loss": 3.7041, "step": 40150 }, { "epoch": 0.45162703695589895, "grad_norm": 1.0635584592819214, "learning_rate": 1.8194200716765344e-05, "loss": 3.7893, "step": 40200 }, { "epoch": 0.4521887621262421, "grad_norm": 1.0225001573562622, "learning_rate": 1.8191953803462493e-05, "loss": 3.834, "step": 40250 }, { "epoch": 0.4527504872965853, "grad_norm": 1.1304889917373657, "learning_rate": 1.8189706890159643e-05, "loss": 3.7892, "step": 40300 }, { "epoch": 0.45331221246692843, "grad_norm": 1.0044043064117432, "learning_rate": 1.8187459976856795e-05, "loss": 3.8025, "step": 40350 }, { "epoch": 0.4538739376372716, "grad_norm": 1.046651005744934, "learning_rate": 1.8185213063553945e-05, "loss": 3.7732, "step": 40400 }, { "epoch": 0.45443566280761477, "grad_norm": 1.2860900163650513, "learning_rate": 1.8182966150251094e-05, "loss": 3.779, "step": 40450 }, { "epoch": 0.4549973879779579, "grad_norm": 1.0236151218414307, "learning_rate": 1.8180719236948243e-05, "loss": 3.8071, "step": 40500 }, { "epoch": 0.45555911314830105, "grad_norm": 1.234114170074463, "learning_rate": 1.8178472323645393e-05, "loss": 3.7356, "step": 40550 }, { "epoch": 0.4561208383186442, "grad_norm": 1.2366036176681519, "learning_rate": 1.8176225410342542e-05, "loss": 3.7333, "step": 40600 }, { "epoch": 0.4566825634889874, "grad_norm": 1.0069408416748047, "learning_rate": 1.817397849703969e-05, "loss": 3.7977, "step": 40650 }, { "epoch": 0.45724428865933053, "grad_norm": 1.099448561668396, "learning_rate": 1.8171731583736844e-05, "loss": 3.8713, "step": 40700 }, { "epoch": 0.4578060138296737, "grad_norm": 0.9119434952735901, "learning_rate": 1.8169484670433993e-05, "loss": 3.7438, "step": 40750 }, { "epoch": 0.45836773900001687, "grad_norm": 1.0150121450424194, "learning_rate": 1.8167237757131143e-05, "loss": 3.7982, "step": 40800 }, { "epoch": 0.45892946417036, "grad_norm": 1.1290061473846436, "learning_rate": 1.8164990843828292e-05, "loss": 3.8241, "step": 40850 }, { "epoch": 0.45949118934070315, "grad_norm": 1.3187247514724731, "learning_rate": 1.816274393052544e-05, "loss": 3.881, "step": 40900 }, { "epoch": 0.46005291451104635, "grad_norm": 0.955877423286438, "learning_rate": 1.816049701722259e-05, "loss": 3.7177, "step": 40950 }, { "epoch": 0.4606146396813895, "grad_norm": 1.0432859659194946, "learning_rate": 1.815825010391974e-05, "loss": 3.8475, "step": 41000 }, { "epoch": 0.46117636485173263, "grad_norm": 1.1149173974990845, "learning_rate": 1.8156003190616893e-05, "loss": 3.8116, "step": 41050 }, { "epoch": 0.4617380900220758, "grad_norm": 1.0391672849655151, "learning_rate": 1.8153756277314042e-05, "loss": 3.8382, "step": 41100 }, { "epoch": 0.462299815192419, "grad_norm": 1.134096384048462, "learning_rate": 1.815150936401119e-05, "loss": 3.827, "step": 41150 }, { "epoch": 0.4628615403627621, "grad_norm": 1.1240386962890625, "learning_rate": 1.814926245070834e-05, "loss": 3.8302, "step": 41200 }, { "epoch": 0.46342326553310526, "grad_norm": 1.013810634613037, "learning_rate": 1.814701553740549e-05, "loss": 3.726, "step": 41250 }, { "epoch": 0.46398499070344845, "grad_norm": 1.0930726528167725, "learning_rate": 1.814476862410264e-05, "loss": 3.7472, "step": 41300 }, { "epoch": 0.4645467158737916, "grad_norm": 1.162300944328308, "learning_rate": 1.814252171079979e-05, "loss": 3.8044, "step": 41350 }, { "epoch": 0.46510844104413473, "grad_norm": 1.0755993127822876, "learning_rate": 1.8140274797496938e-05, "loss": 3.8546, "step": 41400 }, { "epoch": 0.46567016621447793, "grad_norm": 1.058703064918518, "learning_rate": 1.813802788419409e-05, "loss": 3.768, "step": 41450 }, { "epoch": 0.4662318913848211, "grad_norm": 1.0064525604248047, "learning_rate": 1.813578097089124e-05, "loss": 3.7942, "step": 41500 }, { "epoch": 0.4667936165551642, "grad_norm": 1.0544313192367554, "learning_rate": 1.813353405758839e-05, "loss": 3.7394, "step": 41550 }, { "epoch": 0.46735534172550736, "grad_norm": 1.0749855041503906, "learning_rate": 1.813128714428554e-05, "loss": 3.7752, "step": 41600 }, { "epoch": 0.46791706689585055, "grad_norm": 1.1062192916870117, "learning_rate": 1.8129040230982688e-05, "loss": 3.758, "step": 41650 }, { "epoch": 0.4684787920661937, "grad_norm": 1.165113091468811, "learning_rate": 1.8126793317679837e-05, "loss": 3.8017, "step": 41700 }, { "epoch": 0.46904051723653684, "grad_norm": 1.0622942447662354, "learning_rate": 1.8124546404376986e-05, "loss": 3.8276, "step": 41750 }, { "epoch": 0.46960224240688003, "grad_norm": 1.0217134952545166, "learning_rate": 1.812229949107414e-05, "loss": 3.8924, "step": 41800 }, { "epoch": 0.4701639675772232, "grad_norm": 1.0252145528793335, "learning_rate": 1.812005257777129e-05, "loss": 3.793, "step": 41850 }, { "epoch": 0.4707256927475663, "grad_norm": 1.1097480058670044, "learning_rate": 1.8117805664468438e-05, "loss": 3.8127, "step": 41900 }, { "epoch": 0.47128741791790946, "grad_norm": 1.2562998533248901, "learning_rate": 1.8115558751165587e-05, "loss": 3.7614, "step": 41950 }, { "epoch": 0.47184914308825265, "grad_norm": 0.9875652194023132, "learning_rate": 1.8113311837862736e-05, "loss": 3.8162, "step": 42000 }, { "epoch": 0.4724108682585958, "grad_norm": 1.0522080659866333, "learning_rate": 1.8111064924559886e-05, "loss": 3.7795, "step": 42050 }, { "epoch": 0.47297259342893894, "grad_norm": 1.094172477722168, "learning_rate": 1.8108818011257035e-05, "loss": 3.7722, "step": 42100 }, { "epoch": 0.47353431859928213, "grad_norm": 1.3399633169174194, "learning_rate": 1.8106571097954188e-05, "loss": 3.6965, "step": 42150 }, { "epoch": 0.4740960437696253, "grad_norm": 1.1942780017852783, "learning_rate": 1.8104324184651337e-05, "loss": 3.7639, "step": 42200 }, { "epoch": 0.4746577689399684, "grad_norm": 1.0598344802856445, "learning_rate": 1.8102077271348486e-05, "loss": 3.7796, "step": 42250 }, { "epoch": 0.4752194941103116, "grad_norm": 1.0890198945999146, "learning_rate": 1.8099830358045636e-05, "loss": 3.7223, "step": 42300 }, { "epoch": 0.47578121928065475, "grad_norm": 1.029687523841858, "learning_rate": 1.809762838300884e-05, "loss": 3.7558, "step": 42350 }, { "epoch": 0.4763429444509979, "grad_norm": 0.9146989583969116, "learning_rate": 1.809538146970599e-05, "loss": 3.702, "step": 42400 }, { "epoch": 0.47690466962134104, "grad_norm": 1.1156704425811768, "learning_rate": 1.8093134556403144e-05, "loss": 3.8005, "step": 42450 }, { "epoch": 0.47746639479168423, "grad_norm": 1.1274996995925903, "learning_rate": 1.8090887643100293e-05, "loss": 3.8125, "step": 42500 }, { "epoch": 0.4780281199620274, "grad_norm": 1.1434170007705688, "learning_rate": 1.8088640729797442e-05, "loss": 3.7812, "step": 42550 }, { "epoch": 0.4785898451323705, "grad_norm": 1.032583236694336, "learning_rate": 1.808639381649459e-05, "loss": 3.7945, "step": 42600 }, { "epoch": 0.4791515703027137, "grad_norm": 1.0314099788665771, "learning_rate": 1.808414690319174e-05, "loss": 3.7249, "step": 42650 }, { "epoch": 0.47971329547305686, "grad_norm": 1.1356045007705688, "learning_rate": 1.808189998988889e-05, "loss": 3.8384, "step": 42700 }, { "epoch": 0.4802750206434, "grad_norm": 1.1715441942214966, "learning_rate": 1.807965307658604e-05, "loss": 3.858, "step": 42750 }, { "epoch": 0.4808367458137432, "grad_norm": 1.0327473878860474, "learning_rate": 1.8077406163283192e-05, "loss": 3.8273, "step": 42800 }, { "epoch": 0.48139847098408634, "grad_norm": 1.1157915592193604, "learning_rate": 1.807515924998034e-05, "loss": 3.7637, "step": 42850 }, { "epoch": 0.4819601961544295, "grad_norm": 1.0877025127410889, "learning_rate": 1.807291233667749e-05, "loss": 3.7895, "step": 42900 }, { "epoch": 0.4825219213247726, "grad_norm": 1.0560516119003296, "learning_rate": 1.807066542337464e-05, "loss": 3.7441, "step": 42950 }, { "epoch": 0.4830836464951158, "grad_norm": 1.0265017747879028, "learning_rate": 1.806841851007179e-05, "loss": 3.8135, "step": 43000 }, { "epoch": 0.48364537166545896, "grad_norm": 1.2721837759017944, "learning_rate": 1.806617159676894e-05, "loss": 3.7467, "step": 43050 }, { "epoch": 0.4842070968358021, "grad_norm": 1.0625399351119995, "learning_rate": 1.8063924683466088e-05, "loss": 3.694, "step": 43100 }, { "epoch": 0.4847688220061453, "grad_norm": 1.0817184448242188, "learning_rate": 1.806167777016324e-05, "loss": 3.7768, "step": 43150 }, { "epoch": 0.48533054717648844, "grad_norm": 1.1214500665664673, "learning_rate": 1.805943085686039e-05, "loss": 3.833, "step": 43200 }, { "epoch": 0.4858922723468316, "grad_norm": 1.082459568977356, "learning_rate": 1.805718394355754e-05, "loss": 3.8078, "step": 43250 }, { "epoch": 0.4864539975171748, "grad_norm": 1.3082919120788574, "learning_rate": 1.805493703025469e-05, "loss": 3.7968, "step": 43300 }, { "epoch": 0.4870157226875179, "grad_norm": 1.1481602191925049, "learning_rate": 1.8052690116951838e-05, "loss": 3.7631, "step": 43350 }, { "epoch": 0.48757744785786106, "grad_norm": 1.0289119482040405, "learning_rate": 1.8050443203648987e-05, "loss": 3.8111, "step": 43400 }, { "epoch": 0.4881391730282042, "grad_norm": 1.1134016513824463, "learning_rate": 1.8048196290346137e-05, "loss": 3.7552, "step": 43450 }, { "epoch": 0.4887008981985474, "grad_norm": 1.1744118928909302, "learning_rate": 1.804594937704329e-05, "loss": 3.9021, "step": 43500 }, { "epoch": 0.48926262336889054, "grad_norm": 1.0428119897842407, "learning_rate": 1.804370246374044e-05, "loss": 3.7142, "step": 43550 }, { "epoch": 0.4898243485392337, "grad_norm": 1.2769163846969604, "learning_rate": 1.8041455550437588e-05, "loss": 3.7732, "step": 43600 }, { "epoch": 0.4903860737095769, "grad_norm": 1.1050018072128296, "learning_rate": 1.8039208637134737e-05, "loss": 3.8573, "step": 43650 }, { "epoch": 0.49094779887992, "grad_norm": 1.1701105833053589, "learning_rate": 1.8036961723831887e-05, "loss": 3.7084, "step": 43700 }, { "epoch": 0.49150952405026316, "grad_norm": 1.3127405643463135, "learning_rate": 1.8034714810529036e-05, "loss": 3.834, "step": 43750 }, { "epoch": 0.4920712492206063, "grad_norm": 1.1261199712753296, "learning_rate": 1.8032467897226185e-05, "loss": 3.7859, "step": 43800 }, { "epoch": 0.4926329743909495, "grad_norm": 0.8711647987365723, "learning_rate": 1.8030220983923338e-05, "loss": 3.7959, "step": 43850 }, { "epoch": 0.49319469956129264, "grad_norm": 1.015077829360962, "learning_rate": 1.8027974070620487e-05, "loss": 3.78, "step": 43900 }, { "epoch": 0.4937564247316358, "grad_norm": 1.1011428833007812, "learning_rate": 1.8025727157317637e-05, "loss": 3.8081, "step": 43950 }, { "epoch": 0.494318149901979, "grad_norm": 1.3075041770935059, "learning_rate": 1.8023480244014786e-05, "loss": 3.7885, "step": 44000 }, { "epoch": 0.4948798750723221, "grad_norm": 1.1464548110961914, "learning_rate": 1.8021233330711935e-05, "loss": 3.8184, "step": 44050 }, { "epoch": 0.49544160024266526, "grad_norm": 1.1513046026229858, "learning_rate": 1.8018986417409085e-05, "loss": 3.7232, "step": 44100 }, { "epoch": 0.49600332541300846, "grad_norm": 1.0444048643112183, "learning_rate": 1.8016739504106234e-05, "loss": 3.7568, "step": 44150 }, { "epoch": 0.4965650505833516, "grad_norm": 1.0750826597213745, "learning_rate": 1.8014492590803387e-05, "loss": 3.7301, "step": 44200 }, { "epoch": 0.49712677575369474, "grad_norm": 1.0664238929748535, "learning_rate": 1.8012245677500536e-05, "loss": 3.8694, "step": 44250 }, { "epoch": 0.4976885009240379, "grad_norm": 0.9873380661010742, "learning_rate": 1.8009998764197685e-05, "loss": 3.8639, "step": 44300 }, { "epoch": 0.4982502260943811, "grad_norm": 1.2076796293258667, "learning_rate": 1.8007751850894835e-05, "loss": 3.7719, "step": 44350 }, { "epoch": 0.4988119512647242, "grad_norm": 1.0296562910079956, "learning_rate": 1.8005504937591984e-05, "loss": 3.7266, "step": 44400 }, { "epoch": 0.49937367643506736, "grad_norm": 1.1904340982437134, "learning_rate": 1.8003258024289133e-05, "loss": 3.7871, "step": 44450 }, { "epoch": 0.49993540160541056, "grad_norm": 1.0401132106781006, "learning_rate": 1.8001011110986283e-05, "loss": 3.7817, "step": 44500 }, { "epoch": 0.5004971267757536, "grad_norm": 1.9463406801223755, "learning_rate": 1.7998764197683435e-05, "loss": 3.8001, "step": 44550 }, { "epoch": 0.5010588519460969, "grad_norm": 1.1187658309936523, "learning_rate": 1.7996517284380585e-05, "loss": 3.7387, "step": 44600 }, { "epoch": 0.50162057711644, "grad_norm": 1.1454840898513794, "learning_rate": 1.7994270371077734e-05, "loss": 3.8329, "step": 44650 }, { "epoch": 0.5021823022867832, "grad_norm": 1.2378987073898315, "learning_rate": 1.7992023457774883e-05, "loss": 3.8162, "step": 44700 }, { "epoch": 0.5027440274571263, "grad_norm": 2.6603729724884033, "learning_rate": 1.7989776544472033e-05, "loss": 3.7931, "step": 44750 }, { "epoch": 0.5033057526274695, "grad_norm": 1.2083017826080322, "learning_rate": 1.7987529631169182e-05, "loss": 3.863, "step": 44800 }, { "epoch": 0.5038674777978126, "grad_norm": 1.057800054550171, "learning_rate": 1.798528271786633e-05, "loss": 3.7743, "step": 44850 }, { "epoch": 0.5044292029681559, "grad_norm": 1.142313838005066, "learning_rate": 1.7983035804563484e-05, "loss": 3.7603, "step": 44900 }, { "epoch": 0.504990928138499, "grad_norm": 1.0015513896942139, "learning_rate": 1.7980788891260633e-05, "loss": 3.7382, "step": 44950 }, { "epoch": 0.5055526533088421, "grad_norm": 1.2800809144973755, "learning_rate": 1.7978541977957783e-05, "loss": 3.7812, "step": 45000 }, { "epoch": 0.5061143784791853, "grad_norm": 1.1067944765090942, "learning_rate": 1.7976295064654932e-05, "loss": 3.7519, "step": 45050 }, { "epoch": 0.5066761036495284, "grad_norm": 0.9286227822303772, "learning_rate": 1.797404815135208e-05, "loss": 3.703, "step": 45100 }, { "epoch": 0.5072378288198716, "grad_norm": 1.068844199180603, "learning_rate": 1.797180123804923e-05, "loss": 3.8061, "step": 45150 }, { "epoch": 0.5077995539902147, "grad_norm": 1.1424062252044678, "learning_rate": 1.796955432474638e-05, "loss": 3.7488, "step": 45200 }, { "epoch": 0.508361279160558, "grad_norm": 1.1277310848236084, "learning_rate": 1.7967307411443532e-05, "loss": 3.6984, "step": 45250 }, { "epoch": 0.5089230043309011, "grad_norm": 1.0205165147781372, "learning_rate": 1.7965060498140682e-05, "loss": 3.7432, "step": 45300 }, { "epoch": 0.5094847295012442, "grad_norm": 1.3567684888839722, "learning_rate": 1.796281358483783e-05, "loss": 3.8955, "step": 45350 }, { "epoch": 0.5100464546715874, "grad_norm": 0.9939897656440735, "learning_rate": 1.796056667153498e-05, "loss": 3.8553, "step": 45400 }, { "epoch": 0.5106081798419305, "grad_norm": 1.4429136514663696, "learning_rate": 1.795831975823213e-05, "loss": 3.7886, "step": 45450 }, { "epoch": 0.5111699050122737, "grad_norm": 1.3088704347610474, "learning_rate": 1.7956117783195336e-05, "loss": 3.7935, "step": 45500 }, { "epoch": 0.5117316301826168, "grad_norm": 1.0963215827941895, "learning_rate": 1.795387086989249e-05, "loss": 3.7274, "step": 45550 }, { "epoch": 0.5122933553529601, "grad_norm": 1.1551778316497803, "learning_rate": 1.7951623956589638e-05, "loss": 3.7619, "step": 45600 }, { "epoch": 0.5128550805233032, "grad_norm": 1.2554855346679688, "learning_rate": 1.7949377043286787e-05, "loss": 3.8272, "step": 45650 }, { "epoch": 0.5134168056936463, "grad_norm": 1.3378087282180786, "learning_rate": 1.7947130129983936e-05, "loss": 3.7626, "step": 45700 }, { "epoch": 0.5139785308639895, "grad_norm": 1.0493336915969849, "learning_rate": 1.7944883216681086e-05, "loss": 3.8167, "step": 45750 }, { "epoch": 0.5145402560343326, "grad_norm": 1.1279903650283813, "learning_rate": 1.7942636303378235e-05, "loss": 3.7878, "step": 45800 }, { "epoch": 0.5151019812046758, "grad_norm": 1.1540268659591675, "learning_rate": 1.7940389390075384e-05, "loss": 3.8052, "step": 45850 }, { "epoch": 0.5156637063750189, "grad_norm": 1.316269040107727, "learning_rate": 1.7938142476772537e-05, "loss": 3.832, "step": 45900 }, { "epoch": 0.5162254315453622, "grad_norm": 0.9910030364990234, "learning_rate": 1.7935895563469686e-05, "loss": 3.7452, "step": 45950 }, { "epoch": 0.5167871567157053, "grad_norm": 1.1497697830200195, "learning_rate": 1.7933648650166836e-05, "loss": 3.7542, "step": 46000 }, { "epoch": 0.5173488818860484, "grad_norm": 1.08818781375885, "learning_rate": 1.7931401736863985e-05, "loss": 3.8004, "step": 46050 }, { "epoch": 0.5179106070563916, "grad_norm": 1.0237677097320557, "learning_rate": 1.7929154823561134e-05, "loss": 3.8064, "step": 46100 }, { "epoch": 0.5184723322267347, "grad_norm": 1.1072641611099243, "learning_rate": 1.7926907910258284e-05, "loss": 3.7539, "step": 46150 }, { "epoch": 0.5190340573970779, "grad_norm": 1.2030876874923706, "learning_rate": 1.7924660996955433e-05, "loss": 3.8923, "step": 46200 }, { "epoch": 0.5195957825674211, "grad_norm": 1.003875732421875, "learning_rate": 1.7922414083652586e-05, "loss": 3.7836, "step": 46250 }, { "epoch": 0.5201575077377643, "grad_norm": 1.058693766593933, "learning_rate": 1.7920167170349735e-05, "loss": 3.757, "step": 46300 }, { "epoch": 0.5207192329081074, "grad_norm": 1.0645219087600708, "learning_rate": 1.7917920257046884e-05, "loss": 3.8424, "step": 46350 }, { "epoch": 0.5212809580784505, "grad_norm": 1.0608023405075073, "learning_rate": 1.7915673343744034e-05, "loss": 3.7468, "step": 46400 }, { "epoch": 0.5218426832487937, "grad_norm": 1.049337387084961, "learning_rate": 1.7913426430441183e-05, "loss": 3.8215, "step": 46450 }, { "epoch": 0.5224044084191368, "grad_norm": 1.2163463830947876, "learning_rate": 1.7911179517138332e-05, "loss": 3.7684, "step": 46500 }, { "epoch": 0.52296613358948, "grad_norm": 1.433159589767456, "learning_rate": 1.790893260383548e-05, "loss": 3.7887, "step": 46550 }, { "epoch": 0.5235278587598232, "grad_norm": 2.4778525829315186, "learning_rate": 1.790668569053263e-05, "loss": 3.8918, "step": 46600 }, { "epoch": 0.5240895839301664, "grad_norm": 1.0728590488433838, "learning_rate": 1.7904438777229784e-05, "loss": 3.7463, "step": 46650 }, { "epoch": 0.5246513091005095, "grad_norm": 1.0942366123199463, "learning_rate": 1.7902191863926933e-05, "loss": 3.8691, "step": 46700 }, { "epoch": 0.5252130342708526, "grad_norm": 1.2694611549377441, "learning_rate": 1.7899944950624082e-05, "loss": 3.7357, "step": 46750 }, { "epoch": 0.5257747594411958, "grad_norm": 0.9541887044906616, "learning_rate": 1.789769803732123e-05, "loss": 3.8223, "step": 46800 }, { "epoch": 0.5263364846115389, "grad_norm": 1.0459247827529907, "learning_rate": 1.789545112401838e-05, "loss": 3.811, "step": 46850 }, { "epoch": 0.5268982097818821, "grad_norm": 1.215408444404602, "learning_rate": 1.789320421071553e-05, "loss": 3.8257, "step": 46900 }, { "epoch": 0.5274599349522253, "grad_norm": 1.2091480493545532, "learning_rate": 1.789095729741268e-05, "loss": 3.7794, "step": 46950 }, { "epoch": 0.5280216601225685, "grad_norm": 1.0477527379989624, "learning_rate": 1.7888710384109832e-05, "loss": 3.7567, "step": 47000 }, { "epoch": 0.5285833852929116, "grad_norm": 1.0070216655731201, "learning_rate": 1.788646347080698e-05, "loss": 3.8107, "step": 47050 }, { "epoch": 0.5291451104632547, "grad_norm": 0.9893909096717834, "learning_rate": 1.788421655750413e-05, "loss": 3.7764, "step": 47100 }, { "epoch": 0.5297068356335979, "grad_norm": 0.9789925217628479, "learning_rate": 1.788196964420128e-05, "loss": 3.822, "step": 47150 }, { "epoch": 0.530268560803941, "grad_norm": 0.9236642122268677, "learning_rate": 1.787972273089843e-05, "loss": 3.8091, "step": 47200 }, { "epoch": 0.5308302859742843, "grad_norm": 0.9914894700050354, "learning_rate": 1.787747581759558e-05, "loss": 3.7204, "step": 47250 }, { "epoch": 0.5313920111446274, "grad_norm": 1.0681378841400146, "learning_rate": 1.7875228904292728e-05, "loss": 3.778, "step": 47300 }, { "epoch": 0.5319537363149706, "grad_norm": 0.9172595739364624, "learning_rate": 1.787298199098988e-05, "loss": 3.7245, "step": 47350 }, { "epoch": 0.5325154614853137, "grad_norm": 1.0290454626083374, "learning_rate": 1.787073507768703e-05, "loss": 3.6851, "step": 47400 }, { "epoch": 0.5330771866556568, "grad_norm": 1.009787678718567, "learning_rate": 1.786848816438418e-05, "loss": 3.7659, "step": 47450 }, { "epoch": 0.533638911826, "grad_norm": 1.0187406539916992, "learning_rate": 1.786624125108133e-05, "loss": 3.8141, "step": 47500 }, { "epoch": 0.5342006369963431, "grad_norm": 1.3476841449737549, "learning_rate": 1.7863994337778478e-05, "loss": 3.7867, "step": 47550 }, { "epoch": 0.5347623621666864, "grad_norm": 0.9151012897491455, "learning_rate": 1.7861747424475627e-05, "loss": 3.7697, "step": 47600 }, { "epoch": 0.5353240873370295, "grad_norm": 1.2090747356414795, "learning_rate": 1.7859500511172777e-05, "loss": 3.7745, "step": 47650 }, { "epoch": 0.5358858125073727, "grad_norm": 1.0624113082885742, "learning_rate": 1.785725359786993e-05, "loss": 3.8045, "step": 47700 }, { "epoch": 0.5364475376777158, "grad_norm": 0.9345954656600952, "learning_rate": 1.785500668456708e-05, "loss": 3.7731, "step": 47750 }, { "epoch": 0.5370092628480589, "grad_norm": 1.0432896614074707, "learning_rate": 1.7852759771264228e-05, "loss": 3.7274, "step": 47800 }, { "epoch": 0.5375709880184021, "grad_norm": 1.1865384578704834, "learning_rate": 1.7850512857961377e-05, "loss": 3.808, "step": 47850 }, { "epoch": 0.5381327131887452, "grad_norm": 1.299694538116455, "learning_rate": 1.7848265944658527e-05, "loss": 3.8216, "step": 47900 }, { "epoch": 0.5386944383590885, "grad_norm": 1.0396870374679565, "learning_rate": 1.7846019031355676e-05, "loss": 3.7653, "step": 47950 }, { "epoch": 0.5392561635294316, "grad_norm": 1.0869770050048828, "learning_rate": 1.7843772118052825e-05, "loss": 3.7826, "step": 48000 }, { "epoch": 0.5398178886997748, "grad_norm": 1.1290099620819092, "learning_rate": 1.7841525204749978e-05, "loss": 3.7563, "step": 48050 }, { "epoch": 0.5403796138701179, "grad_norm": 1.1860477924346924, "learning_rate": 1.7839278291447127e-05, "loss": 3.7272, "step": 48100 }, { "epoch": 0.540941339040461, "grad_norm": 0.928475022315979, "learning_rate": 1.7837031378144277e-05, "loss": 3.75, "step": 48150 }, { "epoch": 0.5415030642108042, "grad_norm": 1.1937905550003052, "learning_rate": 1.7834784464841426e-05, "loss": 3.729, "step": 48200 }, { "epoch": 0.5420647893811473, "grad_norm": 0.9828348159790039, "learning_rate": 1.7832537551538575e-05, "loss": 3.7127, "step": 48250 }, { "epoch": 0.5426265145514906, "grad_norm": 1.0111702680587769, "learning_rate": 1.7830290638235725e-05, "loss": 3.7807, "step": 48300 }, { "epoch": 0.5431882397218337, "grad_norm": 1.058048963546753, "learning_rate": 1.7828043724932874e-05, "loss": 3.8394, "step": 48350 }, { "epoch": 0.5437499648921769, "grad_norm": 1.034980058670044, "learning_rate": 1.7825796811630027e-05, "loss": 3.8054, "step": 48400 }, { "epoch": 0.54431169006252, "grad_norm": 1.2437156438827515, "learning_rate": 1.7823549898327176e-05, "loss": 3.7458, "step": 48450 }, { "epoch": 0.5448734152328631, "grad_norm": 1.2050033807754517, "learning_rate": 1.7821302985024325e-05, "loss": 3.8883, "step": 48500 }, { "epoch": 0.5454351404032063, "grad_norm": 1.2797744274139404, "learning_rate": 1.7819056071721474e-05, "loss": 3.8127, "step": 48550 }, { "epoch": 0.5459968655735495, "grad_norm": 1.0819027423858643, "learning_rate": 1.7816809158418624e-05, "loss": 3.7676, "step": 48600 }, { "epoch": 0.5465585907438927, "grad_norm": 1.1629184484481812, "learning_rate": 1.7814562245115773e-05, "loss": 3.7014, "step": 48650 }, { "epoch": 0.5471203159142358, "grad_norm": 1.2764919996261597, "learning_rate": 1.7812315331812922e-05, "loss": 3.7961, "step": 48700 }, { "epoch": 0.547682041084579, "grad_norm": 1.063462734222412, "learning_rate": 1.7810068418510075e-05, "loss": 3.79, "step": 48750 }, { "epoch": 0.5482437662549221, "grad_norm": 1.0746026039123535, "learning_rate": 1.7807821505207224e-05, "loss": 3.7543, "step": 48800 }, { "epoch": 0.5488054914252652, "grad_norm": 1.0190949440002441, "learning_rate": 1.7805574591904374e-05, "loss": 3.7683, "step": 48850 }, { "epoch": 0.5493672165956084, "grad_norm": 1.0866377353668213, "learning_rate": 1.7803327678601523e-05, "loss": 3.8017, "step": 48900 }, { "epoch": 0.5499289417659516, "grad_norm": 1.1085233688354492, "learning_rate": 1.7801080765298672e-05, "loss": 3.7577, "step": 48950 }, { "epoch": 0.5504906669362948, "grad_norm": 1.2025694847106934, "learning_rate": 1.7798833851995822e-05, "loss": 3.7937, "step": 49000 }, { "epoch": 0.5510523921066379, "grad_norm": 1.199409008026123, "learning_rate": 1.779658693869297e-05, "loss": 3.7602, "step": 49050 }, { "epoch": 0.5516141172769811, "grad_norm": 1.1424399614334106, "learning_rate": 1.7794340025390124e-05, "loss": 3.7481, "step": 49100 }, { "epoch": 0.5521758424473242, "grad_norm": 0.9761651754379272, "learning_rate": 1.7792093112087273e-05, "loss": 3.7907, "step": 49150 }, { "epoch": 0.5527375676176673, "grad_norm": 1.0852487087249756, "learning_rate": 1.7789846198784422e-05, "loss": 3.7912, "step": 49200 }, { "epoch": 0.5532992927880105, "grad_norm": 1.1241354942321777, "learning_rate": 1.7787599285481572e-05, "loss": 3.7338, "step": 49250 }, { "epoch": 0.5538610179583537, "grad_norm": 1.0911507606506348, "learning_rate": 1.778535237217872e-05, "loss": 3.6668, "step": 49300 }, { "epoch": 0.5544227431286969, "grad_norm": 1.608834981918335, "learning_rate": 1.778310545887587e-05, "loss": 3.7726, "step": 49350 }, { "epoch": 0.55498446829904, "grad_norm": 1.217059850692749, "learning_rate": 1.778085854557302e-05, "loss": 3.7954, "step": 49400 }, { "epoch": 0.5555461934693832, "grad_norm": 1.0991884469985962, "learning_rate": 1.7778611632270172e-05, "loss": 3.7368, "step": 49450 }, { "epoch": 0.5561079186397263, "grad_norm": 0.9799400568008423, "learning_rate": 1.777636471896732e-05, "loss": 3.7249, "step": 49500 }, { "epoch": 0.5566696438100694, "grad_norm": 1.1263720989227295, "learning_rate": 1.777411780566447e-05, "loss": 3.8127, "step": 49550 }, { "epoch": 0.5572313689804126, "grad_norm": 1.0370090007781982, "learning_rate": 1.777187089236162e-05, "loss": 3.6818, "step": 49600 }, { "epoch": 0.5577930941507558, "grad_norm": 2.1102616786956787, "learning_rate": 1.7769668917324826e-05, "loss": 3.7238, "step": 49650 }, { "epoch": 0.558354819321099, "grad_norm": 1.13498055934906, "learning_rate": 1.7767422004021976e-05, "loss": 3.7812, "step": 49700 }, { "epoch": 0.5589165444914421, "grad_norm": 0.9529072046279907, "learning_rate": 1.7765175090719128e-05, "loss": 3.7319, "step": 49750 }, { "epoch": 0.5594782696617853, "grad_norm": 1.0441597700119019, "learning_rate": 1.7762928177416278e-05, "loss": 3.7676, "step": 49800 }, { "epoch": 0.5600399948321284, "grad_norm": 1.1571311950683594, "learning_rate": 1.7760681264113427e-05, "loss": 3.7601, "step": 49850 }, { "epoch": 0.5606017200024715, "grad_norm": 1.0752143859863281, "learning_rate": 1.7758434350810576e-05, "loss": 3.7677, "step": 49900 }, { "epoch": 0.5611634451728148, "grad_norm": 0.9833121299743652, "learning_rate": 1.7756187437507726e-05, "loss": 3.8865, "step": 49950 }, { "epoch": 0.5617251703431579, "grad_norm": 1.0245425701141357, "learning_rate": 1.7753940524204875e-05, "loss": 3.8149, "step": 50000 }, { "epoch": 0.5622868955135011, "grad_norm": 1.0369062423706055, "learning_rate": 1.7751693610902024e-05, "loss": 3.7772, "step": 50050 }, { "epoch": 0.5628486206838442, "grad_norm": 1.0579601526260376, "learning_rate": 1.7749446697599177e-05, "loss": 3.722, "step": 50100 }, { "epoch": 0.5634103458541874, "grad_norm": 1.058449625968933, "learning_rate": 1.7747199784296326e-05, "loss": 3.8031, "step": 50150 }, { "epoch": 0.5639720710245305, "grad_norm": 1.1619818210601807, "learning_rate": 1.7744952870993475e-05, "loss": 3.8059, "step": 50200 }, { "epoch": 0.5645337961948736, "grad_norm": 1.1637179851531982, "learning_rate": 1.7742705957690625e-05, "loss": 3.7732, "step": 50250 }, { "epoch": 0.5650955213652169, "grad_norm": 1.051743745803833, "learning_rate": 1.7740459044387774e-05, "loss": 3.7137, "step": 50300 }, { "epoch": 0.56565724653556, "grad_norm": 1.0153290033340454, "learning_rate": 1.7738212131084923e-05, "loss": 3.7306, "step": 50350 }, { "epoch": 0.5662189717059032, "grad_norm": 1.2448453903198242, "learning_rate": 1.7735965217782073e-05, "loss": 3.8135, "step": 50400 }, { "epoch": 0.5667806968762463, "grad_norm": 1.077943205833435, "learning_rate": 1.7733718304479225e-05, "loss": 3.7605, "step": 50450 }, { "epoch": 0.5673424220465895, "grad_norm": 1.080708384513855, "learning_rate": 1.7731471391176375e-05, "loss": 3.8126, "step": 50500 }, { "epoch": 0.5679041472169326, "grad_norm": 1.1805051565170288, "learning_rate": 1.7729224477873524e-05, "loss": 3.7539, "step": 50550 }, { "epoch": 0.5684658723872758, "grad_norm": 1.0887198448181152, "learning_rate": 1.7726977564570673e-05, "loss": 3.8005, "step": 50600 }, { "epoch": 0.569027597557619, "grad_norm": 1.1345741748809814, "learning_rate": 1.7724730651267823e-05, "loss": 3.8272, "step": 50650 }, { "epoch": 0.5695893227279621, "grad_norm": 0.9996209740638733, "learning_rate": 1.7722483737964972e-05, "loss": 3.7027, "step": 50700 }, { "epoch": 0.5701510478983053, "grad_norm": 1.140913724899292, "learning_rate": 1.772023682466212e-05, "loss": 3.8024, "step": 50750 }, { "epoch": 0.5707127730686484, "grad_norm": 1.1962040662765503, "learning_rate": 1.7717989911359274e-05, "loss": 3.8789, "step": 50800 }, { "epoch": 0.5712744982389916, "grad_norm": 1.0278069972991943, "learning_rate": 1.7715742998056423e-05, "loss": 3.7437, "step": 50850 }, { "epoch": 0.5718362234093347, "grad_norm": 1.2031197547912598, "learning_rate": 1.7713496084753573e-05, "loss": 3.7805, "step": 50900 }, { "epoch": 0.572397948579678, "grad_norm": 1.1851876974105835, "learning_rate": 1.7711249171450722e-05, "loss": 3.8535, "step": 50950 }, { "epoch": 0.5729596737500211, "grad_norm": 1.0673991441726685, "learning_rate": 1.770900225814787e-05, "loss": 3.7427, "step": 51000 }, { "epoch": 0.5735213989203642, "grad_norm": 1.0498384237289429, "learning_rate": 1.770675534484502e-05, "loss": 3.7233, "step": 51050 }, { "epoch": 0.5740831240907074, "grad_norm": 1.1012871265411377, "learning_rate": 1.770450843154217e-05, "loss": 3.7578, "step": 51100 }, { "epoch": 0.5746448492610505, "grad_norm": 1.0736808776855469, "learning_rate": 1.770226151823932e-05, "loss": 3.8299, "step": 51150 }, { "epoch": 0.5752065744313937, "grad_norm": 1.1782079935073853, "learning_rate": 1.7700014604936472e-05, "loss": 3.8076, "step": 51200 }, { "epoch": 0.5757682996017368, "grad_norm": 1.1583595275878906, "learning_rate": 1.769776769163362e-05, "loss": 3.8288, "step": 51250 }, { "epoch": 0.5763300247720801, "grad_norm": 0.9925355911254883, "learning_rate": 1.769552077833077e-05, "loss": 3.7869, "step": 51300 }, { "epoch": 0.5768917499424232, "grad_norm": 1.0739439725875854, "learning_rate": 1.769327386502792e-05, "loss": 3.7328, "step": 51350 }, { "epoch": 0.5774534751127663, "grad_norm": 1.141987681388855, "learning_rate": 1.769102695172507e-05, "loss": 3.7463, "step": 51400 }, { "epoch": 0.5780152002831095, "grad_norm": 1.0860463380813599, "learning_rate": 1.768878003842222e-05, "loss": 3.7705, "step": 51450 }, { "epoch": 0.5785769254534526, "grad_norm": 1.0678012371063232, "learning_rate": 1.7686533125119368e-05, "loss": 3.745, "step": 51500 }, { "epoch": 0.5791386506237958, "grad_norm": 1.1007298231124878, "learning_rate": 1.768428621181652e-05, "loss": 3.7283, "step": 51550 }, { "epoch": 0.5797003757941389, "grad_norm": 1.037177562713623, "learning_rate": 1.768203929851367e-05, "loss": 3.823, "step": 51600 }, { "epoch": 0.5802621009644822, "grad_norm": 1.0051183700561523, "learning_rate": 1.767979238521082e-05, "loss": 3.7693, "step": 51650 }, { "epoch": 0.5808238261348253, "grad_norm": 1.1470272541046143, "learning_rate": 1.767754547190797e-05, "loss": 3.7568, "step": 51700 }, { "epoch": 0.5813855513051684, "grad_norm": 1.1250337362289429, "learning_rate": 1.7675298558605118e-05, "loss": 3.7366, "step": 51750 }, { "epoch": 0.5819472764755116, "grad_norm": 0.9637058973312378, "learning_rate": 1.7673051645302267e-05, "loss": 3.7163, "step": 51800 }, { "epoch": 0.5825090016458547, "grad_norm": 1.10556161403656, "learning_rate": 1.7670804731999416e-05, "loss": 3.8689, "step": 51850 }, { "epoch": 0.5830707268161979, "grad_norm": 0.9998056292533875, "learning_rate": 1.766855781869657e-05, "loss": 3.8945, "step": 51900 }, { "epoch": 0.583632451986541, "grad_norm": 1.2014474868774414, "learning_rate": 1.766631090539372e-05, "loss": 3.7499, "step": 51950 }, { "epoch": 0.5841941771568843, "grad_norm": 1.1735973358154297, "learning_rate": 1.7664063992090868e-05, "loss": 3.825, "step": 52000 }, { "epoch": 0.5847559023272274, "grad_norm": 1.0286104679107666, "learning_rate": 1.7661817078788017e-05, "loss": 3.7428, "step": 52050 }, { "epoch": 0.5853176274975705, "grad_norm": 1.306618094444275, "learning_rate": 1.7659570165485166e-05, "loss": 3.7499, "step": 52100 }, { "epoch": 0.5858793526679137, "grad_norm": 1.021728515625, "learning_rate": 1.7657323252182316e-05, "loss": 3.758, "step": 52150 }, { "epoch": 0.5864410778382568, "grad_norm": 1.2071304321289062, "learning_rate": 1.7655076338879465e-05, "loss": 3.772, "step": 52200 }, { "epoch": 0.5870028030086, "grad_norm": 1.4372752904891968, "learning_rate": 1.7652829425576618e-05, "loss": 3.8094, "step": 52250 }, { "epoch": 0.5875645281789432, "grad_norm": 1.0974149703979492, "learning_rate": 1.7650582512273767e-05, "loss": 3.8042, "step": 52300 }, { "epoch": 0.5881262533492864, "grad_norm": 1.1430492401123047, "learning_rate": 1.7648335598970916e-05, "loss": 3.7804, "step": 52350 }, { "epoch": 0.5886879785196295, "grad_norm": 1.0837328433990479, "learning_rate": 1.7646088685668066e-05, "loss": 3.724, "step": 52400 }, { "epoch": 0.5892497036899726, "grad_norm": 1.0210705995559692, "learning_rate": 1.7643841772365215e-05, "loss": 3.8235, "step": 52450 }, { "epoch": 0.5898114288603158, "grad_norm": 1.0554842948913574, "learning_rate": 1.7641594859062364e-05, "loss": 3.7633, "step": 52500 }, { "epoch": 0.5903731540306589, "grad_norm": 1.0610668659210205, "learning_rate": 1.7639347945759514e-05, "loss": 3.7446, "step": 52550 }, { "epoch": 0.5909348792010021, "grad_norm": 1.0977150201797485, "learning_rate": 1.7637101032456666e-05, "loss": 3.648, "step": 52600 }, { "epoch": 0.5914966043713453, "grad_norm": 1.229286789894104, "learning_rate": 1.7634854119153816e-05, "loss": 3.8027, "step": 52650 }, { "epoch": 0.5920583295416885, "grad_norm": 1.0663516521453857, "learning_rate": 1.7632607205850965e-05, "loss": 3.7696, "step": 52700 }, { "epoch": 0.5926200547120316, "grad_norm": 1.1510690450668335, "learning_rate": 1.7630360292548114e-05, "loss": 3.8121, "step": 52750 }, { "epoch": 0.5931817798823747, "grad_norm": 1.0974013805389404, "learning_rate": 1.7628113379245264e-05, "loss": 3.8296, "step": 52800 }, { "epoch": 0.5937435050527179, "grad_norm": 1.0291469097137451, "learning_rate": 1.7625866465942413e-05, "loss": 3.7295, "step": 52850 }, { "epoch": 0.594305230223061, "grad_norm": 1.1655592918395996, "learning_rate": 1.7623619552639562e-05, "loss": 3.7481, "step": 52900 }, { "epoch": 0.5948669553934042, "grad_norm": 1.3898149728775024, "learning_rate": 1.7621372639336715e-05, "loss": 3.8032, "step": 52950 }, { "epoch": 0.5954286805637474, "grad_norm": 0.9948254823684692, "learning_rate": 1.7619125726033864e-05, "loss": 3.7734, "step": 53000 }, { "epoch": 0.5959904057340906, "grad_norm": 1.1202493906021118, "learning_rate": 1.7616878812731014e-05, "loss": 3.7293, "step": 53050 }, { "epoch": 0.5965521309044337, "grad_norm": 1.0994091033935547, "learning_rate": 1.7614631899428163e-05, "loss": 3.7558, "step": 53100 }, { "epoch": 0.5971138560747768, "grad_norm": 1.0649958848953247, "learning_rate": 1.7612384986125312e-05, "loss": 3.7704, "step": 53150 }, { "epoch": 0.59767558124512, "grad_norm": 1.0792980194091797, "learning_rate": 1.761013807282246e-05, "loss": 3.8017, "step": 53200 }, { "epoch": 0.5982373064154631, "grad_norm": 1.0352792739868164, "learning_rate": 1.760789115951961e-05, "loss": 3.7618, "step": 53250 }, { "epoch": 0.5987990315858063, "grad_norm": 1.2318084239959717, "learning_rate": 1.7605644246216764e-05, "loss": 3.7041, "step": 53300 }, { "epoch": 0.5993607567561495, "grad_norm": 1.151821255683899, "learning_rate": 1.7603397332913913e-05, "loss": 3.792, "step": 53350 }, { "epoch": 0.5999224819264927, "grad_norm": 1.0399080514907837, "learning_rate": 1.7601150419611062e-05, "loss": 3.8003, "step": 53400 }, { "epoch": 0.6004842070968358, "grad_norm": 1.2682665586471558, "learning_rate": 1.759890350630821e-05, "loss": 3.7432, "step": 53450 }, { "epoch": 0.601045932267179, "grad_norm": 1.3421366214752197, "learning_rate": 1.759665659300536e-05, "loss": 3.8493, "step": 53500 }, { "epoch": 0.6016076574375221, "grad_norm": 1.2913321256637573, "learning_rate": 1.759440967970251e-05, "loss": 3.7462, "step": 53550 }, { "epoch": 0.6021693826078652, "grad_norm": 1.2670401334762573, "learning_rate": 1.759216276639966e-05, "loss": 3.8123, "step": 53600 }, { "epoch": 0.6027311077782085, "grad_norm": 0.8991925716400146, "learning_rate": 1.7589915853096812e-05, "loss": 3.767, "step": 53650 }, { "epoch": 0.6032928329485516, "grad_norm": 1.1282976865768433, "learning_rate": 1.758766893979396e-05, "loss": 3.7155, "step": 53700 }, { "epoch": 0.6038545581188948, "grad_norm": 1.3516185283660889, "learning_rate": 1.7585466964757164e-05, "loss": 3.7661, "step": 53750 }, { "epoch": 0.6044162832892379, "grad_norm": 1.5720571279525757, "learning_rate": 1.7583220051454313e-05, "loss": 3.7899, "step": 53800 }, { "epoch": 0.604978008459581, "grad_norm": 1.2103759050369263, "learning_rate": 1.7580973138151466e-05, "loss": 3.7905, "step": 53850 }, { "epoch": 0.6055397336299242, "grad_norm": 1.0293008089065552, "learning_rate": 1.7578726224848615e-05, "loss": 3.7961, "step": 53900 }, { "epoch": 0.6061014588002673, "grad_norm": 1.0778840780258179, "learning_rate": 1.7576479311545765e-05, "loss": 3.8131, "step": 53950 }, { "epoch": 0.6066631839706106, "grad_norm": 1.012994647026062, "learning_rate": 1.7574232398242914e-05, "loss": 3.7049, "step": 54000 }, { "epoch": 0.6072249091409537, "grad_norm": 1.2515206336975098, "learning_rate": 1.7571985484940063e-05, "loss": 3.7537, "step": 54050 }, { "epoch": 0.6077866343112969, "grad_norm": 1.3033223152160645, "learning_rate": 1.7569738571637213e-05, "loss": 3.8329, "step": 54100 }, { "epoch": 0.60834835948164, "grad_norm": 1.1931025981903076, "learning_rate": 1.7567491658334362e-05, "loss": 3.822, "step": 54150 }, { "epoch": 0.6089100846519832, "grad_norm": 0.8393592238426208, "learning_rate": 1.756528968329757e-05, "loss": 3.6741, "step": 54200 }, { "epoch": 0.6094718098223263, "grad_norm": 1.1479541063308716, "learning_rate": 1.756304276999472e-05, "loss": 3.6592, "step": 54250 }, { "epoch": 0.6100335349926694, "grad_norm": 1.1731626987457275, "learning_rate": 1.756079585669187e-05, "loss": 3.7719, "step": 54300 }, { "epoch": 0.6105952601630127, "grad_norm": 1.626962661743164, "learning_rate": 1.755854894338902e-05, "loss": 3.8338, "step": 54350 }, { "epoch": 0.6111569853333558, "grad_norm": 0.9360986351966858, "learning_rate": 1.755630203008617e-05, "loss": 3.758, "step": 54400 }, { "epoch": 0.611718710503699, "grad_norm": 1.112356185913086, "learning_rate": 1.7554055116783318e-05, "loss": 3.8045, "step": 54450 }, { "epoch": 0.6122804356740421, "grad_norm": 1.0053960084915161, "learning_rate": 1.755180820348047e-05, "loss": 3.7159, "step": 54500 }, { "epoch": 0.6128421608443853, "grad_norm": 1.3038718700408936, "learning_rate": 1.754956129017762e-05, "loss": 3.7571, "step": 54550 }, { "epoch": 0.6134038860147284, "grad_norm": 1.116718053817749, "learning_rate": 1.754731437687477e-05, "loss": 3.7673, "step": 54600 }, { "epoch": 0.6139656111850716, "grad_norm": 1.0109951496124268, "learning_rate": 1.754506746357192e-05, "loss": 3.8169, "step": 54650 }, { "epoch": 0.6145273363554148, "grad_norm": 1.0003641843795776, "learning_rate": 1.7542820550269068e-05, "loss": 3.6633, "step": 54700 }, { "epoch": 0.6150890615257579, "grad_norm": 1.0796260833740234, "learning_rate": 1.7540573636966217e-05, "loss": 3.7145, "step": 54750 }, { "epoch": 0.6156507866961011, "grad_norm": 1.163537859916687, "learning_rate": 1.7538326723663366e-05, "loss": 3.8101, "step": 54800 }, { "epoch": 0.6162125118664442, "grad_norm": 1.0152901411056519, "learning_rate": 1.753607981036052e-05, "loss": 3.7383, "step": 54850 }, { "epoch": 0.6167742370367874, "grad_norm": 0.9968041777610779, "learning_rate": 1.753383289705767e-05, "loss": 3.7473, "step": 54900 }, { "epoch": 0.6173359622071305, "grad_norm": 1.4523382186889648, "learning_rate": 1.7531585983754818e-05, "loss": 3.7098, "step": 54950 }, { "epoch": 0.6178976873774737, "grad_norm": 1.306656002998352, "learning_rate": 1.7529339070451967e-05, "loss": 3.6992, "step": 55000 }, { "epoch": 0.6184594125478169, "grad_norm": 1.0262917280197144, "learning_rate": 1.7527092157149116e-05, "loss": 3.826, "step": 55050 }, { "epoch": 0.61902113771816, "grad_norm": 1.0945398807525635, "learning_rate": 1.7524845243846266e-05, "loss": 3.836, "step": 55100 }, { "epoch": 0.6195828628885032, "grad_norm": 1.0034478902816772, "learning_rate": 1.7522598330543415e-05, "loss": 3.7401, "step": 55150 }, { "epoch": 0.6201445880588463, "grad_norm": 1.073586106300354, "learning_rate": 1.7520351417240568e-05, "loss": 3.7507, "step": 55200 }, { "epoch": 0.6207063132291895, "grad_norm": 1.35128653049469, "learning_rate": 1.7518104503937717e-05, "loss": 3.8347, "step": 55250 }, { "epoch": 0.6212680383995326, "grad_norm": 1.1628330945968628, "learning_rate": 1.7515857590634866e-05, "loss": 3.7984, "step": 55300 }, { "epoch": 0.6218297635698758, "grad_norm": 1.1050388813018799, "learning_rate": 1.7513610677332016e-05, "loss": 3.679, "step": 55350 }, { "epoch": 0.622391488740219, "grad_norm": 1.391716718673706, "learning_rate": 1.7511363764029165e-05, "loss": 3.7539, "step": 55400 }, { "epoch": 0.6229532139105621, "grad_norm": 1.0777093172073364, "learning_rate": 1.7509116850726314e-05, "loss": 3.8025, "step": 55450 }, { "epoch": 0.6235149390809053, "grad_norm": 1.2782025337219238, "learning_rate": 1.7506869937423464e-05, "loss": 3.7464, "step": 55500 }, { "epoch": 0.6240766642512484, "grad_norm": 0.9614410996437073, "learning_rate": 1.7504623024120616e-05, "loss": 3.7748, "step": 55550 }, { "epoch": 0.6246383894215916, "grad_norm": 1.1084448099136353, "learning_rate": 1.7502376110817766e-05, "loss": 3.7589, "step": 55600 }, { "epoch": 0.6252001145919347, "grad_norm": 1.0383925437927246, "learning_rate": 1.7500129197514915e-05, "loss": 3.7624, "step": 55650 }, { "epoch": 0.625761839762278, "grad_norm": 1.2714895009994507, "learning_rate": 1.7497882284212064e-05, "loss": 3.8049, "step": 55700 }, { "epoch": 0.6263235649326211, "grad_norm": 1.1721174716949463, "learning_rate": 1.7495635370909214e-05, "loss": 3.7391, "step": 55750 }, { "epoch": 0.6268852901029642, "grad_norm": 1.386572241783142, "learning_rate": 1.7493388457606363e-05, "loss": 3.6544, "step": 55800 }, { "epoch": 0.6274470152733074, "grad_norm": 1.186914086341858, "learning_rate": 1.7491141544303512e-05, "loss": 3.7147, "step": 55850 }, { "epoch": 0.6280087404436505, "grad_norm": 1.1338549852371216, "learning_rate": 1.7488894631000665e-05, "loss": 3.7277, "step": 55900 }, { "epoch": 0.6285704656139937, "grad_norm": 1.3115663528442383, "learning_rate": 1.7486647717697814e-05, "loss": 3.7401, "step": 55950 }, { "epoch": 0.6291321907843369, "grad_norm": 1.172100305557251, "learning_rate": 1.7484400804394964e-05, "loss": 3.8122, "step": 56000 }, { "epoch": 0.62969391595468, "grad_norm": 1.0340574979782104, "learning_rate": 1.7482153891092113e-05, "loss": 3.8087, "step": 56050 }, { "epoch": 0.6302556411250232, "grad_norm": 1.0701868534088135, "learning_rate": 1.7479906977789262e-05, "loss": 3.7115, "step": 56100 }, { "epoch": 0.6308173662953663, "grad_norm": 1.0869747400283813, "learning_rate": 1.747766006448641e-05, "loss": 3.7915, "step": 56150 }, { "epoch": 0.6313790914657095, "grad_norm": 1.1002542972564697, "learning_rate": 1.747541315118356e-05, "loss": 3.6988, "step": 56200 }, { "epoch": 0.6319408166360526, "grad_norm": 1.124596357345581, "learning_rate": 1.7473166237880714e-05, "loss": 3.78, "step": 56250 }, { "epoch": 0.6325025418063958, "grad_norm": 1.2207372188568115, "learning_rate": 1.7470919324577863e-05, "loss": 3.7913, "step": 56300 }, { "epoch": 0.633064266976739, "grad_norm": 1.2835536003112793, "learning_rate": 1.7468672411275012e-05, "loss": 3.7761, "step": 56350 }, { "epoch": 0.6336259921470822, "grad_norm": 1.2327477931976318, "learning_rate": 1.746642549797216e-05, "loss": 3.8194, "step": 56400 }, { "epoch": 0.6341877173174253, "grad_norm": 1.2116644382476807, "learning_rate": 1.7464223522935367e-05, "loss": 3.8478, "step": 56450 }, { "epoch": 0.6347494424877684, "grad_norm": 1.1993951797485352, "learning_rate": 1.7461976609632517e-05, "loss": 3.7858, "step": 56500 }, { "epoch": 0.6353111676581116, "grad_norm": 1.223915457725525, "learning_rate": 1.745972969632967e-05, "loss": 3.7529, "step": 56550 }, { "epoch": 0.6358728928284547, "grad_norm": 1.1108373403549194, "learning_rate": 1.745748278302682e-05, "loss": 3.7742, "step": 56600 }, { "epoch": 0.6364346179987979, "grad_norm": 1.0354841947555542, "learning_rate": 1.7455235869723968e-05, "loss": 3.7672, "step": 56650 }, { "epoch": 0.6369963431691411, "grad_norm": 1.059706687927246, "learning_rate": 1.7452988956421117e-05, "loss": 3.7729, "step": 56700 }, { "epoch": 0.6375580683394843, "grad_norm": 1.0827964544296265, "learning_rate": 1.7450742043118267e-05, "loss": 3.6867, "step": 56750 }, { "epoch": 0.6381197935098274, "grad_norm": 1.078853964805603, "learning_rate": 1.7448495129815416e-05, "loss": 3.7249, "step": 56800 }, { "epoch": 0.6386815186801705, "grad_norm": 1.1734760999679565, "learning_rate": 1.7446248216512565e-05, "loss": 3.8154, "step": 56850 }, { "epoch": 0.6392432438505137, "grad_norm": 1.1998199224472046, "learning_rate": 1.7444001303209718e-05, "loss": 3.7453, "step": 56900 }, { "epoch": 0.6398049690208568, "grad_norm": 1.1486512422561646, "learning_rate": 1.7441754389906867e-05, "loss": 3.7046, "step": 56950 }, { "epoch": 0.6403666941912, "grad_norm": 2.4619200229644775, "learning_rate": 1.7439507476604017e-05, "loss": 3.7254, "step": 57000 }, { "epoch": 0.6409284193615432, "grad_norm": 1.2601149082183838, "learning_rate": 1.7437260563301166e-05, "loss": 3.7013, "step": 57050 }, { "epoch": 0.6414901445318864, "grad_norm": 1.07093346118927, "learning_rate": 1.7435013649998315e-05, "loss": 3.8067, "step": 57100 }, { "epoch": 0.6420518697022295, "grad_norm": 1.187065839767456, "learning_rate": 1.7432766736695465e-05, "loss": 3.7688, "step": 57150 }, { "epoch": 0.6426135948725726, "grad_norm": 1.0475225448608398, "learning_rate": 1.7430519823392614e-05, "loss": 3.7031, "step": 57200 }, { "epoch": 0.6431753200429158, "grad_norm": 1.1916700601577759, "learning_rate": 1.7428272910089767e-05, "loss": 3.7313, "step": 57250 }, { "epoch": 0.6437370452132589, "grad_norm": 1.2383705377578735, "learning_rate": 1.7426025996786916e-05, "loss": 3.8268, "step": 57300 }, { "epoch": 0.6442987703836022, "grad_norm": 1.1119415760040283, "learning_rate": 1.7423779083484065e-05, "loss": 3.7676, "step": 57350 }, { "epoch": 0.6448604955539453, "grad_norm": 1.2077629566192627, "learning_rate": 1.7421532170181215e-05, "loss": 3.8255, "step": 57400 }, { "epoch": 0.6454222207242885, "grad_norm": 1.078245997428894, "learning_rate": 1.7419285256878364e-05, "loss": 3.7174, "step": 57450 }, { "epoch": 0.6459839458946316, "grad_norm": 1.1462548971176147, "learning_rate": 1.7417038343575513e-05, "loss": 3.8094, "step": 57500 }, { "epoch": 0.6465456710649747, "grad_norm": 1.1107356548309326, "learning_rate": 1.7414791430272663e-05, "loss": 3.8238, "step": 57550 }, { "epoch": 0.6471073962353179, "grad_norm": 0.9890028834342957, "learning_rate": 1.7412544516969815e-05, "loss": 3.8116, "step": 57600 }, { "epoch": 0.647669121405661, "grad_norm": 1.0611302852630615, "learning_rate": 1.7410297603666965e-05, "loss": 3.7203, "step": 57650 }, { "epoch": 0.6482308465760043, "grad_norm": 1.0499742031097412, "learning_rate": 1.7408050690364114e-05, "loss": 3.7378, "step": 57700 }, { "epoch": 0.6487925717463474, "grad_norm": 1.1752382516860962, "learning_rate": 1.7405803777061263e-05, "loss": 3.7246, "step": 57750 }, { "epoch": 0.6493542969166906, "grad_norm": 1.149659514427185, "learning_rate": 1.7403556863758413e-05, "loss": 3.7695, "step": 57800 }, { "epoch": 0.6499160220870337, "grad_norm": 0.9258421063423157, "learning_rate": 1.740135488872162e-05, "loss": 3.7976, "step": 57850 }, { "epoch": 0.6504777472573768, "grad_norm": 1.2829738855361938, "learning_rate": 1.739910797541877e-05, "loss": 3.7877, "step": 57900 }, { "epoch": 0.65103947242772, "grad_norm": 1.0089658498764038, "learning_rate": 1.739686106211592e-05, "loss": 3.7311, "step": 57950 }, { "epoch": 0.6516011975980631, "grad_norm": 1.102944016456604, "learning_rate": 1.739461414881307e-05, "loss": 3.7644, "step": 58000 }, { "epoch": 0.6521629227684064, "grad_norm": 1.213724136352539, "learning_rate": 1.739236723551022e-05, "loss": 3.7867, "step": 58050 }, { "epoch": 0.6527246479387495, "grad_norm": 1.135790467262268, "learning_rate": 1.739012032220737e-05, "loss": 3.7394, "step": 58100 }, { "epoch": 0.6532863731090927, "grad_norm": 1.1218922138214111, "learning_rate": 1.7387873408904518e-05, "loss": 3.8348, "step": 58150 }, { "epoch": 0.6538480982794358, "grad_norm": 1.5928314924240112, "learning_rate": 1.7385626495601667e-05, "loss": 3.6883, "step": 58200 }, { "epoch": 0.6544098234497789, "grad_norm": 1.0138171911239624, "learning_rate": 1.738337958229882e-05, "loss": 3.6912, "step": 58250 }, { "epoch": 0.6549715486201221, "grad_norm": 1.1871720552444458, "learning_rate": 1.738113266899597e-05, "loss": 3.7674, "step": 58300 }, { "epoch": 0.6555332737904653, "grad_norm": 1.1729800701141357, "learning_rate": 1.737888575569312e-05, "loss": 3.7707, "step": 58350 }, { "epoch": 0.6560949989608085, "grad_norm": 0.9070509076118469, "learning_rate": 1.7376638842390268e-05, "loss": 3.8358, "step": 58400 }, { "epoch": 0.6566567241311516, "grad_norm": 1.1585700511932373, "learning_rate": 1.7374391929087417e-05, "loss": 3.7481, "step": 58450 }, { "epoch": 0.6572184493014948, "grad_norm": 1.192552924156189, "learning_rate": 1.7372145015784566e-05, "loss": 3.7324, "step": 58500 }, { "epoch": 0.6577801744718379, "grad_norm": 1.0995116233825684, "learning_rate": 1.7369898102481716e-05, "loss": 3.814, "step": 58550 }, { "epoch": 0.658341899642181, "grad_norm": 1.1598093509674072, "learning_rate": 1.736765118917887e-05, "loss": 3.8403, "step": 58600 }, { "epoch": 0.6589036248125242, "grad_norm": 0.9769778847694397, "learning_rate": 1.7365404275876018e-05, "loss": 3.744, "step": 58650 }, { "epoch": 0.6594653499828674, "grad_norm": 1.102225661277771, "learning_rate": 1.7363157362573167e-05, "loss": 3.8054, "step": 58700 }, { "epoch": 0.6600270751532106, "grad_norm": 1.0175129175186157, "learning_rate": 1.7360910449270316e-05, "loss": 3.6721, "step": 58750 }, { "epoch": 0.6605888003235537, "grad_norm": 1.1308871507644653, "learning_rate": 1.7358663535967466e-05, "loss": 3.7017, "step": 58800 }, { "epoch": 0.6611505254938969, "grad_norm": 1.059259057044983, "learning_rate": 1.7356416622664615e-05, "loss": 3.7659, "step": 58850 }, { "epoch": 0.66171225066424, "grad_norm": 0.8596951961517334, "learning_rate": 1.7354169709361764e-05, "loss": 3.6726, "step": 58900 }, { "epoch": 0.6622739758345831, "grad_norm": 1.2974934577941895, "learning_rate": 1.7351922796058917e-05, "loss": 3.74, "step": 58950 }, { "epoch": 0.6628357010049263, "grad_norm": 1.1300395727157593, "learning_rate": 1.7349675882756066e-05, "loss": 3.7653, "step": 59000 }, { "epoch": 0.6633974261752695, "grad_norm": 1.097085952758789, "learning_rate": 1.7347428969453216e-05, "loss": 3.8094, "step": 59050 }, { "epoch": 0.6639591513456127, "grad_norm": 1.1728578805923462, "learning_rate": 1.7345182056150365e-05, "loss": 3.794, "step": 59100 }, { "epoch": 0.6645208765159558, "grad_norm": 1.0365574359893799, "learning_rate": 1.7342935142847514e-05, "loss": 3.7189, "step": 59150 }, { "epoch": 0.665082601686299, "grad_norm": 1.1213483810424805, "learning_rate": 1.7340688229544664e-05, "loss": 3.7355, "step": 59200 }, { "epoch": 0.6656443268566421, "grad_norm": 1.0311239957809448, "learning_rate": 1.7338441316241813e-05, "loss": 3.7611, "step": 59250 }, { "epoch": 0.6662060520269852, "grad_norm": 1.113905906677246, "learning_rate": 1.7336194402938966e-05, "loss": 3.7474, "step": 59300 }, { "epoch": 0.6667677771973284, "grad_norm": 0.9911109209060669, "learning_rate": 1.7333947489636115e-05, "loss": 3.751, "step": 59350 }, { "epoch": 0.6673295023676716, "grad_norm": 1.1860027313232422, "learning_rate": 1.7331700576333264e-05, "loss": 3.7477, "step": 59400 }, { "epoch": 0.6678912275380148, "grad_norm": 1.041394591331482, "learning_rate": 1.7329453663030414e-05, "loss": 3.6717, "step": 59450 }, { "epoch": 0.6684529527083579, "grad_norm": 1.0377204418182373, "learning_rate": 1.7327206749727563e-05, "loss": 3.7745, "step": 59500 }, { "epoch": 0.6690146778787011, "grad_norm": 1.1058553457260132, "learning_rate": 1.7324959836424712e-05, "loss": 3.7069, "step": 59550 }, { "epoch": 0.6695764030490442, "grad_norm": 1.2549515962600708, "learning_rate": 1.732275786138792e-05, "loss": 3.7848, "step": 59600 }, { "epoch": 0.6701381282193873, "grad_norm": 1.1881179809570312, "learning_rate": 1.732051094808507e-05, "loss": 3.8035, "step": 59650 }, { "epoch": 0.6706998533897306, "grad_norm": 1.087872862815857, "learning_rate": 1.731826403478222e-05, "loss": 3.7718, "step": 59700 }, { "epoch": 0.6712615785600737, "grad_norm": 1.0337735414505005, "learning_rate": 1.731601712147937e-05, "loss": 3.7516, "step": 59750 }, { "epoch": 0.6718233037304169, "grad_norm": 1.0685954093933105, "learning_rate": 1.731377020817652e-05, "loss": 3.7406, "step": 59800 }, { "epoch": 0.67238502890076, "grad_norm": 1.3235671520233154, "learning_rate": 1.7311523294873668e-05, "loss": 3.7622, "step": 59850 }, { "epoch": 0.6729467540711032, "grad_norm": 1.0132572650909424, "learning_rate": 1.7309276381570817e-05, "loss": 3.7176, "step": 59900 }, { "epoch": 0.6735084792414463, "grad_norm": 1.0859169960021973, "learning_rate": 1.730702946826797e-05, "loss": 3.7555, "step": 59950 }, { "epoch": 0.6740702044117894, "grad_norm": 1.0595098733901978, "learning_rate": 1.730478255496512e-05, "loss": 3.7683, "step": 60000 }, { "epoch": 0.6746319295821327, "grad_norm": 1.2074511051177979, "learning_rate": 1.730253564166227e-05, "loss": 3.7437, "step": 60050 }, { "epoch": 0.6751936547524758, "grad_norm": 1.5620605945587158, "learning_rate": 1.7300288728359418e-05, "loss": 3.7324, "step": 60100 }, { "epoch": 0.675755379922819, "grad_norm": 1.1325360536575317, "learning_rate": 1.7298041815056567e-05, "loss": 3.7337, "step": 60150 }, { "epoch": 0.6763171050931621, "grad_norm": 1.1669329404830933, "learning_rate": 1.7295794901753717e-05, "loss": 3.7386, "step": 60200 }, { "epoch": 0.6768788302635053, "grad_norm": 1.217720866203308, "learning_rate": 1.7293547988450866e-05, "loss": 3.7491, "step": 60250 }, { "epoch": 0.6774405554338484, "grad_norm": 1.1027276515960693, "learning_rate": 1.729130107514802e-05, "loss": 3.828, "step": 60300 }, { "epoch": 0.6780022806041915, "grad_norm": 1.3892951011657715, "learning_rate": 1.7289054161845168e-05, "loss": 3.8093, "step": 60350 }, { "epoch": 0.6785640057745348, "grad_norm": 1.090294599533081, "learning_rate": 1.7286807248542317e-05, "loss": 3.8163, "step": 60400 }, { "epoch": 0.6791257309448779, "grad_norm": 1.0623295307159424, "learning_rate": 1.7284560335239467e-05, "loss": 3.832, "step": 60450 }, { "epoch": 0.6796874561152211, "grad_norm": 1.1373815536499023, "learning_rate": 1.7282313421936616e-05, "loss": 3.7288, "step": 60500 }, { "epoch": 0.6802491812855642, "grad_norm": 1.2594444751739502, "learning_rate": 1.7280066508633765e-05, "loss": 3.8319, "step": 60550 }, { "epoch": 0.6808109064559074, "grad_norm": 1.963243007659912, "learning_rate": 1.7277819595330915e-05, "loss": 3.6892, "step": 60600 }, { "epoch": 0.6813726316262505, "grad_norm": 1.0691497325897217, "learning_rate": 1.7275572682028064e-05, "loss": 3.7616, "step": 60650 }, { "epoch": 0.6819343567965936, "grad_norm": 1.0307921171188354, "learning_rate": 1.7273325768725217e-05, "loss": 3.7629, "step": 60700 }, { "epoch": 0.6824960819669369, "grad_norm": 1.1926541328430176, "learning_rate": 1.7271078855422366e-05, "loss": 3.7875, "step": 60750 }, { "epoch": 0.68305780713728, "grad_norm": 1.0767431259155273, "learning_rate": 1.7268831942119515e-05, "loss": 3.6975, "step": 60800 }, { "epoch": 0.6836195323076232, "grad_norm": 1.0437264442443848, "learning_rate": 1.7266585028816665e-05, "loss": 3.7654, "step": 60850 }, { "epoch": 0.6841812574779663, "grad_norm": 1.0138980150222778, "learning_rate": 1.7264338115513814e-05, "loss": 3.766, "step": 60900 }, { "epoch": 0.6847429826483095, "grad_norm": 1.0664023160934448, "learning_rate": 1.7262091202210963e-05, "loss": 3.8365, "step": 60950 }, { "epoch": 0.6853047078186526, "grad_norm": 1.0536384582519531, "learning_rate": 1.7259844288908113e-05, "loss": 3.8036, "step": 61000 }, { "epoch": 0.6858664329889959, "grad_norm": 1.0200104713439941, "learning_rate": 1.7257597375605265e-05, "loss": 3.8067, "step": 61050 }, { "epoch": 0.686428158159339, "grad_norm": 1.2039310932159424, "learning_rate": 1.7255350462302415e-05, "loss": 3.8287, "step": 61100 }, { "epoch": 0.6869898833296821, "grad_norm": 1.0688422918319702, "learning_rate": 1.7253103548999564e-05, "loss": 3.6781, "step": 61150 }, { "epoch": 0.6875516085000253, "grad_norm": 1.2913188934326172, "learning_rate": 1.7250856635696713e-05, "loss": 3.7841, "step": 61200 }, { "epoch": 0.6881133336703684, "grad_norm": 0.9989462494850159, "learning_rate": 1.7248609722393863e-05, "loss": 3.7842, "step": 61250 }, { "epoch": 0.6886750588407116, "grad_norm": 1.056571125984192, "learning_rate": 1.7246362809091012e-05, "loss": 3.7838, "step": 61300 }, { "epoch": 0.6892367840110547, "grad_norm": 1.1829471588134766, "learning_rate": 1.724411589578816e-05, "loss": 3.7119, "step": 61350 }, { "epoch": 0.689798509181398, "grad_norm": 1.0671151876449585, "learning_rate": 1.7241868982485314e-05, "loss": 3.7321, "step": 61400 }, { "epoch": 0.6903602343517411, "grad_norm": 1.1066913604736328, "learning_rate": 1.7239622069182463e-05, "loss": 3.7692, "step": 61450 }, { "epoch": 0.6909219595220842, "grad_norm": 1.166175127029419, "learning_rate": 1.7237375155879612e-05, "loss": 3.736, "step": 61500 }, { "epoch": 0.6914836846924274, "grad_norm": 1.100232481956482, "learning_rate": 1.7235128242576762e-05, "loss": 3.8028, "step": 61550 }, { "epoch": 0.6920454098627705, "grad_norm": 1.2112360000610352, "learning_rate": 1.723288132927391e-05, "loss": 3.7698, "step": 61600 }, { "epoch": 0.6926071350331137, "grad_norm": 1.086358904838562, "learning_rate": 1.723063441597106e-05, "loss": 3.7637, "step": 61650 }, { "epoch": 0.6931688602034568, "grad_norm": 1.1099216938018799, "learning_rate": 1.722838750266821e-05, "loss": 3.7888, "step": 61700 }, { "epoch": 0.6937305853738001, "grad_norm": 0.988438606262207, "learning_rate": 1.7226140589365362e-05, "loss": 3.8313, "step": 61750 }, { "epoch": 0.6942923105441432, "grad_norm": 1.0678561925888062, "learning_rate": 1.7223893676062512e-05, "loss": 3.7451, "step": 61800 }, { "epoch": 0.6948540357144863, "grad_norm": 1.293940782546997, "learning_rate": 1.722164676275966e-05, "loss": 3.7087, "step": 61850 }, { "epoch": 0.6954157608848295, "grad_norm": 1.0884907245635986, "learning_rate": 1.721939984945681e-05, "loss": 3.7037, "step": 61900 }, { "epoch": 0.6959774860551726, "grad_norm": 1.0982439517974854, "learning_rate": 1.721715293615396e-05, "loss": 3.6668, "step": 61950 }, { "epoch": 0.6965392112255158, "grad_norm": 1.0753116607666016, "learning_rate": 1.721490602285111e-05, "loss": 3.6831, "step": 62000 }, { "epoch": 0.697100936395859, "grad_norm": 1.009289026260376, "learning_rate": 1.721265910954826e-05, "loss": 3.7106, "step": 62050 }, { "epoch": 0.6976626615662022, "grad_norm": 1.2232242822647095, "learning_rate": 1.721041219624541e-05, "loss": 3.7235, "step": 62100 }, { "epoch": 0.6982243867365453, "grad_norm": 1.4713053703308105, "learning_rate": 1.720816528294256e-05, "loss": 3.7499, "step": 62150 }, { "epoch": 0.6987861119068884, "grad_norm": 1.074111819267273, "learning_rate": 1.720591836963971e-05, "loss": 3.8075, "step": 62200 }, { "epoch": 0.6993478370772316, "grad_norm": 1.2166283130645752, "learning_rate": 1.720367145633686e-05, "loss": 3.6972, "step": 62250 }, { "epoch": 0.6999095622475747, "grad_norm": 1.212631344795227, "learning_rate": 1.720142454303401e-05, "loss": 3.7652, "step": 62300 }, { "epoch": 0.7004712874179179, "grad_norm": 0.9701681137084961, "learning_rate": 1.7199177629731158e-05, "loss": 3.7186, "step": 62350 }, { "epoch": 0.7010330125882611, "grad_norm": 1.0456335544586182, "learning_rate": 1.7196930716428307e-05, "loss": 3.8562, "step": 62400 }, { "epoch": 0.7015947377586043, "grad_norm": 1.0489026308059692, "learning_rate": 1.719468380312546e-05, "loss": 3.6795, "step": 62450 }, { "epoch": 0.7021564629289474, "grad_norm": 1.1351799964904785, "learning_rate": 1.719243688982261e-05, "loss": 3.766, "step": 62500 }, { "epoch": 0.7027181880992905, "grad_norm": 1.2965370416641235, "learning_rate": 1.7190189976519758e-05, "loss": 3.8151, "step": 62550 }, { "epoch": 0.7032799132696337, "grad_norm": 0.9872881174087524, "learning_rate": 1.7187943063216908e-05, "loss": 3.7155, "step": 62600 }, { "epoch": 0.7038416384399768, "grad_norm": 1.045161247253418, "learning_rate": 1.7185696149914057e-05, "loss": 3.8091, "step": 62650 }, { "epoch": 0.70440336361032, "grad_norm": 1.2162364721298218, "learning_rate": 1.7183449236611206e-05, "loss": 3.8024, "step": 62700 }, { "epoch": 0.7049650887806632, "grad_norm": 1.1977596282958984, "learning_rate": 1.7181202323308356e-05, "loss": 3.8153, "step": 62750 }, { "epoch": 0.7055268139510064, "grad_norm": 1.1208105087280273, "learning_rate": 1.7178955410005508e-05, "loss": 3.778, "step": 62800 }, { "epoch": 0.7060885391213495, "grad_norm": 1.4856350421905518, "learning_rate": 1.7176708496702658e-05, "loss": 3.7663, "step": 62850 }, { "epoch": 0.7066502642916926, "grad_norm": 1.2106389999389648, "learning_rate": 1.7174461583399807e-05, "loss": 3.7157, "step": 62900 }, { "epoch": 0.7072119894620358, "grad_norm": 1.0710458755493164, "learning_rate": 1.7172214670096956e-05, "loss": 3.7688, "step": 62950 }, { "epoch": 0.7077737146323789, "grad_norm": 1.059740424156189, "learning_rate": 1.7169967756794106e-05, "loss": 3.6703, "step": 63000 }, { "epoch": 0.7083354398027221, "grad_norm": 1.0456651449203491, "learning_rate": 1.7167720843491255e-05, "loss": 3.7715, "step": 63050 }, { "epoch": 0.7088971649730653, "grad_norm": 1.1711243391036987, "learning_rate": 1.7165473930188404e-05, "loss": 3.7545, "step": 63100 }, { "epoch": 0.7094588901434085, "grad_norm": 1.1514251232147217, "learning_rate": 1.7163227016885557e-05, "loss": 3.7125, "step": 63150 }, { "epoch": 0.7100206153137516, "grad_norm": 1.2115432024002075, "learning_rate": 1.7160980103582706e-05, "loss": 3.7972, "step": 63200 }, { "epoch": 0.7105823404840947, "grad_norm": 1.1310622692108154, "learning_rate": 1.7158733190279855e-05, "loss": 3.7626, "step": 63250 }, { "epoch": 0.7111440656544379, "grad_norm": 1.085196614265442, "learning_rate": 1.7156486276977005e-05, "loss": 3.8308, "step": 63300 }, { "epoch": 0.711705790824781, "grad_norm": 1.0386687517166138, "learning_rate": 1.7154239363674154e-05, "loss": 3.8163, "step": 63350 }, { "epoch": 0.7122675159951243, "grad_norm": 1.1654423475265503, "learning_rate": 1.7151992450371303e-05, "loss": 3.6759, "step": 63400 }, { "epoch": 0.7128292411654674, "grad_norm": 1.0651174783706665, "learning_rate": 1.7149745537068453e-05, "loss": 3.717, "step": 63450 }, { "epoch": 0.7133909663358106, "grad_norm": 1.1307576894760132, "learning_rate": 1.7147498623765605e-05, "loss": 3.8315, "step": 63500 }, { "epoch": 0.7139526915061537, "grad_norm": 1.1386432647705078, "learning_rate": 1.7145251710462755e-05, "loss": 3.7392, "step": 63550 }, { "epoch": 0.7145144166764968, "grad_norm": 1.1393775939941406, "learning_rate": 1.7143004797159904e-05, "loss": 3.705, "step": 63600 }, { "epoch": 0.71507614184684, "grad_norm": 1.118025779724121, "learning_rate": 1.7140757883857053e-05, "loss": 3.699, "step": 63650 }, { "epoch": 0.7156378670171831, "grad_norm": 0.9614808559417725, "learning_rate": 1.7138510970554203e-05, "loss": 3.7437, "step": 63700 }, { "epoch": 0.7161995921875264, "grad_norm": 1.0394705533981323, "learning_rate": 1.7136264057251352e-05, "loss": 3.7325, "step": 63750 }, { "epoch": 0.7167613173578695, "grad_norm": 1.0388416051864624, "learning_rate": 1.71340171439485e-05, "loss": 3.7791, "step": 63800 }, { "epoch": 0.7173230425282127, "grad_norm": 1.1777559518814087, "learning_rate": 1.713177023064565e-05, "loss": 3.7322, "step": 63850 }, { "epoch": 0.7178847676985558, "grad_norm": 1.1553236246109009, "learning_rate": 1.7129523317342803e-05, "loss": 3.7408, "step": 63900 }, { "epoch": 0.718446492868899, "grad_norm": 1.136378526687622, "learning_rate": 1.7127276404039953e-05, "loss": 3.8013, "step": 63950 }, { "epoch": 0.7190082180392421, "grad_norm": 1.2489324808120728, "learning_rate": 1.7125029490737102e-05, "loss": 3.8406, "step": 64000 }, { "epoch": 0.7195699432095852, "grad_norm": 1.0770765542984009, "learning_rate": 1.712278257743425e-05, "loss": 3.7281, "step": 64050 }, { "epoch": 0.7201316683799285, "grad_norm": 0.9865077137947083, "learning_rate": 1.71205356641314e-05, "loss": 3.7968, "step": 64100 }, { "epoch": 0.7206933935502716, "grad_norm": 1.0182303190231323, "learning_rate": 1.711828875082855e-05, "loss": 3.6896, "step": 64150 }, { "epoch": 0.7212551187206148, "grad_norm": 0.9671421051025391, "learning_rate": 1.71160418375257e-05, "loss": 3.8275, "step": 64200 }, { "epoch": 0.7218168438909579, "grad_norm": 1.1231694221496582, "learning_rate": 1.7113794924222852e-05, "loss": 3.6555, "step": 64250 }, { "epoch": 0.722378569061301, "grad_norm": 1.1389505863189697, "learning_rate": 1.711154801092e-05, "loss": 3.7869, "step": 64300 }, { "epoch": 0.7229402942316442, "grad_norm": 1.0590343475341797, "learning_rate": 1.710930109761715e-05, "loss": 3.6597, "step": 64350 }, { "epoch": 0.7235020194019873, "grad_norm": 1.0033551454544067, "learning_rate": 1.71070541843143e-05, "loss": 3.8088, "step": 64400 }, { "epoch": 0.7240637445723306, "grad_norm": 1.258637547492981, "learning_rate": 1.710480727101145e-05, "loss": 3.734, "step": 64450 }, { "epoch": 0.7246254697426737, "grad_norm": 1.2087608575820923, "learning_rate": 1.71025603577086e-05, "loss": 3.7397, "step": 64500 }, { "epoch": 0.7251871949130169, "grad_norm": 1.1399991512298584, "learning_rate": 1.7100313444405748e-05, "loss": 3.762, "step": 64550 }, { "epoch": 0.72574892008336, "grad_norm": 0.8821276426315308, "learning_rate": 1.70980665311029e-05, "loss": 3.6886, "step": 64600 }, { "epoch": 0.7263106452537031, "grad_norm": 1.2095016241073608, "learning_rate": 1.709581961780005e-05, "loss": 3.7076, "step": 64650 }, { "epoch": 0.7268723704240463, "grad_norm": 1.2419086694717407, "learning_rate": 1.70935727044972e-05, "loss": 3.8058, "step": 64700 }, { "epoch": 0.7274340955943895, "grad_norm": 1.1642667055130005, "learning_rate": 1.709132579119435e-05, "loss": 3.6804, "step": 64750 }, { "epoch": 0.7279958207647327, "grad_norm": 1.1226531267166138, "learning_rate": 1.7089078877891498e-05, "loss": 3.693, "step": 64800 }, { "epoch": 0.7285575459350758, "grad_norm": 0.9891392588615417, "learning_rate": 1.7086831964588647e-05, "loss": 3.7193, "step": 64850 }, { "epoch": 0.729119271105419, "grad_norm": 1.0887752771377563, "learning_rate": 1.7084585051285796e-05, "loss": 3.765, "step": 64900 }, { "epoch": 0.7296809962757621, "grad_norm": 1.1410701274871826, "learning_rate": 1.708233813798295e-05, "loss": 3.7251, "step": 64950 }, { "epoch": 0.7302427214461052, "grad_norm": 1.1968011856079102, "learning_rate": 1.70800912246801e-05, "loss": 3.6613, "step": 65000 }, { "epoch": 0.7308044466164484, "grad_norm": 1.0664294958114624, "learning_rate": 1.7077844311377248e-05, "loss": 3.7821, "step": 65050 }, { "epoch": 0.7313661717867916, "grad_norm": 1.238824725151062, "learning_rate": 1.7075597398074397e-05, "loss": 3.7638, "step": 65100 }, { "epoch": 0.7319278969571348, "grad_norm": 1.074196457862854, "learning_rate": 1.7073350484771546e-05, "loss": 3.7331, "step": 65150 }, { "epoch": 0.7324896221274779, "grad_norm": 1.1788239479064941, "learning_rate": 1.7071103571468696e-05, "loss": 3.6592, "step": 65200 }, { "epoch": 0.7330513472978211, "grad_norm": 1.2634332180023193, "learning_rate": 1.7068856658165845e-05, "loss": 3.7664, "step": 65250 }, { "epoch": 0.7336130724681642, "grad_norm": 0.9647150635719299, "learning_rate": 1.7066609744862998e-05, "loss": 3.7154, "step": 65300 }, { "epoch": 0.7341747976385073, "grad_norm": 1.1323342323303223, "learning_rate": 1.7064362831560147e-05, "loss": 3.7684, "step": 65350 }, { "epoch": 0.7347365228088505, "grad_norm": 1.0603891611099243, "learning_rate": 1.7062115918257296e-05, "loss": 3.8007, "step": 65400 }, { "epoch": 0.7352982479791937, "grad_norm": 1.0535929203033447, "learning_rate": 1.7059869004954446e-05, "loss": 3.7819, "step": 65450 }, { "epoch": 0.7358599731495369, "grad_norm": 1.1961077451705933, "learning_rate": 1.7057622091651595e-05, "loss": 3.8177, "step": 65500 }, { "epoch": 0.73642169831988, "grad_norm": 1.076874017715454, "learning_rate": 1.7055375178348744e-05, "loss": 3.6888, "step": 65550 }, { "epoch": 0.7369834234902232, "grad_norm": 1.0798044204711914, "learning_rate": 1.7053128265045894e-05, "loss": 3.8633, "step": 65600 }, { "epoch": 0.7375451486605663, "grad_norm": 1.0618581771850586, "learning_rate": 1.7050881351743046e-05, "loss": 3.7193, "step": 65650 }, { "epoch": 0.7381068738309094, "grad_norm": 1.2281302213668823, "learning_rate": 1.7048634438440196e-05, "loss": 3.6743, "step": 65700 }, { "epoch": 0.7386685990012527, "grad_norm": 1.03543221950531, "learning_rate": 1.7046387525137345e-05, "loss": 3.8051, "step": 65750 }, { "epoch": 0.7392303241715958, "grad_norm": 0.8722898364067078, "learning_rate": 1.7044140611834494e-05, "loss": 3.7792, "step": 65800 }, { "epoch": 0.739792049341939, "grad_norm": 0.9191735982894897, "learning_rate": 1.7041893698531644e-05, "loss": 3.8145, "step": 65850 }, { "epoch": 0.7403537745122821, "grad_norm": 0.9913333058357239, "learning_rate": 1.7039646785228793e-05, "loss": 3.7993, "step": 65900 }, { "epoch": 0.7409154996826253, "grad_norm": 1.242875576019287, "learning_rate": 1.7037399871925942e-05, "loss": 3.6952, "step": 65950 }, { "epoch": 0.7414772248529684, "grad_norm": 1.1154929399490356, "learning_rate": 1.7035152958623095e-05, "loss": 3.7458, "step": 66000 }, { "epoch": 0.7420389500233115, "grad_norm": 1.2310267686843872, "learning_rate": 1.7032906045320244e-05, "loss": 3.7183, "step": 66050 }, { "epoch": 0.7426006751936548, "grad_norm": 1.2243191003799438, "learning_rate": 1.7030659132017394e-05, "loss": 3.662, "step": 66100 }, { "epoch": 0.7431624003639979, "grad_norm": 0.9692927002906799, "learning_rate": 1.7028412218714543e-05, "loss": 3.6818, "step": 66150 }, { "epoch": 0.7437241255343411, "grad_norm": 1.0961863994598389, "learning_rate": 1.7026165305411692e-05, "loss": 3.7774, "step": 66200 }, { "epoch": 0.7442858507046842, "grad_norm": 1.1476986408233643, "learning_rate": 1.702391839210884e-05, "loss": 3.779, "step": 66250 }, { "epoch": 0.7448475758750274, "grad_norm": 1.0727260112762451, "learning_rate": 1.702167147880599e-05, "loss": 3.7399, "step": 66300 }, { "epoch": 0.7454093010453705, "grad_norm": 1.2328996658325195, "learning_rate": 1.7019424565503144e-05, "loss": 3.681, "step": 66350 }, { "epoch": 0.7459710262157137, "grad_norm": 1.0276544094085693, "learning_rate": 1.7017177652200293e-05, "loss": 3.7449, "step": 66400 }, { "epoch": 0.7465327513860569, "grad_norm": 1.0378201007843018, "learning_rate": 1.7014930738897442e-05, "loss": 3.7523, "step": 66450 }, { "epoch": 0.7470944765564, "grad_norm": 0.9312450885772705, "learning_rate": 1.701268382559459e-05, "loss": 3.7095, "step": 66500 }, { "epoch": 0.7476562017267432, "grad_norm": 1.17537522315979, "learning_rate": 1.701043691229174e-05, "loss": 3.7707, "step": 66550 }, { "epoch": 0.7482179268970863, "grad_norm": 0.9856059551239014, "learning_rate": 1.700818999898889e-05, "loss": 3.7125, "step": 66600 }, { "epoch": 0.7487796520674295, "grad_norm": 1.0982396602630615, "learning_rate": 1.700594308568604e-05, "loss": 3.6904, "step": 66650 }, { "epoch": 0.7493413772377726, "grad_norm": 1.0933263301849365, "learning_rate": 1.7003696172383192e-05, "loss": 3.8068, "step": 66700 }, { "epoch": 0.7499031024081158, "grad_norm": 1.0702201128005981, "learning_rate": 1.700144925908034e-05, "loss": 3.7596, "step": 66750 }, { "epoch": 0.750464827578459, "grad_norm": 1.0075763463974, "learning_rate": 1.699920234577749e-05, "loss": 3.7165, "step": 66800 }, { "epoch": 0.7510265527488021, "grad_norm": 1.0466583967208862, "learning_rate": 1.699695543247464e-05, "loss": 3.727, "step": 66850 }, { "epoch": 0.7515882779191453, "grad_norm": 1.0358221530914307, "learning_rate": 1.699470851917179e-05, "loss": 3.78, "step": 66900 }, { "epoch": 0.7521500030894884, "grad_norm": 1.0360231399536133, "learning_rate": 1.699246160586894e-05, "loss": 3.781, "step": 66950 }, { "epoch": 0.7527117282598316, "grad_norm": 1.3516156673431396, "learning_rate": 1.6990214692566088e-05, "loss": 3.7346, "step": 67000 }, { "epoch": 0.7532734534301747, "grad_norm": 1.1824400424957275, "learning_rate": 1.698796777926324e-05, "loss": 3.7438, "step": 67050 }, { "epoch": 0.753835178600518, "grad_norm": 1.1349762678146362, "learning_rate": 1.698572086596039e-05, "loss": 3.7134, "step": 67100 }, { "epoch": 0.7543969037708611, "grad_norm": 1.0137215852737427, "learning_rate": 1.698347395265754e-05, "loss": 3.7641, "step": 67150 }, { "epoch": 0.7549586289412042, "grad_norm": 1.0727646350860596, "learning_rate": 1.698122703935469e-05, "loss": 3.7501, "step": 67200 }, { "epoch": 0.7555203541115474, "grad_norm": 1.266788125038147, "learning_rate": 1.6979025064317895e-05, "loss": 3.7392, "step": 67250 }, { "epoch": 0.7560820792818905, "grad_norm": 1.1174174547195435, "learning_rate": 1.6976778151015044e-05, "loss": 3.7231, "step": 67300 }, { "epoch": 0.7566438044522337, "grad_norm": 1.0874195098876953, "learning_rate": 1.6974531237712197e-05, "loss": 3.7797, "step": 67350 }, { "epoch": 0.7572055296225768, "grad_norm": 1.1917657852172852, "learning_rate": 1.6972284324409346e-05, "loss": 3.6537, "step": 67400 }, { "epoch": 0.7577672547929201, "grad_norm": 1.0290495157241821, "learning_rate": 1.6970037411106495e-05, "loss": 3.6701, "step": 67450 }, { "epoch": 0.7583289799632632, "grad_norm": 1.1170862913131714, "learning_rate": 1.6967790497803645e-05, "loss": 3.7691, "step": 67500 }, { "epoch": 0.7588907051336063, "grad_norm": 1.0585973262786865, "learning_rate": 1.6965543584500794e-05, "loss": 3.7705, "step": 67550 }, { "epoch": 0.7594524303039495, "grad_norm": 1.1106197834014893, "learning_rate": 1.6963296671197943e-05, "loss": 3.7491, "step": 67600 }, { "epoch": 0.7600141554742926, "grad_norm": 1.280634880065918, "learning_rate": 1.6961049757895093e-05, "loss": 3.7504, "step": 67650 }, { "epoch": 0.7605758806446358, "grad_norm": 1.1722393035888672, "learning_rate": 1.6958802844592245e-05, "loss": 3.8211, "step": 67700 }, { "epoch": 0.7611376058149789, "grad_norm": 1.1163361072540283, "learning_rate": 1.6956555931289395e-05, "loss": 3.6904, "step": 67750 }, { "epoch": 0.7616993309853222, "grad_norm": 1.0139960050582886, "learning_rate": 1.6954309017986544e-05, "loss": 3.7935, "step": 67800 }, { "epoch": 0.7622610561556653, "grad_norm": 1.1248722076416016, "learning_rate": 1.6952062104683693e-05, "loss": 3.6819, "step": 67850 }, { "epoch": 0.7628227813260084, "grad_norm": 1.1061235666275024, "learning_rate": 1.6949815191380843e-05, "loss": 3.6705, "step": 67900 }, { "epoch": 0.7633845064963516, "grad_norm": 1.1143214702606201, "learning_rate": 1.6947568278077992e-05, "loss": 3.7279, "step": 67950 }, { "epoch": 0.7639462316666947, "grad_norm": 1.142398476600647, "learning_rate": 1.694532136477514e-05, "loss": 3.6923, "step": 68000 }, { "epoch": 0.7645079568370379, "grad_norm": 1.0227819681167603, "learning_rate": 1.6943074451472294e-05, "loss": 3.7693, "step": 68050 }, { "epoch": 0.765069682007381, "grad_norm": 1.331196665763855, "learning_rate": 1.6940827538169443e-05, "loss": 3.773, "step": 68100 }, { "epoch": 0.7656314071777243, "grad_norm": 0.9705712199211121, "learning_rate": 1.6938580624866593e-05, "loss": 3.7086, "step": 68150 }, { "epoch": 0.7661931323480674, "grad_norm": 1.227713942527771, "learning_rate": 1.6936333711563742e-05, "loss": 3.7594, "step": 68200 }, { "epoch": 0.7667548575184105, "grad_norm": 1.119600534439087, "learning_rate": 1.693408679826089e-05, "loss": 3.8176, "step": 68250 }, { "epoch": 0.7673165826887537, "grad_norm": 1.046105146408081, "learning_rate": 1.693183988495804e-05, "loss": 3.7069, "step": 68300 }, { "epoch": 0.7678783078590968, "grad_norm": 1.4448554515838623, "learning_rate": 1.692959297165519e-05, "loss": 3.6717, "step": 68350 }, { "epoch": 0.76844003302944, "grad_norm": 1.1133235692977905, "learning_rate": 1.692734605835234e-05, "loss": 3.709, "step": 68400 }, { "epoch": 0.7690017581997832, "grad_norm": 1.3116734027862549, "learning_rate": 1.6925099145049492e-05, "loss": 3.7271, "step": 68450 }, { "epoch": 0.7695634833701264, "grad_norm": 1.1115169525146484, "learning_rate": 1.692285223174664e-05, "loss": 3.7566, "step": 68500 }, { "epoch": 0.7701252085404695, "grad_norm": 1.179801106452942, "learning_rate": 1.692060531844379e-05, "loss": 3.786, "step": 68550 }, { "epoch": 0.7706869337108126, "grad_norm": 0.9484862089157104, "learning_rate": 1.691835840514094e-05, "loss": 3.7471, "step": 68600 }, { "epoch": 0.7712486588811558, "grad_norm": 1.0878593921661377, "learning_rate": 1.691611149183809e-05, "loss": 3.7765, "step": 68650 }, { "epoch": 0.7718103840514989, "grad_norm": 1.1715834140777588, "learning_rate": 1.691386457853524e-05, "loss": 3.7477, "step": 68700 }, { "epoch": 0.7723721092218421, "grad_norm": 1.1169553995132446, "learning_rate": 1.6911617665232388e-05, "loss": 3.7611, "step": 68750 }, { "epoch": 0.7729338343921853, "grad_norm": 1.0869131088256836, "learning_rate": 1.690937075192954e-05, "loss": 3.7331, "step": 68800 }, { "epoch": 0.7734955595625285, "grad_norm": 1.1528838872909546, "learning_rate": 1.690712383862669e-05, "loss": 3.7113, "step": 68850 }, { "epoch": 0.7740572847328716, "grad_norm": 1.2826603651046753, "learning_rate": 1.690487692532384e-05, "loss": 3.7586, "step": 68900 }, { "epoch": 0.7746190099032148, "grad_norm": 1.2303800582885742, "learning_rate": 1.690263001202099e-05, "loss": 3.7164, "step": 68950 }, { "epoch": 0.7751807350735579, "grad_norm": 0.8959895968437195, "learning_rate": 1.6900428036984194e-05, "loss": 3.7172, "step": 69000 }, { "epoch": 0.775742460243901, "grad_norm": 0.96584552526474, "learning_rate": 1.6898181123681344e-05, "loss": 3.6761, "step": 69050 }, { "epoch": 0.7763041854142442, "grad_norm": 1.37136971950531, "learning_rate": 1.6895934210378493e-05, "loss": 3.755, "step": 69100 }, { "epoch": 0.7768659105845874, "grad_norm": 2.480153799057007, "learning_rate": 1.6893687297075642e-05, "loss": 3.7863, "step": 69150 }, { "epoch": 0.7774276357549306, "grad_norm": 1.323038101196289, "learning_rate": 1.689144038377279e-05, "loss": 3.7252, "step": 69200 }, { "epoch": 0.7779893609252737, "grad_norm": 1.1072900295257568, "learning_rate": 1.688919347046994e-05, "loss": 3.6942, "step": 69250 }, { "epoch": 0.7785510860956169, "grad_norm": 1.2455112934112549, "learning_rate": 1.6886946557167094e-05, "loss": 3.7065, "step": 69300 }, { "epoch": 0.77911281126596, "grad_norm": 1.305863618850708, "learning_rate": 1.6884699643864243e-05, "loss": 3.7533, "step": 69350 }, { "epoch": 0.7796745364363031, "grad_norm": 1.0050400495529175, "learning_rate": 1.6882452730561392e-05, "loss": 3.7948, "step": 69400 }, { "epoch": 0.7802362616066464, "grad_norm": 0.9906814098358154, "learning_rate": 1.688020581725854e-05, "loss": 3.7531, "step": 69450 }, { "epoch": 0.7807979867769895, "grad_norm": 1.166979432106018, "learning_rate": 1.687795890395569e-05, "loss": 3.721, "step": 69500 }, { "epoch": 0.7813597119473327, "grad_norm": 1.264424443244934, "learning_rate": 1.687571199065284e-05, "loss": 3.7448, "step": 69550 }, { "epoch": 0.7819214371176758, "grad_norm": 1.1750078201293945, "learning_rate": 1.687346507734999e-05, "loss": 3.7879, "step": 69600 }, { "epoch": 0.782483162288019, "grad_norm": 1.0952597856521606, "learning_rate": 1.6871218164047142e-05, "loss": 3.7952, "step": 69650 }, { "epoch": 0.7830448874583621, "grad_norm": 1.0783863067626953, "learning_rate": 1.686897125074429e-05, "loss": 3.7517, "step": 69700 }, { "epoch": 0.7836066126287052, "grad_norm": 1.161342740058899, "learning_rate": 1.686672433744144e-05, "loss": 3.753, "step": 69750 }, { "epoch": 0.7841683377990485, "grad_norm": 1.1949788331985474, "learning_rate": 1.686447742413859e-05, "loss": 3.7667, "step": 69800 }, { "epoch": 0.7847300629693916, "grad_norm": 1.2240240573883057, "learning_rate": 1.686223051083574e-05, "loss": 3.7418, "step": 69850 }, { "epoch": 0.7852917881397348, "grad_norm": 1.1848735809326172, "learning_rate": 1.685998359753289e-05, "loss": 3.7769, "step": 69900 }, { "epoch": 0.7858535133100779, "grad_norm": 0.9923997521400452, "learning_rate": 1.6857736684230038e-05, "loss": 3.7339, "step": 69950 }, { "epoch": 0.786415238480421, "grad_norm": 1.0285292863845825, "learning_rate": 1.685548977092719e-05, "loss": 3.6968, "step": 70000 }, { "epoch": 0.7869769636507642, "grad_norm": 1.2908881902694702, "learning_rate": 1.685324285762434e-05, "loss": 3.7946, "step": 70050 }, { "epoch": 0.7875386888211073, "grad_norm": 1.19662344455719, "learning_rate": 1.685099594432149e-05, "loss": 3.7436, "step": 70100 }, { "epoch": 0.7881004139914506, "grad_norm": 1.0358531475067139, "learning_rate": 1.684874903101864e-05, "loss": 3.7474, "step": 70150 }, { "epoch": 0.7886621391617937, "grad_norm": 1.1050183773040771, "learning_rate": 1.6846502117715788e-05, "loss": 3.6748, "step": 70200 }, { "epoch": 0.7892238643321369, "grad_norm": 1.2210628986358643, "learning_rate": 1.6844255204412937e-05, "loss": 3.6993, "step": 70250 }, { "epoch": 0.78978558950248, "grad_norm": 1.192805528640747, "learning_rate": 1.6842008291110087e-05, "loss": 3.7257, "step": 70300 }, { "epoch": 0.7903473146728232, "grad_norm": 1.0933042764663696, "learning_rate": 1.683976137780724e-05, "loss": 3.6633, "step": 70350 }, { "epoch": 0.7909090398431663, "grad_norm": 1.2619465589523315, "learning_rate": 1.683751446450439e-05, "loss": 3.8322, "step": 70400 }, { "epoch": 0.7914707650135094, "grad_norm": 1.2451398372650146, "learning_rate": 1.6835267551201538e-05, "loss": 3.7935, "step": 70450 }, { "epoch": 0.7920324901838527, "grad_norm": 1.280688762664795, "learning_rate": 1.6833020637898687e-05, "loss": 3.7188, "step": 70500 }, { "epoch": 0.7925942153541958, "grad_norm": 1.074777364730835, "learning_rate": 1.6830773724595837e-05, "loss": 3.8526, "step": 70550 }, { "epoch": 0.793155940524539, "grad_norm": 1.1776268482208252, "learning_rate": 1.6828526811292986e-05, "loss": 3.7506, "step": 70600 }, { "epoch": 0.7937176656948821, "grad_norm": 0.954605758190155, "learning_rate": 1.6826279897990135e-05, "loss": 3.749, "step": 70650 }, { "epoch": 0.7942793908652253, "grad_norm": 1.364682674407959, "learning_rate": 1.6824032984687285e-05, "loss": 3.7188, "step": 70700 }, { "epoch": 0.7948411160355684, "grad_norm": 1.1577703952789307, "learning_rate": 1.6821786071384437e-05, "loss": 3.6976, "step": 70750 }, { "epoch": 0.7954028412059116, "grad_norm": 1.2413133382797241, "learning_rate": 1.6819539158081587e-05, "loss": 3.8732, "step": 70800 }, { "epoch": 0.7959645663762548, "grad_norm": 0.9502367973327637, "learning_rate": 1.6817292244778736e-05, "loss": 3.7627, "step": 70850 }, { "epoch": 0.7965262915465979, "grad_norm": 1.2539441585540771, "learning_rate": 1.6815045331475885e-05, "loss": 3.796, "step": 70900 }, { "epoch": 0.7970880167169411, "grad_norm": 3.9363365173339844, "learning_rate": 1.6812798418173035e-05, "loss": 3.7875, "step": 70950 }, { "epoch": 0.7976497418872842, "grad_norm": 1.077473759651184, "learning_rate": 1.6810551504870184e-05, "loss": 3.7428, "step": 71000 }, { "epoch": 0.7982114670576274, "grad_norm": 1.1866124868392944, "learning_rate": 1.6808304591567333e-05, "loss": 3.7312, "step": 71050 }, { "epoch": 0.7987731922279705, "grad_norm": 1.7925058603286743, "learning_rate": 1.6806057678264486e-05, "loss": 3.794, "step": 71100 }, { "epoch": 0.7993349173983137, "grad_norm": 1.7285633087158203, "learning_rate": 1.6803810764961635e-05, "loss": 3.8283, "step": 71150 }, { "epoch": 0.7998966425686569, "grad_norm": 1.1912788152694702, "learning_rate": 1.6801563851658785e-05, "loss": 3.7616, "step": 71200 }, { "epoch": 0.800458367739, "grad_norm": 0.989124059677124, "learning_rate": 1.6799316938355934e-05, "loss": 3.7482, "step": 71250 }, { "epoch": 0.8010200929093432, "grad_norm": 1.122393250465393, "learning_rate": 1.6797070025053083e-05, "loss": 3.7222, "step": 71300 }, { "epoch": 0.8015818180796863, "grad_norm": 0.9863786101341248, "learning_rate": 1.6794823111750233e-05, "loss": 3.6859, "step": 71350 }, { "epoch": 0.8021435432500295, "grad_norm": 1.1153448820114136, "learning_rate": 1.6792576198447382e-05, "loss": 3.7667, "step": 71400 }, { "epoch": 0.8027052684203726, "grad_norm": 1.0489387512207031, "learning_rate": 1.6790329285144535e-05, "loss": 3.6631, "step": 71450 }, { "epoch": 0.8032669935907158, "grad_norm": 1.1601120233535767, "learning_rate": 1.6788082371841684e-05, "loss": 3.7535, "step": 71500 }, { "epoch": 0.803828718761059, "grad_norm": 1.1417239904403687, "learning_rate": 1.6785835458538833e-05, "loss": 3.6421, "step": 71550 }, { "epoch": 0.8043904439314021, "grad_norm": 1.6279891729354858, "learning_rate": 1.6783588545235983e-05, "loss": 3.8245, "step": 71600 }, { "epoch": 0.8049521691017453, "grad_norm": 0.9623803496360779, "learning_rate": 1.6781341631933132e-05, "loss": 3.7142, "step": 71650 }, { "epoch": 0.8055138942720884, "grad_norm": 1.1247252225875854, "learning_rate": 1.677909471863028e-05, "loss": 3.7255, "step": 71700 }, { "epoch": 0.8060756194424316, "grad_norm": 1.153218388557434, "learning_rate": 1.677684780532743e-05, "loss": 3.691, "step": 71750 }, { "epoch": 0.8066373446127748, "grad_norm": 1.188510775566101, "learning_rate": 1.6774600892024583e-05, "loss": 3.7395, "step": 71800 }, { "epoch": 0.807199069783118, "grad_norm": 1.0337402820587158, "learning_rate": 1.6772353978721732e-05, "loss": 3.7563, "step": 71850 }, { "epoch": 0.8077607949534611, "grad_norm": 1.3426690101623535, "learning_rate": 1.6770107065418882e-05, "loss": 3.7542, "step": 71900 }, { "epoch": 0.8083225201238042, "grad_norm": 1.0045791864395142, "learning_rate": 1.676786015211603e-05, "loss": 3.7913, "step": 71950 }, { "epoch": 0.8088842452941474, "grad_norm": 1.1135921478271484, "learning_rate": 1.676561323881318e-05, "loss": 3.7338, "step": 72000 }, { "epoch": 0.8094459704644905, "grad_norm": 0.9275521039962769, "learning_rate": 1.676336632551033e-05, "loss": 3.6551, "step": 72050 }, { "epoch": 0.8100076956348337, "grad_norm": 1.1273943185806274, "learning_rate": 1.676111941220748e-05, "loss": 3.6895, "step": 72100 }, { "epoch": 0.8105694208051769, "grad_norm": 1.1047978401184082, "learning_rate": 1.6758872498904632e-05, "loss": 3.6553, "step": 72150 }, { "epoch": 0.81113114597552, "grad_norm": 1.4752873182296753, "learning_rate": 1.675662558560178e-05, "loss": 3.7694, "step": 72200 }, { "epoch": 0.8116928711458632, "grad_norm": 1.2518274784088135, "learning_rate": 1.675437867229893e-05, "loss": 3.6904, "step": 72250 }, { "epoch": 0.8122545963162063, "grad_norm": 1.0234527587890625, "learning_rate": 1.675213175899608e-05, "loss": 3.7434, "step": 72300 }, { "epoch": 0.8128163214865495, "grad_norm": 1.005708932876587, "learning_rate": 1.674988484569323e-05, "loss": 3.7529, "step": 72350 }, { "epoch": 0.8133780466568926, "grad_norm": 1.0410258769989014, "learning_rate": 1.674763793239038e-05, "loss": 3.7049, "step": 72400 }, { "epoch": 0.8139397718272358, "grad_norm": 1.0871193408966064, "learning_rate": 1.6745391019087528e-05, "loss": 3.7236, "step": 72450 }, { "epoch": 0.814501496997579, "grad_norm": 1.1557812690734863, "learning_rate": 1.674314410578468e-05, "loss": 3.7443, "step": 72500 }, { "epoch": 0.8150632221679222, "grad_norm": 1.1970670223236084, "learning_rate": 1.674089719248183e-05, "loss": 3.7485, "step": 72550 }, { "epoch": 0.8156249473382653, "grad_norm": 1.3610422611236572, "learning_rate": 1.673865027917898e-05, "loss": 3.7038, "step": 72600 }, { "epoch": 0.8161866725086084, "grad_norm": 1.1171669960021973, "learning_rate": 1.673640336587613e-05, "loss": 3.7223, "step": 72650 }, { "epoch": 0.8167483976789516, "grad_norm": 1.2198951244354248, "learning_rate": 1.6734156452573278e-05, "loss": 3.7455, "step": 72700 }, { "epoch": 0.8173101228492947, "grad_norm": 1.0705602169036865, "learning_rate": 1.6731909539270427e-05, "loss": 3.7316, "step": 72750 }, { "epoch": 0.8178718480196379, "grad_norm": 1.0684411525726318, "learning_rate": 1.6729662625967576e-05, "loss": 3.74, "step": 72800 }, { "epoch": 0.8184335731899811, "grad_norm": 1.1701427698135376, "learning_rate": 1.672741571266473e-05, "loss": 3.7103, "step": 72850 }, { "epoch": 0.8189952983603243, "grad_norm": 1.0043283700942993, "learning_rate": 1.6725168799361878e-05, "loss": 3.7895, "step": 72900 }, { "epoch": 0.8195570235306674, "grad_norm": 1.2557294368743896, "learning_rate": 1.6722921886059028e-05, "loss": 3.7005, "step": 72950 }, { "epoch": 0.8201187487010105, "grad_norm": 1.1772818565368652, "learning_rate": 1.6720674972756177e-05, "loss": 3.6787, "step": 73000 }, { "epoch": 0.8206804738713537, "grad_norm": 1.0652531385421753, "learning_rate": 1.6718428059453326e-05, "loss": 3.6785, "step": 73050 }, { "epoch": 0.8212421990416968, "grad_norm": 1.3456592559814453, "learning_rate": 1.6716181146150476e-05, "loss": 3.8048, "step": 73100 }, { "epoch": 0.8218039242120401, "grad_norm": 1.377454161643982, "learning_rate": 1.6713934232847625e-05, "loss": 3.7803, "step": 73150 }, { "epoch": 0.8223656493823832, "grad_norm": 1.0337969064712524, "learning_rate": 1.6711687319544778e-05, "loss": 3.778, "step": 73200 }, { "epoch": 0.8229273745527264, "grad_norm": 1.2524175643920898, "learning_rate": 1.6709440406241927e-05, "loss": 3.6722, "step": 73250 }, { "epoch": 0.8234890997230695, "grad_norm": 1.074506163597107, "learning_rate": 1.6707193492939076e-05, "loss": 3.602, "step": 73300 }, { "epoch": 0.8240508248934126, "grad_norm": 1.0358381271362305, "learning_rate": 1.6704946579636226e-05, "loss": 3.7684, "step": 73350 }, { "epoch": 0.8246125500637558, "grad_norm": 1.223362922668457, "learning_rate": 1.6702699666333375e-05, "loss": 3.7429, "step": 73400 }, { "epoch": 0.8251742752340989, "grad_norm": 1.2116211652755737, "learning_rate": 1.6700452753030524e-05, "loss": 3.7784, "step": 73450 }, { "epoch": 0.8257360004044422, "grad_norm": 1.03231942653656, "learning_rate": 1.6698205839727673e-05, "loss": 3.7228, "step": 73500 }, { "epoch": 0.8262977255747853, "grad_norm": 1.0249361991882324, "learning_rate": 1.6695958926424826e-05, "loss": 3.6967, "step": 73550 }, { "epoch": 0.8268594507451285, "grad_norm": 1.1133832931518555, "learning_rate": 1.6693756951388032e-05, "loss": 3.7659, "step": 73600 }, { "epoch": 0.8274211759154716, "grad_norm": 1.1546299457550049, "learning_rate": 1.669151003808518e-05, "loss": 3.6815, "step": 73650 }, { "epoch": 0.8279829010858147, "grad_norm": 1.14987313747406, "learning_rate": 1.668926312478233e-05, "loss": 3.7479, "step": 73700 }, { "epoch": 0.8285446262561579, "grad_norm": 1.496004343032837, "learning_rate": 1.668701621147948e-05, "loss": 3.7435, "step": 73750 }, { "epoch": 0.829106351426501, "grad_norm": 1.0407507419586182, "learning_rate": 1.668476929817663e-05, "loss": 3.8376, "step": 73800 }, { "epoch": 0.8296680765968443, "grad_norm": 1.1707595586776733, "learning_rate": 1.6682522384873782e-05, "loss": 3.6745, "step": 73850 }, { "epoch": 0.8302298017671874, "grad_norm": 1.1068360805511475, "learning_rate": 1.668027547157093e-05, "loss": 3.6911, "step": 73900 }, { "epoch": 0.8307915269375306, "grad_norm": 0.9708052277565002, "learning_rate": 1.667802855826808e-05, "loss": 3.7387, "step": 73950 }, { "epoch": 0.8313532521078737, "grad_norm": 1.101502537727356, "learning_rate": 1.667578164496523e-05, "loss": 3.7328, "step": 74000 }, { "epoch": 0.8319149772782168, "grad_norm": 1.345021367073059, "learning_rate": 1.667353473166238e-05, "loss": 3.7497, "step": 74050 }, { "epoch": 0.83247670244856, "grad_norm": 1.2573903799057007, "learning_rate": 1.667128781835953e-05, "loss": 3.8261, "step": 74100 }, { "epoch": 0.8330384276189031, "grad_norm": 1.2525533437728882, "learning_rate": 1.6669085843322738e-05, "loss": 3.7412, "step": 74150 }, { "epoch": 0.8336001527892464, "grad_norm": 1.1022635698318481, "learning_rate": 1.6666838930019887e-05, "loss": 3.7042, "step": 74200 }, { "epoch": 0.8341618779595895, "grad_norm": 1.1093451976776123, "learning_rate": 1.6664592016717037e-05, "loss": 3.714, "step": 74250 }, { "epoch": 0.8347236031299327, "grad_norm": 1.1967084407806396, "learning_rate": 1.6662345103414186e-05, "loss": 3.7316, "step": 74300 }, { "epoch": 0.8352853283002758, "grad_norm": 1.1166884899139404, "learning_rate": 1.6660098190111335e-05, "loss": 3.7497, "step": 74350 }, { "epoch": 0.8358470534706189, "grad_norm": 1.2759777307510376, "learning_rate": 1.6657851276808485e-05, "loss": 3.7377, "step": 74400 }, { "epoch": 0.8364087786409621, "grad_norm": 1.3579269647598267, "learning_rate": 1.6655604363505634e-05, "loss": 3.7868, "step": 74450 }, { "epoch": 0.8369705038113053, "grad_norm": 1.010847568511963, "learning_rate": 1.6653357450202787e-05, "loss": 3.7852, "step": 74500 }, { "epoch": 0.8375322289816485, "grad_norm": 1.3050994873046875, "learning_rate": 1.6651110536899936e-05, "loss": 3.7719, "step": 74550 }, { "epoch": 0.8380939541519916, "grad_norm": 1.1435143947601318, "learning_rate": 1.6648863623597085e-05, "loss": 3.7672, "step": 74600 }, { "epoch": 0.8386556793223348, "grad_norm": 1.064864993095398, "learning_rate": 1.6646616710294235e-05, "loss": 3.6735, "step": 74650 }, { "epoch": 0.8392174044926779, "grad_norm": 1.2209954261779785, "learning_rate": 1.6644369796991384e-05, "loss": 3.6891, "step": 74700 }, { "epoch": 0.839779129663021, "grad_norm": 1.390276312828064, "learning_rate": 1.6642122883688533e-05, "loss": 3.7265, "step": 74750 }, { "epoch": 0.8403408548333642, "grad_norm": 1.1255176067352295, "learning_rate": 1.6639875970385682e-05, "loss": 3.7504, "step": 74800 }, { "epoch": 0.8409025800037074, "grad_norm": 1.1659913063049316, "learning_rate": 1.6637629057082835e-05, "loss": 3.7539, "step": 74850 }, { "epoch": 0.8414643051740506, "grad_norm": 1.150168776512146, "learning_rate": 1.6635382143779984e-05, "loss": 3.7142, "step": 74900 }, { "epoch": 0.8420260303443937, "grad_norm": 1.237899899482727, "learning_rate": 1.6633135230477134e-05, "loss": 3.7639, "step": 74950 }, { "epoch": 0.8425877555147369, "grad_norm": 1.293717384338379, "learning_rate": 1.6630888317174283e-05, "loss": 3.7224, "step": 75000 }, { "epoch": 0.84314948068508, "grad_norm": 0.9956775903701782, "learning_rate": 1.6628641403871432e-05, "loss": 3.7572, "step": 75050 }, { "epoch": 0.8437112058554231, "grad_norm": 1.0118999481201172, "learning_rate": 1.6626394490568582e-05, "loss": 3.7923, "step": 75100 }, { "epoch": 0.8442729310257663, "grad_norm": 1.2486600875854492, "learning_rate": 1.662414757726573e-05, "loss": 3.7677, "step": 75150 }, { "epoch": 0.8448346561961095, "grad_norm": 1.0034174919128418, "learning_rate": 1.6621900663962884e-05, "loss": 3.7479, "step": 75200 }, { "epoch": 0.8453963813664527, "grad_norm": 1.0463954210281372, "learning_rate": 1.6619653750660033e-05, "loss": 3.6765, "step": 75250 }, { "epoch": 0.8459581065367958, "grad_norm": 1.0026686191558838, "learning_rate": 1.6617406837357182e-05, "loss": 3.6909, "step": 75300 }, { "epoch": 0.846519831707139, "grad_norm": 1.2389321327209473, "learning_rate": 1.6615159924054332e-05, "loss": 3.8334, "step": 75350 }, { "epoch": 0.8470815568774821, "grad_norm": 1.0024304389953613, "learning_rate": 1.661291301075148e-05, "loss": 3.6617, "step": 75400 }, { "epoch": 0.8476432820478252, "grad_norm": 1.1129672527313232, "learning_rate": 1.661066609744863e-05, "loss": 3.7321, "step": 75450 }, { "epoch": 0.8482050072181685, "grad_norm": 1.1076396703720093, "learning_rate": 1.660841918414578e-05, "loss": 3.6988, "step": 75500 }, { "epoch": 0.8487667323885116, "grad_norm": 1.0923964977264404, "learning_rate": 1.6606172270842932e-05, "loss": 3.7346, "step": 75550 }, { "epoch": 0.8493284575588548, "grad_norm": 1.1150474548339844, "learning_rate": 1.6603925357540082e-05, "loss": 3.7394, "step": 75600 }, { "epoch": 0.8498901827291979, "grad_norm": 1.1458702087402344, "learning_rate": 1.660167844423723e-05, "loss": 3.7441, "step": 75650 }, { "epoch": 0.8504519078995411, "grad_norm": 1.1019251346588135, "learning_rate": 1.659943153093438e-05, "loss": 3.7684, "step": 75700 }, { "epoch": 0.8510136330698842, "grad_norm": 1.418558120727539, "learning_rate": 1.659718461763153e-05, "loss": 3.7951, "step": 75750 }, { "epoch": 0.8515753582402273, "grad_norm": 1.058002233505249, "learning_rate": 1.659493770432868e-05, "loss": 3.7021, "step": 75800 }, { "epoch": 0.8521370834105706, "grad_norm": 1.3031290769577026, "learning_rate": 1.6592690791025828e-05, "loss": 3.7322, "step": 75850 }, { "epoch": 0.8526988085809137, "grad_norm": 1.255540132522583, "learning_rate": 1.6590443877722978e-05, "loss": 3.7647, "step": 75900 }, { "epoch": 0.8532605337512569, "grad_norm": 1.2593520879745483, "learning_rate": 1.658819696442013e-05, "loss": 3.6771, "step": 75950 }, { "epoch": 0.8538222589216, "grad_norm": 1.0368095636367798, "learning_rate": 1.658595005111728e-05, "loss": 3.6938, "step": 76000 }, { "epoch": 0.8543839840919432, "grad_norm": 1.133135199546814, "learning_rate": 1.658370313781443e-05, "loss": 3.7244, "step": 76050 }, { "epoch": 0.8549457092622863, "grad_norm": 1.1056394577026367, "learning_rate": 1.6581456224511578e-05, "loss": 3.6896, "step": 76100 }, { "epoch": 0.8555074344326294, "grad_norm": 1.148077368736267, "learning_rate": 1.6579209311208728e-05, "loss": 3.7582, "step": 76150 }, { "epoch": 0.8560691596029727, "grad_norm": 1.1761956214904785, "learning_rate": 1.6576962397905877e-05, "loss": 3.738, "step": 76200 }, { "epoch": 0.8566308847733158, "grad_norm": 1.1236742734909058, "learning_rate": 1.6574715484603026e-05, "loss": 3.7053, "step": 76250 }, { "epoch": 0.857192609943659, "grad_norm": 1.1957411766052246, "learning_rate": 1.657246857130018e-05, "loss": 3.8153, "step": 76300 }, { "epoch": 0.8577543351140021, "grad_norm": 1.0819793939590454, "learning_rate": 1.6570221657997328e-05, "loss": 3.7302, "step": 76350 }, { "epoch": 0.8583160602843453, "grad_norm": 1.1566054821014404, "learning_rate": 1.6567974744694478e-05, "loss": 3.7149, "step": 76400 }, { "epoch": 0.8588777854546884, "grad_norm": 1.377386450767517, "learning_rate": 1.6565727831391627e-05, "loss": 3.6913, "step": 76450 }, { "epoch": 0.8594395106250315, "grad_norm": 1.092625617980957, "learning_rate": 1.6563480918088776e-05, "loss": 3.6786, "step": 76500 }, { "epoch": 0.8600012357953748, "grad_norm": 1.1253376007080078, "learning_rate": 1.6561234004785925e-05, "loss": 3.7482, "step": 76550 }, { "epoch": 0.8605629609657179, "grad_norm": 1.1006630659103394, "learning_rate": 1.6558987091483075e-05, "loss": 3.7111, "step": 76600 }, { "epoch": 0.8611246861360611, "grad_norm": 1.213733196258545, "learning_rate": 1.6556740178180228e-05, "loss": 3.6594, "step": 76650 }, { "epoch": 0.8616864113064042, "grad_norm": 1.0177408456802368, "learning_rate": 1.6554493264877377e-05, "loss": 3.7734, "step": 76700 }, { "epoch": 0.8622481364767474, "grad_norm": 1.1692383289337158, "learning_rate": 1.6552246351574526e-05, "loss": 3.6968, "step": 76750 }, { "epoch": 0.8628098616470905, "grad_norm": 1.2537058591842651, "learning_rate": 1.6549999438271675e-05, "loss": 3.7714, "step": 76800 }, { "epoch": 0.8633715868174338, "grad_norm": 1.1537922620773315, "learning_rate": 1.6547752524968825e-05, "loss": 3.7202, "step": 76850 }, { "epoch": 0.8639333119877769, "grad_norm": 1.0910826921463013, "learning_rate": 1.6545505611665974e-05, "loss": 3.7392, "step": 76900 }, { "epoch": 0.86449503715812, "grad_norm": 1.245719075202942, "learning_rate": 1.6543258698363123e-05, "loss": 3.7756, "step": 76950 }, { "epoch": 0.8650567623284632, "grad_norm": 1.1607590913772583, "learning_rate": 1.6541011785060276e-05, "loss": 3.7172, "step": 77000 }, { "epoch": 0.8656184874988063, "grad_norm": 1.1951005458831787, "learning_rate": 1.6538764871757425e-05, "loss": 3.7168, "step": 77050 }, { "epoch": 0.8661802126691495, "grad_norm": 1.0030970573425293, "learning_rate": 1.6536517958454575e-05, "loss": 3.7622, "step": 77100 }, { "epoch": 0.8667419378394926, "grad_norm": 1.0266594886779785, "learning_rate": 1.6534271045151724e-05, "loss": 3.7146, "step": 77150 }, { "epoch": 0.8673036630098359, "grad_norm": 1.2694624662399292, "learning_rate": 1.6532024131848873e-05, "loss": 3.7644, "step": 77200 }, { "epoch": 0.867865388180179, "grad_norm": 1.3341777324676514, "learning_rate": 1.6529777218546023e-05, "loss": 3.7883, "step": 77250 }, { "epoch": 0.8684271133505221, "grad_norm": 1.3228938579559326, "learning_rate": 1.6527530305243172e-05, "loss": 3.7568, "step": 77300 }, { "epoch": 0.8689888385208653, "grad_norm": 1.1019830703735352, "learning_rate": 1.6525283391940325e-05, "loss": 3.7358, "step": 77350 }, { "epoch": 0.8695505636912084, "grad_norm": 1.214781641960144, "learning_rate": 1.6523036478637474e-05, "loss": 3.6864, "step": 77400 }, { "epoch": 0.8701122888615516, "grad_norm": 1.3680715560913086, "learning_rate": 1.6520789565334623e-05, "loss": 3.7394, "step": 77450 }, { "epoch": 0.8706740140318947, "grad_norm": 1.1794883012771606, "learning_rate": 1.6518542652031773e-05, "loss": 3.7502, "step": 77500 }, { "epoch": 0.871235739202238, "grad_norm": 1.2064045667648315, "learning_rate": 1.6516295738728922e-05, "loss": 3.7202, "step": 77550 }, { "epoch": 0.8717974643725811, "grad_norm": 1.033135175704956, "learning_rate": 1.651404882542607e-05, "loss": 3.6429, "step": 77600 }, { "epoch": 0.8723591895429242, "grad_norm": 1.030489206314087, "learning_rate": 1.651180191212322e-05, "loss": 3.7168, "step": 77650 }, { "epoch": 0.8729209147132674, "grad_norm": 1.1649249792099, "learning_rate": 1.6509554998820373e-05, "loss": 3.7608, "step": 77700 }, { "epoch": 0.8734826398836105, "grad_norm": 1.425357460975647, "learning_rate": 1.6507308085517523e-05, "loss": 3.857, "step": 77750 }, { "epoch": 0.8740443650539537, "grad_norm": 1.0895726680755615, "learning_rate": 1.6505061172214672e-05, "loss": 3.7146, "step": 77800 }, { "epoch": 0.8746060902242968, "grad_norm": 1.160264253616333, "learning_rate": 1.650281425891182e-05, "loss": 3.7892, "step": 77850 }, { "epoch": 0.8751678153946401, "grad_norm": 1.16293203830719, "learning_rate": 1.650056734560897e-05, "loss": 3.7371, "step": 77900 }, { "epoch": 0.8757295405649832, "grad_norm": 1.2008315324783325, "learning_rate": 1.649832043230612e-05, "loss": 3.6677, "step": 77950 }, { "epoch": 0.8762912657353263, "grad_norm": 1.1021394729614258, "learning_rate": 1.649607351900327e-05, "loss": 3.7296, "step": 78000 }, { "epoch": 0.8768529909056695, "grad_norm": 1.0809792280197144, "learning_rate": 1.6493826605700422e-05, "loss": 3.7592, "step": 78050 }, { "epoch": 0.8774147160760126, "grad_norm": 1.251360535621643, "learning_rate": 1.649157969239757e-05, "loss": 3.7237, "step": 78100 }, { "epoch": 0.8779764412463558, "grad_norm": 1.233972430229187, "learning_rate": 1.648933277909472e-05, "loss": 3.6951, "step": 78150 }, { "epoch": 0.878538166416699, "grad_norm": 1.0440303087234497, "learning_rate": 1.648708586579187e-05, "loss": 3.7149, "step": 78200 }, { "epoch": 0.8790998915870422, "grad_norm": 1.253597378730774, "learning_rate": 1.648483895248902e-05, "loss": 3.7138, "step": 78250 }, { "epoch": 0.8796616167573853, "grad_norm": 1.2321425676345825, "learning_rate": 1.648259203918617e-05, "loss": 3.714, "step": 78300 }, { "epoch": 0.8802233419277284, "grad_norm": 1.1274776458740234, "learning_rate": 1.6480345125883318e-05, "loss": 3.7227, "step": 78350 }, { "epoch": 0.8807850670980716, "grad_norm": 1.1239445209503174, "learning_rate": 1.647809821258047e-05, "loss": 3.6846, "step": 78400 }, { "epoch": 0.8813467922684147, "grad_norm": 0.9932996034622192, "learning_rate": 1.647585129927762e-05, "loss": 3.7678, "step": 78450 }, { "epoch": 0.8819085174387579, "grad_norm": 1.1654386520385742, "learning_rate": 1.647360438597477e-05, "loss": 3.7241, "step": 78500 }, { "epoch": 0.8824702426091011, "grad_norm": 1.1453014612197876, "learning_rate": 1.647135747267192e-05, "loss": 3.7153, "step": 78550 }, { "epoch": 0.8830319677794443, "grad_norm": 1.1624643802642822, "learning_rate": 1.6469110559369068e-05, "loss": 3.8312, "step": 78600 }, { "epoch": 0.8835936929497874, "grad_norm": 1.084373116493225, "learning_rate": 1.6466863646066217e-05, "loss": 3.7449, "step": 78650 }, { "epoch": 0.8841554181201305, "grad_norm": 0.9288941025733948, "learning_rate": 1.6464616732763366e-05, "loss": 3.7183, "step": 78700 }, { "epoch": 0.8847171432904737, "grad_norm": 0.990007221698761, "learning_rate": 1.646236981946052e-05, "loss": 3.6809, "step": 78750 }, { "epoch": 0.8852788684608168, "grad_norm": 1.068748116493225, "learning_rate": 1.646012290615767e-05, "loss": 3.8225, "step": 78800 }, { "epoch": 0.88584059363116, "grad_norm": 1.0568158626556396, "learning_rate": 1.6457875992854818e-05, "loss": 3.7485, "step": 78850 }, { "epoch": 0.8864023188015032, "grad_norm": 1.3895962238311768, "learning_rate": 1.6455629079551967e-05, "loss": 3.6925, "step": 78900 }, { "epoch": 0.8869640439718464, "grad_norm": 1.1887255907058716, "learning_rate": 1.6453382166249116e-05, "loss": 3.7772, "step": 78950 }, { "epoch": 0.8875257691421895, "grad_norm": 1.065321445465088, "learning_rate": 1.6451135252946266e-05, "loss": 3.702, "step": 79000 }, { "epoch": 0.8880874943125326, "grad_norm": 1.207621455192566, "learning_rate": 1.6448888339643415e-05, "loss": 3.7194, "step": 79050 }, { "epoch": 0.8886492194828758, "grad_norm": 2.230903148651123, "learning_rate": 1.6446641426340564e-05, "loss": 3.7189, "step": 79100 }, { "epoch": 0.8892109446532189, "grad_norm": 1.2009443044662476, "learning_rate": 1.6444394513037717e-05, "loss": 3.7664, "step": 79150 }, { "epoch": 0.8897726698235622, "grad_norm": 1.0513954162597656, "learning_rate": 1.6442192538000923e-05, "loss": 3.7869, "step": 79200 }, { "epoch": 0.8903343949939053, "grad_norm": 1.2765376567840576, "learning_rate": 1.6439945624698072e-05, "loss": 3.6946, "step": 79250 }, { "epoch": 0.8908961201642485, "grad_norm": 1.2688416242599487, "learning_rate": 1.643769871139522e-05, "loss": 3.7589, "step": 79300 }, { "epoch": 0.8914578453345916, "grad_norm": 0.9216362237930298, "learning_rate": 1.643545179809237e-05, "loss": 3.7159, "step": 79350 }, { "epoch": 0.8920195705049347, "grad_norm": 1.1441956758499146, "learning_rate": 1.6433204884789524e-05, "loss": 3.7159, "step": 79400 }, { "epoch": 0.8925812956752779, "grad_norm": 1.182159662246704, "learning_rate": 1.6430957971486673e-05, "loss": 3.8107, "step": 79450 }, { "epoch": 0.893143020845621, "grad_norm": 1.177890419960022, "learning_rate": 1.6428711058183822e-05, "loss": 3.7341, "step": 79500 }, { "epoch": 0.8937047460159643, "grad_norm": 1.2310272455215454, "learning_rate": 1.642646414488097e-05, "loss": 3.7716, "step": 79550 }, { "epoch": 0.8942664711863074, "grad_norm": 1.9204705953598022, "learning_rate": 1.642421723157812e-05, "loss": 3.7327, "step": 79600 }, { "epoch": 0.8948281963566506, "grad_norm": 1.085312843322754, "learning_rate": 1.642197031827527e-05, "loss": 3.7847, "step": 79650 }, { "epoch": 0.8953899215269937, "grad_norm": 1.079974889755249, "learning_rate": 1.641972340497242e-05, "loss": 3.7471, "step": 79700 }, { "epoch": 0.8959516466973368, "grad_norm": 1.1707812547683716, "learning_rate": 1.6417476491669572e-05, "loss": 3.7397, "step": 79750 }, { "epoch": 0.89651337186768, "grad_norm": 0.9927946329116821, "learning_rate": 1.641522957836672e-05, "loss": 3.7296, "step": 79800 }, { "epoch": 0.8970750970380231, "grad_norm": 1.0839712619781494, "learning_rate": 1.641298266506387e-05, "loss": 3.7817, "step": 79850 }, { "epoch": 0.8976368222083664, "grad_norm": 1.1929811239242554, "learning_rate": 1.641073575176102e-05, "loss": 3.7673, "step": 79900 }, { "epoch": 0.8981985473787095, "grad_norm": 1.1976163387298584, "learning_rate": 1.640848883845817e-05, "loss": 3.737, "step": 79950 }, { "epoch": 0.8987602725490527, "grad_norm": 1.006429672241211, "learning_rate": 1.640624192515532e-05, "loss": 3.7814, "step": 80000 }, { "epoch": 0.8993219977193958, "grad_norm": 1.101015567779541, "learning_rate": 1.6403995011852468e-05, "loss": 3.8, "step": 80050 }, { "epoch": 0.899883722889739, "grad_norm": 1.1051702499389648, "learning_rate": 1.6401748098549617e-05, "loss": 3.7417, "step": 80100 }, { "epoch": 0.9004454480600821, "grad_norm": 1.4870634078979492, "learning_rate": 1.639950118524677e-05, "loss": 3.6325, "step": 80150 }, { "epoch": 0.9010071732304252, "grad_norm": 1.0054829120635986, "learning_rate": 1.639725427194392e-05, "loss": 3.6966, "step": 80200 }, { "epoch": 0.9015688984007685, "grad_norm": 1.2589412927627563, "learning_rate": 1.639500735864107e-05, "loss": 3.7446, "step": 80250 }, { "epoch": 0.9021306235711116, "grad_norm": 1.1754649877548218, "learning_rate": 1.6392760445338218e-05, "loss": 3.6498, "step": 80300 }, { "epoch": 0.9026923487414548, "grad_norm": 0.9390884637832642, "learning_rate": 1.6390513532035367e-05, "loss": 3.7443, "step": 80350 }, { "epoch": 0.9032540739117979, "grad_norm": 0.9853662848472595, "learning_rate": 1.6388266618732517e-05, "loss": 3.7775, "step": 80400 }, { "epoch": 0.903815799082141, "grad_norm": 1.0741585493087769, "learning_rate": 1.6386019705429666e-05, "loss": 3.7435, "step": 80450 }, { "epoch": 0.9043775242524842, "grad_norm": 1.0210248231887817, "learning_rate": 1.638377279212682e-05, "loss": 3.696, "step": 80500 }, { "epoch": 0.9049392494228274, "grad_norm": 1.2705175876617432, "learning_rate": 1.6381525878823968e-05, "loss": 3.7158, "step": 80550 }, { "epoch": 0.9055009745931706, "grad_norm": 1.2171190977096558, "learning_rate": 1.6379278965521117e-05, "loss": 3.7859, "step": 80600 }, { "epoch": 0.9060626997635137, "grad_norm": 0.9464622735977173, "learning_rate": 1.6377032052218267e-05, "loss": 3.7325, "step": 80650 }, { "epoch": 0.9066244249338569, "grad_norm": 1.2099945545196533, "learning_rate": 1.6374785138915416e-05, "loss": 3.6913, "step": 80700 }, { "epoch": 0.9071861501042, "grad_norm": 1.1484709978103638, "learning_rate": 1.6372538225612565e-05, "loss": 3.7389, "step": 80750 }, { "epoch": 0.9077478752745431, "grad_norm": 1.155481219291687, "learning_rate": 1.6370291312309715e-05, "loss": 3.7722, "step": 80800 }, { "epoch": 0.9083096004448863, "grad_norm": 1.1051075458526611, "learning_rate": 1.6368044399006867e-05, "loss": 3.7531, "step": 80850 }, { "epoch": 0.9088713256152295, "grad_norm": 1.0399267673492432, "learning_rate": 1.6365797485704017e-05, "loss": 3.6485, "step": 80900 }, { "epoch": 0.9094330507855727, "grad_norm": 1.14091157913208, "learning_rate": 1.6363550572401166e-05, "loss": 3.6963, "step": 80950 }, { "epoch": 0.9099947759559158, "grad_norm": 2.3374216556549072, "learning_rate": 1.6361303659098315e-05, "loss": 3.7764, "step": 81000 }, { "epoch": 0.910556501126259, "grad_norm": 1.1379660367965698, "learning_rate": 1.635910168406152e-05, "loss": 3.6665, "step": 81050 }, { "epoch": 0.9111182262966021, "grad_norm": 1.128388524055481, "learning_rate": 1.635685477075867e-05, "loss": 3.8047, "step": 81100 }, { "epoch": 0.9116799514669452, "grad_norm": 1.1033473014831543, "learning_rate": 1.6354607857455823e-05, "loss": 3.7733, "step": 81150 }, { "epoch": 0.9122416766372884, "grad_norm": 1.1441519260406494, "learning_rate": 1.6352360944152973e-05, "loss": 3.7884, "step": 81200 }, { "epoch": 0.9128034018076316, "grad_norm": 1.1598007678985596, "learning_rate": 1.6350114030850122e-05, "loss": 3.7307, "step": 81250 }, { "epoch": 0.9133651269779748, "grad_norm": 1.1273629665374756, "learning_rate": 1.634786711754727e-05, "loss": 3.7872, "step": 81300 }, { "epoch": 0.9139268521483179, "grad_norm": 0.9979346394538879, "learning_rate": 1.634562020424442e-05, "loss": 3.7104, "step": 81350 }, { "epoch": 0.9144885773186611, "grad_norm": 0.9176252484321594, "learning_rate": 1.634337329094157e-05, "loss": 3.693, "step": 81400 }, { "epoch": 0.9150503024890042, "grad_norm": 1.0392706394195557, "learning_rate": 1.634112637763872e-05, "loss": 3.7285, "step": 81450 }, { "epoch": 0.9156120276593473, "grad_norm": 1.0022395849227905, "learning_rate": 1.6338879464335872e-05, "loss": 3.6837, "step": 81500 }, { "epoch": 0.9161737528296905, "grad_norm": 1.1621699333190918, "learning_rate": 1.633663255103302e-05, "loss": 3.6972, "step": 81550 }, { "epoch": 0.9167354780000337, "grad_norm": 1.1367850303649902, "learning_rate": 1.633438563773017e-05, "loss": 3.6952, "step": 81600 }, { "epoch": 0.9172972031703769, "grad_norm": 1.3023808002471924, "learning_rate": 1.633213872442732e-05, "loss": 3.7635, "step": 81650 }, { "epoch": 0.91785892834072, "grad_norm": 1.0885933637619019, "learning_rate": 1.632989181112447e-05, "loss": 3.7296, "step": 81700 }, { "epoch": 0.9184206535110632, "grad_norm": 1.1319997310638428, "learning_rate": 1.632764489782162e-05, "loss": 3.7595, "step": 81750 }, { "epoch": 0.9189823786814063, "grad_norm": 1.052115797996521, "learning_rate": 1.6325397984518768e-05, "loss": 3.7607, "step": 81800 }, { "epoch": 0.9195441038517495, "grad_norm": 1.0686609745025635, "learning_rate": 1.632315107121592e-05, "loss": 3.7502, "step": 81850 }, { "epoch": 0.9201058290220927, "grad_norm": 1.1252344846725464, "learning_rate": 1.632090415791307e-05, "loss": 3.6367, "step": 81900 }, { "epoch": 0.9206675541924358, "grad_norm": 1.3658348321914673, "learning_rate": 1.631865724461022e-05, "loss": 3.7372, "step": 81950 }, { "epoch": 0.921229279362779, "grad_norm": 1.0420020818710327, "learning_rate": 1.631641033130737e-05, "loss": 3.7205, "step": 82000 }, { "epoch": 0.9217910045331221, "grad_norm": 1.2236835956573486, "learning_rate": 1.6314163418004518e-05, "loss": 3.7831, "step": 82050 }, { "epoch": 0.9223527297034653, "grad_norm": 1.1498215198516846, "learning_rate": 1.6311916504701667e-05, "loss": 3.7857, "step": 82100 }, { "epoch": 0.9229144548738084, "grad_norm": 1.0487319231033325, "learning_rate": 1.6309669591398816e-05, "loss": 3.7117, "step": 82150 }, { "epoch": 0.9234761800441516, "grad_norm": 0.9913235902786255, "learning_rate": 1.630742267809597e-05, "loss": 3.7913, "step": 82200 }, { "epoch": 0.9240379052144948, "grad_norm": 1.0620747804641724, "learning_rate": 1.630517576479312e-05, "loss": 3.7876, "step": 82250 }, { "epoch": 0.924599630384838, "grad_norm": 1.0348880290985107, "learning_rate": 1.6302928851490268e-05, "loss": 3.8007, "step": 82300 }, { "epoch": 0.9251613555551811, "grad_norm": 1.0195879936218262, "learning_rate": 1.6300681938187417e-05, "loss": 3.7364, "step": 82350 }, { "epoch": 0.9257230807255242, "grad_norm": 1.4597419500350952, "learning_rate": 1.6298435024884566e-05, "loss": 3.7365, "step": 82400 }, { "epoch": 0.9262848058958674, "grad_norm": 1.0478322505950928, "learning_rate": 1.6296188111581716e-05, "loss": 3.6908, "step": 82450 }, { "epoch": 0.9268465310662105, "grad_norm": 1.2558234930038452, "learning_rate": 1.6293941198278865e-05, "loss": 3.7282, "step": 82500 }, { "epoch": 0.9274082562365537, "grad_norm": 1.243863821029663, "learning_rate": 1.6291694284976018e-05, "loss": 3.7268, "step": 82550 }, { "epoch": 0.9279699814068969, "grad_norm": 1.3925235271453857, "learning_rate": 1.6289447371673167e-05, "loss": 3.7183, "step": 82600 }, { "epoch": 0.92853170657724, "grad_norm": 1.1192623376846313, "learning_rate": 1.6287200458370316e-05, "loss": 3.6893, "step": 82650 }, { "epoch": 0.9290934317475832, "grad_norm": 1.2902791500091553, "learning_rate": 1.6284953545067466e-05, "loss": 3.764, "step": 82700 }, { "epoch": 0.9296551569179263, "grad_norm": 1.401732325553894, "learning_rate": 1.6282706631764615e-05, "loss": 3.7047, "step": 82750 }, { "epoch": 0.9302168820882695, "grad_norm": 1.1082121133804321, "learning_rate": 1.6280459718461764e-05, "loss": 3.7185, "step": 82800 }, { "epoch": 0.9307786072586126, "grad_norm": 1.0722180604934692, "learning_rate": 1.6278212805158914e-05, "loss": 3.7609, "step": 82850 }, { "epoch": 0.9313403324289559, "grad_norm": 1.0038912296295166, "learning_rate": 1.6275965891856066e-05, "loss": 3.7067, "step": 82900 }, { "epoch": 0.931902057599299, "grad_norm": 1.0864737033843994, "learning_rate": 1.6273718978553216e-05, "loss": 3.7562, "step": 82950 }, { "epoch": 0.9324637827696421, "grad_norm": 1.1519027948379517, "learning_rate": 1.6271472065250365e-05, "loss": 3.7136, "step": 83000 }, { "epoch": 0.9330255079399853, "grad_norm": 1.0338419675827026, "learning_rate": 1.6269225151947514e-05, "loss": 3.7762, "step": 83050 }, { "epoch": 0.9335872331103284, "grad_norm": 1.1340597867965698, "learning_rate": 1.6266978238644664e-05, "loss": 3.7631, "step": 83100 }, { "epoch": 0.9341489582806716, "grad_norm": 1.0290595293045044, "learning_rate": 1.6264731325341813e-05, "loss": 3.6794, "step": 83150 }, { "epoch": 0.9347106834510147, "grad_norm": 1.1557437181472778, "learning_rate": 1.6262484412038962e-05, "loss": 3.7217, "step": 83200 }, { "epoch": 0.935272408621358, "grad_norm": 1.2205008268356323, "learning_rate": 1.6260237498736115e-05, "loss": 3.7248, "step": 83250 }, { "epoch": 0.9358341337917011, "grad_norm": 1.0577982664108276, "learning_rate": 1.6257990585433264e-05, "loss": 3.7641, "step": 83300 }, { "epoch": 0.9363958589620442, "grad_norm": 1.2078427076339722, "learning_rate": 1.6255743672130414e-05, "loss": 3.7042, "step": 83350 }, { "epoch": 0.9369575841323874, "grad_norm": 1.1405781507492065, "learning_rate": 1.6253496758827563e-05, "loss": 3.7604, "step": 83400 }, { "epoch": 0.9375193093027305, "grad_norm": 1.126896619796753, "learning_rate": 1.6251249845524712e-05, "loss": 3.7297, "step": 83450 }, { "epoch": 0.9380810344730737, "grad_norm": 2.030507802963257, "learning_rate": 1.624900293222186e-05, "loss": 3.7266, "step": 83500 }, { "epoch": 0.9386427596434168, "grad_norm": 0.8830623626708984, "learning_rate": 1.624675601891901e-05, "loss": 3.6499, "step": 83550 }, { "epoch": 0.9392044848137601, "grad_norm": 1.0999085903167725, "learning_rate": 1.6244509105616163e-05, "loss": 3.8124, "step": 83600 }, { "epoch": 0.9397662099841032, "grad_norm": 1.0144646167755127, "learning_rate": 1.6242262192313313e-05, "loss": 3.693, "step": 83650 }, { "epoch": 0.9403279351544463, "grad_norm": 1.2930108308792114, "learning_rate": 1.6240015279010462e-05, "loss": 3.7773, "step": 83700 }, { "epoch": 0.9408896603247895, "grad_norm": 1.0382893085479736, "learning_rate": 1.623776836570761e-05, "loss": 3.7728, "step": 83750 }, { "epoch": 0.9414513854951326, "grad_norm": 1.1282553672790527, "learning_rate": 1.623552145240476e-05, "loss": 3.745, "step": 83800 }, { "epoch": 0.9420131106654758, "grad_norm": 1.1230621337890625, "learning_rate": 1.623327453910191e-05, "loss": 3.6773, "step": 83850 }, { "epoch": 0.9425748358358189, "grad_norm": 1.0702792406082153, "learning_rate": 1.623102762579906e-05, "loss": 3.6582, "step": 83900 }, { "epoch": 0.9431365610061622, "grad_norm": 1.0770357847213745, "learning_rate": 1.6228780712496212e-05, "loss": 3.7332, "step": 83950 }, { "epoch": 0.9436982861765053, "grad_norm": 1.0595725774765015, "learning_rate": 1.622653379919336e-05, "loss": 3.664, "step": 84000 }, { "epoch": 0.9442600113468484, "grad_norm": 1.3068342208862305, "learning_rate": 1.622428688589051e-05, "loss": 3.7584, "step": 84050 }, { "epoch": 0.9448217365171916, "grad_norm": 1.1843857765197754, "learning_rate": 1.622203997258766e-05, "loss": 3.7288, "step": 84100 }, { "epoch": 0.9453834616875347, "grad_norm": 1.2081139087677002, "learning_rate": 1.621979305928481e-05, "loss": 3.7039, "step": 84150 }, { "epoch": 0.9459451868578779, "grad_norm": 1.5447063446044922, "learning_rate": 1.621754614598196e-05, "loss": 3.7121, "step": 84200 }, { "epoch": 0.9465069120282211, "grad_norm": 1.1818842887878418, "learning_rate": 1.6215299232679108e-05, "loss": 3.731, "step": 84250 }, { "epoch": 0.9470686371985643, "grad_norm": 1.17287278175354, "learning_rate": 1.621305231937626e-05, "loss": 3.6707, "step": 84300 }, { "epoch": 0.9476303623689074, "grad_norm": 1.114976167678833, "learning_rate": 1.621080540607341e-05, "loss": 3.724, "step": 84350 }, { "epoch": 0.9481920875392505, "grad_norm": 1.4023295640945435, "learning_rate": 1.620855849277056e-05, "loss": 3.837, "step": 84400 }, { "epoch": 0.9487538127095937, "grad_norm": 1.0715686082839966, "learning_rate": 1.620631157946771e-05, "loss": 3.7689, "step": 84450 }, { "epoch": 0.9493155378799368, "grad_norm": 1.0490070581436157, "learning_rate": 1.6204064666164858e-05, "loss": 3.7099, "step": 84500 }, { "epoch": 0.94987726305028, "grad_norm": 1.0359930992126465, "learning_rate": 1.6201817752862007e-05, "loss": 3.6603, "step": 84550 }, { "epoch": 0.9504389882206232, "grad_norm": 1.259287714958191, "learning_rate": 1.6199570839559157e-05, "loss": 3.8236, "step": 84600 }, { "epoch": 0.9510007133909664, "grad_norm": 1.167106032371521, "learning_rate": 1.6197323926256306e-05, "loss": 3.7038, "step": 84650 }, { "epoch": 0.9515624385613095, "grad_norm": 1.2514828443527222, "learning_rate": 1.619507701295346e-05, "loss": 3.8219, "step": 84700 }, { "epoch": 0.9521241637316527, "grad_norm": 1.179477572441101, "learning_rate": 1.6192830099650608e-05, "loss": 3.6342, "step": 84750 }, { "epoch": 0.9526858889019958, "grad_norm": 1.272379994392395, "learning_rate": 1.6190583186347757e-05, "loss": 3.6968, "step": 84800 }, { "epoch": 0.9532476140723389, "grad_norm": 1.0392554998397827, "learning_rate": 1.6188336273044907e-05, "loss": 3.741, "step": 84850 }, { "epoch": 0.9538093392426821, "grad_norm": 1.019022822380066, "learning_rate": 1.6186089359742056e-05, "loss": 3.7318, "step": 84900 }, { "epoch": 0.9543710644130253, "grad_norm": 1.1417900323867798, "learning_rate": 1.6183842446439205e-05, "loss": 3.6845, "step": 84950 }, { "epoch": 0.9549327895833685, "grad_norm": 1.342798113822937, "learning_rate": 1.6181595533136355e-05, "loss": 3.7322, "step": 85000 }, { "epoch": 0.9554945147537116, "grad_norm": 1.011507511138916, "learning_rate": 1.6179348619833507e-05, "loss": 3.7205, "step": 85050 }, { "epoch": 0.9560562399240548, "grad_norm": 1.0429478883743286, "learning_rate": 1.6177101706530657e-05, "loss": 3.7183, "step": 85100 }, { "epoch": 0.9566179650943979, "grad_norm": 1.3029417991638184, "learning_rate": 1.6174854793227806e-05, "loss": 3.7278, "step": 85150 }, { "epoch": 0.957179690264741, "grad_norm": 1.1098041534423828, "learning_rate": 1.6172607879924955e-05, "loss": 3.7224, "step": 85200 }, { "epoch": 0.9577414154350842, "grad_norm": 1.0785588026046753, "learning_rate": 1.6170360966622104e-05, "loss": 3.7441, "step": 85250 }, { "epoch": 0.9583031406054274, "grad_norm": 1.1768405437469482, "learning_rate": 1.6168114053319254e-05, "loss": 3.7158, "step": 85300 }, { "epoch": 0.9588648657757706, "grad_norm": 0.9101014137268066, "learning_rate": 1.6165867140016403e-05, "loss": 3.684, "step": 85350 }, { "epoch": 0.9594265909461137, "grad_norm": 1.017488956451416, "learning_rate": 1.6163620226713556e-05, "loss": 3.7257, "step": 85400 }, { "epoch": 0.9599883161164569, "grad_norm": 1.0188761949539185, "learning_rate": 1.6161373313410705e-05, "loss": 3.7245, "step": 85450 }, { "epoch": 0.9605500412868, "grad_norm": 1.5345085859298706, "learning_rate": 1.6159126400107854e-05, "loss": 3.8303, "step": 85500 }, { "epoch": 0.9611117664571431, "grad_norm": 0.9981528520584106, "learning_rate": 1.6156879486805004e-05, "loss": 3.7071, "step": 85550 }, { "epoch": 0.9616734916274864, "grad_norm": 1.1540861129760742, "learning_rate": 1.6154632573502153e-05, "loss": 3.7012, "step": 85600 }, { "epoch": 0.9622352167978295, "grad_norm": 1.0878418684005737, "learning_rate": 1.6152385660199302e-05, "loss": 3.7355, "step": 85650 }, { "epoch": 0.9627969419681727, "grad_norm": 1.034525752067566, "learning_rate": 1.6150138746896452e-05, "loss": 3.7422, "step": 85700 }, { "epoch": 0.9633586671385158, "grad_norm": 1.1636149883270264, "learning_rate": 1.6147891833593604e-05, "loss": 3.7703, "step": 85750 }, { "epoch": 0.963920392308859, "grad_norm": 1.3120676279067993, "learning_rate": 1.6145689858556807e-05, "loss": 3.6778, "step": 85800 }, { "epoch": 0.9644821174792021, "grad_norm": 1.1310409307479858, "learning_rate": 1.6143442945253956e-05, "loss": 3.7145, "step": 85850 }, { "epoch": 0.9650438426495452, "grad_norm": 1.0255100727081299, "learning_rate": 1.614119603195111e-05, "loss": 3.8163, "step": 85900 }, { "epoch": 0.9656055678198885, "grad_norm": 0.9281632304191589, "learning_rate": 1.6138949118648258e-05, "loss": 3.7115, "step": 85950 }, { "epoch": 0.9661672929902316, "grad_norm": 1.1930404901504517, "learning_rate": 1.6136702205345408e-05, "loss": 3.8068, "step": 86000 }, { "epoch": 0.9667290181605748, "grad_norm": 1.111885666847229, "learning_rate": 1.6134455292042557e-05, "loss": 3.7737, "step": 86050 }, { "epoch": 0.9672907433309179, "grad_norm": 1.0258991718292236, "learning_rate": 1.6132208378739706e-05, "loss": 3.8056, "step": 86100 }, { "epoch": 0.967852468501261, "grad_norm": 1.169977068901062, "learning_rate": 1.6129961465436856e-05, "loss": 3.7443, "step": 86150 }, { "epoch": 0.9684141936716042, "grad_norm": 1.0789185762405396, "learning_rate": 1.6127714552134005e-05, "loss": 3.7529, "step": 86200 }, { "epoch": 0.9689759188419473, "grad_norm": 1.2293628454208374, "learning_rate": 1.6125467638831158e-05, "loss": 3.722, "step": 86250 }, { "epoch": 0.9695376440122906, "grad_norm": 1.1016701459884644, "learning_rate": 1.6123220725528307e-05, "loss": 3.7634, "step": 86300 }, { "epoch": 0.9700993691826337, "grad_norm": 1.0682452917099, "learning_rate": 1.6120973812225456e-05, "loss": 3.6436, "step": 86350 }, { "epoch": 0.9706610943529769, "grad_norm": 1.1063158512115479, "learning_rate": 1.6118726898922606e-05, "loss": 3.6602, "step": 86400 }, { "epoch": 0.97122281952332, "grad_norm": 1.049170732498169, "learning_rate": 1.6116479985619755e-05, "loss": 3.7842, "step": 86450 }, { "epoch": 0.9717845446936632, "grad_norm": 1.108414888381958, "learning_rate": 1.611427801058296e-05, "loss": 3.7661, "step": 86500 }, { "epoch": 0.9723462698640063, "grad_norm": 1.3780474662780762, "learning_rate": 1.6112031097280113e-05, "loss": 3.7555, "step": 86550 }, { "epoch": 0.9729079950343495, "grad_norm": 1.1071162223815918, "learning_rate": 1.6109784183977263e-05, "loss": 3.7328, "step": 86600 }, { "epoch": 0.9734697202046927, "grad_norm": 1.064674973487854, "learning_rate": 1.6107537270674412e-05, "loss": 3.6932, "step": 86650 }, { "epoch": 0.9740314453750358, "grad_norm": 1.2865511178970337, "learning_rate": 1.610529035737156e-05, "loss": 3.6596, "step": 86700 }, { "epoch": 0.974593170545379, "grad_norm": 1.1827284097671509, "learning_rate": 1.610304344406871e-05, "loss": 3.7418, "step": 86750 }, { "epoch": 0.9751548957157221, "grad_norm": 1.0069527626037598, "learning_rate": 1.610079653076586e-05, "loss": 3.74, "step": 86800 }, { "epoch": 0.9757166208860653, "grad_norm": 0.9784291386604309, "learning_rate": 1.609854961746301e-05, "loss": 3.7252, "step": 86850 }, { "epoch": 0.9762783460564084, "grad_norm": 1.0610507726669312, "learning_rate": 1.6096302704160162e-05, "loss": 3.7043, "step": 86900 }, { "epoch": 0.9768400712267516, "grad_norm": 1.321476936340332, "learning_rate": 1.609405579085731e-05, "loss": 3.7311, "step": 86950 }, { "epoch": 0.9774017963970948, "grad_norm": 1.2549364566802979, "learning_rate": 1.609180887755446e-05, "loss": 3.7484, "step": 87000 }, { "epoch": 0.9779635215674379, "grad_norm": 1.0170053243637085, "learning_rate": 1.608956196425161e-05, "loss": 3.6022, "step": 87050 }, { "epoch": 0.9785252467377811, "grad_norm": 1.1055805683135986, "learning_rate": 1.608731505094876e-05, "loss": 3.7055, "step": 87100 }, { "epoch": 0.9790869719081242, "grad_norm": 1.103136420249939, "learning_rate": 1.608506813764591e-05, "loss": 3.7579, "step": 87150 }, { "epoch": 0.9796486970784674, "grad_norm": 1.2995116710662842, "learning_rate": 1.6082821224343058e-05, "loss": 3.7244, "step": 87200 }, { "epoch": 0.9802104222488105, "grad_norm": 1.2343697547912598, "learning_rate": 1.608057431104021e-05, "loss": 3.7127, "step": 87250 }, { "epoch": 0.9807721474191538, "grad_norm": 1.2577475309371948, "learning_rate": 1.607832739773736e-05, "loss": 3.8002, "step": 87300 }, { "epoch": 0.9813338725894969, "grad_norm": 1.1043496131896973, "learning_rate": 1.607608048443451e-05, "loss": 3.7542, "step": 87350 }, { "epoch": 0.98189559775984, "grad_norm": 1.2996944189071655, "learning_rate": 1.607383357113166e-05, "loss": 3.7628, "step": 87400 }, { "epoch": 0.9824573229301832, "grad_norm": 1.0178141593933105, "learning_rate": 1.6071586657828808e-05, "loss": 3.7823, "step": 87450 }, { "epoch": 0.9830190481005263, "grad_norm": 1.2404489517211914, "learning_rate": 1.6069339744525957e-05, "loss": 3.6674, "step": 87500 }, { "epoch": 0.9835807732708695, "grad_norm": 1.1711283922195435, "learning_rate": 1.6067092831223107e-05, "loss": 3.7598, "step": 87550 }, { "epoch": 0.9841424984412126, "grad_norm": 1.113303542137146, "learning_rate": 1.606484591792026e-05, "loss": 3.7011, "step": 87600 }, { "epoch": 0.9847042236115559, "grad_norm": 1.2303305864334106, "learning_rate": 1.606259900461741e-05, "loss": 3.7933, "step": 87650 }, { "epoch": 0.985265948781899, "grad_norm": 1.190563440322876, "learning_rate": 1.6060352091314558e-05, "loss": 3.7997, "step": 87700 }, { "epoch": 0.9858276739522421, "grad_norm": 1.2915029525756836, "learning_rate": 1.6058105178011707e-05, "loss": 3.6851, "step": 87750 }, { "epoch": 0.9863893991225853, "grad_norm": 0.9873474836349487, "learning_rate": 1.6055858264708857e-05, "loss": 3.7403, "step": 87800 }, { "epoch": 0.9869511242929284, "grad_norm": 1.1766434907913208, "learning_rate": 1.6053611351406006e-05, "loss": 3.7529, "step": 87850 }, { "epoch": 0.9875128494632716, "grad_norm": 1.0172233581542969, "learning_rate": 1.6051364438103155e-05, "loss": 3.766, "step": 87900 }, { "epoch": 0.9880745746336148, "grad_norm": 1.124140739440918, "learning_rate": 1.6049117524800305e-05, "loss": 3.7028, "step": 87950 }, { "epoch": 0.988636299803958, "grad_norm": 1.1819666624069214, "learning_rate": 1.6046870611497457e-05, "loss": 3.729, "step": 88000 }, { "epoch": 0.9891980249743011, "grad_norm": 1.2545924186706543, "learning_rate": 1.6044623698194607e-05, "loss": 3.6809, "step": 88050 }, { "epoch": 0.9897597501446442, "grad_norm": 1.0095438957214355, "learning_rate": 1.6042376784891756e-05, "loss": 3.7897, "step": 88100 }, { "epoch": 0.9903214753149874, "grad_norm": 1.0697457790374756, "learning_rate": 1.6040129871588905e-05, "loss": 3.7674, "step": 88150 }, { "epoch": 0.9908832004853305, "grad_norm": 1.0836268663406372, "learning_rate": 1.6037882958286054e-05, "loss": 3.6559, "step": 88200 }, { "epoch": 0.9914449256556737, "grad_norm": 1.0364786386489868, "learning_rate": 1.6035636044983204e-05, "loss": 3.6434, "step": 88250 }, { "epoch": 0.9920066508260169, "grad_norm": 1.1839394569396973, "learning_rate": 1.6033389131680353e-05, "loss": 3.6774, "step": 88300 }, { "epoch": 0.99256837599636, "grad_norm": 1.2167590856552124, "learning_rate": 1.6031142218377506e-05, "loss": 3.7092, "step": 88350 }, { "epoch": 0.9931301011667032, "grad_norm": 1.0059152841567993, "learning_rate": 1.6028895305074655e-05, "loss": 3.6941, "step": 88400 }, { "epoch": 0.9936918263370463, "grad_norm": 1.1846799850463867, "learning_rate": 1.6026648391771804e-05, "loss": 3.7267, "step": 88450 }, { "epoch": 0.9942535515073895, "grad_norm": 1.0491869449615479, "learning_rate": 1.6024401478468954e-05, "loss": 3.7686, "step": 88500 }, { "epoch": 0.9948152766777326, "grad_norm": 1.0813937187194824, "learning_rate": 1.6022154565166103e-05, "loss": 3.696, "step": 88550 }, { "epoch": 0.9953770018480758, "grad_norm": 1.1521028280258179, "learning_rate": 1.6019907651863252e-05, "loss": 3.6952, "step": 88600 }, { "epoch": 0.995938727018419, "grad_norm": 1.0822917222976685, "learning_rate": 1.6017705676826462e-05, "loss": 3.7029, "step": 88650 }, { "epoch": 0.9965004521887622, "grad_norm": 1.2565577030181885, "learning_rate": 1.601545876352361e-05, "loss": 3.7513, "step": 88700 }, { "epoch": 0.9970621773591053, "grad_norm": 1.0764832496643066, "learning_rate": 1.601321185022076e-05, "loss": 3.7125, "step": 88750 }, { "epoch": 0.9976239025294484, "grad_norm": 1.0988057851791382, "learning_rate": 1.601096493691791e-05, "loss": 3.7222, "step": 88800 }, { "epoch": 0.9981856276997916, "grad_norm": 1.1058125495910645, "learning_rate": 1.600871802361506e-05, "loss": 3.7659, "step": 88850 }, { "epoch": 0.9987473528701347, "grad_norm": 2.2041337490081787, "learning_rate": 1.6006471110312208e-05, "loss": 3.7178, "step": 88900 }, { "epoch": 0.9993090780404779, "grad_norm": 1.0233266353607178, "learning_rate": 1.6004224197009358e-05, "loss": 3.8086, "step": 88950 }, { "epoch": 0.9998708032108211, "grad_norm": 1.1076833009719849, "learning_rate": 1.600197728370651e-05, "loss": 3.709, "step": 89000 }, { "epoch": 0.9999943827482966, "eval_bleu": 5.7847, "eval_gen_len": 21.1148, "eval_loss": 3.5926201343536377, "eval_runtime": 23818.0921, "eval_samples_per_second": 14.949, "eval_steps_per_second": 0.934, "step": 89011 }, { "epoch": 1.0004325283811641, "grad_norm": 1.1763598918914795, "learning_rate": 1.599973037040366e-05, "loss": 3.6897, "step": 89050 }, { "epoch": 1.0009942535515073, "grad_norm": 1.2548326253890991, "learning_rate": 1.599748345710081e-05, "loss": 3.658, "step": 89100 }, { "epoch": 1.0015559787218506, "grad_norm": 1.1703482866287231, "learning_rate": 1.5995236543797958e-05, "loss": 3.6923, "step": 89150 }, { "epoch": 1.0021177038921938, "grad_norm": 0.9730241298675537, "learning_rate": 1.5992989630495108e-05, "loss": 3.6708, "step": 89200 }, { "epoch": 1.002679429062537, "grad_norm": 1.160300850868225, "learning_rate": 1.5990742717192257e-05, "loss": 3.7001, "step": 89250 }, { "epoch": 1.00324115423288, "grad_norm": 1.2727965116500854, "learning_rate": 1.5988495803889406e-05, "loss": 3.7369, "step": 89300 }, { "epoch": 1.0038028794032232, "grad_norm": 1.0181578397750854, "learning_rate": 1.598624889058656e-05, "loss": 3.7247, "step": 89350 }, { "epoch": 1.0043646045735664, "grad_norm": 1.1146141290664673, "learning_rate": 1.5984001977283708e-05, "loss": 3.7194, "step": 89400 }, { "epoch": 1.0049263297439095, "grad_norm": 1.2422945499420166, "learning_rate": 1.5981755063980858e-05, "loss": 3.7583, "step": 89450 }, { "epoch": 1.0054880549142526, "grad_norm": 1.0126484632492065, "learning_rate": 1.5979508150678007e-05, "loss": 3.6875, "step": 89500 }, { "epoch": 1.0060497800845958, "grad_norm": 1.0848461389541626, "learning_rate": 1.5977261237375156e-05, "loss": 3.7259, "step": 89550 }, { "epoch": 1.006611505254939, "grad_norm": 1.1333413124084473, "learning_rate": 1.5975014324072306e-05, "loss": 3.6812, "step": 89600 }, { "epoch": 1.007173230425282, "grad_norm": 1.2641425132751465, "learning_rate": 1.5972767410769455e-05, "loss": 3.6745, "step": 89650 }, { "epoch": 1.0077349555956252, "grad_norm": 1.11570143699646, "learning_rate": 1.5970520497466608e-05, "loss": 3.7447, "step": 89700 }, { "epoch": 1.0082966807659683, "grad_norm": 1.2924162149429321, "learning_rate": 1.5968273584163757e-05, "loss": 3.795, "step": 89750 }, { "epoch": 1.0088584059363117, "grad_norm": 1.1690458059310913, "learning_rate": 1.5966026670860906e-05, "loss": 3.6647, "step": 89800 }, { "epoch": 1.0094201311066548, "grad_norm": 1.1844213008880615, "learning_rate": 1.5963779757558055e-05, "loss": 3.6823, "step": 89850 }, { "epoch": 1.009981856276998, "grad_norm": 1.0748149156570435, "learning_rate": 1.5961532844255205e-05, "loss": 3.7264, "step": 89900 }, { "epoch": 1.0105435814473411, "grad_norm": 1.9574732780456543, "learning_rate": 1.5959285930952354e-05, "loss": 3.7581, "step": 89950 }, { "epoch": 1.0111053066176843, "grad_norm": 1.1267497539520264, "learning_rate": 1.5957039017649503e-05, "loss": 3.7218, "step": 90000 }, { "epoch": 1.0116670317880274, "grad_norm": 1.1397321224212646, "learning_rate": 1.5954792104346656e-05, "loss": 3.7047, "step": 90050 }, { "epoch": 1.0122287569583706, "grad_norm": 1.2505513429641724, "learning_rate": 1.5952545191043805e-05, "loss": 3.6746, "step": 90100 }, { "epoch": 1.0127904821287137, "grad_norm": 1.0324023962020874, "learning_rate": 1.5950298277740955e-05, "loss": 3.8013, "step": 90150 }, { "epoch": 1.0133522072990568, "grad_norm": 1.238370418548584, "learning_rate": 1.5948051364438104e-05, "loss": 3.764, "step": 90200 }, { "epoch": 1.0139139324694, "grad_norm": 1.2217398881912231, "learning_rate": 1.5945804451135253e-05, "loss": 3.7278, "step": 90250 }, { "epoch": 1.0144756576397431, "grad_norm": 1.2465476989746094, "learning_rate": 1.5943557537832403e-05, "loss": 3.6837, "step": 90300 }, { "epoch": 1.0150373828100863, "grad_norm": 1.1466426849365234, "learning_rate": 1.5941310624529552e-05, "loss": 3.8206, "step": 90350 }, { "epoch": 1.0155991079804294, "grad_norm": 1.092600703239441, "learning_rate": 1.5939063711226705e-05, "loss": 3.6569, "step": 90400 }, { "epoch": 1.0161608331507725, "grad_norm": 1.0314606428146362, "learning_rate": 1.5936816797923854e-05, "loss": 3.7336, "step": 90450 }, { "epoch": 1.016722558321116, "grad_norm": 1.1504980325698853, "learning_rate": 1.5934569884621003e-05, "loss": 3.6885, "step": 90500 }, { "epoch": 1.017284283491459, "grad_norm": 1.2484686374664307, "learning_rate": 1.5932322971318153e-05, "loss": 3.7622, "step": 90550 }, { "epoch": 1.0178460086618022, "grad_norm": 1.1624548435211182, "learning_rate": 1.5930076058015302e-05, "loss": 3.7382, "step": 90600 }, { "epoch": 1.0184077338321453, "grad_norm": 1.129113793373108, "learning_rate": 1.592782914471245e-05, "loss": 3.7568, "step": 90650 }, { "epoch": 1.0189694590024885, "grad_norm": 0.9480738043785095, "learning_rate": 1.59255822314096e-05, "loss": 3.7107, "step": 90700 }, { "epoch": 1.0195311841728316, "grad_norm": 0.9821470379829407, "learning_rate": 1.5923335318106753e-05, "loss": 3.7356, "step": 90750 }, { "epoch": 1.0200929093431748, "grad_norm": 1.2413480281829834, "learning_rate": 1.5921088404803903e-05, "loss": 3.7106, "step": 90800 }, { "epoch": 1.020654634513518, "grad_norm": 0.9626780152320862, "learning_rate": 1.5918841491501052e-05, "loss": 3.728, "step": 90850 }, { "epoch": 1.021216359683861, "grad_norm": 1.038010835647583, "learning_rate": 1.59165945781982e-05, "loss": 3.7322, "step": 90900 }, { "epoch": 1.0217780848542042, "grad_norm": 1.1623151302337646, "learning_rate": 1.591434766489535e-05, "loss": 3.7519, "step": 90950 }, { "epoch": 1.0223398100245473, "grad_norm": 1.1785417795181274, "learning_rate": 1.59121007515925e-05, "loss": 3.7059, "step": 91000 }, { "epoch": 1.0229015351948905, "grad_norm": 1.0354048013687134, "learning_rate": 1.590985383828965e-05, "loss": 3.6771, "step": 91050 }, { "epoch": 1.0234632603652336, "grad_norm": 1.1689685583114624, "learning_rate": 1.5907606924986802e-05, "loss": 3.7238, "step": 91100 }, { "epoch": 1.024024985535577, "grad_norm": 1.102995753288269, "learning_rate": 1.590536001168395e-05, "loss": 3.721, "step": 91150 }, { "epoch": 1.0245867107059201, "grad_norm": 1.0524019002914429, "learning_rate": 1.59031130983811e-05, "loss": 3.6991, "step": 91200 }, { "epoch": 1.0251484358762633, "grad_norm": 1.2444039583206177, "learning_rate": 1.590086618507825e-05, "loss": 3.7336, "step": 91250 }, { "epoch": 1.0257101610466064, "grad_norm": 1.0841864347457886, "learning_rate": 1.58986192717754e-05, "loss": 3.8199, "step": 91300 }, { "epoch": 1.0262718862169495, "grad_norm": 1.1749396324157715, "learning_rate": 1.589637235847255e-05, "loss": 3.6557, "step": 91350 }, { "epoch": 1.0268336113872927, "grad_norm": 1.1323601007461548, "learning_rate": 1.5894125445169698e-05, "loss": 3.7425, "step": 91400 }, { "epoch": 1.0273953365576358, "grad_norm": 1.066879153251648, "learning_rate": 1.589187853186685e-05, "loss": 3.6942, "step": 91450 }, { "epoch": 1.027957061727979, "grad_norm": 0.929542601108551, "learning_rate": 1.5889631618564e-05, "loss": 3.6912, "step": 91500 }, { "epoch": 1.028518786898322, "grad_norm": 1.0617517232894897, "learning_rate": 1.588738470526115e-05, "loss": 3.6954, "step": 91550 }, { "epoch": 1.0290805120686652, "grad_norm": 0.9447341561317444, "learning_rate": 1.58851377919583e-05, "loss": 3.685, "step": 91600 }, { "epoch": 1.0296422372390084, "grad_norm": 1.0214871168136597, "learning_rate": 1.5882890878655448e-05, "loss": 3.581, "step": 91650 }, { "epoch": 1.0302039624093515, "grad_norm": 1.0436146259307861, "learning_rate": 1.5880643965352597e-05, "loss": 3.7354, "step": 91700 }, { "epoch": 1.0307656875796947, "grad_norm": 1.1642152070999146, "learning_rate": 1.5878397052049746e-05, "loss": 3.7397, "step": 91750 }, { "epoch": 1.0313274127500378, "grad_norm": 1.1631594896316528, "learning_rate": 1.58761501387469e-05, "loss": 3.7611, "step": 91800 }, { "epoch": 1.0318891379203812, "grad_norm": 1.2050690650939941, "learning_rate": 1.587390322544405e-05, "loss": 3.6778, "step": 91850 }, { "epoch": 1.0324508630907243, "grad_norm": 1.268229365348816, "learning_rate": 1.5871656312141198e-05, "loss": 3.6934, "step": 91900 }, { "epoch": 1.0330125882610675, "grad_norm": 1.3370764255523682, "learning_rate": 1.5869409398838347e-05, "loss": 3.7889, "step": 91950 }, { "epoch": 1.0335743134314106, "grad_norm": 1.07756769657135, "learning_rate": 1.5867162485535496e-05, "loss": 3.6994, "step": 92000 }, { "epoch": 1.0341360386017537, "grad_norm": 1.1191749572753906, "learning_rate": 1.5864915572232646e-05, "loss": 3.7082, "step": 92050 }, { "epoch": 1.0346977637720969, "grad_norm": 1.1558022499084473, "learning_rate": 1.5862668658929795e-05, "loss": 3.729, "step": 92100 }, { "epoch": 1.03525948894244, "grad_norm": 1.1125577688217163, "learning_rate": 1.5860421745626944e-05, "loss": 3.7722, "step": 92150 }, { "epoch": 1.0358212141127832, "grad_norm": 1.2622685432434082, "learning_rate": 1.5858174832324097e-05, "loss": 3.7259, "step": 92200 }, { "epoch": 1.0363829392831263, "grad_norm": 1.3767738342285156, "learning_rate": 1.5855927919021246e-05, "loss": 3.7607, "step": 92250 }, { "epoch": 1.0369446644534694, "grad_norm": 0.9928566813468933, "learning_rate": 1.5853681005718396e-05, "loss": 3.7339, "step": 92300 }, { "epoch": 1.0375063896238126, "grad_norm": 1.0747640132904053, "learning_rate": 1.5851434092415545e-05, "loss": 3.8035, "step": 92350 }, { "epoch": 1.0380681147941557, "grad_norm": 1.2698166370391846, "learning_rate": 1.5849187179112694e-05, "loss": 3.7692, "step": 92400 }, { "epoch": 1.0386298399644989, "grad_norm": 1.2842738628387451, "learning_rate": 1.5846940265809844e-05, "loss": 3.8173, "step": 92450 }, { "epoch": 1.0391915651348422, "grad_norm": 1.1700290441513062, "learning_rate": 1.5844693352506993e-05, "loss": 3.7536, "step": 92500 }, { "epoch": 1.0397532903051854, "grad_norm": 1.02366304397583, "learning_rate": 1.5842446439204146e-05, "loss": 3.7264, "step": 92550 }, { "epoch": 1.0403150154755285, "grad_norm": 1.1328957080841064, "learning_rate": 1.5840199525901295e-05, "loss": 3.6595, "step": 92600 }, { "epoch": 1.0408767406458717, "grad_norm": 1.1032906770706177, "learning_rate": 1.5837952612598444e-05, "loss": 3.7593, "step": 92650 }, { "epoch": 1.0414384658162148, "grad_norm": 1.2312544584274292, "learning_rate": 1.5835705699295594e-05, "loss": 3.7379, "step": 92700 }, { "epoch": 1.042000190986558, "grad_norm": 1.1248546838760376, "learning_rate": 1.5833458785992743e-05, "loss": 3.669, "step": 92750 }, { "epoch": 1.042561916156901, "grad_norm": 1.175838828086853, "learning_rate": 1.5831211872689892e-05, "loss": 3.7091, "step": 92800 }, { "epoch": 1.0431236413272442, "grad_norm": 1.394729733467102, "learning_rate": 1.582896495938704e-05, "loss": 3.7474, "step": 92850 }, { "epoch": 1.0436853664975874, "grad_norm": 1.2979602813720703, "learning_rate": 1.5826718046084194e-05, "loss": 3.6861, "step": 92900 }, { "epoch": 1.0442470916679305, "grad_norm": 1.091827392578125, "learning_rate": 1.5824471132781344e-05, "loss": 3.6909, "step": 92950 }, { "epoch": 1.0448088168382736, "grad_norm": 1.2266827821731567, "learning_rate": 1.5822224219478493e-05, "loss": 3.7223, "step": 93000 }, { "epoch": 1.0453705420086168, "grad_norm": 1.0776731967926025, "learning_rate": 1.5819977306175642e-05, "loss": 3.7334, "step": 93050 }, { "epoch": 1.04593226717896, "grad_norm": 1.119864583015442, "learning_rate": 1.581773039287279e-05, "loss": 3.7641, "step": 93100 }, { "epoch": 1.046493992349303, "grad_norm": 1.124421238899231, "learning_rate": 1.581548347956994e-05, "loss": 3.6658, "step": 93150 }, { "epoch": 1.0470557175196464, "grad_norm": 1.103226900100708, "learning_rate": 1.581323656626709e-05, "loss": 3.6809, "step": 93200 }, { "epoch": 1.0476174426899896, "grad_norm": 0.9896043539047241, "learning_rate": 1.5810989652964243e-05, "loss": 3.6122, "step": 93250 }, { "epoch": 1.0481791678603327, "grad_norm": 1.226040244102478, "learning_rate": 1.5808742739661392e-05, "loss": 3.6669, "step": 93300 }, { "epoch": 1.0487408930306759, "grad_norm": 1.0165162086486816, "learning_rate": 1.580649582635854e-05, "loss": 3.6696, "step": 93350 }, { "epoch": 1.049302618201019, "grad_norm": 1.2807351350784302, "learning_rate": 1.580424891305569e-05, "loss": 3.7671, "step": 93400 }, { "epoch": 1.0498643433713621, "grad_norm": 1.1978585720062256, "learning_rate": 1.580200199975284e-05, "loss": 3.77, "step": 93450 }, { "epoch": 1.0504260685417053, "grad_norm": 1.1764203310012817, "learning_rate": 1.579975508644999e-05, "loss": 3.8034, "step": 93500 }, { "epoch": 1.0509877937120484, "grad_norm": 1.1233216524124146, "learning_rate": 1.579750817314714e-05, "loss": 3.7798, "step": 93550 }, { "epoch": 1.0515495188823916, "grad_norm": 1.6650383472442627, "learning_rate": 1.579526125984429e-05, "loss": 3.7086, "step": 93600 }, { "epoch": 1.0521112440527347, "grad_norm": 1.1246583461761475, "learning_rate": 1.579301434654144e-05, "loss": 3.8124, "step": 93650 }, { "epoch": 1.0526729692230778, "grad_norm": 1.054193377494812, "learning_rate": 1.579076743323859e-05, "loss": 3.7576, "step": 93700 }, { "epoch": 1.053234694393421, "grad_norm": 1.1520317792892456, "learning_rate": 1.578852051993574e-05, "loss": 3.7045, "step": 93750 }, { "epoch": 1.0537964195637641, "grad_norm": 1.1220756769180298, "learning_rate": 1.578627360663289e-05, "loss": 3.7785, "step": 93800 }, { "epoch": 1.0543581447341075, "grad_norm": 1.1155728101730347, "learning_rate": 1.5784026693330038e-05, "loss": 3.7105, "step": 93850 }, { "epoch": 1.0549198699044506, "grad_norm": 1.0775452852249146, "learning_rate": 1.5781824718293247e-05, "loss": 3.7609, "step": 93900 }, { "epoch": 1.0554815950747938, "grad_norm": 1.1409608125686646, "learning_rate": 1.5779577804990397e-05, "loss": 3.7306, "step": 93950 }, { "epoch": 1.056043320245137, "grad_norm": 1.1192364692687988, "learning_rate": 1.5777330891687546e-05, "loss": 3.6886, "step": 94000 }, { "epoch": 1.05660504541548, "grad_norm": 1.3993706703186035, "learning_rate": 1.5775083978384695e-05, "loss": 3.6814, "step": 94050 }, { "epoch": 1.0571667705858232, "grad_norm": 1.189325213432312, "learning_rate": 1.5772837065081845e-05, "loss": 3.7275, "step": 94100 }, { "epoch": 1.0577284957561663, "grad_norm": 1.2303613424301147, "learning_rate": 1.5770590151778994e-05, "loss": 3.7163, "step": 94150 }, { "epoch": 1.0582902209265095, "grad_norm": 1.0467129945755005, "learning_rate": 1.5768343238476143e-05, "loss": 3.6759, "step": 94200 }, { "epoch": 1.0588519460968526, "grad_norm": 1.060932993888855, "learning_rate": 1.5766096325173296e-05, "loss": 3.6594, "step": 94250 }, { "epoch": 1.0594136712671958, "grad_norm": 1.2679494619369507, "learning_rate": 1.5763849411870445e-05, "loss": 3.7668, "step": 94300 }, { "epoch": 1.059975396437539, "grad_norm": 1.0520027875900269, "learning_rate": 1.5761602498567595e-05, "loss": 3.6711, "step": 94350 }, { "epoch": 1.060537121607882, "grad_norm": 1.1975017786026, "learning_rate": 1.5759355585264744e-05, "loss": 3.7632, "step": 94400 }, { "epoch": 1.0610988467782252, "grad_norm": 1.310758352279663, "learning_rate": 1.5757108671961893e-05, "loss": 3.736, "step": 94450 }, { "epoch": 1.0616605719485683, "grad_norm": 1.2200254201889038, "learning_rate": 1.5754861758659043e-05, "loss": 3.766, "step": 94500 }, { "epoch": 1.0622222971189117, "grad_norm": 1.3379442691802979, "learning_rate": 1.5752614845356192e-05, "loss": 3.7193, "step": 94550 }, { "epoch": 1.0627840222892548, "grad_norm": 1.1302127838134766, "learning_rate": 1.5750367932053345e-05, "loss": 3.7355, "step": 94600 }, { "epoch": 1.063345747459598, "grad_norm": 1.0796693563461304, "learning_rate": 1.5748121018750494e-05, "loss": 3.7235, "step": 94650 }, { "epoch": 1.0639074726299411, "grad_norm": 1.236109972000122, "learning_rate": 1.5745874105447643e-05, "loss": 3.7059, "step": 94700 }, { "epoch": 1.0644691978002843, "grad_norm": 1.2734140157699585, "learning_rate": 1.5743627192144793e-05, "loss": 3.7032, "step": 94750 }, { "epoch": 1.0650309229706274, "grad_norm": 1.3345608711242676, "learning_rate": 1.5741380278841942e-05, "loss": 3.7339, "step": 94800 }, { "epoch": 1.0655926481409705, "grad_norm": 1.1744345426559448, "learning_rate": 1.573913336553909e-05, "loss": 3.6792, "step": 94850 }, { "epoch": 1.0661543733113137, "grad_norm": 1.0625845193862915, "learning_rate": 1.573688645223624e-05, "loss": 3.718, "step": 94900 }, { "epoch": 1.0667160984816568, "grad_norm": 1.1035292148590088, "learning_rate": 1.5734639538933393e-05, "loss": 3.731, "step": 94950 }, { "epoch": 1.067277823652, "grad_norm": 1.2014189958572388, "learning_rate": 1.5732392625630543e-05, "loss": 3.686, "step": 95000 }, { "epoch": 1.067839548822343, "grad_norm": 1.1344666481018066, "learning_rate": 1.5730145712327692e-05, "loss": 3.7133, "step": 95050 }, { "epoch": 1.0684012739926863, "grad_norm": 1.1860787868499756, "learning_rate": 1.572789879902484e-05, "loss": 3.6945, "step": 95100 }, { "epoch": 1.0689629991630294, "grad_norm": 1.129529356956482, "learning_rate": 1.572565188572199e-05, "loss": 3.6903, "step": 95150 }, { "epoch": 1.0695247243333728, "grad_norm": 0.9301499128341675, "learning_rate": 1.572340497241914e-05, "loss": 3.6702, "step": 95200 }, { "epoch": 1.070086449503716, "grad_norm": 1.171042799949646, "learning_rate": 1.572115805911629e-05, "loss": 3.6832, "step": 95250 }, { "epoch": 1.070648174674059, "grad_norm": 1.0683860778808594, "learning_rate": 1.5718911145813442e-05, "loss": 3.7034, "step": 95300 }, { "epoch": 1.0712098998444022, "grad_norm": 1.2346833944320679, "learning_rate": 1.571666423251059e-05, "loss": 3.7396, "step": 95350 }, { "epoch": 1.0717716250147453, "grad_norm": 1.0209951400756836, "learning_rate": 1.571441731920774e-05, "loss": 3.7822, "step": 95400 }, { "epoch": 1.0723333501850885, "grad_norm": 1.2905466556549072, "learning_rate": 1.571217040590489e-05, "loss": 3.7594, "step": 95450 }, { "epoch": 1.0728950753554316, "grad_norm": 1.3073861598968506, "learning_rate": 1.570992349260204e-05, "loss": 3.7016, "step": 95500 }, { "epoch": 1.0734568005257747, "grad_norm": 1.2471473217010498, "learning_rate": 1.570767657929919e-05, "loss": 3.6912, "step": 95550 }, { "epoch": 1.0740185256961179, "grad_norm": 1.2150790691375732, "learning_rate": 1.5705429665996338e-05, "loss": 3.7118, "step": 95600 }, { "epoch": 1.074580250866461, "grad_norm": 1.1582289934158325, "learning_rate": 1.570318275269349e-05, "loss": 3.7255, "step": 95650 }, { "epoch": 1.0751419760368042, "grad_norm": 1.164430022239685, "learning_rate": 1.570093583939064e-05, "loss": 3.7233, "step": 95700 }, { "epoch": 1.0757037012071473, "grad_norm": 1.222058892250061, "learning_rate": 1.569868892608779e-05, "loss": 3.774, "step": 95750 }, { "epoch": 1.0762654263774905, "grad_norm": 1.12148118019104, "learning_rate": 1.569644201278494e-05, "loss": 3.7889, "step": 95800 }, { "epoch": 1.0768271515478336, "grad_norm": 0.9750103950500488, "learning_rate": 1.5694195099482088e-05, "loss": 3.728, "step": 95850 }, { "epoch": 1.077388876718177, "grad_norm": 1.0557671785354614, "learning_rate": 1.5691948186179237e-05, "loss": 3.7635, "step": 95900 }, { "epoch": 1.07795060188852, "grad_norm": 1.0505820512771606, "learning_rate": 1.5689701272876386e-05, "loss": 3.7686, "step": 95950 }, { "epoch": 1.0785123270588632, "grad_norm": 1.1918795108795166, "learning_rate": 1.568745435957354e-05, "loss": 3.6623, "step": 96000 }, { "epoch": 1.0790740522292064, "grad_norm": 1.1036181449890137, "learning_rate": 1.568520744627069e-05, "loss": 3.624, "step": 96050 }, { "epoch": 1.0796357773995495, "grad_norm": 1.1361887454986572, "learning_rate": 1.5683005471233894e-05, "loss": 3.7652, "step": 96100 }, { "epoch": 1.0801975025698927, "grad_norm": 1.04814612865448, "learning_rate": 1.5680758557931044e-05, "loss": 3.6505, "step": 96150 }, { "epoch": 1.0807592277402358, "grad_norm": 1.2747437953948975, "learning_rate": 1.5678511644628193e-05, "loss": 3.6925, "step": 96200 }, { "epoch": 1.081320952910579, "grad_norm": 0.9882274866104126, "learning_rate": 1.5676264731325342e-05, "loss": 3.7606, "step": 96250 }, { "epoch": 1.081882678080922, "grad_norm": 1.2443149089813232, "learning_rate": 1.5674017818022495e-05, "loss": 3.7596, "step": 96300 }, { "epoch": 1.0824444032512652, "grad_norm": 1.0911964178085327, "learning_rate": 1.5671770904719644e-05, "loss": 3.6851, "step": 96350 }, { "epoch": 1.0830061284216084, "grad_norm": 1.175838828086853, "learning_rate": 1.5669523991416794e-05, "loss": 3.7198, "step": 96400 }, { "epoch": 1.0835678535919515, "grad_norm": 1.2125855684280396, "learning_rate": 1.5667277078113943e-05, "loss": 3.7662, "step": 96450 }, { "epoch": 1.0841295787622949, "grad_norm": 1.248186707496643, "learning_rate": 1.5665030164811092e-05, "loss": 3.6982, "step": 96500 }, { "epoch": 1.084691303932638, "grad_norm": 1.143099308013916, "learning_rate": 1.566278325150824e-05, "loss": 3.8005, "step": 96550 }, { "epoch": 1.0852530291029812, "grad_norm": 1.080655574798584, "learning_rate": 1.566053633820539e-05, "loss": 3.7345, "step": 96600 }, { "epoch": 1.0858147542733243, "grad_norm": 1.2310138940811157, "learning_rate": 1.5658289424902544e-05, "loss": 3.6589, "step": 96650 }, { "epoch": 1.0863764794436674, "grad_norm": 1.3286763429641724, "learning_rate": 1.5656042511599693e-05, "loss": 3.7821, "step": 96700 }, { "epoch": 1.0869382046140106, "grad_norm": 1.1189380884170532, "learning_rate": 1.5653795598296842e-05, "loss": 3.6979, "step": 96750 }, { "epoch": 1.0874999297843537, "grad_norm": 0.9450409412384033, "learning_rate": 1.565154868499399e-05, "loss": 3.7887, "step": 96800 }, { "epoch": 1.0880616549546969, "grad_norm": 1.258392572402954, "learning_rate": 1.564930177169114e-05, "loss": 3.7496, "step": 96850 }, { "epoch": 1.08862338012504, "grad_norm": 1.073094367980957, "learning_rate": 1.564705485838829e-05, "loss": 3.7592, "step": 96900 }, { "epoch": 1.0891851052953831, "grad_norm": 1.1871005296707153, "learning_rate": 1.564480794508544e-05, "loss": 3.6373, "step": 96950 }, { "epoch": 1.0897468304657263, "grad_norm": 1.2509921789169312, "learning_rate": 1.5642561031782592e-05, "loss": 3.8089, "step": 97000 }, { "epoch": 1.0903085556360694, "grad_norm": 1.1163924932479858, "learning_rate": 1.564031411847974e-05, "loss": 3.7584, "step": 97050 }, { "epoch": 1.0908702808064126, "grad_norm": 1.2207306623458862, "learning_rate": 1.563806720517689e-05, "loss": 3.7147, "step": 97100 }, { "epoch": 1.0914320059767557, "grad_norm": 1.2841033935546875, "learning_rate": 1.563582029187404e-05, "loss": 3.8064, "step": 97150 }, { "epoch": 1.0919937311470989, "grad_norm": 1.4015285968780518, "learning_rate": 1.563357337857119e-05, "loss": 3.6844, "step": 97200 }, { "epoch": 1.0925554563174422, "grad_norm": 1.2208726406097412, "learning_rate": 1.563132646526834e-05, "loss": 3.7054, "step": 97250 }, { "epoch": 1.0931171814877854, "grad_norm": 1.1583671569824219, "learning_rate": 1.5629079551965488e-05, "loss": 3.644, "step": 97300 }, { "epoch": 1.0936789066581285, "grad_norm": 1.1001847982406616, "learning_rate": 1.5626832638662637e-05, "loss": 3.8188, "step": 97350 }, { "epoch": 1.0942406318284716, "grad_norm": 1.0487697124481201, "learning_rate": 1.562458572535979e-05, "loss": 3.7028, "step": 97400 }, { "epoch": 1.0948023569988148, "grad_norm": 0.9969462752342224, "learning_rate": 1.562233881205694e-05, "loss": 3.7467, "step": 97450 }, { "epoch": 1.095364082169158, "grad_norm": 1.1875361204147339, "learning_rate": 1.562009189875409e-05, "loss": 3.6715, "step": 97500 }, { "epoch": 1.095925807339501, "grad_norm": 1.092423915863037, "learning_rate": 1.5617844985451238e-05, "loss": 3.7414, "step": 97550 }, { "epoch": 1.0964875325098442, "grad_norm": 1.3828538656234741, "learning_rate": 1.5615598072148387e-05, "loss": 3.7184, "step": 97600 }, { "epoch": 1.0970492576801874, "grad_norm": 1.3633829355239868, "learning_rate": 1.5613351158845537e-05, "loss": 3.7591, "step": 97650 }, { "epoch": 1.0976109828505305, "grad_norm": 0.983437716960907, "learning_rate": 1.5611104245542686e-05, "loss": 3.6628, "step": 97700 }, { "epoch": 1.0981727080208736, "grad_norm": 1.1300578117370605, "learning_rate": 1.560885733223984e-05, "loss": 3.7409, "step": 97750 }, { "epoch": 1.0987344331912168, "grad_norm": 1.1857285499572754, "learning_rate": 1.5606610418936988e-05, "loss": 3.6844, "step": 97800 }, { "epoch": 1.0992961583615601, "grad_norm": 1.1955124139785767, "learning_rate": 1.5604363505634137e-05, "loss": 3.7202, "step": 97850 }, { "epoch": 1.0998578835319033, "grad_norm": 0.9688912630081177, "learning_rate": 1.5602116592331287e-05, "loss": 3.7509, "step": 97900 }, { "epoch": 1.1004196087022464, "grad_norm": 1.1819149255752563, "learning_rate": 1.5599869679028436e-05, "loss": 3.7001, "step": 97950 }, { "epoch": 1.1009813338725896, "grad_norm": 1.1008272171020508, "learning_rate": 1.5597622765725585e-05, "loss": 3.7148, "step": 98000 }, { "epoch": 1.1015430590429327, "grad_norm": 1.1886111497879028, "learning_rate": 1.5595375852422735e-05, "loss": 3.7624, "step": 98050 }, { "epoch": 1.1021047842132758, "grad_norm": 0.9735630750656128, "learning_rate": 1.5593128939119887e-05, "loss": 3.7987, "step": 98100 }, { "epoch": 1.102666509383619, "grad_norm": 0.9906808733940125, "learning_rate": 1.5590882025817037e-05, "loss": 3.7065, "step": 98150 }, { "epoch": 1.1032282345539621, "grad_norm": 1.0516225099563599, "learning_rate": 1.5588635112514186e-05, "loss": 3.7193, "step": 98200 }, { "epoch": 1.1037899597243053, "grad_norm": 1.4463839530944824, "learning_rate": 1.5586388199211335e-05, "loss": 3.7413, "step": 98250 }, { "epoch": 1.1043516848946484, "grad_norm": 1.1022162437438965, "learning_rate": 1.5584141285908484e-05, "loss": 3.671, "step": 98300 }, { "epoch": 1.1049134100649916, "grad_norm": 1.0157989263534546, "learning_rate": 1.5581894372605634e-05, "loss": 3.671, "step": 98350 }, { "epoch": 1.1054751352353347, "grad_norm": 1.3087482452392578, "learning_rate": 1.5579647459302783e-05, "loss": 3.7227, "step": 98400 }, { "epoch": 1.1060368604056778, "grad_norm": 1.2353326082229614, "learning_rate": 1.5577400545999936e-05, "loss": 3.6886, "step": 98450 }, { "epoch": 1.106598585576021, "grad_norm": 1.1072086095809937, "learning_rate": 1.5575153632697085e-05, "loss": 3.7556, "step": 98500 }, { "epoch": 1.1071603107463641, "grad_norm": 1.0533146858215332, "learning_rate": 1.5572906719394234e-05, "loss": 3.6718, "step": 98550 }, { "epoch": 1.1077220359167075, "grad_norm": 1.1676750183105469, "learning_rate": 1.5570659806091384e-05, "loss": 3.7657, "step": 98600 }, { "epoch": 1.1082837610870506, "grad_norm": 1.3429557085037231, "learning_rate": 1.5568412892788533e-05, "loss": 3.7055, "step": 98650 }, { "epoch": 1.1088454862573938, "grad_norm": 1.054917573928833, "learning_rate": 1.5566165979485682e-05, "loss": 3.7481, "step": 98700 }, { "epoch": 1.109407211427737, "grad_norm": 1.1266285181045532, "learning_rate": 1.5563919066182832e-05, "loss": 3.6911, "step": 98750 }, { "epoch": 1.10996893659808, "grad_norm": 1.0789703130722046, "learning_rate": 1.5561672152879984e-05, "loss": 3.762, "step": 98800 }, { "epoch": 1.1105306617684232, "grad_norm": 1.295448660850525, "learning_rate": 1.5559425239577134e-05, "loss": 3.7771, "step": 98850 }, { "epoch": 1.1110923869387663, "grad_norm": 1.0544154644012451, "learning_rate": 1.5557178326274283e-05, "loss": 3.6607, "step": 98900 }, { "epoch": 1.1116541121091095, "grad_norm": 1.0492252111434937, "learning_rate": 1.5554931412971432e-05, "loss": 3.7173, "step": 98950 }, { "epoch": 1.1122158372794526, "grad_norm": 1.2639553546905518, "learning_rate": 1.5552684499668582e-05, "loss": 3.7465, "step": 99000 }, { "epoch": 1.1127775624497958, "grad_norm": 1.0499306917190552, "learning_rate": 1.555043758636573e-05, "loss": 3.6379, "step": 99050 }, { "epoch": 1.113339287620139, "grad_norm": 1.1774410009384155, "learning_rate": 1.554819067306288e-05, "loss": 3.7211, "step": 99100 }, { "epoch": 1.113901012790482, "grad_norm": 1.1847046613693237, "learning_rate": 1.5545943759760033e-05, "loss": 3.7337, "step": 99150 }, { "epoch": 1.1144627379608254, "grad_norm": 0.9903504252433777, "learning_rate": 1.5543696846457182e-05, "loss": 3.7566, "step": 99200 }, { "epoch": 1.1150244631311685, "grad_norm": 1.2131892442703247, "learning_rate": 1.554144993315433e-05, "loss": 3.8094, "step": 99250 }, { "epoch": 1.1155861883015117, "grad_norm": 1.0200257301330566, "learning_rate": 1.553920301985148e-05, "loss": 3.7359, "step": 99300 }, { "epoch": 1.1161479134718548, "grad_norm": 1.1790155172348022, "learning_rate": 1.553695610654863e-05, "loss": 3.6969, "step": 99350 }, { "epoch": 1.116709638642198, "grad_norm": 1.2871062755584717, "learning_rate": 1.553470919324578e-05, "loss": 3.7804, "step": 99400 }, { "epoch": 1.117271363812541, "grad_norm": 1.1888140439987183, "learning_rate": 1.553246227994293e-05, "loss": 3.7267, "step": 99450 }, { "epoch": 1.1178330889828842, "grad_norm": 1.1498525142669678, "learning_rate": 1.553021536664008e-05, "loss": 3.7379, "step": 99500 }, { "epoch": 1.1183948141532274, "grad_norm": 1.5620925426483154, "learning_rate": 1.552796845333723e-05, "loss": 3.6527, "step": 99550 }, { "epoch": 1.1189565393235705, "grad_norm": 1.0870367288589478, "learning_rate": 1.552572154003438e-05, "loss": 3.7163, "step": 99600 }, { "epoch": 1.1195182644939137, "grad_norm": 1.0825954675674438, "learning_rate": 1.552347462673153e-05, "loss": 3.667, "step": 99650 }, { "epoch": 1.1200799896642568, "grad_norm": 1.2793874740600586, "learning_rate": 1.552122771342868e-05, "loss": 3.6803, "step": 99700 }, { "epoch": 1.1206417148346, "grad_norm": 1.1869399547576904, "learning_rate": 1.5518980800125828e-05, "loss": 3.7173, "step": 99750 }, { "epoch": 1.121203440004943, "grad_norm": 1.2226241827011108, "learning_rate": 1.5516733886822978e-05, "loss": 3.6696, "step": 99800 }, { "epoch": 1.1217651651752862, "grad_norm": 1.0593713521957397, "learning_rate": 1.551448697352013e-05, "loss": 3.7269, "step": 99850 }, { "epoch": 1.1223268903456294, "grad_norm": 1.244052767753601, "learning_rate": 1.551224006021728e-05, "loss": 3.749, "step": 99900 }, { "epoch": 1.1228886155159727, "grad_norm": 0.9682693481445312, "learning_rate": 1.550999314691443e-05, "loss": 3.7029, "step": 99950 }, { "epoch": 1.1234503406863159, "grad_norm": 1.2320550680160522, "learning_rate": 1.5507746233611578e-05, "loss": 3.6564, "step": 100000 }, { "epoch": 1.124012065856659, "grad_norm": 1.021205186843872, "learning_rate": 1.5505499320308728e-05, "loss": 3.6825, "step": 100050 }, { "epoch": 1.1245737910270022, "grad_norm": 1.2643808126449585, "learning_rate": 1.5503252407005877e-05, "loss": 3.75, "step": 100100 }, { "epoch": 1.1251355161973453, "grad_norm": 1.2180043458938599, "learning_rate": 1.5501005493703026e-05, "loss": 3.7175, "step": 100150 }, { "epoch": 1.1256972413676885, "grad_norm": 1.088374137878418, "learning_rate": 1.549875858040018e-05, "loss": 3.7444, "step": 100200 }, { "epoch": 1.1262589665380316, "grad_norm": 1.0124397277832031, "learning_rate": 1.5496556605363385e-05, "loss": 3.633, "step": 100250 }, { "epoch": 1.1268206917083747, "grad_norm": 1.0882328748703003, "learning_rate": 1.5494309692060534e-05, "loss": 3.699, "step": 100300 }, { "epoch": 1.1273824168787179, "grad_norm": 1.055010199546814, "learning_rate": 1.5492062778757683e-05, "loss": 3.6626, "step": 100350 }, { "epoch": 1.127944142049061, "grad_norm": 1.2953052520751953, "learning_rate": 1.5489815865454833e-05, "loss": 3.75, "step": 100400 }, { "epoch": 1.1285058672194042, "grad_norm": 1.1922369003295898, "learning_rate": 1.5487568952151982e-05, "loss": 3.6626, "step": 100450 }, { "epoch": 1.1290675923897473, "grad_norm": 1.0497500896453857, "learning_rate": 1.5485322038849135e-05, "loss": 3.7183, "step": 100500 }, { "epoch": 1.1296293175600907, "grad_norm": 1.1612207889556885, "learning_rate": 1.5483075125546284e-05, "loss": 3.7178, "step": 100550 }, { "epoch": 1.1301910427304338, "grad_norm": 1.0290371179580688, "learning_rate": 1.5480828212243433e-05, "loss": 3.783, "step": 100600 }, { "epoch": 1.130752767900777, "grad_norm": 1.1671088933944702, "learning_rate": 1.5478581298940583e-05, "loss": 3.7467, "step": 100650 }, { "epoch": 1.13131449307112, "grad_norm": 1.2563223838806152, "learning_rate": 1.5476334385637732e-05, "loss": 3.6544, "step": 100700 }, { "epoch": 1.1318762182414632, "grad_norm": 1.2090439796447754, "learning_rate": 1.547408747233488e-05, "loss": 3.7014, "step": 100750 }, { "epoch": 1.1324379434118064, "grad_norm": 1.234808325767517, "learning_rate": 1.547184055903203e-05, "loss": 3.6535, "step": 100800 }, { "epoch": 1.1329996685821495, "grad_norm": 1.047635793685913, "learning_rate": 1.5469593645729183e-05, "loss": 3.6579, "step": 100850 }, { "epoch": 1.1335613937524927, "grad_norm": 1.0435529947280884, "learning_rate": 1.5467346732426333e-05, "loss": 3.6524, "step": 100900 }, { "epoch": 1.1341231189228358, "grad_norm": 1.3722610473632812, "learning_rate": 1.5465099819123482e-05, "loss": 3.7106, "step": 100950 }, { "epoch": 1.134684844093179, "grad_norm": 1.0594531297683716, "learning_rate": 1.546285290582063e-05, "loss": 3.6417, "step": 101000 }, { "epoch": 1.135246569263522, "grad_norm": 0.9762507677078247, "learning_rate": 1.546060599251778e-05, "loss": 3.7199, "step": 101050 }, { "epoch": 1.1358082944338652, "grad_norm": 1.0277429819107056, "learning_rate": 1.545835907921493e-05, "loss": 3.6726, "step": 101100 }, { "epoch": 1.1363700196042084, "grad_norm": 1.122733473777771, "learning_rate": 1.545611216591208e-05, "loss": 3.6344, "step": 101150 }, { "epoch": 1.1369317447745515, "grad_norm": 1.0020467042922974, "learning_rate": 1.5453910190875285e-05, "loss": 3.6257, "step": 101200 }, { "epoch": 1.1374934699448946, "grad_norm": 1.2245641946792603, "learning_rate": 1.5451663277572434e-05, "loss": 3.7534, "step": 101250 }, { "epoch": 1.138055195115238, "grad_norm": 1.1989482641220093, "learning_rate": 1.5449416364269584e-05, "loss": 3.68, "step": 101300 }, { "epoch": 1.1386169202855811, "grad_norm": 1.059341311454773, "learning_rate": 1.5447169450966733e-05, "loss": 3.7051, "step": 101350 }, { "epoch": 1.1391786454559243, "grad_norm": 1.4291677474975586, "learning_rate": 1.5444922537663886e-05, "loss": 3.7544, "step": 101400 }, { "epoch": 1.1397403706262674, "grad_norm": 1.1562519073486328, "learning_rate": 1.5442675624361035e-05, "loss": 3.7279, "step": 101450 }, { "epoch": 1.1403020957966106, "grad_norm": 1.0684752464294434, "learning_rate": 1.5440428711058184e-05, "loss": 3.7251, "step": 101500 }, { "epoch": 1.1408638209669537, "grad_norm": 1.074190616607666, "learning_rate": 1.5438181797755334e-05, "loss": 3.7438, "step": 101550 }, { "epoch": 1.1414255461372969, "grad_norm": 1.1465007066726685, "learning_rate": 1.5435934884452483e-05, "loss": 3.7327, "step": 101600 }, { "epoch": 1.14198727130764, "grad_norm": 1.063234567642212, "learning_rate": 1.5433687971149632e-05, "loss": 3.69, "step": 101650 }, { "epoch": 1.1425489964779831, "grad_norm": 1.1495132446289062, "learning_rate": 1.5431441057846782e-05, "loss": 3.7358, "step": 101700 }, { "epoch": 1.1431107216483263, "grad_norm": 1.066072702407837, "learning_rate": 1.5429194144543934e-05, "loss": 3.6587, "step": 101750 }, { "epoch": 1.1436724468186694, "grad_norm": 1.6146560907363892, "learning_rate": 1.5426947231241084e-05, "loss": 3.7906, "step": 101800 }, { "epoch": 1.1442341719890126, "grad_norm": 1.0587708950042725, "learning_rate": 1.5424700317938233e-05, "loss": 3.6794, "step": 101850 }, { "epoch": 1.144795897159356, "grad_norm": 1.157436728477478, "learning_rate": 1.5422453404635382e-05, "loss": 3.7041, "step": 101900 }, { "epoch": 1.145357622329699, "grad_norm": 1.051474690437317, "learning_rate": 1.5420206491332532e-05, "loss": 3.6958, "step": 101950 }, { "epoch": 1.1459193475000422, "grad_norm": 1.2667258977890015, "learning_rate": 1.541795957802968e-05, "loss": 3.7161, "step": 102000 }, { "epoch": 1.1464810726703853, "grad_norm": 1.2712479829788208, "learning_rate": 1.541571266472683e-05, "loss": 3.7091, "step": 102050 }, { "epoch": 1.1470427978407285, "grad_norm": 1.0357364416122437, "learning_rate": 1.5413465751423983e-05, "loss": 3.7355, "step": 102100 }, { "epoch": 1.1476045230110716, "grad_norm": 1.2836031913757324, "learning_rate": 1.5411218838121132e-05, "loss": 3.7377, "step": 102150 }, { "epoch": 1.1481662481814148, "grad_norm": 1.155813217163086, "learning_rate": 1.540897192481828e-05, "loss": 3.7245, "step": 102200 }, { "epoch": 1.148727973351758, "grad_norm": 1.5667455196380615, "learning_rate": 1.540672501151543e-05, "loss": 3.6165, "step": 102250 }, { "epoch": 1.149289698522101, "grad_norm": 1.0747267007827759, "learning_rate": 1.540447809821258e-05, "loss": 3.6887, "step": 102300 }, { "epoch": 1.1498514236924442, "grad_norm": 1.7879501581192017, "learning_rate": 1.540223118490973e-05, "loss": 3.6892, "step": 102350 }, { "epoch": 1.1504131488627873, "grad_norm": 1.2933627367019653, "learning_rate": 1.539998427160688e-05, "loss": 3.7902, "step": 102400 }, { "epoch": 1.1509748740331305, "grad_norm": 1.0486758947372437, "learning_rate": 1.539773735830403e-05, "loss": 3.6981, "step": 102450 }, { "epoch": 1.1515365992034736, "grad_norm": 1.2376704216003418, "learning_rate": 1.539549044500118e-05, "loss": 3.7477, "step": 102500 }, { "epoch": 1.1520983243738168, "grad_norm": 1.3571228981018066, "learning_rate": 1.539324353169833e-05, "loss": 3.7425, "step": 102550 }, { "epoch": 1.15266004954416, "grad_norm": 1.027738094329834, "learning_rate": 1.539099661839548e-05, "loss": 3.7008, "step": 102600 }, { "epoch": 1.1532217747145033, "grad_norm": 1.1306037902832031, "learning_rate": 1.538874970509263e-05, "loss": 3.6775, "step": 102650 }, { "epoch": 1.1537834998848464, "grad_norm": 1.0368131399154663, "learning_rate": 1.5386502791789778e-05, "loss": 3.6883, "step": 102700 }, { "epoch": 1.1543452250551895, "grad_norm": 1.1494563817977905, "learning_rate": 1.5384255878486928e-05, "loss": 3.6778, "step": 102750 }, { "epoch": 1.1549069502255327, "grad_norm": 1.192533254623413, "learning_rate": 1.538200896518408e-05, "loss": 3.6954, "step": 102800 }, { "epoch": 1.1554686753958758, "grad_norm": 1.1109049320220947, "learning_rate": 1.537976205188123e-05, "loss": 3.7542, "step": 102850 }, { "epoch": 1.156030400566219, "grad_norm": 1.0070918798446655, "learning_rate": 1.537751513857838e-05, "loss": 3.8029, "step": 102900 }, { "epoch": 1.1565921257365621, "grad_norm": 0.9579331874847412, "learning_rate": 1.5375268225275528e-05, "loss": 3.7035, "step": 102950 }, { "epoch": 1.1571538509069053, "grad_norm": 1.4610426425933838, "learning_rate": 1.5373021311972678e-05, "loss": 3.6512, "step": 103000 }, { "epoch": 1.1577155760772484, "grad_norm": 1.1504992246627808, "learning_rate": 1.5370774398669827e-05, "loss": 3.6739, "step": 103050 }, { "epoch": 1.1582773012475915, "grad_norm": 2.7192490100860596, "learning_rate": 1.5368527485366976e-05, "loss": 3.7345, "step": 103100 }, { "epoch": 1.1588390264179347, "grad_norm": 1.1754087209701538, "learning_rate": 1.536628057206413e-05, "loss": 3.7255, "step": 103150 }, { "epoch": 1.1594007515882778, "grad_norm": 1.2351810932159424, "learning_rate": 1.5364033658761278e-05, "loss": 3.6805, "step": 103200 }, { "epoch": 1.1599624767586212, "grad_norm": 1.058080792427063, "learning_rate": 1.5361786745458427e-05, "loss": 3.6916, "step": 103250 }, { "epoch": 1.1605242019289643, "grad_norm": 1.3002619743347168, "learning_rate": 1.5359539832155577e-05, "loss": 3.7551, "step": 103300 }, { "epoch": 1.1610859270993075, "grad_norm": 1.1462116241455078, "learning_rate": 1.5357292918852726e-05, "loss": 3.5689, "step": 103350 }, { "epoch": 1.1616476522696506, "grad_norm": 1.3079191446304321, "learning_rate": 1.5355046005549875e-05, "loss": 3.7718, "step": 103400 }, { "epoch": 1.1622093774399938, "grad_norm": 1.3035516738891602, "learning_rate": 1.5352799092247025e-05, "loss": 3.632, "step": 103450 }, { "epoch": 1.162771102610337, "grad_norm": 1.0845363140106201, "learning_rate": 1.5350552178944177e-05, "loss": 3.5912, "step": 103500 }, { "epoch": 1.16333282778068, "grad_norm": 1.0748491287231445, "learning_rate": 1.5348305265641327e-05, "loss": 3.6705, "step": 103550 }, { "epoch": 1.1638945529510232, "grad_norm": 1.2455357313156128, "learning_rate": 1.5346058352338476e-05, "loss": 3.7515, "step": 103600 }, { "epoch": 1.1644562781213663, "grad_norm": 1.014829397201538, "learning_rate": 1.5343811439035625e-05, "loss": 3.6542, "step": 103650 }, { "epoch": 1.1650180032917095, "grad_norm": 1.1986323595046997, "learning_rate": 1.5341564525732775e-05, "loss": 3.7361, "step": 103700 }, { "epoch": 1.1655797284620526, "grad_norm": 1.2283251285552979, "learning_rate": 1.5339317612429924e-05, "loss": 3.7181, "step": 103750 }, { "epoch": 1.1661414536323957, "grad_norm": 1.078841209411621, "learning_rate": 1.5337070699127073e-05, "loss": 3.6782, "step": 103800 }, { "epoch": 1.1667031788027389, "grad_norm": 1.1124110221862793, "learning_rate": 1.5334823785824226e-05, "loss": 3.7446, "step": 103850 }, { "epoch": 1.167264903973082, "grad_norm": 1.2355866432189941, "learning_rate": 1.5332621810787432e-05, "loss": 3.6567, "step": 103900 }, { "epoch": 1.1678266291434252, "grad_norm": 1.1304376125335693, "learning_rate": 1.533037489748458e-05, "loss": 3.7139, "step": 103950 }, { "epoch": 1.1683883543137685, "grad_norm": 1.1909832954406738, "learning_rate": 1.532812798418173e-05, "loss": 3.6523, "step": 104000 }, { "epoch": 1.1689500794841117, "grad_norm": 1.1495308876037598, "learning_rate": 1.532588107087888e-05, "loss": 3.6911, "step": 104050 }, { "epoch": 1.1695118046544548, "grad_norm": 1.2217122316360474, "learning_rate": 1.532363415757603e-05, "loss": 3.7879, "step": 104100 }, { "epoch": 1.170073529824798, "grad_norm": 1.1243988275527954, "learning_rate": 1.5321387244273182e-05, "loss": 3.7405, "step": 104150 }, { "epoch": 1.170635254995141, "grad_norm": 1.328789234161377, "learning_rate": 1.531914033097033e-05, "loss": 3.7426, "step": 104200 }, { "epoch": 1.1711969801654842, "grad_norm": 1.166298270225525, "learning_rate": 1.531689341766748e-05, "loss": 3.6431, "step": 104250 }, { "epoch": 1.1717587053358274, "grad_norm": 1.3083101511001587, "learning_rate": 1.531464650436463e-05, "loss": 3.7106, "step": 104300 }, { "epoch": 1.1723204305061705, "grad_norm": 1.0733588933944702, "learning_rate": 1.531239959106178e-05, "loss": 3.6376, "step": 104350 }, { "epoch": 1.1728821556765137, "grad_norm": 1.035312533378601, "learning_rate": 1.531015267775893e-05, "loss": 3.7374, "step": 104400 }, { "epoch": 1.1734438808468568, "grad_norm": 0.9976434707641602, "learning_rate": 1.5307905764456078e-05, "loss": 3.7363, "step": 104450 }, { "epoch": 1.1740056060172, "grad_norm": 1.200663685798645, "learning_rate": 1.530565885115323e-05, "loss": 3.8055, "step": 104500 }, { "epoch": 1.1745673311875433, "grad_norm": 1.2299480438232422, "learning_rate": 1.5303456876116436e-05, "loss": 3.743, "step": 104550 }, { "epoch": 1.1751290563578864, "grad_norm": 1.0745593309402466, "learning_rate": 1.5301209962813586e-05, "loss": 3.6963, "step": 104600 }, { "epoch": 1.1756907815282296, "grad_norm": 1.1543883085250854, "learning_rate": 1.5298963049510735e-05, "loss": 3.7487, "step": 104650 }, { "epoch": 1.1762525066985727, "grad_norm": 1.074861764907837, "learning_rate": 1.5296716136207884e-05, "loss": 3.7323, "step": 104700 }, { "epoch": 1.1768142318689159, "grad_norm": 1.0769885778427124, "learning_rate": 1.5294469222905034e-05, "loss": 3.7242, "step": 104750 }, { "epoch": 1.177375957039259, "grad_norm": 1.5533466339111328, "learning_rate": 1.5292222309602186e-05, "loss": 3.7333, "step": 104800 }, { "epoch": 1.1779376822096022, "grad_norm": 1.3578367233276367, "learning_rate": 1.5289975396299336e-05, "loss": 3.6818, "step": 104850 }, { "epoch": 1.1784994073799453, "grad_norm": 1.153058648109436, "learning_rate": 1.5287728482996485e-05, "loss": 3.6829, "step": 104900 }, { "epoch": 1.1790611325502884, "grad_norm": 1.275603175163269, "learning_rate": 1.5285481569693634e-05, "loss": 3.6589, "step": 104950 }, { "epoch": 1.1796228577206316, "grad_norm": 1.2551240921020508, "learning_rate": 1.5283234656390784e-05, "loss": 3.7048, "step": 105000 }, { "epoch": 1.1801845828909747, "grad_norm": 1.1533915996551514, "learning_rate": 1.5280987743087933e-05, "loss": 3.7098, "step": 105050 }, { "epoch": 1.1807463080613179, "grad_norm": 1.2769936323165894, "learning_rate": 1.5278740829785082e-05, "loss": 3.6012, "step": 105100 }, { "epoch": 1.181308033231661, "grad_norm": 1.0044163465499878, "learning_rate": 1.5276493916482235e-05, "loss": 3.7165, "step": 105150 }, { "epoch": 1.1818697584020041, "grad_norm": 1.1639456748962402, "learning_rate": 1.5274247003179384e-05, "loss": 3.7855, "step": 105200 }, { "epoch": 1.1824314835723473, "grad_norm": 1.2102593183517456, "learning_rate": 1.5272000089876534e-05, "loss": 3.6921, "step": 105250 }, { "epoch": 1.1829932087426904, "grad_norm": 1.1959712505340576, "learning_rate": 1.5269753176573683e-05, "loss": 3.714, "step": 105300 }, { "epoch": 1.1835549339130338, "grad_norm": 0.9928191304206848, "learning_rate": 1.5267506263270832e-05, "loss": 3.725, "step": 105350 }, { "epoch": 1.184116659083377, "grad_norm": 1.190778374671936, "learning_rate": 1.526525934996798e-05, "loss": 3.7029, "step": 105400 }, { "epoch": 1.18467838425372, "grad_norm": 1.3059523105621338, "learning_rate": 1.526301243666513e-05, "loss": 3.6104, "step": 105450 }, { "epoch": 1.1852401094240632, "grad_norm": 1.4401763677597046, "learning_rate": 1.5260765523362284e-05, "loss": 3.7813, "step": 105500 }, { "epoch": 1.1858018345944064, "grad_norm": 1.1502825021743774, "learning_rate": 1.5258518610059431e-05, "loss": 3.7401, "step": 105550 }, { "epoch": 1.1863635597647495, "grad_norm": 1.2435052394866943, "learning_rate": 1.5256271696756582e-05, "loss": 3.7393, "step": 105600 }, { "epoch": 1.1869252849350926, "grad_norm": 2.0304958820343018, "learning_rate": 1.5254024783453732e-05, "loss": 3.7324, "step": 105650 }, { "epoch": 1.1874870101054358, "grad_norm": 1.141411542892456, "learning_rate": 1.5251777870150881e-05, "loss": 3.655, "step": 105700 }, { "epoch": 1.188048735275779, "grad_norm": 1.2367336750030518, "learning_rate": 1.524953095684803e-05, "loss": 3.6525, "step": 105750 }, { "epoch": 1.188610460446122, "grad_norm": 1.052817463874817, "learning_rate": 1.5247284043545181e-05, "loss": 3.7261, "step": 105800 }, { "epoch": 1.1891721856164652, "grad_norm": 1.1793750524520874, "learning_rate": 1.524503713024233e-05, "loss": 3.7451, "step": 105850 }, { "epoch": 1.1897339107868086, "grad_norm": 1.0441009998321533, "learning_rate": 1.524279021693948e-05, "loss": 3.6667, "step": 105900 }, { "epoch": 1.1902956359571517, "grad_norm": 1.2185910940170288, "learning_rate": 1.5240543303636631e-05, "loss": 3.7966, "step": 105950 }, { "epoch": 1.1908573611274949, "grad_norm": 1.138037919998169, "learning_rate": 1.523829639033378e-05, "loss": 3.6983, "step": 106000 }, { "epoch": 1.191419086297838, "grad_norm": 1.0619055032730103, "learning_rate": 1.523604947703093e-05, "loss": 3.7142, "step": 106050 }, { "epoch": 1.1919808114681811, "grad_norm": 1.0743881464004517, "learning_rate": 1.5233802563728079e-05, "loss": 3.7762, "step": 106100 }, { "epoch": 1.1925425366385243, "grad_norm": 1.375406265258789, "learning_rate": 1.523155565042523e-05, "loss": 3.736, "step": 106150 }, { "epoch": 1.1931042618088674, "grad_norm": 1.1640788316726685, "learning_rate": 1.522930873712238e-05, "loss": 3.7453, "step": 106200 }, { "epoch": 1.1936659869792106, "grad_norm": 1.2602263689041138, "learning_rate": 1.5227061823819528e-05, "loss": 3.6928, "step": 106250 }, { "epoch": 1.1942277121495537, "grad_norm": 1.1936029195785522, "learning_rate": 1.522481491051668e-05, "loss": 3.6771, "step": 106300 }, { "epoch": 1.1947894373198968, "grad_norm": 1.2278850078582764, "learning_rate": 1.5222567997213829e-05, "loss": 3.7069, "step": 106350 }, { "epoch": 1.19535116249024, "grad_norm": 1.2584307193756104, "learning_rate": 1.5220321083910978e-05, "loss": 3.6803, "step": 106400 }, { "epoch": 1.1959128876605831, "grad_norm": 1.0643031597137451, "learning_rate": 1.5218074170608127e-05, "loss": 3.6955, "step": 106450 }, { "epoch": 1.1964746128309263, "grad_norm": 1.1142960786819458, "learning_rate": 1.5215827257305278e-05, "loss": 3.7017, "step": 106500 }, { "epoch": 1.1970363380012694, "grad_norm": 1.0555589199066162, "learning_rate": 1.5213580344002428e-05, "loss": 3.6561, "step": 106550 }, { "epoch": 1.1975980631716125, "grad_norm": 0.980929434299469, "learning_rate": 1.5211333430699577e-05, "loss": 3.7023, "step": 106600 }, { "epoch": 1.198159788341956, "grad_norm": 1.0431848764419556, "learning_rate": 1.5209086517396726e-05, "loss": 3.6619, "step": 106650 }, { "epoch": 1.198721513512299, "grad_norm": 0.95896315574646, "learning_rate": 1.5206839604093877e-05, "loss": 3.7181, "step": 106700 }, { "epoch": 1.1992832386826422, "grad_norm": 1.2175319194793701, "learning_rate": 1.5204592690791027e-05, "loss": 3.6748, "step": 106750 }, { "epoch": 1.1998449638529853, "grad_norm": 1.2603718042373657, "learning_rate": 1.5202345777488176e-05, "loss": 3.7236, "step": 106800 }, { "epoch": 1.2004066890233285, "grad_norm": 1.4466156959533691, "learning_rate": 1.5200098864185327e-05, "loss": 3.6938, "step": 106850 }, { "epoch": 1.2009684141936716, "grad_norm": 1.2379579544067383, "learning_rate": 1.5197851950882476e-05, "loss": 3.7017, "step": 106900 }, { "epoch": 1.2015301393640148, "grad_norm": 1.505018711090088, "learning_rate": 1.5195605037579626e-05, "loss": 3.6715, "step": 106950 }, { "epoch": 1.202091864534358, "grad_norm": 1.1268073320388794, "learning_rate": 1.5193358124276775e-05, "loss": 3.6606, "step": 107000 }, { "epoch": 1.202653589704701, "grad_norm": 1.1279513835906982, "learning_rate": 1.5191111210973926e-05, "loss": 3.6994, "step": 107050 }, { "epoch": 1.2032153148750442, "grad_norm": 1.2397137880325317, "learning_rate": 1.5188864297671075e-05, "loss": 3.7105, "step": 107100 }, { "epoch": 1.2037770400453873, "grad_norm": 1.1287885904312134, "learning_rate": 1.5186617384368225e-05, "loss": 3.6917, "step": 107150 }, { "epoch": 1.2043387652157305, "grad_norm": 1.0952703952789307, "learning_rate": 1.5184370471065376e-05, "loss": 3.7041, "step": 107200 }, { "epoch": 1.2049004903860738, "grad_norm": 1.8485311269760132, "learning_rate": 1.5182123557762525e-05, "loss": 3.7, "step": 107250 }, { "epoch": 1.205462215556417, "grad_norm": 1.0265604257583618, "learning_rate": 1.5179876644459674e-05, "loss": 3.7924, "step": 107300 }, { "epoch": 1.2060239407267601, "grad_norm": 1.2580769062042236, "learning_rate": 1.5177629731156824e-05, "loss": 3.7509, "step": 107350 }, { "epoch": 1.2065856658971033, "grad_norm": 1.2376199960708618, "learning_rate": 1.5175382817853975e-05, "loss": 3.7154, "step": 107400 }, { "epoch": 1.2071473910674464, "grad_norm": 1.0325915813446045, "learning_rate": 1.5173135904551124e-05, "loss": 3.7247, "step": 107450 }, { "epoch": 1.2077091162377895, "grad_norm": 1.171482801437378, "learning_rate": 1.5170888991248273e-05, "loss": 3.6708, "step": 107500 }, { "epoch": 1.2082708414081327, "grad_norm": 1.3903834819793701, "learning_rate": 1.5168642077945424e-05, "loss": 3.6801, "step": 107550 }, { "epoch": 1.2088325665784758, "grad_norm": 1.2108118534088135, "learning_rate": 1.5166395164642574e-05, "loss": 3.7365, "step": 107600 }, { "epoch": 1.209394291748819, "grad_norm": 1.1592521667480469, "learning_rate": 1.5164148251339723e-05, "loss": 3.6927, "step": 107650 }, { "epoch": 1.209956016919162, "grad_norm": 1.0881690979003906, "learning_rate": 1.5161901338036872e-05, "loss": 3.682, "step": 107700 }, { "epoch": 1.2105177420895052, "grad_norm": 1.1194993257522583, "learning_rate": 1.5159654424734023e-05, "loss": 3.6634, "step": 107750 }, { "epoch": 1.2110794672598484, "grad_norm": 1.1075609922409058, "learning_rate": 1.5157407511431173e-05, "loss": 3.7438, "step": 107800 }, { "epoch": 1.2116411924301915, "grad_norm": 1.0043065547943115, "learning_rate": 1.5155160598128322e-05, "loss": 3.7448, "step": 107850 }, { "epoch": 1.2122029176005347, "grad_norm": 1.1683133840560913, "learning_rate": 1.5152913684825473e-05, "loss": 3.6245, "step": 107900 }, { "epoch": 1.2127646427708778, "grad_norm": 1.2724297046661377, "learning_rate": 1.5150666771522622e-05, "loss": 3.6912, "step": 107950 }, { "epoch": 1.2133263679412212, "grad_norm": 1.1075135469436646, "learning_rate": 1.5148419858219772e-05, "loss": 3.7175, "step": 108000 }, { "epoch": 1.2138880931115643, "grad_norm": 1.2676115036010742, "learning_rate": 1.514617294491692e-05, "loss": 3.7054, "step": 108050 }, { "epoch": 1.2144498182819075, "grad_norm": 1.432173490524292, "learning_rate": 1.5143926031614072e-05, "loss": 3.67, "step": 108100 }, { "epoch": 1.2150115434522506, "grad_norm": 1.1345535516738892, "learning_rate": 1.5141679118311221e-05, "loss": 3.704, "step": 108150 }, { "epoch": 1.2155732686225937, "grad_norm": 1.0536555051803589, "learning_rate": 1.513943220500837e-05, "loss": 3.7215, "step": 108200 }, { "epoch": 1.2161349937929369, "grad_norm": 1.1665971279144287, "learning_rate": 1.5137185291705521e-05, "loss": 3.7347, "step": 108250 }, { "epoch": 1.21669671896328, "grad_norm": 1.1783932447433472, "learning_rate": 1.513493837840267e-05, "loss": 3.6768, "step": 108300 }, { "epoch": 1.2172584441336232, "grad_norm": 0.9993147253990173, "learning_rate": 1.513269146509982e-05, "loss": 3.7377, "step": 108350 }, { "epoch": 1.2178201693039663, "grad_norm": 1.104299783706665, "learning_rate": 1.513044455179697e-05, "loss": 3.7368, "step": 108400 }, { "epoch": 1.2183818944743094, "grad_norm": 1.1289198398590088, "learning_rate": 1.512819763849412e-05, "loss": 3.6997, "step": 108450 }, { "epoch": 1.2189436196446526, "grad_norm": 1.1112141609191895, "learning_rate": 1.512595072519127e-05, "loss": 3.7008, "step": 108500 }, { "epoch": 1.2195053448149957, "grad_norm": 1.1366983652114868, "learning_rate": 1.5123703811888419e-05, "loss": 3.6892, "step": 108550 }, { "epoch": 1.220067069985339, "grad_norm": 1.1454689502716064, "learning_rate": 1.5121456898585568e-05, "loss": 3.767, "step": 108600 }, { "epoch": 1.2206287951556822, "grad_norm": 1.21815824508667, "learning_rate": 1.511920998528272e-05, "loss": 3.724, "step": 108650 }, { "epoch": 1.2211905203260254, "grad_norm": 1.1037623882293701, "learning_rate": 1.5116963071979869e-05, "loss": 3.6244, "step": 108700 }, { "epoch": 1.2217522454963685, "grad_norm": 1.2103321552276611, "learning_rate": 1.5114716158677018e-05, "loss": 3.753, "step": 108750 }, { "epoch": 1.2223139706667117, "grad_norm": 1.3215216398239136, "learning_rate": 1.5112469245374169e-05, "loss": 3.6574, "step": 108800 }, { "epoch": 1.2228756958370548, "grad_norm": 1.1635072231292725, "learning_rate": 1.5110222332071318e-05, "loss": 3.6211, "step": 108850 }, { "epoch": 1.223437421007398, "grad_norm": 1.132358193397522, "learning_rate": 1.5107975418768468e-05, "loss": 3.6865, "step": 108900 }, { "epoch": 1.223999146177741, "grad_norm": 1.190798282623291, "learning_rate": 1.5105728505465617e-05, "loss": 3.7067, "step": 108950 }, { "epoch": 1.2245608713480842, "grad_norm": 1.2921621799468994, "learning_rate": 1.5103481592162768e-05, "loss": 3.7109, "step": 109000 }, { "epoch": 1.2251225965184274, "grad_norm": 1.2048287391662598, "learning_rate": 1.5101234678859917e-05, "loss": 3.6675, "step": 109050 }, { "epoch": 1.2256843216887705, "grad_norm": 1.2336902618408203, "learning_rate": 1.5098987765557067e-05, "loss": 3.6948, "step": 109100 }, { "epoch": 1.2262460468591136, "grad_norm": 1.1223293542861938, "learning_rate": 1.5096740852254218e-05, "loss": 3.6968, "step": 109150 }, { "epoch": 1.2268077720294568, "grad_norm": 1.248460054397583, "learning_rate": 1.5094493938951367e-05, "loss": 3.6594, "step": 109200 }, { "epoch": 1.2273694971998, "grad_norm": 1.082493782043457, "learning_rate": 1.5092291963914575e-05, "loss": 3.7333, "step": 109250 }, { "epoch": 1.227931222370143, "grad_norm": 1.0764968395233154, "learning_rate": 1.5090045050611724e-05, "loss": 3.7282, "step": 109300 }, { "epoch": 1.2284929475404864, "grad_norm": 1.0554648637771606, "learning_rate": 1.5087798137308873e-05, "loss": 3.7384, "step": 109350 }, { "epoch": 1.2290546727108296, "grad_norm": 1.110594630241394, "learning_rate": 1.5085551224006023e-05, "loss": 3.7397, "step": 109400 }, { "epoch": 1.2296163978811727, "grad_norm": 1.0039312839508057, "learning_rate": 1.5083304310703174e-05, "loss": 3.6706, "step": 109450 }, { "epoch": 1.2301781230515159, "grad_norm": 1.1047683954238892, "learning_rate": 1.5081057397400323e-05, "loss": 3.7168, "step": 109500 }, { "epoch": 1.230739848221859, "grad_norm": 1.0532225370407104, "learning_rate": 1.5078810484097472e-05, "loss": 3.6769, "step": 109550 }, { "epoch": 1.2313015733922021, "grad_norm": 1.2127481698989868, "learning_rate": 1.5076563570794621e-05, "loss": 3.7215, "step": 109600 }, { "epoch": 1.2318632985625453, "grad_norm": 1.2573966979980469, "learning_rate": 1.5074316657491773e-05, "loss": 3.6805, "step": 109650 }, { "epoch": 1.2324250237328884, "grad_norm": 1.2704925537109375, "learning_rate": 1.5072069744188922e-05, "loss": 3.7492, "step": 109700 }, { "epoch": 1.2329867489032316, "grad_norm": 1.2403591871261597, "learning_rate": 1.5069822830886071e-05, "loss": 3.7066, "step": 109750 }, { "epoch": 1.2335484740735747, "grad_norm": 1.384395718574524, "learning_rate": 1.5067575917583222e-05, "loss": 3.6857, "step": 109800 }, { "epoch": 1.2341101992439178, "grad_norm": 1.0223932266235352, "learning_rate": 1.5065329004280371e-05, "loss": 3.6618, "step": 109850 }, { "epoch": 1.234671924414261, "grad_norm": 1.2047230005264282, "learning_rate": 1.506308209097752e-05, "loss": 3.7276, "step": 109900 }, { "epoch": 1.2352336495846044, "grad_norm": 0.9568697810173035, "learning_rate": 1.506083517767467e-05, "loss": 3.6579, "step": 109950 }, { "epoch": 1.2357953747549475, "grad_norm": 1.0862027406692505, "learning_rate": 1.5058588264371821e-05, "loss": 3.7278, "step": 110000 }, { "epoch": 1.2363570999252906, "grad_norm": 1.1824320554733276, "learning_rate": 1.505634135106897e-05, "loss": 3.7534, "step": 110050 }, { "epoch": 1.2369188250956338, "grad_norm": 1.0982342958450317, "learning_rate": 1.505409443776612e-05, "loss": 3.7264, "step": 110100 }, { "epoch": 1.237480550265977, "grad_norm": 0.9904276728630066, "learning_rate": 1.505184752446327e-05, "loss": 3.7695, "step": 110150 }, { "epoch": 1.23804227543632, "grad_norm": 0.9733591079711914, "learning_rate": 1.504960061116042e-05, "loss": 3.7002, "step": 110200 }, { "epoch": 1.2386040006066632, "grad_norm": 1.2126539945602417, "learning_rate": 1.504735369785757e-05, "loss": 3.7215, "step": 110250 }, { "epoch": 1.2391657257770063, "grad_norm": 1.3110642433166504, "learning_rate": 1.5045106784554719e-05, "loss": 3.6464, "step": 110300 }, { "epoch": 1.2397274509473495, "grad_norm": 1.0849037170410156, "learning_rate": 1.504285987125187e-05, "loss": 3.6448, "step": 110350 }, { "epoch": 1.2402891761176926, "grad_norm": 1.155277967453003, "learning_rate": 1.5040612957949019e-05, "loss": 3.6634, "step": 110400 }, { "epoch": 1.2408509012880358, "grad_norm": 1.2127948999404907, "learning_rate": 1.5038366044646168e-05, "loss": 3.7335, "step": 110450 }, { "epoch": 1.241412626458379, "grad_norm": 1.2872679233551025, "learning_rate": 1.503611913134332e-05, "loss": 3.6575, "step": 110500 }, { "epoch": 1.241974351628722, "grad_norm": 1.0627508163452148, "learning_rate": 1.5033872218040469e-05, "loss": 3.7264, "step": 110550 }, { "epoch": 1.2425360767990652, "grad_norm": 1.176890254020691, "learning_rate": 1.5031625304737618e-05, "loss": 3.6949, "step": 110600 }, { "epoch": 1.2430978019694083, "grad_norm": 1.2959498167037964, "learning_rate": 1.5029378391434767e-05, "loss": 3.7166, "step": 110650 }, { "epoch": 1.2436595271397517, "grad_norm": 1.1180050373077393, "learning_rate": 1.5027131478131918e-05, "loss": 3.7013, "step": 110700 }, { "epoch": 1.2442212523100948, "grad_norm": 1.2117652893066406, "learning_rate": 1.5024884564829068e-05, "loss": 3.6865, "step": 110750 }, { "epoch": 1.244782977480438, "grad_norm": 1.0430887937545776, "learning_rate": 1.5022637651526217e-05, "loss": 3.658, "step": 110800 }, { "epoch": 1.2453447026507811, "grad_norm": 1.1503196954727173, "learning_rate": 1.5020390738223368e-05, "loss": 3.7274, "step": 110850 }, { "epoch": 1.2459064278211243, "grad_norm": 1.1620378494262695, "learning_rate": 1.5018143824920517e-05, "loss": 3.6921, "step": 110900 }, { "epoch": 1.2464681529914674, "grad_norm": 1.1004583835601807, "learning_rate": 1.5015896911617667e-05, "loss": 3.6466, "step": 110950 }, { "epoch": 1.2470298781618105, "grad_norm": 1.1144731044769287, "learning_rate": 1.5013649998314816e-05, "loss": 3.6323, "step": 111000 }, { "epoch": 1.2475916033321537, "grad_norm": 1.1629180908203125, "learning_rate": 1.5011403085011967e-05, "loss": 3.5816, "step": 111050 }, { "epoch": 1.2481533285024968, "grad_norm": 1.1811234951019287, "learning_rate": 1.5009156171709116e-05, "loss": 3.7079, "step": 111100 }, { "epoch": 1.24871505367284, "grad_norm": 1.2378838062286377, "learning_rate": 1.5006909258406266e-05, "loss": 3.6832, "step": 111150 }, { "epoch": 1.2492767788431831, "grad_norm": 1.2465800046920776, "learning_rate": 1.5004662345103415e-05, "loss": 3.6852, "step": 111200 }, { "epoch": 1.2498385040135263, "grad_norm": 1.2350941896438599, "learning_rate": 1.5002415431800566e-05, "loss": 3.7366, "step": 111250 }, { "epoch": 1.2504002291838696, "grad_norm": 1.2513706684112549, "learning_rate": 1.5000213456763772e-05, "loss": 3.7345, "step": 111300 }, { "epoch": 1.2509619543542128, "grad_norm": 1.3040194511413574, "learning_rate": 1.4997966543460923e-05, "loss": 3.7018, "step": 111350 }, { "epoch": 1.251523679524556, "grad_norm": 1.1939553022384644, "learning_rate": 1.4995719630158072e-05, "loss": 3.7485, "step": 111400 }, { "epoch": 1.252085404694899, "grad_norm": 1.2655975818634033, "learning_rate": 1.4993472716855221e-05, "loss": 3.7011, "step": 111450 }, { "epoch": 1.2526471298652422, "grad_norm": 1.149431824684143, "learning_rate": 1.4991225803552372e-05, "loss": 3.6237, "step": 111500 }, { "epoch": 1.2532088550355853, "grad_norm": 1.2270399332046509, "learning_rate": 1.4988978890249522e-05, "loss": 3.6909, "step": 111550 }, { "epoch": 1.2537705802059285, "grad_norm": 1.2324990034103394, "learning_rate": 1.4986731976946671e-05, "loss": 3.7021, "step": 111600 }, { "epoch": 1.2543323053762716, "grad_norm": 1.0944278240203857, "learning_rate": 1.498448506364382e-05, "loss": 3.7399, "step": 111650 }, { "epoch": 1.2548940305466147, "grad_norm": 1.1698552370071411, "learning_rate": 1.4982238150340971e-05, "loss": 3.7134, "step": 111700 }, { "epoch": 1.2554557557169579, "grad_norm": 1.2568777799606323, "learning_rate": 1.497999123703812e-05, "loss": 3.6943, "step": 111750 }, { "epoch": 1.256017480887301, "grad_norm": 1.315553069114685, "learning_rate": 1.497774432373527e-05, "loss": 3.7308, "step": 111800 }, { "epoch": 1.2565792060576442, "grad_norm": 1.1367578506469727, "learning_rate": 1.4975497410432421e-05, "loss": 3.6847, "step": 111850 }, { "epoch": 1.2571409312279873, "grad_norm": 1.129918098449707, "learning_rate": 1.497325049712957e-05, "loss": 3.6771, "step": 111900 }, { "epoch": 1.2577026563983305, "grad_norm": 1.112339735031128, "learning_rate": 1.497100358382672e-05, "loss": 3.6944, "step": 111950 }, { "epoch": 1.2582643815686736, "grad_norm": 1.0623482465744019, "learning_rate": 1.4968756670523869e-05, "loss": 3.7297, "step": 112000 }, { "epoch": 1.2588261067390167, "grad_norm": 1.1499406099319458, "learning_rate": 1.496650975722102e-05, "loss": 3.6691, "step": 112050 }, { "epoch": 1.25938783190936, "grad_norm": 1.095597505569458, "learning_rate": 1.496426284391817e-05, "loss": 3.6609, "step": 112100 }, { "epoch": 1.2599495570797032, "grad_norm": 1.269573450088501, "learning_rate": 1.4962015930615319e-05, "loss": 3.659, "step": 112150 }, { "epoch": 1.2605112822500464, "grad_norm": 1.1126152276992798, "learning_rate": 1.4959769017312468e-05, "loss": 3.6726, "step": 112200 }, { "epoch": 1.2610730074203895, "grad_norm": 1.1701678037643433, "learning_rate": 1.4957522104009619e-05, "loss": 3.6442, "step": 112250 }, { "epoch": 1.2616347325907327, "grad_norm": 1.4589455127716064, "learning_rate": 1.4955275190706768e-05, "loss": 3.7182, "step": 112300 }, { "epoch": 1.2621964577610758, "grad_norm": 1.2170416116714478, "learning_rate": 1.4953028277403918e-05, "loss": 3.6852, "step": 112350 }, { "epoch": 1.262758182931419, "grad_norm": 0.9736093282699585, "learning_rate": 1.4950781364101069e-05, "loss": 3.6293, "step": 112400 }, { "epoch": 1.263319908101762, "grad_norm": 1.2244106531143188, "learning_rate": 1.4948534450798218e-05, "loss": 3.669, "step": 112450 }, { "epoch": 1.2638816332721052, "grad_norm": 1.2522999048233032, "learning_rate": 1.4946287537495367e-05, "loss": 3.7332, "step": 112500 }, { "epoch": 1.2644433584424484, "grad_norm": 1.381110429763794, "learning_rate": 1.4944040624192517e-05, "loss": 3.6822, "step": 112550 }, { "epoch": 1.2650050836127917, "grad_norm": 1.1141459941864014, "learning_rate": 1.4941793710889668e-05, "loss": 3.7492, "step": 112600 }, { "epoch": 1.2655668087831349, "grad_norm": 0.9814095497131348, "learning_rate": 1.4939546797586817e-05, "loss": 3.7012, "step": 112650 }, { "epoch": 1.266128533953478, "grad_norm": 1.2120487689971924, "learning_rate": 1.4937299884283966e-05, "loss": 3.6702, "step": 112700 }, { "epoch": 1.2666902591238212, "grad_norm": 1.2875162363052368, "learning_rate": 1.4935052970981117e-05, "loss": 3.6214, "step": 112750 }, { "epoch": 1.2672519842941643, "grad_norm": 1.1989301443099976, "learning_rate": 1.4932806057678267e-05, "loss": 3.6967, "step": 112800 }, { "epoch": 1.2678137094645074, "grad_norm": 1.0975011587142944, "learning_rate": 1.4930559144375416e-05, "loss": 3.71, "step": 112850 }, { "epoch": 1.2683754346348506, "grad_norm": 1.4478185176849365, "learning_rate": 1.4928312231072565e-05, "loss": 3.6802, "step": 112900 }, { "epoch": 1.2689371598051937, "grad_norm": 1.5076407194137573, "learning_rate": 1.4926065317769716e-05, "loss": 3.7218, "step": 112950 }, { "epoch": 1.2694988849755369, "grad_norm": 1.088319182395935, "learning_rate": 1.4923818404466866e-05, "loss": 3.7645, "step": 113000 }, { "epoch": 1.27006061014588, "grad_norm": 1.0324077606201172, "learning_rate": 1.4921571491164015e-05, "loss": 3.6841, "step": 113050 }, { "epoch": 1.2706223353162232, "grad_norm": 1.0946040153503418, "learning_rate": 1.4919324577861166e-05, "loss": 3.7037, "step": 113100 }, { "epoch": 1.2711840604865663, "grad_norm": 1.1757967472076416, "learning_rate": 1.4917077664558315e-05, "loss": 3.8264, "step": 113150 }, { "epoch": 1.2717457856569094, "grad_norm": 1.2199627161026, "learning_rate": 1.4914830751255464e-05, "loss": 3.7236, "step": 113200 }, { "epoch": 1.2723075108272526, "grad_norm": 1.1769957542419434, "learning_rate": 1.4912583837952614e-05, "loss": 3.7198, "step": 113250 }, { "epoch": 1.2728692359975957, "grad_norm": 1.118649959564209, "learning_rate": 1.4910336924649765e-05, "loss": 3.7062, "step": 113300 }, { "epoch": 1.2734309611679389, "grad_norm": 1.1054401397705078, "learning_rate": 1.4908090011346914e-05, "loss": 3.7428, "step": 113350 }, { "epoch": 1.273992686338282, "grad_norm": 1.1560803651809692, "learning_rate": 1.4905843098044063e-05, "loss": 3.7675, "step": 113400 }, { "epoch": 1.2745544115086254, "grad_norm": 1.1944975852966309, "learning_rate": 1.4903596184741214e-05, "loss": 3.693, "step": 113450 }, { "epoch": 1.2751161366789685, "grad_norm": 1.030148983001709, "learning_rate": 1.4901349271438364e-05, "loss": 3.713, "step": 113500 }, { "epoch": 1.2756778618493116, "grad_norm": 1.2132092714309692, "learning_rate": 1.4899102358135513e-05, "loss": 3.6853, "step": 113550 }, { "epoch": 1.2762395870196548, "grad_norm": 1.1267964839935303, "learning_rate": 1.4896855444832662e-05, "loss": 3.8054, "step": 113600 }, { "epoch": 1.276801312189998, "grad_norm": 1.0857360363006592, "learning_rate": 1.4894608531529813e-05, "loss": 3.7694, "step": 113650 }, { "epoch": 1.277363037360341, "grad_norm": 1.1156096458435059, "learning_rate": 1.4892361618226963e-05, "loss": 3.6494, "step": 113700 }, { "epoch": 1.2779247625306842, "grad_norm": 1.1470805406570435, "learning_rate": 1.4890114704924112e-05, "loss": 3.7046, "step": 113750 }, { "epoch": 1.2784864877010274, "grad_norm": 1.1375935077667236, "learning_rate": 1.4887867791621261e-05, "loss": 3.655, "step": 113800 }, { "epoch": 1.2790482128713705, "grad_norm": 1.490735411643982, "learning_rate": 1.4885620878318412e-05, "loss": 3.7503, "step": 113850 }, { "epoch": 1.2796099380417136, "grad_norm": 1.081573486328125, "learning_rate": 1.4883373965015562e-05, "loss": 3.6072, "step": 113900 }, { "epoch": 1.280171663212057, "grad_norm": 1.0996750593185425, "learning_rate": 1.4881127051712711e-05, "loss": 3.6634, "step": 113950 }, { "epoch": 1.2807333883824001, "grad_norm": 1.2041651010513306, "learning_rate": 1.4878880138409862e-05, "loss": 3.6514, "step": 114000 }, { "epoch": 1.2812951135527433, "grad_norm": 1.0212005376815796, "learning_rate": 1.4876633225107011e-05, "loss": 3.7077, "step": 114050 }, { "epoch": 1.2818568387230864, "grad_norm": 1.223238229751587, "learning_rate": 1.487438631180416e-05, "loss": 3.6684, "step": 114100 }, { "epoch": 1.2824185638934296, "grad_norm": 1.146500587463379, "learning_rate": 1.487213939850131e-05, "loss": 3.6591, "step": 114150 }, { "epoch": 1.2829802890637727, "grad_norm": 1.1014022827148438, "learning_rate": 1.4869892485198461e-05, "loss": 3.7355, "step": 114200 }, { "epoch": 1.2835420142341158, "grad_norm": 1.3315798044204712, "learning_rate": 1.486764557189561e-05, "loss": 3.7325, "step": 114250 }, { "epoch": 1.284103739404459, "grad_norm": 1.263845682144165, "learning_rate": 1.486539865859276e-05, "loss": 3.6997, "step": 114300 }, { "epoch": 1.2846654645748021, "grad_norm": 1.3485934734344482, "learning_rate": 1.4863196683555967e-05, "loss": 3.7422, "step": 114350 }, { "epoch": 1.2852271897451453, "grad_norm": 1.086905598640442, "learning_rate": 1.4860949770253117e-05, "loss": 3.7186, "step": 114400 }, { "epoch": 1.2857889149154884, "grad_norm": 1.1574716567993164, "learning_rate": 1.4858702856950268e-05, "loss": 3.698, "step": 114450 }, { "epoch": 1.2863506400858316, "grad_norm": 5.0376482009887695, "learning_rate": 1.4856455943647417e-05, "loss": 3.6421, "step": 114500 }, { "epoch": 1.2869123652561747, "grad_norm": 1.0967539548873901, "learning_rate": 1.4854209030344566e-05, "loss": 3.7413, "step": 114550 }, { "epoch": 1.2874740904265178, "grad_norm": 1.210237741470337, "learning_rate": 1.4851962117041715e-05, "loss": 3.7403, "step": 114600 }, { "epoch": 1.288035815596861, "grad_norm": 1.0677835941314697, "learning_rate": 1.4849715203738867e-05, "loss": 3.7031, "step": 114650 }, { "epoch": 1.2885975407672041, "grad_norm": 1.0944105386734009, "learning_rate": 1.4847468290436016e-05, "loss": 3.6474, "step": 114700 }, { "epoch": 1.2891592659375473, "grad_norm": 1.1223304271697998, "learning_rate": 1.4845221377133165e-05, "loss": 3.6784, "step": 114750 }, { "epoch": 1.2897209911078906, "grad_norm": 1.1143394708633423, "learning_rate": 1.4842974463830314e-05, "loss": 3.7114, "step": 114800 }, { "epoch": 1.2902827162782338, "grad_norm": 1.536241888999939, "learning_rate": 1.4840727550527465e-05, "loss": 3.6856, "step": 114850 }, { "epoch": 1.290844441448577, "grad_norm": 0.9196121096611023, "learning_rate": 1.4838480637224615e-05, "loss": 3.7058, "step": 114900 }, { "epoch": 1.29140616661892, "grad_norm": 1.2177156209945679, "learning_rate": 1.4836233723921764e-05, "loss": 3.6414, "step": 114950 }, { "epoch": 1.2919678917892632, "grad_norm": 1.143907904624939, "learning_rate": 1.4833986810618915e-05, "loss": 3.7765, "step": 115000 }, { "epoch": 1.2925296169596063, "grad_norm": 1.2396646738052368, "learning_rate": 1.4831739897316064e-05, "loss": 3.7571, "step": 115050 }, { "epoch": 1.2930913421299495, "grad_norm": 1.1673306226730347, "learning_rate": 1.4829492984013214e-05, "loss": 3.704, "step": 115100 }, { "epoch": 1.2936530673002926, "grad_norm": 1.3346099853515625, "learning_rate": 1.4827246070710363e-05, "loss": 3.66, "step": 115150 }, { "epoch": 1.2942147924706358, "grad_norm": 1.3197540044784546, "learning_rate": 1.4824999157407514e-05, "loss": 3.6604, "step": 115200 }, { "epoch": 1.294776517640979, "grad_norm": 1.0987499952316284, "learning_rate": 1.4822752244104663e-05, "loss": 3.6903, "step": 115250 }, { "epoch": 1.2953382428113223, "grad_norm": 1.1738516092300415, "learning_rate": 1.4820505330801813e-05, "loss": 3.7153, "step": 115300 }, { "epoch": 1.2958999679816654, "grad_norm": 1.0726885795593262, "learning_rate": 1.4818258417498964e-05, "loss": 3.7386, "step": 115350 }, { "epoch": 1.2964616931520085, "grad_norm": 1.132994294166565, "learning_rate": 1.4816011504196113e-05, "loss": 3.5907, "step": 115400 }, { "epoch": 1.2970234183223517, "grad_norm": 0.9900727868080139, "learning_rate": 1.4813764590893262e-05, "loss": 3.7105, "step": 115450 }, { "epoch": 1.2975851434926948, "grad_norm": 1.0517420768737793, "learning_rate": 1.4811517677590412e-05, "loss": 3.7059, "step": 115500 }, { "epoch": 1.298146868663038, "grad_norm": 1.2015070915222168, "learning_rate": 1.4809270764287563e-05, "loss": 3.7225, "step": 115550 }, { "epoch": 1.298708593833381, "grad_norm": 1.1541849374771118, "learning_rate": 1.4807023850984712e-05, "loss": 3.6733, "step": 115600 }, { "epoch": 1.2992703190037242, "grad_norm": 1.1331740617752075, "learning_rate": 1.4804776937681861e-05, "loss": 3.66, "step": 115650 }, { "epoch": 1.2998320441740674, "grad_norm": 1.0337045192718506, "learning_rate": 1.4802530024379012e-05, "loss": 3.7144, "step": 115700 }, { "epoch": 1.3003937693444105, "grad_norm": 1.3465909957885742, "learning_rate": 1.4800283111076162e-05, "loss": 3.7363, "step": 115750 }, { "epoch": 1.3009554945147537, "grad_norm": 1.2244303226470947, "learning_rate": 1.4798036197773311e-05, "loss": 3.7488, "step": 115800 }, { "epoch": 1.3015172196850968, "grad_norm": 1.0829503536224365, "learning_rate": 1.479578928447046e-05, "loss": 3.7574, "step": 115850 }, { "epoch": 1.30207894485544, "grad_norm": 1.0618500709533691, "learning_rate": 1.4793542371167611e-05, "loss": 3.7326, "step": 115900 }, { "epoch": 1.302640670025783, "grad_norm": 1.1596113443374634, "learning_rate": 1.479129545786476e-05, "loss": 3.6051, "step": 115950 }, { "epoch": 1.3032023951961262, "grad_norm": 1.2710692882537842, "learning_rate": 1.478904854456191e-05, "loss": 3.6918, "step": 116000 }, { "epoch": 1.3037641203664694, "grad_norm": 1.3042854070663452, "learning_rate": 1.4786801631259061e-05, "loss": 3.6836, "step": 116050 }, { "epoch": 1.3043258455368125, "grad_norm": 1.170735478401184, "learning_rate": 1.478455471795621e-05, "loss": 3.7783, "step": 116100 }, { "epoch": 1.3048875707071559, "grad_norm": 1.127317190170288, "learning_rate": 1.478230780465336e-05, "loss": 3.7091, "step": 116150 }, { "epoch": 1.305449295877499, "grad_norm": 1.0851610898971558, "learning_rate": 1.4780060891350509e-05, "loss": 3.7688, "step": 116200 }, { "epoch": 1.3060110210478422, "grad_norm": 1.5472756624221802, "learning_rate": 1.477781397804766e-05, "loss": 3.6622, "step": 116250 }, { "epoch": 1.3065727462181853, "grad_norm": 1.0288634300231934, "learning_rate": 1.477556706474481e-05, "loss": 3.7608, "step": 116300 }, { "epoch": 1.3071344713885285, "grad_norm": 1.2104113101959229, "learning_rate": 1.4773320151441959e-05, "loss": 3.7222, "step": 116350 }, { "epoch": 1.3076961965588716, "grad_norm": 1.206800103187561, "learning_rate": 1.4771073238139108e-05, "loss": 3.672, "step": 116400 }, { "epoch": 1.3082579217292147, "grad_norm": 1.1641842126846313, "learning_rate": 1.4768826324836259e-05, "loss": 3.7283, "step": 116450 }, { "epoch": 1.3088196468995579, "grad_norm": 1.1273703575134277, "learning_rate": 1.4766579411533408e-05, "loss": 3.716, "step": 116500 }, { "epoch": 1.309381372069901, "grad_norm": 1.2490730285644531, "learning_rate": 1.4764332498230557e-05, "loss": 3.735, "step": 116550 }, { "epoch": 1.3099430972402442, "grad_norm": 1.1901054382324219, "learning_rate": 1.4762085584927708e-05, "loss": 3.6776, "step": 116600 }, { "epoch": 1.3105048224105875, "grad_norm": 1.13470458984375, "learning_rate": 1.4759838671624858e-05, "loss": 3.6582, "step": 116650 }, { "epoch": 1.3110665475809307, "grad_norm": 1.322670817375183, "learning_rate": 1.4757591758322007e-05, "loss": 3.7346, "step": 116700 }, { "epoch": 1.3116282727512738, "grad_norm": 0.9900648593902588, "learning_rate": 1.4755344845019156e-05, "loss": 3.6795, "step": 116750 }, { "epoch": 1.312189997921617, "grad_norm": 1.1879056692123413, "learning_rate": 1.4753097931716307e-05, "loss": 3.7087, "step": 116800 }, { "epoch": 1.31275172309196, "grad_norm": 0.9856444597244263, "learning_rate": 1.4750851018413457e-05, "loss": 3.7151, "step": 116850 }, { "epoch": 1.3133134482623032, "grad_norm": 1.1498546600341797, "learning_rate": 1.4748604105110606e-05, "loss": 3.6927, "step": 116900 }, { "epoch": 1.3138751734326464, "grad_norm": 1.1568413972854614, "learning_rate": 1.4746357191807757e-05, "loss": 3.6835, "step": 116950 }, { "epoch": 1.3144368986029895, "grad_norm": 1.1762259006500244, "learning_rate": 1.4744110278504906e-05, "loss": 3.7163, "step": 117000 }, { "epoch": 1.3149986237733327, "grad_norm": 1.0447081327438354, "learning_rate": 1.4741863365202056e-05, "loss": 3.6836, "step": 117050 }, { "epoch": 1.3155603489436758, "grad_norm": 1.3564331531524658, "learning_rate": 1.4739616451899205e-05, "loss": 3.7426, "step": 117100 }, { "epoch": 1.316122074114019, "grad_norm": 1.373300313949585, "learning_rate": 1.4737369538596356e-05, "loss": 3.7362, "step": 117150 }, { "epoch": 1.316683799284362, "grad_norm": 1.2786941528320312, "learning_rate": 1.4735122625293505e-05, "loss": 3.6725, "step": 117200 }, { "epoch": 1.3172455244547052, "grad_norm": 1.2043522596359253, "learning_rate": 1.4732875711990655e-05, "loss": 3.6457, "step": 117250 }, { "epoch": 1.3178072496250484, "grad_norm": 1.0595104694366455, "learning_rate": 1.4730628798687806e-05, "loss": 3.6185, "step": 117300 }, { "epoch": 1.3183689747953915, "grad_norm": 1.272301435470581, "learning_rate": 1.4728381885384955e-05, "loss": 3.7125, "step": 117350 }, { "epoch": 1.3189306999657346, "grad_norm": 1.266560435295105, "learning_rate": 1.4726134972082104e-05, "loss": 3.7035, "step": 117400 }, { "epoch": 1.3194924251360778, "grad_norm": 1.0973302125930786, "learning_rate": 1.4723888058779254e-05, "loss": 3.7357, "step": 117450 }, { "epoch": 1.3200541503064211, "grad_norm": 1.0538254976272583, "learning_rate": 1.4721641145476405e-05, "loss": 3.7578, "step": 117500 }, { "epoch": 1.3206158754767643, "grad_norm": 1.1436649560928345, "learning_rate": 1.4719394232173552e-05, "loss": 3.6825, "step": 117550 }, { "epoch": 1.3211776006471074, "grad_norm": 1.0678046941757202, "learning_rate": 1.4717147318870702e-05, "loss": 3.7167, "step": 117600 }, { "epoch": 1.3217393258174506, "grad_norm": 1.199366569519043, "learning_rate": 1.4714900405567851e-05, "loss": 3.7284, "step": 117650 }, { "epoch": 1.3223010509877937, "grad_norm": 1.2246750593185425, "learning_rate": 1.4712653492265e-05, "loss": 3.7814, "step": 117700 }, { "epoch": 1.3228627761581369, "grad_norm": 1.0405752658843994, "learning_rate": 1.4710406578962151e-05, "loss": 3.7945, "step": 117750 }, { "epoch": 1.32342450132848, "grad_norm": 1.1649138927459717, "learning_rate": 1.47081596656593e-05, "loss": 3.7168, "step": 117800 }, { "epoch": 1.3239862264988231, "grad_norm": 1.1335917711257935, "learning_rate": 1.470591275235645e-05, "loss": 3.6777, "step": 117850 }, { "epoch": 1.3245479516691663, "grad_norm": 1.229051113128662, "learning_rate": 1.4703665839053601e-05, "loss": 3.6618, "step": 117900 }, { "epoch": 1.3251096768395094, "grad_norm": 1.164178729057312, "learning_rate": 1.470141892575075e-05, "loss": 3.6298, "step": 117950 }, { "epoch": 1.3256714020098528, "grad_norm": 1.0775868892669678, "learning_rate": 1.46991720124479e-05, "loss": 3.7261, "step": 118000 }, { "epoch": 1.326233127180196, "grad_norm": 1.1860066652297974, "learning_rate": 1.4696925099145049e-05, "loss": 3.7335, "step": 118050 }, { "epoch": 1.326794852350539, "grad_norm": 1.2197784185409546, "learning_rate": 1.46946781858422e-05, "loss": 3.7328, "step": 118100 }, { "epoch": 1.3273565775208822, "grad_norm": 1.1759244203567505, "learning_rate": 1.469243127253935e-05, "loss": 3.6612, "step": 118150 }, { "epoch": 1.3279183026912253, "grad_norm": 1.2106807231903076, "learning_rate": 1.4690184359236498e-05, "loss": 3.6147, "step": 118200 }, { "epoch": 1.3284800278615685, "grad_norm": 1.2478992938995361, "learning_rate": 1.468793744593365e-05, "loss": 3.6817, "step": 118250 }, { "epoch": 1.3290417530319116, "grad_norm": 1.6169747114181519, "learning_rate": 1.4685690532630799e-05, "loss": 3.6878, "step": 118300 }, { "epoch": 1.3296034782022548, "grad_norm": 1.3529175519943237, "learning_rate": 1.4683488557594006e-05, "loss": 3.8043, "step": 118350 }, { "epoch": 1.330165203372598, "grad_norm": 0.9039379358291626, "learning_rate": 1.4681241644291156e-05, "loss": 3.7257, "step": 118400 }, { "epoch": 1.330726928542941, "grad_norm": 1.121334433555603, "learning_rate": 1.4678994730988305e-05, "loss": 3.7019, "step": 118450 }, { "epoch": 1.3312886537132842, "grad_norm": 1.0345427989959717, "learning_rate": 1.4676747817685454e-05, "loss": 3.6703, "step": 118500 }, { "epoch": 1.3318503788836273, "grad_norm": 1.2660282850265503, "learning_rate": 1.4674500904382605e-05, "loss": 3.6788, "step": 118550 }, { "epoch": 1.3324121040539705, "grad_norm": 1.0917772054672241, "learning_rate": 1.4672253991079755e-05, "loss": 3.68, "step": 118600 }, { "epoch": 1.3329738292243136, "grad_norm": 1.1843641996383667, "learning_rate": 1.4670007077776904e-05, "loss": 3.6719, "step": 118650 }, { "epoch": 1.3335355543946568, "grad_norm": 1.2436368465423584, "learning_rate": 1.4667760164474053e-05, "loss": 3.7044, "step": 118700 }, { "epoch": 1.334097279565, "grad_norm": 1.1449768543243408, "learning_rate": 1.4665513251171204e-05, "loss": 3.7275, "step": 118750 }, { "epoch": 1.3346590047353433, "grad_norm": 1.2294001579284668, "learning_rate": 1.4663266337868354e-05, "loss": 3.6284, "step": 118800 }, { "epoch": 1.3352207299056864, "grad_norm": 1.0378721952438354, "learning_rate": 1.4661019424565503e-05, "loss": 3.7391, "step": 118850 }, { "epoch": 1.3357824550760296, "grad_norm": 0.9750446081161499, "learning_rate": 1.4658772511262654e-05, "loss": 3.7817, "step": 118900 }, { "epoch": 1.3363441802463727, "grad_norm": 1.2073137760162354, "learning_rate": 1.4656525597959803e-05, "loss": 3.7439, "step": 118950 }, { "epoch": 1.3369059054167158, "grad_norm": 1.1098533868789673, "learning_rate": 1.4654278684656953e-05, "loss": 3.7667, "step": 119000 }, { "epoch": 1.337467630587059, "grad_norm": 1.1253433227539062, "learning_rate": 1.4652031771354102e-05, "loss": 3.673, "step": 119050 }, { "epoch": 1.3380293557574021, "grad_norm": 1.296921968460083, "learning_rate": 1.4649784858051253e-05, "loss": 3.7251, "step": 119100 }, { "epoch": 1.3385910809277453, "grad_norm": 0.9098636507987976, "learning_rate": 1.4647537944748402e-05, "loss": 3.6904, "step": 119150 }, { "epoch": 1.3391528060980884, "grad_norm": 1.0598982572555542, "learning_rate": 1.4645291031445552e-05, "loss": 3.65, "step": 119200 }, { "epoch": 1.3397145312684315, "grad_norm": 0.9995057582855225, "learning_rate": 1.4643044118142703e-05, "loss": 3.6988, "step": 119250 }, { "epoch": 1.3402762564387747, "grad_norm": 1.265258550643921, "learning_rate": 1.4640797204839852e-05, "loss": 3.6759, "step": 119300 }, { "epoch": 1.340837981609118, "grad_norm": 1.3248783349990845, "learning_rate": 1.4638550291537001e-05, "loss": 3.5958, "step": 119350 }, { "epoch": 1.3413997067794612, "grad_norm": 1.2543925046920776, "learning_rate": 1.463630337823415e-05, "loss": 3.6777, "step": 119400 }, { "epoch": 1.3419614319498043, "grad_norm": 1.18267023563385, "learning_rate": 1.4634056464931302e-05, "loss": 3.6751, "step": 119450 }, { "epoch": 1.3425231571201475, "grad_norm": 1.44830322265625, "learning_rate": 1.4631809551628451e-05, "loss": 3.6802, "step": 119500 }, { "epoch": 1.3430848822904906, "grad_norm": 1.311497449874878, "learning_rate": 1.46295626383256e-05, "loss": 3.7528, "step": 119550 }, { "epoch": 1.3436466074608338, "grad_norm": 1.0765646696090698, "learning_rate": 1.4627315725022751e-05, "loss": 3.7024, "step": 119600 }, { "epoch": 1.344208332631177, "grad_norm": 1.1783418655395508, "learning_rate": 1.46250688117199e-05, "loss": 3.7267, "step": 119650 }, { "epoch": 1.34477005780152, "grad_norm": 1.1875866651535034, "learning_rate": 1.462282189841705e-05, "loss": 3.6513, "step": 119700 }, { "epoch": 1.3453317829718632, "grad_norm": 1.1184399127960205, "learning_rate": 1.4620574985114199e-05, "loss": 3.75, "step": 119750 }, { "epoch": 1.3458935081422063, "grad_norm": 1.2593485116958618, "learning_rate": 1.461832807181135e-05, "loss": 3.6672, "step": 119800 }, { "epoch": 1.3464552333125495, "grad_norm": 1.0915449857711792, "learning_rate": 1.46160811585085e-05, "loss": 3.656, "step": 119850 }, { "epoch": 1.3470169584828926, "grad_norm": 1.3104792833328247, "learning_rate": 1.4613834245205649e-05, "loss": 3.686, "step": 119900 }, { "epoch": 1.3475786836532357, "grad_norm": 1.1535950899124146, "learning_rate": 1.46115873319028e-05, "loss": 3.7439, "step": 119950 }, { "epoch": 1.3481404088235789, "grad_norm": 1.2658485174179077, "learning_rate": 1.4609340418599949e-05, "loss": 3.7157, "step": 120000 }, { "epoch": 1.348702133993922, "grad_norm": 1.2296286821365356, "learning_rate": 1.4607093505297098e-05, "loss": 3.6688, "step": 120050 }, { "epoch": 1.3492638591642652, "grad_norm": 1.166917085647583, "learning_rate": 1.4604846591994248e-05, "loss": 3.6548, "step": 120100 }, { "epoch": 1.3498255843346085, "grad_norm": 1.2615032196044922, "learning_rate": 1.4602599678691399e-05, "loss": 3.6502, "step": 120150 }, { "epoch": 1.3503873095049517, "grad_norm": 1.0976001024246216, "learning_rate": 1.4600352765388548e-05, "loss": 3.6865, "step": 120200 }, { "epoch": 1.3509490346752948, "grad_norm": 1.0847792625427246, "learning_rate": 1.4598105852085697e-05, "loss": 3.677, "step": 120250 }, { "epoch": 1.351510759845638, "grad_norm": 1.1527408361434937, "learning_rate": 1.4595858938782847e-05, "loss": 3.6671, "step": 120300 }, { "epoch": 1.352072485015981, "grad_norm": 1.0550224781036377, "learning_rate": 1.4593612025479998e-05, "loss": 3.6829, "step": 120350 }, { "epoch": 1.3526342101863242, "grad_norm": 1.3266491889953613, "learning_rate": 1.4591365112177147e-05, "loss": 3.6975, "step": 120400 }, { "epoch": 1.3531959353566674, "grad_norm": 1.0050033330917358, "learning_rate": 1.4589118198874296e-05, "loss": 3.6553, "step": 120450 }, { "epoch": 1.3537576605270105, "grad_norm": 1.0242691040039062, "learning_rate": 1.4586916223837504e-05, "loss": 3.7015, "step": 120500 }, { "epoch": 1.3543193856973537, "grad_norm": 1.1202960014343262, "learning_rate": 1.4584669310534653e-05, "loss": 3.6946, "step": 120550 }, { "epoch": 1.3548811108676968, "grad_norm": 0.9728813767433167, "learning_rate": 1.4582422397231804e-05, "loss": 3.6664, "step": 120600 }, { "epoch": 1.3554428360380402, "grad_norm": 1.2763217687606812, "learning_rate": 1.4580175483928954e-05, "loss": 3.7196, "step": 120650 }, { "epoch": 1.3560045612083833, "grad_norm": 0.9913472533226013, "learning_rate": 1.4577928570626103e-05, "loss": 3.68, "step": 120700 }, { "epoch": 1.3565662863787264, "grad_norm": 1.1339112520217896, "learning_rate": 1.4575681657323252e-05, "loss": 3.6267, "step": 120750 }, { "epoch": 1.3571280115490696, "grad_norm": 1.1435879468917847, "learning_rate": 1.4573434744020403e-05, "loss": 3.6202, "step": 120800 }, { "epoch": 1.3576897367194127, "grad_norm": 1.1710400581359863, "learning_rate": 1.4571187830717553e-05, "loss": 3.7836, "step": 120850 }, { "epoch": 1.3582514618897559, "grad_norm": 1.1049959659576416, "learning_rate": 1.4568940917414702e-05, "loss": 3.6587, "step": 120900 }, { "epoch": 1.358813187060099, "grad_norm": 0.995599091053009, "learning_rate": 1.4566694004111853e-05, "loss": 3.6778, "step": 120950 }, { "epoch": 1.3593749122304422, "grad_norm": 1.1417511701583862, "learning_rate": 1.4564536967341117e-05, "loss": 3.7144, "step": 121000 }, { "epoch": 1.3599366374007853, "grad_norm": 1.1375656127929688, "learning_rate": 1.4562290054038266e-05, "loss": 3.7073, "step": 121050 }, { "epoch": 1.3604983625711284, "grad_norm": 1.0259642601013184, "learning_rate": 1.4560043140735416e-05, "loss": 3.6656, "step": 121100 }, { "epoch": 1.3610600877414716, "grad_norm": 1.1820317506790161, "learning_rate": 1.4557796227432565e-05, "loss": 3.6812, "step": 121150 }, { "epoch": 1.3616218129118147, "grad_norm": 1.1481682062149048, "learning_rate": 1.4555549314129716e-05, "loss": 3.7106, "step": 121200 }, { "epoch": 1.3621835380821579, "grad_norm": 1.3173041343688965, "learning_rate": 1.4553302400826865e-05, "loss": 3.6936, "step": 121250 }, { "epoch": 1.362745263252501, "grad_norm": 1.0805565118789673, "learning_rate": 1.4551055487524015e-05, "loss": 3.6706, "step": 121300 }, { "epoch": 1.3633069884228441, "grad_norm": 1.17776620388031, "learning_rate": 1.4548808574221166e-05, "loss": 3.7006, "step": 121350 }, { "epoch": 1.3638687135931873, "grad_norm": 1.0657659769058228, "learning_rate": 1.4546561660918315e-05, "loss": 3.748, "step": 121400 }, { "epoch": 1.3644304387635304, "grad_norm": 1.0460538864135742, "learning_rate": 1.4544314747615464e-05, "loss": 3.704, "step": 121450 }, { "epoch": 1.3649921639338738, "grad_norm": 1.2596518993377686, "learning_rate": 1.4542067834312614e-05, "loss": 3.6839, "step": 121500 }, { "epoch": 1.365553889104217, "grad_norm": 1.397570252418518, "learning_rate": 1.4539820921009765e-05, "loss": 3.7181, "step": 121550 }, { "epoch": 1.36611561427456, "grad_norm": 1.280348300933838, "learning_rate": 1.4537574007706914e-05, "loss": 3.7099, "step": 121600 }, { "epoch": 1.3666773394449032, "grad_norm": 1.201006531715393, "learning_rate": 1.4535327094404063e-05, "loss": 3.7533, "step": 121650 }, { "epoch": 1.3672390646152464, "grad_norm": 1.2775695323944092, "learning_rate": 1.4533080181101213e-05, "loss": 3.6657, "step": 121700 }, { "epoch": 1.3678007897855895, "grad_norm": 1.1215566396713257, "learning_rate": 1.4530833267798364e-05, "loss": 3.6661, "step": 121750 }, { "epoch": 1.3683625149559326, "grad_norm": 1.0805236101150513, "learning_rate": 1.4528586354495513e-05, "loss": 3.7707, "step": 121800 }, { "epoch": 1.3689242401262758, "grad_norm": 1.1486221551895142, "learning_rate": 1.4526339441192662e-05, "loss": 3.7298, "step": 121850 }, { "epoch": 1.369485965296619, "grad_norm": 1.1970341205596924, "learning_rate": 1.4524092527889813e-05, "loss": 3.7511, "step": 121900 }, { "epoch": 1.370047690466962, "grad_norm": 1.014717936515808, "learning_rate": 1.4521845614586963e-05, "loss": 3.6929, "step": 121950 }, { "epoch": 1.3706094156373054, "grad_norm": 1.0722380876541138, "learning_rate": 1.4519598701284112e-05, "loss": 3.6561, "step": 122000 }, { "epoch": 1.3711711408076486, "grad_norm": 1.07088041305542, "learning_rate": 1.4517351787981261e-05, "loss": 3.8093, "step": 122050 }, { "epoch": 1.3717328659779917, "grad_norm": 1.2771564722061157, "learning_rate": 1.4515104874678412e-05, "loss": 3.67, "step": 122100 }, { "epoch": 1.3722945911483349, "grad_norm": 1.1219714879989624, "learning_rate": 1.4512857961375562e-05, "loss": 3.6666, "step": 122150 }, { "epoch": 1.372856316318678, "grad_norm": 1.1300429105758667, "learning_rate": 1.4510611048072711e-05, "loss": 3.6496, "step": 122200 }, { "epoch": 1.3734180414890211, "grad_norm": 2.0728774070739746, "learning_rate": 1.4508364134769862e-05, "loss": 3.775, "step": 122250 }, { "epoch": 1.3739797666593643, "grad_norm": 1.1185555458068848, "learning_rate": 1.4506117221467011e-05, "loss": 3.629, "step": 122300 }, { "epoch": 1.3745414918297074, "grad_norm": 1.1981322765350342, "learning_rate": 1.450387030816416e-05, "loss": 3.6022, "step": 122350 }, { "epoch": 1.3751032170000506, "grad_norm": 1.2505836486816406, "learning_rate": 1.450162339486131e-05, "loss": 3.671, "step": 122400 }, { "epoch": 1.3756649421703937, "grad_norm": 1.1908420324325562, "learning_rate": 1.4499421419824517e-05, "loss": 3.6314, "step": 122450 }, { "epoch": 1.3762266673407368, "grad_norm": 1.2761627435684204, "learning_rate": 1.4497174506521667e-05, "loss": 3.7494, "step": 122500 }, { "epoch": 1.37678839251108, "grad_norm": 1.2110997438430786, "learning_rate": 1.4494927593218818e-05, "loss": 3.6841, "step": 122550 }, { "epoch": 1.3773501176814231, "grad_norm": 1.1500811576843262, "learning_rate": 1.4492680679915967e-05, "loss": 3.6617, "step": 122600 }, { "epoch": 1.3779118428517663, "grad_norm": 1.080535888671875, "learning_rate": 1.4490433766613116e-05, "loss": 3.7118, "step": 122650 }, { "epoch": 1.3784735680221094, "grad_norm": 1.1847634315490723, "learning_rate": 1.4488186853310266e-05, "loss": 3.6507, "step": 122700 }, { "epoch": 1.3790352931924525, "grad_norm": 1.1334607601165771, "learning_rate": 1.4485939940007417e-05, "loss": 3.6664, "step": 122750 }, { "epoch": 1.3795970183627957, "grad_norm": 1.0726972818374634, "learning_rate": 1.4483693026704566e-05, "loss": 3.7032, "step": 122800 }, { "epoch": 1.380158743533139, "grad_norm": 1.2592869997024536, "learning_rate": 1.4481446113401715e-05, "loss": 3.6406, "step": 122850 }, { "epoch": 1.3807204687034822, "grad_norm": 1.1625162363052368, "learning_rate": 1.4479199200098866e-05, "loss": 3.8139, "step": 122900 }, { "epoch": 1.3812821938738253, "grad_norm": 1.1738134622573853, "learning_rate": 1.4476952286796016e-05, "loss": 3.7118, "step": 122950 }, { "epoch": 1.3818439190441685, "grad_norm": 0.9934506416320801, "learning_rate": 1.4474705373493165e-05, "loss": 3.7588, "step": 123000 }, { "epoch": 1.3824056442145116, "grad_norm": 1.1412187814712524, "learning_rate": 1.4472458460190314e-05, "loss": 3.7243, "step": 123050 }, { "epoch": 1.3829673693848548, "grad_norm": 1.0966845750808716, "learning_rate": 1.4470211546887465e-05, "loss": 3.6771, "step": 123100 }, { "epoch": 1.383529094555198, "grad_norm": 1.2062981128692627, "learning_rate": 1.4467964633584615e-05, "loss": 3.7103, "step": 123150 }, { "epoch": 1.384090819725541, "grad_norm": 0.9867963194847107, "learning_rate": 1.4465717720281764e-05, "loss": 3.7591, "step": 123200 }, { "epoch": 1.3846525448958842, "grad_norm": 1.0068353414535522, "learning_rate": 1.4463470806978915e-05, "loss": 3.6298, "step": 123250 }, { "epoch": 1.3852142700662273, "grad_norm": 1.4179811477661133, "learning_rate": 1.4461223893676064e-05, "loss": 3.6966, "step": 123300 }, { "epoch": 1.3857759952365707, "grad_norm": 1.0487943887710571, "learning_rate": 1.4458976980373214e-05, "loss": 3.736, "step": 123350 }, { "epoch": 1.3863377204069138, "grad_norm": 1.301311731338501, "learning_rate": 1.4456730067070363e-05, "loss": 3.6252, "step": 123400 }, { "epoch": 1.386899445577257, "grad_norm": 1.1093934774398804, "learning_rate": 1.4454483153767514e-05, "loss": 3.7113, "step": 123450 }, { "epoch": 1.3874611707476001, "grad_norm": 1.0409111976623535, "learning_rate": 1.4452236240464663e-05, "loss": 3.6539, "step": 123500 }, { "epoch": 1.3880228959179433, "grad_norm": 1.1051313877105713, "learning_rate": 1.4449989327161813e-05, "loss": 3.7126, "step": 123550 }, { "epoch": 1.3885846210882864, "grad_norm": 1.1085457801818848, "learning_rate": 1.4447742413858964e-05, "loss": 3.692, "step": 123600 }, { "epoch": 1.3891463462586295, "grad_norm": 1.0587685108184814, "learning_rate": 1.4445495500556113e-05, "loss": 3.7159, "step": 123650 }, { "epoch": 1.3897080714289727, "grad_norm": 1.2765042781829834, "learning_rate": 1.4443248587253262e-05, "loss": 3.6629, "step": 123700 }, { "epoch": 1.3902697965993158, "grad_norm": 1.0963952541351318, "learning_rate": 1.4441001673950412e-05, "loss": 3.6346, "step": 123750 }, { "epoch": 1.390831521769659, "grad_norm": 1.1037063598632812, "learning_rate": 1.4438754760647563e-05, "loss": 3.6834, "step": 123800 }, { "epoch": 1.391393246940002, "grad_norm": 1.283635139465332, "learning_rate": 1.4436507847344712e-05, "loss": 3.6974, "step": 123850 }, { "epoch": 1.3919549721103452, "grad_norm": 1.2471383810043335, "learning_rate": 1.4434260934041861e-05, "loss": 3.7738, "step": 123900 }, { "epoch": 1.3925166972806884, "grad_norm": 1.1486215591430664, "learning_rate": 1.4432014020739012e-05, "loss": 3.6628, "step": 123950 }, { "epoch": 1.3930784224510315, "grad_norm": 1.244869351387024, "learning_rate": 1.4429767107436162e-05, "loss": 3.6546, "step": 124000 }, { "epoch": 1.3936401476213747, "grad_norm": 1.3446252346038818, "learning_rate": 1.442752019413331e-05, "loss": 3.7014, "step": 124050 }, { "epoch": 1.3942018727917178, "grad_norm": 1.0424251556396484, "learning_rate": 1.442527328083046e-05, "loss": 3.7377, "step": 124100 }, { "epoch": 1.394763597962061, "grad_norm": 1.0342272520065308, "learning_rate": 1.4423026367527611e-05, "loss": 3.7084, "step": 124150 }, { "epoch": 1.3953253231324043, "grad_norm": 1.1792279481887817, "learning_rate": 1.442077945422476e-05, "loss": 3.6867, "step": 124200 }, { "epoch": 1.3958870483027475, "grad_norm": 1.029192328453064, "learning_rate": 1.441853254092191e-05, "loss": 3.6794, "step": 124250 }, { "epoch": 1.3964487734730906, "grad_norm": 1.1158950328826904, "learning_rate": 1.4416285627619059e-05, "loss": 3.6644, "step": 124300 }, { "epoch": 1.3970104986434337, "grad_norm": 1.1530020236968994, "learning_rate": 1.441403871431621e-05, "loss": 3.6921, "step": 124350 }, { "epoch": 1.3975722238137769, "grad_norm": 1.100235939025879, "learning_rate": 1.441179180101336e-05, "loss": 3.6874, "step": 124400 }, { "epoch": 1.39813394898412, "grad_norm": 1.1759610176086426, "learning_rate": 1.4409544887710509e-05, "loss": 3.7239, "step": 124450 }, { "epoch": 1.3986956741544632, "grad_norm": 1.1375986337661743, "learning_rate": 1.440729797440766e-05, "loss": 3.693, "step": 124500 }, { "epoch": 1.3992573993248063, "grad_norm": 1.2448017597198486, "learning_rate": 1.4405051061104809e-05, "loss": 3.649, "step": 124550 }, { "epoch": 1.3998191244951494, "grad_norm": 1.1365571022033691, "learning_rate": 1.4402804147801958e-05, "loss": 3.6916, "step": 124600 }, { "epoch": 1.4003808496654926, "grad_norm": 1.1279441118240356, "learning_rate": 1.4400557234499108e-05, "loss": 3.6958, "step": 124650 }, { "epoch": 1.400942574835836, "grad_norm": 1.1293400526046753, "learning_rate": 1.4398310321196259e-05, "loss": 3.6368, "step": 124700 }, { "epoch": 1.401504300006179, "grad_norm": 1.1918554306030273, "learning_rate": 1.4396063407893408e-05, "loss": 3.7152, "step": 124750 }, { "epoch": 1.4020660251765222, "grad_norm": 1.128774881362915, "learning_rate": 1.4393816494590557e-05, "loss": 3.7738, "step": 124800 }, { "epoch": 1.4026277503468654, "grad_norm": 1.1045053005218506, "learning_rate": 1.4391569581287708e-05, "loss": 3.6975, "step": 124850 }, { "epoch": 1.4031894755172085, "grad_norm": 1.0207412242889404, "learning_rate": 1.4389322667984858e-05, "loss": 3.7152, "step": 124900 }, { "epoch": 1.4037512006875517, "grad_norm": 1.2564656734466553, "learning_rate": 1.4387075754682007e-05, "loss": 3.7022, "step": 124950 }, { "epoch": 1.4043129258578948, "grad_norm": 1.2463675737380981, "learning_rate": 1.4384828841379156e-05, "loss": 3.6474, "step": 125000 }, { "epoch": 1.404874651028238, "grad_norm": 1.220080852508545, "learning_rate": 1.4382581928076307e-05, "loss": 3.6833, "step": 125050 }, { "epoch": 1.405436376198581, "grad_norm": 1.3084275722503662, "learning_rate": 1.4380335014773457e-05, "loss": 3.7149, "step": 125100 }, { "epoch": 1.4059981013689242, "grad_norm": 1.0215991735458374, "learning_rate": 1.4378088101470606e-05, "loss": 3.68, "step": 125150 }, { "epoch": 1.4065598265392674, "grad_norm": 1.144219994544983, "learning_rate": 1.4375841188167757e-05, "loss": 3.6669, "step": 125200 }, { "epoch": 1.4071215517096105, "grad_norm": 1.0633809566497803, "learning_rate": 1.4373594274864906e-05, "loss": 3.679, "step": 125250 }, { "epoch": 1.4076832768799536, "grad_norm": 1.0320639610290527, "learning_rate": 1.4371347361562056e-05, "loss": 3.7242, "step": 125300 }, { "epoch": 1.4082450020502968, "grad_norm": 1.2658512592315674, "learning_rate": 1.4369100448259205e-05, "loss": 3.6426, "step": 125350 }, { "epoch": 1.40880672722064, "grad_norm": 1.3241974115371704, "learning_rate": 1.4366853534956356e-05, "loss": 3.6962, "step": 125400 }, { "epoch": 1.409368452390983, "grad_norm": 1.142940878868103, "learning_rate": 1.4364606621653505e-05, "loss": 3.7261, "step": 125450 }, { "epoch": 1.4099301775613262, "grad_norm": 1.0860401391983032, "learning_rate": 1.4362359708350655e-05, "loss": 3.6748, "step": 125500 }, { "epoch": 1.4104919027316696, "grad_norm": 1.0803844928741455, "learning_rate": 1.4360112795047806e-05, "loss": 3.6706, "step": 125550 }, { "epoch": 1.4110536279020127, "grad_norm": 1.0579431056976318, "learning_rate": 1.4357865881744955e-05, "loss": 3.6161, "step": 125600 }, { "epoch": 1.4116153530723559, "grad_norm": 1.117509126663208, "learning_rate": 1.4355618968442104e-05, "loss": 3.6908, "step": 125650 }, { "epoch": 1.412177078242699, "grad_norm": 0.9918485283851624, "learning_rate": 1.4353372055139254e-05, "loss": 3.6771, "step": 125700 }, { "epoch": 1.4127388034130421, "grad_norm": 1.1350773572921753, "learning_rate": 1.4351125141836405e-05, "loss": 3.6657, "step": 125750 }, { "epoch": 1.4133005285833853, "grad_norm": 1.097361445426941, "learning_rate": 1.4348878228533554e-05, "loss": 3.6585, "step": 125800 }, { "epoch": 1.4138622537537284, "grad_norm": 1.077668309211731, "learning_rate": 1.4346631315230703e-05, "loss": 3.6636, "step": 125850 }, { "epoch": 1.4144239789240716, "grad_norm": 1.076488971710205, "learning_rate": 1.4344384401927852e-05, "loss": 3.7048, "step": 125900 }, { "epoch": 1.4149857040944147, "grad_norm": 1.0363222360610962, "learning_rate": 1.4342137488625003e-05, "loss": 3.7138, "step": 125950 }, { "epoch": 1.4155474292647579, "grad_norm": 1.0795608758926392, "learning_rate": 1.4339890575322153e-05, "loss": 3.7081, "step": 126000 }, { "epoch": 1.4161091544351012, "grad_norm": 1.0292444229125977, "learning_rate": 1.4337643662019302e-05, "loss": 3.698, "step": 126050 }, { "epoch": 1.4166708796054444, "grad_norm": 1.1306686401367188, "learning_rate": 1.4335396748716453e-05, "loss": 3.7045, "step": 126100 }, { "epoch": 1.4172326047757875, "grad_norm": 1.0027133226394653, "learning_rate": 1.4333149835413602e-05, "loss": 3.7866, "step": 126150 }, { "epoch": 1.4177943299461306, "grad_norm": 1.2080094814300537, "learning_rate": 1.4330902922110752e-05, "loss": 3.7534, "step": 126200 }, { "epoch": 1.4183560551164738, "grad_norm": 1.169230341911316, "learning_rate": 1.4328656008807901e-05, "loss": 3.6231, "step": 126250 }, { "epoch": 1.418917780286817, "grad_norm": 1.258392572402954, "learning_rate": 1.4326409095505052e-05, "loss": 3.6981, "step": 126300 }, { "epoch": 1.41947950545716, "grad_norm": 1.195525050163269, "learning_rate": 1.4324162182202201e-05, "loss": 3.7301, "step": 126350 }, { "epoch": 1.4200412306275032, "grad_norm": 1.0613311529159546, "learning_rate": 1.432191526889935e-05, "loss": 3.6765, "step": 126400 }, { "epoch": 1.4206029557978463, "grad_norm": 1.3959957361221313, "learning_rate": 1.4319668355596502e-05, "loss": 3.7244, "step": 126450 }, { "epoch": 1.4211646809681895, "grad_norm": 1.214530348777771, "learning_rate": 1.4317421442293651e-05, "loss": 3.7079, "step": 126500 }, { "epoch": 1.4217264061385326, "grad_norm": 1.0872678756713867, "learning_rate": 1.43151745289908e-05, "loss": 3.7104, "step": 126550 }, { "epoch": 1.4222881313088758, "grad_norm": 1.122354507446289, "learning_rate": 1.431292761568795e-05, "loss": 3.6687, "step": 126600 }, { "epoch": 1.422849856479219, "grad_norm": 1.29712975025177, "learning_rate": 1.43106807023851e-05, "loss": 3.6975, "step": 126650 }, { "epoch": 1.423411581649562, "grad_norm": 1.2663835287094116, "learning_rate": 1.430843378908225e-05, "loss": 3.5613, "step": 126700 }, { "epoch": 1.4239733068199052, "grad_norm": 1.0217410326004028, "learning_rate": 1.43061868757794e-05, "loss": 3.7152, "step": 126750 }, { "epoch": 1.4245350319902483, "grad_norm": 1.2005146741867065, "learning_rate": 1.430393996247655e-05, "loss": 3.6889, "step": 126800 }, { "epoch": 1.4250967571605915, "grad_norm": 1.1882201433181763, "learning_rate": 1.43016930491737e-05, "loss": 3.617, "step": 126850 }, { "epoch": 1.4256584823309348, "grad_norm": 1.1847543716430664, "learning_rate": 1.4299446135870849e-05, "loss": 3.6783, "step": 126900 }, { "epoch": 1.426220207501278, "grad_norm": 1.4459447860717773, "learning_rate": 1.4297199222567998e-05, "loss": 3.7479, "step": 126950 }, { "epoch": 1.4267819326716211, "grad_norm": 1.077094554901123, "learning_rate": 1.429495230926515e-05, "loss": 3.6993, "step": 127000 }, { "epoch": 1.4273436578419643, "grad_norm": 0.9693183302879333, "learning_rate": 1.4292705395962299e-05, "loss": 3.5469, "step": 127050 }, { "epoch": 1.4279053830123074, "grad_norm": 0.9766404032707214, "learning_rate": 1.4290458482659448e-05, "loss": 3.721, "step": 127100 }, { "epoch": 1.4284671081826505, "grad_norm": 1.2357624769210815, "learning_rate": 1.4288211569356599e-05, "loss": 3.6714, "step": 127150 }, { "epoch": 1.4290288333529937, "grad_norm": 1.0561624765396118, "learning_rate": 1.4285964656053748e-05, "loss": 3.743, "step": 127200 }, { "epoch": 1.4295905585233368, "grad_norm": 1.4372925758361816, "learning_rate": 1.4283717742750898e-05, "loss": 3.6237, "step": 127250 }, { "epoch": 1.43015228369368, "grad_norm": 1.37134850025177, "learning_rate": 1.4281470829448047e-05, "loss": 3.7034, "step": 127300 }, { "epoch": 1.4307140088640231, "grad_norm": 1.3327314853668213, "learning_rate": 1.4279223916145198e-05, "loss": 3.6874, "step": 127350 }, { "epoch": 1.4312757340343665, "grad_norm": 1.2093868255615234, "learning_rate": 1.4276977002842347e-05, "loss": 3.6656, "step": 127400 }, { "epoch": 1.4318374592047096, "grad_norm": 1.1644083261489868, "learning_rate": 1.4274730089539497e-05, "loss": 3.6415, "step": 127450 }, { "epoch": 1.4323991843750528, "grad_norm": 1.0986287593841553, "learning_rate": 1.4272483176236646e-05, "loss": 3.7425, "step": 127500 }, { "epoch": 1.432960909545396, "grad_norm": 1.1844991445541382, "learning_rate": 1.4270236262933797e-05, "loss": 3.6582, "step": 127550 }, { "epoch": 1.433522634715739, "grad_norm": 1.1148576736450195, "learning_rate": 1.4267989349630946e-05, "loss": 3.637, "step": 127600 }, { "epoch": 1.4340843598860822, "grad_norm": 1.2863951921463013, "learning_rate": 1.4265787374594154e-05, "loss": 3.6883, "step": 127650 }, { "epoch": 1.4346460850564253, "grad_norm": 1.1279274225234985, "learning_rate": 1.4263540461291303e-05, "loss": 3.6854, "step": 127700 }, { "epoch": 1.4352078102267685, "grad_norm": 1.2286314964294434, "learning_rate": 1.4261293547988452e-05, "loss": 3.6296, "step": 127750 }, { "epoch": 1.4357695353971116, "grad_norm": 1.189313530921936, "learning_rate": 1.4259046634685603e-05, "loss": 3.7342, "step": 127800 }, { "epoch": 1.4363312605674547, "grad_norm": 1.071767807006836, "learning_rate": 1.4256799721382753e-05, "loss": 3.7375, "step": 127850 }, { "epoch": 1.436892985737798, "grad_norm": 1.6580266952514648, "learning_rate": 1.4254552808079902e-05, "loss": 3.6849, "step": 127900 }, { "epoch": 1.437454710908141, "grad_norm": 1.3203271627426147, "learning_rate": 1.4252305894777051e-05, "loss": 3.7026, "step": 127950 }, { "epoch": 1.4380164360784842, "grad_norm": 1.336090087890625, "learning_rate": 1.4250058981474202e-05, "loss": 3.7083, "step": 128000 }, { "epoch": 1.4385781612488273, "grad_norm": 1.0259944200515747, "learning_rate": 1.4247812068171352e-05, "loss": 3.7734, "step": 128050 }, { "epoch": 1.4391398864191705, "grad_norm": 1.3454315662384033, "learning_rate": 1.4245565154868501e-05, "loss": 3.6751, "step": 128100 }, { "epoch": 1.4397016115895136, "grad_norm": 1.3078041076660156, "learning_rate": 1.4243318241565652e-05, "loss": 3.6879, "step": 128150 }, { "epoch": 1.4402633367598567, "grad_norm": 1.1114927530288696, "learning_rate": 1.4241071328262801e-05, "loss": 3.6901, "step": 128200 }, { "epoch": 1.4408250619302, "grad_norm": 1.1944338083267212, "learning_rate": 1.423882441495995e-05, "loss": 3.6592, "step": 128250 }, { "epoch": 1.4413867871005432, "grad_norm": 1.1394671201705933, "learning_rate": 1.42365775016571e-05, "loss": 3.7205, "step": 128300 }, { "epoch": 1.4419485122708864, "grad_norm": 1.316195011138916, "learning_rate": 1.4234330588354251e-05, "loss": 3.6758, "step": 128350 }, { "epoch": 1.4425102374412295, "grad_norm": 1.1554863452911377, "learning_rate": 1.42320836750514e-05, "loss": 3.7103, "step": 128400 }, { "epoch": 1.4430719626115727, "grad_norm": 1.152806043624878, "learning_rate": 1.422983676174855e-05, "loss": 3.7541, "step": 128450 }, { "epoch": 1.4436336877819158, "grad_norm": 1.1429660320281982, "learning_rate": 1.4227589848445699e-05, "loss": 3.7786, "step": 128500 }, { "epoch": 1.444195412952259, "grad_norm": 1.2185338735580444, "learning_rate": 1.422534293514285e-05, "loss": 3.6484, "step": 128550 }, { "epoch": 1.444757138122602, "grad_norm": 1.334708333015442, "learning_rate": 1.422309602184e-05, "loss": 3.6916, "step": 128600 }, { "epoch": 1.4453188632929452, "grad_norm": 1.0479364395141602, "learning_rate": 1.4220849108537149e-05, "loss": 3.7259, "step": 128650 }, { "epoch": 1.4458805884632884, "grad_norm": 1.2427948713302612, "learning_rate": 1.42186021952343e-05, "loss": 3.7372, "step": 128700 }, { "epoch": 1.4464423136336317, "grad_norm": 1.1542428731918335, "learning_rate": 1.4216355281931449e-05, "loss": 3.6836, "step": 128750 }, { "epoch": 1.4470040388039749, "grad_norm": 1.1932610273361206, "learning_rate": 1.4214108368628598e-05, "loss": 3.7243, "step": 128800 }, { "epoch": 1.447565763974318, "grad_norm": 1.0314894914627075, "learning_rate": 1.4211861455325748e-05, "loss": 3.6924, "step": 128850 }, { "epoch": 1.4481274891446612, "grad_norm": 1.1394891738891602, "learning_rate": 1.4209614542022899e-05, "loss": 3.6632, "step": 128900 }, { "epoch": 1.4486892143150043, "grad_norm": 1.2204251289367676, "learning_rate": 1.4207367628720048e-05, "loss": 3.7129, "step": 128950 }, { "epoch": 1.4492509394853474, "grad_norm": 1.5714738368988037, "learning_rate": 1.4205120715417197e-05, "loss": 3.6579, "step": 129000 }, { "epoch": 1.4498126646556906, "grad_norm": 1.2154221534729004, "learning_rate": 1.4202873802114348e-05, "loss": 3.6245, "step": 129050 }, { "epoch": 1.4503743898260337, "grad_norm": 1.0103734731674194, "learning_rate": 1.4200626888811498e-05, "loss": 3.6631, "step": 129100 }, { "epoch": 1.4509361149963769, "grad_norm": 1.0002745389938354, "learning_rate": 1.4198379975508647e-05, "loss": 3.7259, "step": 129150 }, { "epoch": 1.45149784016672, "grad_norm": 1.2197633981704712, "learning_rate": 1.4196133062205796e-05, "loss": 3.6225, "step": 129200 }, { "epoch": 1.4520595653370632, "grad_norm": 1.1183518171310425, "learning_rate": 1.4193886148902947e-05, "loss": 3.6419, "step": 129250 }, { "epoch": 1.4526212905074063, "grad_norm": 1.2335314750671387, "learning_rate": 1.4191639235600096e-05, "loss": 3.5827, "step": 129300 }, { "epoch": 1.4531830156777494, "grad_norm": 1.0736817121505737, "learning_rate": 1.41894372605633e-05, "loss": 3.6258, "step": 129350 }, { "epoch": 1.4537447408480926, "grad_norm": 1.283150315284729, "learning_rate": 1.4187190347260452e-05, "loss": 3.7434, "step": 129400 }, { "epoch": 1.4543064660184357, "grad_norm": 1.1960936784744263, "learning_rate": 1.4184943433957601e-05, "loss": 3.7223, "step": 129450 }, { "epoch": 1.4548681911887789, "grad_norm": 1.3931291103363037, "learning_rate": 1.418269652065475e-05, "loss": 3.6696, "step": 129500 }, { "epoch": 1.455429916359122, "grad_norm": 1.104906439781189, "learning_rate": 1.41804496073519e-05, "loss": 3.7111, "step": 129550 }, { "epoch": 1.4559916415294654, "grad_norm": 1.100758671760559, "learning_rate": 1.417820269404905e-05, "loss": 3.6306, "step": 129600 }, { "epoch": 1.4565533666998085, "grad_norm": 1.404305100440979, "learning_rate": 1.41759557807462e-05, "loss": 3.6803, "step": 129650 }, { "epoch": 1.4571150918701516, "grad_norm": 1.3029145002365112, "learning_rate": 1.417370886744335e-05, "loss": 3.6495, "step": 129700 }, { "epoch": 1.4576768170404948, "grad_norm": 1.038677453994751, "learning_rate": 1.41714619541405e-05, "loss": 3.6658, "step": 129750 }, { "epoch": 1.458238542210838, "grad_norm": 1.1654647588729858, "learning_rate": 1.416921504083765e-05, "loss": 3.7818, "step": 129800 }, { "epoch": 1.458800267381181, "grad_norm": 1.0105221271514893, "learning_rate": 1.4166968127534799e-05, "loss": 3.7315, "step": 129850 }, { "epoch": 1.4593619925515242, "grad_norm": 1.0998730659484863, "learning_rate": 1.4164721214231948e-05, "loss": 3.7785, "step": 129900 }, { "epoch": 1.4599237177218674, "grad_norm": 1.0934780836105347, "learning_rate": 1.41624743009291e-05, "loss": 3.6549, "step": 129950 }, { "epoch": 1.4604854428922105, "grad_norm": 1.2375984191894531, "learning_rate": 1.4160227387626249e-05, "loss": 3.7077, "step": 130000 }, { "epoch": 1.4610471680625536, "grad_norm": 1.1644408702850342, "learning_rate": 1.4157980474323398e-05, "loss": 3.6903, "step": 130050 }, { "epoch": 1.461608893232897, "grad_norm": 1.0978128910064697, "learning_rate": 1.4155733561020549e-05, "loss": 3.6279, "step": 130100 }, { "epoch": 1.4621706184032401, "grad_norm": 1.1605157852172852, "learning_rate": 1.4153486647717698e-05, "loss": 3.7902, "step": 130150 }, { "epoch": 1.4627323435735833, "grad_norm": 1.271424412727356, "learning_rate": 1.4151239734414848e-05, "loss": 3.7077, "step": 130200 }, { "epoch": 1.4632940687439264, "grad_norm": 1.1201791763305664, "learning_rate": 1.4148992821111997e-05, "loss": 3.6611, "step": 130250 }, { "epoch": 1.4638557939142696, "grad_norm": 1.347887396812439, "learning_rate": 1.4146745907809148e-05, "loss": 3.697, "step": 130300 }, { "epoch": 1.4644175190846127, "grad_norm": 1.1798382997512817, "learning_rate": 1.4144498994506297e-05, "loss": 3.626, "step": 130350 }, { "epoch": 1.4649792442549558, "grad_norm": 3.724522829055786, "learning_rate": 1.4142252081203447e-05, "loss": 3.6137, "step": 130400 }, { "epoch": 1.465540969425299, "grad_norm": 1.0803759098052979, "learning_rate": 1.4140005167900598e-05, "loss": 3.6533, "step": 130450 }, { "epoch": 1.4661026945956421, "grad_norm": 1.1397240161895752, "learning_rate": 1.4137758254597747e-05, "loss": 3.7668, "step": 130500 }, { "epoch": 1.4666644197659853, "grad_norm": 1.0932416915893555, "learning_rate": 1.4135511341294896e-05, "loss": 3.7927, "step": 130550 }, { "epoch": 1.4672261449363284, "grad_norm": 1.1760075092315674, "learning_rate": 1.4133264427992046e-05, "loss": 3.6757, "step": 130600 }, { "epoch": 1.4677878701066716, "grad_norm": 1.1441518068313599, "learning_rate": 1.4131017514689197e-05, "loss": 3.7135, "step": 130650 }, { "epoch": 1.4683495952770147, "grad_norm": 1.0260703563690186, "learning_rate": 1.4128770601386346e-05, "loss": 3.644, "step": 130700 }, { "epoch": 1.4689113204473578, "grad_norm": 1.1579915285110474, "learning_rate": 1.4126523688083495e-05, "loss": 3.6791, "step": 130750 }, { "epoch": 1.469473045617701, "grad_norm": 1.1837948560714722, "learning_rate": 1.4124276774780644e-05, "loss": 3.6832, "step": 130800 }, { "epoch": 1.4700347707880441, "grad_norm": 1.105905294418335, "learning_rate": 1.4122029861477795e-05, "loss": 3.6534, "step": 130850 }, { "epoch": 1.4705964959583873, "grad_norm": 1.1882402896881104, "learning_rate": 1.4119782948174945e-05, "loss": 3.5552, "step": 130900 }, { "epoch": 1.4711582211287306, "grad_norm": 1.0834487676620483, "learning_rate": 1.4117536034872094e-05, "loss": 3.6657, "step": 130950 }, { "epoch": 1.4717199462990738, "grad_norm": 1.0879491567611694, "learning_rate": 1.4115289121569245e-05, "loss": 3.6612, "step": 131000 }, { "epoch": 1.472281671469417, "grad_norm": 1.14711332321167, "learning_rate": 1.4113042208266394e-05, "loss": 3.6724, "step": 131050 }, { "epoch": 1.47284339663976, "grad_norm": 1.1432496309280396, "learning_rate": 1.4110795294963544e-05, "loss": 3.69, "step": 131100 }, { "epoch": 1.4734051218101032, "grad_norm": 1.0857371091842651, "learning_rate": 1.4108548381660693e-05, "loss": 3.6368, "step": 131150 }, { "epoch": 1.4739668469804463, "grad_norm": 1.097549557685852, "learning_rate": 1.4106301468357844e-05, "loss": 3.6635, "step": 131200 }, { "epoch": 1.4745285721507895, "grad_norm": 1.0512332916259766, "learning_rate": 1.4104054555054993e-05, "loss": 3.7048, "step": 131250 }, { "epoch": 1.4750902973211326, "grad_norm": 1.1470379829406738, "learning_rate": 1.4101807641752143e-05, "loss": 3.7138, "step": 131300 }, { "epoch": 1.4756520224914758, "grad_norm": 1.0969492197036743, "learning_rate": 1.4099560728449294e-05, "loss": 3.7089, "step": 131350 }, { "epoch": 1.476213747661819, "grad_norm": 1.1652694940567017, "learning_rate": 1.4097313815146443e-05, "loss": 3.6553, "step": 131400 }, { "epoch": 1.4767754728321623, "grad_norm": 1.1556792259216309, "learning_rate": 1.4095066901843592e-05, "loss": 3.582, "step": 131450 }, { "epoch": 1.4773371980025054, "grad_norm": 1.1985116004943848, "learning_rate": 1.4092819988540742e-05, "loss": 3.635, "step": 131500 }, { "epoch": 1.4778989231728485, "grad_norm": 1.3924946784973145, "learning_rate": 1.4090573075237893e-05, "loss": 3.7242, "step": 131550 }, { "epoch": 1.4784606483431917, "grad_norm": 1.2415711879730225, "learning_rate": 1.4088326161935042e-05, "loss": 3.752, "step": 131600 }, { "epoch": 1.4790223735135348, "grad_norm": 1.3485480546951294, "learning_rate": 1.4086079248632191e-05, "loss": 3.6784, "step": 131650 }, { "epoch": 1.479584098683878, "grad_norm": 1.1684798002243042, "learning_rate": 1.4083832335329342e-05, "loss": 3.6775, "step": 131700 }, { "epoch": 1.480145823854221, "grad_norm": 1.3369027376174927, "learning_rate": 1.4081585422026492e-05, "loss": 3.7093, "step": 131750 }, { "epoch": 1.4807075490245643, "grad_norm": 1.2231554985046387, "learning_rate": 1.4079338508723641e-05, "loss": 3.7172, "step": 131800 }, { "epoch": 1.4812692741949074, "grad_norm": 1.0951569080352783, "learning_rate": 1.407709159542079e-05, "loss": 3.7198, "step": 131850 }, { "epoch": 1.4818309993652505, "grad_norm": 1.049019455909729, "learning_rate": 1.4074844682117941e-05, "loss": 3.7794, "step": 131900 }, { "epoch": 1.4823927245355937, "grad_norm": 1.3006795644760132, "learning_rate": 1.407259776881509e-05, "loss": 3.6084, "step": 131950 }, { "epoch": 1.4829544497059368, "grad_norm": 1.2521191835403442, "learning_rate": 1.407035085551224e-05, "loss": 3.7331, "step": 132000 }, { "epoch": 1.48351617487628, "grad_norm": 1.125417709350586, "learning_rate": 1.4068103942209391e-05, "loss": 3.6835, "step": 132050 }, { "epoch": 1.484077900046623, "grad_norm": 1.1457781791687012, "learning_rate": 1.406585702890654e-05, "loss": 3.7033, "step": 132100 }, { "epoch": 1.4846396252169662, "grad_norm": 1.1376407146453857, "learning_rate": 1.406361011560369e-05, "loss": 3.7074, "step": 132150 }, { "epoch": 1.4852013503873094, "grad_norm": 1.1445938348770142, "learning_rate": 1.4061363202300839e-05, "loss": 3.6893, "step": 132200 }, { "epoch": 1.4857630755576527, "grad_norm": 1.0046522617340088, "learning_rate": 1.405911628899799e-05, "loss": 3.6741, "step": 132250 }, { "epoch": 1.4863248007279959, "grad_norm": 1.265594482421875, "learning_rate": 1.405686937569514e-05, "loss": 3.6365, "step": 132300 }, { "epoch": 1.486886525898339, "grad_norm": 1.1877540349960327, "learning_rate": 1.4054622462392289e-05, "loss": 3.6737, "step": 132350 }, { "epoch": 1.4874482510686822, "grad_norm": 1.0919491052627563, "learning_rate": 1.4052375549089438e-05, "loss": 3.7282, "step": 132400 }, { "epoch": 1.4880099762390253, "grad_norm": 1.1396973133087158, "learning_rate": 1.4050128635786589e-05, "loss": 3.6833, "step": 132450 }, { "epoch": 1.4885717014093685, "grad_norm": 1.05573570728302, "learning_rate": 1.4047881722483738e-05, "loss": 3.7685, "step": 132500 }, { "epoch": 1.4891334265797116, "grad_norm": 0.9932793974876404, "learning_rate": 1.4045634809180887e-05, "loss": 3.6699, "step": 132550 }, { "epoch": 1.4896951517500547, "grad_norm": 1.2047851085662842, "learning_rate": 1.4043387895878038e-05, "loss": 3.6439, "step": 132600 }, { "epoch": 1.4902568769203979, "grad_norm": 2.069572687149048, "learning_rate": 1.4041140982575188e-05, "loss": 3.6966, "step": 132650 }, { "epoch": 1.490818602090741, "grad_norm": 1.0004225969314575, "learning_rate": 1.4038894069272337e-05, "loss": 3.7124, "step": 132700 }, { "epoch": 1.4913803272610842, "grad_norm": 1.1484732627868652, "learning_rate": 1.4036647155969486e-05, "loss": 3.7173, "step": 132750 }, { "epoch": 1.4919420524314275, "grad_norm": 1.0647058486938477, "learning_rate": 1.4034400242666637e-05, "loss": 3.7109, "step": 132800 }, { "epoch": 1.4925037776017707, "grad_norm": 1.1456990242004395, "learning_rate": 1.4032153329363787e-05, "loss": 3.6422, "step": 132850 }, { "epoch": 1.4930655027721138, "grad_norm": 1.2056528329849243, "learning_rate": 1.4029906416060936e-05, "loss": 3.6598, "step": 132900 }, { "epoch": 1.493627227942457, "grad_norm": 1.307550072669983, "learning_rate": 1.4027659502758087e-05, "loss": 3.6904, "step": 132950 }, { "epoch": 1.4941889531128, "grad_norm": 1.2377219200134277, "learning_rate": 1.4025412589455236e-05, "loss": 3.6669, "step": 133000 }, { "epoch": 1.4947506782831432, "grad_norm": 1.174829363822937, "learning_rate": 1.4023165676152386e-05, "loss": 3.6073, "step": 133050 }, { "epoch": 1.4953124034534864, "grad_norm": 1.102279782295227, "learning_rate": 1.4020918762849535e-05, "loss": 3.701, "step": 133100 }, { "epoch": 1.4958741286238295, "grad_norm": 1.1181694269180298, "learning_rate": 1.4018671849546686e-05, "loss": 3.6779, "step": 133150 }, { "epoch": 1.4964358537941727, "grad_norm": 1.0357697010040283, "learning_rate": 1.4016424936243835e-05, "loss": 3.6923, "step": 133200 }, { "epoch": 1.4969975789645158, "grad_norm": 0.9432259202003479, "learning_rate": 1.4014178022940985e-05, "loss": 3.7048, "step": 133250 }, { "epoch": 1.497559304134859, "grad_norm": 1.0554338693618774, "learning_rate": 1.4011931109638136e-05, "loss": 3.7422, "step": 133300 }, { "epoch": 1.498121029305202, "grad_norm": 1.1480789184570312, "learning_rate": 1.4009684196335285e-05, "loss": 3.7092, "step": 133350 }, { "epoch": 1.4986827544755452, "grad_norm": 1.0501972436904907, "learning_rate": 1.4007437283032434e-05, "loss": 3.7159, "step": 133400 }, { "epoch": 1.4992444796458884, "grad_norm": 1.296221375465393, "learning_rate": 1.4005190369729584e-05, "loss": 3.6433, "step": 133450 }, { "epoch": 1.4998062048162315, "grad_norm": 1.0956120491027832, "learning_rate": 1.4002943456426735e-05, "loss": 3.6807, "step": 133500 }, { "epoch": 1.5003679299865746, "grad_norm": 1.1169849634170532, "learning_rate": 1.4000696543123884e-05, "loss": 3.6143, "step": 133550 }, { "epoch": 1.5009296551569178, "grad_norm": 1.1566176414489746, "learning_rate": 1.3998449629821033e-05, "loss": 3.7322, "step": 133600 }, { "epoch": 1.501491380327261, "grad_norm": 1.040257215499878, "learning_rate": 1.3996202716518184e-05, "loss": 3.6666, "step": 133650 }, { "epoch": 1.5020531054976043, "grad_norm": 1.1366089582443237, "learning_rate": 1.3993955803215334e-05, "loss": 3.7142, "step": 133700 }, { "epoch": 1.5026148306679474, "grad_norm": 1.131287693977356, "learning_rate": 1.3991708889912483e-05, "loss": 3.7191, "step": 133750 }, { "epoch": 1.5031765558382906, "grad_norm": 1.1907302141189575, "learning_rate": 1.3989461976609632e-05, "loss": 3.6475, "step": 133800 }, { "epoch": 1.5037382810086337, "grad_norm": 1.2153143882751465, "learning_rate": 1.3987215063306783e-05, "loss": 3.7205, "step": 133850 }, { "epoch": 1.5043000061789769, "grad_norm": 1.1694841384887695, "learning_rate": 1.3984968150003933e-05, "loss": 3.6523, "step": 133900 }, { "epoch": 1.50486173134932, "grad_norm": 1.141675591468811, "learning_rate": 1.3982721236701082e-05, "loss": 3.6907, "step": 133950 }, { "epoch": 1.5054234565196631, "grad_norm": 1.1201070547103882, "learning_rate": 1.3980474323398233e-05, "loss": 3.752, "step": 134000 }, { "epoch": 1.5059851816900065, "grad_norm": 1.5073152780532837, "learning_rate": 1.3978227410095382e-05, "loss": 3.6264, "step": 134050 }, { "epoch": 1.5065469068603496, "grad_norm": 1.4463489055633545, "learning_rate": 1.3975980496792532e-05, "loss": 3.636, "step": 134100 }, { "epoch": 1.5071086320306928, "grad_norm": 1.1376241445541382, "learning_rate": 1.3973733583489681e-05, "loss": 3.6875, "step": 134150 }, { "epoch": 1.507670357201036, "grad_norm": 1.2015286684036255, "learning_rate": 1.3971486670186832e-05, "loss": 3.6775, "step": 134200 }, { "epoch": 1.508232082371379, "grad_norm": 1.2428511381149292, "learning_rate": 1.3969239756883981e-05, "loss": 3.7064, "step": 134250 }, { "epoch": 1.5087938075417222, "grad_norm": 1.1819654703140259, "learning_rate": 1.396699284358113e-05, "loss": 3.6566, "step": 134300 }, { "epoch": 1.5093555327120654, "grad_norm": 1.1704370975494385, "learning_rate": 1.396474593027828e-05, "loss": 3.6326, "step": 134350 }, { "epoch": 1.5099172578824085, "grad_norm": 1.0891886949539185, "learning_rate": 1.396249901697543e-05, "loss": 3.7003, "step": 134400 }, { "epoch": 1.5104789830527516, "grad_norm": 1.1705445051193237, "learning_rate": 1.396025210367258e-05, "loss": 3.669, "step": 134450 }, { "epoch": 1.5110407082230948, "grad_norm": 1.3058868646621704, "learning_rate": 1.395800519036973e-05, "loss": 3.6446, "step": 134500 }, { "epoch": 1.511602433393438, "grad_norm": 1.1677734851837158, "learning_rate": 1.395575827706688e-05, "loss": 3.6687, "step": 134550 }, { "epoch": 1.512164158563781, "grad_norm": 1.088308334350586, "learning_rate": 1.3953556302030086e-05, "loss": 3.6781, "step": 134600 }, { "epoch": 1.5127258837341242, "grad_norm": 1.1753487586975098, "learning_rate": 1.3951309388727237e-05, "loss": 3.7159, "step": 134650 }, { "epoch": 1.5132876089044673, "grad_norm": 1.3727972507476807, "learning_rate": 1.3949062475424387e-05, "loss": 3.6943, "step": 134700 }, { "epoch": 1.5138493340748105, "grad_norm": 1.4732756614685059, "learning_rate": 1.3946815562121536e-05, "loss": 3.717, "step": 134750 }, { "epoch": 1.5144110592451536, "grad_norm": 1.0472688674926758, "learning_rate": 1.3944568648818685e-05, "loss": 3.6123, "step": 134800 }, { "epoch": 1.5149727844154968, "grad_norm": 1.0159178972244263, "learning_rate": 1.3942321735515836e-05, "loss": 3.6726, "step": 134850 }, { "epoch": 1.51553450958584, "grad_norm": 0.938441812992096, "learning_rate": 1.3940074822212986e-05, "loss": 3.6414, "step": 134900 }, { "epoch": 1.516096234756183, "grad_norm": 1.1761131286621094, "learning_rate": 1.3937827908910135e-05, "loss": 3.6857, "step": 134950 }, { "epoch": 1.5166579599265262, "grad_norm": 1.54920494556427, "learning_rate": 1.3935580995607286e-05, "loss": 3.6467, "step": 135000 }, { "epoch": 1.5172196850968696, "grad_norm": 1.1692527532577515, "learning_rate": 1.3933334082304435e-05, "loss": 3.71, "step": 135050 }, { "epoch": 1.5177814102672127, "grad_norm": 1.1783121824264526, "learning_rate": 1.3931087169001585e-05, "loss": 3.6882, "step": 135100 }, { "epoch": 1.5183431354375558, "grad_norm": 1.5568922758102417, "learning_rate": 1.3928840255698734e-05, "loss": 3.7029, "step": 135150 }, { "epoch": 1.518904860607899, "grad_norm": 1.1709709167480469, "learning_rate": 1.3926593342395885e-05, "loss": 3.6893, "step": 135200 }, { "epoch": 1.5194665857782421, "grad_norm": 1.2699800729751587, "learning_rate": 1.3924346429093034e-05, "loss": 3.6815, "step": 135250 }, { "epoch": 1.5200283109485853, "grad_norm": 1.2915416955947876, "learning_rate": 1.3922099515790184e-05, "loss": 3.6813, "step": 135300 }, { "epoch": 1.5205900361189284, "grad_norm": 1.2163134813308716, "learning_rate": 1.3919852602487333e-05, "loss": 3.7156, "step": 135350 }, { "epoch": 1.5211517612892718, "grad_norm": 1.2914841175079346, "learning_rate": 1.3917605689184484e-05, "loss": 3.6713, "step": 135400 }, { "epoch": 1.521713486459615, "grad_norm": 1.1673322916030884, "learning_rate": 1.3915358775881633e-05, "loss": 3.7357, "step": 135450 }, { "epoch": 1.522275211629958, "grad_norm": 1.0428760051727295, "learning_rate": 1.3913111862578783e-05, "loss": 3.7632, "step": 135500 }, { "epoch": 1.5228369368003012, "grad_norm": 1.1311368942260742, "learning_rate": 1.3910864949275934e-05, "loss": 3.6775, "step": 135550 }, { "epoch": 1.5233986619706443, "grad_norm": 1.162413477897644, "learning_rate": 1.3908618035973083e-05, "loss": 3.7379, "step": 135600 }, { "epoch": 1.5239603871409875, "grad_norm": 1.0958778858184814, "learning_rate": 1.3906371122670232e-05, "loss": 3.7454, "step": 135650 }, { "epoch": 1.5245221123113306, "grad_norm": 1.0307263135910034, "learning_rate": 1.3904124209367382e-05, "loss": 3.7756, "step": 135700 }, { "epoch": 1.5250838374816738, "grad_norm": 1.3076707124710083, "learning_rate": 1.3901877296064533e-05, "loss": 3.7433, "step": 135750 }, { "epoch": 1.525645562652017, "grad_norm": 1.247527837753296, "learning_rate": 1.3899630382761682e-05, "loss": 3.6535, "step": 135800 }, { "epoch": 1.52620728782236, "grad_norm": 1.0686296224594116, "learning_rate": 1.3897383469458831e-05, "loss": 3.6597, "step": 135850 }, { "epoch": 1.5267690129927032, "grad_norm": 1.087568998336792, "learning_rate": 1.3895136556155982e-05, "loss": 3.6908, "step": 135900 }, { "epoch": 1.5273307381630463, "grad_norm": 0.9681762456893921, "learning_rate": 1.3892889642853131e-05, "loss": 3.6816, "step": 135950 }, { "epoch": 1.5278924633333895, "grad_norm": 1.1612462997436523, "learning_rate": 1.389064272955028e-05, "loss": 3.7534, "step": 136000 }, { "epoch": 1.5284541885037326, "grad_norm": 1.172121286392212, "learning_rate": 1.388839581624743e-05, "loss": 3.7198, "step": 136050 }, { "epoch": 1.5290159136740757, "grad_norm": 1.2644882202148438, "learning_rate": 1.3886148902944581e-05, "loss": 3.5908, "step": 136100 }, { "epoch": 1.5295776388444189, "grad_norm": 3.783116340637207, "learning_rate": 1.388390198964173e-05, "loss": 3.7086, "step": 136150 }, { "epoch": 1.530139364014762, "grad_norm": 1.1201610565185547, "learning_rate": 1.388165507633888e-05, "loss": 3.716, "step": 136200 }, { "epoch": 1.5307010891851052, "grad_norm": 1.1066677570343018, "learning_rate": 1.387940816303603e-05, "loss": 3.6825, "step": 136250 }, { "epoch": 1.5312628143554483, "grad_norm": 1.046709418296814, "learning_rate": 1.387716124973318e-05, "loss": 3.6776, "step": 136300 }, { "epoch": 1.5318245395257915, "grad_norm": 1.0653623342514038, "learning_rate": 1.387491433643033e-05, "loss": 3.6368, "step": 136350 }, { "epoch": 1.5323862646961348, "grad_norm": 1.281095027923584, "learning_rate": 1.3872667423127479e-05, "loss": 3.6537, "step": 136400 }, { "epoch": 1.532947989866478, "grad_norm": 1.406171441078186, "learning_rate": 1.387042050982463e-05, "loss": 3.6694, "step": 136450 }, { "epoch": 1.533509715036821, "grad_norm": 1.2122948169708252, "learning_rate": 1.3868173596521779e-05, "loss": 3.7041, "step": 136500 }, { "epoch": 1.5340714402071642, "grad_norm": 1.148941159248352, "learning_rate": 1.3865926683218928e-05, "loss": 3.7345, "step": 136550 }, { "epoch": 1.5346331653775074, "grad_norm": 1.2091238498687744, "learning_rate": 1.386367976991608e-05, "loss": 3.7257, "step": 136600 }, { "epoch": 1.5351948905478505, "grad_norm": 1.314763069152832, "learning_rate": 1.3861432856613229e-05, "loss": 3.7493, "step": 136650 }, { "epoch": 1.5357566157181937, "grad_norm": 1.0040805339813232, "learning_rate": 1.3859185943310378e-05, "loss": 3.6682, "step": 136700 }, { "epoch": 1.536318340888537, "grad_norm": 1.187330722808838, "learning_rate": 1.3856939030007527e-05, "loss": 3.6458, "step": 136750 }, { "epoch": 1.5368800660588802, "grad_norm": 1.1977338790893555, "learning_rate": 1.3854692116704678e-05, "loss": 3.7833, "step": 136800 }, { "epoch": 1.5374417912292233, "grad_norm": 1.2390897274017334, "learning_rate": 1.3852445203401828e-05, "loss": 3.705, "step": 136850 }, { "epoch": 1.5380035163995665, "grad_norm": 1.155185341835022, "learning_rate": 1.3850198290098977e-05, "loss": 3.6979, "step": 136900 }, { "epoch": 1.5385652415699096, "grad_norm": 1.2628538608551025, "learning_rate": 1.3847951376796126e-05, "loss": 3.6829, "step": 136950 }, { "epoch": 1.5391269667402527, "grad_norm": 1.5286192893981934, "learning_rate": 1.3845704463493277e-05, "loss": 3.7026, "step": 137000 }, { "epoch": 1.5396886919105959, "grad_norm": 1.2452843189239502, "learning_rate": 1.3843457550190427e-05, "loss": 3.6803, "step": 137050 }, { "epoch": 1.540250417080939, "grad_norm": 1.0073351860046387, "learning_rate": 1.3841210636887576e-05, "loss": 3.6946, "step": 137100 }, { "epoch": 1.5408121422512822, "grad_norm": 1.2091684341430664, "learning_rate": 1.3838963723584727e-05, "loss": 3.5651, "step": 137150 }, { "epoch": 1.5413738674216253, "grad_norm": 1.0364580154418945, "learning_rate": 1.3836716810281876e-05, "loss": 3.7116, "step": 137200 }, { "epoch": 1.5419355925919684, "grad_norm": 1.0992523431777954, "learning_rate": 1.3834469896979026e-05, "loss": 3.606, "step": 137250 }, { "epoch": 1.5424973177623116, "grad_norm": 1.0807006359100342, "learning_rate": 1.3832222983676175e-05, "loss": 3.6177, "step": 137300 }, { "epoch": 1.5430590429326547, "grad_norm": 1.2157042026519775, "learning_rate": 1.3829976070373326e-05, "loss": 3.6924, "step": 137350 }, { "epoch": 1.5436207681029979, "grad_norm": 1.1488664150238037, "learning_rate": 1.3827729157070475e-05, "loss": 3.7056, "step": 137400 }, { "epoch": 1.544182493273341, "grad_norm": 1.1668779850006104, "learning_rate": 1.3825482243767625e-05, "loss": 3.7133, "step": 137450 }, { "epoch": 1.5447442184436841, "grad_norm": 1.0572956800460815, "learning_rate": 1.3823235330464776e-05, "loss": 3.6649, "step": 137500 }, { "epoch": 1.5453059436140273, "grad_norm": 1.1351521015167236, "learning_rate": 1.3820988417161925e-05, "loss": 3.6284, "step": 137550 }, { "epoch": 1.5458676687843704, "grad_norm": 1.3236806392669678, "learning_rate": 1.3818741503859074e-05, "loss": 3.6578, "step": 137600 }, { "epoch": 1.5464293939547136, "grad_norm": 1.1036694049835205, "learning_rate": 1.3816494590556224e-05, "loss": 3.6787, "step": 137650 }, { "epoch": 1.5469911191250567, "grad_norm": 1.151566743850708, "learning_rate": 1.3814247677253375e-05, "loss": 3.7113, "step": 137700 }, { "epoch": 1.5475528442954, "grad_norm": 0.9375407695770264, "learning_rate": 1.3812000763950524e-05, "loss": 3.6449, "step": 137750 }, { "epoch": 1.5481145694657432, "grad_norm": 1.1477205753326416, "learning_rate": 1.3809753850647673e-05, "loss": 3.6459, "step": 137800 }, { "epoch": 1.5486762946360864, "grad_norm": 1.1529345512390137, "learning_rate": 1.3807506937344824e-05, "loss": 3.715, "step": 137850 }, { "epoch": 1.5492380198064295, "grad_norm": 1.256088137626648, "learning_rate": 1.3805260024041973e-05, "loss": 3.8204, "step": 137900 }, { "epoch": 1.5497997449767726, "grad_norm": 1.1161493062973022, "learning_rate": 1.3803013110739123e-05, "loss": 3.6332, "step": 137950 }, { "epoch": 1.5503614701471158, "grad_norm": 1.0601900815963745, "learning_rate": 1.3800766197436272e-05, "loss": 3.6654, "step": 138000 }, { "epoch": 1.550923195317459, "grad_norm": 1.0600312948226929, "learning_rate": 1.3798519284133423e-05, "loss": 3.627, "step": 138050 }, { "epoch": 1.5514849204878023, "grad_norm": 1.2241134643554688, "learning_rate": 1.3796272370830572e-05, "loss": 3.668, "step": 138100 }, { "epoch": 1.5520466456581454, "grad_norm": 1.3569682836532593, "learning_rate": 1.3794025457527722e-05, "loss": 3.6421, "step": 138150 }, { "epoch": 1.5526083708284886, "grad_norm": 1.1638623476028442, "learning_rate": 1.3791778544224873e-05, "loss": 3.6432, "step": 138200 }, { "epoch": 1.5531700959988317, "grad_norm": 1.006299614906311, "learning_rate": 1.3789531630922022e-05, "loss": 3.6522, "step": 138250 }, { "epoch": 1.5537318211691749, "grad_norm": 1.0800831317901611, "learning_rate": 1.3787284717619171e-05, "loss": 3.6606, "step": 138300 }, { "epoch": 1.554293546339518, "grad_norm": 1.030090093612671, "learning_rate": 1.378503780431632e-05, "loss": 3.6854, "step": 138350 }, { "epoch": 1.5548552715098611, "grad_norm": 1.1533602476119995, "learning_rate": 1.3782790891013472e-05, "loss": 3.6155, "step": 138400 }, { "epoch": 1.5554169966802043, "grad_norm": 1.198823094367981, "learning_rate": 1.3780543977710621e-05, "loss": 3.8225, "step": 138450 }, { "epoch": 1.5559787218505474, "grad_norm": 3.5411183834075928, "learning_rate": 1.377829706440777e-05, "loss": 3.6767, "step": 138500 }, { "epoch": 1.5565404470208906, "grad_norm": 1.1365792751312256, "learning_rate": 1.377605015110492e-05, "loss": 3.6306, "step": 138550 }, { "epoch": 1.5571021721912337, "grad_norm": 1.0482217073440552, "learning_rate": 1.377380323780207e-05, "loss": 3.6779, "step": 138600 }, { "epoch": 1.5576638973615768, "grad_norm": 1.256150484085083, "learning_rate": 1.377155632449922e-05, "loss": 3.6576, "step": 138650 }, { "epoch": 1.55822562253192, "grad_norm": 1.284239649772644, "learning_rate": 1.3769354349462428e-05, "loss": 3.602, "step": 138700 }, { "epoch": 1.5587873477022631, "grad_norm": 1.1476000547409058, "learning_rate": 1.3767107436159577e-05, "loss": 3.7112, "step": 138750 }, { "epoch": 1.5593490728726063, "grad_norm": 0.9629308581352234, "learning_rate": 1.3764860522856726e-05, "loss": 3.6444, "step": 138800 }, { "epoch": 1.5599107980429494, "grad_norm": 1.180854082107544, "learning_rate": 1.3762613609553877e-05, "loss": 3.6665, "step": 138850 }, { "epoch": 1.5604725232132926, "grad_norm": 1.1548141241073608, "learning_rate": 1.3760366696251027e-05, "loss": 3.6188, "step": 138900 }, { "epoch": 1.5610342483836357, "grad_norm": 1.1530333757400513, "learning_rate": 1.3758119782948176e-05, "loss": 3.6888, "step": 138950 }, { "epoch": 1.5615959735539788, "grad_norm": 1.10718834400177, "learning_rate": 1.3755872869645325e-05, "loss": 3.6498, "step": 139000 }, { "epoch": 1.562157698724322, "grad_norm": 1.4435906410217285, "learning_rate": 1.3753625956342476e-05, "loss": 3.7191, "step": 139050 }, { "epoch": 1.5627194238946653, "grad_norm": 1.1973544359207153, "learning_rate": 1.3751379043039626e-05, "loss": 3.674, "step": 139100 }, { "epoch": 1.5632811490650085, "grad_norm": 1.0264248847961426, "learning_rate": 1.3749132129736775e-05, "loss": 3.6218, "step": 139150 }, { "epoch": 1.5638428742353516, "grad_norm": 1.1173083782196045, "learning_rate": 1.3746885216433926e-05, "loss": 3.6692, "step": 139200 }, { "epoch": 1.5644045994056948, "grad_norm": 1.0052393674850464, "learning_rate": 1.3744638303131075e-05, "loss": 3.7372, "step": 139250 }, { "epoch": 1.564966324576038, "grad_norm": 1.0546742677688599, "learning_rate": 1.3742391389828224e-05, "loss": 3.7064, "step": 139300 }, { "epoch": 1.565528049746381, "grad_norm": 1.112864375114441, "learning_rate": 1.3740144476525374e-05, "loss": 3.6565, "step": 139350 }, { "epoch": 1.5660897749167244, "grad_norm": 1.0232577323913574, "learning_rate": 1.3737897563222525e-05, "loss": 3.7068, "step": 139400 }, { "epoch": 1.5666515000870676, "grad_norm": 1.1858417987823486, "learning_rate": 1.3735650649919674e-05, "loss": 3.7506, "step": 139450 }, { "epoch": 1.5672132252574107, "grad_norm": 1.325637936592102, "learning_rate": 1.3733403736616823e-05, "loss": 3.6607, "step": 139500 }, { "epoch": 1.5677749504277538, "grad_norm": 1.0840829610824585, "learning_rate": 1.3731156823313973e-05, "loss": 3.7168, "step": 139550 }, { "epoch": 1.568336675598097, "grad_norm": 1.2827153205871582, "learning_rate": 1.3728909910011124e-05, "loss": 3.6482, "step": 139600 }, { "epoch": 1.5688984007684401, "grad_norm": 1.183478593826294, "learning_rate": 1.3726662996708273e-05, "loss": 3.6704, "step": 139650 }, { "epoch": 1.5694601259387833, "grad_norm": 1.265271544456482, "learning_rate": 1.3724416083405422e-05, "loss": 3.7252, "step": 139700 }, { "epoch": 1.5700218511091264, "grad_norm": 1.3925718069076538, "learning_rate": 1.3722169170102573e-05, "loss": 3.698, "step": 139750 }, { "epoch": 1.5705835762794695, "grad_norm": 1.0602258443832397, "learning_rate": 1.3719922256799723e-05, "loss": 3.6162, "step": 139800 }, { "epoch": 1.5711453014498127, "grad_norm": 1.487443447113037, "learning_rate": 1.3717675343496872e-05, "loss": 3.6364, "step": 139850 }, { "epoch": 1.5717070266201558, "grad_norm": 1.189277172088623, "learning_rate": 1.3715428430194021e-05, "loss": 3.618, "step": 139900 }, { "epoch": 1.572268751790499, "grad_norm": 1.0855302810668945, "learning_rate": 1.3713181516891172e-05, "loss": 3.6734, "step": 139950 }, { "epoch": 1.572830476960842, "grad_norm": 1.330960988998413, "learning_rate": 1.3710934603588322e-05, "loss": 3.6967, "step": 140000 }, { "epoch": 1.5733922021311852, "grad_norm": 1.1523526906967163, "learning_rate": 1.3708687690285471e-05, "loss": 3.6636, "step": 140050 }, { "epoch": 1.5739539273015284, "grad_norm": 1.225419044494629, "learning_rate": 1.3706440776982622e-05, "loss": 3.7668, "step": 140100 }, { "epoch": 1.5745156524718715, "grad_norm": 1.1650224924087524, "learning_rate": 1.3704193863679771e-05, "loss": 3.6304, "step": 140150 }, { "epoch": 1.5750773776422147, "grad_norm": 1.2172008752822876, "learning_rate": 1.370194695037692e-05, "loss": 3.7296, "step": 140200 }, { "epoch": 1.5756391028125578, "grad_norm": 1.1196541786193848, "learning_rate": 1.369970003707407e-05, "loss": 3.69, "step": 140250 }, { "epoch": 1.576200827982901, "grad_norm": 1.1689870357513428, "learning_rate": 1.3697453123771221e-05, "loss": 3.7265, "step": 140300 }, { "epoch": 1.576762553153244, "grad_norm": 1.0626927614212036, "learning_rate": 1.369520621046837e-05, "loss": 3.6962, "step": 140350 }, { "epoch": 1.5773242783235872, "grad_norm": 1.0944437980651855, "learning_rate": 1.369295929716552e-05, "loss": 3.7403, "step": 140400 }, { "epoch": 1.5778860034939306, "grad_norm": 1.046596646308899, "learning_rate": 1.369071238386267e-05, "loss": 3.6516, "step": 140450 }, { "epoch": 1.5784477286642737, "grad_norm": 1.0734217166900635, "learning_rate": 1.368846547055982e-05, "loss": 3.756, "step": 140500 }, { "epoch": 1.5790094538346169, "grad_norm": 1.1881511211395264, "learning_rate": 1.368621855725697e-05, "loss": 3.6859, "step": 140550 }, { "epoch": 1.57957117900496, "grad_norm": 1.231651782989502, "learning_rate": 1.3683971643954119e-05, "loss": 3.6863, "step": 140600 }, { "epoch": 1.5801329041753032, "grad_norm": 1.2965707778930664, "learning_rate": 1.368172473065127e-05, "loss": 3.6765, "step": 140650 }, { "epoch": 1.5806946293456463, "grad_norm": 1.2120448350906372, "learning_rate": 1.3679477817348419e-05, "loss": 3.6023, "step": 140700 }, { "epoch": 1.5812563545159897, "grad_norm": 1.1119015216827393, "learning_rate": 1.3677230904045568e-05, "loss": 3.7126, "step": 140750 }, { "epoch": 1.5818180796863328, "grad_norm": 1.1401653289794922, "learning_rate": 1.367498399074272e-05, "loss": 3.6261, "step": 140800 }, { "epoch": 1.582379804856676, "grad_norm": 1.1538084745407104, "learning_rate": 1.3672737077439869e-05, "loss": 3.7259, "step": 140850 }, { "epoch": 1.582941530027019, "grad_norm": 1.398902177810669, "learning_rate": 1.3670490164137018e-05, "loss": 3.7078, "step": 140900 }, { "epoch": 1.5835032551973622, "grad_norm": 1.173081398010254, "learning_rate": 1.3668243250834167e-05, "loss": 3.701, "step": 140950 }, { "epoch": 1.5840649803677054, "grad_norm": 1.103658676147461, "learning_rate": 1.3665996337531318e-05, "loss": 3.6483, "step": 141000 }, { "epoch": 1.5846267055380485, "grad_norm": 1.2359012365341187, "learning_rate": 1.3663749424228468e-05, "loss": 3.6826, "step": 141050 }, { "epoch": 1.5851884307083917, "grad_norm": 1.2011862993240356, "learning_rate": 1.3661502510925617e-05, "loss": 3.7179, "step": 141100 }, { "epoch": 1.5857501558787348, "grad_norm": 1.0387769937515259, "learning_rate": 1.3659255597622766e-05, "loss": 3.7604, "step": 141150 }, { "epoch": 1.586311881049078, "grad_norm": 1.2249939441680908, "learning_rate": 1.3657008684319917e-05, "loss": 3.6343, "step": 141200 }, { "epoch": 1.586873606219421, "grad_norm": 1.3379638195037842, "learning_rate": 1.3654761771017066e-05, "loss": 3.7245, "step": 141250 }, { "epoch": 1.5874353313897642, "grad_norm": 1.1127783060073853, "learning_rate": 1.3652559795980274e-05, "loss": 3.6998, "step": 141300 }, { "epoch": 1.5879970565601074, "grad_norm": 1.1300557851791382, "learning_rate": 1.3650312882677423e-05, "loss": 3.5868, "step": 141350 }, { "epoch": 1.5885587817304505, "grad_norm": 1.0959243774414062, "learning_rate": 1.3648065969374573e-05, "loss": 3.6953, "step": 141400 }, { "epoch": 1.5891205069007937, "grad_norm": 1.0239161252975464, "learning_rate": 1.3645819056071724e-05, "loss": 3.6807, "step": 141450 }, { "epoch": 1.5896822320711368, "grad_norm": 1.115818977355957, "learning_rate": 1.3643572142768873e-05, "loss": 3.6034, "step": 141500 }, { "epoch": 1.59024395724148, "grad_norm": 1.1584968566894531, "learning_rate": 1.3641325229466022e-05, "loss": 3.6404, "step": 141550 }, { "epoch": 1.590805682411823, "grad_norm": 1.4609769582748413, "learning_rate": 1.3639078316163172e-05, "loss": 3.7024, "step": 141600 }, { "epoch": 1.5913674075821662, "grad_norm": 1.1476662158966064, "learning_rate": 1.3636831402860323e-05, "loss": 3.6814, "step": 141650 }, { "epoch": 1.5919291327525094, "grad_norm": 1.1682904958724976, "learning_rate": 1.3634584489557472e-05, "loss": 3.7362, "step": 141700 }, { "epoch": 1.5924908579228527, "grad_norm": 1.1838876008987427, "learning_rate": 1.3632337576254621e-05, "loss": 3.6828, "step": 141750 }, { "epoch": 1.5930525830931959, "grad_norm": 1.0836089849472046, "learning_rate": 1.3630090662951772e-05, "loss": 3.727, "step": 141800 }, { "epoch": 1.593614308263539, "grad_norm": 1.0461901426315308, "learning_rate": 1.3627843749648922e-05, "loss": 3.6681, "step": 141850 }, { "epoch": 1.5941760334338821, "grad_norm": 1.067426323890686, "learning_rate": 1.3625596836346071e-05, "loss": 3.6756, "step": 141900 }, { "epoch": 1.5947377586042253, "grad_norm": 1.1616348028182983, "learning_rate": 1.362334992304322e-05, "loss": 3.6586, "step": 141950 }, { "epoch": 1.5952994837745684, "grad_norm": 1.3039299249649048, "learning_rate": 1.3621103009740371e-05, "loss": 3.5232, "step": 142000 }, { "epoch": 1.5958612089449116, "grad_norm": 1.2461251020431519, "learning_rate": 1.361885609643752e-05, "loss": 3.6559, "step": 142050 }, { "epoch": 1.596422934115255, "grad_norm": 1.3545992374420166, "learning_rate": 1.361660918313467e-05, "loss": 3.7266, "step": 142100 }, { "epoch": 1.596984659285598, "grad_norm": 2.3668854236602783, "learning_rate": 1.361436226983182e-05, "loss": 3.692, "step": 142150 }, { "epoch": 1.5975463844559412, "grad_norm": 1.298037052154541, "learning_rate": 1.361211535652897e-05, "loss": 3.6174, "step": 142200 }, { "epoch": 1.5981081096262844, "grad_norm": 1.0800282955169678, "learning_rate": 1.360986844322612e-05, "loss": 3.6606, "step": 142250 }, { "epoch": 1.5986698347966275, "grad_norm": 1.1888720989227295, "learning_rate": 1.3607621529923269e-05, "loss": 3.6515, "step": 142300 }, { "epoch": 1.5992315599669706, "grad_norm": 1.1780089139938354, "learning_rate": 1.360537461662042e-05, "loss": 3.7173, "step": 142350 }, { "epoch": 1.5997932851373138, "grad_norm": 1.1620633602142334, "learning_rate": 1.360312770331757e-05, "loss": 3.6857, "step": 142400 }, { "epoch": 1.600355010307657, "grad_norm": 1.112202525138855, "learning_rate": 1.3600880790014719e-05, "loss": 3.6677, "step": 142450 }, { "epoch": 1.600916735478, "grad_norm": 1.1175018548965454, "learning_rate": 1.3598633876711868e-05, "loss": 3.6737, "step": 142500 }, { "epoch": 1.6014784606483432, "grad_norm": 1.0645023584365845, "learning_rate": 1.3596386963409019e-05, "loss": 3.6471, "step": 142550 }, { "epoch": 1.6020401858186863, "grad_norm": 1.084029197692871, "learning_rate": 1.3594140050106168e-05, "loss": 3.6971, "step": 142600 }, { "epoch": 1.6026019109890295, "grad_norm": 1.1282392740249634, "learning_rate": 1.3591893136803318e-05, "loss": 3.7043, "step": 142650 }, { "epoch": 1.6031636361593726, "grad_norm": 1.0407531261444092, "learning_rate": 1.3589646223500469e-05, "loss": 3.6152, "step": 142700 }, { "epoch": 1.6037253613297158, "grad_norm": 1.0244865417480469, "learning_rate": 1.3587399310197618e-05, "loss": 3.7487, "step": 142750 }, { "epoch": 1.604287086500059, "grad_norm": 1.1283087730407715, "learning_rate": 1.3585152396894767e-05, "loss": 3.6107, "step": 142800 }, { "epoch": 1.604848811670402, "grad_norm": 1.0483598709106445, "learning_rate": 1.3582905483591916e-05, "loss": 3.6978, "step": 142850 }, { "epoch": 1.6054105368407452, "grad_norm": 1.1567394733428955, "learning_rate": 1.3580658570289067e-05, "loss": 3.6641, "step": 142900 }, { "epoch": 1.6059722620110883, "grad_norm": 1.137740969657898, "learning_rate": 1.3578456595252273e-05, "loss": 3.6434, "step": 142950 }, { "epoch": 1.6065339871814315, "grad_norm": 1.0882683992385864, "learning_rate": 1.3576209681949424e-05, "loss": 3.7948, "step": 143000 }, { "epoch": 1.6070957123517746, "grad_norm": 1.1528105735778809, "learning_rate": 1.3573962768646574e-05, "loss": 3.6705, "step": 143050 }, { "epoch": 1.607657437522118, "grad_norm": 1.1751548051834106, "learning_rate": 1.3571715855343723e-05, "loss": 3.636, "step": 143100 }, { "epoch": 1.6082191626924611, "grad_norm": 1.5588665008544922, "learning_rate": 1.3569468942040872e-05, "loss": 3.7752, "step": 143150 }, { "epoch": 1.6087808878628043, "grad_norm": 1.3791288137435913, "learning_rate": 1.3567222028738023e-05, "loss": 3.6663, "step": 143200 }, { "epoch": 1.6093426130331474, "grad_norm": 1.1276522874832153, "learning_rate": 1.3564975115435173e-05, "loss": 3.593, "step": 143250 }, { "epoch": 1.6099043382034905, "grad_norm": 1.1951262950897217, "learning_rate": 1.3562728202132322e-05, "loss": 3.6767, "step": 143300 }, { "epoch": 1.6104660633738337, "grad_norm": 1.2520290613174438, "learning_rate": 1.3560481288829473e-05, "loss": 3.6266, "step": 143350 }, { "epoch": 1.6110277885441768, "grad_norm": 1.223700761795044, "learning_rate": 1.3558234375526622e-05, "loss": 3.6326, "step": 143400 }, { "epoch": 1.6115895137145202, "grad_norm": 1.3785969018936157, "learning_rate": 1.3555987462223772e-05, "loss": 3.6384, "step": 143450 }, { "epoch": 1.6121512388848633, "grad_norm": 1.168166995048523, "learning_rate": 1.3553740548920921e-05, "loss": 3.672, "step": 143500 }, { "epoch": 1.6127129640552065, "grad_norm": 1.169132113456726, "learning_rate": 1.3551493635618072e-05, "loss": 3.6733, "step": 143550 }, { "epoch": 1.6132746892255496, "grad_norm": 1.4326705932617188, "learning_rate": 1.3549246722315221e-05, "loss": 3.7418, "step": 143600 }, { "epoch": 1.6138364143958928, "grad_norm": 1.1875486373901367, "learning_rate": 1.354699980901237e-05, "loss": 3.6875, "step": 143650 }, { "epoch": 1.614398139566236, "grad_norm": 0.9022040367126465, "learning_rate": 1.3544752895709522e-05, "loss": 3.7225, "step": 143700 }, { "epoch": 1.614959864736579, "grad_norm": 1.1078884601593018, "learning_rate": 1.3542505982406671e-05, "loss": 3.6703, "step": 143750 }, { "epoch": 1.6155215899069222, "grad_norm": 1.1090278625488281, "learning_rate": 1.354025906910382e-05, "loss": 3.6247, "step": 143800 }, { "epoch": 1.6160833150772653, "grad_norm": 1.1982507705688477, "learning_rate": 1.353801215580097e-05, "loss": 3.6545, "step": 143850 }, { "epoch": 1.6166450402476085, "grad_norm": 1.2535675764083862, "learning_rate": 1.353576524249812e-05, "loss": 3.7138, "step": 143900 }, { "epoch": 1.6172067654179516, "grad_norm": 1.0719192028045654, "learning_rate": 1.353351832919527e-05, "loss": 3.6688, "step": 143950 }, { "epoch": 1.6177684905882947, "grad_norm": 0.994109034538269, "learning_rate": 1.353127141589242e-05, "loss": 3.7049, "step": 144000 }, { "epoch": 1.618330215758638, "grad_norm": 1.3408006429672241, "learning_rate": 1.352902450258957e-05, "loss": 3.6536, "step": 144050 }, { "epoch": 1.618891940928981, "grad_norm": 1.1289949417114258, "learning_rate": 1.352677758928672e-05, "loss": 3.6695, "step": 144100 }, { "epoch": 1.6194536660993242, "grad_norm": 1.5051370859146118, "learning_rate": 1.3524530675983869e-05, "loss": 3.6928, "step": 144150 }, { "epoch": 1.6200153912696673, "grad_norm": 1.1607725620269775, "learning_rate": 1.3522283762681018e-05, "loss": 3.6913, "step": 144200 }, { "epoch": 1.6205771164400105, "grad_norm": 1.049977421760559, "learning_rate": 1.352003684937817e-05, "loss": 3.6361, "step": 144250 }, { "epoch": 1.6211388416103536, "grad_norm": 0.9868359565734863, "learning_rate": 1.3517789936075318e-05, "loss": 3.7028, "step": 144300 }, { "epoch": 1.6217005667806967, "grad_norm": 1.1460314989089966, "learning_rate": 1.3515543022772468e-05, "loss": 3.7152, "step": 144350 }, { "epoch": 1.6222622919510399, "grad_norm": 1.1550520658493042, "learning_rate": 1.3513296109469619e-05, "loss": 3.7497, "step": 144400 }, { "epoch": 1.6228240171213832, "grad_norm": 1.182363510131836, "learning_rate": 1.3511049196166768e-05, "loss": 3.6572, "step": 144450 }, { "epoch": 1.6233857422917264, "grad_norm": 1.1829805374145508, "learning_rate": 1.3508802282863917e-05, "loss": 3.633, "step": 144500 }, { "epoch": 1.6239474674620695, "grad_norm": 1.1781424283981323, "learning_rate": 1.3506555369561067e-05, "loss": 3.7259, "step": 144550 }, { "epoch": 1.6245091926324127, "grad_norm": 1.1826103925704956, "learning_rate": 1.3504308456258218e-05, "loss": 3.6989, "step": 144600 }, { "epoch": 1.6250709178027558, "grad_norm": 1.1167058944702148, "learning_rate": 1.3502061542955367e-05, "loss": 3.7153, "step": 144650 }, { "epoch": 1.625632642973099, "grad_norm": 1.5774495601654053, "learning_rate": 1.3499814629652516e-05, "loss": 3.6014, "step": 144700 }, { "epoch": 1.626194368143442, "grad_norm": 1.1393110752105713, "learning_rate": 1.3497567716349666e-05, "loss": 3.7552, "step": 144750 }, { "epoch": 1.6267560933137855, "grad_norm": 1.0390650033950806, "learning_rate": 1.3495320803046817e-05, "loss": 3.6646, "step": 144800 }, { "epoch": 1.6273178184841286, "grad_norm": 1.3771274089813232, "learning_rate": 1.3493073889743966e-05, "loss": 3.7414, "step": 144850 }, { "epoch": 1.6278795436544717, "grad_norm": 1.2277251482009888, "learning_rate": 1.3490826976441115e-05, "loss": 3.7624, "step": 144900 }, { "epoch": 1.6284412688248149, "grad_norm": 1.057714581489563, "learning_rate": 1.3488580063138266e-05, "loss": 3.747, "step": 144950 }, { "epoch": 1.629002993995158, "grad_norm": 1.2887024879455566, "learning_rate": 1.3486333149835416e-05, "loss": 3.6581, "step": 145000 }, { "epoch": 1.6295647191655012, "grad_norm": 1.094120740890503, "learning_rate": 1.3484086236532565e-05, "loss": 3.6541, "step": 145050 }, { "epoch": 1.6301264443358443, "grad_norm": 1.047278881072998, "learning_rate": 1.3481839323229714e-05, "loss": 3.6114, "step": 145100 }, { "epoch": 1.6306881695061874, "grad_norm": 1.2887330055236816, "learning_rate": 1.3479592409926865e-05, "loss": 3.7136, "step": 145150 }, { "epoch": 1.6312498946765306, "grad_norm": 1.118496298789978, "learning_rate": 1.3477345496624015e-05, "loss": 3.7317, "step": 145200 }, { "epoch": 1.6318116198468737, "grad_norm": 1.2175357341766357, "learning_rate": 1.3475098583321164e-05, "loss": 3.6609, "step": 145250 }, { "epoch": 1.6323733450172169, "grad_norm": 1.1313681602478027, "learning_rate": 1.3472851670018315e-05, "loss": 3.6874, "step": 145300 }, { "epoch": 1.63293507018756, "grad_norm": 1.1100246906280518, "learning_rate": 1.3470604756715464e-05, "loss": 3.6398, "step": 145350 }, { "epoch": 1.6334967953579032, "grad_norm": 1.0380334854125977, "learning_rate": 1.3468357843412614e-05, "loss": 3.6961, "step": 145400 }, { "epoch": 1.6340585205282463, "grad_norm": 1.1742682456970215, "learning_rate": 1.3466110930109763e-05, "loss": 3.6451, "step": 145450 }, { "epoch": 1.6346202456985894, "grad_norm": 1.14945650100708, "learning_rate": 1.3463864016806914e-05, "loss": 3.6867, "step": 145500 }, { "epoch": 1.6351819708689326, "grad_norm": 1.1547523736953735, "learning_rate": 1.3461617103504063e-05, "loss": 3.7471, "step": 145550 }, { "epoch": 1.6357436960392757, "grad_norm": 1.1640186309814453, "learning_rate": 1.3459370190201213e-05, "loss": 3.6456, "step": 145600 }, { "epoch": 1.6363054212096189, "grad_norm": 1.2586885690689087, "learning_rate": 1.3457123276898364e-05, "loss": 3.721, "step": 145650 }, { "epoch": 1.636867146379962, "grad_norm": 1.0116692781448364, "learning_rate": 1.3454876363595513e-05, "loss": 3.6787, "step": 145700 }, { "epoch": 1.6374288715503051, "grad_norm": 1.1636160612106323, "learning_rate": 1.3452629450292662e-05, "loss": 3.6964, "step": 145750 }, { "epoch": 1.6379905967206485, "grad_norm": 1.192798376083374, "learning_rate": 1.3450382536989812e-05, "loss": 3.6242, "step": 145800 }, { "epoch": 1.6385523218909916, "grad_norm": 1.0549873113632202, "learning_rate": 1.3448135623686963e-05, "loss": 3.7125, "step": 145850 }, { "epoch": 1.6391140470613348, "grad_norm": 1.1213629245758057, "learning_rate": 1.3445888710384112e-05, "loss": 3.6865, "step": 145900 }, { "epoch": 1.639675772231678, "grad_norm": 1.334212303161621, "learning_rate": 1.3443641797081261e-05, "loss": 3.6126, "step": 145950 }, { "epoch": 1.640237497402021, "grad_norm": 1.2307395935058594, "learning_rate": 1.3441394883778412e-05, "loss": 3.6608, "step": 146000 }, { "epoch": 1.6407992225723642, "grad_norm": 1.046105980873108, "learning_rate": 1.3439147970475562e-05, "loss": 3.6599, "step": 146050 }, { "epoch": 1.6413609477427074, "grad_norm": 1.116716980934143, "learning_rate": 1.343690105717271e-05, "loss": 3.6798, "step": 146100 }, { "epoch": 1.6419226729130507, "grad_norm": 1.1969423294067383, "learning_rate": 1.343465414386986e-05, "loss": 3.612, "step": 146150 }, { "epoch": 1.6424843980833939, "grad_norm": 1.1656075716018677, "learning_rate": 1.3432407230567011e-05, "loss": 3.6648, "step": 146200 }, { "epoch": 1.643046123253737, "grad_norm": 1.1140774488449097, "learning_rate": 1.343016031726416e-05, "loss": 3.6866, "step": 146250 }, { "epoch": 1.6436078484240801, "grad_norm": 1.15651273727417, "learning_rate": 1.342791340396131e-05, "loss": 3.6857, "step": 146300 }, { "epoch": 1.6441695735944233, "grad_norm": 1.2643494606018066, "learning_rate": 1.3425666490658459e-05, "loss": 3.6299, "step": 146350 }, { "epoch": 1.6447312987647664, "grad_norm": 1.1954561471939087, "learning_rate": 1.342341957735561e-05, "loss": 3.5863, "step": 146400 }, { "epoch": 1.6452930239351096, "grad_norm": 1.0443463325500488, "learning_rate": 1.342117266405276e-05, "loss": 3.6834, "step": 146450 }, { "epoch": 1.6458547491054527, "grad_norm": 1.190408706665039, "learning_rate": 1.3418925750749909e-05, "loss": 3.7262, "step": 146500 }, { "epoch": 1.6464164742757958, "grad_norm": 1.1993037462234497, "learning_rate": 1.341667883744706e-05, "loss": 3.6069, "step": 146550 }, { "epoch": 1.646978199446139, "grad_norm": 1.260898470878601, "learning_rate": 1.3414431924144209e-05, "loss": 3.6386, "step": 146600 }, { "epoch": 1.6475399246164821, "grad_norm": 1.1906691789627075, "learning_rate": 1.3412185010841358e-05, "loss": 3.7299, "step": 146650 }, { "epoch": 1.6481016497868253, "grad_norm": 0.9800334572792053, "learning_rate": 1.3409938097538508e-05, "loss": 3.6911, "step": 146700 }, { "epoch": 1.6486633749571684, "grad_norm": 1.2064788341522217, "learning_rate": 1.3407691184235659e-05, "loss": 3.6661, "step": 146750 }, { "epoch": 1.6492251001275116, "grad_norm": 1.5114845037460327, "learning_rate": 1.3405444270932808e-05, "loss": 3.67, "step": 146800 }, { "epoch": 1.6497868252978547, "grad_norm": 1.1322213411331177, "learning_rate": 1.3403197357629957e-05, "loss": 3.6547, "step": 146850 }, { "epoch": 1.6503485504681978, "grad_norm": 1.2612017393112183, "learning_rate": 1.3400950444327108e-05, "loss": 3.7193, "step": 146900 }, { "epoch": 1.650910275638541, "grad_norm": 1.2148751020431519, "learning_rate": 1.3398703531024258e-05, "loss": 3.6716, "step": 146950 }, { "epoch": 1.6514720008088841, "grad_norm": 1.3447052240371704, "learning_rate": 1.3396456617721407e-05, "loss": 3.6779, "step": 147000 }, { "epoch": 1.6520337259792273, "grad_norm": 1.0935733318328857, "learning_rate": 1.3394209704418556e-05, "loss": 3.6086, "step": 147050 }, { "epoch": 1.6525954511495704, "grad_norm": 1.6140910387039185, "learning_rate": 1.3392007729381762e-05, "loss": 3.7257, "step": 147100 }, { "epoch": 1.6531571763199138, "grad_norm": 1.2870099544525146, "learning_rate": 1.3389760816078912e-05, "loss": 3.6534, "step": 147150 }, { "epoch": 1.653718901490257, "grad_norm": 1.1969412565231323, "learning_rate": 1.3387513902776061e-05, "loss": 3.608, "step": 147200 }, { "epoch": 1.6542806266606, "grad_norm": 1.3302042484283447, "learning_rate": 1.3385266989473212e-05, "loss": 3.6281, "step": 147250 }, { "epoch": 1.6548423518309432, "grad_norm": 1.2988100051879883, "learning_rate": 1.3383020076170361e-05, "loss": 3.6913, "step": 147300 }, { "epoch": 1.6554040770012863, "grad_norm": 1.3460708856582642, "learning_rate": 1.338077316286751e-05, "loss": 3.659, "step": 147350 }, { "epoch": 1.6559658021716295, "grad_norm": 1.122812271118164, "learning_rate": 1.337852624956466e-05, "loss": 3.722, "step": 147400 }, { "epoch": 1.6565275273419726, "grad_norm": 1.312804937362671, "learning_rate": 1.337627933626181e-05, "loss": 3.6989, "step": 147450 }, { "epoch": 1.657089252512316, "grad_norm": 1.2745600938796997, "learning_rate": 1.337403242295896e-05, "loss": 3.7124, "step": 147500 }, { "epoch": 1.6576509776826591, "grad_norm": 1.1472702026367188, "learning_rate": 1.337178550965611e-05, "loss": 3.7748, "step": 147550 }, { "epoch": 1.6582127028530023, "grad_norm": 1.2738385200500488, "learning_rate": 1.336953859635326e-05, "loss": 3.7381, "step": 147600 }, { "epoch": 1.6587744280233454, "grad_norm": 1.1442394256591797, "learning_rate": 1.336729168305041e-05, "loss": 3.777, "step": 147650 }, { "epoch": 1.6593361531936885, "grad_norm": 1.043246865272522, "learning_rate": 1.3365044769747559e-05, "loss": 3.7133, "step": 147700 }, { "epoch": 1.6598978783640317, "grad_norm": 1.2269457578659058, "learning_rate": 1.3362797856444708e-05, "loss": 3.7473, "step": 147750 }, { "epoch": 1.6604596035343748, "grad_norm": 1.331426739692688, "learning_rate": 1.336055094314186e-05, "loss": 3.7658, "step": 147800 }, { "epoch": 1.661021328704718, "grad_norm": 1.048108458518982, "learning_rate": 1.3358304029839009e-05, "loss": 3.6972, "step": 147850 }, { "epoch": 1.6615830538750611, "grad_norm": 1.126671314239502, "learning_rate": 1.3356057116536158e-05, "loss": 3.643, "step": 147900 }, { "epoch": 1.6621447790454043, "grad_norm": 1.2158805131912231, "learning_rate": 1.3353810203233309e-05, "loss": 3.6418, "step": 147950 }, { "epoch": 1.6627065042157474, "grad_norm": 1.0428258180618286, "learning_rate": 1.3351563289930458e-05, "loss": 3.6805, "step": 148000 }, { "epoch": 1.6632682293860905, "grad_norm": 1.3096319437026978, "learning_rate": 1.3349316376627608e-05, "loss": 3.6689, "step": 148050 }, { "epoch": 1.6638299545564337, "grad_norm": 1.0405086278915405, "learning_rate": 1.3347069463324757e-05, "loss": 3.7214, "step": 148100 }, { "epoch": 1.6643916797267768, "grad_norm": 1.9383026361465454, "learning_rate": 1.3344822550021908e-05, "loss": 3.6636, "step": 148150 }, { "epoch": 1.66495340489712, "grad_norm": 1.1549087762832642, "learning_rate": 1.3342575636719057e-05, "loss": 3.6556, "step": 148200 }, { "epoch": 1.665515130067463, "grad_norm": 1.1671701669692993, "learning_rate": 1.3340328723416207e-05, "loss": 3.6536, "step": 148250 }, { "epoch": 1.6660768552378062, "grad_norm": 1.178333044052124, "learning_rate": 1.3338081810113358e-05, "loss": 3.6632, "step": 148300 }, { "epoch": 1.6666385804081494, "grad_norm": 1.2279329299926758, "learning_rate": 1.3335834896810507e-05, "loss": 3.7197, "step": 148350 }, { "epoch": 1.6672003055784925, "grad_norm": 1.4624500274658203, "learning_rate": 1.3333587983507656e-05, "loss": 3.6903, "step": 148400 }, { "epoch": 1.6677620307488357, "grad_norm": 1.1612122058868408, "learning_rate": 1.3331341070204806e-05, "loss": 3.6885, "step": 148450 }, { "epoch": 1.668323755919179, "grad_norm": 1.1029609441757202, "learning_rate": 1.3329094156901957e-05, "loss": 3.6618, "step": 148500 }, { "epoch": 1.6688854810895222, "grad_norm": 1.0497239828109741, "learning_rate": 1.3326847243599106e-05, "loss": 3.6376, "step": 148550 }, { "epoch": 1.6694472062598653, "grad_norm": 1.1928558349609375, "learning_rate": 1.3324600330296255e-05, "loss": 3.6673, "step": 148600 }, { "epoch": 1.6700089314302085, "grad_norm": 1.1223574876785278, "learning_rate": 1.3322353416993406e-05, "loss": 3.6929, "step": 148650 }, { "epoch": 1.6705706566005516, "grad_norm": 1.122258186340332, "learning_rate": 1.3320106503690556e-05, "loss": 3.6343, "step": 148700 }, { "epoch": 1.6711323817708947, "grad_norm": 1.1371307373046875, "learning_rate": 1.3317859590387705e-05, "loss": 3.5773, "step": 148750 }, { "epoch": 1.6716941069412379, "grad_norm": 1.0262322425842285, "learning_rate": 1.3315612677084854e-05, "loss": 3.7727, "step": 148800 }, { "epoch": 1.6722558321115812, "grad_norm": 1.0024133920669556, "learning_rate": 1.3313365763782005e-05, "loss": 3.7303, "step": 148850 }, { "epoch": 1.6728175572819244, "grad_norm": 1.2706538438796997, "learning_rate": 1.3311118850479155e-05, "loss": 3.7602, "step": 148900 }, { "epoch": 1.6733792824522675, "grad_norm": 1.0769323110580444, "learning_rate": 1.3308871937176304e-05, "loss": 3.6039, "step": 148950 }, { "epoch": 1.6739410076226107, "grad_norm": 0.9288907051086426, "learning_rate": 1.3306625023873453e-05, "loss": 3.6582, "step": 149000 }, { "epoch": 1.6745027327929538, "grad_norm": 1.0951144695281982, "learning_rate": 1.3304378110570604e-05, "loss": 3.6777, "step": 149050 }, { "epoch": 1.675064457963297, "grad_norm": 1.2050976753234863, "learning_rate": 1.3302131197267754e-05, "loss": 3.6267, "step": 149100 }, { "epoch": 1.67562618313364, "grad_norm": 1.0500702857971191, "learning_rate": 1.3299884283964903e-05, "loss": 3.6249, "step": 149150 }, { "epoch": 1.6761879083039832, "grad_norm": 1.0992419719696045, "learning_rate": 1.3297637370662054e-05, "loss": 3.7171, "step": 149200 }, { "epoch": 1.6767496334743264, "grad_norm": 1.1110466718673706, "learning_rate": 1.3295390457359203e-05, "loss": 3.6624, "step": 149250 }, { "epoch": 1.6773113586446695, "grad_norm": 1.0994672775268555, "learning_rate": 1.3293143544056352e-05, "loss": 3.6212, "step": 149300 }, { "epoch": 1.6778730838150127, "grad_norm": 1.1601837873458862, "learning_rate": 1.3290896630753502e-05, "loss": 3.5943, "step": 149350 }, { "epoch": 1.6784348089853558, "grad_norm": 1.0538126230239868, "learning_rate": 1.3288649717450653e-05, "loss": 3.7741, "step": 149400 }, { "epoch": 1.678996534155699, "grad_norm": 1.2923500537872314, "learning_rate": 1.3286402804147802e-05, "loss": 3.6315, "step": 149450 }, { "epoch": 1.679558259326042, "grad_norm": 1.4426004886627197, "learning_rate": 1.3284155890844951e-05, "loss": 3.6831, "step": 149500 }, { "epoch": 1.6801199844963852, "grad_norm": 1.1323589086532593, "learning_rate": 1.3281908977542102e-05, "loss": 3.6512, "step": 149550 }, { "epoch": 1.6806817096667284, "grad_norm": 1.238003134727478, "learning_rate": 1.3279662064239252e-05, "loss": 3.711, "step": 149600 }, { "epoch": 1.6812434348370715, "grad_norm": 1.192116379737854, "learning_rate": 1.3277415150936401e-05, "loss": 3.7677, "step": 149650 }, { "epoch": 1.6818051600074146, "grad_norm": 1.178747296333313, "learning_rate": 1.327516823763355e-05, "loss": 3.6192, "step": 149700 }, { "epoch": 1.6823668851777578, "grad_norm": 1.0546813011169434, "learning_rate": 1.3272921324330701e-05, "loss": 3.6233, "step": 149750 }, { "epoch": 1.682928610348101, "grad_norm": 1.1599066257476807, "learning_rate": 1.327067441102785e-05, "loss": 3.7189, "step": 149800 }, { "epoch": 1.6834903355184443, "grad_norm": 1.2025222778320312, "learning_rate": 1.3268427497725e-05, "loss": 3.6428, "step": 149850 }, { "epoch": 1.6840520606887874, "grad_norm": 1.1144520044326782, "learning_rate": 1.3266180584422151e-05, "loss": 3.6962, "step": 149900 }, { "epoch": 1.6846137858591306, "grad_norm": 1.1810729503631592, "learning_rate": 1.32639336711193e-05, "loss": 3.6183, "step": 149950 }, { "epoch": 1.6851755110294737, "grad_norm": 3.7173421382904053, "learning_rate": 1.326168675781645e-05, "loss": 3.6547, "step": 150000 }, { "epoch": 1.6857372361998169, "grad_norm": 1.180238962173462, "learning_rate": 1.3259439844513599e-05, "loss": 3.69, "step": 150050 }, { "epoch": 1.68629896137016, "grad_norm": 1.4270219802856445, "learning_rate": 1.325719293121075e-05, "loss": 3.6633, "step": 150100 }, { "epoch": 1.6868606865405031, "grad_norm": 1.2268694639205933, "learning_rate": 1.32549460179079e-05, "loss": 3.6613, "step": 150150 }, { "epoch": 1.6874224117108465, "grad_norm": 1.1365429162979126, "learning_rate": 1.3252699104605049e-05, "loss": 3.6612, "step": 150200 }, { "epoch": 1.6879841368811896, "grad_norm": 1.1073280572891235, "learning_rate": 1.32504521913022e-05, "loss": 3.5813, "step": 150250 }, { "epoch": 1.6885458620515328, "grad_norm": 1.258264422416687, "learning_rate": 1.3248205277999349e-05, "loss": 3.6542, "step": 150300 }, { "epoch": 1.689107587221876, "grad_norm": 1.1778461933135986, "learning_rate": 1.3245958364696498e-05, "loss": 3.6699, "step": 150350 }, { "epoch": 1.689669312392219, "grad_norm": 1.2182080745697021, "learning_rate": 1.3243711451393648e-05, "loss": 3.6747, "step": 150400 }, { "epoch": 1.6902310375625622, "grad_norm": 1.2940605878829956, "learning_rate": 1.3241464538090799e-05, "loss": 3.6862, "step": 150450 }, { "epoch": 1.6907927627329054, "grad_norm": 1.1755141019821167, "learning_rate": 1.3239217624787948e-05, "loss": 3.7816, "step": 150500 }, { "epoch": 1.6913544879032485, "grad_norm": 1.064827799797058, "learning_rate": 1.3236970711485097e-05, "loss": 3.6505, "step": 150550 }, { "epoch": 1.6919162130735916, "grad_norm": 1.185843825340271, "learning_rate": 1.3234723798182247e-05, "loss": 3.6531, "step": 150600 }, { "epoch": 1.6924779382439348, "grad_norm": 1.4173972606658936, "learning_rate": 1.3232476884879398e-05, "loss": 3.747, "step": 150650 }, { "epoch": 1.693039663414278, "grad_norm": 1.2260687351226807, "learning_rate": 1.3230229971576547e-05, "loss": 3.6408, "step": 150700 }, { "epoch": 1.693601388584621, "grad_norm": 1.0594072341918945, "learning_rate": 1.3227983058273696e-05, "loss": 3.6188, "step": 150750 }, { "epoch": 1.6941631137549642, "grad_norm": 1.0781843662261963, "learning_rate": 1.3225736144970847e-05, "loss": 3.6484, "step": 150800 }, { "epoch": 1.6947248389253073, "grad_norm": 1.171908974647522, "learning_rate": 1.3223489231667997e-05, "loss": 3.7537, "step": 150850 }, { "epoch": 1.6952865640956505, "grad_norm": 1.2834751605987549, "learning_rate": 1.3221242318365146e-05, "loss": 3.65, "step": 150900 }, { "epoch": 1.6958482892659936, "grad_norm": 1.113500952720642, "learning_rate": 1.3218995405062295e-05, "loss": 3.6374, "step": 150950 }, { "epoch": 1.6964100144363368, "grad_norm": 1.1537258625030518, "learning_rate": 1.3216748491759446e-05, "loss": 3.6043, "step": 151000 }, { "epoch": 1.69697173960668, "grad_norm": 1.1523864269256592, "learning_rate": 1.3214501578456596e-05, "loss": 3.6814, "step": 151050 }, { "epoch": 1.697533464777023, "grad_norm": 1.1361587047576904, "learning_rate": 1.3212254665153745e-05, "loss": 3.6928, "step": 151100 }, { "epoch": 1.6980951899473662, "grad_norm": 1.2203460931777954, "learning_rate": 1.3210007751850896e-05, "loss": 3.7051, "step": 151150 }, { "epoch": 1.6986569151177096, "grad_norm": 1.1210293769836426, "learning_rate": 1.3207760838548045e-05, "loss": 3.7301, "step": 151200 }, { "epoch": 1.6992186402880527, "grad_norm": 1.1691347360610962, "learning_rate": 1.3205513925245194e-05, "loss": 3.6296, "step": 151250 }, { "epoch": 1.6997803654583958, "grad_norm": 1.1081463098526, "learning_rate": 1.3203267011942344e-05, "loss": 3.7019, "step": 151300 }, { "epoch": 1.700342090628739, "grad_norm": 1.0898524522781372, "learning_rate": 1.3201020098639495e-05, "loss": 3.7002, "step": 151350 }, { "epoch": 1.7009038157990821, "grad_norm": 1.0776554346084595, "learning_rate": 1.3198773185336644e-05, "loss": 3.6315, "step": 151400 }, { "epoch": 1.7014655409694253, "grad_norm": 1.0310733318328857, "learning_rate": 1.3196526272033793e-05, "loss": 3.6213, "step": 151450 }, { "epoch": 1.7020272661397684, "grad_norm": 1.294753074645996, "learning_rate": 1.3194324296997001e-05, "loss": 3.7238, "step": 151500 }, { "epoch": 1.7025889913101118, "grad_norm": 1.391838788986206, "learning_rate": 1.319207738369415e-05, "loss": 3.7018, "step": 151550 }, { "epoch": 1.703150716480455, "grad_norm": 1.1529507637023926, "learning_rate": 1.31898304703913e-05, "loss": 3.6026, "step": 151600 }, { "epoch": 1.703712441650798, "grad_norm": 1.0914785861968994, "learning_rate": 1.318758355708845e-05, "loss": 3.7089, "step": 151650 }, { "epoch": 1.7042741668211412, "grad_norm": 1.33309805393219, "learning_rate": 1.31853366437856e-05, "loss": 3.6505, "step": 151700 }, { "epoch": 1.7048358919914843, "grad_norm": 1.1748844385147095, "learning_rate": 1.318308973048275e-05, "loss": 3.6378, "step": 151750 }, { "epoch": 1.7053976171618275, "grad_norm": 1.3727186918258667, "learning_rate": 1.31808428171799e-05, "loss": 3.7217, "step": 151800 }, { "epoch": 1.7059593423321706, "grad_norm": 1.2234445810317993, "learning_rate": 1.317859590387705e-05, "loss": 3.7484, "step": 151850 }, { "epoch": 1.7065210675025138, "grad_norm": 1.3959254026412964, "learning_rate": 1.3176348990574199e-05, "loss": 3.7239, "step": 151900 }, { "epoch": 1.707082792672857, "grad_norm": 1.5000845193862915, "learning_rate": 1.3174102077271348e-05, "loss": 3.6651, "step": 151950 }, { "epoch": 1.7076445178432, "grad_norm": 1.2455960512161255, "learning_rate": 1.31718551639685e-05, "loss": 3.6371, "step": 152000 }, { "epoch": 1.7082062430135432, "grad_norm": 1.1509108543395996, "learning_rate": 1.3169608250665649e-05, "loss": 3.695, "step": 152050 }, { "epoch": 1.7087679681838863, "grad_norm": 1.1868137121200562, "learning_rate": 1.3167361337362798e-05, "loss": 3.6248, "step": 152100 }, { "epoch": 1.7093296933542295, "grad_norm": 1.268827199935913, "learning_rate": 1.3165114424059949e-05, "loss": 3.6964, "step": 152150 }, { "epoch": 1.7098914185245726, "grad_norm": 1.41181480884552, "learning_rate": 1.3162867510757098e-05, "loss": 3.6524, "step": 152200 }, { "epoch": 1.7104531436949157, "grad_norm": 1.2418874502182007, "learning_rate": 1.3160620597454248e-05, "loss": 3.6359, "step": 152250 }, { "epoch": 1.7110148688652589, "grad_norm": 1.390811562538147, "learning_rate": 1.3158373684151397e-05, "loss": 3.6681, "step": 152300 }, { "epoch": 1.711576594035602, "grad_norm": 1.1755359172821045, "learning_rate": 1.3156126770848548e-05, "loss": 3.7523, "step": 152350 }, { "epoch": 1.7121383192059452, "grad_norm": 1.3521742820739746, "learning_rate": 1.3153879857545697e-05, "loss": 3.6319, "step": 152400 }, { "epoch": 1.7127000443762883, "grad_norm": 1.1433719396591187, "learning_rate": 1.3151632944242847e-05, "loss": 3.6684, "step": 152450 }, { "epoch": 1.7132617695466315, "grad_norm": 1.052834391593933, "learning_rate": 1.3149386030939998e-05, "loss": 3.63, "step": 152500 }, { "epoch": 1.7138234947169748, "grad_norm": 1.0887142419815063, "learning_rate": 1.3147184055903203e-05, "loss": 3.6899, "step": 152550 }, { "epoch": 1.714385219887318, "grad_norm": 1.1405411958694458, "learning_rate": 1.3144937142600353e-05, "loss": 3.7286, "step": 152600 }, { "epoch": 1.714946945057661, "grad_norm": 0.9536702632904053, "learning_rate": 1.3142690229297504e-05, "loss": 3.7611, "step": 152650 }, { "epoch": 1.7155086702280042, "grad_norm": 1.643053650856018, "learning_rate": 1.3140443315994653e-05, "loss": 3.6665, "step": 152700 }, { "epoch": 1.7160703953983474, "grad_norm": 1.278144121170044, "learning_rate": 1.3138196402691802e-05, "loss": 3.6664, "step": 152750 }, { "epoch": 1.7166321205686905, "grad_norm": 1.1933716535568237, "learning_rate": 1.3135949489388953e-05, "loss": 3.6371, "step": 152800 }, { "epoch": 1.7171938457390339, "grad_norm": 1.2203391790390015, "learning_rate": 1.3133702576086103e-05, "loss": 3.656, "step": 152850 }, { "epoch": 1.717755570909377, "grad_norm": 1.1443779468536377, "learning_rate": 1.3131455662783252e-05, "loss": 3.7061, "step": 152900 }, { "epoch": 1.7183172960797202, "grad_norm": 1.0043830871582031, "learning_rate": 1.3129208749480401e-05, "loss": 3.6858, "step": 152950 }, { "epoch": 1.7188790212500633, "grad_norm": 1.0541163682937622, "learning_rate": 1.3126961836177552e-05, "loss": 3.697, "step": 153000 }, { "epoch": 1.7194407464204065, "grad_norm": 1.1946184635162354, "learning_rate": 1.3124714922874702e-05, "loss": 3.7139, "step": 153050 }, { "epoch": 1.7200024715907496, "grad_norm": 1.1679470539093018, "learning_rate": 1.3122468009571851e-05, "loss": 3.6856, "step": 153100 }, { "epoch": 1.7205641967610927, "grad_norm": 1.140401005744934, "learning_rate": 1.3120221096269002e-05, "loss": 3.6471, "step": 153150 }, { "epoch": 1.7211259219314359, "grad_norm": 1.2564250230789185, "learning_rate": 1.3117974182966151e-05, "loss": 3.6139, "step": 153200 }, { "epoch": 1.721687647101779, "grad_norm": 1.1056748628616333, "learning_rate": 1.31157272696633e-05, "loss": 3.7062, "step": 153250 }, { "epoch": 1.7222493722721222, "grad_norm": 1.2699270248413086, "learning_rate": 1.311348035636045e-05, "loss": 3.6892, "step": 153300 }, { "epoch": 1.7228110974424653, "grad_norm": 1.2607805728912354, "learning_rate": 1.3111233443057601e-05, "loss": 3.6134, "step": 153350 }, { "epoch": 1.7233728226128084, "grad_norm": 1.2254953384399414, "learning_rate": 1.310898652975475e-05, "loss": 3.721, "step": 153400 }, { "epoch": 1.7239345477831516, "grad_norm": 1.3641060590744019, "learning_rate": 1.31067396164519e-05, "loss": 3.6985, "step": 153450 }, { "epoch": 1.7244962729534947, "grad_norm": 1.052706003189087, "learning_rate": 1.310449270314905e-05, "loss": 3.7281, "step": 153500 }, { "epoch": 1.7250579981238379, "grad_norm": 1.120184063911438, "learning_rate": 1.31022457898462e-05, "loss": 3.6994, "step": 153550 }, { "epoch": 1.725619723294181, "grad_norm": 1.2567594051361084, "learning_rate": 1.309999887654335e-05, "loss": 3.6809, "step": 153600 }, { "epoch": 1.7261814484645241, "grad_norm": 1.1006348133087158, "learning_rate": 1.3097751963240499e-05, "loss": 3.689, "step": 153650 }, { "epoch": 1.7267431736348673, "grad_norm": 1.1645714044570923, "learning_rate": 1.309550504993765e-05, "loss": 3.6636, "step": 153700 }, { "epoch": 1.7273048988052104, "grad_norm": 1.1966338157653809, "learning_rate": 1.3093258136634799e-05, "loss": 3.6868, "step": 153750 }, { "epoch": 1.7278666239755536, "grad_norm": 1.1202161312103271, "learning_rate": 1.3091011223331948e-05, "loss": 3.6664, "step": 153800 }, { "epoch": 1.7284283491458967, "grad_norm": 1.236364483833313, "learning_rate": 1.30887643100291e-05, "loss": 3.5549, "step": 153850 }, { "epoch": 1.72899007431624, "grad_norm": 1.3125675916671753, "learning_rate": 1.3086517396726249e-05, "loss": 3.6765, "step": 153900 }, { "epoch": 1.7295517994865832, "grad_norm": 1.2397410869598389, "learning_rate": 1.3084270483423398e-05, "loss": 3.6706, "step": 153950 }, { "epoch": 1.7301135246569264, "grad_norm": 1.2448431253433228, "learning_rate": 1.3082023570120547e-05, "loss": 3.7243, "step": 154000 }, { "epoch": 1.7306752498272695, "grad_norm": 1.0740070343017578, "learning_rate": 1.3079776656817698e-05, "loss": 3.6927, "step": 154050 }, { "epoch": 1.7312369749976126, "grad_norm": 1.0397754907608032, "learning_rate": 1.3077529743514848e-05, "loss": 3.6253, "step": 154100 }, { "epoch": 1.7317987001679558, "grad_norm": 0.9703950881958008, "learning_rate": 1.3075282830211997e-05, "loss": 3.6631, "step": 154150 }, { "epoch": 1.7323604253382991, "grad_norm": 1.2280876636505127, "learning_rate": 1.3073035916909146e-05, "loss": 3.6468, "step": 154200 }, { "epoch": 1.7329221505086423, "grad_norm": 1.0809295177459717, "learning_rate": 1.3070789003606297e-05, "loss": 3.6998, "step": 154250 }, { "epoch": 1.7334838756789854, "grad_norm": 1.2361363172531128, "learning_rate": 1.3068542090303446e-05, "loss": 3.6474, "step": 154300 }, { "epoch": 1.7340456008493286, "grad_norm": 1.0839402675628662, "learning_rate": 1.3066295177000596e-05, "loss": 3.6696, "step": 154350 }, { "epoch": 1.7346073260196717, "grad_norm": 1.2606266736984253, "learning_rate": 1.3064048263697747e-05, "loss": 3.6083, "step": 154400 }, { "epoch": 1.7351690511900149, "grad_norm": 1.069074034690857, "learning_rate": 1.3061801350394896e-05, "loss": 3.625, "step": 154450 }, { "epoch": 1.735730776360358, "grad_norm": 1.063657283782959, "learning_rate": 1.3059554437092045e-05, "loss": 3.6952, "step": 154500 }, { "epoch": 1.7362925015307011, "grad_norm": 1.1664756536483765, "learning_rate": 1.3057307523789195e-05, "loss": 3.6318, "step": 154550 }, { "epoch": 1.7368542267010443, "grad_norm": 1.219751238822937, "learning_rate": 1.3055060610486346e-05, "loss": 3.727, "step": 154600 }, { "epoch": 1.7374159518713874, "grad_norm": 1.2343968152999878, "learning_rate": 1.3052813697183495e-05, "loss": 3.6172, "step": 154650 }, { "epoch": 1.7379776770417306, "grad_norm": 1.1121562719345093, "learning_rate": 1.3050566783880644e-05, "loss": 3.6709, "step": 154700 }, { "epoch": 1.7385394022120737, "grad_norm": 1.5850425958633423, "learning_rate": 1.3048319870577795e-05, "loss": 3.7197, "step": 154750 }, { "epoch": 1.7391011273824168, "grad_norm": 1.3235431909561157, "learning_rate": 1.3046072957274945e-05, "loss": 3.6213, "step": 154800 }, { "epoch": 1.73966285255276, "grad_norm": 1.1664890050888062, "learning_rate": 1.3043826043972094e-05, "loss": 3.6559, "step": 154850 }, { "epoch": 1.7402245777231031, "grad_norm": 1.647775411605835, "learning_rate": 1.3041579130669243e-05, "loss": 3.64, "step": 154900 }, { "epoch": 1.7407863028934463, "grad_norm": 0.9884029626846313, "learning_rate": 1.3039332217366394e-05, "loss": 3.6411, "step": 154950 }, { "epoch": 1.7413480280637894, "grad_norm": 1.123098611831665, "learning_rate": 1.3037085304063544e-05, "loss": 3.7255, "step": 155000 }, { "epoch": 1.7419097532341326, "grad_norm": 1.176184892654419, "learning_rate": 1.3034883329026751e-05, "loss": 3.7112, "step": 155050 }, { "epoch": 1.7424714784044757, "grad_norm": 1.0983113050460815, "learning_rate": 1.30326364157239e-05, "loss": 3.6281, "step": 155100 }, { "epoch": 1.7430332035748188, "grad_norm": 1.1853564977645874, "learning_rate": 1.303038950242105e-05, "loss": 3.6697, "step": 155150 }, { "epoch": 1.743594928745162, "grad_norm": 1.1369297504425049, "learning_rate": 1.30281425891182e-05, "loss": 3.6517, "step": 155200 }, { "epoch": 1.7441566539155053, "grad_norm": 1.1019175052642822, "learning_rate": 1.302589567581535e-05, "loss": 3.6663, "step": 155250 }, { "epoch": 1.7447183790858485, "grad_norm": 1.0535612106323242, "learning_rate": 1.30236487625125e-05, "loss": 3.6439, "step": 155300 }, { "epoch": 1.7452801042561916, "grad_norm": 1.1677604913711548, "learning_rate": 1.3021401849209649e-05, "loss": 3.707, "step": 155350 }, { "epoch": 1.7458418294265348, "grad_norm": 1.1271241903305054, "learning_rate": 1.30191549359068e-05, "loss": 3.7345, "step": 155400 }, { "epoch": 1.746403554596878, "grad_norm": 1.1681445837020874, "learning_rate": 1.301690802260395e-05, "loss": 3.6706, "step": 155450 }, { "epoch": 1.746965279767221, "grad_norm": 1.0617426633834839, "learning_rate": 1.3014661109301099e-05, "loss": 3.6912, "step": 155500 }, { "epoch": 1.7475270049375644, "grad_norm": 1.1394972801208496, "learning_rate": 1.3012414195998248e-05, "loss": 3.5982, "step": 155550 }, { "epoch": 1.7480887301079076, "grad_norm": 1.078129529953003, "learning_rate": 1.3010167282695399e-05, "loss": 3.6903, "step": 155600 }, { "epoch": 1.7486504552782507, "grad_norm": 1.2922347784042358, "learning_rate": 1.3007920369392548e-05, "loss": 3.6904, "step": 155650 }, { "epoch": 1.7492121804485938, "grad_norm": 1.561508297920227, "learning_rate": 1.3005673456089698e-05, "loss": 3.6604, "step": 155700 }, { "epoch": 1.749773905618937, "grad_norm": 1.1499099731445312, "learning_rate": 1.3003426542786849e-05, "loss": 3.6863, "step": 155750 }, { "epoch": 1.7503356307892801, "grad_norm": 1.0464122295379639, "learning_rate": 1.3001179629483998e-05, "loss": 3.6086, "step": 155800 }, { "epoch": 1.7508973559596233, "grad_norm": 1.1842803955078125, "learning_rate": 1.2998932716181147e-05, "loss": 3.653, "step": 155850 }, { "epoch": 1.7514590811299664, "grad_norm": 1.3405252695083618, "learning_rate": 1.2996685802878296e-05, "loss": 3.699, "step": 155900 }, { "epoch": 1.7520208063003095, "grad_norm": 1.129227876663208, "learning_rate": 1.2994438889575447e-05, "loss": 3.6578, "step": 155950 }, { "epoch": 1.7525825314706527, "grad_norm": 1.137797236442566, "learning_rate": 1.2992191976272597e-05, "loss": 3.7331, "step": 156000 }, { "epoch": 1.7531442566409958, "grad_norm": 1.041616439819336, "learning_rate": 1.2989945062969746e-05, "loss": 3.6837, "step": 156050 }, { "epoch": 1.753705981811339, "grad_norm": 1.1781810522079468, "learning_rate": 1.2987698149666897e-05, "loss": 3.635, "step": 156100 }, { "epoch": 1.754267706981682, "grad_norm": 1.1395310163497925, "learning_rate": 1.2985451236364046e-05, "loss": 3.7524, "step": 156150 }, { "epoch": 1.7548294321520252, "grad_norm": 1.4573463201522827, "learning_rate": 1.2983204323061196e-05, "loss": 3.7249, "step": 156200 }, { "epoch": 1.7553911573223684, "grad_norm": 1.1571025848388672, "learning_rate": 1.2980957409758345e-05, "loss": 3.7074, "step": 156250 }, { "epoch": 1.7559528824927115, "grad_norm": 1.0512663125991821, "learning_rate": 1.2978710496455496e-05, "loss": 3.6184, "step": 156300 }, { "epoch": 1.7565146076630547, "grad_norm": 1.0732377767562866, "learning_rate": 1.2976463583152645e-05, "loss": 3.6676, "step": 156350 }, { "epoch": 1.7570763328333978, "grad_norm": 1.127360463142395, "learning_rate": 1.2974216669849795e-05, "loss": 3.6764, "step": 156400 }, { "epoch": 1.757638058003741, "grad_norm": 1.111916184425354, "learning_rate": 1.2971969756546946e-05, "loss": 3.6875, "step": 156450 }, { "epoch": 1.758199783174084, "grad_norm": 1.1683317422866821, "learning_rate": 1.2969722843244095e-05, "loss": 3.7127, "step": 156500 }, { "epoch": 1.7587615083444275, "grad_norm": 1.1425973176956177, "learning_rate": 1.2967520868207301e-05, "loss": 3.638, "step": 156550 }, { "epoch": 1.7593232335147706, "grad_norm": 1.2834800481796265, "learning_rate": 1.2965273954904452e-05, "loss": 3.6476, "step": 156600 }, { "epoch": 1.7598849586851137, "grad_norm": 1.1782361268997192, "learning_rate": 1.2963027041601601e-05, "loss": 3.6163, "step": 156650 }, { "epoch": 1.7604466838554569, "grad_norm": 1.069303035736084, "learning_rate": 1.296078012829875e-05, "loss": 3.7065, "step": 156700 }, { "epoch": 1.7610084090258, "grad_norm": 1.0960201025009155, "learning_rate": 1.2958533214995902e-05, "loss": 3.678, "step": 156750 }, { "epoch": 1.7615701341961432, "grad_norm": 1.3260595798492432, "learning_rate": 1.2956286301693051e-05, "loss": 3.6791, "step": 156800 }, { "epoch": 1.7621318593664863, "grad_norm": 1.1478679180145264, "learning_rate": 1.29540393883902e-05, "loss": 3.6592, "step": 156850 }, { "epoch": 1.7626935845368297, "grad_norm": 1.0272244215011597, "learning_rate": 1.295179247508735e-05, "loss": 3.6611, "step": 156900 }, { "epoch": 1.7632553097071728, "grad_norm": 1.1543618440628052, "learning_rate": 1.29495455617845e-05, "loss": 3.6443, "step": 156950 }, { "epoch": 1.763817034877516, "grad_norm": 1.275540828704834, "learning_rate": 1.294729864848165e-05, "loss": 3.7771, "step": 157000 }, { "epoch": 1.764378760047859, "grad_norm": 1.168523907661438, "learning_rate": 1.29450517351788e-05, "loss": 3.6167, "step": 157050 }, { "epoch": 1.7649404852182022, "grad_norm": 0.9482042193412781, "learning_rate": 1.294280482187595e-05, "loss": 3.6669, "step": 157100 }, { "epoch": 1.7655022103885454, "grad_norm": 1.0952187776565552, "learning_rate": 1.29405579085731e-05, "loss": 3.682, "step": 157150 }, { "epoch": 1.7660639355588885, "grad_norm": 1.2059918642044067, "learning_rate": 1.2938310995270249e-05, "loss": 3.5579, "step": 157200 }, { "epoch": 1.7666256607292317, "grad_norm": 1.1145679950714111, "learning_rate": 1.2936064081967398e-05, "loss": 3.6652, "step": 157250 }, { "epoch": 1.7671873858995748, "grad_norm": 1.1019964218139648, "learning_rate": 1.293381716866455e-05, "loss": 3.6714, "step": 157300 }, { "epoch": 1.767749111069918, "grad_norm": 1.5606224536895752, "learning_rate": 1.2931570255361699e-05, "loss": 3.6421, "step": 157350 }, { "epoch": 1.768310836240261, "grad_norm": 1.1179559230804443, "learning_rate": 1.2929323342058848e-05, "loss": 3.6624, "step": 157400 }, { "epoch": 1.7688725614106042, "grad_norm": 1.2087434530258179, "learning_rate": 1.2927076428755999e-05, "loss": 3.6901, "step": 157450 }, { "epoch": 1.7694342865809474, "grad_norm": 1.0996601581573486, "learning_rate": 1.2924829515453148e-05, "loss": 3.6827, "step": 157500 }, { "epoch": 1.7699960117512905, "grad_norm": 1.1604089736938477, "learning_rate": 1.2922582602150297e-05, "loss": 3.6265, "step": 157550 }, { "epoch": 1.7705577369216337, "grad_norm": 1.0635164976119995, "learning_rate": 1.2920335688847447e-05, "loss": 3.7781, "step": 157600 }, { "epoch": 1.7711194620919768, "grad_norm": 1.26372230052948, "learning_rate": 1.2918088775544598e-05, "loss": 3.6933, "step": 157650 }, { "epoch": 1.77168118726232, "grad_norm": 1.1672794818878174, "learning_rate": 1.2915841862241747e-05, "loss": 3.6697, "step": 157700 }, { "epoch": 1.772242912432663, "grad_norm": 1.0433988571166992, "learning_rate": 1.2913594948938896e-05, "loss": 3.6325, "step": 157750 }, { "epoch": 1.7728046376030062, "grad_norm": 1.1019052267074585, "learning_rate": 1.2911348035636046e-05, "loss": 3.6513, "step": 157800 }, { "epoch": 1.7733663627733494, "grad_norm": 1.1907793283462524, "learning_rate": 1.2909101122333197e-05, "loss": 3.7005, "step": 157850 }, { "epoch": 1.7739280879436927, "grad_norm": 1.1211351156234741, "learning_rate": 1.2906854209030346e-05, "loss": 3.7068, "step": 157900 }, { "epoch": 1.7744898131140359, "grad_norm": 1.1437180042266846, "learning_rate": 1.2904607295727495e-05, "loss": 3.693, "step": 157950 }, { "epoch": 1.775051538284379, "grad_norm": 1.1743581295013428, "learning_rate": 1.2902360382424646e-05, "loss": 3.6189, "step": 158000 }, { "epoch": 1.7756132634547221, "grad_norm": 1.1614916324615479, "learning_rate": 1.2900113469121796e-05, "loss": 3.7102, "step": 158050 }, { "epoch": 1.7761749886250653, "grad_norm": 1.1317543983459473, "learning_rate": 1.2897866555818945e-05, "loss": 3.6501, "step": 158100 }, { "epoch": 1.7767367137954084, "grad_norm": 1.259522557258606, "learning_rate": 1.2895664580782153e-05, "loss": 3.7159, "step": 158150 }, { "epoch": 1.7772984389657516, "grad_norm": 1.4388420581817627, "learning_rate": 1.2893417667479302e-05, "loss": 3.6617, "step": 158200 }, { "epoch": 1.777860164136095, "grad_norm": 1.279236912727356, "learning_rate": 1.2891170754176451e-05, "loss": 3.751, "step": 158250 }, { "epoch": 1.778421889306438, "grad_norm": 1.0801838636398315, "learning_rate": 1.2888923840873602e-05, "loss": 3.6969, "step": 158300 }, { "epoch": 1.7789836144767812, "grad_norm": 1.2110668420791626, "learning_rate": 1.2886676927570752e-05, "loss": 3.6743, "step": 158350 }, { "epoch": 1.7795453396471244, "grad_norm": 1.1318589448928833, "learning_rate": 1.2884430014267901e-05, "loss": 3.6516, "step": 158400 }, { "epoch": 1.7801070648174675, "grad_norm": 1.0363426208496094, "learning_rate": 1.2882183100965052e-05, "loss": 3.6595, "step": 158450 }, { "epoch": 1.7806687899878106, "grad_norm": 1.2623485326766968, "learning_rate": 1.2879936187662201e-05, "loss": 3.6201, "step": 158500 }, { "epoch": 1.7812305151581538, "grad_norm": 1.0081738233566284, "learning_rate": 1.287768927435935e-05, "loss": 3.7067, "step": 158550 }, { "epoch": 1.781792240328497, "grad_norm": 1.1923474073410034, "learning_rate": 1.28754423610565e-05, "loss": 3.6912, "step": 158600 }, { "epoch": 1.78235396549884, "grad_norm": 1.139208436012268, "learning_rate": 1.2873195447753651e-05, "loss": 3.6359, "step": 158650 }, { "epoch": 1.7829156906691832, "grad_norm": 1.2712637186050415, "learning_rate": 1.28709485344508e-05, "loss": 3.7443, "step": 158700 }, { "epoch": 1.7834774158395263, "grad_norm": 1.1234378814697266, "learning_rate": 1.286870162114795e-05, "loss": 3.6251, "step": 158750 }, { "epoch": 1.7840391410098695, "grad_norm": 1.275105595588684, "learning_rate": 1.2866454707845099e-05, "loss": 3.6907, "step": 158800 }, { "epoch": 1.7846008661802126, "grad_norm": 1.3843269348144531, "learning_rate": 1.286420779454225e-05, "loss": 3.6619, "step": 158850 }, { "epoch": 1.7851625913505558, "grad_norm": 1.107105016708374, "learning_rate": 1.28619608812394e-05, "loss": 3.7125, "step": 158900 }, { "epoch": 1.785724316520899, "grad_norm": 1.0981186628341675, "learning_rate": 1.2859713967936548e-05, "loss": 3.6826, "step": 158950 }, { "epoch": 1.786286041691242, "grad_norm": 1.2366962432861328, "learning_rate": 1.28574670546337e-05, "loss": 3.6565, "step": 159000 }, { "epoch": 1.7868477668615852, "grad_norm": 1.1007949113845825, "learning_rate": 1.2855220141330849e-05, "loss": 3.7028, "step": 159050 }, { "epoch": 1.7874094920319283, "grad_norm": 1.1660207509994507, "learning_rate": 1.2852973228027998e-05, "loss": 3.6584, "step": 159100 }, { "epoch": 1.7879712172022715, "grad_norm": 1.135389804840088, "learning_rate": 1.2850726314725147e-05, "loss": 3.6954, "step": 159150 }, { "epoch": 1.7885329423726146, "grad_norm": 1.2505366802215576, "learning_rate": 1.2848479401422298e-05, "loss": 3.7322, "step": 159200 }, { "epoch": 1.789094667542958, "grad_norm": 1.289759874343872, "learning_rate": 1.2846232488119448e-05, "loss": 3.7352, "step": 159250 }, { "epoch": 1.7896563927133011, "grad_norm": 1.5705466270446777, "learning_rate": 1.2843985574816597e-05, "loss": 3.6966, "step": 159300 }, { "epoch": 1.7902181178836443, "grad_norm": 1.1748906373977661, "learning_rate": 1.2841738661513748e-05, "loss": 3.6485, "step": 159350 }, { "epoch": 1.7907798430539874, "grad_norm": 1.2288804054260254, "learning_rate": 1.2839491748210897e-05, "loss": 3.7492, "step": 159400 }, { "epoch": 1.7913415682243305, "grad_norm": 1.230589747428894, "learning_rate": 1.2837244834908047e-05, "loss": 3.7004, "step": 159450 }, { "epoch": 1.7919032933946737, "grad_norm": 1.1219701766967773, "learning_rate": 1.2834997921605196e-05, "loss": 3.6086, "step": 159500 }, { "epoch": 1.7924650185650168, "grad_norm": 1.1636079549789429, "learning_rate": 1.2832751008302347e-05, "loss": 3.6157, "step": 159550 }, { "epoch": 1.7930267437353602, "grad_norm": 1.3621783256530762, "learning_rate": 1.2830504094999496e-05, "loss": 3.6461, "step": 159600 }, { "epoch": 1.7935884689057033, "grad_norm": 1.0143169164657593, "learning_rate": 1.2828257181696646e-05, "loss": 3.7079, "step": 159650 }, { "epoch": 1.7941501940760465, "grad_norm": 1.0718719959259033, "learning_rate": 1.2826010268393797e-05, "loss": 3.6971, "step": 159700 }, { "epoch": 1.7947119192463896, "grad_norm": 0.9802167415618896, "learning_rate": 1.2823763355090946e-05, "loss": 3.6349, "step": 159750 }, { "epoch": 1.7952736444167328, "grad_norm": 1.4045599699020386, "learning_rate": 1.2821516441788095e-05, "loss": 3.7373, "step": 159800 }, { "epoch": 1.795835369587076, "grad_norm": 1.0568664073944092, "learning_rate": 1.2819269528485245e-05, "loss": 3.6243, "step": 159850 }, { "epoch": 1.796397094757419, "grad_norm": 1.1222189664840698, "learning_rate": 1.2817022615182396e-05, "loss": 3.6887, "step": 159900 }, { "epoch": 1.7969588199277622, "grad_norm": 1.0660656690597534, "learning_rate": 1.2814775701879545e-05, "loss": 3.6987, "step": 159950 }, { "epoch": 1.7975205450981053, "grad_norm": 1.1978665590286255, "learning_rate": 1.2812528788576694e-05, "loss": 3.7301, "step": 160000 }, { "epoch": 1.7980822702684485, "grad_norm": 1.1811360120773315, "learning_rate": 1.2810281875273845e-05, "loss": 3.6331, "step": 160050 }, { "epoch": 1.7986439954387916, "grad_norm": 1.0400501489639282, "learning_rate": 1.2808034961970995e-05, "loss": 3.6449, "step": 160100 }, { "epoch": 1.7992057206091348, "grad_norm": 1.3442927598953247, "learning_rate": 1.2805788048668144e-05, "loss": 3.6649, "step": 160150 }, { "epoch": 1.799767445779478, "grad_norm": 1.1856480836868286, "learning_rate": 1.2803541135365293e-05, "loss": 3.7003, "step": 160200 }, { "epoch": 1.800329170949821, "grad_norm": 1.2278929948806763, "learning_rate": 1.2801294222062444e-05, "loss": 3.6596, "step": 160250 }, { "epoch": 1.8008908961201642, "grad_norm": 1.3022801876068115, "learning_rate": 1.2799047308759594e-05, "loss": 3.6942, "step": 160300 }, { "epoch": 1.8014526212905073, "grad_norm": 1.1130034923553467, "learning_rate": 1.2796800395456743e-05, "loss": 3.6764, "step": 160350 }, { "epoch": 1.8020143464608505, "grad_norm": 1.1749629974365234, "learning_rate": 1.2794553482153892e-05, "loss": 3.6354, "step": 160400 }, { "epoch": 1.8025760716311936, "grad_norm": 1.0336318016052246, "learning_rate": 1.2792306568851043e-05, "loss": 3.6024, "step": 160450 }, { "epoch": 1.8031377968015367, "grad_norm": 1.0810723304748535, "learning_rate": 1.2790059655548193e-05, "loss": 3.602, "step": 160500 }, { "epoch": 1.8036995219718799, "grad_norm": 1.1231781244277954, "learning_rate": 1.2787812742245342e-05, "loss": 3.6974, "step": 160550 }, { "epoch": 1.8042612471422232, "grad_norm": 1.1039625406265259, "learning_rate": 1.2785565828942493e-05, "loss": 3.6196, "step": 160600 }, { "epoch": 1.8048229723125664, "grad_norm": 1.094585657119751, "learning_rate": 1.2783318915639642e-05, "loss": 3.6405, "step": 160650 }, { "epoch": 1.8053846974829095, "grad_norm": 1.246232271194458, "learning_rate": 1.2781072002336792e-05, "loss": 3.6121, "step": 160700 }, { "epoch": 1.8059464226532527, "grad_norm": 1.2481576204299927, "learning_rate": 1.277882508903394e-05, "loss": 3.5926, "step": 160750 }, { "epoch": 1.8065081478235958, "grad_norm": 1.3390722274780273, "learning_rate": 1.2776578175731092e-05, "loss": 3.7399, "step": 160800 }, { "epoch": 1.807069872993939, "grad_norm": 1.299284815788269, "learning_rate": 1.2774331262428241e-05, "loss": 3.6092, "step": 160850 }, { "epoch": 1.807631598164282, "grad_norm": 1.2629300355911255, "learning_rate": 1.277208434912539e-05, "loss": 3.6756, "step": 160900 }, { "epoch": 1.8081933233346255, "grad_norm": 1.2889231443405151, "learning_rate": 1.2769837435822541e-05, "loss": 3.6702, "step": 160950 }, { "epoch": 1.8087550485049686, "grad_norm": 1.0715359449386597, "learning_rate": 1.276759052251969e-05, "loss": 3.6998, "step": 161000 }, { "epoch": 1.8093167736753117, "grad_norm": 1.08761465549469, "learning_rate": 1.276534360921684e-05, "loss": 3.667, "step": 161050 }, { "epoch": 1.8098784988456549, "grad_norm": 1.1700549125671387, "learning_rate": 1.276309669591399e-05, "loss": 3.659, "step": 161100 }, { "epoch": 1.810440224015998, "grad_norm": 1.126656174659729, "learning_rate": 1.276084978261114e-05, "loss": 3.7039, "step": 161150 }, { "epoch": 1.8110019491863412, "grad_norm": 1.1791270971298218, "learning_rate": 1.275860286930829e-05, "loss": 3.6929, "step": 161200 }, { "epoch": 1.8115636743566843, "grad_norm": 1.0086944103240967, "learning_rate": 1.2756355956005439e-05, "loss": 3.7241, "step": 161250 }, { "epoch": 1.8121253995270274, "grad_norm": 1.1907039880752563, "learning_rate": 1.275410904270259e-05, "loss": 3.6694, "step": 161300 }, { "epoch": 1.8126871246973706, "grad_norm": 1.202735185623169, "learning_rate": 1.275186212939974e-05, "loss": 3.6575, "step": 161350 }, { "epoch": 1.8132488498677137, "grad_norm": 1.0789495706558228, "learning_rate": 1.2749615216096889e-05, "loss": 3.621, "step": 161400 }, { "epoch": 1.8138105750380569, "grad_norm": 1.6504883766174316, "learning_rate": 1.2747368302794038e-05, "loss": 3.5736, "step": 161450 }, { "epoch": 1.8143723002084, "grad_norm": 1.1836166381835938, "learning_rate": 1.2745121389491189e-05, "loss": 3.7147, "step": 161500 }, { "epoch": 1.8149340253787432, "grad_norm": 1.2792590856552124, "learning_rate": 1.2742874476188338e-05, "loss": 3.6766, "step": 161550 }, { "epoch": 1.8154957505490863, "grad_norm": 1.2648366689682007, "learning_rate": 1.2740627562885488e-05, "loss": 3.714, "step": 161600 }, { "epoch": 1.8160574757194294, "grad_norm": 1.0378963947296143, "learning_rate": 1.2738380649582639e-05, "loss": 3.6357, "step": 161650 }, { "epoch": 1.8166192008897726, "grad_norm": 1.0107951164245605, "learning_rate": 1.2736133736279788e-05, "loss": 3.7196, "step": 161700 }, { "epoch": 1.8171809260601157, "grad_norm": 1.324353814125061, "learning_rate": 1.2733886822976937e-05, "loss": 3.6357, "step": 161750 }, { "epoch": 1.8177426512304589, "grad_norm": 1.1754910945892334, "learning_rate": 1.2731639909674087e-05, "loss": 3.6433, "step": 161800 }, { "epoch": 1.818304376400802, "grad_norm": 1.1803934574127197, "learning_rate": 1.2729392996371238e-05, "loss": 3.6711, "step": 161850 }, { "epoch": 1.8188661015711451, "grad_norm": 1.0327410697937012, "learning_rate": 1.2727146083068387e-05, "loss": 3.7195, "step": 161900 }, { "epoch": 1.8194278267414885, "grad_norm": 1.205920934677124, "learning_rate": 1.2724899169765536e-05, "loss": 3.6471, "step": 161950 }, { "epoch": 1.8199895519118316, "grad_norm": 0.9720444083213806, "learning_rate": 1.2722652256462686e-05, "loss": 3.6269, "step": 162000 }, { "epoch": 1.8205512770821748, "grad_norm": 1.2071832418441772, "learning_rate": 1.2720405343159837e-05, "loss": 3.6919, "step": 162050 }, { "epoch": 1.821113002252518, "grad_norm": 1.2657171487808228, "learning_rate": 1.2718158429856986e-05, "loss": 3.6709, "step": 162100 }, { "epoch": 1.821674727422861, "grad_norm": 1.1440403461456299, "learning_rate": 1.2715911516554135e-05, "loss": 3.705, "step": 162150 }, { "epoch": 1.8222364525932042, "grad_norm": 1.1335704326629639, "learning_rate": 1.2713664603251286e-05, "loss": 3.6568, "step": 162200 }, { "epoch": 1.8227981777635474, "grad_norm": 1.2695765495300293, "learning_rate": 1.2711417689948436e-05, "loss": 3.6917, "step": 162250 }, { "epoch": 1.8233599029338907, "grad_norm": 1.1169294118881226, "learning_rate": 1.2709170776645585e-05, "loss": 3.6422, "step": 162300 }, { "epoch": 1.8239216281042339, "grad_norm": 1.1687711477279663, "learning_rate": 1.2706923863342734e-05, "loss": 3.7103, "step": 162350 }, { "epoch": 1.824483353274577, "grad_norm": 1.162584662437439, "learning_rate": 1.2704676950039885e-05, "loss": 3.6889, "step": 162400 }, { "epoch": 1.8250450784449201, "grad_norm": 1.252883791923523, "learning_rate": 1.2702430036737035e-05, "loss": 3.6662, "step": 162450 }, { "epoch": 1.8256068036152633, "grad_norm": 1.131441593170166, "learning_rate": 1.2700183123434184e-05, "loss": 3.7217, "step": 162500 }, { "epoch": 1.8261685287856064, "grad_norm": 1.153075098991394, "learning_rate": 1.2697936210131335e-05, "loss": 3.6241, "step": 162550 }, { "epoch": 1.8267302539559496, "grad_norm": 1.070625901222229, "learning_rate": 1.2695689296828484e-05, "loss": 3.6846, "step": 162600 }, { "epoch": 1.8272919791262927, "grad_norm": 1.016002893447876, "learning_rate": 1.2693442383525633e-05, "loss": 3.6648, "step": 162650 }, { "epoch": 1.8278537042966359, "grad_norm": 1.0782279968261719, "learning_rate": 1.2691195470222783e-05, "loss": 3.6122, "step": 162700 }, { "epoch": 1.828415429466979, "grad_norm": 1.3411065340042114, "learning_rate": 1.2688948556919934e-05, "loss": 3.6722, "step": 162750 }, { "epoch": 1.8289771546373221, "grad_norm": 0.9754913449287415, "learning_rate": 1.2686701643617083e-05, "loss": 3.6901, "step": 162800 }, { "epoch": 1.8295388798076653, "grad_norm": 1.3152989149093628, "learning_rate": 1.2684454730314232e-05, "loss": 3.7504, "step": 162850 }, { "epoch": 1.8301006049780084, "grad_norm": 1.2586301565170288, "learning_rate": 1.2682207817011383e-05, "loss": 3.6349, "step": 162900 }, { "epoch": 1.8306623301483516, "grad_norm": 1.1225285530090332, "learning_rate": 1.2679960903708533e-05, "loss": 3.7205, "step": 162950 }, { "epoch": 1.8312240553186947, "grad_norm": 1.2042137384414673, "learning_rate": 1.2677713990405682e-05, "loss": 3.613, "step": 163000 }, { "epoch": 1.8317857804890378, "grad_norm": 1.0179029703140259, "learning_rate": 1.2675467077102831e-05, "loss": 3.7395, "step": 163050 }, { "epoch": 1.832347505659381, "grad_norm": 1.2871730327606201, "learning_rate": 1.2673220163799982e-05, "loss": 3.6364, "step": 163100 }, { "epoch": 1.8329092308297241, "grad_norm": 0.9877760410308838, "learning_rate": 1.2670973250497132e-05, "loss": 3.6152, "step": 163150 }, { "epoch": 1.8334709560000673, "grad_norm": 1.0872858762741089, "learning_rate": 1.2668726337194281e-05, "loss": 3.6376, "step": 163200 }, { "epoch": 1.8340326811704104, "grad_norm": 1.5374056100845337, "learning_rate": 1.2666479423891432e-05, "loss": 3.6909, "step": 163250 }, { "epoch": 1.8345944063407538, "grad_norm": 0.9771916270256042, "learning_rate": 1.2664232510588581e-05, "loss": 3.6638, "step": 163300 }, { "epoch": 1.835156131511097, "grad_norm": 1.1272979974746704, "learning_rate": 1.266198559728573e-05, "loss": 3.6806, "step": 163350 }, { "epoch": 1.83571785668144, "grad_norm": 1.0883663892745972, "learning_rate": 1.2659783622248935e-05, "loss": 3.596, "step": 163400 }, { "epoch": 1.8362795818517832, "grad_norm": 1.067847490310669, "learning_rate": 1.2657536708946086e-05, "loss": 3.7046, "step": 163450 }, { "epoch": 1.8368413070221263, "grad_norm": 2.010690212249756, "learning_rate": 1.2655289795643235e-05, "loss": 3.6551, "step": 163500 }, { "epoch": 1.8374030321924695, "grad_norm": 1.235400915145874, "learning_rate": 1.2653042882340385e-05, "loss": 3.6273, "step": 163550 }, { "epoch": 1.8379647573628126, "grad_norm": 1.1057569980621338, "learning_rate": 1.2650795969037536e-05, "loss": 3.5601, "step": 163600 }, { "epoch": 1.838526482533156, "grad_norm": 1.0582396984100342, "learning_rate": 1.2648549055734685e-05, "loss": 3.69, "step": 163650 }, { "epoch": 1.8390882077034991, "grad_norm": 1.1918880939483643, "learning_rate": 1.2646302142431834e-05, "loss": 3.7185, "step": 163700 }, { "epoch": 1.8396499328738423, "grad_norm": 1.1677442789077759, "learning_rate": 1.2644055229128984e-05, "loss": 3.6586, "step": 163750 }, { "epoch": 1.8402116580441854, "grad_norm": 1.2134548425674438, "learning_rate": 1.2641808315826135e-05, "loss": 3.8218, "step": 163800 }, { "epoch": 1.8407733832145285, "grad_norm": 1.1581451892852783, "learning_rate": 1.2639561402523284e-05, "loss": 3.7075, "step": 163850 }, { "epoch": 1.8413351083848717, "grad_norm": 1.1051182746887207, "learning_rate": 1.2637314489220433e-05, "loss": 3.7662, "step": 163900 }, { "epoch": 1.8418968335552148, "grad_norm": 1.2703778743743896, "learning_rate": 1.2635067575917584e-05, "loss": 3.6097, "step": 163950 }, { "epoch": 1.842458558725558, "grad_norm": 1.1580830812454224, "learning_rate": 1.2632820662614734e-05, "loss": 3.7451, "step": 164000 }, { "epoch": 1.8430202838959011, "grad_norm": 1.3673245906829834, "learning_rate": 1.2630573749311883e-05, "loss": 3.7244, "step": 164050 }, { "epoch": 1.8435820090662443, "grad_norm": 1.2850085496902466, "learning_rate": 1.2628326836009032e-05, "loss": 3.669, "step": 164100 }, { "epoch": 1.8441437342365874, "grad_norm": 1.1741001605987549, "learning_rate": 1.2626079922706183e-05, "loss": 3.7257, "step": 164150 }, { "epoch": 1.8447054594069305, "grad_norm": 1.191554069519043, "learning_rate": 1.2623833009403332e-05, "loss": 3.6091, "step": 164200 }, { "epoch": 1.8452671845772737, "grad_norm": 1.0448939800262451, "learning_rate": 1.2621586096100482e-05, "loss": 3.6468, "step": 164250 }, { "epoch": 1.8458289097476168, "grad_norm": 1.1517430543899536, "learning_rate": 1.2619339182797631e-05, "loss": 3.7041, "step": 164300 }, { "epoch": 1.84639063491796, "grad_norm": 1.2172197103500366, "learning_rate": 1.2617092269494782e-05, "loss": 3.6273, "step": 164350 }, { "epoch": 1.846952360088303, "grad_norm": 1.2285085916519165, "learning_rate": 1.2614845356191931e-05, "loss": 3.5998, "step": 164400 }, { "epoch": 1.8475140852586462, "grad_norm": 1.0099074840545654, "learning_rate": 1.261259844288908e-05, "loss": 3.6736, "step": 164450 }, { "epoch": 1.8480758104289894, "grad_norm": 1.1768802404403687, "learning_rate": 1.2610351529586232e-05, "loss": 3.6699, "step": 164500 }, { "epoch": 1.8486375355993325, "grad_norm": 1.1922498941421509, "learning_rate": 1.2608104616283381e-05, "loss": 3.7537, "step": 164550 }, { "epoch": 1.8491992607696757, "grad_norm": 1.1091722249984741, "learning_rate": 1.260585770298053e-05, "loss": 3.6485, "step": 164600 }, { "epoch": 1.849760985940019, "grad_norm": 1.1923208236694336, "learning_rate": 1.260361078967768e-05, "loss": 3.706, "step": 164650 }, { "epoch": 1.8503227111103622, "grad_norm": 1.2572157382965088, "learning_rate": 1.260136387637483e-05, "loss": 3.6803, "step": 164700 }, { "epoch": 1.8508844362807053, "grad_norm": 1.1056233644485474, "learning_rate": 1.259911696307198e-05, "loss": 3.6433, "step": 164750 }, { "epoch": 1.8514461614510485, "grad_norm": 1.1612497568130493, "learning_rate": 1.259687004976913e-05, "loss": 3.6575, "step": 164800 }, { "epoch": 1.8520078866213916, "grad_norm": 1.1800098419189453, "learning_rate": 1.259462313646628e-05, "loss": 3.6447, "step": 164850 }, { "epoch": 1.8525696117917347, "grad_norm": 1.2071192264556885, "learning_rate": 1.259237622316343e-05, "loss": 3.6733, "step": 164900 }, { "epoch": 1.8531313369620779, "grad_norm": 1.476125717163086, "learning_rate": 1.2590129309860579e-05, "loss": 3.6673, "step": 164950 }, { "epoch": 1.8536930621324212, "grad_norm": 1.2621724605560303, "learning_rate": 1.2587882396557728e-05, "loss": 3.6506, "step": 165000 }, { "epoch": 1.8542547873027644, "grad_norm": 1.1475720405578613, "learning_rate": 1.258563548325488e-05, "loss": 3.6903, "step": 165050 }, { "epoch": 1.8548165124731075, "grad_norm": 1.089660406112671, "learning_rate": 1.2583388569952029e-05, "loss": 3.6402, "step": 165100 }, { "epoch": 1.8553782376434507, "grad_norm": 1.3196265697479248, "learning_rate": 1.2581141656649178e-05, "loss": 3.6875, "step": 165150 }, { "epoch": 1.8559399628137938, "grad_norm": 1.2380402088165283, "learning_rate": 1.2578894743346329e-05, "loss": 3.6614, "step": 165200 }, { "epoch": 1.856501687984137, "grad_norm": 1.473062515258789, "learning_rate": 1.2576647830043478e-05, "loss": 3.7173, "step": 165250 }, { "epoch": 1.85706341315448, "grad_norm": 1.1862760782241821, "learning_rate": 1.2574400916740628e-05, "loss": 3.6043, "step": 165300 }, { "epoch": 1.8576251383248232, "grad_norm": 1.079772710800171, "learning_rate": 1.2572154003437777e-05, "loss": 3.7339, "step": 165350 }, { "epoch": 1.8581868634951664, "grad_norm": 1.1612881422042847, "learning_rate": 1.2569907090134928e-05, "loss": 3.6729, "step": 165400 }, { "epoch": 1.8587485886655095, "grad_norm": 1.1970138549804688, "learning_rate": 1.2567660176832077e-05, "loss": 3.671, "step": 165450 }, { "epoch": 1.8593103138358527, "grad_norm": 1.266577124595642, "learning_rate": 1.2565413263529227e-05, "loss": 3.6625, "step": 165500 }, { "epoch": 1.8598720390061958, "grad_norm": 1.0181468725204468, "learning_rate": 1.2563211288492434e-05, "loss": 3.6748, "step": 165550 }, { "epoch": 1.860433764176539, "grad_norm": 1.2418451309204102, "learning_rate": 1.2560964375189583e-05, "loss": 3.6395, "step": 165600 }, { "epoch": 1.860995489346882, "grad_norm": 2.6148860454559326, "learning_rate": 1.2558717461886733e-05, "loss": 3.6926, "step": 165650 }, { "epoch": 1.8615572145172252, "grad_norm": 1.2383267879486084, "learning_rate": 1.2556470548583884e-05, "loss": 3.708, "step": 165700 }, { "epoch": 1.8621189396875684, "grad_norm": 1.0923413038253784, "learning_rate": 1.2554223635281033e-05, "loss": 3.6836, "step": 165750 }, { "epoch": 1.8626806648579115, "grad_norm": 1.1667050123214722, "learning_rate": 1.2551976721978182e-05, "loss": 3.6588, "step": 165800 }, { "epoch": 1.8632423900282546, "grad_norm": 1.1292152404785156, "learning_rate": 1.2549729808675333e-05, "loss": 3.6692, "step": 165850 }, { "epoch": 1.8638041151985978, "grad_norm": 1.0999101400375366, "learning_rate": 1.2547482895372483e-05, "loss": 3.6453, "step": 165900 }, { "epoch": 1.864365840368941, "grad_norm": 1.339193344116211, "learning_rate": 1.2545235982069632e-05, "loss": 3.6964, "step": 165950 }, { "epoch": 1.8649275655392843, "grad_norm": 1.148620843887329, "learning_rate": 1.2542989068766781e-05, "loss": 3.7522, "step": 166000 }, { "epoch": 1.8654892907096274, "grad_norm": 1.2183562517166138, "learning_rate": 1.2540742155463932e-05, "loss": 3.6897, "step": 166050 }, { "epoch": 1.8660510158799706, "grad_norm": 1.0874114036560059, "learning_rate": 1.2538495242161082e-05, "loss": 3.7102, "step": 166100 }, { "epoch": 1.8666127410503137, "grad_norm": 1.3484716415405273, "learning_rate": 1.2536248328858231e-05, "loss": 3.6323, "step": 166150 }, { "epoch": 1.8671744662206569, "grad_norm": 1.0507627725601196, "learning_rate": 1.2534001415555382e-05, "loss": 3.6477, "step": 166200 }, { "epoch": 1.867736191391, "grad_norm": 0.9727892279624939, "learning_rate": 1.2531754502252531e-05, "loss": 3.7327, "step": 166250 }, { "epoch": 1.8682979165613434, "grad_norm": 1.1271682977676392, "learning_rate": 1.252950758894968e-05, "loss": 3.6228, "step": 166300 }, { "epoch": 1.8688596417316865, "grad_norm": 1.0812020301818848, "learning_rate": 1.252726067564683e-05, "loss": 3.7272, "step": 166350 }, { "epoch": 1.8694213669020296, "grad_norm": 1.0943026542663574, "learning_rate": 1.2525013762343981e-05, "loss": 3.6819, "step": 166400 }, { "epoch": 1.8699830920723728, "grad_norm": 1.1824041604995728, "learning_rate": 1.252276684904113e-05, "loss": 3.6363, "step": 166450 }, { "epoch": 1.870544817242716, "grad_norm": 1.389378547668457, "learning_rate": 1.252051993573828e-05, "loss": 3.659, "step": 166500 }, { "epoch": 1.871106542413059, "grad_norm": 1.2918018102645874, "learning_rate": 1.251827302243543e-05, "loss": 3.604, "step": 166550 }, { "epoch": 1.8716682675834022, "grad_norm": 1.1217737197875977, "learning_rate": 1.251602610913258e-05, "loss": 3.7202, "step": 166600 }, { "epoch": 1.8722299927537454, "grad_norm": 1.286453366279602, "learning_rate": 1.251377919582973e-05, "loss": 3.6355, "step": 166650 }, { "epoch": 1.8727917179240885, "grad_norm": 1.2219747304916382, "learning_rate": 1.2511532282526879e-05, "loss": 3.7565, "step": 166700 }, { "epoch": 1.8733534430944316, "grad_norm": 1.135813593864441, "learning_rate": 1.250928536922403e-05, "loss": 3.6647, "step": 166750 }, { "epoch": 1.8739151682647748, "grad_norm": 1.8363410234451294, "learning_rate": 1.2507038455921179e-05, "loss": 3.6416, "step": 166800 }, { "epoch": 1.874476893435118, "grad_norm": 1.0600069761276245, "learning_rate": 1.2504791542618328e-05, "loss": 3.7299, "step": 166850 }, { "epoch": 1.875038618605461, "grad_norm": 0.9839280247688293, "learning_rate": 1.2502544629315478e-05, "loss": 3.7078, "step": 166900 }, { "epoch": 1.8756003437758042, "grad_norm": 1.0196044445037842, "learning_rate": 1.2500297716012629e-05, "loss": 3.6207, "step": 166950 }, { "epoch": 1.8761620689461473, "grad_norm": 1.2017905712127686, "learning_rate": 1.2498050802709778e-05, "loss": 3.6814, "step": 167000 }, { "epoch": 1.8767237941164905, "grad_norm": 1.1838090419769287, "learning_rate": 1.2495803889406927e-05, "loss": 3.6566, "step": 167050 }, { "epoch": 1.8772855192868336, "grad_norm": 1.3739231824874878, "learning_rate": 1.2493556976104078e-05, "loss": 3.6778, "step": 167100 }, { "epoch": 1.8778472444571768, "grad_norm": 1.1968698501586914, "learning_rate": 1.2491310062801228e-05, "loss": 3.6274, "step": 167150 }, { "epoch": 1.87840896962752, "grad_norm": 1.2168124914169312, "learning_rate": 1.2489063149498377e-05, "loss": 3.7561, "step": 167200 }, { "epoch": 1.878970694797863, "grad_norm": 0.9981439709663391, "learning_rate": 1.2486816236195526e-05, "loss": 3.6272, "step": 167250 }, { "epoch": 1.8795324199682062, "grad_norm": 1.1597669124603271, "learning_rate": 1.2484569322892677e-05, "loss": 3.6901, "step": 167300 }, { "epoch": 1.8800941451385496, "grad_norm": 1.1445658206939697, "learning_rate": 1.2482322409589827e-05, "loss": 3.6727, "step": 167350 }, { "epoch": 1.8806558703088927, "grad_norm": 1.12496817111969, "learning_rate": 1.2480075496286976e-05, "loss": 3.6599, "step": 167400 }, { "epoch": 1.8812175954792358, "grad_norm": 0.900623619556427, "learning_rate": 1.2477828582984127e-05, "loss": 3.6356, "step": 167450 }, { "epoch": 1.881779320649579, "grad_norm": 1.2634716033935547, "learning_rate": 1.2475581669681276e-05, "loss": 3.6375, "step": 167500 }, { "epoch": 1.8823410458199221, "grad_norm": 1.1005879640579224, "learning_rate": 1.2473334756378425e-05, "loss": 3.665, "step": 167550 }, { "epoch": 1.8829027709902653, "grad_norm": 1.0364843606948853, "learning_rate": 1.2471087843075575e-05, "loss": 3.6104, "step": 167600 }, { "epoch": 1.8834644961606086, "grad_norm": 1.0718927383422852, "learning_rate": 1.2468840929772726e-05, "loss": 3.6778, "step": 167650 }, { "epoch": 1.8840262213309518, "grad_norm": 1.133893609046936, "learning_rate": 1.2466594016469875e-05, "loss": 3.6997, "step": 167700 }, { "epoch": 1.884587946501295, "grad_norm": 1.0615676641464233, "learning_rate": 1.2464347103167024e-05, "loss": 3.6485, "step": 167750 }, { "epoch": 1.885149671671638, "grad_norm": 1.2539339065551758, "learning_rate": 1.2462100189864175e-05, "loss": 3.6845, "step": 167800 }, { "epoch": 1.8857113968419812, "grad_norm": 1.1269609928131104, "learning_rate": 1.2459853276561325e-05, "loss": 3.6504, "step": 167850 }, { "epoch": 1.8862731220123243, "grad_norm": 1.220188856124878, "learning_rate": 1.2457606363258474e-05, "loss": 3.6054, "step": 167900 }, { "epoch": 1.8868348471826675, "grad_norm": 1.1640559434890747, "learning_rate": 1.2455359449955623e-05, "loss": 3.6642, "step": 167950 }, { "epoch": 1.8873965723530106, "grad_norm": 1.1580772399902344, "learning_rate": 1.2453112536652774e-05, "loss": 3.683, "step": 168000 }, { "epoch": 1.8879582975233538, "grad_norm": 1.1342999935150146, "learning_rate": 1.2450865623349924e-05, "loss": 3.6598, "step": 168050 }, { "epoch": 1.888520022693697, "grad_norm": 1.1601992845535278, "learning_rate": 1.2448618710047073e-05, "loss": 3.6451, "step": 168100 }, { "epoch": 1.88908174786404, "grad_norm": 1.161476969718933, "learning_rate": 1.2446371796744224e-05, "loss": 3.6168, "step": 168150 }, { "epoch": 1.8896434730343832, "grad_norm": 1.0610288381576538, "learning_rate": 1.2444124883441373e-05, "loss": 3.7246, "step": 168200 }, { "epoch": 1.8902051982047263, "grad_norm": 1.1727944612503052, "learning_rate": 1.2441877970138523e-05, "loss": 3.7116, "step": 168250 }, { "epoch": 1.8907669233750695, "grad_norm": 1.1720994710922241, "learning_rate": 1.2439631056835672e-05, "loss": 3.6688, "step": 168300 }, { "epoch": 1.8913286485454126, "grad_norm": 1.2003391981124878, "learning_rate": 1.2437384143532823e-05, "loss": 3.5629, "step": 168350 }, { "epoch": 1.8918903737157557, "grad_norm": 1.2217501401901245, "learning_rate": 1.2435137230229972e-05, "loss": 3.6789, "step": 168400 }, { "epoch": 1.8924520988860989, "grad_norm": 1.2102724313735962, "learning_rate": 1.2432890316927122e-05, "loss": 3.698, "step": 168450 }, { "epoch": 1.893013824056442, "grad_norm": 1.123992919921875, "learning_rate": 1.2430643403624273e-05, "loss": 3.6312, "step": 168500 }, { "epoch": 1.8935755492267852, "grad_norm": 1.128458023071289, "learning_rate": 1.2428396490321422e-05, "loss": 3.7031, "step": 168550 }, { "epoch": 1.8941372743971283, "grad_norm": 1.181666612625122, "learning_rate": 1.2426149577018571e-05, "loss": 3.5914, "step": 168600 }, { "epoch": 1.8946989995674715, "grad_norm": 1.11942458152771, "learning_rate": 1.242390266371572e-05, "loss": 3.7215, "step": 168650 }, { "epoch": 1.8952607247378148, "grad_norm": 1.2772291898727417, "learning_rate": 1.2421655750412872e-05, "loss": 3.6783, "step": 168700 }, { "epoch": 1.895822449908158, "grad_norm": 1.3013349771499634, "learning_rate": 1.2419408837110021e-05, "loss": 3.686, "step": 168750 }, { "epoch": 1.896384175078501, "grad_norm": 0.983788251876831, "learning_rate": 1.241716192380717e-05, "loss": 3.6762, "step": 168800 }, { "epoch": 1.8969459002488442, "grad_norm": 1.247036099433899, "learning_rate": 1.241491501050432e-05, "loss": 3.6746, "step": 168850 }, { "epoch": 1.8975076254191874, "grad_norm": 1.238627314567566, "learning_rate": 1.241266809720147e-05, "loss": 3.6322, "step": 168900 }, { "epoch": 1.8980693505895305, "grad_norm": 1.1986321210861206, "learning_rate": 1.241042118389862e-05, "loss": 3.7118, "step": 168950 }, { "epoch": 1.8986310757598739, "grad_norm": 1.5572178363800049, "learning_rate": 1.240817427059577e-05, "loss": 3.6297, "step": 169000 }, { "epoch": 1.899192800930217, "grad_norm": 1.256276249885559, "learning_rate": 1.240592735729292e-05, "loss": 3.6844, "step": 169050 }, { "epoch": 1.8997545261005602, "grad_norm": 1.120082974433899, "learning_rate": 1.240368044399007e-05, "loss": 3.6543, "step": 169100 }, { "epoch": 1.9003162512709033, "grad_norm": 1.2404358386993408, "learning_rate": 1.2401433530687219e-05, "loss": 3.6648, "step": 169150 }, { "epoch": 1.9008779764412465, "grad_norm": 1.2647347450256348, "learning_rate": 1.2399186617384368e-05, "loss": 3.7098, "step": 169200 }, { "epoch": 1.9014397016115896, "grad_norm": 1.0600517988204956, "learning_rate": 1.239693970408152e-05, "loss": 3.6644, "step": 169250 }, { "epoch": 1.9020014267819327, "grad_norm": 1.066779613494873, "learning_rate": 1.2394692790778668e-05, "loss": 3.7638, "step": 169300 }, { "epoch": 1.9025631519522759, "grad_norm": 1.0992931127548218, "learning_rate": 1.2392445877475818e-05, "loss": 3.6887, "step": 169350 }, { "epoch": 1.903124877122619, "grad_norm": 1.0564545392990112, "learning_rate": 1.2390198964172969e-05, "loss": 3.6604, "step": 169400 }, { "epoch": 1.9036866022929622, "grad_norm": 1.122517704963684, "learning_rate": 1.2387952050870118e-05, "loss": 3.621, "step": 169450 }, { "epoch": 1.9042483274633053, "grad_norm": 1.1531624794006348, "learning_rate": 1.2385705137567267e-05, "loss": 3.6562, "step": 169500 }, { "epoch": 1.9048100526336484, "grad_norm": 1.1910247802734375, "learning_rate": 1.2383458224264417e-05, "loss": 3.6425, "step": 169550 }, { "epoch": 1.9053717778039916, "grad_norm": 1.2163705825805664, "learning_rate": 1.2381211310961568e-05, "loss": 3.601, "step": 169600 }, { "epoch": 1.9059335029743347, "grad_norm": 1.324793815612793, "learning_rate": 1.2378964397658717e-05, "loss": 3.6598, "step": 169650 }, { "epoch": 1.9064952281446779, "grad_norm": 1.1938973665237427, "learning_rate": 1.2376717484355866e-05, "loss": 3.7127, "step": 169700 }, { "epoch": 1.907056953315021, "grad_norm": 1.1128785610198975, "learning_rate": 1.2374470571053017e-05, "loss": 3.6492, "step": 169750 }, { "epoch": 1.9076186784853641, "grad_norm": 1.3163511753082275, "learning_rate": 1.2372223657750167e-05, "loss": 3.6423, "step": 169800 }, { "epoch": 1.9081804036557073, "grad_norm": 1.1273045539855957, "learning_rate": 1.2369976744447316e-05, "loss": 3.6753, "step": 169850 }, { "epoch": 1.9087421288260504, "grad_norm": 1.101928472518921, "learning_rate": 1.2367729831144465e-05, "loss": 3.7705, "step": 169900 }, { "epoch": 1.9093038539963936, "grad_norm": 1.085390567779541, "learning_rate": 1.2365482917841616e-05, "loss": 3.6355, "step": 169950 }, { "epoch": 1.909865579166737, "grad_norm": 1.0475810766220093, "learning_rate": 1.2363236004538766e-05, "loss": 3.6887, "step": 170000 }, { "epoch": 1.91042730433708, "grad_norm": 1.0383238792419434, "learning_rate": 1.2360989091235915e-05, "loss": 3.6665, "step": 170050 }, { "epoch": 1.9109890295074232, "grad_norm": 1.354358196258545, "learning_rate": 1.2358742177933066e-05, "loss": 3.6792, "step": 170100 }, { "epoch": 1.9115507546777664, "grad_norm": 1.1882213354110718, "learning_rate": 1.2356495264630215e-05, "loss": 3.8028, "step": 170150 }, { "epoch": 1.9121124798481095, "grad_norm": 0.9827480912208557, "learning_rate": 1.2354248351327365e-05, "loss": 3.6659, "step": 170200 }, { "epoch": 1.9126742050184526, "grad_norm": 1.2606571912765503, "learning_rate": 1.2352001438024514e-05, "loss": 3.7601, "step": 170250 }, { "epoch": 1.9132359301887958, "grad_norm": 1.0390775203704834, "learning_rate": 1.2349754524721665e-05, "loss": 3.6812, "step": 170300 }, { "epoch": 1.9137976553591391, "grad_norm": 1.2463587522506714, "learning_rate": 1.2347507611418814e-05, "loss": 3.6176, "step": 170350 }, { "epoch": 1.9143593805294823, "grad_norm": 1.0027610063552856, "learning_rate": 1.2345305636382022e-05, "loss": 3.6706, "step": 170400 }, { "epoch": 1.9149211056998254, "grad_norm": 1.1805909872055054, "learning_rate": 1.2343058723079171e-05, "loss": 3.6551, "step": 170450 }, { "epoch": 1.9154828308701686, "grad_norm": 1.3694796562194824, "learning_rate": 1.234081180977632e-05, "loss": 3.7188, "step": 170500 }, { "epoch": 1.9160445560405117, "grad_norm": 1.2324082851409912, "learning_rate": 1.233856489647347e-05, "loss": 3.646, "step": 170550 }, { "epoch": 1.9166062812108549, "grad_norm": 0.9831752777099609, "learning_rate": 1.2336317983170621e-05, "loss": 3.637, "step": 170600 }, { "epoch": 1.917168006381198, "grad_norm": 1.0351693630218506, "learning_rate": 1.233407106986777e-05, "loss": 3.7079, "step": 170650 }, { "epoch": 1.9177297315515411, "grad_norm": 1.325768232345581, "learning_rate": 1.233182415656492e-05, "loss": 3.752, "step": 170700 }, { "epoch": 1.9182914567218843, "grad_norm": 1.097913146018982, "learning_rate": 1.232957724326207e-05, "loss": 3.6707, "step": 170750 }, { "epoch": 1.9188531818922274, "grad_norm": 1.0953683853149414, "learning_rate": 1.232733032995922e-05, "loss": 3.6086, "step": 170800 }, { "epoch": 1.9194149070625706, "grad_norm": 1.2423237562179565, "learning_rate": 1.2325083416656369e-05, "loss": 3.6191, "step": 170850 }, { "epoch": 1.9199766322329137, "grad_norm": 1.1714104413986206, "learning_rate": 1.2322836503353518e-05, "loss": 3.7158, "step": 170900 }, { "epoch": 1.9205383574032568, "grad_norm": 1.0403722524642944, "learning_rate": 1.2320634528316726e-05, "loss": 3.6497, "step": 170950 }, { "epoch": 1.9211000825736, "grad_norm": 1.111100196838379, "learning_rate": 1.2318387615013875e-05, "loss": 3.6101, "step": 171000 }, { "epoch": 1.9216618077439431, "grad_norm": 1.2852191925048828, "learning_rate": 1.2316140701711026e-05, "loss": 3.6072, "step": 171050 }, { "epoch": 1.9222235329142863, "grad_norm": 1.4478634595870972, "learning_rate": 1.2313893788408176e-05, "loss": 3.702, "step": 171100 }, { "epoch": 1.9227852580846294, "grad_norm": 1.2640856504440308, "learning_rate": 1.2311646875105325e-05, "loss": 3.5434, "step": 171150 }, { "epoch": 1.9233469832549726, "grad_norm": 1.191200852394104, "learning_rate": 1.2309444900068533e-05, "loss": 3.5903, "step": 171200 }, { "epoch": 1.9239087084253157, "grad_norm": 1.1147801876068115, "learning_rate": 1.2307197986765682e-05, "loss": 3.676, "step": 171250 }, { "epoch": 1.9244704335956588, "grad_norm": 1.3612926006317139, "learning_rate": 1.2304951073462831e-05, "loss": 3.7212, "step": 171300 }, { "epoch": 1.9250321587660022, "grad_norm": 1.052769422531128, "learning_rate": 1.2302704160159982e-05, "loss": 3.6766, "step": 171350 }, { "epoch": 1.9255938839363453, "grad_norm": 1.0386183261871338, "learning_rate": 1.2300457246857132e-05, "loss": 3.6387, "step": 171400 }, { "epoch": 1.9261556091066885, "grad_norm": 1.2325382232666016, "learning_rate": 1.2298210333554281e-05, "loss": 3.6958, "step": 171450 }, { "epoch": 1.9267173342770316, "grad_norm": 1.1987472772598267, "learning_rate": 1.229596342025143e-05, "loss": 3.7288, "step": 171500 }, { "epoch": 1.9272790594473748, "grad_norm": 1.0316622257232666, "learning_rate": 1.2293716506948581e-05, "loss": 3.6099, "step": 171550 }, { "epoch": 1.927840784617718, "grad_norm": 1.1573472023010254, "learning_rate": 1.229146959364573e-05, "loss": 3.6843, "step": 171600 }, { "epoch": 1.928402509788061, "grad_norm": 1.351535677909851, "learning_rate": 1.228922268034288e-05, "loss": 3.6412, "step": 171650 }, { "epoch": 1.9289642349584044, "grad_norm": 1.1215472221374512, "learning_rate": 1.2286975767040031e-05, "loss": 3.6187, "step": 171700 }, { "epoch": 1.9295259601287476, "grad_norm": 1.1648633480072021, "learning_rate": 1.228472885373718e-05, "loss": 3.6905, "step": 171750 }, { "epoch": 1.9300876852990907, "grad_norm": 1.2470959424972534, "learning_rate": 1.228248194043433e-05, "loss": 3.6249, "step": 171800 }, { "epoch": 1.9306494104694338, "grad_norm": 2.624309778213501, "learning_rate": 1.2280235027131479e-05, "loss": 3.6018, "step": 171850 }, { "epoch": 1.931211135639777, "grad_norm": 1.2481865882873535, "learning_rate": 1.227798811382863e-05, "loss": 3.653, "step": 171900 }, { "epoch": 1.9317728608101201, "grad_norm": 1.2186285257339478, "learning_rate": 1.227574120052578e-05, "loss": 3.666, "step": 171950 }, { "epoch": 1.9323345859804633, "grad_norm": 1.1497619152069092, "learning_rate": 1.2273494287222929e-05, "loss": 3.676, "step": 172000 }, { "epoch": 1.9328963111508064, "grad_norm": 1.1843317747116089, "learning_rate": 1.227124737392008e-05, "loss": 3.6265, "step": 172050 }, { "epoch": 1.9334580363211495, "grad_norm": 1.1359236240386963, "learning_rate": 1.2269000460617229e-05, "loss": 3.6598, "step": 172100 }, { "epoch": 1.9340197614914927, "grad_norm": 1.17789888381958, "learning_rate": 1.2266753547314378e-05, "loss": 3.6281, "step": 172150 }, { "epoch": 1.9345814866618358, "grad_norm": 1.2440000772476196, "learning_rate": 1.2264506634011527e-05, "loss": 3.6616, "step": 172200 }, { "epoch": 1.935143211832179, "grad_norm": 1.3088947534561157, "learning_rate": 1.2262259720708678e-05, "loss": 3.7083, "step": 172250 }, { "epoch": 1.935704937002522, "grad_norm": 1.0753567218780518, "learning_rate": 1.2260012807405828e-05, "loss": 3.5984, "step": 172300 }, { "epoch": 1.9362666621728652, "grad_norm": 1.1660813093185425, "learning_rate": 1.2257765894102977e-05, "loss": 3.649, "step": 172350 }, { "epoch": 1.9368283873432084, "grad_norm": 1.3445658683776855, "learning_rate": 1.2255518980800128e-05, "loss": 3.7311, "step": 172400 }, { "epoch": 1.9373901125135515, "grad_norm": 1.0266178846359253, "learning_rate": 1.2253272067497277e-05, "loss": 3.5824, "step": 172450 }, { "epoch": 1.9379518376838947, "grad_norm": 1.0390028953552246, "learning_rate": 1.2251025154194427e-05, "loss": 3.6594, "step": 172500 }, { "epoch": 1.9385135628542378, "grad_norm": 1.387713074684143, "learning_rate": 1.2248778240891576e-05, "loss": 3.6387, "step": 172550 }, { "epoch": 1.939075288024581, "grad_norm": 1.2674925327301025, "learning_rate": 1.2246531327588727e-05, "loss": 3.6946, "step": 172600 }, { "epoch": 1.939637013194924, "grad_norm": 1.0736058950424194, "learning_rate": 1.2244284414285876e-05, "loss": 3.7241, "step": 172650 }, { "epoch": 1.9401987383652675, "grad_norm": 1.0840920209884644, "learning_rate": 1.2242037500983026e-05, "loss": 3.643, "step": 172700 }, { "epoch": 1.9407604635356106, "grad_norm": 1.102241039276123, "learning_rate": 1.2239790587680177e-05, "loss": 3.7588, "step": 172750 }, { "epoch": 1.9413221887059537, "grad_norm": 1.0428377389907837, "learning_rate": 1.2237543674377326e-05, "loss": 3.5957, "step": 172800 }, { "epoch": 1.9418839138762969, "grad_norm": 1.10206937789917, "learning_rate": 1.2235296761074475e-05, "loss": 3.6582, "step": 172850 }, { "epoch": 1.94244563904664, "grad_norm": 1.1507540941238403, "learning_rate": 1.2233049847771625e-05, "loss": 3.7045, "step": 172900 }, { "epoch": 1.9430073642169832, "grad_norm": 2.4136881828308105, "learning_rate": 1.2230802934468776e-05, "loss": 3.6613, "step": 172950 }, { "epoch": 1.9435690893873263, "grad_norm": 1.395912528038025, "learning_rate": 1.2228556021165925e-05, "loss": 3.6723, "step": 173000 }, { "epoch": 1.9441308145576697, "grad_norm": 1.20100998878479, "learning_rate": 1.2226309107863074e-05, "loss": 3.6252, "step": 173050 }, { "epoch": 1.9446925397280128, "grad_norm": 1.0831165313720703, "learning_rate": 1.2224062194560224e-05, "loss": 3.6134, "step": 173100 }, { "epoch": 1.945254264898356, "grad_norm": 1.2886075973510742, "learning_rate": 1.2221815281257375e-05, "loss": 3.673, "step": 173150 }, { "epoch": 1.945815990068699, "grad_norm": 1.2429815530776978, "learning_rate": 1.2219568367954524e-05, "loss": 3.6499, "step": 173200 }, { "epoch": 1.9463777152390422, "grad_norm": 1.1157153844833374, "learning_rate": 1.2217321454651673e-05, "loss": 3.6351, "step": 173250 }, { "epoch": 1.9469394404093854, "grad_norm": 1.1003978252410889, "learning_rate": 1.2215074541348824e-05, "loss": 3.6049, "step": 173300 }, { "epoch": 1.9475011655797285, "grad_norm": 1.092509150505066, "learning_rate": 1.2212827628045974e-05, "loss": 3.6381, "step": 173350 }, { "epoch": 1.9480628907500717, "grad_norm": 1.0967395305633545, "learning_rate": 1.2210580714743123e-05, "loss": 3.6851, "step": 173400 }, { "epoch": 1.9486246159204148, "grad_norm": 1.0878044366836548, "learning_rate": 1.2208333801440272e-05, "loss": 3.6377, "step": 173450 }, { "epoch": 1.949186341090758, "grad_norm": 1.120462417602539, "learning_rate": 1.2206086888137423e-05, "loss": 3.709, "step": 173500 }, { "epoch": 1.949748066261101, "grad_norm": 1.0878864526748657, "learning_rate": 1.2203839974834573e-05, "loss": 3.6988, "step": 173550 }, { "epoch": 1.9503097914314442, "grad_norm": 1.0665409564971924, "learning_rate": 1.2201593061531722e-05, "loss": 3.6464, "step": 173600 }, { "epoch": 1.9508715166017874, "grad_norm": 1.1327898502349854, "learning_rate": 1.2199346148228873e-05, "loss": 3.6786, "step": 173650 }, { "epoch": 1.9514332417721305, "grad_norm": 1.3542909622192383, "learning_rate": 1.2197099234926022e-05, "loss": 3.6545, "step": 173700 }, { "epoch": 1.9519949669424737, "grad_norm": 1.3699864149093628, "learning_rate": 1.2194852321623172e-05, "loss": 3.6166, "step": 173750 }, { "epoch": 1.9525566921128168, "grad_norm": 1.2797566652297974, "learning_rate": 1.219260540832032e-05, "loss": 3.6561, "step": 173800 }, { "epoch": 1.95311841728316, "grad_norm": 1.0340110063552856, "learning_rate": 1.2190358495017472e-05, "loss": 3.7516, "step": 173850 }, { "epoch": 1.953680142453503, "grad_norm": 1.2984986305236816, "learning_rate": 1.2188111581714621e-05, "loss": 3.726, "step": 173900 }, { "epoch": 1.9542418676238462, "grad_norm": 1.1449670791625977, "learning_rate": 1.218586466841177e-05, "loss": 3.6731, "step": 173950 }, { "epoch": 1.9548035927941894, "grad_norm": 1.189428448677063, "learning_rate": 1.2183617755108921e-05, "loss": 3.655, "step": 174000 }, { "epoch": 1.9553653179645327, "grad_norm": 1.0728976726531982, "learning_rate": 1.218137084180607e-05, "loss": 3.6844, "step": 174050 }, { "epoch": 1.9559270431348759, "grad_norm": 1.316543698310852, "learning_rate": 1.217912392850322e-05, "loss": 3.685, "step": 174100 }, { "epoch": 1.956488768305219, "grad_norm": 1.0091776847839355, "learning_rate": 1.217687701520037e-05, "loss": 3.7103, "step": 174150 }, { "epoch": 1.9570504934755621, "grad_norm": 1.1065725088119507, "learning_rate": 1.217463010189752e-05, "loss": 3.7193, "step": 174200 }, { "epoch": 1.9576122186459053, "grad_norm": 1.032600998878479, "learning_rate": 1.217238318859467e-05, "loss": 3.6122, "step": 174250 }, { "epoch": 1.9581739438162484, "grad_norm": 1.183090090751648, "learning_rate": 1.2170136275291819e-05, "loss": 3.6051, "step": 174300 }, { "epoch": 1.9587356689865916, "grad_norm": 1.437174677848816, "learning_rate": 1.216788936198897e-05, "loss": 3.6526, "step": 174350 }, { "epoch": 1.959297394156935, "grad_norm": 1.0554615259170532, "learning_rate": 1.216564244868612e-05, "loss": 3.6749, "step": 174400 }, { "epoch": 1.959859119327278, "grad_norm": 1.0573745965957642, "learning_rate": 1.2163395535383269e-05, "loss": 3.6799, "step": 174450 }, { "epoch": 1.9604208444976212, "grad_norm": 1.2770339250564575, "learning_rate": 1.2161148622080418e-05, "loss": 3.6972, "step": 174500 }, { "epoch": 1.9609825696679644, "grad_norm": 1.0049799680709839, "learning_rate": 1.2158901708777569e-05, "loss": 3.6382, "step": 174550 }, { "epoch": 1.9615442948383075, "grad_norm": 1.2158585786819458, "learning_rate": 1.2156654795474718e-05, "loss": 3.6784, "step": 174600 }, { "epoch": 1.9621060200086506, "grad_norm": 1.1403707265853882, "learning_rate": 1.2154407882171868e-05, "loss": 3.6615, "step": 174650 }, { "epoch": 1.9626677451789938, "grad_norm": 1.1854654550552368, "learning_rate": 1.2152160968869019e-05, "loss": 3.6604, "step": 174700 }, { "epoch": 1.963229470349337, "grad_norm": 1.1664259433746338, "learning_rate": 1.2149914055566168e-05, "loss": 3.6374, "step": 174750 }, { "epoch": 1.96379119551968, "grad_norm": 1.2385070323944092, "learning_rate": 1.2147667142263317e-05, "loss": 3.6751, "step": 174800 }, { "epoch": 1.9643529206900232, "grad_norm": 1.2153853178024292, "learning_rate": 1.2145420228960467e-05, "loss": 3.6107, "step": 174850 }, { "epoch": 1.9649146458603663, "grad_norm": 1.2236731052398682, "learning_rate": 1.2143173315657618e-05, "loss": 3.6285, "step": 174900 }, { "epoch": 1.9654763710307095, "grad_norm": 1.0807764530181885, "learning_rate": 1.2140926402354767e-05, "loss": 3.6047, "step": 174950 }, { "epoch": 1.9660380962010526, "grad_norm": 1.1282566785812378, "learning_rate": 1.2138679489051916e-05, "loss": 3.6077, "step": 175000 }, { "epoch": 1.9665998213713958, "grad_norm": 1.238545536994934, "learning_rate": 1.2136432575749066e-05, "loss": 3.7162, "step": 175050 }, { "epoch": 1.967161546541739, "grad_norm": 1.094677448272705, "learning_rate": 1.2134185662446217e-05, "loss": 3.699, "step": 175100 }, { "epoch": 1.967723271712082, "grad_norm": 1.2476550340652466, "learning_rate": 1.2131938749143366e-05, "loss": 3.6716, "step": 175150 }, { "epoch": 1.9682849968824252, "grad_norm": 1.1629317998886108, "learning_rate": 1.2129691835840515e-05, "loss": 3.6337, "step": 175200 }, { "epoch": 1.9688467220527683, "grad_norm": 1.261732816696167, "learning_rate": 1.2127444922537666e-05, "loss": 3.6534, "step": 175250 }, { "epoch": 1.9694084472231115, "grad_norm": 1.1375350952148438, "learning_rate": 1.2125198009234816e-05, "loss": 3.7596, "step": 175300 }, { "epoch": 1.9699701723934546, "grad_norm": 1.0128322839736938, "learning_rate": 1.2122951095931965e-05, "loss": 3.5842, "step": 175350 }, { "epoch": 1.970531897563798, "grad_norm": 1.4134347438812256, "learning_rate": 1.2120704182629114e-05, "loss": 3.7659, "step": 175400 }, { "epoch": 1.9710936227341411, "grad_norm": 1.1948753595352173, "learning_rate": 1.2118457269326265e-05, "loss": 3.6362, "step": 175450 }, { "epoch": 1.9716553479044843, "grad_norm": 1.2572717666625977, "learning_rate": 1.2116210356023415e-05, "loss": 3.7703, "step": 175500 }, { "epoch": 1.9722170730748274, "grad_norm": 1.0441925525665283, "learning_rate": 1.2113963442720564e-05, "loss": 3.6649, "step": 175550 }, { "epoch": 1.9727787982451706, "grad_norm": 1.2285887002944946, "learning_rate": 1.2111716529417715e-05, "loss": 3.6652, "step": 175600 }, { "epoch": 1.9733405234155137, "grad_norm": 1.2192641496658325, "learning_rate": 1.2109469616114864e-05, "loss": 3.6174, "step": 175650 }, { "epoch": 1.9739022485858568, "grad_norm": 1.2564349174499512, "learning_rate": 1.2107222702812014e-05, "loss": 3.6524, "step": 175700 }, { "epoch": 1.9744639737562002, "grad_norm": 1.3048930168151855, "learning_rate": 1.2104975789509163e-05, "loss": 3.7103, "step": 175750 }, { "epoch": 1.9750256989265433, "grad_norm": 1.180362582206726, "learning_rate": 1.2102728876206314e-05, "loss": 3.6536, "step": 175800 }, { "epoch": 1.9755874240968865, "grad_norm": 1.1488772630691528, "learning_rate": 1.210052690116952e-05, "loss": 3.7154, "step": 175850 }, { "epoch": 1.9761491492672296, "grad_norm": 1.1649621725082397, "learning_rate": 1.209827998786667e-05, "loss": 3.5611, "step": 175900 }, { "epoch": 1.9767108744375728, "grad_norm": 1.1073241233825684, "learning_rate": 1.209603307456382e-05, "loss": 3.6528, "step": 175950 }, { "epoch": 1.977272599607916, "grad_norm": 1.1920549869537354, "learning_rate": 1.209378616126097e-05, "loss": 3.6264, "step": 176000 }, { "epoch": 1.977834324778259, "grad_norm": 1.4693737030029297, "learning_rate": 1.2091539247958119e-05, "loss": 3.6645, "step": 176050 }, { "epoch": 1.9783960499486022, "grad_norm": 2.4251062870025635, "learning_rate": 1.208929233465527e-05, "loss": 3.6509, "step": 176100 }, { "epoch": 1.9789577751189453, "grad_norm": 1.1257350444793701, "learning_rate": 1.2087045421352419e-05, "loss": 3.6939, "step": 176150 }, { "epoch": 1.9795195002892885, "grad_norm": 1.2718253135681152, "learning_rate": 1.2084798508049568e-05, "loss": 3.6891, "step": 176200 }, { "epoch": 1.9800812254596316, "grad_norm": 1.1150072813034058, "learning_rate": 1.208255159474672e-05, "loss": 3.6344, "step": 176250 }, { "epoch": 1.9806429506299748, "grad_norm": 1.1545917987823486, "learning_rate": 1.2080304681443869e-05, "loss": 3.5648, "step": 176300 }, { "epoch": 1.981204675800318, "grad_norm": 1.1448496580123901, "learning_rate": 1.2078057768141018e-05, "loss": 3.6085, "step": 176350 }, { "epoch": 1.981766400970661, "grad_norm": 1.113794207572937, "learning_rate": 1.2075810854838167e-05, "loss": 3.5622, "step": 176400 }, { "epoch": 1.9823281261410042, "grad_norm": 1.2571277618408203, "learning_rate": 1.2073563941535318e-05, "loss": 3.5989, "step": 176450 }, { "epoch": 1.9828898513113473, "grad_norm": 1.4230687618255615, "learning_rate": 1.2071317028232468e-05, "loss": 3.6226, "step": 176500 }, { "epoch": 1.9834515764816905, "grad_norm": 1.2205623388290405, "learning_rate": 1.2069070114929617e-05, "loss": 3.7403, "step": 176550 }, { "epoch": 1.9840133016520336, "grad_norm": 1.0938069820404053, "learning_rate": 1.2066823201626768e-05, "loss": 3.7403, "step": 176600 }, { "epoch": 1.9845750268223767, "grad_norm": 1.2703936100006104, "learning_rate": 1.2064576288323917e-05, "loss": 3.6401, "step": 176650 }, { "epoch": 1.9851367519927199, "grad_norm": 1.0708928108215332, "learning_rate": 1.2062329375021067e-05, "loss": 3.6527, "step": 176700 }, { "epoch": 1.9856984771630632, "grad_norm": 1.0272630453109741, "learning_rate": 1.2060082461718216e-05, "loss": 3.7237, "step": 176750 }, { "epoch": 1.9862602023334064, "grad_norm": 1.2990872859954834, "learning_rate": 1.2057835548415367e-05, "loss": 3.5784, "step": 176800 }, { "epoch": 1.9868219275037495, "grad_norm": 1.2626888751983643, "learning_rate": 1.2055588635112516e-05, "loss": 3.6648, "step": 176850 }, { "epoch": 1.9873836526740927, "grad_norm": 1.298158049583435, "learning_rate": 1.2053341721809666e-05, "loss": 3.6625, "step": 176900 }, { "epoch": 1.9879453778444358, "grad_norm": 1.1682379245758057, "learning_rate": 1.2051094808506817e-05, "loss": 3.6812, "step": 176950 }, { "epoch": 1.988507103014779, "grad_norm": 1.0595457553863525, "learning_rate": 1.2048847895203966e-05, "loss": 3.7265, "step": 177000 }, { "epoch": 1.989068828185122, "grad_norm": 1.2153105735778809, "learning_rate": 1.2046600981901115e-05, "loss": 3.6573, "step": 177050 }, { "epoch": 1.9896305533554655, "grad_norm": 1.3306924104690552, "learning_rate": 1.2044354068598265e-05, "loss": 3.6213, "step": 177100 }, { "epoch": 1.9901922785258086, "grad_norm": 1.3015691041946411, "learning_rate": 1.2042107155295416e-05, "loss": 3.6733, "step": 177150 }, { "epoch": 1.9907540036961517, "grad_norm": 1.3520183563232422, "learning_rate": 1.2039860241992565e-05, "loss": 3.5986, "step": 177200 }, { "epoch": 1.9913157288664949, "grad_norm": 1.1858617067337036, "learning_rate": 1.2037613328689714e-05, "loss": 3.6491, "step": 177250 }, { "epoch": 1.991877454036838, "grad_norm": 1.3016369342803955, "learning_rate": 1.2035366415386865e-05, "loss": 3.7812, "step": 177300 }, { "epoch": 1.9924391792071812, "grad_norm": 1.259308099746704, "learning_rate": 1.2033119502084014e-05, "loss": 3.663, "step": 177350 }, { "epoch": 1.9930009043775243, "grad_norm": 0.9826324582099915, "learning_rate": 1.2030872588781164e-05, "loss": 3.6208, "step": 177400 }, { "epoch": 1.9935626295478674, "grad_norm": 1.122003436088562, "learning_rate": 1.2028625675478313e-05, "loss": 3.5831, "step": 177450 }, { "epoch": 1.9941243547182106, "grad_norm": 1.0870612859725952, "learning_rate": 1.2026378762175464e-05, "loss": 3.6434, "step": 177500 }, { "epoch": 1.9946860798885537, "grad_norm": 1.2947250604629517, "learning_rate": 1.2024131848872613e-05, "loss": 3.6164, "step": 177550 }, { "epoch": 1.9952478050588969, "grad_norm": 1.055662751197815, "learning_rate": 1.2021884935569763e-05, "loss": 3.6337, "step": 177600 }, { "epoch": 1.99580953022924, "grad_norm": 1.093037724494934, "learning_rate": 1.2019638022266912e-05, "loss": 3.6574, "step": 177650 }, { "epoch": 1.9963712553995832, "grad_norm": 1.1708205938339233, "learning_rate": 1.2017391108964063e-05, "loss": 3.6793, "step": 177700 }, { "epoch": 1.9969329805699263, "grad_norm": 1.315970778465271, "learning_rate": 1.2015144195661212e-05, "loss": 3.6725, "step": 177750 }, { "epoch": 1.9974947057402694, "grad_norm": 1.1181354522705078, "learning_rate": 1.2012897282358362e-05, "loss": 3.6046, "step": 177800 }, { "epoch": 1.9980564309106126, "grad_norm": 1.1091622114181519, "learning_rate": 1.2010650369055513e-05, "loss": 3.7024, "step": 177850 }, { "epoch": 1.9986181560809557, "grad_norm": 1.0360990762710571, "learning_rate": 1.2008403455752662e-05, "loss": 3.6336, "step": 177900 }, { "epoch": 1.9991798812512989, "grad_norm": 1.18849515914917, "learning_rate": 1.2006156542449811e-05, "loss": 3.6415, "step": 177950 }, { "epoch": 1.999741606421642, "grad_norm": 1.1326818466186523, "learning_rate": 1.200390962914696e-05, "loss": 3.7134, "step": 178000 }, { "epoch": 2.0, "eval_bleu": 5.9547, "eval_gen_len": 21.5378, "eval_loss": 3.5279541015625, "eval_runtime": 25257.1404, "eval_samples_per_second": 14.097, "eval_steps_per_second": 0.881, "step": 178023 }, { "epoch": 2.000303331591985, "grad_norm": 1.4539496898651123, "learning_rate": 1.2001662715844112e-05, "loss": 3.6639, "step": 178050 }, { "epoch": 2.0008650567623283, "grad_norm": 1.0738657712936401, "learning_rate": 1.199941580254126e-05, "loss": 3.6008, "step": 178100 }, { "epoch": 2.0014267819326714, "grad_norm": 1.2506818771362305, "learning_rate": 1.1997168889238409e-05, "loss": 3.6954, "step": 178150 }, { "epoch": 2.0019885071030146, "grad_norm": 1.0013850927352905, "learning_rate": 1.1994921975935558e-05, "loss": 3.6818, "step": 178200 }, { "epoch": 2.002550232273358, "grad_norm": 1.1331974267959595, "learning_rate": 1.1992675062632709e-05, "loss": 3.6265, "step": 178250 }, { "epoch": 2.0031119574437013, "grad_norm": 1.3222802877426147, "learning_rate": 1.1990428149329858e-05, "loss": 3.6171, "step": 178300 }, { "epoch": 2.0036736826140444, "grad_norm": 1.0162914991378784, "learning_rate": 1.1988181236027008e-05, "loss": 3.6243, "step": 178350 }, { "epoch": 2.0042354077843876, "grad_norm": 1.3031641244888306, "learning_rate": 1.1985934322724157e-05, "loss": 3.7056, "step": 178400 }, { "epoch": 2.0047971329547307, "grad_norm": 1.261915922164917, "learning_rate": 1.1983687409421308e-05, "loss": 3.6977, "step": 178450 }, { "epoch": 2.005358858125074, "grad_norm": 1.172046422958374, "learning_rate": 1.1981440496118457e-05, "loss": 3.5842, "step": 178500 }, { "epoch": 2.005920583295417, "grad_norm": 1.1569795608520508, "learning_rate": 1.1979193582815607e-05, "loss": 3.6705, "step": 178550 }, { "epoch": 2.00648230846576, "grad_norm": 1.1903021335601807, "learning_rate": 1.1976946669512758e-05, "loss": 3.682, "step": 178600 }, { "epoch": 2.0070440336361033, "grad_norm": 1.133834958076477, "learning_rate": 1.1974699756209907e-05, "loss": 3.6267, "step": 178650 }, { "epoch": 2.0076057588064464, "grad_norm": 0.9919807314872742, "learning_rate": 1.1972452842907056e-05, "loss": 3.6693, "step": 178700 }, { "epoch": 2.0081674839767896, "grad_norm": 1.1761101484298706, "learning_rate": 1.1970205929604206e-05, "loss": 3.6217, "step": 178750 }, { "epoch": 2.0087292091471327, "grad_norm": 1.1369194984436035, "learning_rate": 1.1968003954567413e-05, "loss": 3.6111, "step": 178800 }, { "epoch": 2.009290934317476, "grad_norm": 1.0703468322753906, "learning_rate": 1.1965757041264562e-05, "loss": 3.677, "step": 178850 }, { "epoch": 2.009852659487819, "grad_norm": 1.1398569345474243, "learning_rate": 1.1963510127961713e-05, "loss": 3.5947, "step": 178900 }, { "epoch": 2.010414384658162, "grad_norm": 1.134208083152771, "learning_rate": 1.1961263214658863e-05, "loss": 3.6926, "step": 178950 }, { "epoch": 2.0109761098285053, "grad_norm": 1.1300833225250244, "learning_rate": 1.1959016301356012e-05, "loss": 3.6604, "step": 179000 }, { "epoch": 2.0115378349988484, "grad_norm": 1.3578588962554932, "learning_rate": 1.1956769388053161e-05, "loss": 3.7019, "step": 179050 }, { "epoch": 2.0120995601691916, "grad_norm": 1.3306846618652344, "learning_rate": 1.1954522474750312e-05, "loss": 3.7365, "step": 179100 }, { "epoch": 2.0126612853395347, "grad_norm": 1.1825461387634277, "learning_rate": 1.1952275561447462e-05, "loss": 3.6839, "step": 179150 }, { "epoch": 2.013223010509878, "grad_norm": 1.1960430145263672, "learning_rate": 1.1950028648144611e-05, "loss": 3.6459, "step": 179200 }, { "epoch": 2.013784735680221, "grad_norm": 1.111112117767334, "learning_rate": 1.1947781734841762e-05, "loss": 3.5988, "step": 179250 }, { "epoch": 2.014346460850564, "grad_norm": 1.2758218050003052, "learning_rate": 1.1945534821538911e-05, "loss": 3.6658, "step": 179300 }, { "epoch": 2.0149081860209073, "grad_norm": 1.2596542835235596, "learning_rate": 1.194328790823606e-05, "loss": 3.6152, "step": 179350 }, { "epoch": 2.0154699111912504, "grad_norm": 1.0755528211593628, "learning_rate": 1.194104099493321e-05, "loss": 3.5839, "step": 179400 }, { "epoch": 2.0160316363615935, "grad_norm": 1.2503666877746582, "learning_rate": 1.1938794081630361e-05, "loss": 3.6438, "step": 179450 }, { "epoch": 2.0165933615319367, "grad_norm": 1.22162663936615, "learning_rate": 1.193654716832751e-05, "loss": 3.6185, "step": 179500 }, { "epoch": 2.01715508670228, "grad_norm": 1.2688690423965454, "learning_rate": 1.193430025502466e-05, "loss": 3.6753, "step": 179550 }, { "epoch": 2.0177168118726234, "grad_norm": 1.0360270738601685, "learning_rate": 1.193205334172181e-05, "loss": 3.6716, "step": 179600 }, { "epoch": 2.0182785370429666, "grad_norm": 1.278264045715332, "learning_rate": 1.192980642841896e-05, "loss": 3.6364, "step": 179650 }, { "epoch": 2.0188402622133097, "grad_norm": 1.4267991781234741, "learning_rate": 1.192755951511611e-05, "loss": 3.6839, "step": 179700 }, { "epoch": 2.019401987383653, "grad_norm": 1.0959867238998413, "learning_rate": 1.1925312601813259e-05, "loss": 3.6806, "step": 179750 }, { "epoch": 2.019963712553996, "grad_norm": 1.009420394897461, "learning_rate": 1.192306568851041e-05, "loss": 3.6266, "step": 179800 }, { "epoch": 2.020525437724339, "grad_norm": 1.0204261541366577, "learning_rate": 1.1920818775207559e-05, "loss": 3.6664, "step": 179850 }, { "epoch": 2.0210871628946823, "grad_norm": 1.2288864850997925, "learning_rate": 1.1918571861904708e-05, "loss": 3.6486, "step": 179900 }, { "epoch": 2.0216488880650254, "grad_norm": 1.0590049028396606, "learning_rate": 1.1916324948601858e-05, "loss": 3.651, "step": 179950 }, { "epoch": 2.0222106132353685, "grad_norm": 1.2189126014709473, "learning_rate": 1.1914078035299009e-05, "loss": 3.6959, "step": 180000 }, { "epoch": 2.0227723384057117, "grad_norm": 1.2482175827026367, "learning_rate": 1.1911831121996158e-05, "loss": 3.715, "step": 180050 }, { "epoch": 2.023334063576055, "grad_norm": 1.0789780616760254, "learning_rate": 1.1909584208693307e-05, "loss": 3.6599, "step": 180100 }, { "epoch": 2.023895788746398, "grad_norm": 1.0925908088684082, "learning_rate": 1.1907337295390458e-05, "loss": 3.6499, "step": 180150 }, { "epoch": 2.024457513916741, "grad_norm": 1.1479415893554688, "learning_rate": 1.1905090382087608e-05, "loss": 3.7215, "step": 180200 }, { "epoch": 2.0250192390870843, "grad_norm": 1.3037455081939697, "learning_rate": 1.1902843468784757e-05, "loss": 3.6362, "step": 180250 }, { "epoch": 2.0255809642574274, "grad_norm": 1.1210711002349854, "learning_rate": 1.1900596555481906e-05, "loss": 3.6479, "step": 180300 }, { "epoch": 2.0261426894277705, "grad_norm": 1.0629558563232422, "learning_rate": 1.1898349642179057e-05, "loss": 3.6371, "step": 180350 }, { "epoch": 2.0267044145981137, "grad_norm": 1.052786946296692, "learning_rate": 1.1896102728876207e-05, "loss": 3.712, "step": 180400 }, { "epoch": 2.027266139768457, "grad_norm": 1.3238290548324585, "learning_rate": 1.1893855815573356e-05, "loss": 3.6417, "step": 180450 }, { "epoch": 2.0278278649388, "grad_norm": 0.9874387383460999, "learning_rate": 1.1891608902270507e-05, "loss": 3.6863, "step": 180500 }, { "epoch": 2.028389590109143, "grad_norm": 1.063574194908142, "learning_rate": 1.1889361988967656e-05, "loss": 3.6966, "step": 180550 }, { "epoch": 2.0289513152794862, "grad_norm": 1.119516372680664, "learning_rate": 1.1887115075664805e-05, "loss": 3.6497, "step": 180600 }, { "epoch": 2.0295130404498294, "grad_norm": 1.3257489204406738, "learning_rate": 1.1884868162361955e-05, "loss": 3.5998, "step": 180650 }, { "epoch": 2.0300747656201725, "grad_norm": 1.4712659120559692, "learning_rate": 1.1882621249059106e-05, "loss": 3.7032, "step": 180700 }, { "epoch": 2.0306364907905157, "grad_norm": 1.154805302619934, "learning_rate": 1.1880374335756255e-05, "loss": 3.6574, "step": 180750 }, { "epoch": 2.031198215960859, "grad_norm": 1.1678481101989746, "learning_rate": 1.1878127422453404e-05, "loss": 3.7243, "step": 180800 }, { "epoch": 2.031759941131202, "grad_norm": 1.2385306358337402, "learning_rate": 1.1875880509150555e-05, "loss": 3.6573, "step": 180850 }, { "epoch": 2.032321666301545, "grad_norm": 1.1208651065826416, "learning_rate": 1.1873633595847705e-05, "loss": 3.6187, "step": 180900 }, { "epoch": 2.0328833914718887, "grad_norm": 1.2408610582351685, "learning_rate": 1.1871386682544854e-05, "loss": 3.7124, "step": 180950 }, { "epoch": 2.033445116642232, "grad_norm": 1.1662001609802246, "learning_rate": 1.1869139769242003e-05, "loss": 3.6918, "step": 181000 }, { "epoch": 2.034006841812575, "grad_norm": 1.667796015739441, "learning_rate": 1.1866892855939154e-05, "loss": 3.6008, "step": 181050 }, { "epoch": 2.034568566982918, "grad_norm": 1.2891185283660889, "learning_rate": 1.1864645942636304e-05, "loss": 3.692, "step": 181100 }, { "epoch": 2.0351302921532612, "grad_norm": 1.0154170989990234, "learning_rate": 1.1862399029333453e-05, "loss": 3.7017, "step": 181150 }, { "epoch": 2.0356920173236044, "grad_norm": 1.350826382637024, "learning_rate": 1.1860152116030604e-05, "loss": 3.6805, "step": 181200 }, { "epoch": 2.0362537424939475, "grad_norm": 1.127942442893982, "learning_rate": 1.185795014099381e-05, "loss": 3.6208, "step": 181250 }, { "epoch": 2.0368154676642907, "grad_norm": 1.4027514457702637, "learning_rate": 1.185570322769096e-05, "loss": 3.612, "step": 181300 }, { "epoch": 2.037377192834634, "grad_norm": 1.2916492223739624, "learning_rate": 1.185345631438811e-05, "loss": 3.6641, "step": 181350 }, { "epoch": 2.037938918004977, "grad_norm": 1.3983794450759888, "learning_rate": 1.185120940108526e-05, "loss": 3.6643, "step": 181400 }, { "epoch": 2.03850064317532, "grad_norm": 1.2258174419403076, "learning_rate": 1.1848962487782409e-05, "loss": 3.6637, "step": 181450 }, { "epoch": 2.0390623683456632, "grad_norm": 1.2498546838760376, "learning_rate": 1.184671557447956e-05, "loss": 3.6297, "step": 181500 }, { "epoch": 2.0396240935160064, "grad_norm": 1.1554557085037231, "learning_rate": 1.184446866117671e-05, "loss": 3.7559, "step": 181550 }, { "epoch": 2.0401858186863495, "grad_norm": 1.154598355293274, "learning_rate": 1.1842221747873859e-05, "loss": 3.6876, "step": 181600 }, { "epoch": 2.0407475438566927, "grad_norm": 1.2266391515731812, "learning_rate": 1.1839974834571008e-05, "loss": 3.6902, "step": 181650 }, { "epoch": 2.041309269027036, "grad_norm": 1.2957992553710938, "learning_rate": 1.1837727921268159e-05, "loss": 3.6062, "step": 181700 }, { "epoch": 2.041870994197379, "grad_norm": 0.9922723770141602, "learning_rate": 1.1835481007965308e-05, "loss": 3.5309, "step": 181750 }, { "epoch": 2.042432719367722, "grad_norm": 1.1541030406951904, "learning_rate": 1.1833234094662458e-05, "loss": 3.6341, "step": 181800 }, { "epoch": 2.0429944445380652, "grad_norm": 1.2921535968780518, "learning_rate": 1.1830987181359609e-05, "loss": 3.6493, "step": 181850 }, { "epoch": 2.0435561697084084, "grad_norm": 1.2229830026626587, "learning_rate": 1.1828740268056758e-05, "loss": 3.7006, "step": 181900 }, { "epoch": 2.0441178948787515, "grad_norm": 1.2174954414367676, "learning_rate": 1.1826493354753907e-05, "loss": 3.6276, "step": 181950 }, { "epoch": 2.0446796200490946, "grad_norm": 1.0949270725250244, "learning_rate": 1.1824246441451057e-05, "loss": 3.6625, "step": 182000 }, { "epoch": 2.045241345219438, "grad_norm": 1.0617477893829346, "learning_rate": 1.1821999528148208e-05, "loss": 3.6732, "step": 182050 }, { "epoch": 2.045803070389781, "grad_norm": 1.1221001148223877, "learning_rate": 1.1819752614845357e-05, "loss": 3.609, "step": 182100 }, { "epoch": 2.046364795560124, "grad_norm": 1.1541255712509155, "learning_rate": 1.1817505701542506e-05, "loss": 3.683, "step": 182150 }, { "epoch": 2.046926520730467, "grad_norm": 1.1322633028030396, "learning_rate": 1.1815258788239657e-05, "loss": 3.6044, "step": 182200 }, { "epoch": 2.0474882459008104, "grad_norm": 1.0582863092422485, "learning_rate": 1.1813011874936806e-05, "loss": 3.6516, "step": 182250 }, { "epoch": 2.048049971071154, "grad_norm": 1.0736987590789795, "learning_rate": 1.1810764961633956e-05, "loss": 3.6536, "step": 182300 }, { "epoch": 2.048611696241497, "grad_norm": 1.0526435375213623, "learning_rate": 1.1808518048331105e-05, "loss": 3.6763, "step": 182350 }, { "epoch": 2.0491734214118402, "grad_norm": 1.4919220209121704, "learning_rate": 1.1806271135028256e-05, "loss": 3.6844, "step": 182400 }, { "epoch": 2.0497351465821834, "grad_norm": 1.0388952493667603, "learning_rate": 1.1804024221725405e-05, "loss": 3.6663, "step": 182450 }, { "epoch": 2.0502968717525265, "grad_norm": 1.1224077939987183, "learning_rate": 1.1801777308422555e-05, "loss": 3.6662, "step": 182500 }, { "epoch": 2.0508585969228696, "grad_norm": 1.2052985429763794, "learning_rate": 1.1799530395119704e-05, "loss": 3.589, "step": 182550 }, { "epoch": 2.051420322093213, "grad_norm": 1.1244508028030396, "learning_rate": 1.1797283481816855e-05, "loss": 3.7222, "step": 182600 }, { "epoch": 2.051982047263556, "grad_norm": 1.211931824684143, "learning_rate": 1.1795036568514004e-05, "loss": 3.6968, "step": 182650 }, { "epoch": 2.052543772433899, "grad_norm": 1.1259714365005493, "learning_rate": 1.1792789655211154e-05, "loss": 3.7227, "step": 182700 }, { "epoch": 2.053105497604242, "grad_norm": 1.1324070692062378, "learning_rate": 1.1790542741908305e-05, "loss": 3.6529, "step": 182750 }, { "epoch": 2.0536672227745854, "grad_norm": 1.025359869003296, "learning_rate": 1.1788295828605454e-05, "loss": 3.6157, "step": 182800 }, { "epoch": 2.0542289479449285, "grad_norm": 1.016520380973816, "learning_rate": 1.1786048915302603e-05, "loss": 3.6116, "step": 182850 }, { "epoch": 2.0547906731152716, "grad_norm": 1.1791082620620728, "learning_rate": 1.1783802001999753e-05, "loss": 3.6682, "step": 182900 }, { "epoch": 2.055352398285615, "grad_norm": 1.1866360902786255, "learning_rate": 1.1781555088696904e-05, "loss": 3.6468, "step": 182950 }, { "epoch": 2.055914123455958, "grad_norm": 1.1281777620315552, "learning_rate": 1.1779308175394053e-05, "loss": 3.6572, "step": 183000 }, { "epoch": 2.056475848626301, "grad_norm": 1.286525845527649, "learning_rate": 1.1777061262091202e-05, "loss": 3.6769, "step": 183050 }, { "epoch": 2.057037573796644, "grad_norm": 1.0334420204162598, "learning_rate": 1.177485928705441e-05, "loss": 3.6316, "step": 183100 }, { "epoch": 2.0575992989669873, "grad_norm": 1.4133697748184204, "learning_rate": 1.177261237375156e-05, "loss": 3.6962, "step": 183150 }, { "epoch": 2.0581610241373305, "grad_norm": 1.1942658424377441, "learning_rate": 1.177036546044871e-05, "loss": 3.7566, "step": 183200 }, { "epoch": 2.0587227493076736, "grad_norm": 1.0915184020996094, "learning_rate": 1.176811854714586e-05, "loss": 3.5812, "step": 183250 }, { "epoch": 2.0592844744780168, "grad_norm": 1.1224225759506226, "learning_rate": 1.1765871633843009e-05, "loss": 3.6077, "step": 183300 }, { "epoch": 2.05984619964836, "grad_norm": 1.21129310131073, "learning_rate": 1.1763624720540158e-05, "loss": 3.6295, "step": 183350 }, { "epoch": 2.060407924818703, "grad_norm": 1.2197438478469849, "learning_rate": 1.176137780723731e-05, "loss": 3.6664, "step": 183400 }, { "epoch": 2.060969649989046, "grad_norm": 1.0733387470245361, "learning_rate": 1.1759130893934459e-05, "loss": 3.6526, "step": 183450 }, { "epoch": 2.0615313751593893, "grad_norm": 1.2364966869354248, "learning_rate": 1.1756883980631608e-05, "loss": 3.6675, "step": 183500 }, { "epoch": 2.0620931003297325, "grad_norm": 1.0612620115280151, "learning_rate": 1.1754637067328757e-05, "loss": 3.6113, "step": 183550 }, { "epoch": 2.0626548255000756, "grad_norm": 1.0921076536178589, "learning_rate": 1.1752390154025908e-05, "loss": 3.6378, "step": 183600 }, { "epoch": 2.063216550670419, "grad_norm": 1.0724772214889526, "learning_rate": 1.1750143240723057e-05, "loss": 3.6179, "step": 183650 }, { "epoch": 2.0637782758407623, "grad_norm": 1.191307783126831, "learning_rate": 1.1747896327420207e-05, "loss": 3.6389, "step": 183700 }, { "epoch": 2.0643400010111055, "grad_norm": 1.1037898063659668, "learning_rate": 1.1745649414117358e-05, "loss": 3.6549, "step": 183750 }, { "epoch": 2.0649017261814486, "grad_norm": 1.1357640027999878, "learning_rate": 1.1743402500814507e-05, "loss": 3.73, "step": 183800 }, { "epoch": 2.0654634513517918, "grad_norm": 1.1400688886642456, "learning_rate": 1.1741155587511656e-05, "loss": 3.6958, "step": 183850 }, { "epoch": 2.066025176522135, "grad_norm": 1.2211834192276, "learning_rate": 1.1738908674208806e-05, "loss": 3.6394, "step": 183900 }, { "epoch": 2.066586901692478, "grad_norm": 1.3305988311767578, "learning_rate": 1.1736661760905957e-05, "loss": 3.6558, "step": 183950 }, { "epoch": 2.067148626862821, "grad_norm": 1.3925904035568237, "learning_rate": 1.1734414847603106e-05, "loss": 3.5734, "step": 184000 }, { "epoch": 2.0677103520331643, "grad_norm": 1.1970486640930176, "learning_rate": 1.1732167934300255e-05, "loss": 3.6224, "step": 184050 }, { "epoch": 2.0682720772035075, "grad_norm": 1.1737490892410278, "learning_rate": 1.1729921020997406e-05, "loss": 3.7166, "step": 184100 }, { "epoch": 2.0688338023738506, "grad_norm": 1.2897924184799194, "learning_rate": 1.1727674107694556e-05, "loss": 3.7132, "step": 184150 }, { "epoch": 2.0693955275441938, "grad_norm": 1.4397519826889038, "learning_rate": 1.1725427194391705e-05, "loss": 3.6388, "step": 184200 }, { "epoch": 2.069957252714537, "grad_norm": 1.2425822019577026, "learning_rate": 1.1723180281088854e-05, "loss": 3.6275, "step": 184250 }, { "epoch": 2.07051897788488, "grad_norm": 1.143621802330017, "learning_rate": 1.1720933367786005e-05, "loss": 3.6994, "step": 184300 }, { "epoch": 2.071080703055223, "grad_norm": 1.2500793933868408, "learning_rate": 1.1718686454483155e-05, "loss": 3.6829, "step": 184350 }, { "epoch": 2.0716424282255663, "grad_norm": 1.2162973880767822, "learning_rate": 1.1716439541180304e-05, "loss": 3.6092, "step": 184400 }, { "epoch": 2.0722041533959095, "grad_norm": 1.1496464014053345, "learning_rate": 1.1714192627877455e-05, "loss": 3.6196, "step": 184450 }, { "epoch": 2.0727658785662526, "grad_norm": 1.0511910915374756, "learning_rate": 1.1711945714574604e-05, "loss": 3.6539, "step": 184500 }, { "epoch": 2.0733276037365957, "grad_norm": 1.103660225868225, "learning_rate": 1.1709698801271754e-05, "loss": 3.6648, "step": 184550 }, { "epoch": 2.073889328906939, "grad_norm": 1.0108928680419922, "learning_rate": 1.1707451887968903e-05, "loss": 3.7308, "step": 184600 }, { "epoch": 2.074451054077282, "grad_norm": 1.1093605756759644, "learning_rate": 1.1705204974666054e-05, "loss": 3.6693, "step": 184650 }, { "epoch": 2.075012779247625, "grad_norm": 1.1762808561325073, "learning_rate": 1.1702958061363203e-05, "loss": 3.6235, "step": 184700 }, { "epoch": 2.0755745044179683, "grad_norm": 1.3621156215667725, "learning_rate": 1.1700711148060353e-05, "loss": 3.5564, "step": 184750 }, { "epoch": 2.0761362295883115, "grad_norm": 1.2420388460159302, "learning_rate": 1.1698464234757504e-05, "loss": 3.6166, "step": 184800 }, { "epoch": 2.0766979547586546, "grad_norm": 1.2055463790893555, "learning_rate": 1.1696217321454653e-05, "loss": 3.7241, "step": 184850 }, { "epoch": 2.0772596799289977, "grad_norm": 1.192975401878357, "learning_rate": 1.1693970408151802e-05, "loss": 3.7106, "step": 184900 }, { "epoch": 2.077821405099341, "grad_norm": 1.2383537292480469, "learning_rate": 1.1691723494848952e-05, "loss": 3.6599, "step": 184950 }, { "epoch": 2.0783831302696845, "grad_norm": 1.301489233970642, "learning_rate": 1.1689476581546103e-05, "loss": 3.6764, "step": 185000 }, { "epoch": 2.0789448554400276, "grad_norm": 1.3114230632781982, "learning_rate": 1.1687229668243252e-05, "loss": 3.6454, "step": 185050 }, { "epoch": 2.0795065806103707, "grad_norm": 1.1451566219329834, "learning_rate": 1.1684982754940401e-05, "loss": 3.6942, "step": 185100 }, { "epoch": 2.080068305780714, "grad_norm": 1.0022562742233276, "learning_rate": 1.168273584163755e-05, "loss": 3.6398, "step": 185150 }, { "epoch": 2.080630030951057, "grad_norm": 1.2197297811508179, "learning_rate": 1.1680488928334702e-05, "loss": 3.6824, "step": 185200 }, { "epoch": 2.0811917561214, "grad_norm": 1.25179123878479, "learning_rate": 1.1678242015031851e-05, "loss": 3.6963, "step": 185250 }, { "epoch": 2.0817534812917433, "grad_norm": 1.1346317529678345, "learning_rate": 1.1675995101729e-05, "loss": 3.731, "step": 185300 }, { "epoch": 2.0823152064620865, "grad_norm": 1.211344838142395, "learning_rate": 1.1673748188426151e-05, "loss": 3.6652, "step": 185350 }, { "epoch": 2.0828769316324296, "grad_norm": 1.2521421909332275, "learning_rate": 1.16715012751233e-05, "loss": 3.7182, "step": 185400 }, { "epoch": 2.0834386568027727, "grad_norm": 1.1842113733291626, "learning_rate": 1.166925436182045e-05, "loss": 3.6045, "step": 185450 }, { "epoch": 2.084000381973116, "grad_norm": 1.0520801544189453, "learning_rate": 1.1667007448517599e-05, "loss": 3.6785, "step": 185500 }, { "epoch": 2.084562107143459, "grad_norm": 1.1933637857437134, "learning_rate": 1.166476053521475e-05, "loss": 3.5994, "step": 185550 }, { "epoch": 2.085123832313802, "grad_norm": 1.1843327283859253, "learning_rate": 1.16625136219119e-05, "loss": 3.688, "step": 185600 }, { "epoch": 2.0856855574841453, "grad_norm": 1.059524655342102, "learning_rate": 1.1660266708609049e-05, "loss": 3.6427, "step": 185650 }, { "epoch": 2.0862472826544884, "grad_norm": 1.3903207778930664, "learning_rate": 1.16580197953062e-05, "loss": 3.6132, "step": 185700 }, { "epoch": 2.0868090078248316, "grad_norm": 1.3105016946792603, "learning_rate": 1.1655772882003349e-05, "loss": 3.6907, "step": 185750 }, { "epoch": 2.0873707329951747, "grad_norm": 1.014356255531311, "learning_rate": 1.1653525968700498e-05, "loss": 3.6845, "step": 185800 }, { "epoch": 2.087932458165518, "grad_norm": 1.4541751146316528, "learning_rate": 1.1651279055397648e-05, "loss": 3.6286, "step": 185850 }, { "epoch": 2.088494183335861, "grad_norm": 1.0511747598648071, "learning_rate": 1.1649032142094799e-05, "loss": 3.7113, "step": 185900 }, { "epoch": 2.089055908506204, "grad_norm": 1.2117619514465332, "learning_rate": 1.1646785228791948e-05, "loss": 3.6283, "step": 185950 }, { "epoch": 2.0896176336765473, "grad_norm": 1.050269365310669, "learning_rate": 1.1644538315489097e-05, "loss": 3.6544, "step": 186000 }, { "epoch": 2.0901793588468904, "grad_norm": 1.167660117149353, "learning_rate": 1.1642291402186248e-05, "loss": 3.6782, "step": 186050 }, { "epoch": 2.0907410840172336, "grad_norm": 1.171392560005188, "learning_rate": 1.1640044488883398e-05, "loss": 3.6661, "step": 186100 }, { "epoch": 2.0913028091875767, "grad_norm": 1.2858030796051025, "learning_rate": 1.1637797575580547e-05, "loss": 3.6055, "step": 186150 }, { "epoch": 2.09186453435792, "grad_norm": 1.2579023838043213, "learning_rate": 1.1635550662277696e-05, "loss": 3.6953, "step": 186200 }, { "epoch": 2.092426259528263, "grad_norm": 1.1274867057800293, "learning_rate": 1.1633303748974847e-05, "loss": 3.5906, "step": 186250 }, { "epoch": 2.092987984698606, "grad_norm": 1.3431787490844727, "learning_rate": 1.1631056835671997e-05, "loss": 3.6835, "step": 186300 }, { "epoch": 2.0935497098689497, "grad_norm": 1.0384411811828613, "learning_rate": 1.1628809922369146e-05, "loss": 3.6966, "step": 186350 }, { "epoch": 2.094111435039293, "grad_norm": 1.216713547706604, "learning_rate": 1.1626563009066297e-05, "loss": 3.6826, "step": 186400 }, { "epoch": 2.094673160209636, "grad_norm": 1.3694875240325928, "learning_rate": 1.1624316095763446e-05, "loss": 3.6066, "step": 186450 }, { "epoch": 2.095234885379979, "grad_norm": 1.1893595457077026, "learning_rate": 1.1622069182460596e-05, "loss": 3.6106, "step": 186500 }, { "epoch": 2.0957966105503223, "grad_norm": 1.1873623132705688, "learning_rate": 1.1619822269157745e-05, "loss": 3.7164, "step": 186550 }, { "epoch": 2.0963583357206654, "grad_norm": 1.196428656578064, "learning_rate": 1.1617575355854896e-05, "loss": 3.5973, "step": 186600 }, { "epoch": 2.0969200608910086, "grad_norm": 1.1118874549865723, "learning_rate": 1.1615328442552045e-05, "loss": 3.6804, "step": 186650 }, { "epoch": 2.0974817860613517, "grad_norm": 1.3892502784729004, "learning_rate": 1.1613081529249195e-05, "loss": 3.7129, "step": 186700 }, { "epoch": 2.098043511231695, "grad_norm": 1.0746665000915527, "learning_rate": 1.1610879554212402e-05, "loss": 3.6837, "step": 186750 }, { "epoch": 2.098605236402038, "grad_norm": 1.3109077215194702, "learning_rate": 1.1608632640909552e-05, "loss": 3.6782, "step": 186800 }, { "epoch": 2.099166961572381, "grad_norm": 1.3797475099563599, "learning_rate": 1.1606385727606701e-05, "loss": 3.7002, "step": 186850 }, { "epoch": 2.0997286867427243, "grad_norm": 1.1863510608673096, "learning_rate": 1.1604138814303852e-05, "loss": 3.6765, "step": 186900 }, { "epoch": 2.1002904119130674, "grad_norm": 1.1817659139633179, "learning_rate": 1.1601891901001001e-05, "loss": 3.698, "step": 186950 }, { "epoch": 2.1008521370834106, "grad_norm": 1.0148324966430664, "learning_rate": 1.159964498769815e-05, "loss": 3.6419, "step": 187000 }, { "epoch": 2.1014138622537537, "grad_norm": 1.1663976907730103, "learning_rate": 1.1597398074395302e-05, "loss": 3.6567, "step": 187050 }, { "epoch": 2.101975587424097, "grad_norm": 1.1736410856246948, "learning_rate": 1.159515116109245e-05, "loss": 3.6098, "step": 187100 }, { "epoch": 2.10253731259444, "grad_norm": 1.2112184762954712, "learning_rate": 1.15929042477896e-05, "loss": 3.5847, "step": 187150 }, { "epoch": 2.103099037764783, "grad_norm": 1.570095419883728, "learning_rate": 1.159065733448675e-05, "loss": 3.6716, "step": 187200 }, { "epoch": 2.1036607629351263, "grad_norm": 1.0069478750228882, "learning_rate": 1.15884104211839e-05, "loss": 3.6609, "step": 187250 }, { "epoch": 2.1042224881054694, "grad_norm": 0.9928694367408752, "learning_rate": 1.158616350788105e-05, "loss": 3.6824, "step": 187300 }, { "epoch": 2.1047842132758126, "grad_norm": 1.1236019134521484, "learning_rate": 1.1583916594578199e-05, "loss": 3.6956, "step": 187350 }, { "epoch": 2.1053459384461557, "grad_norm": 1.145443081855774, "learning_rate": 1.158166968127535e-05, "loss": 3.5721, "step": 187400 }, { "epoch": 2.105907663616499, "grad_norm": 1.148145318031311, "learning_rate": 1.15794227679725e-05, "loss": 3.5823, "step": 187450 }, { "epoch": 2.106469388786842, "grad_norm": 1.2535157203674316, "learning_rate": 1.1577175854669649e-05, "loss": 3.6493, "step": 187500 }, { "epoch": 2.107031113957185, "grad_norm": 2.8420257568359375, "learning_rate": 1.1574928941366798e-05, "loss": 3.6648, "step": 187550 }, { "epoch": 2.1075928391275283, "grad_norm": 1.2739931344985962, "learning_rate": 1.1572682028063949e-05, "loss": 3.5955, "step": 187600 }, { "epoch": 2.1081545642978714, "grad_norm": 1.2763255834579468, "learning_rate": 1.1570435114761098e-05, "loss": 3.6826, "step": 187650 }, { "epoch": 2.108716289468215, "grad_norm": 1.135329246520996, "learning_rate": 1.1568188201458248e-05, "loss": 3.5985, "step": 187700 }, { "epoch": 2.109278014638558, "grad_norm": 1.0852805376052856, "learning_rate": 1.1565941288155397e-05, "loss": 3.7168, "step": 187750 }, { "epoch": 2.1098397398089013, "grad_norm": 1.18838632106781, "learning_rate": 1.1563694374852548e-05, "loss": 3.5967, "step": 187800 }, { "epoch": 2.1104014649792444, "grad_norm": 1.0903185606002808, "learning_rate": 1.1561447461549697e-05, "loss": 3.6357, "step": 187850 }, { "epoch": 2.1109631901495876, "grad_norm": 1.1050207614898682, "learning_rate": 1.1559200548246847e-05, "loss": 3.6751, "step": 187900 }, { "epoch": 2.1115249153199307, "grad_norm": 1.1145532131195068, "learning_rate": 1.1556953634943998e-05, "loss": 3.6524, "step": 187950 }, { "epoch": 2.112086640490274, "grad_norm": 1.222307562828064, "learning_rate": 1.1554706721641147e-05, "loss": 3.667, "step": 188000 }, { "epoch": 2.112648365660617, "grad_norm": 1.4427876472473145, "learning_rate": 1.1552459808338296e-05, "loss": 3.6596, "step": 188050 }, { "epoch": 2.11321009083096, "grad_norm": 1.021989107131958, "learning_rate": 1.1550212895035446e-05, "loss": 3.6318, "step": 188100 }, { "epoch": 2.1137718160013033, "grad_norm": 1.2242029905319214, "learning_rate": 1.1547965981732597e-05, "loss": 3.594, "step": 188150 }, { "epoch": 2.1143335411716464, "grad_norm": 1.1787387132644653, "learning_rate": 1.1545719068429746e-05, "loss": 3.6721, "step": 188200 }, { "epoch": 2.1148952663419895, "grad_norm": 1.0349905490875244, "learning_rate": 1.1543472155126895e-05, "loss": 3.7216, "step": 188250 }, { "epoch": 2.1154569915123327, "grad_norm": 1.1641631126403809, "learning_rate": 1.1541225241824046e-05, "loss": 3.6352, "step": 188300 }, { "epoch": 2.116018716682676, "grad_norm": 1.2305693626403809, "learning_rate": 1.1538978328521196e-05, "loss": 3.6903, "step": 188350 }, { "epoch": 2.116580441853019, "grad_norm": 1.3482508659362793, "learning_rate": 1.1536731415218345e-05, "loss": 3.6506, "step": 188400 }, { "epoch": 2.117142167023362, "grad_norm": 1.0359673500061035, "learning_rate": 1.1534484501915494e-05, "loss": 3.6182, "step": 188450 }, { "epoch": 2.1177038921937053, "grad_norm": 1.2034860849380493, "learning_rate": 1.1532237588612645e-05, "loss": 3.6963, "step": 188500 }, { "epoch": 2.1182656173640484, "grad_norm": 1.143668293952942, "learning_rate": 1.1529990675309795e-05, "loss": 3.7738, "step": 188550 }, { "epoch": 2.1188273425343915, "grad_norm": 1.3331931829452515, "learning_rate": 1.1527743762006944e-05, "loss": 3.5772, "step": 188600 }, { "epoch": 2.1193890677047347, "grad_norm": 1.3241474628448486, "learning_rate": 1.1525496848704095e-05, "loss": 3.653, "step": 188650 }, { "epoch": 2.119950792875078, "grad_norm": 1.1467458009719849, "learning_rate": 1.1523249935401244e-05, "loss": 3.7084, "step": 188700 }, { "epoch": 2.120512518045421, "grad_norm": 1.1851401329040527, "learning_rate": 1.1521003022098394e-05, "loss": 3.6817, "step": 188750 }, { "epoch": 2.121074243215764, "grad_norm": 1.1613062620162964, "learning_rate": 1.1518756108795543e-05, "loss": 3.6369, "step": 188800 }, { "epoch": 2.1216359683861072, "grad_norm": 1.0273969173431396, "learning_rate": 1.1516509195492694e-05, "loss": 3.7027, "step": 188850 }, { "epoch": 2.1221976935564504, "grad_norm": 1.080709457397461, "learning_rate": 1.1514262282189843e-05, "loss": 3.6969, "step": 188900 }, { "epoch": 2.1227594187267935, "grad_norm": 1.1772719621658325, "learning_rate": 1.1512015368886992e-05, "loss": 3.6234, "step": 188950 }, { "epoch": 2.1233211438971367, "grad_norm": 1.0060964822769165, "learning_rate": 1.1509768455584143e-05, "loss": 3.6852, "step": 189000 }, { "epoch": 2.1238828690674803, "grad_norm": 1.2354568243026733, "learning_rate": 1.1507521542281293e-05, "loss": 3.6297, "step": 189050 }, { "epoch": 2.1244445942378234, "grad_norm": 1.106758952140808, "learning_rate": 1.1505274628978442e-05, "loss": 3.6663, "step": 189100 }, { "epoch": 2.1250063194081665, "grad_norm": 1.177374243736267, "learning_rate": 1.1503027715675591e-05, "loss": 3.7331, "step": 189150 }, { "epoch": 2.1255680445785097, "grad_norm": 1.298965334892273, "learning_rate": 1.1500780802372742e-05, "loss": 3.5916, "step": 189200 }, { "epoch": 2.126129769748853, "grad_norm": 1.2143000364303589, "learning_rate": 1.1498533889069892e-05, "loss": 3.7088, "step": 189250 }, { "epoch": 2.126691494919196, "grad_norm": 1.0102218389511108, "learning_rate": 1.1496286975767041e-05, "loss": 3.6115, "step": 189300 }, { "epoch": 2.127253220089539, "grad_norm": 1.1801633834838867, "learning_rate": 1.149404006246419e-05, "loss": 3.6492, "step": 189350 }, { "epoch": 2.1278149452598822, "grad_norm": 1.0461994409561157, "learning_rate": 1.1491793149161341e-05, "loss": 3.6076, "step": 189400 }, { "epoch": 2.1283766704302254, "grad_norm": 1.0831447839736938, "learning_rate": 1.148954623585849e-05, "loss": 3.6448, "step": 189450 }, { "epoch": 2.1289383956005685, "grad_norm": 1.1427998542785645, "learning_rate": 1.148729932255564e-05, "loss": 3.5827, "step": 189500 }, { "epoch": 2.1295001207709117, "grad_norm": 1.3421205282211304, "learning_rate": 1.1485052409252791e-05, "loss": 3.6507, "step": 189550 }, { "epoch": 2.130061845941255, "grad_norm": 1.284712553024292, "learning_rate": 1.148280549594994e-05, "loss": 3.6311, "step": 189600 }, { "epoch": 2.130623571111598, "grad_norm": 1.7064614295959473, "learning_rate": 1.148055858264709e-05, "loss": 3.6555, "step": 189650 }, { "epoch": 2.131185296281941, "grad_norm": 1.0825332403182983, "learning_rate": 1.1478311669344239e-05, "loss": 3.6579, "step": 189700 }, { "epoch": 2.1317470214522842, "grad_norm": 1.303828477859497, "learning_rate": 1.1476109694307447e-05, "loss": 3.6228, "step": 189750 }, { "epoch": 2.1323087466226274, "grad_norm": 1.2739299535751343, "learning_rate": 1.1473862781004596e-05, "loss": 3.7096, "step": 189800 }, { "epoch": 2.1328704717929705, "grad_norm": 1.215766191482544, "learning_rate": 1.1471615867701747e-05, "loss": 3.7074, "step": 189850 }, { "epoch": 2.1334321969633137, "grad_norm": 1.0622475147247314, "learning_rate": 1.1469368954398896e-05, "loss": 3.7047, "step": 189900 }, { "epoch": 2.133993922133657, "grad_norm": 1.1313506364822388, "learning_rate": 1.1467122041096046e-05, "loss": 3.7185, "step": 189950 }, { "epoch": 2.134555647304, "grad_norm": 1.1207643747329712, "learning_rate": 1.1464875127793197e-05, "loss": 3.6292, "step": 190000 }, { "epoch": 2.135117372474343, "grad_norm": 1.1433231830596924, "learning_rate": 1.1462628214490346e-05, "loss": 3.6774, "step": 190050 }, { "epoch": 2.135679097644686, "grad_norm": 1.3061046600341797, "learning_rate": 1.1460381301187495e-05, "loss": 3.6434, "step": 190100 }, { "epoch": 2.1362408228150294, "grad_norm": 1.1008479595184326, "learning_rate": 1.1458134387884645e-05, "loss": 3.6726, "step": 190150 }, { "epoch": 2.1368025479853725, "grad_norm": 1.1142629384994507, "learning_rate": 1.1455887474581796e-05, "loss": 3.6476, "step": 190200 }, { "epoch": 2.1373642731557156, "grad_norm": 1.2122528553009033, "learning_rate": 1.1453640561278945e-05, "loss": 3.7287, "step": 190250 }, { "epoch": 2.137925998326059, "grad_norm": 1.0298256874084473, "learning_rate": 1.1451393647976094e-05, "loss": 3.6936, "step": 190300 }, { "epoch": 2.138487723496402, "grad_norm": 1.2197506427764893, "learning_rate": 1.1449146734673244e-05, "loss": 3.698, "step": 190350 }, { "epoch": 2.1390494486667455, "grad_norm": 1.2474559545516968, "learning_rate": 1.1446899821370395e-05, "loss": 3.6674, "step": 190400 }, { "epoch": 2.1396111738370887, "grad_norm": 1.1839038133621216, "learning_rate": 1.1444652908067544e-05, "loss": 3.65, "step": 190450 }, { "epoch": 2.140172899007432, "grad_norm": 1.0742826461791992, "learning_rate": 1.1442405994764693e-05, "loss": 3.7048, "step": 190500 }, { "epoch": 2.140734624177775, "grad_norm": 1.9982452392578125, "learning_rate": 1.1440159081461844e-05, "loss": 3.6968, "step": 190550 }, { "epoch": 2.141296349348118, "grad_norm": 1.16074800491333, "learning_rate": 1.1437912168158993e-05, "loss": 3.5948, "step": 190600 }, { "epoch": 2.141858074518461, "grad_norm": 1.2510814666748047, "learning_rate": 1.1435665254856143e-05, "loss": 3.6732, "step": 190650 }, { "epoch": 2.1424197996888044, "grad_norm": 1.1811882257461548, "learning_rate": 1.1433418341553292e-05, "loss": 3.7217, "step": 190700 }, { "epoch": 2.1429815248591475, "grad_norm": 1.0726606845855713, "learning_rate": 1.1431171428250443e-05, "loss": 3.6865, "step": 190750 }, { "epoch": 2.1435432500294906, "grad_norm": 1.2404667139053345, "learning_rate": 1.1428924514947592e-05, "loss": 3.639, "step": 190800 }, { "epoch": 2.144104975199834, "grad_norm": 1.1091114282608032, "learning_rate": 1.1426677601644742e-05, "loss": 3.6803, "step": 190850 }, { "epoch": 2.144666700370177, "grad_norm": 1.1242750883102417, "learning_rate": 1.1424430688341893e-05, "loss": 3.6574, "step": 190900 }, { "epoch": 2.14522842554052, "grad_norm": 0.9307460784912109, "learning_rate": 1.1422183775039042e-05, "loss": 3.7273, "step": 190950 }, { "epoch": 2.145790150710863, "grad_norm": 1.123029112815857, "learning_rate": 1.1419936861736191e-05, "loss": 3.6071, "step": 191000 }, { "epoch": 2.1463518758812064, "grad_norm": 1.162349820137024, "learning_rate": 1.141768994843334e-05, "loss": 3.687, "step": 191050 }, { "epoch": 2.1469136010515495, "grad_norm": 1.164348840713501, "learning_rate": 1.1415443035130492e-05, "loss": 3.6717, "step": 191100 }, { "epoch": 2.1474753262218926, "grad_norm": 1.2084416151046753, "learning_rate": 1.1413196121827641e-05, "loss": 3.6662, "step": 191150 }, { "epoch": 2.1480370513922358, "grad_norm": 1.1981921195983887, "learning_rate": 1.141094920852479e-05, "loss": 3.6786, "step": 191200 }, { "epoch": 2.148598776562579, "grad_norm": 1.047624111175537, "learning_rate": 1.1408702295221941e-05, "loss": 3.7191, "step": 191250 }, { "epoch": 2.149160501732922, "grad_norm": 1.154442548751831, "learning_rate": 1.140645538191909e-05, "loss": 3.6479, "step": 191300 }, { "epoch": 2.149722226903265, "grad_norm": 1.0728384256362915, "learning_rate": 1.140420846861624e-05, "loss": 3.6669, "step": 191350 }, { "epoch": 2.1502839520736083, "grad_norm": 1.2294445037841797, "learning_rate": 1.140196155531339e-05, "loss": 3.6674, "step": 191400 }, { "epoch": 2.1508456772439515, "grad_norm": 1.1478815078735352, "learning_rate": 1.139971464201054e-05, "loss": 3.5864, "step": 191450 }, { "epoch": 2.1514074024142946, "grad_norm": 1.1214373111724854, "learning_rate": 1.139746772870769e-05, "loss": 3.7497, "step": 191500 }, { "epoch": 2.1519691275846378, "grad_norm": 1.183599829673767, "learning_rate": 1.1395220815404839e-05, "loss": 3.6244, "step": 191550 }, { "epoch": 2.152530852754981, "grad_norm": 1.1228039264678955, "learning_rate": 1.139297390210199e-05, "loss": 3.617, "step": 191600 }, { "epoch": 2.1530925779253245, "grad_norm": 1.067888617515564, "learning_rate": 1.139072698879914e-05, "loss": 3.6954, "step": 191650 }, { "epoch": 2.153654303095667, "grad_norm": 1.085471510887146, "learning_rate": 1.1388480075496289e-05, "loss": 3.7293, "step": 191700 }, { "epoch": 2.1542160282660108, "grad_norm": 1.1760141849517822, "learning_rate": 1.1386233162193438e-05, "loss": 3.6734, "step": 191750 }, { "epoch": 2.154777753436354, "grad_norm": 1.1908788681030273, "learning_rate": 1.1383986248890589e-05, "loss": 3.6836, "step": 191800 }, { "epoch": 2.155339478606697, "grad_norm": 1.1994457244873047, "learning_rate": 1.1381739335587738e-05, "loss": 3.5693, "step": 191850 }, { "epoch": 2.15590120377704, "grad_norm": 1.2013858556747437, "learning_rate": 1.1379492422284888e-05, "loss": 3.6542, "step": 191900 }, { "epoch": 2.1564629289473833, "grad_norm": 1.1068100929260254, "learning_rate": 1.1377245508982037e-05, "loss": 3.633, "step": 191950 }, { "epoch": 2.1570246541177265, "grad_norm": 1.1132248640060425, "learning_rate": 1.1374998595679188e-05, "loss": 3.6764, "step": 192000 }, { "epoch": 2.1575863792880696, "grad_norm": 1.1618642807006836, "learning_rate": 1.1372751682376337e-05, "loss": 3.6839, "step": 192050 }, { "epoch": 2.1581481044584128, "grad_norm": 1.1055786609649658, "learning_rate": 1.1370504769073487e-05, "loss": 3.6788, "step": 192100 }, { "epoch": 2.158709829628756, "grad_norm": 1.046073317527771, "learning_rate": 1.1368257855770638e-05, "loss": 3.6824, "step": 192150 }, { "epoch": 2.159271554799099, "grad_norm": 1.0207507610321045, "learning_rate": 1.1366010942467787e-05, "loss": 3.6136, "step": 192200 }, { "epoch": 2.159833279969442, "grad_norm": 1.2055473327636719, "learning_rate": 1.1363764029164936e-05, "loss": 3.6755, "step": 192250 }, { "epoch": 2.1603950051397853, "grad_norm": 1.1619362831115723, "learning_rate": 1.1361517115862085e-05, "loss": 3.6353, "step": 192300 }, { "epoch": 2.1609567303101285, "grad_norm": 1.0099173784255981, "learning_rate": 1.1359270202559236e-05, "loss": 3.6397, "step": 192350 }, { "epoch": 2.1615184554804716, "grad_norm": 1.1548542976379395, "learning_rate": 1.1357023289256386e-05, "loss": 3.7073, "step": 192400 }, { "epoch": 2.1620801806508148, "grad_norm": 1.1104629039764404, "learning_rate": 1.1354776375953535e-05, "loss": 3.7453, "step": 192450 }, { "epoch": 2.162641905821158, "grad_norm": 3.932023525238037, "learning_rate": 1.1352529462650686e-05, "loss": 3.6289, "step": 192500 }, { "epoch": 2.163203630991501, "grad_norm": 1.2070088386535645, "learning_rate": 1.1350282549347835e-05, "loss": 3.701, "step": 192550 }, { "epoch": 2.163765356161844, "grad_norm": 1.2627888917922974, "learning_rate": 1.1348035636044985e-05, "loss": 3.6332, "step": 192600 }, { "epoch": 2.1643270813321873, "grad_norm": 1.2985186576843262, "learning_rate": 1.1345788722742134e-05, "loss": 3.5978, "step": 192650 }, { "epoch": 2.1648888065025305, "grad_norm": 1.1906691789627075, "learning_rate": 1.1343541809439285e-05, "loss": 3.6652, "step": 192700 }, { "epoch": 2.1654505316728736, "grad_norm": 1.1376667022705078, "learning_rate": 1.1341294896136434e-05, "loss": 3.6271, "step": 192750 }, { "epoch": 2.1660122568432167, "grad_norm": 1.0888322591781616, "learning_rate": 1.1339047982833584e-05, "loss": 3.5866, "step": 192800 }, { "epoch": 2.16657398201356, "grad_norm": 1.0450643301010132, "learning_rate": 1.1336801069530735e-05, "loss": 3.6099, "step": 192850 }, { "epoch": 2.167135707183903, "grad_norm": 1.1386446952819824, "learning_rate": 1.1334554156227884e-05, "loss": 3.5828, "step": 192900 }, { "epoch": 2.167697432354246, "grad_norm": 1.1101601123809814, "learning_rate": 1.1332307242925033e-05, "loss": 3.6742, "step": 192950 }, { "epoch": 2.1682591575245898, "grad_norm": 1.1992783546447754, "learning_rate": 1.1330060329622183e-05, "loss": 3.6665, "step": 193000 }, { "epoch": 2.1688208826949325, "grad_norm": 1.1518826484680176, "learning_rate": 1.1327813416319334e-05, "loss": 3.6617, "step": 193050 }, { "epoch": 2.169382607865276, "grad_norm": 1.4648700952529907, "learning_rate": 1.1325566503016483e-05, "loss": 3.6676, "step": 193100 }, { "epoch": 2.169944333035619, "grad_norm": 1.2937272787094116, "learning_rate": 1.1323319589713632e-05, "loss": 3.7151, "step": 193150 }, { "epoch": 2.1705060582059623, "grad_norm": 1.221721887588501, "learning_rate": 1.1321072676410783e-05, "loss": 3.6795, "step": 193200 }, { "epoch": 2.1710677833763055, "grad_norm": 1.5011725425720215, "learning_rate": 1.1318825763107933e-05, "loss": 3.6012, "step": 193250 }, { "epoch": 2.1716295085466486, "grad_norm": 1.0371367931365967, "learning_rate": 1.1316578849805082e-05, "loss": 3.635, "step": 193300 }, { "epoch": 2.1721912337169917, "grad_norm": 1.062644124031067, "learning_rate": 1.1314331936502231e-05, "loss": 3.7003, "step": 193350 }, { "epoch": 2.172752958887335, "grad_norm": 1.191779613494873, "learning_rate": 1.1312085023199382e-05, "loss": 3.6577, "step": 193400 }, { "epoch": 2.173314684057678, "grad_norm": 1.106966257095337, "learning_rate": 1.1309838109896532e-05, "loss": 3.702, "step": 193450 }, { "epoch": 2.173876409228021, "grad_norm": 1.474577784538269, "learning_rate": 1.1307591196593681e-05, "loss": 3.7128, "step": 193500 }, { "epoch": 2.1744381343983643, "grad_norm": 0.9508491158485413, "learning_rate": 1.1305344283290832e-05, "loss": 3.6307, "step": 193550 }, { "epoch": 2.1749998595687074, "grad_norm": 1.0904008150100708, "learning_rate": 1.1303097369987981e-05, "loss": 3.6612, "step": 193600 }, { "epoch": 2.1755615847390506, "grad_norm": 1.1697651147842407, "learning_rate": 1.130085045668513e-05, "loss": 3.7044, "step": 193650 }, { "epoch": 2.1761233099093937, "grad_norm": 1.2298531532287598, "learning_rate": 1.129860354338228e-05, "loss": 3.6504, "step": 193700 }, { "epoch": 2.176685035079737, "grad_norm": 1.0923755168914795, "learning_rate": 1.1296356630079431e-05, "loss": 3.6549, "step": 193750 }, { "epoch": 2.17724676025008, "grad_norm": 1.1315332651138306, "learning_rate": 1.1294154655042635e-05, "loss": 3.7113, "step": 193800 }, { "epoch": 2.177808485420423, "grad_norm": 0.9924259781837463, "learning_rate": 1.1291907741739784e-05, "loss": 3.6409, "step": 193850 }, { "epoch": 2.1783702105907663, "grad_norm": 1.2110371589660645, "learning_rate": 1.1289660828436935e-05, "loss": 3.6688, "step": 193900 }, { "epoch": 2.1789319357611094, "grad_norm": 1.2357207536697388, "learning_rate": 1.1287413915134085e-05, "loss": 3.6865, "step": 193950 }, { "epoch": 2.1794936609314526, "grad_norm": 1.288679838180542, "learning_rate": 1.1285167001831234e-05, "loss": 3.6362, "step": 194000 }, { "epoch": 2.1800553861017957, "grad_norm": 1.194150686264038, "learning_rate": 1.1282920088528383e-05, "loss": 3.7321, "step": 194050 }, { "epoch": 2.180617111272139, "grad_norm": 1.1033546924591064, "learning_rate": 1.1280673175225534e-05, "loss": 3.6577, "step": 194100 }, { "epoch": 2.181178836442482, "grad_norm": 1.285281777381897, "learning_rate": 1.1278426261922684e-05, "loss": 3.525, "step": 194150 }, { "epoch": 2.181740561612825, "grad_norm": 1.0182749032974243, "learning_rate": 1.1276179348619833e-05, "loss": 3.7052, "step": 194200 }, { "epoch": 2.1823022867831683, "grad_norm": 1.1671384572982788, "learning_rate": 1.1273932435316984e-05, "loss": 3.6362, "step": 194250 }, { "epoch": 2.1828640119535114, "grad_norm": 1.1830589771270752, "learning_rate": 1.1271685522014133e-05, "loss": 3.6723, "step": 194300 }, { "epoch": 2.183425737123855, "grad_norm": 1.2304086685180664, "learning_rate": 1.1269438608711283e-05, "loss": 3.6114, "step": 194350 }, { "epoch": 2.1839874622941977, "grad_norm": 3.535710573196411, "learning_rate": 1.1267191695408432e-05, "loss": 3.6511, "step": 194400 }, { "epoch": 2.1845491874645413, "grad_norm": 1.1054655313491821, "learning_rate": 1.1264944782105583e-05, "loss": 3.5475, "step": 194450 }, { "epoch": 2.1851109126348844, "grad_norm": 1.1622523069381714, "learning_rate": 1.1262697868802732e-05, "loss": 3.6228, "step": 194500 }, { "epoch": 2.1856726378052276, "grad_norm": 1.2083780765533447, "learning_rate": 1.1260450955499882e-05, "loss": 3.6528, "step": 194550 }, { "epoch": 2.1862343629755707, "grad_norm": 1.095506191253662, "learning_rate": 1.1258204042197031e-05, "loss": 3.7116, "step": 194600 }, { "epoch": 2.186796088145914, "grad_norm": 1.1502091884613037, "learning_rate": 1.1255957128894182e-05, "loss": 3.6603, "step": 194650 }, { "epoch": 2.187357813316257, "grad_norm": 1.1696006059646606, "learning_rate": 1.1253710215591331e-05, "loss": 3.6242, "step": 194700 }, { "epoch": 2.1879195384866, "grad_norm": 1.3300443887710571, "learning_rate": 1.125146330228848e-05, "loss": 3.625, "step": 194750 }, { "epoch": 2.1884812636569433, "grad_norm": 1.1268310546875, "learning_rate": 1.1249216388985632e-05, "loss": 3.6589, "step": 194800 }, { "epoch": 2.1890429888272864, "grad_norm": 1.2160589694976807, "learning_rate": 1.1246969475682781e-05, "loss": 3.5945, "step": 194850 }, { "epoch": 2.1896047139976296, "grad_norm": 1.0838061571121216, "learning_rate": 1.124472256237993e-05, "loss": 3.6457, "step": 194900 }, { "epoch": 2.1901664391679727, "grad_norm": 1.2142306566238403, "learning_rate": 1.124247564907708e-05, "loss": 3.6174, "step": 194950 }, { "epoch": 2.190728164338316, "grad_norm": 1.091429591178894, "learning_rate": 1.124022873577423e-05, "loss": 3.6036, "step": 195000 }, { "epoch": 2.191289889508659, "grad_norm": 1.0193670988082886, "learning_rate": 1.123798182247138e-05, "loss": 3.6599, "step": 195050 }, { "epoch": 2.191851614679002, "grad_norm": 1.2449196577072144, "learning_rate": 1.123573490916853e-05, "loss": 3.7411, "step": 195100 }, { "epoch": 2.1924133398493453, "grad_norm": 1.2533265352249146, "learning_rate": 1.123348799586568e-05, "loss": 3.6701, "step": 195150 }, { "epoch": 2.1929750650196884, "grad_norm": 1.0579513311386108, "learning_rate": 1.1231286020828886e-05, "loss": 3.5372, "step": 195200 }, { "epoch": 2.1935367901900316, "grad_norm": 1.1262048482894897, "learning_rate": 1.1229039107526035e-05, "loss": 3.6519, "step": 195250 }, { "epoch": 2.1940985153603747, "grad_norm": 1.3239110708236694, "learning_rate": 1.1226792194223186e-05, "loss": 3.674, "step": 195300 }, { "epoch": 2.194660240530718, "grad_norm": 1.1064801216125488, "learning_rate": 1.1224545280920336e-05, "loss": 3.5913, "step": 195350 }, { "epoch": 2.195221965701061, "grad_norm": 1.1677799224853516, "learning_rate": 1.1222298367617485e-05, "loss": 3.6571, "step": 195400 }, { "epoch": 2.195783690871404, "grad_norm": 1.2749618291854858, "learning_rate": 1.1220051454314636e-05, "loss": 3.6984, "step": 195450 }, { "epoch": 2.1963454160417473, "grad_norm": 1.1262245178222656, "learning_rate": 1.1217804541011785e-05, "loss": 3.6573, "step": 195500 }, { "epoch": 2.1969071412120904, "grad_norm": 1.2008533477783203, "learning_rate": 1.1215557627708935e-05, "loss": 3.6885, "step": 195550 }, { "epoch": 2.1974688663824335, "grad_norm": 1.3368501663208008, "learning_rate": 1.1213310714406084e-05, "loss": 3.56, "step": 195600 }, { "epoch": 2.1980305915527767, "grad_norm": 1.2217975854873657, "learning_rate": 1.1211063801103235e-05, "loss": 3.6528, "step": 195650 }, { "epoch": 2.1985923167231203, "grad_norm": 1.2406367063522339, "learning_rate": 1.1208816887800384e-05, "loss": 3.67, "step": 195700 }, { "epoch": 2.199154041893463, "grad_norm": 1.1667741537094116, "learning_rate": 1.1206569974497534e-05, "loss": 3.6043, "step": 195750 }, { "epoch": 2.1997157670638066, "grad_norm": 1.0640195608139038, "learning_rate": 1.1204323061194685e-05, "loss": 3.6652, "step": 195800 }, { "epoch": 2.2002774922341497, "grad_norm": 1.1555191278457642, "learning_rate": 1.1202076147891834e-05, "loss": 3.6219, "step": 195850 }, { "epoch": 2.200839217404493, "grad_norm": 1.0477038621902466, "learning_rate": 1.1199829234588983e-05, "loss": 3.5922, "step": 195900 }, { "epoch": 2.201400942574836, "grad_norm": 1.1948083639144897, "learning_rate": 1.1197582321286133e-05, "loss": 3.6632, "step": 195950 }, { "epoch": 2.201962667745179, "grad_norm": 1.0949130058288574, "learning_rate": 1.1195335407983284e-05, "loss": 3.5816, "step": 196000 }, { "epoch": 2.2025243929155223, "grad_norm": 1.1196852922439575, "learning_rate": 1.1193088494680433e-05, "loss": 3.6256, "step": 196050 }, { "epoch": 2.2030861180858654, "grad_norm": 1.0841516256332397, "learning_rate": 1.1190841581377582e-05, "loss": 3.6894, "step": 196100 }, { "epoch": 2.2036478432562085, "grad_norm": 1.2244213819503784, "learning_rate": 1.1188594668074733e-05, "loss": 3.6416, "step": 196150 }, { "epoch": 2.2042095684265517, "grad_norm": 1.1902669668197632, "learning_rate": 1.1186347754771883e-05, "loss": 3.6649, "step": 196200 }, { "epoch": 2.204771293596895, "grad_norm": 1.4555490016937256, "learning_rate": 1.1184100841469032e-05, "loss": 3.7151, "step": 196250 }, { "epoch": 2.205333018767238, "grad_norm": 1.236116647720337, "learning_rate": 1.1181853928166181e-05, "loss": 3.5958, "step": 196300 }, { "epoch": 2.205894743937581, "grad_norm": 1.2802221775054932, "learning_rate": 1.1179607014863332e-05, "loss": 3.5966, "step": 196350 }, { "epoch": 2.2064564691079243, "grad_norm": 1.3653600215911865, "learning_rate": 1.1177360101560482e-05, "loss": 3.6607, "step": 196400 }, { "epoch": 2.2070181942782674, "grad_norm": 0.9888893365859985, "learning_rate": 1.1175113188257631e-05, "loss": 3.6654, "step": 196450 }, { "epoch": 2.2075799194486105, "grad_norm": 1.0868043899536133, "learning_rate": 1.1172866274954782e-05, "loss": 3.6432, "step": 196500 }, { "epoch": 2.2081416446189537, "grad_norm": 1.1037492752075195, "learning_rate": 1.1170619361651931e-05, "loss": 3.6218, "step": 196550 }, { "epoch": 2.208703369789297, "grad_norm": 1.244435429573059, "learning_rate": 1.116837244834908e-05, "loss": 3.6909, "step": 196600 }, { "epoch": 2.20926509495964, "grad_norm": 1.148008108139038, "learning_rate": 1.116612553504623e-05, "loss": 3.646, "step": 196650 }, { "epoch": 2.209826820129983, "grad_norm": 1.1879488229751587, "learning_rate": 1.1163878621743381e-05, "loss": 3.7235, "step": 196700 }, { "epoch": 2.2103885453003262, "grad_norm": 1.0928137302398682, "learning_rate": 1.116163170844053e-05, "loss": 3.6463, "step": 196750 }, { "epoch": 2.2109502704706694, "grad_norm": 1.0006128549575806, "learning_rate": 1.115938479513768e-05, "loss": 3.7025, "step": 196800 }, { "epoch": 2.2115119956410125, "grad_norm": 1.2530834674835205, "learning_rate": 1.115713788183483e-05, "loss": 3.6741, "step": 196850 }, { "epoch": 2.2120737208113557, "grad_norm": 1.4288831949234009, "learning_rate": 1.115489096853198e-05, "loss": 3.6525, "step": 196900 }, { "epoch": 2.212635445981699, "grad_norm": 1.1534000635147095, "learning_rate": 1.115264405522913e-05, "loss": 3.6866, "step": 196950 }, { "epoch": 2.213197171152042, "grad_norm": 1.12404465675354, "learning_rate": 1.1150397141926278e-05, "loss": 3.7836, "step": 197000 }, { "epoch": 2.2137588963223855, "grad_norm": 1.221832275390625, "learning_rate": 1.114815022862343e-05, "loss": 3.7044, "step": 197050 }, { "epoch": 2.2143206214927282, "grad_norm": 1.2092190980911255, "learning_rate": 1.1145903315320579e-05, "loss": 3.6073, "step": 197100 }, { "epoch": 2.214882346663072, "grad_norm": 1.0969897508621216, "learning_rate": 1.1143656402017728e-05, "loss": 3.6569, "step": 197150 }, { "epoch": 2.215444071833415, "grad_norm": 1.1661725044250488, "learning_rate": 1.1141409488714877e-05, "loss": 3.6218, "step": 197200 }, { "epoch": 2.216005797003758, "grad_norm": 3.4693822860717773, "learning_rate": 1.1139162575412028e-05, "loss": 3.6296, "step": 197250 }, { "epoch": 2.2165675221741012, "grad_norm": 1.1358059644699097, "learning_rate": 1.1136915662109178e-05, "loss": 3.5186, "step": 197300 }, { "epoch": 2.2171292473444444, "grad_norm": 1.1608294248580933, "learning_rate": 1.1134668748806327e-05, "loss": 3.6262, "step": 197350 }, { "epoch": 2.2176909725147875, "grad_norm": 1.3317348957061768, "learning_rate": 1.1132421835503478e-05, "loss": 3.602, "step": 197400 }, { "epoch": 2.2182526976851307, "grad_norm": 1.1245685815811157, "learning_rate": 1.1130174922200627e-05, "loss": 3.7152, "step": 197450 }, { "epoch": 2.218814422855474, "grad_norm": 1.0280276536941528, "learning_rate": 1.1127928008897777e-05, "loss": 3.5854, "step": 197500 }, { "epoch": 2.219376148025817, "grad_norm": 1.0780091285705566, "learning_rate": 1.1125681095594926e-05, "loss": 3.6664, "step": 197550 }, { "epoch": 2.21993787319616, "grad_norm": 1.0677632093429565, "learning_rate": 1.1123434182292077e-05, "loss": 3.683, "step": 197600 }, { "epoch": 2.2204995983665032, "grad_norm": 1.2442424297332764, "learning_rate": 1.1121187268989226e-05, "loss": 3.6738, "step": 197650 }, { "epoch": 2.2210613235368464, "grad_norm": 1.1354444026947021, "learning_rate": 1.1118940355686376e-05, "loss": 3.6153, "step": 197700 }, { "epoch": 2.2216230487071895, "grad_norm": 1.225928544998169, "learning_rate": 1.1116693442383527e-05, "loss": 3.6901, "step": 197750 }, { "epoch": 2.2221847738775327, "grad_norm": 1.3080968856811523, "learning_rate": 1.1114446529080676e-05, "loss": 3.6879, "step": 197800 }, { "epoch": 2.222746499047876, "grad_norm": 1.0205680131912231, "learning_rate": 1.1112199615777825e-05, "loss": 3.6512, "step": 197850 }, { "epoch": 2.223308224218219, "grad_norm": 1.0382564067840576, "learning_rate": 1.1109952702474975e-05, "loss": 3.7031, "step": 197900 }, { "epoch": 2.223869949388562, "grad_norm": 1.2341996431350708, "learning_rate": 1.1107705789172126e-05, "loss": 3.7014, "step": 197950 }, { "epoch": 2.2244316745589052, "grad_norm": 1.1369259357452393, "learning_rate": 1.1105458875869275e-05, "loss": 3.7022, "step": 198000 }, { "epoch": 2.2249933997292484, "grad_norm": 1.226236343383789, "learning_rate": 1.1103211962566424e-05, "loss": 3.5697, "step": 198050 }, { "epoch": 2.2255551248995915, "grad_norm": 1.083716630935669, "learning_rate": 1.1100965049263575e-05, "loss": 3.5827, "step": 198100 }, { "epoch": 2.2261168500699346, "grad_norm": 1.1048554182052612, "learning_rate": 1.1098718135960725e-05, "loss": 3.5583, "step": 198150 }, { "epoch": 2.226678575240278, "grad_norm": 1.3492372035980225, "learning_rate": 1.1096471222657874e-05, "loss": 3.6775, "step": 198200 }, { "epoch": 2.227240300410621, "grad_norm": 1.2858997583389282, "learning_rate": 1.1094224309355023e-05, "loss": 3.6434, "step": 198250 }, { "epoch": 2.227802025580964, "grad_norm": 1.2529743909835815, "learning_rate": 1.1091977396052174e-05, "loss": 3.5534, "step": 198300 }, { "epoch": 2.228363750751307, "grad_norm": 1.1517821550369263, "learning_rate": 1.1089730482749324e-05, "loss": 3.6645, "step": 198350 }, { "epoch": 2.228925475921651, "grad_norm": 1.4201101064682007, "learning_rate": 1.1087483569446473e-05, "loss": 3.7434, "step": 198400 }, { "epoch": 2.2294872010919935, "grad_norm": 1.2251349687576294, "learning_rate": 1.1085236656143624e-05, "loss": 3.6577, "step": 198450 }, { "epoch": 2.230048926262337, "grad_norm": 1.2064093351364136, "learning_rate": 1.1082989742840773e-05, "loss": 3.57, "step": 198500 }, { "epoch": 2.2306106514326802, "grad_norm": 1.2292957305908203, "learning_rate": 1.1080742829537923e-05, "loss": 3.6438, "step": 198550 }, { "epoch": 2.2311723766030234, "grad_norm": 1.165260910987854, "learning_rate": 1.1078495916235072e-05, "loss": 3.6598, "step": 198600 }, { "epoch": 2.2317341017733665, "grad_norm": 1.052223563194275, "learning_rate": 1.1076249002932223e-05, "loss": 3.7084, "step": 198650 }, { "epoch": 2.2322958269437096, "grad_norm": 1.179465413093567, "learning_rate": 1.1074002089629372e-05, "loss": 3.7048, "step": 198700 }, { "epoch": 2.232857552114053, "grad_norm": 1.2978014945983887, "learning_rate": 1.1071755176326522e-05, "loss": 3.6179, "step": 198750 }, { "epoch": 2.233419277284396, "grad_norm": 1.4667948484420776, "learning_rate": 1.106950826302367e-05, "loss": 3.6239, "step": 198800 }, { "epoch": 2.233981002454739, "grad_norm": 1.246903657913208, "learning_rate": 1.1067261349720822e-05, "loss": 3.6396, "step": 198850 }, { "epoch": 2.234542727625082, "grad_norm": 0.9928773045539856, "learning_rate": 1.1065014436417971e-05, "loss": 3.6633, "step": 198900 }, { "epoch": 2.2351044527954254, "grad_norm": 1.1761093139648438, "learning_rate": 1.106276752311512e-05, "loss": 3.6865, "step": 198950 }, { "epoch": 2.2356661779657685, "grad_norm": 1.2706598043441772, "learning_rate": 1.1060520609812271e-05, "loss": 3.654, "step": 199000 }, { "epoch": 2.2362279031361116, "grad_norm": 1.2302583456039429, "learning_rate": 1.105827369650942e-05, "loss": 3.6131, "step": 199050 }, { "epoch": 2.236789628306455, "grad_norm": 1.1114083528518677, "learning_rate": 1.105602678320657e-05, "loss": 3.6453, "step": 199100 }, { "epoch": 2.237351353476798, "grad_norm": 1.1350913047790527, "learning_rate": 1.105377986990372e-05, "loss": 3.676, "step": 199150 }, { "epoch": 2.237913078647141, "grad_norm": 1.2548173666000366, "learning_rate": 1.105153295660087e-05, "loss": 3.6188, "step": 199200 }, { "epoch": 2.238474803817484, "grad_norm": 1.1997712850570679, "learning_rate": 1.104928604329802e-05, "loss": 3.6246, "step": 199250 }, { "epoch": 2.2390365289878273, "grad_norm": 1.1969828605651855, "learning_rate": 1.1047039129995169e-05, "loss": 3.704, "step": 199300 }, { "epoch": 2.2395982541581705, "grad_norm": 1.2156494855880737, "learning_rate": 1.104479221669232e-05, "loss": 3.6953, "step": 199350 }, { "epoch": 2.2401599793285136, "grad_norm": 1.2109017372131348, "learning_rate": 1.104254530338947e-05, "loss": 3.7038, "step": 199400 }, { "epoch": 2.2407217044988568, "grad_norm": 1.0635199546813965, "learning_rate": 1.1040298390086619e-05, "loss": 3.6131, "step": 199450 }, { "epoch": 2.2412834296692, "grad_norm": 1.0498710870742798, "learning_rate": 1.1038051476783768e-05, "loss": 3.6186, "step": 199500 }, { "epoch": 2.241845154839543, "grad_norm": 1.3620375394821167, "learning_rate": 1.1035804563480919e-05, "loss": 3.6428, "step": 199550 }, { "epoch": 2.242406880009886, "grad_norm": 1.0943694114685059, "learning_rate": 1.1033557650178068e-05, "loss": 3.6925, "step": 199600 }, { "epoch": 2.2429686051802293, "grad_norm": 1.103467345237732, "learning_rate": 1.1031310736875218e-05, "loss": 3.6864, "step": 199650 }, { "epoch": 2.2435303303505725, "grad_norm": 1.0348925590515137, "learning_rate": 1.1029063823572369e-05, "loss": 3.6311, "step": 199700 }, { "epoch": 2.244092055520916, "grad_norm": 1.1905165910720825, "learning_rate": 1.1026816910269518e-05, "loss": 3.659, "step": 199750 }, { "epoch": 2.2446537806912588, "grad_norm": 1.3361129760742188, "learning_rate": 1.1024569996966667e-05, "loss": 3.6541, "step": 199800 }, { "epoch": 2.2452155058616023, "grad_norm": 1.1166316270828247, "learning_rate": 1.1022323083663817e-05, "loss": 3.5581, "step": 199850 }, { "epoch": 2.2457772310319455, "grad_norm": 1.2856205701828003, "learning_rate": 1.1020076170360968e-05, "loss": 3.6256, "step": 199900 }, { "epoch": 2.2463389562022886, "grad_norm": 1.3031456470489502, "learning_rate": 1.1017829257058117e-05, "loss": 3.6146, "step": 199950 }, { "epoch": 2.2469006813726318, "grad_norm": 1.0035897493362427, "learning_rate": 1.1015582343755266e-05, "loss": 3.7184, "step": 200000 }, { "epoch": 2.247462406542975, "grad_norm": 1.3020460605621338, "learning_rate": 1.1013335430452417e-05, "loss": 3.6684, "step": 200050 }, { "epoch": 2.248024131713318, "grad_norm": 1.1402523517608643, "learning_rate": 1.1011088517149567e-05, "loss": 3.6102, "step": 200100 }, { "epoch": 2.248585856883661, "grad_norm": 1.2541253566741943, "learning_rate": 1.1008841603846716e-05, "loss": 3.6028, "step": 200150 }, { "epoch": 2.2491475820540043, "grad_norm": 1.1429928541183472, "learning_rate": 1.1006594690543865e-05, "loss": 3.664, "step": 200200 }, { "epoch": 2.2497093072243475, "grad_norm": 1.4887467622756958, "learning_rate": 1.1004392715507073e-05, "loss": 3.6599, "step": 200250 }, { "epoch": 2.2502710323946906, "grad_norm": 1.1874189376831055, "learning_rate": 1.1002145802204222e-05, "loss": 3.7105, "step": 200300 }, { "epoch": 2.2508327575650338, "grad_norm": 1.4966840744018555, "learning_rate": 1.0999898888901373e-05, "loss": 3.6475, "step": 200350 }, { "epoch": 2.251394482735377, "grad_norm": 1.068434476852417, "learning_rate": 1.0997651975598523e-05, "loss": 3.6619, "step": 200400 }, { "epoch": 2.25195620790572, "grad_norm": 1.2671818733215332, "learning_rate": 1.0995405062295672e-05, "loss": 3.6184, "step": 200450 }, { "epoch": 2.252517933076063, "grad_norm": 1.0523030757904053, "learning_rate": 1.0993158148992821e-05, "loss": 3.6023, "step": 200500 }, { "epoch": 2.2530796582464063, "grad_norm": 1.0349535942077637, "learning_rate": 1.0990911235689972e-05, "loss": 3.6532, "step": 200550 }, { "epoch": 2.2536413834167495, "grad_norm": 0.9204732179641724, "learning_rate": 1.0988664322387121e-05, "loss": 3.6567, "step": 200600 }, { "epoch": 2.2542031085870926, "grad_norm": 1.0933384895324707, "learning_rate": 1.098641740908427e-05, "loss": 3.6977, "step": 200650 }, { "epoch": 2.2547648337574357, "grad_norm": 1.305959939956665, "learning_rate": 1.0984170495781422e-05, "loss": 3.6408, "step": 200700 }, { "epoch": 2.255326558927779, "grad_norm": 1.1364576816558838, "learning_rate": 1.0981923582478571e-05, "loss": 3.5922, "step": 200750 }, { "epoch": 2.255888284098122, "grad_norm": 1.2036141157150269, "learning_rate": 1.097967666917572e-05, "loss": 3.6742, "step": 200800 }, { "epoch": 2.256450009268465, "grad_norm": 1.1628385782241821, "learning_rate": 1.097742975587287e-05, "loss": 3.7339, "step": 200850 }, { "epoch": 2.2570117344388083, "grad_norm": 1.0342094898223877, "learning_rate": 1.097518284257002e-05, "loss": 3.6503, "step": 200900 }, { "epoch": 2.2575734596091515, "grad_norm": 1.196022629737854, "learning_rate": 1.097293592926717e-05, "loss": 3.6804, "step": 200950 }, { "epoch": 2.2581351847794946, "grad_norm": 1.245009422302246, "learning_rate": 1.097068901596432e-05, "loss": 3.5853, "step": 201000 }, { "epoch": 2.2586969099498377, "grad_norm": 1.203466534614563, "learning_rate": 1.096844210266147e-05, "loss": 3.6283, "step": 201050 }, { "epoch": 2.2592586351201813, "grad_norm": 1.1776548624038696, "learning_rate": 1.096619518935862e-05, "loss": 3.637, "step": 201100 }, { "epoch": 2.259820360290524, "grad_norm": 1.0840448141098022, "learning_rate": 1.0963948276055769e-05, "loss": 3.595, "step": 201150 }, { "epoch": 2.2603820854608676, "grad_norm": 1.3171910047531128, "learning_rate": 1.0961701362752918e-05, "loss": 3.6929, "step": 201200 }, { "epoch": 2.2609438106312107, "grad_norm": 0.9727421402931213, "learning_rate": 1.095945444945007e-05, "loss": 3.5353, "step": 201250 }, { "epoch": 2.261505535801554, "grad_norm": 1.2239476442337036, "learning_rate": 1.0957207536147219e-05, "loss": 3.6409, "step": 201300 }, { "epoch": 2.262067260971897, "grad_norm": 1.076357364654541, "learning_rate": 1.0954960622844368e-05, "loss": 3.6531, "step": 201350 }, { "epoch": 2.26262898614224, "grad_norm": 1.1421380043029785, "learning_rate": 1.0952713709541517e-05, "loss": 3.6225, "step": 201400 }, { "epoch": 2.2631907113125833, "grad_norm": 1.158110499382019, "learning_rate": 1.0950466796238668e-05, "loss": 3.6719, "step": 201450 }, { "epoch": 2.2637524364829265, "grad_norm": 1.3112207651138306, "learning_rate": 1.0948219882935818e-05, "loss": 3.6902, "step": 201500 }, { "epoch": 2.2643141616532696, "grad_norm": 1.178525686264038, "learning_rate": 1.0945972969632967e-05, "loss": 3.6311, "step": 201550 }, { "epoch": 2.2648758868236127, "grad_norm": 1.0799179077148438, "learning_rate": 1.0943726056330118e-05, "loss": 3.6396, "step": 201600 }, { "epoch": 2.265437611993956, "grad_norm": 1.1785821914672852, "learning_rate": 1.0941479143027267e-05, "loss": 3.6488, "step": 201650 }, { "epoch": 2.265999337164299, "grad_norm": 1.0482996702194214, "learning_rate": 1.0939232229724417e-05, "loss": 3.6388, "step": 201700 }, { "epoch": 2.266561062334642, "grad_norm": 1.2701849937438965, "learning_rate": 1.0936985316421566e-05, "loss": 3.6097, "step": 201750 }, { "epoch": 2.2671227875049853, "grad_norm": 1.0427109003067017, "learning_rate": 1.0934738403118717e-05, "loss": 3.5898, "step": 201800 }, { "epoch": 2.2676845126753284, "grad_norm": 1.1787889003753662, "learning_rate": 1.0932491489815866e-05, "loss": 3.5461, "step": 201850 }, { "epoch": 2.2682462378456716, "grad_norm": 1.0887303352355957, "learning_rate": 1.0930244576513016e-05, "loss": 3.6988, "step": 201900 }, { "epoch": 2.2688079630160147, "grad_norm": 1.3267873525619507, "learning_rate": 1.0927997663210167e-05, "loss": 3.6635, "step": 201950 }, { "epoch": 2.269369688186358, "grad_norm": 1.1822171211242676, "learning_rate": 1.0925750749907316e-05, "loss": 3.596, "step": 202000 }, { "epoch": 2.269931413356701, "grad_norm": 1.2599691152572632, "learning_rate": 1.0923503836604465e-05, "loss": 3.6804, "step": 202050 }, { "epoch": 2.270493138527044, "grad_norm": 1.1934468746185303, "learning_rate": 1.0921256923301615e-05, "loss": 3.66, "step": 202100 }, { "epoch": 2.2710548636973873, "grad_norm": 1.2453982830047607, "learning_rate": 1.0919010009998766e-05, "loss": 3.6684, "step": 202150 }, { "epoch": 2.2716165888677304, "grad_norm": 0.9536491632461548, "learning_rate": 1.0916763096695915e-05, "loss": 3.6246, "step": 202200 }, { "epoch": 2.2721783140380736, "grad_norm": 1.0329339504241943, "learning_rate": 1.0914516183393064e-05, "loss": 3.7197, "step": 202250 }, { "epoch": 2.2727400392084167, "grad_norm": 1.1337761878967285, "learning_rate": 1.0912269270090215e-05, "loss": 3.6023, "step": 202300 }, { "epoch": 2.27330176437876, "grad_norm": 0.9923123717308044, "learning_rate": 1.0910022356787364e-05, "loss": 3.6429, "step": 202350 }, { "epoch": 2.273863489549103, "grad_norm": 1.7720701694488525, "learning_rate": 1.0907775443484514e-05, "loss": 3.6781, "step": 202400 }, { "epoch": 2.2744252147194466, "grad_norm": 1.190482258796692, "learning_rate": 1.0905573468447721e-05, "loss": 3.6465, "step": 202450 }, { "epoch": 2.2749869398897893, "grad_norm": 1.2313224077224731, "learning_rate": 1.090332655514487e-05, "loss": 3.6782, "step": 202500 }, { "epoch": 2.275548665060133, "grad_norm": 1.0463004112243652, "learning_rate": 1.090107964184202e-05, "loss": 3.7569, "step": 202550 }, { "epoch": 2.276110390230476, "grad_norm": 1.1098483800888062, "learning_rate": 1.0898832728539171e-05, "loss": 3.5805, "step": 202600 }, { "epoch": 2.276672115400819, "grad_norm": 1.1954679489135742, "learning_rate": 1.089658581523632e-05, "loss": 3.6629, "step": 202650 }, { "epoch": 2.2772338405711623, "grad_norm": 1.240631341934204, "learning_rate": 1.089433890193347e-05, "loss": 3.6044, "step": 202700 }, { "epoch": 2.2777955657415054, "grad_norm": 1.313819169998169, "learning_rate": 1.0892091988630619e-05, "loss": 3.6146, "step": 202750 }, { "epoch": 2.2783572909118486, "grad_norm": 1.261030912399292, "learning_rate": 1.088984507532777e-05, "loss": 3.6696, "step": 202800 }, { "epoch": 2.2789190160821917, "grad_norm": 1.1849219799041748, "learning_rate": 1.088759816202492e-05, "loss": 3.7237, "step": 202850 }, { "epoch": 2.279480741252535, "grad_norm": 1.3644628524780273, "learning_rate": 1.0885351248722069e-05, "loss": 3.6499, "step": 202900 }, { "epoch": 2.280042466422878, "grad_norm": 1.3659747838974, "learning_rate": 1.088310433541922e-05, "loss": 3.66, "step": 202950 }, { "epoch": 2.280604191593221, "grad_norm": 1.1541392803192139, "learning_rate": 1.0880857422116369e-05, "loss": 3.7414, "step": 203000 }, { "epoch": 2.2811659167635643, "grad_norm": 1.039172649383545, "learning_rate": 1.0878610508813518e-05, "loss": 3.6835, "step": 203050 }, { "epoch": 2.2817276419339074, "grad_norm": 1.0046977996826172, "learning_rate": 1.0876363595510668e-05, "loss": 3.6234, "step": 203100 }, { "epoch": 2.2822893671042506, "grad_norm": 1.396032452583313, "learning_rate": 1.0874116682207819e-05, "loss": 3.6203, "step": 203150 }, { "epoch": 2.2828510922745937, "grad_norm": 1.0471729040145874, "learning_rate": 1.0871869768904968e-05, "loss": 3.6176, "step": 203200 }, { "epoch": 2.283412817444937, "grad_norm": 1.3599485158920288, "learning_rate": 1.0869622855602117e-05, "loss": 3.6163, "step": 203250 }, { "epoch": 2.28397454261528, "grad_norm": 1.3429630994796753, "learning_rate": 1.0867375942299268e-05, "loss": 3.6847, "step": 203300 }, { "epoch": 2.284536267785623, "grad_norm": 1.2321631908416748, "learning_rate": 1.0865129028996418e-05, "loss": 3.6505, "step": 203350 }, { "epoch": 2.2850979929559663, "grad_norm": 1.2308857440948486, "learning_rate": 1.0862882115693567e-05, "loss": 3.6571, "step": 203400 }, { "epoch": 2.2856597181263094, "grad_norm": 1.1756349802017212, "learning_rate": 1.0860635202390716e-05, "loss": 3.6543, "step": 203450 }, { "epoch": 2.2862214432966526, "grad_norm": 1.045255422592163, "learning_rate": 1.0858388289087867e-05, "loss": 3.6134, "step": 203500 }, { "epoch": 2.2867831684669957, "grad_norm": 1.206047534942627, "learning_rate": 1.0856141375785017e-05, "loss": 3.591, "step": 203550 }, { "epoch": 2.287344893637339, "grad_norm": 1.1834707260131836, "learning_rate": 1.0853894462482166e-05, "loss": 3.6628, "step": 203600 }, { "epoch": 2.287906618807682, "grad_norm": 0.9967352151870728, "learning_rate": 1.0851647549179317e-05, "loss": 3.6352, "step": 203650 }, { "epoch": 2.288468343978025, "grad_norm": 1.293893814086914, "learning_rate": 1.0849400635876466e-05, "loss": 3.6823, "step": 203700 }, { "epoch": 2.2890300691483683, "grad_norm": 1.2545384168624878, "learning_rate": 1.0847153722573616e-05, "loss": 3.6486, "step": 203750 }, { "epoch": 2.289591794318712, "grad_norm": 1.224974513053894, "learning_rate": 1.0844906809270765e-05, "loss": 3.7078, "step": 203800 }, { "epoch": 2.2901535194890545, "grad_norm": 1.20431649684906, "learning_rate": 1.0842659895967916e-05, "loss": 3.702, "step": 203850 }, { "epoch": 2.290715244659398, "grad_norm": 1.206586241722107, "learning_rate": 1.0840412982665065e-05, "loss": 3.6062, "step": 203900 }, { "epoch": 2.2912769698297413, "grad_norm": 1.0869084596633911, "learning_rate": 1.0838166069362214e-05, "loss": 3.6111, "step": 203950 }, { "epoch": 2.2918386950000844, "grad_norm": 1.0805386304855347, "learning_rate": 1.0835919156059364e-05, "loss": 3.645, "step": 204000 }, { "epoch": 2.2924004201704276, "grad_norm": 1.3582491874694824, "learning_rate": 1.0833672242756515e-05, "loss": 3.7398, "step": 204050 }, { "epoch": 2.2929621453407707, "grad_norm": 1.1402126550674438, "learning_rate": 1.0831425329453664e-05, "loss": 3.5855, "step": 204100 }, { "epoch": 2.293523870511114, "grad_norm": 1.2263522148132324, "learning_rate": 1.0829178416150813e-05, "loss": 3.6099, "step": 204150 }, { "epoch": 2.294085595681457, "grad_norm": 1.1827775239944458, "learning_rate": 1.0826931502847964e-05, "loss": 3.6069, "step": 204200 }, { "epoch": 2.2946473208518, "grad_norm": 1.1820076704025269, "learning_rate": 1.0824684589545114e-05, "loss": 3.5847, "step": 204250 }, { "epoch": 2.2952090460221433, "grad_norm": 1.3091599941253662, "learning_rate": 1.0822437676242263e-05, "loss": 3.6601, "step": 204300 }, { "epoch": 2.2957707711924864, "grad_norm": 1.1638920307159424, "learning_rate": 1.0820190762939412e-05, "loss": 3.6105, "step": 204350 }, { "epoch": 2.2963324963628295, "grad_norm": 1.0978766679763794, "learning_rate": 1.0817943849636563e-05, "loss": 3.6773, "step": 204400 }, { "epoch": 2.2968942215331727, "grad_norm": 1.1125136613845825, "learning_rate": 1.0815696936333713e-05, "loss": 3.6585, "step": 204450 }, { "epoch": 2.297455946703516, "grad_norm": 1.2809821367263794, "learning_rate": 1.0813450023030862e-05, "loss": 3.6271, "step": 204500 }, { "epoch": 2.298017671873859, "grad_norm": 1.2620893716812134, "learning_rate": 1.0811203109728013e-05, "loss": 3.648, "step": 204550 }, { "epoch": 2.298579397044202, "grad_norm": 1.2011334896087646, "learning_rate": 1.0808956196425162e-05, "loss": 3.6975, "step": 204600 }, { "epoch": 2.2991411222145453, "grad_norm": 1.1442394256591797, "learning_rate": 1.0806709283122312e-05, "loss": 3.5874, "step": 204650 }, { "epoch": 2.2997028473848884, "grad_norm": 1.2460062503814697, "learning_rate": 1.0804462369819461e-05, "loss": 3.6958, "step": 204700 }, { "epoch": 2.3002645725552315, "grad_norm": 1.1312328577041626, "learning_rate": 1.0802215456516612e-05, "loss": 3.6445, "step": 204750 }, { "epoch": 2.3008262977255747, "grad_norm": 1.2142772674560547, "learning_rate": 1.0799968543213761e-05, "loss": 3.6515, "step": 204800 }, { "epoch": 2.301388022895918, "grad_norm": 1.2299082279205322, "learning_rate": 1.079772162991091e-05, "loss": 3.6667, "step": 204850 }, { "epoch": 2.301949748066261, "grad_norm": 1.1363869905471802, "learning_rate": 1.0795474716608062e-05, "loss": 3.692, "step": 204900 }, { "epoch": 2.302511473236604, "grad_norm": 1.0149507522583008, "learning_rate": 1.0793227803305211e-05, "loss": 3.6699, "step": 204950 }, { "epoch": 2.3030731984069472, "grad_norm": 1.3276242017745972, "learning_rate": 1.079098089000236e-05, "loss": 3.6628, "step": 205000 }, { "epoch": 2.3036349235772904, "grad_norm": 1.2210326194763184, "learning_rate": 1.078873397669951e-05, "loss": 3.6833, "step": 205050 }, { "epoch": 2.3041966487476335, "grad_norm": 1.1588329076766968, "learning_rate": 1.078648706339666e-05, "loss": 3.6682, "step": 205100 }, { "epoch": 2.304758373917977, "grad_norm": 1.1705402135849, "learning_rate": 1.078424015009381e-05, "loss": 3.6154, "step": 205150 }, { "epoch": 2.30532009908832, "grad_norm": 1.0182251930236816, "learning_rate": 1.078199323679096e-05, "loss": 3.6416, "step": 205200 }, { "epoch": 2.3058818242586634, "grad_norm": 1.4481021165847778, "learning_rate": 1.077974632348811e-05, "loss": 3.5957, "step": 205250 }, { "epoch": 2.3064435494290065, "grad_norm": 1.7506201267242432, "learning_rate": 1.077749941018526e-05, "loss": 3.6368, "step": 205300 }, { "epoch": 2.3070052745993497, "grad_norm": 1.2117962837219238, "learning_rate": 1.0775252496882409e-05, "loss": 3.6373, "step": 205350 }, { "epoch": 2.307566999769693, "grad_norm": 1.046059012413025, "learning_rate": 1.0773005583579558e-05, "loss": 3.6224, "step": 205400 }, { "epoch": 2.308128724940036, "grad_norm": 1.2020015716552734, "learning_rate": 1.077075867027671e-05, "loss": 3.6837, "step": 205450 }, { "epoch": 2.308690450110379, "grad_norm": 1.1153267621994019, "learning_rate": 1.0768511756973859e-05, "loss": 3.6947, "step": 205500 }, { "epoch": 2.3092521752807222, "grad_norm": 1.2907932996749878, "learning_rate": 1.0766264843671008e-05, "loss": 3.7269, "step": 205550 }, { "epoch": 2.3098139004510654, "grad_norm": 1.1694223880767822, "learning_rate": 1.0764017930368157e-05, "loss": 3.6491, "step": 205600 }, { "epoch": 2.3103756256214085, "grad_norm": 1.2663607597351074, "learning_rate": 1.0761771017065308e-05, "loss": 3.6718, "step": 205650 }, { "epoch": 2.3109373507917517, "grad_norm": 1.1763992309570312, "learning_rate": 1.0759524103762457e-05, "loss": 3.5671, "step": 205700 }, { "epoch": 2.311499075962095, "grad_norm": 1.0478515625, "learning_rate": 1.0757277190459607e-05, "loss": 3.6266, "step": 205750 }, { "epoch": 2.312060801132438, "grad_norm": 1.0747405290603638, "learning_rate": 1.0755030277156758e-05, "loss": 3.6551, "step": 205800 }, { "epoch": 2.312622526302781, "grad_norm": 1.0794366598129272, "learning_rate": 1.0752783363853907e-05, "loss": 3.6115, "step": 205850 }, { "epoch": 2.3131842514731242, "grad_norm": 1.2503647804260254, "learning_rate": 1.0750536450551056e-05, "loss": 3.6626, "step": 205900 }, { "epoch": 2.3137459766434674, "grad_norm": 1.3416721820831299, "learning_rate": 1.0748289537248206e-05, "loss": 3.6613, "step": 205950 }, { "epoch": 2.3143077018138105, "grad_norm": 1.128549337387085, "learning_rate": 1.0746042623945357e-05, "loss": 3.6929, "step": 206000 }, { "epoch": 2.3148694269841537, "grad_norm": 1.3488574028015137, "learning_rate": 1.0743795710642506e-05, "loss": 3.6999, "step": 206050 }, { "epoch": 2.315431152154497, "grad_norm": 1.2450639009475708, "learning_rate": 1.0741548797339655e-05, "loss": 3.6602, "step": 206100 }, { "epoch": 2.31599287732484, "grad_norm": 1.137138843536377, "learning_rate": 1.0739301884036806e-05, "loss": 3.6707, "step": 206150 }, { "epoch": 2.316554602495183, "grad_norm": 1.1490094661712646, "learning_rate": 1.073714484726607e-05, "loss": 3.6099, "step": 206200 }, { "epoch": 2.317116327665526, "grad_norm": 1.1616986989974976, "learning_rate": 1.073489793396322e-05, "loss": 3.658, "step": 206250 }, { "epoch": 2.3176780528358694, "grad_norm": 1.285482406616211, "learning_rate": 1.073265102066037e-05, "loss": 3.6447, "step": 206300 }, { "epoch": 2.3182397780062125, "grad_norm": 1.3026942014694214, "learning_rate": 1.0730404107357519e-05, "loss": 3.7069, "step": 206350 }, { "epoch": 2.3188015031765556, "grad_norm": 1.1542354822158813, "learning_rate": 1.072815719405467e-05, "loss": 3.724, "step": 206400 }, { "epoch": 2.319363228346899, "grad_norm": 1.1903376579284668, "learning_rate": 1.0725910280751819e-05, "loss": 3.5683, "step": 206450 }, { "epoch": 2.3199249535172424, "grad_norm": 1.0221582651138306, "learning_rate": 1.0723663367448968e-05, "loss": 3.6657, "step": 206500 }, { "epoch": 2.320486678687585, "grad_norm": 1.1555572748184204, "learning_rate": 1.072141645414612e-05, "loss": 3.6727, "step": 206550 }, { "epoch": 2.3210484038579287, "grad_norm": 1.0467361211776733, "learning_rate": 1.0719169540843269e-05, "loss": 3.6765, "step": 206600 }, { "epoch": 2.321610129028272, "grad_norm": 1.1794931888580322, "learning_rate": 1.0716922627540418e-05, "loss": 3.7091, "step": 206650 }, { "epoch": 2.322171854198615, "grad_norm": 1.255722999572754, "learning_rate": 1.0714675714237567e-05, "loss": 3.6278, "step": 206700 }, { "epoch": 2.322733579368958, "grad_norm": 1.2511638402938843, "learning_rate": 1.0712428800934718e-05, "loss": 3.672, "step": 206750 }, { "epoch": 2.323295304539301, "grad_norm": 1.034964680671692, "learning_rate": 1.0710181887631868e-05, "loss": 3.6508, "step": 206800 }, { "epoch": 2.3238570297096444, "grad_norm": 1.3474657535552979, "learning_rate": 1.0707934974329017e-05, "loss": 3.6476, "step": 206850 }, { "epoch": 2.3244187548799875, "grad_norm": 1.443680763244629, "learning_rate": 1.0705688061026168e-05, "loss": 3.6276, "step": 206900 }, { "epoch": 2.3249804800503306, "grad_norm": 1.1959564685821533, "learning_rate": 1.0703441147723317e-05, "loss": 3.6052, "step": 206950 }, { "epoch": 2.325542205220674, "grad_norm": 1.2000470161437988, "learning_rate": 1.0701194234420466e-05, "loss": 3.6673, "step": 207000 }, { "epoch": 2.326103930391017, "grad_norm": 1.1703729629516602, "learning_rate": 1.0698947321117616e-05, "loss": 3.6643, "step": 207050 }, { "epoch": 2.32666565556136, "grad_norm": 1.043171763420105, "learning_rate": 1.0696700407814767e-05, "loss": 3.65, "step": 207100 }, { "epoch": 2.327227380731703, "grad_norm": 1.255172610282898, "learning_rate": 1.0694453494511916e-05, "loss": 3.6703, "step": 207150 }, { "epoch": 2.3277891059020464, "grad_norm": 1.2360597848892212, "learning_rate": 1.0692206581209065e-05, "loss": 3.6407, "step": 207200 }, { "epoch": 2.3283508310723895, "grad_norm": 1.3001174926757812, "learning_rate": 1.0689959667906216e-05, "loss": 3.6984, "step": 207250 }, { "epoch": 2.3289125562427326, "grad_norm": 1.0854785442352295, "learning_rate": 1.0687712754603366e-05, "loss": 3.5829, "step": 207300 }, { "epoch": 2.3294742814130758, "grad_norm": 1.53227698802948, "learning_rate": 1.0685465841300515e-05, "loss": 3.6344, "step": 207350 }, { "epoch": 2.330036006583419, "grad_norm": 1.0187820196151733, "learning_rate": 1.0683218927997664e-05, "loss": 3.6426, "step": 207400 }, { "epoch": 2.330597731753762, "grad_norm": 1.227591633796692, "learning_rate": 1.0680972014694815e-05, "loss": 3.6018, "step": 207450 }, { "epoch": 2.331159456924105, "grad_norm": 1.2064884901046753, "learning_rate": 1.0678725101391965e-05, "loss": 3.6709, "step": 207500 }, { "epoch": 2.3317211820944483, "grad_norm": 1.5153018236160278, "learning_rate": 1.0676478188089114e-05, "loss": 3.6663, "step": 207550 }, { "epoch": 2.3322829072647915, "grad_norm": 0.9832055568695068, "learning_rate": 1.0674231274786263e-05, "loss": 3.6747, "step": 207600 }, { "epoch": 2.3328446324351346, "grad_norm": 1.329431176185608, "learning_rate": 1.0671984361483414e-05, "loss": 3.6594, "step": 207650 }, { "epoch": 2.3334063576054778, "grad_norm": 1.0778201818466187, "learning_rate": 1.0669737448180564e-05, "loss": 3.6171, "step": 207700 }, { "epoch": 2.3339680827758214, "grad_norm": 1.1410647630691528, "learning_rate": 1.0667490534877713e-05, "loss": 3.64, "step": 207750 }, { "epoch": 2.334529807946164, "grad_norm": 1.1544653177261353, "learning_rate": 1.0665243621574864e-05, "loss": 3.6708, "step": 207800 }, { "epoch": 2.3350915331165076, "grad_norm": 1.1542446613311768, "learning_rate": 1.0662996708272013e-05, "loss": 3.6195, "step": 207850 }, { "epoch": 2.3356532582868503, "grad_norm": 0.9414635896682739, "learning_rate": 1.0660749794969163e-05, "loss": 3.7031, "step": 207900 }, { "epoch": 2.336214983457194, "grad_norm": 1.4521923065185547, "learning_rate": 1.0658502881666312e-05, "loss": 3.6119, "step": 207950 }, { "epoch": 2.336776708627537, "grad_norm": 1.123447299003601, "learning_rate": 1.0656255968363463e-05, "loss": 3.6402, "step": 208000 }, { "epoch": 2.33733843379788, "grad_norm": 1.0561689138412476, "learning_rate": 1.0654009055060612e-05, "loss": 3.6748, "step": 208050 }, { "epoch": 2.3379001589682233, "grad_norm": 1.0808101892471313, "learning_rate": 1.0651762141757762e-05, "loss": 3.6409, "step": 208100 }, { "epoch": 2.3384618841385665, "grad_norm": 1.194089651107788, "learning_rate": 1.0649515228454913e-05, "loss": 3.6534, "step": 208150 }, { "epoch": 2.3390236093089096, "grad_norm": 2.9633257389068604, "learning_rate": 1.0647268315152062e-05, "loss": 3.6217, "step": 208200 }, { "epoch": 2.3395853344792528, "grad_norm": 1.3349636793136597, "learning_rate": 1.0645021401849211e-05, "loss": 3.6271, "step": 208250 }, { "epoch": 2.340147059649596, "grad_norm": 1.3026089668273926, "learning_rate": 1.064277448854636e-05, "loss": 3.6668, "step": 208300 }, { "epoch": 2.340708784819939, "grad_norm": 4.803406715393066, "learning_rate": 1.0640527575243512e-05, "loss": 3.6234, "step": 208350 }, { "epoch": 2.341270509990282, "grad_norm": 1.1711561679840088, "learning_rate": 1.0638280661940661e-05, "loss": 3.6529, "step": 208400 }, { "epoch": 2.3418322351606253, "grad_norm": 1.007407784461975, "learning_rate": 1.063603374863781e-05, "loss": 3.646, "step": 208450 }, { "epoch": 2.3423939603309685, "grad_norm": 1.0475504398345947, "learning_rate": 1.0633786835334961e-05, "loss": 3.767, "step": 208500 }, { "epoch": 2.3429556855013116, "grad_norm": 1.1315566301345825, "learning_rate": 1.063153992203211e-05, "loss": 3.6716, "step": 208550 }, { "epoch": 2.3435174106716548, "grad_norm": 1.1114345788955688, "learning_rate": 1.062929300872926e-05, "loss": 3.6099, "step": 208600 }, { "epoch": 2.344079135841998, "grad_norm": 1.3237587213516235, "learning_rate": 1.062704609542641e-05, "loss": 3.697, "step": 208650 }, { "epoch": 2.344640861012341, "grad_norm": 1.151031255722046, "learning_rate": 1.062479918212356e-05, "loss": 3.6629, "step": 208700 }, { "epoch": 2.345202586182684, "grad_norm": 1.1777541637420654, "learning_rate": 1.062255226882071e-05, "loss": 3.676, "step": 208750 }, { "epoch": 2.3457643113530273, "grad_norm": 1.126303791999817, "learning_rate": 1.0620305355517859e-05, "loss": 3.554, "step": 208800 }, { "epoch": 2.3463260365233705, "grad_norm": 1.1741201877593994, "learning_rate": 1.061805844221501e-05, "loss": 3.6529, "step": 208850 }, { "epoch": 2.3468877616937136, "grad_norm": 1.0780304670333862, "learning_rate": 1.061581152891216e-05, "loss": 3.6431, "step": 208900 }, { "epoch": 2.3474494868640567, "grad_norm": 1.1062053442001343, "learning_rate": 1.0613564615609308e-05, "loss": 3.6463, "step": 208950 }, { "epoch": 2.3480112120344, "grad_norm": 1.2727903127670288, "learning_rate": 1.0611317702306458e-05, "loss": 3.6653, "step": 209000 }, { "epoch": 2.348572937204743, "grad_norm": 1.1275520324707031, "learning_rate": 1.0609070789003609e-05, "loss": 3.6461, "step": 209050 }, { "epoch": 2.3491346623750866, "grad_norm": 1.3311638832092285, "learning_rate": 1.0606823875700758e-05, "loss": 3.702, "step": 209100 }, { "epoch": 2.3496963875454293, "grad_norm": 1.1495685577392578, "learning_rate": 1.0604576962397907e-05, "loss": 3.6863, "step": 209150 }, { "epoch": 2.350258112715773, "grad_norm": 1.2845324277877808, "learning_rate": 1.0602330049095057e-05, "loss": 3.6757, "step": 209200 }, { "epoch": 2.3508198378861156, "grad_norm": 5.180753231048584, "learning_rate": 1.0600083135792208e-05, "loss": 3.6989, "step": 209250 }, { "epoch": 2.351381563056459, "grad_norm": 1.39224374294281, "learning_rate": 1.0597836222489357e-05, "loss": 3.5864, "step": 209300 }, { "epoch": 2.3519432882268023, "grad_norm": 1.1731806993484497, "learning_rate": 1.0595589309186506e-05, "loss": 3.6332, "step": 209350 }, { "epoch": 2.3525050133971455, "grad_norm": 1.3183012008666992, "learning_rate": 1.0593342395883657e-05, "loss": 3.6614, "step": 209400 }, { "epoch": 2.3530667385674886, "grad_norm": 1.3669489622116089, "learning_rate": 1.0591095482580807e-05, "loss": 3.6488, "step": 209450 }, { "epoch": 2.3536284637378317, "grad_norm": 1.0421273708343506, "learning_rate": 1.0588848569277956e-05, "loss": 3.5317, "step": 209500 }, { "epoch": 2.354190188908175, "grad_norm": 1.057143211364746, "learning_rate": 1.0586601655975105e-05, "loss": 3.6176, "step": 209550 }, { "epoch": 2.354751914078518, "grad_norm": 1.0931198596954346, "learning_rate": 1.0584354742672256e-05, "loss": 3.6886, "step": 209600 }, { "epoch": 2.355313639248861, "grad_norm": 1.4499849081039429, "learning_rate": 1.0582107829369406e-05, "loss": 3.7207, "step": 209650 }, { "epoch": 2.3558753644192043, "grad_norm": 1.3522409200668335, "learning_rate": 1.0579860916066555e-05, "loss": 3.6915, "step": 209700 }, { "epoch": 2.3564370895895475, "grad_norm": 1.271995186805725, "learning_rate": 1.0577614002763706e-05, "loss": 3.6232, "step": 209750 }, { "epoch": 2.3569988147598906, "grad_norm": 1.169212818145752, "learning_rate": 1.0575367089460855e-05, "loss": 3.7393, "step": 209800 }, { "epoch": 2.3575605399302337, "grad_norm": 1.1130293607711792, "learning_rate": 1.0573120176158005e-05, "loss": 3.6652, "step": 209850 }, { "epoch": 2.358122265100577, "grad_norm": 1.2910044193267822, "learning_rate": 1.0570873262855154e-05, "loss": 3.7016, "step": 209900 }, { "epoch": 2.35868399027092, "grad_norm": 1.1066601276397705, "learning_rate": 1.0568626349552305e-05, "loss": 3.6889, "step": 209950 }, { "epoch": 2.359245715441263, "grad_norm": 1.2973527908325195, "learning_rate": 1.0566379436249454e-05, "loss": 3.6981, "step": 210000 }, { "epoch": 2.3598074406116063, "grad_norm": 1.340416431427002, "learning_rate": 1.0564132522946604e-05, "loss": 3.5788, "step": 210050 }, { "epoch": 2.3603691657819494, "grad_norm": 1.1649587154388428, "learning_rate": 1.0561885609643755e-05, "loss": 3.6245, "step": 210100 }, { "epoch": 2.3609308909522926, "grad_norm": 1.2052050828933716, "learning_rate": 1.0559638696340902e-05, "loss": 3.7734, "step": 210150 }, { "epoch": 2.3614926161226357, "grad_norm": 1.1914728879928589, "learning_rate": 1.0557391783038052e-05, "loss": 3.6131, "step": 210200 }, { "epoch": 2.362054341292979, "grad_norm": 1.1473121643066406, "learning_rate": 1.0555144869735201e-05, "loss": 3.6201, "step": 210250 }, { "epoch": 2.362616066463322, "grad_norm": 1.0467044115066528, "learning_rate": 1.055289795643235e-05, "loss": 3.6894, "step": 210300 }, { "epoch": 2.363177791633665, "grad_norm": 1.103670358657837, "learning_rate": 1.0550651043129501e-05, "loss": 3.6476, "step": 210350 }, { "epoch": 2.3637395168040083, "grad_norm": 1.2465933561325073, "learning_rate": 1.054840412982665e-05, "loss": 3.5981, "step": 210400 }, { "epoch": 2.364301241974352, "grad_norm": 1.1501760482788086, "learning_rate": 1.05461572165238e-05, "loss": 3.6431, "step": 210450 }, { "epoch": 2.3648629671446946, "grad_norm": 1.0407079458236694, "learning_rate": 1.054391030322095e-05, "loss": 3.6348, "step": 210500 }, { "epoch": 2.365424692315038, "grad_norm": 1.2164947986602783, "learning_rate": 1.05416633899181e-05, "loss": 3.6887, "step": 210550 }, { "epoch": 2.365986417485381, "grad_norm": 1.007479190826416, "learning_rate": 1.053941647661525e-05, "loss": 3.5831, "step": 210600 }, { "epoch": 2.3665481426557244, "grad_norm": 1.0290968418121338, "learning_rate": 1.0537169563312399e-05, "loss": 3.5973, "step": 210650 }, { "epoch": 2.3671098678260676, "grad_norm": 1.297382116317749, "learning_rate": 1.053492265000955e-05, "loss": 3.7142, "step": 210700 }, { "epoch": 2.3676715929964107, "grad_norm": 1.215598464012146, "learning_rate": 1.0532675736706699e-05, "loss": 3.6544, "step": 210750 }, { "epoch": 2.368233318166754, "grad_norm": 1.5987659692764282, "learning_rate": 1.0530428823403848e-05, "loss": 3.6862, "step": 210800 }, { "epoch": 2.368795043337097, "grad_norm": 1.0816415548324585, "learning_rate": 1.0528181910100998e-05, "loss": 3.5947, "step": 210850 }, { "epoch": 2.36935676850744, "grad_norm": 1.255016565322876, "learning_rate": 1.0525934996798149e-05, "loss": 3.5878, "step": 210900 }, { "epoch": 2.3699184936777833, "grad_norm": 1.2578915357589722, "learning_rate": 1.0523688083495298e-05, "loss": 3.6189, "step": 210950 }, { "epoch": 2.3704802188481264, "grad_norm": 1.1681090593338013, "learning_rate": 1.0521441170192447e-05, "loss": 3.6204, "step": 211000 }, { "epoch": 2.3710419440184696, "grad_norm": 1.3190033435821533, "learning_rate": 1.0519194256889598e-05, "loss": 3.6094, "step": 211050 }, { "epoch": 2.3716036691888127, "grad_norm": 1.0447425842285156, "learning_rate": 1.0516947343586748e-05, "loss": 3.6688, "step": 211100 }, { "epoch": 2.372165394359156, "grad_norm": 1.2852200269699097, "learning_rate": 1.0514700430283897e-05, "loss": 3.639, "step": 211150 }, { "epoch": 2.372727119529499, "grad_norm": 1.0362699031829834, "learning_rate": 1.0512453516981046e-05, "loss": 3.6384, "step": 211200 }, { "epoch": 2.373288844699842, "grad_norm": 1.2051905393600464, "learning_rate": 1.0510206603678197e-05, "loss": 3.6728, "step": 211250 }, { "epoch": 2.3738505698701853, "grad_norm": 1.088995337486267, "learning_rate": 1.0507959690375347e-05, "loss": 3.6041, "step": 211300 }, { "epoch": 2.3744122950405284, "grad_norm": 1.1478296518325806, "learning_rate": 1.0505712777072496e-05, "loss": 3.6943, "step": 211350 }, { "epoch": 2.3749740202108716, "grad_norm": 1.334021806716919, "learning_rate": 1.0503465863769647e-05, "loss": 3.6199, "step": 211400 }, { "epoch": 2.3755357453812147, "grad_norm": 1.2082432508468628, "learning_rate": 1.0501218950466796e-05, "loss": 3.7031, "step": 211450 }, { "epoch": 2.376097470551558, "grad_norm": 1.133233666419983, "learning_rate": 1.0498972037163946e-05, "loss": 3.7362, "step": 211500 }, { "epoch": 2.376659195721901, "grad_norm": 1.2972453832626343, "learning_rate": 1.0496725123861095e-05, "loss": 3.6044, "step": 211550 }, { "epoch": 2.377220920892244, "grad_norm": 1.0553796291351318, "learning_rate": 1.0494478210558246e-05, "loss": 3.6411, "step": 211600 }, { "epoch": 2.3777826460625873, "grad_norm": 1.4199177026748657, "learning_rate": 1.0492231297255395e-05, "loss": 3.6847, "step": 211650 }, { "epoch": 2.3783443712329304, "grad_norm": 1.332186222076416, "learning_rate": 1.0489984383952545e-05, "loss": 3.645, "step": 211700 }, { "epoch": 2.3789060964032736, "grad_norm": 1.034148097038269, "learning_rate": 1.0487737470649696e-05, "loss": 3.6861, "step": 211750 }, { "epoch": 2.379467821573617, "grad_norm": 1.1468724012374878, "learning_rate": 1.0485490557346845e-05, "loss": 3.6651, "step": 211800 }, { "epoch": 2.38002954674396, "grad_norm": 1.186020851135254, "learning_rate": 1.0483243644043994e-05, "loss": 3.6186, "step": 211850 }, { "epoch": 2.3805912719143034, "grad_norm": 1.140351414680481, "learning_rate": 1.0480996730741144e-05, "loss": 3.6236, "step": 211900 }, { "epoch": 2.381152997084646, "grad_norm": 1.1734734773635864, "learning_rate": 1.0478749817438295e-05, "loss": 3.6766, "step": 211950 }, { "epoch": 2.3817147222549897, "grad_norm": 1.1378458738327026, "learning_rate": 1.0476502904135444e-05, "loss": 3.6499, "step": 212000 }, { "epoch": 2.382276447425333, "grad_norm": 1.2333191633224487, "learning_rate": 1.0474255990832593e-05, "loss": 3.7074, "step": 212050 }, { "epoch": 2.382838172595676, "grad_norm": 1.0585025548934937, "learning_rate": 1.0472009077529744e-05, "loss": 3.6026, "step": 212100 }, { "epoch": 2.383399897766019, "grad_norm": 1.2075213193893433, "learning_rate": 1.0469762164226894e-05, "loss": 3.7076, "step": 212150 }, { "epoch": 2.3839616229363623, "grad_norm": 1.148812174797058, "learning_rate": 1.0467515250924043e-05, "loss": 3.6372, "step": 212200 }, { "epoch": 2.3845233481067054, "grad_norm": 1.156119704246521, "learning_rate": 1.0465268337621192e-05, "loss": 3.6752, "step": 212250 }, { "epoch": 2.3850850732770486, "grad_norm": 1.3167558908462524, "learning_rate": 1.0463021424318343e-05, "loss": 3.653, "step": 212300 }, { "epoch": 2.3856467984473917, "grad_norm": 1.27252995967865, "learning_rate": 1.0460774511015492e-05, "loss": 3.6471, "step": 212350 }, { "epoch": 2.386208523617735, "grad_norm": 1.3370920419692993, "learning_rate": 1.0458527597712642e-05, "loss": 3.6483, "step": 212400 }, { "epoch": 2.386770248788078, "grad_norm": 1.3032448291778564, "learning_rate": 1.0456280684409791e-05, "loss": 3.6389, "step": 212450 }, { "epoch": 2.387331973958421, "grad_norm": 1.0964500904083252, "learning_rate": 1.0454033771106942e-05, "loss": 3.6809, "step": 212500 }, { "epoch": 2.3878936991287643, "grad_norm": 1.074172019958496, "learning_rate": 1.0451786857804091e-05, "loss": 3.7075, "step": 212550 }, { "epoch": 2.3884554242991074, "grad_norm": 1.2807329893112183, "learning_rate": 1.044953994450124e-05, "loss": 3.6838, "step": 212600 }, { "epoch": 2.3890171494694505, "grad_norm": 1.1744275093078613, "learning_rate": 1.0447293031198392e-05, "loss": 3.6299, "step": 212650 }, { "epoch": 2.3895788746397937, "grad_norm": 1.3053351640701294, "learning_rate": 1.0445046117895541e-05, "loss": 3.6463, "step": 212700 }, { "epoch": 2.390140599810137, "grad_norm": 1.2043471336364746, "learning_rate": 1.044279920459269e-05, "loss": 3.6402, "step": 212750 }, { "epoch": 2.39070232498048, "grad_norm": 1.0350993871688843, "learning_rate": 1.044055229128984e-05, "loss": 3.6565, "step": 212800 }, { "epoch": 2.391264050150823, "grad_norm": 1.1305289268493652, "learning_rate": 1.0438350316253047e-05, "loss": 3.7486, "step": 212850 }, { "epoch": 2.3918257753211662, "grad_norm": 1.2447079420089722, "learning_rate": 1.0436103402950197e-05, "loss": 3.6249, "step": 212900 }, { "epoch": 2.3923875004915094, "grad_norm": 1.3462215662002563, "learning_rate": 1.0433856489647348e-05, "loss": 3.6136, "step": 212950 }, { "epoch": 2.3929492256618525, "grad_norm": 1.3543200492858887, "learning_rate": 1.0431609576344497e-05, "loss": 3.6838, "step": 213000 }, { "epoch": 2.3935109508321957, "grad_norm": 1.217747449874878, "learning_rate": 1.0429362663041646e-05, "loss": 3.6542, "step": 213050 }, { "epoch": 2.394072676002539, "grad_norm": 1.1525721549987793, "learning_rate": 1.0427115749738797e-05, "loss": 3.6055, "step": 213100 }, { "epoch": 2.3946344011728824, "grad_norm": 1.1056920289993286, "learning_rate": 1.0424868836435947e-05, "loss": 3.6782, "step": 213150 }, { "epoch": 2.395196126343225, "grad_norm": 0.9926613569259644, "learning_rate": 1.0422621923133096e-05, "loss": 3.6243, "step": 213200 }, { "epoch": 2.3957578515135687, "grad_norm": 1.0209110975265503, "learning_rate": 1.0420419948096304e-05, "loss": 3.6288, "step": 213250 }, { "epoch": 2.396319576683912, "grad_norm": 1.0990540981292725, "learning_rate": 1.0418173034793453e-05, "loss": 3.6718, "step": 213300 }, { "epoch": 2.396881301854255, "grad_norm": 1.2863000631332397, "learning_rate": 1.0415926121490602e-05, "loss": 3.6767, "step": 213350 }, { "epoch": 2.397443027024598, "grad_norm": 1.0847902297973633, "learning_rate": 1.0413679208187753e-05, "loss": 3.671, "step": 213400 }, { "epoch": 2.3980047521949412, "grad_norm": 1.220410943031311, "learning_rate": 1.0411432294884903e-05, "loss": 3.5818, "step": 213450 }, { "epoch": 2.3985664773652844, "grad_norm": 2.0756099224090576, "learning_rate": 1.0409185381582052e-05, "loss": 3.6215, "step": 213500 }, { "epoch": 2.3991282025356275, "grad_norm": 1.3049434423446655, "learning_rate": 1.0406938468279201e-05, "loss": 3.6558, "step": 213550 }, { "epoch": 2.3996899277059707, "grad_norm": 1.2601306438446045, "learning_rate": 1.0404691554976352e-05, "loss": 3.5866, "step": 213600 }, { "epoch": 2.400251652876314, "grad_norm": 1.2760636806488037, "learning_rate": 1.0402444641673501e-05, "loss": 3.6674, "step": 213650 }, { "epoch": 2.400813378046657, "grad_norm": 1.0744049549102783, "learning_rate": 1.040019772837065e-05, "loss": 3.6744, "step": 213700 }, { "epoch": 2.401375103217, "grad_norm": 1.2868483066558838, "learning_rate": 1.0397950815067802e-05, "loss": 3.6368, "step": 213750 }, { "epoch": 2.4019368283873432, "grad_norm": 1.3159387111663818, "learning_rate": 1.0395703901764951e-05, "loss": 3.6556, "step": 213800 }, { "epoch": 2.4024985535576864, "grad_norm": 1.1377818584442139, "learning_rate": 1.03934569884621e-05, "loss": 3.6512, "step": 213850 }, { "epoch": 2.4030602787280295, "grad_norm": 1.1791696548461914, "learning_rate": 1.039121007515925e-05, "loss": 3.6349, "step": 213900 }, { "epoch": 2.4036220038983727, "grad_norm": 1.1945469379425049, "learning_rate": 1.03889631618564e-05, "loss": 3.6634, "step": 213950 }, { "epoch": 2.404183729068716, "grad_norm": 1.2033038139343262, "learning_rate": 1.038671624855355e-05, "loss": 3.639, "step": 214000 }, { "epoch": 2.404745454239059, "grad_norm": 1.6592442989349365, "learning_rate": 1.03844693352507e-05, "loss": 3.6357, "step": 214050 }, { "epoch": 2.405307179409402, "grad_norm": 1.1835962533950806, "learning_rate": 1.038222242194785e-05, "loss": 3.6082, "step": 214100 }, { "epoch": 2.4058689045797452, "grad_norm": 1.2160457372665405, "learning_rate": 1.0379975508645e-05, "loss": 3.6424, "step": 214150 }, { "epoch": 2.4064306297500884, "grad_norm": 1.13961923122406, "learning_rate": 1.0377773533608206e-05, "loss": 3.6699, "step": 214200 }, { "epoch": 2.4069923549204315, "grad_norm": 1.2178468704223633, "learning_rate": 1.0375526620305357e-05, "loss": 3.6302, "step": 214250 }, { "epoch": 2.4075540800907747, "grad_norm": 1.180449366569519, "learning_rate": 1.0373279707002506e-05, "loss": 3.6606, "step": 214300 }, { "epoch": 2.408115805261118, "grad_norm": 1.0457371473312378, "learning_rate": 1.0371032793699655e-05, "loss": 3.6764, "step": 214350 }, { "epoch": 2.408677530431461, "grad_norm": 1.032880425453186, "learning_rate": 1.0368785880396806e-05, "loss": 3.6936, "step": 214400 }, { "epoch": 2.409239255601804, "grad_norm": 1.0636664628982544, "learning_rate": 1.0366538967093956e-05, "loss": 3.605, "step": 214450 }, { "epoch": 2.4098009807721477, "grad_norm": 1.0582681894302368, "learning_rate": 1.0364292053791105e-05, "loss": 3.7026, "step": 214500 }, { "epoch": 2.4103627059424904, "grad_norm": 1.1255106925964355, "learning_rate": 1.0362045140488254e-05, "loss": 3.6631, "step": 214550 }, { "epoch": 2.410924431112834, "grad_norm": 1.17404305934906, "learning_rate": 1.0359798227185405e-05, "loss": 3.6122, "step": 214600 }, { "epoch": 2.411486156283177, "grad_norm": 1.4032615423202515, "learning_rate": 1.0357551313882555e-05, "loss": 3.6654, "step": 214650 }, { "epoch": 2.4120478814535202, "grad_norm": 1.0625853538513184, "learning_rate": 1.0355304400579704e-05, "loss": 3.7238, "step": 214700 }, { "epoch": 2.4126096066238634, "grad_norm": 1.2413190603256226, "learning_rate": 1.0353057487276855e-05, "loss": 3.7059, "step": 214750 }, { "epoch": 2.4131713317942065, "grad_norm": 0.9958818554878235, "learning_rate": 1.0350810573974004e-05, "loss": 3.6565, "step": 214800 }, { "epoch": 2.4137330569645497, "grad_norm": 1.2933051586151123, "learning_rate": 1.0348563660671154e-05, "loss": 3.649, "step": 214850 }, { "epoch": 2.414294782134893, "grad_norm": 1.0213230848312378, "learning_rate": 1.0346316747368303e-05, "loss": 3.6311, "step": 214900 }, { "epoch": 2.414856507305236, "grad_norm": 1.2674108743667603, "learning_rate": 1.0344069834065454e-05, "loss": 3.7243, "step": 214950 }, { "epoch": 2.415418232475579, "grad_norm": 1.2302976846694946, "learning_rate": 1.0341822920762603e-05, "loss": 3.7059, "step": 215000 }, { "epoch": 2.415979957645922, "grad_norm": 1.2191627025604248, "learning_rate": 1.0339576007459753e-05, "loss": 3.6319, "step": 215050 }, { "epoch": 2.4165416828162654, "grad_norm": 1.215530514717102, "learning_rate": 1.0337329094156902e-05, "loss": 3.6553, "step": 215100 }, { "epoch": 2.4171034079866085, "grad_norm": 1.2192349433898926, "learning_rate": 1.0335082180854053e-05, "loss": 3.6413, "step": 215150 }, { "epoch": 2.4176651331569516, "grad_norm": 1.132501244544983, "learning_rate": 1.0332835267551202e-05, "loss": 3.6787, "step": 215200 }, { "epoch": 2.418226858327295, "grad_norm": 1.065586805343628, "learning_rate": 1.0330588354248351e-05, "loss": 3.5863, "step": 215250 }, { "epoch": 2.418788583497638, "grad_norm": 1.0597398281097412, "learning_rate": 1.0328341440945502e-05, "loss": 3.6315, "step": 215300 }, { "epoch": 2.419350308667981, "grad_norm": 1.0700304508209229, "learning_rate": 1.0326094527642652e-05, "loss": 3.7679, "step": 215350 }, { "epoch": 2.419912033838324, "grad_norm": 1.3021875619888306, "learning_rate": 1.0323847614339801e-05, "loss": 3.7039, "step": 215400 }, { "epoch": 2.4204737590086673, "grad_norm": 1.3365675210952759, "learning_rate": 1.032160070103695e-05, "loss": 3.6874, "step": 215450 }, { "epoch": 2.4210354841790105, "grad_norm": 1.261286735534668, "learning_rate": 1.0319353787734101e-05, "loss": 3.599, "step": 215500 }, { "epoch": 2.4215972093493536, "grad_norm": 1.128183364868164, "learning_rate": 1.031710687443125e-05, "loss": 3.6616, "step": 215550 }, { "epoch": 2.4221589345196968, "grad_norm": 1.0821489095687866, "learning_rate": 1.03148599611284e-05, "loss": 3.6034, "step": 215600 }, { "epoch": 2.42272065969004, "grad_norm": 1.2359782457351685, "learning_rate": 1.0312613047825551e-05, "loss": 3.6662, "step": 215650 }, { "epoch": 2.423282384860383, "grad_norm": 1.2059345245361328, "learning_rate": 1.03103661345227e-05, "loss": 3.5647, "step": 215700 }, { "epoch": 2.423844110030726, "grad_norm": 1.7044291496276855, "learning_rate": 1.030811922121985e-05, "loss": 3.6759, "step": 215750 }, { "epoch": 2.4244058352010693, "grad_norm": 1.320499062538147, "learning_rate": 1.0305872307916999e-05, "loss": 3.6642, "step": 215800 }, { "epoch": 2.424967560371413, "grad_norm": 1.1735167503356934, "learning_rate": 1.030362539461415e-05, "loss": 3.6223, "step": 215850 }, { "epoch": 2.4255292855417556, "grad_norm": 1.0095165967941284, "learning_rate": 1.03013784813113e-05, "loss": 3.6482, "step": 215900 }, { "epoch": 2.426091010712099, "grad_norm": 1.0431257486343384, "learning_rate": 1.0299131568008449e-05, "loss": 3.5853, "step": 215950 }, { "epoch": 2.4266527358824423, "grad_norm": 1.2413547039031982, "learning_rate": 1.02968846547056e-05, "loss": 3.6254, "step": 216000 }, { "epoch": 2.4272144610527855, "grad_norm": 1.143583059310913, "learning_rate": 1.0294637741402749e-05, "loss": 3.6449, "step": 216050 }, { "epoch": 2.4277761862231286, "grad_norm": 1.152782678604126, "learning_rate": 1.0292390828099898e-05, "loss": 3.6702, "step": 216100 }, { "epoch": 2.4283379113934718, "grad_norm": 1.3686628341674805, "learning_rate": 1.0290143914797048e-05, "loss": 3.6322, "step": 216150 }, { "epoch": 2.428899636563815, "grad_norm": 1.1874475479125977, "learning_rate": 1.0287897001494199e-05, "loss": 3.6287, "step": 216200 }, { "epoch": 2.429461361734158, "grad_norm": 1.4857885837554932, "learning_rate": 1.0285650088191348e-05, "loss": 3.6677, "step": 216250 }, { "epoch": 2.430023086904501, "grad_norm": 1.2705439329147339, "learning_rate": 1.0283403174888497e-05, "loss": 3.6905, "step": 216300 }, { "epoch": 2.4305848120748443, "grad_norm": 1.2015371322631836, "learning_rate": 1.0281156261585648e-05, "loss": 3.6073, "step": 216350 }, { "epoch": 2.4311465372451875, "grad_norm": 1.1872402429580688, "learning_rate": 1.0278909348282798e-05, "loss": 3.6097, "step": 216400 }, { "epoch": 2.4317082624155306, "grad_norm": 1.4883981943130493, "learning_rate": 1.0276662434979947e-05, "loss": 3.6469, "step": 216450 }, { "epoch": 2.4322699875858738, "grad_norm": 1.1034952402114868, "learning_rate": 1.0274415521677096e-05, "loss": 3.7157, "step": 216500 }, { "epoch": 2.432831712756217, "grad_norm": 1.0829899311065674, "learning_rate": 1.0272168608374247e-05, "loss": 3.6375, "step": 216550 }, { "epoch": 2.43339343792656, "grad_norm": 1.2546882629394531, "learning_rate": 1.0269921695071397e-05, "loss": 3.6124, "step": 216600 }, { "epoch": 2.433955163096903, "grad_norm": 1.122017741203308, "learning_rate": 1.0267674781768546e-05, "loss": 3.6131, "step": 216650 }, { "epoch": 2.4345168882672463, "grad_norm": 1.322969675064087, "learning_rate": 1.0265427868465697e-05, "loss": 3.6663, "step": 216700 }, { "epoch": 2.4350786134375895, "grad_norm": 1.2598247528076172, "learning_rate": 1.0263180955162846e-05, "loss": 3.6465, "step": 216750 }, { "epoch": 2.4356403386079326, "grad_norm": 1.169589877128601, "learning_rate": 1.0260934041859996e-05, "loss": 3.6928, "step": 216800 }, { "epoch": 2.4362020637782758, "grad_norm": 1.1836838722229004, "learning_rate": 1.0258687128557145e-05, "loss": 3.7588, "step": 216850 }, { "epoch": 2.436763788948619, "grad_norm": 1.3690922260284424, "learning_rate": 1.0256440215254296e-05, "loss": 3.7039, "step": 216900 }, { "epoch": 2.437325514118962, "grad_norm": 1.2842140197753906, "learning_rate": 1.0254193301951445e-05, "loss": 3.5944, "step": 216950 }, { "epoch": 2.437887239289305, "grad_norm": 1.17047119140625, "learning_rate": 1.0251946388648594e-05, "loss": 3.6459, "step": 217000 }, { "epoch": 2.4384489644596483, "grad_norm": 1.1833974123001099, "learning_rate": 1.0249699475345744e-05, "loss": 3.6239, "step": 217050 }, { "epoch": 2.4390106896299915, "grad_norm": 1.0102674961090088, "learning_rate": 1.0247452562042895e-05, "loss": 3.6602, "step": 217100 }, { "epoch": 2.4395724148003346, "grad_norm": 1.1659404039382935, "learning_rate": 1.0245205648740044e-05, "loss": 3.6185, "step": 217150 }, { "epoch": 2.440134139970678, "grad_norm": 1.080694317817688, "learning_rate": 1.0242958735437193e-05, "loss": 3.7509, "step": 217200 }, { "epoch": 2.440695865141021, "grad_norm": 1.0265114307403564, "learning_rate": 1.0240711822134344e-05, "loss": 3.6257, "step": 217250 }, { "epoch": 2.4412575903113645, "grad_norm": 1.1171821355819702, "learning_rate": 1.0238464908831494e-05, "loss": 3.6605, "step": 217300 }, { "epoch": 2.4418193154817076, "grad_norm": 1.123995304107666, "learning_rate": 1.0236217995528643e-05, "loss": 3.6906, "step": 217350 }, { "epoch": 2.4423810406520507, "grad_norm": 1.3015533685684204, "learning_rate": 1.0233971082225792e-05, "loss": 3.6372, "step": 217400 }, { "epoch": 2.442942765822394, "grad_norm": 1.2588542699813843, "learning_rate": 1.0231724168922943e-05, "loss": 3.6604, "step": 217450 }, { "epoch": 2.443504490992737, "grad_norm": 1.0011076927185059, "learning_rate": 1.0229477255620093e-05, "loss": 3.62, "step": 217500 }, { "epoch": 2.44406621616308, "grad_norm": 1.1005232334136963, "learning_rate": 1.0227230342317242e-05, "loss": 3.6412, "step": 217550 }, { "epoch": 2.4446279413334233, "grad_norm": 1.3701766729354858, "learning_rate": 1.0224983429014393e-05, "loss": 3.682, "step": 217600 }, { "epoch": 2.4451896665037665, "grad_norm": 1.1460520029067993, "learning_rate": 1.0222736515711542e-05, "loss": 3.6485, "step": 217650 }, { "epoch": 2.4457513916741096, "grad_norm": 1.1530455350875854, "learning_rate": 1.0220489602408692e-05, "loss": 3.6716, "step": 217700 }, { "epoch": 2.4463131168444527, "grad_norm": 1.154485821723938, "learning_rate": 1.0218242689105841e-05, "loss": 3.6729, "step": 217750 }, { "epoch": 2.446874842014796, "grad_norm": 1.321520209312439, "learning_rate": 1.0215995775802992e-05, "loss": 3.582, "step": 217800 }, { "epoch": 2.447436567185139, "grad_norm": 1.4846552610397339, "learning_rate": 1.0213748862500141e-05, "loss": 3.7066, "step": 217850 }, { "epoch": 2.447998292355482, "grad_norm": 2.0846121311187744, "learning_rate": 1.021150194919729e-05, "loss": 3.6042, "step": 217900 }, { "epoch": 2.4485600175258253, "grad_norm": 1.1489448547363281, "learning_rate": 1.0209255035894442e-05, "loss": 3.6159, "step": 217950 }, { "epoch": 2.4491217426961684, "grad_norm": 1.273450493812561, "learning_rate": 1.0207008122591591e-05, "loss": 3.5722, "step": 218000 }, { "epoch": 2.4496834678665116, "grad_norm": 1.0643285512924194, "learning_rate": 1.020476120928874e-05, "loss": 3.6386, "step": 218050 }, { "epoch": 2.4502451930368547, "grad_norm": 1.552664875984192, "learning_rate": 1.020251429598589e-05, "loss": 3.6878, "step": 218100 }, { "epoch": 2.450806918207198, "grad_norm": 1.3038806915283203, "learning_rate": 1.020026738268304e-05, "loss": 3.6344, "step": 218150 }, { "epoch": 2.451368643377541, "grad_norm": 1.5481359958648682, "learning_rate": 1.019802046938019e-05, "loss": 3.634, "step": 218200 }, { "epoch": 2.451930368547884, "grad_norm": 0.9933804273605347, "learning_rate": 1.019577355607734e-05, "loss": 3.5348, "step": 218250 }, { "epoch": 2.4524920937182273, "grad_norm": 0.9886470437049866, "learning_rate": 1.019352664277449e-05, "loss": 3.6177, "step": 218300 }, { "epoch": 2.4530538188885704, "grad_norm": 1.268970251083374, "learning_rate": 1.019127972947164e-05, "loss": 3.7232, "step": 218350 }, { "epoch": 2.4536155440589136, "grad_norm": 1.159010648727417, "learning_rate": 1.0189032816168789e-05, "loss": 3.7073, "step": 218400 }, { "epoch": 2.4541772692292567, "grad_norm": 1.1107714176177979, "learning_rate": 1.0186785902865938e-05, "loss": 3.5887, "step": 218450 }, { "epoch": 2.4547389943996, "grad_norm": 1.1322340965270996, "learning_rate": 1.018453898956309e-05, "loss": 3.6622, "step": 218500 }, { "epoch": 2.4553007195699434, "grad_norm": 1.1565299034118652, "learning_rate": 1.0182292076260239e-05, "loss": 3.6327, "step": 218550 }, { "epoch": 2.455862444740286, "grad_norm": 1.2363173961639404, "learning_rate": 1.0180045162957388e-05, "loss": 3.6067, "step": 218600 }, { "epoch": 2.4564241699106297, "grad_norm": 1.071761965751648, "learning_rate": 1.0177798249654537e-05, "loss": 3.6748, "step": 218650 }, { "epoch": 2.456985895080973, "grad_norm": 1.1876356601715088, "learning_rate": 1.0175551336351688e-05, "loss": 3.7111, "step": 218700 }, { "epoch": 2.457547620251316, "grad_norm": 1.0165085792541504, "learning_rate": 1.0173304423048838e-05, "loss": 3.7227, "step": 218750 }, { "epoch": 2.458109345421659, "grad_norm": 1.1822094917297363, "learning_rate": 1.0171102448012045e-05, "loss": 3.6007, "step": 218800 }, { "epoch": 2.4586710705920023, "grad_norm": 1.026947021484375, "learning_rate": 1.0168900472975251e-05, "loss": 3.6632, "step": 218850 }, { "epoch": 2.4592327957623454, "grad_norm": 1.3460332155227661, "learning_rate": 1.0166653559672402e-05, "loss": 3.6523, "step": 218900 }, { "epoch": 2.4597945209326886, "grad_norm": 1.2409924268722534, "learning_rate": 1.0164406646369551e-05, "loss": 3.6089, "step": 218950 }, { "epoch": 2.4603562461030317, "grad_norm": 1.1835811138153076, "learning_rate": 1.01621597330667e-05, "loss": 3.6637, "step": 219000 }, { "epoch": 2.460917971273375, "grad_norm": 1.5220344066619873, "learning_rate": 1.015991281976385e-05, "loss": 3.6492, "step": 219050 }, { "epoch": 2.461479696443718, "grad_norm": 1.036289930343628, "learning_rate": 1.0157665906461001e-05, "loss": 3.6771, "step": 219100 }, { "epoch": 2.462041421614061, "grad_norm": 1.3220967054367065, "learning_rate": 1.015541899315815e-05, "loss": 3.6313, "step": 219150 }, { "epoch": 2.4626031467844043, "grad_norm": 1.1639056205749512, "learning_rate": 1.01531720798553e-05, "loss": 3.6877, "step": 219200 }, { "epoch": 2.4631648719547474, "grad_norm": 1.1655687093734741, "learning_rate": 1.015092516655245e-05, "loss": 3.6249, "step": 219250 }, { "epoch": 2.4637265971250906, "grad_norm": 1.2269138097763062, "learning_rate": 1.01486782532496e-05, "loss": 3.6885, "step": 219300 }, { "epoch": 2.4642883222954337, "grad_norm": 1.293755292892456, "learning_rate": 1.014643133994675e-05, "loss": 3.5937, "step": 219350 }, { "epoch": 2.464850047465777, "grad_norm": 1.0094069242477417, "learning_rate": 1.0144184426643899e-05, "loss": 3.6618, "step": 219400 }, { "epoch": 2.46541177263612, "grad_norm": 1.3151395320892334, "learning_rate": 1.014193751334105e-05, "loss": 3.6901, "step": 219450 }, { "epoch": 2.465973497806463, "grad_norm": 1.2603576183319092, "learning_rate": 1.0139690600038199e-05, "loss": 3.637, "step": 219500 }, { "epoch": 2.4665352229768063, "grad_norm": 1.1873706579208374, "learning_rate": 1.0137443686735348e-05, "loss": 3.6228, "step": 219550 }, { "epoch": 2.4670969481471494, "grad_norm": 1.1385982036590576, "learning_rate": 1.01351967734325e-05, "loss": 3.6278, "step": 219600 }, { "epoch": 2.4676586733174926, "grad_norm": 1.1630982160568237, "learning_rate": 1.0132949860129649e-05, "loss": 3.6334, "step": 219650 }, { "epoch": 2.4682203984878357, "grad_norm": 1.1185778379440308, "learning_rate": 1.0130702946826798e-05, "loss": 3.6418, "step": 219700 }, { "epoch": 2.468782123658179, "grad_norm": 1.111441969871521, "learning_rate": 1.0128456033523947e-05, "loss": 3.668, "step": 219750 }, { "epoch": 2.469343848828522, "grad_norm": 1.442638635635376, "learning_rate": 1.0126209120221098e-05, "loss": 3.6877, "step": 219800 }, { "epoch": 2.469905573998865, "grad_norm": 1.118112325668335, "learning_rate": 1.0123962206918248e-05, "loss": 3.5624, "step": 219850 }, { "epoch": 2.4704672991692087, "grad_norm": 1.1270848512649536, "learning_rate": 1.0121715293615397e-05, "loss": 3.6369, "step": 219900 }, { "epoch": 2.4710290243395514, "grad_norm": 1.3634648323059082, "learning_rate": 1.0119468380312548e-05, "loss": 3.6662, "step": 219950 }, { "epoch": 2.471590749509895, "grad_norm": 1.1991126537322998, "learning_rate": 1.0117221467009697e-05, "loss": 3.6581, "step": 220000 }, { "epoch": 2.472152474680238, "grad_norm": 1.2159489393234253, "learning_rate": 1.0114974553706847e-05, "loss": 3.5836, "step": 220050 }, { "epoch": 2.4727141998505813, "grad_norm": 1.057368516921997, "learning_rate": 1.0112727640403996e-05, "loss": 3.6606, "step": 220100 }, { "epoch": 2.4732759250209244, "grad_norm": 1.4949387311935425, "learning_rate": 1.0110480727101147e-05, "loss": 3.7073, "step": 220150 }, { "epoch": 2.4738376501912676, "grad_norm": 1.039035677909851, "learning_rate": 1.0108233813798296e-05, "loss": 3.6282, "step": 220200 }, { "epoch": 2.4743993753616107, "grad_norm": 1.3212014436721802, "learning_rate": 1.0105986900495445e-05, "loss": 3.6479, "step": 220250 }, { "epoch": 2.474961100531954, "grad_norm": 1.0305522680282593, "learning_rate": 1.0103739987192596e-05, "loss": 3.6493, "step": 220300 }, { "epoch": 2.475522825702297, "grad_norm": 1.081343412399292, "learning_rate": 1.0101493073889746e-05, "loss": 3.6103, "step": 220350 }, { "epoch": 2.47608455087264, "grad_norm": 1.2654993534088135, "learning_rate": 1.0099246160586895e-05, "loss": 3.6338, "step": 220400 }, { "epoch": 2.4766462760429833, "grad_norm": 2.206638813018799, "learning_rate": 1.0096999247284044e-05, "loss": 3.657, "step": 220450 }, { "epoch": 2.4772080012133264, "grad_norm": 1.1787418127059937, "learning_rate": 1.0094752333981195e-05, "loss": 3.601, "step": 220500 }, { "epoch": 2.4777697263836695, "grad_norm": 1.1602827310562134, "learning_rate": 1.0092505420678345e-05, "loss": 3.7573, "step": 220550 }, { "epoch": 2.4783314515540127, "grad_norm": 1.0930272340774536, "learning_rate": 1.0090258507375494e-05, "loss": 3.5823, "step": 220600 }, { "epoch": 2.478893176724356, "grad_norm": 1.2273497581481934, "learning_rate": 1.0088011594072643e-05, "loss": 3.593, "step": 220650 }, { "epoch": 2.479454901894699, "grad_norm": 1.2213597297668457, "learning_rate": 1.0085764680769794e-05, "loss": 3.7434, "step": 220700 }, { "epoch": 2.480016627065042, "grad_norm": 1.196616530418396, "learning_rate": 1.0083517767466944e-05, "loss": 3.5832, "step": 220750 }, { "epoch": 2.4805783522353853, "grad_norm": 1.21841299533844, "learning_rate": 1.0081270854164093e-05, "loss": 3.7306, "step": 220800 }, { "epoch": 2.4811400774057284, "grad_norm": 1.2195571660995483, "learning_rate": 1.0079023940861244e-05, "loss": 3.6752, "step": 220850 }, { "epoch": 2.4817018025760715, "grad_norm": 1.2555859088897705, "learning_rate": 1.0076777027558393e-05, "loss": 3.6822, "step": 220900 }, { "epoch": 2.4822635277464147, "grad_norm": 1.1204763650894165, "learning_rate": 1.0074530114255543e-05, "loss": 3.7223, "step": 220950 }, { "epoch": 2.482825252916758, "grad_norm": 1.1556209325790405, "learning_rate": 1.0072283200952692e-05, "loss": 3.6588, "step": 221000 }, { "epoch": 2.483386978087101, "grad_norm": 1.1292698383331299, "learning_rate": 1.0070036287649843e-05, "loss": 3.6401, "step": 221050 }, { "epoch": 2.483948703257444, "grad_norm": 1.0915005207061768, "learning_rate": 1.0067789374346992e-05, "loss": 3.5738, "step": 221100 }, { "epoch": 2.4845104284277872, "grad_norm": 1.2218918800354004, "learning_rate": 1.0065542461044142e-05, "loss": 3.6765, "step": 221150 }, { "epoch": 2.4850721535981304, "grad_norm": 1.2162284851074219, "learning_rate": 1.0063295547741293e-05, "loss": 3.5944, "step": 221200 }, { "epoch": 2.485633878768474, "grad_norm": 1.1479686498641968, "learning_rate": 1.0061048634438442e-05, "loss": 3.6261, "step": 221250 }, { "epoch": 2.4861956039388167, "grad_norm": 1.1043280363082886, "learning_rate": 1.0058801721135591e-05, "loss": 3.6859, "step": 221300 }, { "epoch": 2.4867573291091603, "grad_norm": 1.1242334842681885, "learning_rate": 1.005655480783274e-05, "loss": 3.6959, "step": 221350 }, { "epoch": 2.4873190542795034, "grad_norm": 1.1252726316452026, "learning_rate": 1.0054307894529892e-05, "loss": 3.6519, "step": 221400 }, { "epoch": 2.4878807794498465, "grad_norm": 1.095293641090393, "learning_rate": 1.0052060981227041e-05, "loss": 3.6378, "step": 221450 }, { "epoch": 2.4884425046201897, "grad_norm": 1.1695057153701782, "learning_rate": 1.004981406792419e-05, "loss": 3.6797, "step": 221500 }, { "epoch": 2.489004229790533, "grad_norm": 1.2546207904815674, "learning_rate": 1.0047567154621341e-05, "loss": 3.6618, "step": 221550 }, { "epoch": 2.489565954960876, "grad_norm": 1.0123082399368286, "learning_rate": 1.004532024131849e-05, "loss": 3.6444, "step": 221600 }, { "epoch": 2.490127680131219, "grad_norm": 1.2161242961883545, "learning_rate": 1.004307332801564e-05, "loss": 3.5852, "step": 221650 }, { "epoch": 2.4906894053015622, "grad_norm": 0.9452434182167053, "learning_rate": 1.004082641471279e-05, "loss": 3.5919, "step": 221700 }, { "epoch": 2.4912511304719054, "grad_norm": 1.1768102645874023, "learning_rate": 1.003857950140994e-05, "loss": 3.6244, "step": 221750 }, { "epoch": 2.4918128556422485, "grad_norm": 1.2206100225448608, "learning_rate": 1.003633258810709e-05, "loss": 3.6372, "step": 221800 }, { "epoch": 2.4923745808125917, "grad_norm": 1.3399410247802734, "learning_rate": 1.0034085674804239e-05, "loss": 3.6186, "step": 221850 }, { "epoch": 2.492936305982935, "grad_norm": 1.1862549781799316, "learning_rate": 1.003183876150139e-05, "loss": 3.5653, "step": 221900 }, { "epoch": 2.493498031153278, "grad_norm": 1.1225323677062988, "learning_rate": 1.002959184819854e-05, "loss": 3.6313, "step": 221950 }, { "epoch": 2.494059756323621, "grad_norm": 1.3176363706588745, "learning_rate": 1.0027344934895688e-05, "loss": 3.6931, "step": 222000 }, { "epoch": 2.4946214814939642, "grad_norm": 1.2463792562484741, "learning_rate": 1.0025098021592838e-05, "loss": 3.6632, "step": 222050 }, { "epoch": 2.4951832066643074, "grad_norm": 1.0954043865203857, "learning_rate": 1.0022851108289989e-05, "loss": 3.6704, "step": 222100 }, { "epoch": 2.4957449318346505, "grad_norm": 1.0819978713989258, "learning_rate": 1.0020604194987138e-05, "loss": 3.6596, "step": 222150 }, { "epoch": 2.4963066570049937, "grad_norm": 1.0964871644973755, "learning_rate": 1.0018357281684287e-05, "loss": 3.6577, "step": 222200 }, { "epoch": 2.496868382175337, "grad_norm": 1.0642293691635132, "learning_rate": 1.0016110368381437e-05, "loss": 3.6343, "step": 222250 }, { "epoch": 2.49743010734568, "grad_norm": 1.3439793586730957, "learning_rate": 1.0013863455078588e-05, "loss": 3.6691, "step": 222300 }, { "epoch": 2.497991832516023, "grad_norm": 1.1293972730636597, "learning_rate": 1.0011616541775737e-05, "loss": 3.6225, "step": 222350 }, { "epoch": 2.4985535576863662, "grad_norm": 2.4220151901245117, "learning_rate": 1.0009369628472886e-05, "loss": 3.6662, "step": 222400 }, { "epoch": 2.4991152828567094, "grad_norm": 1.14895498752594, "learning_rate": 1.0007122715170037e-05, "loss": 3.7285, "step": 222450 }, { "epoch": 2.4996770080270525, "grad_norm": 1.3529636859893799, "learning_rate": 1.0004875801867187e-05, "loss": 3.6247, "step": 222500 }, { "epoch": 2.5002387331973956, "grad_norm": 1.074212670326233, "learning_rate": 1.0002628888564336e-05, "loss": 3.6574, "step": 222550 }, { "epoch": 2.5008004583677392, "grad_norm": 1.0517715215682983, "learning_rate": 1.0000381975261485e-05, "loss": 3.6177, "step": 222600 }, { "epoch": 2.501362183538082, "grad_norm": 1.2327373027801514, "learning_rate": 9.998135061958635e-06, "loss": 3.6466, "step": 222650 }, { "epoch": 2.5019239087084255, "grad_norm": 1.2518898248672485, "learning_rate": 9.995888148655784e-06, "loss": 3.5565, "step": 222700 }, { "epoch": 2.502485633878768, "grad_norm": 1.2579315900802612, "learning_rate": 9.993641235352935e-06, "loss": 3.6041, "step": 222750 }, { "epoch": 2.503047359049112, "grad_norm": 1.1023578643798828, "learning_rate": 9.991394322050084e-06, "loss": 3.5945, "step": 222800 }, { "epoch": 2.503609084219455, "grad_norm": 1.2034040689468384, "learning_rate": 9.989147408747234e-06, "loss": 3.7039, "step": 222850 }, { "epoch": 2.504170809389798, "grad_norm": 1.1180808544158936, "learning_rate": 9.986900495444383e-06, "loss": 3.6345, "step": 222900 }, { "epoch": 2.504732534560141, "grad_norm": 1.0925536155700684, "learning_rate": 9.984653582141534e-06, "loss": 3.6766, "step": 222950 }, { "epoch": 2.5052942597304844, "grad_norm": 1.1931333541870117, "learning_rate": 9.982406668838683e-06, "loss": 3.7071, "step": 223000 }, { "epoch": 2.5058559849008275, "grad_norm": 1.371894359588623, "learning_rate": 9.980159755535833e-06, "loss": 3.6939, "step": 223050 }, { "epoch": 2.5064177100711706, "grad_norm": 1.0637400150299072, "learning_rate": 9.977912842232984e-06, "loss": 3.6487, "step": 223100 }, { "epoch": 2.506979435241514, "grad_norm": 1.3496999740600586, "learning_rate": 9.975665928930133e-06, "loss": 3.6769, "step": 223150 }, { "epoch": 2.507541160411857, "grad_norm": 1.0121572017669678, "learning_rate": 9.973419015627282e-06, "loss": 3.7185, "step": 223200 }, { "epoch": 2.5081028855822, "grad_norm": 1.0683531761169434, "learning_rate": 9.971172102324432e-06, "loss": 3.6865, "step": 223250 }, { "epoch": 2.508664610752543, "grad_norm": 1.2914478778839111, "learning_rate": 9.968925189021583e-06, "loss": 3.6488, "step": 223300 }, { "epoch": 2.5092263359228864, "grad_norm": 1.1898984909057617, "learning_rate": 9.966678275718732e-06, "loss": 3.6985, "step": 223350 }, { "epoch": 2.5097880610932295, "grad_norm": 1.2759250402450562, "learning_rate": 9.964431362415881e-06, "loss": 3.5878, "step": 223400 }, { "epoch": 2.5103497862635726, "grad_norm": 1.2521064281463623, "learning_rate": 9.962184449113032e-06, "loss": 3.5806, "step": 223450 }, { "epoch": 2.5109115114339158, "grad_norm": 1.2106614112854004, "learning_rate": 9.959937535810182e-06, "loss": 3.7408, "step": 223500 }, { "epoch": 2.511473236604259, "grad_norm": 1.4075654745101929, "learning_rate": 9.957690622507331e-06, "loss": 3.6301, "step": 223550 }, { "epoch": 2.512034961774602, "grad_norm": 1.0811665058135986, "learning_rate": 9.95544370920448e-06, "loss": 3.6178, "step": 223600 }, { "epoch": 2.512596686944945, "grad_norm": 1.3429867029190063, "learning_rate": 9.953196795901631e-06, "loss": 3.658, "step": 223650 }, { "epoch": 2.5131584121152883, "grad_norm": 1.1472651958465576, "learning_rate": 9.95094988259878e-06, "loss": 3.6108, "step": 223700 }, { "epoch": 2.5137201372856315, "grad_norm": 1.316075325012207, "learning_rate": 9.94870296929593e-06, "loss": 3.6201, "step": 223750 }, { "epoch": 2.5142818624559746, "grad_norm": 1.3413081169128418, "learning_rate": 9.94645605599308e-06, "loss": 3.5898, "step": 223800 }, { "epoch": 2.514843587626318, "grad_norm": 1.2931952476501465, "learning_rate": 9.94420914269023e-06, "loss": 3.6628, "step": 223850 }, { "epoch": 2.515405312796661, "grad_norm": 1.4174492359161377, "learning_rate": 9.94196222938738e-06, "loss": 3.6464, "step": 223900 }, { "epoch": 2.5159670379670045, "grad_norm": 1.0922467708587646, "learning_rate": 9.939715316084529e-06, "loss": 3.5746, "step": 223950 }, { "epoch": 2.516528763137347, "grad_norm": 1.1403781175613403, "learning_rate": 9.93746840278168e-06, "loss": 3.6885, "step": 224000 }, { "epoch": 2.5170904883076908, "grad_norm": 1.3727272748947144, "learning_rate": 9.935221489478829e-06, "loss": 3.6833, "step": 224050 }, { "epoch": 2.5176522134780335, "grad_norm": 1.3506124019622803, "learning_rate": 9.932974576175978e-06, "loss": 3.6645, "step": 224100 }, { "epoch": 2.518213938648377, "grad_norm": 1.1761014461517334, "learning_rate": 9.93072766287313e-06, "loss": 3.6475, "step": 224150 }, { "epoch": 2.51877566381872, "grad_norm": 1.3753101825714111, "learning_rate": 9.928480749570279e-06, "loss": 3.594, "step": 224200 }, { "epoch": 2.5193373889890633, "grad_norm": 1.7304960489273071, "learning_rate": 9.926233836267428e-06, "loss": 3.645, "step": 224250 }, { "epoch": 2.5198991141594065, "grad_norm": 1.1977565288543701, "learning_rate": 9.923986922964577e-06, "loss": 3.6815, "step": 224300 }, { "epoch": 2.5204608393297496, "grad_norm": 1.3336644172668457, "learning_rate": 9.921740009661728e-06, "loss": 3.6404, "step": 224350 }, { "epoch": 2.5210225645000928, "grad_norm": 1.0081298351287842, "learning_rate": 9.919493096358878e-06, "loss": 3.5817, "step": 224400 }, { "epoch": 2.521584289670436, "grad_norm": 1.0654035806655884, "learning_rate": 9.917246183056027e-06, "loss": 3.6469, "step": 224450 }, { "epoch": 2.522146014840779, "grad_norm": 1.0444693565368652, "learning_rate": 9.914999269753176e-06, "loss": 3.5891, "step": 224500 }, { "epoch": 2.522707740011122, "grad_norm": 1.1029542684555054, "learning_rate": 9.912752356450327e-06, "loss": 3.5915, "step": 224550 }, { "epoch": 2.5232694651814653, "grad_norm": 1.1915572881698608, "learning_rate": 9.910505443147477e-06, "loss": 3.6211, "step": 224600 }, { "epoch": 2.5238311903518085, "grad_norm": 2.208009958267212, "learning_rate": 9.908258529844626e-06, "loss": 3.6802, "step": 224650 }, { "epoch": 2.5243929155221516, "grad_norm": 1.3922756910324097, "learning_rate": 9.906011616541777e-06, "loss": 3.595, "step": 224700 }, { "epoch": 2.5249546406924948, "grad_norm": 1.2410928010940552, "learning_rate": 9.903764703238926e-06, "loss": 3.7118, "step": 224750 }, { "epoch": 2.525516365862838, "grad_norm": 1.3258376121520996, "learning_rate": 9.901517789936076e-06, "loss": 3.6872, "step": 224800 }, { "epoch": 2.526078091033181, "grad_norm": 1.3248200416564941, "learning_rate": 9.899315814899283e-06, "loss": 3.687, "step": 224850 }, { "epoch": 2.526639816203524, "grad_norm": 1.1712646484375, "learning_rate": 9.897068901596433e-06, "loss": 3.6254, "step": 224900 }, { "epoch": 2.5272015413738673, "grad_norm": 1.007655143737793, "learning_rate": 9.894821988293582e-06, "loss": 3.6162, "step": 224950 }, { "epoch": 2.5277632665442105, "grad_norm": 1.2282979488372803, "learning_rate": 9.892575074990733e-06, "loss": 3.7098, "step": 225000 }, { "epoch": 2.5283249917145536, "grad_norm": 1.0925374031066895, "learning_rate": 9.890328161687882e-06, "loss": 3.606, "step": 225050 }, { "epoch": 2.5288867168848967, "grad_norm": 1.2417986392974854, "learning_rate": 9.888081248385032e-06, "loss": 3.5919, "step": 225100 }, { "epoch": 2.52944844205524, "grad_norm": 1.144342303276062, "learning_rate": 9.885834335082183e-06, "loss": 3.6636, "step": 225150 }, { "epoch": 2.5300101672255835, "grad_norm": 1.090048909187317, "learning_rate": 9.883587421779332e-06, "loss": 3.601, "step": 225200 }, { "epoch": 2.530571892395926, "grad_norm": 1.1404286623001099, "learning_rate": 9.881340508476481e-06, "loss": 3.5664, "step": 225250 }, { "epoch": 2.5311336175662698, "grad_norm": 1.2363505363464355, "learning_rate": 9.87909359517363e-06, "loss": 3.6045, "step": 225300 }, { "epoch": 2.5316953427366125, "grad_norm": 1.2350670099258423, "learning_rate": 9.876846681870781e-06, "loss": 3.6679, "step": 225350 }, { "epoch": 2.532257067906956, "grad_norm": 1.3042763471603394, "learning_rate": 9.87459976856793e-06, "loss": 3.6355, "step": 225400 }, { "epoch": 2.5328187930772987, "grad_norm": 1.2034074068069458, "learning_rate": 9.87235285526508e-06, "loss": 3.6217, "step": 225450 }, { "epoch": 2.5333805182476423, "grad_norm": 1.1341825723648071, "learning_rate": 9.87010594196223e-06, "loss": 3.5584, "step": 225500 }, { "epoch": 2.5339422434179855, "grad_norm": 1.2538197040557861, "learning_rate": 9.86785902865938e-06, "loss": 3.6495, "step": 225550 }, { "epoch": 2.5345039685883286, "grad_norm": 1.225101113319397, "learning_rate": 9.86561211535653e-06, "loss": 3.6399, "step": 225600 }, { "epoch": 2.5350656937586717, "grad_norm": 1.322907567024231, "learning_rate": 9.863365202053679e-06, "loss": 3.6577, "step": 225650 }, { "epoch": 2.535627418929015, "grad_norm": 1.14747154712677, "learning_rate": 9.86111828875083e-06, "loss": 3.6627, "step": 225700 }, { "epoch": 2.536189144099358, "grad_norm": 1.3067673444747925, "learning_rate": 9.85887137544798e-06, "loss": 3.6451, "step": 225750 }, { "epoch": 2.536750869269701, "grad_norm": 1.1978440284729004, "learning_rate": 9.856624462145129e-06, "loss": 3.6941, "step": 225800 }, { "epoch": 2.5373125944400443, "grad_norm": 1.2274320125579834, "learning_rate": 9.854377548842278e-06, "loss": 3.5767, "step": 225850 }, { "epoch": 2.5378743196103875, "grad_norm": 1.400112271308899, "learning_rate": 9.852130635539429e-06, "loss": 3.6707, "step": 225900 }, { "epoch": 2.5384360447807306, "grad_norm": 1.0723315477371216, "learning_rate": 9.849883722236578e-06, "loss": 3.6521, "step": 225950 }, { "epoch": 2.5389977699510737, "grad_norm": 1.1016207933425903, "learning_rate": 9.847636808933728e-06, "loss": 3.6211, "step": 226000 }, { "epoch": 2.539559495121417, "grad_norm": 1.2424542903900146, "learning_rate": 9.845389895630879e-06, "loss": 3.6466, "step": 226050 }, { "epoch": 2.54012122029176, "grad_norm": 1.113968014717102, "learning_rate": 9.843142982328028e-06, "loss": 3.6645, "step": 226100 }, { "epoch": 2.540682945462103, "grad_norm": 1.185343623161316, "learning_rate": 9.840896069025177e-06, "loss": 3.6477, "step": 226150 }, { "epoch": 2.5412446706324463, "grad_norm": 1.28215491771698, "learning_rate": 9.838649155722327e-06, "loss": 3.7339, "step": 226200 }, { "epoch": 2.5418063958027894, "grad_norm": 1.1373237371444702, "learning_rate": 9.836402242419478e-06, "loss": 3.5953, "step": 226250 }, { "epoch": 2.5423681209731326, "grad_norm": 1.2297579050064087, "learning_rate": 9.834155329116627e-06, "loss": 3.6917, "step": 226300 }, { "epoch": 2.5429298461434757, "grad_norm": 1.1348350048065186, "learning_rate": 9.831908415813776e-06, "loss": 3.5935, "step": 226350 }, { "epoch": 2.543491571313819, "grad_norm": 1.2301366329193115, "learning_rate": 9.829661502510927e-06, "loss": 3.5674, "step": 226400 }, { "epoch": 2.544053296484162, "grad_norm": 1.1980699300765991, "learning_rate": 9.827414589208077e-06, "loss": 3.6758, "step": 226450 }, { "epoch": 2.544615021654505, "grad_norm": 1.1047879457473755, "learning_rate": 9.825167675905226e-06, "loss": 3.7047, "step": 226500 }, { "epoch": 2.5451767468248487, "grad_norm": 1.1126362085342407, "learning_rate": 9.822920762602375e-06, "loss": 3.5761, "step": 226550 }, { "epoch": 2.5457384719951914, "grad_norm": 1.0520906448364258, "learning_rate": 9.820673849299526e-06, "loss": 3.5795, "step": 226600 }, { "epoch": 2.546300197165535, "grad_norm": 1.475176215171814, "learning_rate": 9.818426935996676e-06, "loss": 3.6411, "step": 226650 }, { "epoch": 2.5468619223358777, "grad_norm": 1.1460368633270264, "learning_rate": 9.816180022693825e-06, "loss": 3.6371, "step": 226700 }, { "epoch": 2.5474236475062213, "grad_norm": 1.2125566005706787, "learning_rate": 9.813933109390976e-06, "loss": 3.6441, "step": 226750 }, { "epoch": 2.547985372676564, "grad_norm": 1.3701915740966797, "learning_rate": 9.811686196088125e-06, "loss": 3.7155, "step": 226800 }, { "epoch": 2.5485470978469076, "grad_norm": 1.1672906875610352, "learning_rate": 9.809439282785275e-06, "loss": 3.6298, "step": 226850 }, { "epoch": 2.5491088230172507, "grad_norm": 1.2414804697036743, "learning_rate": 9.807237307748482e-06, "loss": 3.6688, "step": 226900 }, { "epoch": 2.549670548187594, "grad_norm": 1.0830920934677124, "learning_rate": 9.804990394445631e-06, "loss": 3.5695, "step": 226950 }, { "epoch": 2.550232273357937, "grad_norm": 1.1367555856704712, "learning_rate": 9.80274348114278e-06, "loss": 3.6027, "step": 227000 }, { "epoch": 2.55079399852828, "grad_norm": 1.1189628839492798, "learning_rate": 9.800541506105987e-06, "loss": 3.6176, "step": 227050 }, { "epoch": 2.5513557236986233, "grad_norm": 1.156973958015442, "learning_rate": 9.798294592803138e-06, "loss": 3.6535, "step": 227100 }, { "epoch": 2.5519174488689664, "grad_norm": 1.1597468852996826, "learning_rate": 9.796047679500287e-06, "loss": 3.6358, "step": 227150 }, { "epoch": 2.5524791740393096, "grad_norm": 1.1040409803390503, "learning_rate": 9.793800766197436e-06, "loss": 3.7609, "step": 227200 }, { "epoch": 2.5530408992096527, "grad_norm": 1.3094677925109863, "learning_rate": 9.791553852894586e-06, "loss": 3.71, "step": 227250 }, { "epoch": 2.553602624379996, "grad_norm": 1.2151713371276855, "learning_rate": 9.789306939591737e-06, "loss": 3.6331, "step": 227300 }, { "epoch": 2.554164349550339, "grad_norm": 1.1243503093719482, "learning_rate": 9.787060026288886e-06, "loss": 3.597, "step": 227350 }, { "epoch": 2.554726074720682, "grad_norm": 1.1748740673065186, "learning_rate": 9.784813112986035e-06, "loss": 3.6153, "step": 227400 }, { "epoch": 2.5552877998910253, "grad_norm": 1.1404298543930054, "learning_rate": 9.782566199683186e-06, "loss": 3.6226, "step": 227450 }, { "epoch": 2.5558495250613684, "grad_norm": 1.275992751121521, "learning_rate": 9.780319286380336e-06, "loss": 3.7076, "step": 227500 }, { "epoch": 2.5564112502317116, "grad_norm": 1.2919789552688599, "learning_rate": 9.778072373077485e-06, "loss": 3.5933, "step": 227550 }, { "epoch": 2.5569729754020547, "grad_norm": 1.1128894090652466, "learning_rate": 9.775825459774634e-06, "loss": 3.6055, "step": 227600 }, { "epoch": 2.557534700572398, "grad_norm": 1.2390849590301514, "learning_rate": 9.773578546471785e-06, "loss": 3.6385, "step": 227650 }, { "epoch": 2.558096425742741, "grad_norm": 1.1185600757598877, "learning_rate": 9.771331633168935e-06, "loss": 3.6285, "step": 227700 }, { "epoch": 2.558658150913084, "grad_norm": 1.2565029859542847, "learning_rate": 9.769084719866084e-06, "loss": 3.6374, "step": 227750 }, { "epoch": 2.5592198760834273, "grad_norm": 1.1560763120651245, "learning_rate": 9.766837806563235e-06, "loss": 3.686, "step": 227800 }, { "epoch": 2.5597816012537704, "grad_norm": 1.5770220756530762, "learning_rate": 9.764590893260384e-06, "loss": 3.663, "step": 227850 }, { "epoch": 2.560343326424114, "grad_norm": 1.2783591747283936, "learning_rate": 9.762343979957534e-06, "loss": 3.6719, "step": 227900 }, { "epoch": 2.5609050515944567, "grad_norm": 1.4159845113754272, "learning_rate": 9.760097066654683e-06, "loss": 3.6422, "step": 227950 }, { "epoch": 2.5614667767648003, "grad_norm": 1.1151092052459717, "learning_rate": 9.757850153351834e-06, "loss": 3.5841, "step": 228000 }, { "epoch": 2.562028501935143, "grad_norm": 1.0429368019104004, "learning_rate": 9.755603240048983e-06, "loss": 3.5358, "step": 228050 }, { "epoch": 2.5625902271054866, "grad_norm": 1.4416569471359253, "learning_rate": 9.753356326746133e-06, "loss": 3.6824, "step": 228100 }, { "epoch": 2.5631519522758293, "grad_norm": 1.1972196102142334, "learning_rate": 9.751109413443282e-06, "loss": 3.6269, "step": 228150 }, { "epoch": 2.563713677446173, "grad_norm": 1.1545695066452026, "learning_rate": 9.748862500140433e-06, "loss": 3.626, "step": 228200 }, { "epoch": 2.564275402616516, "grad_norm": 1.23501455783844, "learning_rate": 9.746615586837582e-06, "loss": 3.6907, "step": 228250 }, { "epoch": 2.564837127786859, "grad_norm": 1.1475234031677246, "learning_rate": 9.744368673534731e-06, "loss": 3.7481, "step": 228300 }, { "epoch": 2.5653988529572023, "grad_norm": 1.705857276916504, "learning_rate": 9.742121760231882e-06, "loss": 3.605, "step": 228350 }, { "epoch": 2.5659605781275454, "grad_norm": 1.0402103662490845, "learning_rate": 9.739874846929032e-06, "loss": 3.6349, "step": 228400 }, { "epoch": 2.5665223032978886, "grad_norm": 1.1646581888198853, "learning_rate": 9.737627933626181e-06, "loss": 3.6811, "step": 228450 }, { "epoch": 2.5670840284682317, "grad_norm": 1.327896237373352, "learning_rate": 9.73538102032333e-06, "loss": 3.7385, "step": 228500 }, { "epoch": 2.567645753638575, "grad_norm": 1.330582618713379, "learning_rate": 9.733134107020481e-06, "loss": 3.7422, "step": 228550 }, { "epoch": 2.568207478808918, "grad_norm": 1.0372596979141235, "learning_rate": 9.73088719371763e-06, "loss": 3.6206, "step": 228600 }, { "epoch": 2.568769203979261, "grad_norm": 1.1130669116973877, "learning_rate": 9.72864028041478e-06, "loss": 3.7046, "step": 228650 }, { "epoch": 2.5693309291496043, "grad_norm": 1.196959137916565, "learning_rate": 9.726393367111931e-06, "loss": 3.631, "step": 228700 }, { "epoch": 2.5698926543199474, "grad_norm": 1.3812388181686401, "learning_rate": 9.72414645380908e-06, "loss": 3.6295, "step": 228750 }, { "epoch": 2.5704543794902905, "grad_norm": 1.2576218843460083, "learning_rate": 9.72189954050623e-06, "loss": 3.6193, "step": 228800 }, { "epoch": 2.5710161046606337, "grad_norm": 1.4126763343811035, "learning_rate": 9.719652627203379e-06, "loss": 3.5631, "step": 228850 }, { "epoch": 2.571577829830977, "grad_norm": 1.189589262008667, "learning_rate": 9.71740571390053e-06, "loss": 3.5979, "step": 228900 }, { "epoch": 2.57213955500132, "grad_norm": 1.1494877338409424, "learning_rate": 9.71515880059768e-06, "loss": 3.639, "step": 228950 }, { "epoch": 2.572701280171663, "grad_norm": 1.1925745010375977, "learning_rate": 9.712911887294829e-06, "loss": 3.6837, "step": 229000 }, { "epoch": 2.5732630053420062, "grad_norm": 1.2979105710983276, "learning_rate": 9.71066497399198e-06, "loss": 3.6697, "step": 229050 }, { "epoch": 2.5738247305123494, "grad_norm": 1.1843870878219604, "learning_rate": 9.708418060689129e-06, "loss": 3.6816, "step": 229100 }, { "epoch": 2.5743864556826925, "grad_norm": 0.9702261686325073, "learning_rate": 9.706171147386278e-06, "loss": 3.6774, "step": 229150 }, { "epoch": 2.5749481808530357, "grad_norm": 1.1876819133758545, "learning_rate": 9.703924234083428e-06, "loss": 3.6931, "step": 229200 }, { "epoch": 2.5755099060233793, "grad_norm": 1.1669557094573975, "learning_rate": 9.701722259046635e-06, "loss": 3.6682, "step": 229250 }, { "epoch": 2.576071631193722, "grad_norm": 1.2973331212997437, "learning_rate": 9.699475345743785e-06, "loss": 3.6812, "step": 229300 }, { "epoch": 2.5766333563640655, "grad_norm": 1.2905819416046143, "learning_rate": 9.697228432440936e-06, "loss": 3.7467, "step": 229350 }, { "epoch": 2.5771950815344082, "grad_norm": 1.0218195915222168, "learning_rate": 9.694981519138085e-06, "loss": 3.6171, "step": 229400 }, { "epoch": 2.577756806704752, "grad_norm": 1.2959119081497192, "learning_rate": 9.692734605835234e-06, "loss": 3.6238, "step": 229450 }, { "epoch": 2.5783185318750945, "grad_norm": 1.1232720613479614, "learning_rate": 9.690487692532384e-06, "loss": 3.6235, "step": 229500 }, { "epoch": 2.578880257045438, "grad_norm": 1.2635669708251953, "learning_rate": 9.688240779229535e-06, "loss": 3.6442, "step": 229550 }, { "epoch": 2.5794419822157812, "grad_norm": 1.140675663948059, "learning_rate": 9.685993865926684e-06, "loss": 3.6595, "step": 229600 }, { "epoch": 2.5800037073861244, "grad_norm": 1.2757861614227295, "learning_rate": 9.683746952623833e-06, "loss": 3.6604, "step": 229650 }, { "epoch": 2.5805654325564675, "grad_norm": 1.026766300201416, "learning_rate": 9.681500039320984e-06, "loss": 3.676, "step": 229700 }, { "epoch": 2.5811271577268107, "grad_norm": 1.2279552221298218, "learning_rate": 9.679253126018134e-06, "loss": 3.6094, "step": 229750 }, { "epoch": 2.581688882897154, "grad_norm": 1.172577142715454, "learning_rate": 9.677006212715283e-06, "loss": 3.7079, "step": 229800 }, { "epoch": 2.582250608067497, "grad_norm": 1.140169382095337, "learning_rate": 9.674759299412432e-06, "loss": 3.6405, "step": 229850 }, { "epoch": 2.58281233323784, "grad_norm": 1.3780945539474487, "learning_rate": 9.672512386109583e-06, "loss": 3.6893, "step": 229900 }, { "epoch": 2.5833740584081832, "grad_norm": 1.1179897785186768, "learning_rate": 9.670265472806732e-06, "loss": 3.605, "step": 229950 }, { "epoch": 2.5839357835785264, "grad_norm": 1.0302451848983765, "learning_rate": 9.668018559503882e-06, "loss": 3.5907, "step": 230000 }, { "epoch": 2.5844975087488695, "grad_norm": 1.1761894226074219, "learning_rate": 9.665771646201033e-06, "loss": 3.6571, "step": 230050 }, { "epoch": 2.5850592339192127, "grad_norm": 1.1279114484786987, "learning_rate": 9.663524732898182e-06, "loss": 3.6152, "step": 230100 }, { "epoch": 2.585620959089556, "grad_norm": 1.0731397867202759, "learning_rate": 9.661277819595331e-06, "loss": 3.6051, "step": 230150 }, { "epoch": 2.586182684259899, "grad_norm": 1.0091617107391357, "learning_rate": 9.65903090629248e-06, "loss": 3.6454, "step": 230200 }, { "epoch": 2.586744409430242, "grad_norm": 1.1999599933624268, "learning_rate": 9.656783992989632e-06, "loss": 3.5903, "step": 230250 }, { "epoch": 2.5873061346005852, "grad_norm": 1.1237565279006958, "learning_rate": 9.654537079686781e-06, "loss": 3.6313, "step": 230300 }, { "epoch": 2.5878678597709284, "grad_norm": 1.1436638832092285, "learning_rate": 9.65229016638393e-06, "loss": 3.684, "step": 230350 }, { "epoch": 2.5884295849412715, "grad_norm": 1.419813632965088, "learning_rate": 9.650043253081081e-06, "loss": 3.6183, "step": 230400 }, { "epoch": 2.5889913101116147, "grad_norm": 1.036826252937317, "learning_rate": 9.64779633977823e-06, "loss": 3.6608, "step": 230450 }, { "epoch": 2.589553035281958, "grad_norm": 1.0370091199874878, "learning_rate": 9.64554942647538e-06, "loss": 3.6029, "step": 230500 }, { "epoch": 2.590114760452301, "grad_norm": 1.214907169342041, "learning_rate": 9.64330251317253e-06, "loss": 3.6474, "step": 230550 }, { "epoch": 2.5906764856226445, "grad_norm": 1.1329468488693237, "learning_rate": 9.64105559986968e-06, "loss": 3.6658, "step": 230600 }, { "epoch": 2.591238210792987, "grad_norm": 1.250857949256897, "learning_rate": 9.63880868656683e-06, "loss": 3.7155, "step": 230650 }, { "epoch": 2.591799935963331, "grad_norm": 1.4392398595809937, "learning_rate": 9.636561773263979e-06, "loss": 3.638, "step": 230700 }, { "epoch": 2.5923616611336735, "grad_norm": 1.2033816576004028, "learning_rate": 9.634314859961128e-06, "loss": 3.595, "step": 230750 }, { "epoch": 2.592923386304017, "grad_norm": 1.2748844623565674, "learning_rate": 9.63206794665828e-06, "loss": 3.6759, "step": 230800 }, { "epoch": 2.59348511147436, "grad_norm": 1.1000982522964478, "learning_rate": 9.629821033355429e-06, "loss": 3.6866, "step": 230850 }, { "epoch": 2.5940468366447034, "grad_norm": 1.2321597337722778, "learning_rate": 9.627574120052578e-06, "loss": 3.7767, "step": 230900 }, { "epoch": 2.5946085618150465, "grad_norm": 1.126806378364563, "learning_rate": 9.625327206749729e-06, "loss": 3.7061, "step": 230950 }, { "epoch": 2.5951702869853897, "grad_norm": 1.2558183670043945, "learning_rate": 9.623080293446878e-06, "loss": 3.6879, "step": 231000 }, { "epoch": 2.595732012155733, "grad_norm": 1.1356263160705566, "learning_rate": 9.620833380144028e-06, "loss": 3.6204, "step": 231050 }, { "epoch": 2.596293737326076, "grad_norm": 1.0545427799224854, "learning_rate": 9.618586466841177e-06, "loss": 3.6406, "step": 231100 }, { "epoch": 2.596855462496419, "grad_norm": 1.0609054565429688, "learning_rate": 9.616339553538328e-06, "loss": 3.6456, "step": 231150 }, { "epoch": 2.597417187666762, "grad_norm": 1.2802765369415283, "learning_rate": 9.614092640235477e-06, "loss": 3.6708, "step": 231200 }, { "epoch": 2.5979789128371054, "grad_norm": 1.2823866605758667, "learning_rate": 9.611845726932627e-06, "loss": 3.6163, "step": 231250 }, { "epoch": 2.5985406380074485, "grad_norm": 1.1320018768310547, "learning_rate": 9.609598813629778e-06, "loss": 3.6317, "step": 231300 }, { "epoch": 2.5991023631777916, "grad_norm": 1.2215267419815063, "learning_rate": 9.607351900326927e-06, "loss": 3.7005, "step": 231350 }, { "epoch": 2.599664088348135, "grad_norm": 1.1547940969467163, "learning_rate": 9.605104987024076e-06, "loss": 3.6928, "step": 231400 }, { "epoch": 2.600225813518478, "grad_norm": 1.1811012029647827, "learning_rate": 9.602858073721226e-06, "loss": 3.7021, "step": 231450 }, { "epoch": 2.600787538688821, "grad_norm": 1.4081757068634033, "learning_rate": 9.600611160418377e-06, "loss": 3.6676, "step": 231500 }, { "epoch": 2.601349263859164, "grad_norm": 1.0522280931472778, "learning_rate": 9.598364247115526e-06, "loss": 3.6563, "step": 231550 }, { "epoch": 2.6019109890295073, "grad_norm": 1.44118332862854, "learning_rate": 9.596117333812675e-06, "loss": 3.6493, "step": 231600 }, { "epoch": 2.6024727141998505, "grad_norm": 1.143570899963379, "learning_rate": 9.593870420509826e-06, "loss": 3.6629, "step": 231650 }, { "epoch": 2.6030344393701936, "grad_norm": 1.205647349357605, "learning_rate": 9.591623507206975e-06, "loss": 3.6862, "step": 231700 }, { "epoch": 2.6035961645405368, "grad_norm": 1.1053411960601807, "learning_rate": 9.589376593904125e-06, "loss": 3.7092, "step": 231750 }, { "epoch": 2.60415788971088, "grad_norm": 1.081027626991272, "learning_rate": 9.587129680601274e-06, "loss": 3.6065, "step": 231800 }, { "epoch": 2.604719614881223, "grad_norm": 1.2975106239318848, "learning_rate": 9.584882767298425e-06, "loss": 3.6254, "step": 231850 }, { "epoch": 2.605281340051566, "grad_norm": 1.1592826843261719, "learning_rate": 9.582635853995574e-06, "loss": 3.5702, "step": 231900 }, { "epoch": 2.60584306522191, "grad_norm": 1.1324329376220703, "learning_rate": 9.580388940692724e-06, "loss": 3.6665, "step": 231950 }, { "epoch": 2.6064047903922525, "grad_norm": 1.1697642803192139, "learning_rate": 9.578142027389875e-06, "loss": 3.6271, "step": 232000 }, { "epoch": 2.606966515562596, "grad_norm": 1.2418887615203857, "learning_rate": 9.575895114087024e-06, "loss": 3.6288, "step": 232050 }, { "epoch": 2.6075282407329388, "grad_norm": 1.1859896183013916, "learning_rate": 9.573648200784173e-06, "loss": 3.669, "step": 232100 }, { "epoch": 2.6080899659032823, "grad_norm": 1.1547828912734985, "learning_rate": 9.571401287481323e-06, "loss": 3.7053, "step": 232150 }, { "epoch": 2.608651691073625, "grad_norm": 1.3412789106369019, "learning_rate": 9.569154374178474e-06, "loss": 3.6156, "step": 232200 }, { "epoch": 2.6092134162439686, "grad_norm": 1.2743942737579346, "learning_rate": 9.566907460875623e-06, "loss": 3.6579, "step": 232250 }, { "epoch": 2.6097751414143118, "grad_norm": 1.14712393283844, "learning_rate": 9.564660547572772e-06, "loss": 3.6421, "step": 232300 }, { "epoch": 2.610336866584655, "grad_norm": 1.2016804218292236, "learning_rate": 9.562413634269922e-06, "loss": 3.5883, "step": 232350 }, { "epoch": 2.610898591754998, "grad_norm": 1.2405736446380615, "learning_rate": 9.560166720967073e-06, "loss": 3.6278, "step": 232400 }, { "epoch": 2.611460316925341, "grad_norm": 1.0889028310775757, "learning_rate": 9.557919807664222e-06, "loss": 3.634, "step": 232450 }, { "epoch": 2.6120220420956843, "grad_norm": 1.4650338888168335, "learning_rate": 9.555672894361371e-06, "loss": 3.6748, "step": 232500 }, { "epoch": 2.6125837672660275, "grad_norm": 1.162473440170288, "learning_rate": 9.553425981058522e-06, "loss": 3.6959, "step": 232550 }, { "epoch": 2.6131454924363706, "grad_norm": 1.3220645189285278, "learning_rate": 9.551179067755672e-06, "loss": 3.6216, "step": 232600 }, { "epoch": 2.6137072176067138, "grad_norm": 1.1755160093307495, "learning_rate": 9.548932154452821e-06, "loss": 3.6655, "step": 232650 }, { "epoch": 2.614268942777057, "grad_norm": 1.1696518659591675, "learning_rate": 9.54668524114997e-06, "loss": 3.6851, "step": 232700 }, { "epoch": 2.6148306679474, "grad_norm": 1.214743733406067, "learning_rate": 9.544438327847121e-06, "loss": 3.6141, "step": 232750 }, { "epoch": 2.615392393117743, "grad_norm": 1.0277268886566162, "learning_rate": 9.54219141454427e-06, "loss": 3.6633, "step": 232800 }, { "epoch": 2.6159541182880863, "grad_norm": 1.1339294910430908, "learning_rate": 9.53994450124142e-06, "loss": 3.638, "step": 232850 }, { "epoch": 2.6165158434584295, "grad_norm": 1.1576228141784668, "learning_rate": 9.537697587938571e-06, "loss": 3.5904, "step": 232900 }, { "epoch": 2.6170775686287726, "grad_norm": 1.4079818725585938, "learning_rate": 9.53545067463572e-06, "loss": 3.6654, "step": 232950 }, { "epoch": 2.6176392937991158, "grad_norm": 1.0673800706863403, "learning_rate": 9.53320376133287e-06, "loss": 3.6718, "step": 233000 }, { "epoch": 2.618201018969459, "grad_norm": 1.1846091747283936, "learning_rate": 9.530956848030019e-06, "loss": 3.648, "step": 233050 }, { "epoch": 2.618762744139802, "grad_norm": 1.1678266525268555, "learning_rate": 9.52870993472717e-06, "loss": 3.5739, "step": 233100 }, { "epoch": 2.619324469310145, "grad_norm": 1.2650402784347534, "learning_rate": 9.52646302142432e-06, "loss": 3.6458, "step": 233150 }, { "epoch": 2.6198861944804883, "grad_norm": 1.2277120351791382, "learning_rate": 9.524216108121469e-06, "loss": 3.6615, "step": 233200 }, { "epoch": 2.6204479196508315, "grad_norm": 1.2441250085830688, "learning_rate": 9.52196919481862e-06, "loss": 3.6584, "step": 233250 }, { "epoch": 2.621009644821175, "grad_norm": 1.3057355880737305, "learning_rate": 9.519722281515769e-06, "loss": 3.6591, "step": 233300 }, { "epoch": 2.6215713699915177, "grad_norm": 1.1118351221084595, "learning_rate": 9.517475368212918e-06, "loss": 3.6061, "step": 233350 }, { "epoch": 2.6221330951618613, "grad_norm": 1.1428354978561401, "learning_rate": 9.515228454910068e-06, "loss": 3.6541, "step": 233400 }, { "epoch": 2.622694820332204, "grad_norm": 1.2182830572128296, "learning_rate": 9.512981541607219e-06, "loss": 3.6867, "step": 233450 }, { "epoch": 2.6232565455025476, "grad_norm": 1.158341884613037, "learning_rate": 9.510734628304368e-06, "loss": 3.6546, "step": 233500 }, { "epoch": 2.6238182706728903, "grad_norm": 1.13567054271698, "learning_rate": 9.508487715001517e-06, "loss": 3.6903, "step": 233550 }, { "epoch": 2.624379995843234, "grad_norm": 1.0206648111343384, "learning_rate": 9.506240801698668e-06, "loss": 3.5835, "step": 233600 }, { "epoch": 2.624941721013577, "grad_norm": 1.0175087451934814, "learning_rate": 9.503993888395817e-06, "loss": 3.7287, "step": 233650 }, { "epoch": 2.62550344618392, "grad_norm": 0.9948505163192749, "learning_rate": 9.501746975092967e-06, "loss": 3.6272, "step": 233700 }, { "epoch": 2.6260651713542633, "grad_norm": 1.4096300601959229, "learning_rate": 9.499500061790116e-06, "loss": 3.6122, "step": 233750 }, { "epoch": 2.6266268965246065, "grad_norm": 1.0361156463623047, "learning_rate": 9.497253148487267e-06, "loss": 3.6412, "step": 233800 }, { "epoch": 2.6271886216949496, "grad_norm": 1.1061748266220093, "learning_rate": 9.495006235184416e-06, "loss": 3.5806, "step": 233850 }, { "epoch": 2.6277503468652927, "grad_norm": 1.3449162244796753, "learning_rate": 9.492759321881566e-06, "loss": 3.5348, "step": 233900 }, { "epoch": 2.628312072035636, "grad_norm": 1.7299484014511108, "learning_rate": 9.490512408578717e-06, "loss": 3.6063, "step": 233950 }, { "epoch": 2.628873797205979, "grad_norm": 1.1820323467254639, "learning_rate": 9.488265495275866e-06, "loss": 3.6409, "step": 234000 }, { "epoch": 2.629435522376322, "grad_norm": 1.146406888961792, "learning_rate": 9.486063520239072e-06, "loss": 3.731, "step": 234050 }, { "epoch": 2.6299972475466653, "grad_norm": 1.1521809101104736, "learning_rate": 9.483816606936223e-06, "loss": 3.6015, "step": 234100 }, { "epoch": 2.6305589727170084, "grad_norm": 1.2219223976135254, "learning_rate": 9.481569693633372e-06, "loss": 3.6379, "step": 234150 }, { "epoch": 2.6311206978873516, "grad_norm": 1.103837013244629, "learning_rate": 9.479322780330522e-06, "loss": 3.641, "step": 234200 }, { "epoch": 2.6316824230576947, "grad_norm": 1.1956130266189575, "learning_rate": 9.477075867027673e-06, "loss": 3.5529, "step": 234250 }, { "epoch": 2.632244148228038, "grad_norm": 1.3069959878921509, "learning_rate": 9.474828953724822e-06, "loss": 3.606, "step": 234300 }, { "epoch": 2.632805873398381, "grad_norm": 1.2597901821136475, "learning_rate": 9.472582040421971e-06, "loss": 3.6377, "step": 234350 }, { "epoch": 2.633367598568724, "grad_norm": 1.0955803394317627, "learning_rate": 9.47033512711912e-06, "loss": 3.7162, "step": 234400 }, { "epoch": 2.6339293237390673, "grad_norm": 1.1958327293395996, "learning_rate": 9.468088213816272e-06, "loss": 3.5871, "step": 234450 }, { "epoch": 2.6344910489094104, "grad_norm": 1.2950828075408936, "learning_rate": 9.465841300513421e-06, "loss": 3.6441, "step": 234500 }, { "epoch": 2.6350527740797536, "grad_norm": 0.989854633808136, "learning_rate": 9.46359438721057e-06, "loss": 3.696, "step": 234550 }, { "epoch": 2.6356144992500967, "grad_norm": 1.1626341342926025, "learning_rate": 9.461347473907721e-06, "loss": 3.653, "step": 234600 }, { "epoch": 2.6361762244204403, "grad_norm": 1.0983715057373047, "learning_rate": 9.45910056060487e-06, "loss": 3.6619, "step": 234650 }, { "epoch": 2.636737949590783, "grad_norm": 1.0216072797775269, "learning_rate": 9.45685364730202e-06, "loss": 3.6305, "step": 234700 }, { "epoch": 2.6372996747611266, "grad_norm": 1.0913469791412354, "learning_rate": 9.45460673399917e-06, "loss": 3.6289, "step": 234750 }, { "epoch": 2.6378613999314693, "grad_norm": 1.5100804567337036, "learning_rate": 9.45235982069632e-06, "loss": 3.6328, "step": 234800 }, { "epoch": 2.638423125101813, "grad_norm": 1.1998149156570435, "learning_rate": 9.45011290739347e-06, "loss": 3.7158, "step": 234850 }, { "epoch": 2.6389848502721556, "grad_norm": 1.3535040616989136, "learning_rate": 9.447865994090619e-06, "loss": 3.6558, "step": 234900 }, { "epoch": 2.639546575442499, "grad_norm": 1.1371439695358276, "learning_rate": 9.445619080787768e-06, "loss": 3.6966, "step": 234950 }, { "epoch": 2.6401083006128423, "grad_norm": 1.214625358581543, "learning_rate": 9.44337216748492e-06, "loss": 3.6085, "step": 235000 }, { "epoch": 2.6406700257831854, "grad_norm": 1.1851134300231934, "learning_rate": 9.441125254182069e-06, "loss": 3.6476, "step": 235050 }, { "epoch": 2.6412317509535286, "grad_norm": 1.106297254562378, "learning_rate": 9.438878340879218e-06, "loss": 3.6427, "step": 235100 }, { "epoch": 2.6417934761238717, "grad_norm": 1.156652808189392, "learning_rate": 9.436631427576369e-06, "loss": 3.6887, "step": 235150 }, { "epoch": 2.642355201294215, "grad_norm": 1.2252565622329712, "learning_rate": 9.434384514273518e-06, "loss": 3.6339, "step": 235200 }, { "epoch": 2.642916926464558, "grad_norm": 0.9991495013237, "learning_rate": 9.432137600970667e-06, "loss": 3.591, "step": 235250 }, { "epoch": 2.643478651634901, "grad_norm": 1.096312403678894, "learning_rate": 9.429890687667817e-06, "loss": 3.6028, "step": 235300 }, { "epoch": 2.6440403768052443, "grad_norm": 1.1620893478393555, "learning_rate": 9.427643774364968e-06, "loss": 3.6738, "step": 235350 }, { "epoch": 2.6446021019755874, "grad_norm": 1.4076168537139893, "learning_rate": 9.425396861062117e-06, "loss": 3.6402, "step": 235400 }, { "epoch": 2.6451638271459306, "grad_norm": 1.1057208776474, "learning_rate": 9.423149947759266e-06, "loss": 3.6549, "step": 235450 }, { "epoch": 2.6457255523162737, "grad_norm": 1.0851587057113647, "learning_rate": 9.420903034456417e-06, "loss": 3.6544, "step": 235500 }, { "epoch": 2.646287277486617, "grad_norm": 1.1699048280715942, "learning_rate": 9.418656121153567e-06, "loss": 3.6056, "step": 235550 }, { "epoch": 2.64684900265696, "grad_norm": 1.1354840993881226, "learning_rate": 9.416409207850716e-06, "loss": 3.6982, "step": 235600 }, { "epoch": 2.647410727827303, "grad_norm": 1.257293701171875, "learning_rate": 9.414162294547865e-06, "loss": 3.6762, "step": 235650 }, { "epoch": 2.6479724529976463, "grad_norm": 1.1046998500823975, "learning_rate": 9.411960319511071e-06, "loss": 3.6089, "step": 235700 }, { "epoch": 2.6485341781679894, "grad_norm": 1.2434829473495483, "learning_rate": 9.409713406208222e-06, "loss": 3.6709, "step": 235750 }, { "epoch": 2.6490959033383326, "grad_norm": 1.1442354917526245, "learning_rate": 9.407466492905372e-06, "loss": 3.5999, "step": 235800 }, { "epoch": 2.6496576285086757, "grad_norm": 1.150447964668274, "learning_rate": 9.405219579602521e-06, "loss": 3.6673, "step": 235850 }, { "epoch": 2.650219353679019, "grad_norm": 1.1089730262756348, "learning_rate": 9.402972666299672e-06, "loss": 3.5911, "step": 235900 }, { "epoch": 2.650781078849362, "grad_norm": 1.2682679891586304, "learning_rate": 9.400725752996821e-06, "loss": 3.5348, "step": 235950 }, { "epoch": 2.6513428040197056, "grad_norm": 1.3281710147857666, "learning_rate": 9.39847883969397e-06, "loss": 3.6362, "step": 236000 }, { "epoch": 2.6519045291900483, "grad_norm": 1.3261010646820068, "learning_rate": 9.39623192639112e-06, "loss": 3.6157, "step": 236050 }, { "epoch": 2.652466254360392, "grad_norm": 1.2928086519241333, "learning_rate": 9.393985013088271e-06, "loss": 3.6542, "step": 236100 }, { "epoch": 2.6530279795307345, "grad_norm": 1.4209024906158447, "learning_rate": 9.39173809978542e-06, "loss": 3.7344, "step": 236150 }, { "epoch": 2.653589704701078, "grad_norm": 1.31272554397583, "learning_rate": 9.38949118648257e-06, "loss": 3.7201, "step": 236200 }, { "epoch": 2.6541514298714213, "grad_norm": 1.176992654800415, "learning_rate": 9.38724427317972e-06, "loss": 3.6173, "step": 236250 }, { "epoch": 2.6547131550417644, "grad_norm": 1.160933256149292, "learning_rate": 9.38499735987687e-06, "loss": 3.642, "step": 236300 }, { "epoch": 2.6552748802121076, "grad_norm": 1.5851428508758545, "learning_rate": 9.38275044657402e-06, "loss": 3.7491, "step": 236350 }, { "epoch": 2.6558366053824507, "grad_norm": 1.163303256034851, "learning_rate": 9.380503533271169e-06, "loss": 3.63, "step": 236400 }, { "epoch": 2.656398330552794, "grad_norm": 1.2374213933944702, "learning_rate": 9.37825661996832e-06, "loss": 3.6488, "step": 236450 }, { "epoch": 2.656960055723137, "grad_norm": 1.2023411989212036, "learning_rate": 9.376009706665469e-06, "loss": 3.6569, "step": 236500 }, { "epoch": 2.65752178089348, "grad_norm": 1.4848881959915161, "learning_rate": 9.373762793362618e-06, "loss": 3.6683, "step": 236550 }, { "epoch": 2.6580835060638233, "grad_norm": 2.022050142288208, "learning_rate": 9.371515880059767e-06, "loss": 3.6196, "step": 236600 }, { "epoch": 2.6586452312341664, "grad_norm": 1.1171576976776123, "learning_rate": 9.369268966756918e-06, "loss": 3.5738, "step": 236650 }, { "epoch": 2.6592069564045095, "grad_norm": 1.1062504053115845, "learning_rate": 9.367022053454068e-06, "loss": 3.7431, "step": 236700 }, { "epoch": 2.6597686815748527, "grad_norm": 1.2124685049057007, "learning_rate": 9.364775140151217e-06, "loss": 3.575, "step": 236750 }, { "epoch": 2.660330406745196, "grad_norm": 1.1376625299453735, "learning_rate": 9.362528226848368e-06, "loss": 3.6897, "step": 236800 }, { "epoch": 2.660892131915539, "grad_norm": 1.3555550575256348, "learning_rate": 9.360281313545517e-06, "loss": 3.6666, "step": 236850 }, { "epoch": 2.661453857085882, "grad_norm": 3.010193109512329, "learning_rate": 9.358034400242667e-06, "loss": 3.6706, "step": 236900 }, { "epoch": 2.6620155822562253, "grad_norm": 1.2489869594573975, "learning_rate": 9.355787486939816e-06, "loss": 3.6344, "step": 236950 }, { "epoch": 2.6625773074265684, "grad_norm": 1.212320327758789, "learning_rate": 9.353540573636967e-06, "loss": 3.6851, "step": 237000 }, { "epoch": 2.6631390325969115, "grad_norm": 1.0101131200790405, "learning_rate": 9.351293660334116e-06, "loss": 3.6492, "step": 237050 }, { "epoch": 2.6637007577672547, "grad_norm": 1.0726019144058228, "learning_rate": 9.349046747031266e-06, "loss": 3.6297, "step": 237100 }, { "epoch": 2.664262482937598, "grad_norm": 1.1873642206192017, "learning_rate": 9.346799833728417e-06, "loss": 3.6453, "step": 237150 }, { "epoch": 2.664824208107941, "grad_norm": 1.2106003761291504, "learning_rate": 9.344552920425566e-06, "loss": 3.6412, "step": 237200 }, { "epoch": 2.665385933278284, "grad_norm": 1.0715569257736206, "learning_rate": 9.342306007122715e-06, "loss": 3.6722, "step": 237250 }, { "epoch": 2.6659476584486272, "grad_norm": 1.1939773559570312, "learning_rate": 9.340059093819865e-06, "loss": 3.6001, "step": 237300 }, { "epoch": 2.666509383618971, "grad_norm": 1.160776138305664, "learning_rate": 9.337812180517016e-06, "loss": 3.6837, "step": 237350 }, { "epoch": 2.6670711087893135, "grad_norm": 1.4269455671310425, "learning_rate": 9.335565267214165e-06, "loss": 3.7343, "step": 237400 }, { "epoch": 2.667632833959657, "grad_norm": 1.1893212795257568, "learning_rate": 9.333318353911314e-06, "loss": 3.714, "step": 237450 }, { "epoch": 2.66819455913, "grad_norm": 1.197709321975708, "learning_rate": 9.331071440608465e-06, "loss": 3.6083, "step": 237500 }, { "epoch": 2.6687562843003434, "grad_norm": 1.1480973958969116, "learning_rate": 9.328824527305615e-06, "loss": 3.6273, "step": 237550 }, { "epoch": 2.6693180094706865, "grad_norm": 1.0416712760925293, "learning_rate": 9.326577614002764e-06, "loss": 3.6305, "step": 237600 }, { "epoch": 2.6698797346410297, "grad_norm": 1.234265923500061, "learning_rate": 9.324330700699913e-06, "loss": 3.6808, "step": 237650 }, { "epoch": 2.670441459811373, "grad_norm": 1.2496914863586426, "learning_rate": 9.322083787397064e-06, "loss": 3.5479, "step": 237700 }, { "epoch": 2.671003184981716, "grad_norm": 1.355756163597107, "learning_rate": 9.319836874094214e-06, "loss": 3.5549, "step": 237750 }, { "epoch": 2.671564910152059, "grad_norm": 1.4441754817962646, "learning_rate": 9.317589960791363e-06, "loss": 3.6918, "step": 237800 }, { "epoch": 2.6721266353224022, "grad_norm": 1.2543582916259766, "learning_rate": 9.315343047488514e-06, "loss": 3.6524, "step": 237850 }, { "epoch": 2.6726883604927454, "grad_norm": 1.2836605310440063, "learning_rate": 9.31314107245172e-06, "loss": 3.6425, "step": 237900 }, { "epoch": 2.6732500856630885, "grad_norm": 1.5076261758804321, "learning_rate": 9.31089415914887e-06, "loss": 3.6275, "step": 237950 }, { "epoch": 2.6738118108334317, "grad_norm": 1.1220850944519043, "learning_rate": 9.30864724584602e-06, "loss": 3.6343, "step": 238000 }, { "epoch": 2.674373536003775, "grad_norm": 1.0616127252578735, "learning_rate": 9.30640033254317e-06, "loss": 3.6229, "step": 238050 }, { "epoch": 2.674935261174118, "grad_norm": 1.0698143243789673, "learning_rate": 9.304153419240319e-06, "loss": 3.5899, "step": 238100 }, { "epoch": 2.675496986344461, "grad_norm": 1.200904369354248, "learning_rate": 9.30190650593747e-06, "loss": 3.6681, "step": 238150 }, { "epoch": 2.6760587115148042, "grad_norm": 1.0132920742034912, "learning_rate": 9.299659592634619e-06, "loss": 3.6474, "step": 238200 }, { "epoch": 2.6766204366851474, "grad_norm": 1.3051836490631104, "learning_rate": 9.297412679331768e-06, "loss": 3.609, "step": 238250 }, { "epoch": 2.6771821618554905, "grad_norm": 1.1440683603286743, "learning_rate": 9.295165766028918e-06, "loss": 3.6477, "step": 238300 }, { "epoch": 2.6777438870258337, "grad_norm": 0.9804529547691345, "learning_rate": 9.292918852726069e-06, "loss": 3.6513, "step": 238350 }, { "epoch": 2.678305612196177, "grad_norm": 1.2794501781463623, "learning_rate": 9.290671939423218e-06, "loss": 3.6358, "step": 238400 }, { "epoch": 2.67886733736652, "grad_norm": 1.1855015754699707, "learning_rate": 9.288425026120367e-06, "loss": 3.7059, "step": 238450 }, { "epoch": 2.679429062536863, "grad_norm": 1.282378077507019, "learning_rate": 9.286178112817518e-06, "loss": 3.6186, "step": 238500 }, { "epoch": 2.6799907877072062, "grad_norm": 1.1567116975784302, "learning_rate": 9.283931199514668e-06, "loss": 3.621, "step": 238550 }, { "epoch": 2.6805525128775494, "grad_norm": 1.2169396877288818, "learning_rate": 9.281684286211817e-06, "loss": 3.581, "step": 238600 }, { "epoch": 2.6811142380478925, "grad_norm": 1.1934642791748047, "learning_rate": 9.279437372908966e-06, "loss": 3.6494, "step": 238650 }, { "epoch": 2.681675963218236, "grad_norm": 1.1107144355773926, "learning_rate": 9.277190459606117e-06, "loss": 3.5958, "step": 238700 }, { "epoch": 2.682237688388579, "grad_norm": 1.163450002670288, "learning_rate": 9.274943546303267e-06, "loss": 3.6176, "step": 238750 }, { "epoch": 2.6827994135589224, "grad_norm": 1.4309276342391968, "learning_rate": 9.272696633000416e-06, "loss": 3.6634, "step": 238800 }, { "epoch": 2.683361138729265, "grad_norm": 1.1508939266204834, "learning_rate": 9.270449719697567e-06, "loss": 3.6055, "step": 238850 }, { "epoch": 2.6839228638996087, "grad_norm": 1.1272213459014893, "learning_rate": 9.268202806394716e-06, "loss": 3.608, "step": 238900 }, { "epoch": 2.684484589069952, "grad_norm": 1.311135172843933, "learning_rate": 9.265955893091866e-06, "loss": 3.6426, "step": 238950 }, { "epoch": 2.685046314240295, "grad_norm": 1.0856311321258545, "learning_rate": 9.263708979789015e-06, "loss": 3.5447, "step": 239000 }, { "epoch": 2.685608039410638, "grad_norm": 1.298883318901062, "learning_rate": 9.261462066486166e-06, "loss": 3.6805, "step": 239050 }, { "epoch": 2.6861697645809812, "grad_norm": 1.2146790027618408, "learning_rate": 9.259215153183315e-06, "loss": 3.6835, "step": 239100 }, { "epoch": 2.6867314897513244, "grad_norm": 1.240931510925293, "learning_rate": 9.256968239880465e-06, "loss": 3.5861, "step": 239150 }, { "epoch": 2.6872932149216675, "grad_norm": 1.3513401746749878, "learning_rate": 9.254721326577614e-06, "loss": 3.6177, "step": 239200 }, { "epoch": 2.6878549400920106, "grad_norm": 1.0338592529296875, "learning_rate": 9.252474413274765e-06, "loss": 3.5875, "step": 239250 }, { "epoch": 2.688416665262354, "grad_norm": 1.1467785835266113, "learning_rate": 9.250227499971914e-06, "loss": 3.6338, "step": 239300 }, { "epoch": 2.688978390432697, "grad_norm": 1.279463291168213, "learning_rate": 9.247980586669064e-06, "loss": 3.6839, "step": 239350 }, { "epoch": 2.68954011560304, "grad_norm": 1.1546107530593872, "learning_rate": 9.245733673366215e-06, "loss": 3.6859, "step": 239400 }, { "epoch": 2.690101840773383, "grad_norm": 1.0767256021499634, "learning_rate": 9.243486760063364e-06, "loss": 3.6337, "step": 239450 }, { "epoch": 2.6906635659437264, "grad_norm": 1.2261996269226074, "learning_rate": 9.241239846760513e-06, "loss": 3.6972, "step": 239500 }, { "epoch": 2.6912252911140695, "grad_norm": 1.0836470127105713, "learning_rate": 9.238992933457663e-06, "loss": 3.6677, "step": 239550 }, { "epoch": 2.6917870162844126, "grad_norm": 1.1737556457519531, "learning_rate": 9.236746020154814e-06, "loss": 3.6386, "step": 239600 }, { "epoch": 2.692348741454756, "grad_norm": 1.1945728063583374, "learning_rate": 9.234499106851963e-06, "loss": 3.6118, "step": 239650 }, { "epoch": 2.692910466625099, "grad_norm": 1.3617130517959595, "learning_rate": 9.232252193549112e-06, "loss": 3.5948, "step": 239700 }, { "epoch": 2.693472191795442, "grad_norm": 1.123518943786621, "learning_rate": 9.230005280246263e-06, "loss": 3.6197, "step": 239750 }, { "epoch": 2.694033916965785, "grad_norm": 1.2159085273742676, "learning_rate": 9.227758366943413e-06, "loss": 3.5417, "step": 239800 }, { "epoch": 2.6945956421361283, "grad_norm": 1.347956657409668, "learning_rate": 9.225511453640562e-06, "loss": 3.7454, "step": 239850 }, { "epoch": 2.6951573673064715, "grad_norm": 1.34463632106781, "learning_rate": 9.223264540337711e-06, "loss": 3.6986, "step": 239900 }, { "epoch": 2.695719092476815, "grad_norm": 1.1763185262680054, "learning_rate": 9.221017627034862e-06, "loss": 3.5904, "step": 239950 }, { "epoch": 2.6962808176471578, "grad_norm": 1.2356795072555542, "learning_rate": 9.218770713732011e-06, "loss": 3.6617, "step": 240000 }, { "epoch": 2.6968425428175014, "grad_norm": 2.081676721572876, "learning_rate": 9.21652380042916e-06, "loss": 3.6598, "step": 240050 }, { "epoch": 2.697404267987844, "grad_norm": 1.215590238571167, "learning_rate": 9.214276887126312e-06, "loss": 3.6284, "step": 240100 }, { "epoch": 2.6979659931581876, "grad_norm": 1.2818858623504639, "learning_rate": 9.212029973823461e-06, "loss": 3.7313, "step": 240150 }, { "epoch": 2.6985277183285303, "grad_norm": 1.0892887115478516, "learning_rate": 9.20978306052061e-06, "loss": 3.6084, "step": 240200 }, { "epoch": 2.699089443498874, "grad_norm": 1.2128300666809082, "learning_rate": 9.20753614721776e-06, "loss": 3.6177, "step": 240250 }, { "epoch": 2.699651168669217, "grad_norm": 1.139426589012146, "learning_rate": 9.20528923391491e-06, "loss": 3.5944, "step": 240300 }, { "epoch": 2.70021289383956, "grad_norm": 1.1543854475021362, "learning_rate": 9.20304232061206e-06, "loss": 3.6371, "step": 240350 }, { "epoch": 2.7007746190099033, "grad_norm": 1.1960457563400269, "learning_rate": 9.20079540730921e-06, "loss": 3.6425, "step": 240400 }, { "epoch": 2.7013363441802465, "grad_norm": 1.123557209968567, "learning_rate": 9.19854849400636e-06, "loss": 3.6176, "step": 240450 }, { "epoch": 2.7018980693505896, "grad_norm": 1.2426378726959229, "learning_rate": 9.19630158070351e-06, "loss": 3.578, "step": 240500 }, { "epoch": 2.7024597945209328, "grad_norm": 1.2549535036087036, "learning_rate": 9.194054667400659e-06, "loss": 3.6038, "step": 240550 }, { "epoch": 2.703021519691276, "grad_norm": 1.2187061309814453, "learning_rate": 9.191807754097808e-06, "loss": 3.7156, "step": 240600 }, { "epoch": 2.703583244861619, "grad_norm": 1.3443732261657715, "learning_rate": 9.18956084079496e-06, "loss": 3.72, "step": 240650 }, { "epoch": 2.704144970031962, "grad_norm": 0.9994232058525085, "learning_rate": 9.187313927492109e-06, "loss": 3.6145, "step": 240700 }, { "epoch": 2.7047066952023053, "grad_norm": 1.1979265213012695, "learning_rate": 9.185067014189258e-06, "loss": 3.6338, "step": 240750 }, { "epoch": 2.7052684203726485, "grad_norm": 1.807159662246704, "learning_rate": 9.182820100886409e-06, "loss": 3.6061, "step": 240800 }, { "epoch": 2.7058301455429916, "grad_norm": 1.0491582155227661, "learning_rate": 9.180573187583558e-06, "loss": 3.6531, "step": 240850 }, { "epoch": 2.7063918707133348, "grad_norm": 1.0936228036880493, "learning_rate": 9.178326274280708e-06, "loss": 3.6325, "step": 240900 }, { "epoch": 2.706953595883678, "grad_norm": 1.1391863822937012, "learning_rate": 9.176079360977857e-06, "loss": 3.5716, "step": 240950 }, { "epoch": 2.707515321054021, "grad_norm": 1.4656747579574585, "learning_rate": 9.173832447675008e-06, "loss": 3.592, "step": 241000 }, { "epoch": 2.708077046224364, "grad_norm": 1.1552832126617432, "learning_rate": 9.171585534372157e-06, "loss": 3.5734, "step": 241050 }, { "epoch": 2.7086387713947073, "grad_norm": 1.0767759084701538, "learning_rate": 9.169338621069307e-06, "loss": 3.7697, "step": 241100 }, { "epoch": 2.7092004965650505, "grad_norm": 1.158240795135498, "learning_rate": 9.167091707766456e-06, "loss": 3.6753, "step": 241150 }, { "epoch": 2.7097622217353936, "grad_norm": 1.0932152271270752, "learning_rate": 9.164844794463607e-06, "loss": 3.7098, "step": 241200 }, { "epoch": 2.7103239469057367, "grad_norm": 1.2083481550216675, "learning_rate": 9.162597881160756e-06, "loss": 3.6069, "step": 241250 }, { "epoch": 2.7108856720760803, "grad_norm": 1.191329002380371, "learning_rate": 9.160350967857906e-06, "loss": 3.5987, "step": 241300 }, { "epoch": 2.711447397246423, "grad_norm": 1.1355061531066895, "learning_rate": 9.158104054555057e-06, "loss": 3.5238, "step": 241350 }, { "epoch": 2.7120091224167666, "grad_norm": 1.117626667022705, "learning_rate": 9.155857141252206e-06, "loss": 3.6962, "step": 241400 }, { "epoch": 2.7125708475871093, "grad_norm": 1.2261159420013428, "learning_rate": 9.153610227949355e-06, "loss": 3.6318, "step": 241450 }, { "epoch": 2.713132572757453, "grad_norm": 1.1712418794631958, "learning_rate": 9.151363314646505e-06, "loss": 3.6104, "step": 241500 }, { "epoch": 2.7136942979277956, "grad_norm": 1.111906886100769, "learning_rate": 9.149116401343656e-06, "loss": 3.5495, "step": 241550 }, { "epoch": 2.714256023098139, "grad_norm": 1.1478056907653809, "learning_rate": 9.146869488040805e-06, "loss": 3.6877, "step": 241600 }, { "epoch": 2.7148177482684823, "grad_norm": 1.1401646137237549, "learning_rate": 9.144622574737954e-06, "loss": 3.6927, "step": 241650 }, { "epoch": 2.7153794734388255, "grad_norm": 1.0862573385238647, "learning_rate": 9.142375661435105e-06, "loss": 3.6816, "step": 241700 }, { "epoch": 2.7159411986091686, "grad_norm": 1.3958070278167725, "learning_rate": 9.140128748132255e-06, "loss": 3.6334, "step": 241750 }, { "epoch": 2.7165029237795117, "grad_norm": 1.2459132671356201, "learning_rate": 9.137881834829404e-06, "loss": 3.6812, "step": 241800 }, { "epoch": 2.717064648949855, "grad_norm": 1.3747844696044922, "learning_rate": 9.135634921526553e-06, "loss": 3.6143, "step": 241850 }, { "epoch": 2.717626374120198, "grad_norm": 1.1211518049240112, "learning_rate": 9.133388008223704e-06, "loss": 3.5938, "step": 241900 }, { "epoch": 2.718188099290541, "grad_norm": 1.055198073387146, "learning_rate": 9.131141094920853e-06, "loss": 3.6315, "step": 241950 }, { "epoch": 2.7187498244608843, "grad_norm": 1.279075264930725, "learning_rate": 9.128894181618003e-06, "loss": 3.7427, "step": 242000 }, { "epoch": 2.7193115496312275, "grad_norm": 1.2764612436294556, "learning_rate": 9.126647268315154e-06, "loss": 3.6734, "step": 242050 }, { "epoch": 2.7198732748015706, "grad_norm": 1.2446975708007812, "learning_rate": 9.124400355012303e-06, "loss": 3.6952, "step": 242100 }, { "epoch": 2.7204349999719137, "grad_norm": 1.241504430770874, "learning_rate": 9.122153441709452e-06, "loss": 3.6832, "step": 242150 }, { "epoch": 2.720996725142257, "grad_norm": 1.3896424770355225, "learning_rate": 9.119906528406602e-06, "loss": 3.6839, "step": 242200 }, { "epoch": 2.7215584503126, "grad_norm": 1.1340253353118896, "learning_rate": 9.117659615103753e-06, "loss": 3.6136, "step": 242250 }, { "epoch": 2.722120175482943, "grad_norm": 1.163301944732666, "learning_rate": 9.115412701800902e-06, "loss": 3.5696, "step": 242300 }, { "epoch": 2.7226819006532863, "grad_norm": 1.260135531425476, "learning_rate": 9.113165788498051e-06, "loss": 3.5426, "step": 242350 }, { "epoch": 2.7232436258236294, "grad_norm": 1.1422429084777832, "learning_rate": 9.110918875195202e-06, "loss": 3.7006, "step": 242400 }, { "epoch": 2.7238053509939726, "grad_norm": 1.1454027891159058, "learning_rate": 9.108671961892352e-06, "loss": 3.5909, "step": 242450 }, { "epoch": 2.7243670761643157, "grad_norm": 1.1539671421051025, "learning_rate": 9.106425048589501e-06, "loss": 3.5875, "step": 242500 }, { "epoch": 2.724928801334659, "grad_norm": 1.1983054876327515, "learning_rate": 9.10417813528665e-06, "loss": 3.6367, "step": 242550 }, { "epoch": 2.725490526505002, "grad_norm": 2.16520094871521, "learning_rate": 9.101931221983801e-06, "loss": 3.5879, "step": 242600 }, { "epoch": 2.7260522516753456, "grad_norm": 1.152925968170166, "learning_rate": 9.09968430868095e-06, "loss": 3.6638, "step": 242650 }, { "epoch": 2.7266139768456883, "grad_norm": 1.5511201620101929, "learning_rate": 9.0974373953781e-06, "loss": 3.651, "step": 242700 }, { "epoch": 2.727175702016032, "grad_norm": 1.3489755392074585, "learning_rate": 9.09519048207525e-06, "loss": 3.6506, "step": 242750 }, { "epoch": 2.7277374271863746, "grad_norm": 1.2226667404174805, "learning_rate": 9.0929435687724e-06, "loss": 3.6772, "step": 242800 }, { "epoch": 2.728299152356718, "grad_norm": 1.0761747360229492, "learning_rate": 9.09069665546955e-06, "loss": 3.6534, "step": 242850 }, { "epoch": 2.728860877527061, "grad_norm": 1.2483551502227783, "learning_rate": 9.088449742166699e-06, "loss": 3.7027, "step": 242900 }, { "epoch": 2.7294226026974044, "grad_norm": 1.4201072454452515, "learning_rate": 9.08620282886385e-06, "loss": 3.6614, "step": 242950 }, { "epoch": 2.7299843278677476, "grad_norm": 1.1233983039855957, "learning_rate": 9.083955915561e-06, "loss": 3.6096, "step": 243000 }, { "epoch": 2.7305460530380907, "grad_norm": 1.2304747104644775, "learning_rate": 9.081709002258149e-06, "loss": 3.6144, "step": 243050 }, { "epoch": 2.731107778208434, "grad_norm": 1.2645460367202759, "learning_rate": 9.079462088955298e-06, "loss": 3.6092, "step": 243100 }, { "epoch": 2.731669503378777, "grad_norm": 1.2786263227462769, "learning_rate": 9.077215175652449e-06, "loss": 3.5947, "step": 243150 }, { "epoch": 2.73223122854912, "grad_norm": 1.286088228225708, "learning_rate": 9.074968262349598e-06, "loss": 3.6442, "step": 243200 }, { "epoch": 2.7327929537194633, "grad_norm": 1.1732699871063232, "learning_rate": 9.072721349046748e-06, "loss": 3.6664, "step": 243250 }, { "epoch": 2.7333546788898064, "grad_norm": 1.2338069677352905, "learning_rate": 9.070474435743899e-06, "loss": 3.6205, "step": 243300 }, { "epoch": 2.7339164040601496, "grad_norm": 1.1352485418319702, "learning_rate": 9.068227522441048e-06, "loss": 3.6537, "step": 243350 }, { "epoch": 2.7344781292304927, "grad_norm": 1.456569790840149, "learning_rate": 9.065980609138197e-06, "loss": 3.6952, "step": 243400 }, { "epoch": 2.735039854400836, "grad_norm": 1.1619824171066284, "learning_rate": 9.063733695835347e-06, "loss": 3.6651, "step": 243450 }, { "epoch": 2.735601579571179, "grad_norm": 1.2682414054870605, "learning_rate": 9.061531720798554e-06, "loss": 3.5823, "step": 243500 }, { "epoch": 2.736163304741522, "grad_norm": 1.2526882886886597, "learning_rate": 9.059284807495703e-06, "loss": 3.6704, "step": 243550 }, { "epoch": 2.7367250299118653, "grad_norm": 1.0567783117294312, "learning_rate": 9.057037894192854e-06, "loss": 3.6031, "step": 243600 }, { "epoch": 2.7372867550822084, "grad_norm": 1.2400263547897339, "learning_rate": 9.054790980890004e-06, "loss": 3.6466, "step": 243650 }, { "epoch": 2.7378484802525516, "grad_norm": 1.216665267944336, "learning_rate": 9.052544067587153e-06, "loss": 3.6579, "step": 243700 }, { "epoch": 2.7384102054228947, "grad_norm": 1.3223114013671875, "learning_rate": 9.050297154284302e-06, "loss": 3.617, "step": 243750 }, { "epoch": 2.738971930593238, "grad_norm": 1.1438169479370117, "learning_rate": 9.048050240981453e-06, "loss": 3.6198, "step": 243800 }, { "epoch": 2.739533655763581, "grad_norm": 1.1510319709777832, "learning_rate": 9.045803327678603e-06, "loss": 3.6681, "step": 243850 }, { "epoch": 2.740095380933924, "grad_norm": 1.2513720989227295, "learning_rate": 9.043556414375752e-06, "loss": 3.5245, "step": 243900 }, { "epoch": 2.7406571061042673, "grad_norm": 1.0472750663757324, "learning_rate": 9.041309501072903e-06, "loss": 3.6285, "step": 243950 }, { "epoch": 2.741218831274611, "grad_norm": 1.2637301683425903, "learning_rate": 9.03906258777005e-06, "loss": 3.6431, "step": 244000 }, { "epoch": 2.7417805564449536, "grad_norm": 1.1572527885437012, "learning_rate": 9.036815674467202e-06, "loss": 3.6351, "step": 244050 }, { "epoch": 2.742342281615297, "grad_norm": 1.313968539237976, "learning_rate": 9.034568761164351e-06, "loss": 3.6279, "step": 244100 }, { "epoch": 2.74290400678564, "grad_norm": 1.2717763185501099, "learning_rate": 9.0323218478615e-06, "loss": 3.6438, "step": 244150 }, { "epoch": 2.7434657319559834, "grad_norm": 1.3627235889434814, "learning_rate": 9.03007493455865e-06, "loss": 3.6907, "step": 244200 }, { "epoch": 2.744027457126326, "grad_norm": 1.1469571590423584, "learning_rate": 9.0278280212558e-06, "loss": 3.6161, "step": 244250 }, { "epoch": 2.7445891822966697, "grad_norm": 1.127825140953064, "learning_rate": 9.025626046219007e-06, "loss": 3.6382, "step": 244300 }, { "epoch": 2.745150907467013, "grad_norm": 1.3997374773025513, "learning_rate": 9.023379132916158e-06, "loss": 3.6522, "step": 244350 }, { "epoch": 2.745712632637356, "grad_norm": 1.309731125831604, "learning_rate": 9.021132219613307e-06, "loss": 3.6872, "step": 244400 }, { "epoch": 2.746274357807699, "grad_norm": 1.3515589237213135, "learning_rate": 9.018885306310456e-06, "loss": 3.6513, "step": 244450 }, { "epoch": 2.7468360829780423, "grad_norm": 1.3234989643096924, "learning_rate": 9.016638393007606e-06, "loss": 3.7034, "step": 244500 }, { "epoch": 2.7473978081483854, "grad_norm": 1.0811554193496704, "learning_rate": 9.014391479704757e-06, "loss": 3.6828, "step": 244550 }, { "epoch": 2.7479595333187286, "grad_norm": 1.234450340270996, "learning_rate": 9.012144566401906e-06, "loss": 3.7199, "step": 244600 }, { "epoch": 2.7485212584890717, "grad_norm": 1.0604885816574097, "learning_rate": 9.009897653099055e-06, "loss": 3.6605, "step": 244650 }, { "epoch": 2.749082983659415, "grad_norm": 1.215888500213623, "learning_rate": 9.007650739796206e-06, "loss": 3.5975, "step": 244700 }, { "epoch": 2.749644708829758, "grad_norm": 1.1834688186645508, "learning_rate": 9.005403826493356e-06, "loss": 3.7012, "step": 244750 }, { "epoch": 2.750206434000101, "grad_norm": 1.245441198348999, "learning_rate": 9.003156913190505e-06, "loss": 3.7003, "step": 244800 }, { "epoch": 2.7507681591704443, "grad_norm": 1.0024648904800415, "learning_rate": 9.000909999887654e-06, "loss": 3.6394, "step": 244850 }, { "epoch": 2.7513298843407874, "grad_norm": 1.1968562602996826, "learning_rate": 8.998663086584805e-06, "loss": 3.6342, "step": 244900 }, { "epoch": 2.7518916095111305, "grad_norm": 1.2056657075881958, "learning_rate": 8.996416173281954e-06, "loss": 3.5545, "step": 244950 }, { "epoch": 2.7524533346814737, "grad_norm": 1.3170746564865112, "learning_rate": 8.994169259979104e-06, "loss": 3.7121, "step": 245000 }, { "epoch": 2.753015059851817, "grad_norm": 1.23115074634552, "learning_rate": 8.991922346676255e-06, "loss": 3.5795, "step": 245050 }, { "epoch": 2.75357678502216, "grad_norm": 1.2126127481460571, "learning_rate": 8.989675433373404e-06, "loss": 3.5721, "step": 245100 }, { "epoch": 2.754138510192503, "grad_norm": 1.1153799295425415, "learning_rate": 8.987428520070553e-06, "loss": 3.7114, "step": 245150 }, { "epoch": 2.7547002353628463, "grad_norm": 1.2014546394348145, "learning_rate": 8.985181606767703e-06, "loss": 3.6424, "step": 245200 }, { "epoch": 2.7552619605331894, "grad_norm": 1.1441682577133179, "learning_rate": 8.982934693464854e-06, "loss": 3.5702, "step": 245250 }, { "epoch": 2.7558236857035325, "grad_norm": 1.2664340734481812, "learning_rate": 8.980687780162003e-06, "loss": 3.5781, "step": 245300 }, { "epoch": 2.756385410873876, "grad_norm": 1.2148077487945557, "learning_rate": 8.978440866859152e-06, "loss": 3.6007, "step": 245350 }, { "epoch": 2.756947136044219, "grad_norm": 1.3117756843566895, "learning_rate": 8.976193953556302e-06, "loss": 3.5995, "step": 245400 }, { "epoch": 2.7575088612145624, "grad_norm": 1.3556278944015503, "learning_rate": 8.973947040253453e-06, "loss": 3.6414, "step": 245450 }, { "epoch": 2.758070586384905, "grad_norm": 1.1827054023742676, "learning_rate": 8.971700126950602e-06, "loss": 3.6375, "step": 245500 }, { "epoch": 2.7586323115552487, "grad_norm": 1.2030131816864014, "learning_rate": 8.969453213647751e-06, "loss": 3.671, "step": 245550 }, { "epoch": 2.7591940367255914, "grad_norm": 1.1759666204452515, "learning_rate": 8.967206300344902e-06, "loss": 3.6656, "step": 245600 }, { "epoch": 2.759755761895935, "grad_norm": 1.3242117166519165, "learning_rate": 8.964959387042052e-06, "loss": 3.6524, "step": 245650 }, { "epoch": 2.760317487066278, "grad_norm": 1.2978301048278809, "learning_rate": 8.962712473739201e-06, "loss": 3.6528, "step": 245700 }, { "epoch": 2.7608792122366212, "grad_norm": 1.304042100906372, "learning_rate": 8.96046556043635e-06, "loss": 3.6286, "step": 245750 }, { "epoch": 2.7614409374069644, "grad_norm": 1.214600920677185, "learning_rate": 8.958218647133501e-06, "loss": 3.5943, "step": 245800 }, { "epoch": 2.7620026625773075, "grad_norm": 1.1684232950210571, "learning_rate": 8.95597173383065e-06, "loss": 3.551, "step": 245850 }, { "epoch": 2.7625643877476507, "grad_norm": 1.1153879165649414, "learning_rate": 8.9537248205278e-06, "loss": 3.6145, "step": 245900 }, { "epoch": 2.763126112917994, "grad_norm": 1.3280410766601562, "learning_rate": 8.951477907224951e-06, "loss": 3.6442, "step": 245950 }, { "epoch": 2.763687838088337, "grad_norm": 1.2092448472976685, "learning_rate": 8.9492309939221e-06, "loss": 3.7016, "step": 246000 }, { "epoch": 2.76424956325868, "grad_norm": 1.1283891201019287, "learning_rate": 8.94698408061925e-06, "loss": 3.6287, "step": 246050 }, { "epoch": 2.7648112884290232, "grad_norm": 1.2193617820739746, "learning_rate": 8.944737167316399e-06, "loss": 3.6846, "step": 246100 }, { "epoch": 2.7653730135993664, "grad_norm": 1.167502760887146, "learning_rate": 8.94249025401355e-06, "loss": 3.6715, "step": 246150 }, { "epoch": 2.7659347387697095, "grad_norm": 1.071590542793274, "learning_rate": 8.9402433407107e-06, "loss": 3.586, "step": 246200 }, { "epoch": 2.7664964639400527, "grad_norm": 1.2695493698120117, "learning_rate": 8.937996427407849e-06, "loss": 3.6704, "step": 246250 }, { "epoch": 2.767058189110396, "grad_norm": 1.181056261062622, "learning_rate": 8.935749514105e-06, "loss": 3.6904, "step": 246300 }, { "epoch": 2.767619914280739, "grad_norm": 1.0909191370010376, "learning_rate": 8.933502600802149e-06, "loss": 3.5455, "step": 246350 }, { "epoch": 2.768181639451082, "grad_norm": 1.1767804622650146, "learning_rate": 8.931255687499298e-06, "loss": 3.656, "step": 246400 }, { "epoch": 2.7687433646214252, "grad_norm": 1.1259640455245972, "learning_rate": 8.929008774196448e-06, "loss": 3.5926, "step": 246450 }, { "epoch": 2.7693050897917684, "grad_norm": 1.0753785371780396, "learning_rate": 8.926761860893599e-06, "loss": 3.6134, "step": 246500 }, { "epoch": 2.7698668149621115, "grad_norm": 1.1701672077178955, "learning_rate": 8.924514947590748e-06, "loss": 3.5493, "step": 246550 }, { "epoch": 2.7704285401324547, "grad_norm": 1.2856541872024536, "learning_rate": 8.922268034287897e-06, "loss": 3.7311, "step": 246600 }, { "epoch": 2.770990265302798, "grad_norm": 1.198441505432129, "learning_rate": 8.920021120985048e-06, "loss": 3.6904, "step": 246650 }, { "epoch": 2.7715519904731414, "grad_norm": 1.2065755128860474, "learning_rate": 8.917774207682197e-06, "loss": 3.5943, "step": 246700 }, { "epoch": 2.772113715643484, "grad_norm": 1.2504088878631592, "learning_rate": 8.915527294379347e-06, "loss": 3.5905, "step": 246750 }, { "epoch": 2.7726754408138277, "grad_norm": 1.1688779592514038, "learning_rate": 8.913280381076496e-06, "loss": 3.735, "step": 246800 }, { "epoch": 2.7732371659841704, "grad_norm": 1.1989463567733765, "learning_rate": 8.911033467773647e-06, "loss": 3.5422, "step": 246850 }, { "epoch": 2.773798891154514, "grad_norm": 1.1521341800689697, "learning_rate": 8.908786554470796e-06, "loss": 3.635, "step": 246900 }, { "epoch": 2.7743606163248566, "grad_norm": 1.1931767463684082, "learning_rate": 8.906539641167946e-06, "loss": 3.6291, "step": 246950 }, { "epoch": 2.7749223414952002, "grad_norm": 1.0621672868728638, "learning_rate": 8.904292727865095e-06, "loss": 3.6459, "step": 247000 }, { "epoch": 2.7754840666655434, "grad_norm": 1.0527377128601074, "learning_rate": 8.902045814562246e-06, "loss": 3.6414, "step": 247050 }, { "epoch": 2.7760457918358865, "grad_norm": 1.1050207614898682, "learning_rate": 8.899798901259395e-06, "loss": 3.626, "step": 247100 }, { "epoch": 2.7766075170062297, "grad_norm": 1.1118500232696533, "learning_rate": 8.897551987956545e-06, "loss": 3.6147, "step": 247150 }, { "epoch": 2.777169242176573, "grad_norm": 1.1933377981185913, "learning_rate": 8.895305074653696e-06, "loss": 3.6173, "step": 247200 }, { "epoch": 2.777730967346916, "grad_norm": 1.0509549379348755, "learning_rate": 8.893058161350845e-06, "loss": 3.6746, "step": 247250 }, { "epoch": 2.778292692517259, "grad_norm": 1.2486989498138428, "learning_rate": 8.890811248047994e-06, "loss": 3.6326, "step": 247300 }, { "epoch": 2.778854417687602, "grad_norm": 1.1824915409088135, "learning_rate": 8.888564334745144e-06, "loss": 3.6884, "step": 247350 }, { "epoch": 2.7794161428579454, "grad_norm": 1.0689419507980347, "learning_rate": 8.886317421442295e-06, "loss": 3.6419, "step": 247400 }, { "epoch": 2.7799778680282885, "grad_norm": 1.365525484085083, "learning_rate": 8.884070508139444e-06, "loss": 3.655, "step": 247450 }, { "epoch": 2.7805395931986316, "grad_norm": 1.170935034751892, "learning_rate": 8.881823594836593e-06, "loss": 3.6977, "step": 247500 }, { "epoch": 2.781101318368975, "grad_norm": 1.2151670455932617, "learning_rate": 8.879576681533744e-06, "loss": 3.6232, "step": 247550 }, { "epoch": 2.781663043539318, "grad_norm": 1.2371901273727417, "learning_rate": 8.877329768230894e-06, "loss": 3.6226, "step": 247600 }, { "epoch": 2.782224768709661, "grad_norm": 1.160022497177124, "learning_rate": 8.875082854928043e-06, "loss": 3.5917, "step": 247650 }, { "epoch": 2.782786493880004, "grad_norm": 1.082189679145813, "learning_rate": 8.872835941625192e-06, "loss": 3.5847, "step": 247700 }, { "epoch": 2.7833482190503473, "grad_norm": 1.2220501899719238, "learning_rate": 8.870589028322343e-06, "loss": 3.6836, "step": 247750 }, { "epoch": 2.7839099442206905, "grad_norm": 1.026036024093628, "learning_rate": 8.868342115019493e-06, "loss": 3.6766, "step": 247800 }, { "epoch": 2.7844716693910336, "grad_norm": 1.0759615898132324, "learning_rate": 8.866095201716642e-06, "loss": 3.5203, "step": 247850 }, { "epoch": 2.7850333945613768, "grad_norm": 1.0197969675064087, "learning_rate": 8.863848288413793e-06, "loss": 3.5947, "step": 247900 }, { "epoch": 2.78559511973172, "grad_norm": 1.2278189659118652, "learning_rate": 8.861601375110942e-06, "loss": 3.6332, "step": 247950 }, { "epoch": 2.786156844902063, "grad_norm": 1.2787965536117554, "learning_rate": 8.859354461808092e-06, "loss": 3.553, "step": 248000 }, { "epoch": 2.7867185700724066, "grad_norm": 1.1631124019622803, "learning_rate": 8.857107548505241e-06, "loss": 3.5928, "step": 248050 }, { "epoch": 2.7872802952427493, "grad_norm": 1.2057949304580688, "learning_rate": 8.854860635202392e-06, "loss": 3.5721, "step": 248100 }, { "epoch": 2.787842020413093, "grad_norm": 1.0266140699386597, "learning_rate": 8.852613721899541e-06, "loss": 3.6113, "step": 248150 }, { "epoch": 2.7884037455834356, "grad_norm": 1.293364405632019, "learning_rate": 8.85036680859669e-06, "loss": 3.698, "step": 248200 }, { "epoch": 2.788965470753779, "grad_norm": 1.096808910369873, "learning_rate": 8.848119895293842e-06, "loss": 3.6375, "step": 248250 }, { "epoch": 2.789527195924122, "grad_norm": 1.2148140668869019, "learning_rate": 8.845872981990991e-06, "loss": 3.6571, "step": 248300 }, { "epoch": 2.7900889210944655, "grad_norm": 1.2298568487167358, "learning_rate": 8.84362606868814e-06, "loss": 3.5967, "step": 248350 }, { "epoch": 2.7906506462648086, "grad_norm": 1.1999495029449463, "learning_rate": 8.841424093651348e-06, "loss": 3.6436, "step": 248400 }, { "epoch": 2.7912123714351518, "grad_norm": 1.2359338998794556, "learning_rate": 8.839177180348497e-06, "loss": 3.6498, "step": 248450 }, { "epoch": 2.791774096605495, "grad_norm": 1.1596742868423462, "learning_rate": 8.836930267045646e-06, "loss": 3.6416, "step": 248500 }, { "epoch": 2.792335821775838, "grad_norm": 1.1785552501678467, "learning_rate": 8.834728292008854e-06, "loss": 3.714, "step": 248550 }, { "epoch": 2.792897546946181, "grad_norm": 1.2940049171447754, "learning_rate": 8.832481378706003e-06, "loss": 3.6765, "step": 248600 }, { "epoch": 2.7934592721165243, "grad_norm": 1.2178881168365479, "learning_rate": 8.830234465403154e-06, "loss": 3.6165, "step": 248650 }, { "epoch": 2.7940209972868675, "grad_norm": 1.068784236907959, "learning_rate": 8.827987552100304e-06, "loss": 3.6239, "step": 248700 }, { "epoch": 2.7945827224572106, "grad_norm": 1.4802403450012207, "learning_rate": 8.825740638797453e-06, "loss": 3.6999, "step": 248750 }, { "epoch": 2.7951444476275538, "grad_norm": 1.1822059154510498, "learning_rate": 8.823493725494602e-06, "loss": 3.7425, "step": 248800 }, { "epoch": 2.795706172797897, "grad_norm": 1.0425667762756348, "learning_rate": 8.821246812191753e-06, "loss": 3.7056, "step": 248850 }, { "epoch": 2.79626789796824, "grad_norm": 1.1360570192337036, "learning_rate": 8.818999898888903e-06, "loss": 3.6334, "step": 248900 }, { "epoch": 2.796829623138583, "grad_norm": 1.0428361892700195, "learning_rate": 8.816752985586052e-06, "loss": 3.6138, "step": 248950 }, { "epoch": 2.7973913483089263, "grad_norm": 1.3827158212661743, "learning_rate": 8.814506072283201e-06, "loss": 3.669, "step": 249000 }, { "epoch": 2.7979530734792695, "grad_norm": 1.1668068170547485, "learning_rate": 8.812259158980352e-06, "loss": 3.6294, "step": 249050 }, { "epoch": 2.7985147986496126, "grad_norm": 1.1025125980377197, "learning_rate": 8.810012245677502e-06, "loss": 3.6196, "step": 249100 }, { "epoch": 2.7990765238199558, "grad_norm": 1.403463363647461, "learning_rate": 8.807765332374651e-06, "loss": 3.6405, "step": 249150 }, { "epoch": 2.799638248990299, "grad_norm": 1.200330138206482, "learning_rate": 8.805518419071802e-06, "loss": 3.6801, "step": 249200 }, { "epoch": 2.800199974160642, "grad_norm": 1.2137800455093384, "learning_rate": 8.803271505768951e-06, "loss": 3.7161, "step": 249250 }, { "epoch": 2.800761699330985, "grad_norm": 1.0471996068954468, "learning_rate": 8.8010245924661e-06, "loss": 3.6234, "step": 249300 }, { "epoch": 2.8013234245013283, "grad_norm": 1.1346518993377686, "learning_rate": 8.79877767916325e-06, "loss": 3.7064, "step": 249350 }, { "epoch": 2.801885149671672, "grad_norm": 1.2496943473815918, "learning_rate": 8.796530765860401e-06, "loss": 3.6554, "step": 249400 }, { "epoch": 2.8024468748420146, "grad_norm": 1.213826298713684, "learning_rate": 8.79428385255755e-06, "loss": 3.655, "step": 249450 }, { "epoch": 2.803008600012358, "grad_norm": 1.3619517087936401, "learning_rate": 8.7920369392547e-06, "loss": 3.5986, "step": 249500 }, { "epoch": 2.803570325182701, "grad_norm": 1.109195590019226, "learning_rate": 8.78979002595185e-06, "loss": 3.6476, "step": 249550 }, { "epoch": 2.8041320503530445, "grad_norm": 1.2254453897476196, "learning_rate": 8.787543112649e-06, "loss": 3.6908, "step": 249600 }, { "epoch": 2.804693775523387, "grad_norm": 1.133043885231018, "learning_rate": 8.78529619934615e-06, "loss": 3.704, "step": 249650 }, { "epoch": 2.8052555006937308, "grad_norm": 1.0666509866714478, "learning_rate": 8.783049286043298e-06, "loss": 3.6652, "step": 249700 }, { "epoch": 2.805817225864074, "grad_norm": 1.1767330169677734, "learning_rate": 8.78080237274045e-06, "loss": 3.6484, "step": 249750 }, { "epoch": 2.806378951034417, "grad_norm": 1.2071733474731445, "learning_rate": 8.778555459437599e-06, "loss": 3.5479, "step": 249800 }, { "epoch": 2.80694067620476, "grad_norm": 1.0609846115112305, "learning_rate": 8.776308546134748e-06, "loss": 3.6491, "step": 249850 }, { "epoch": 2.8075024013751033, "grad_norm": 1.076884388923645, "learning_rate": 8.774061632831899e-06, "loss": 3.6264, "step": 249900 }, { "epoch": 2.8080641265454465, "grad_norm": 1.1630980968475342, "learning_rate": 8.771814719529048e-06, "loss": 3.5596, "step": 249950 }, { "epoch": 2.8086258517157896, "grad_norm": 1.1163171529769897, "learning_rate": 8.769567806226198e-06, "loss": 3.6302, "step": 250000 }, { "epoch": 2.8091875768861327, "grad_norm": 1.2870421409606934, "learning_rate": 8.767320892923347e-06, "loss": 3.6759, "step": 250050 }, { "epoch": 2.809749302056476, "grad_norm": 1.2093838453292847, "learning_rate": 8.765073979620498e-06, "loss": 3.6711, "step": 250100 }, { "epoch": 2.810311027226819, "grad_norm": 1.1980772018432617, "learning_rate": 8.762827066317647e-06, "loss": 3.6602, "step": 250150 }, { "epoch": 2.810872752397162, "grad_norm": 1.2051241397857666, "learning_rate": 8.760580153014797e-06, "loss": 3.6949, "step": 250200 }, { "epoch": 2.8114344775675053, "grad_norm": 1.3431698083877563, "learning_rate": 8.758333239711948e-06, "loss": 3.6473, "step": 250250 }, { "epoch": 2.8119962027378484, "grad_norm": 1.0086201429367065, "learning_rate": 8.756086326409097e-06, "loss": 3.6524, "step": 250300 }, { "epoch": 2.8125579279081916, "grad_norm": 1.1246020793914795, "learning_rate": 8.753839413106246e-06, "loss": 3.6584, "step": 250350 }, { "epoch": 2.8131196530785347, "grad_norm": 1.124054193496704, "learning_rate": 8.751592499803396e-06, "loss": 3.6191, "step": 250400 }, { "epoch": 2.813681378248878, "grad_norm": 1.1673346757888794, "learning_rate": 8.749345586500547e-06, "loss": 3.7155, "step": 250450 }, { "epoch": 2.814243103419221, "grad_norm": 1.0143064260482788, "learning_rate": 8.747098673197696e-06, "loss": 3.6322, "step": 250500 }, { "epoch": 2.814804828589564, "grad_norm": 1.2310535907745361, "learning_rate": 8.744851759894845e-06, "loss": 3.6532, "step": 250550 }, { "epoch": 2.8153665537599073, "grad_norm": 1.1990960836410522, "learning_rate": 8.742604846591995e-06, "loss": 3.6442, "step": 250600 }, { "epoch": 2.8159282789302504, "grad_norm": 1.0461411476135254, "learning_rate": 8.740357933289146e-06, "loss": 3.576, "step": 250650 }, { "epoch": 2.8164900041005936, "grad_norm": 1.3315068483352661, "learning_rate": 8.738155958252352e-06, "loss": 3.7084, "step": 250700 }, { "epoch": 2.817051729270937, "grad_norm": 1.2003432512283325, "learning_rate": 8.735909044949501e-06, "loss": 3.7478, "step": 250750 }, { "epoch": 2.81761345444128, "grad_norm": 1.1496334075927734, "learning_rate": 8.73366213164665e-06, "loss": 3.6494, "step": 250800 }, { "epoch": 2.8181751796116234, "grad_norm": 1.3857228755950928, "learning_rate": 8.731415218343801e-06, "loss": 3.7085, "step": 250850 }, { "epoch": 2.818736904781966, "grad_norm": 1.5104680061340332, "learning_rate": 8.72916830504095e-06, "loss": 3.6291, "step": 250900 }, { "epoch": 2.8192986299523097, "grad_norm": 1.228920817375183, "learning_rate": 8.7269213917381e-06, "loss": 3.7008, "step": 250950 }, { "epoch": 2.8198603551226524, "grad_norm": 1.0269227027893066, "learning_rate": 8.72467447843525e-06, "loss": 3.6127, "step": 251000 }, { "epoch": 2.820422080292996, "grad_norm": 1.380207896232605, "learning_rate": 8.7224275651324e-06, "loss": 3.6471, "step": 251050 }, { "epoch": 2.820983805463339, "grad_norm": 1.117279052734375, "learning_rate": 8.72018065182955e-06, "loss": 3.6003, "step": 251100 }, { "epoch": 2.8215455306336823, "grad_norm": 1.062842607498169, "learning_rate": 8.717933738526699e-06, "loss": 3.5711, "step": 251150 }, { "epoch": 2.8221072558040254, "grad_norm": 0.9625957608222961, "learning_rate": 8.71568682522385e-06, "loss": 3.7088, "step": 251200 }, { "epoch": 2.8226689809743686, "grad_norm": 1.1299883127212524, "learning_rate": 8.713439911921e-06, "loss": 3.6721, "step": 251250 }, { "epoch": 2.8232307061447117, "grad_norm": 1.284261703491211, "learning_rate": 8.711192998618148e-06, "loss": 3.6396, "step": 251300 }, { "epoch": 2.823792431315055, "grad_norm": 1.3281031847000122, "learning_rate": 8.708946085315298e-06, "loss": 3.6529, "step": 251350 }, { "epoch": 2.824354156485398, "grad_norm": 1.1197929382324219, "learning_rate": 8.706699172012449e-06, "loss": 3.6708, "step": 251400 }, { "epoch": 2.824915881655741, "grad_norm": 1.1786137819290161, "learning_rate": 8.704452258709598e-06, "loss": 3.616, "step": 251450 }, { "epoch": 2.8254776068260843, "grad_norm": 1.1578930616378784, "learning_rate": 8.702205345406747e-06, "loss": 3.6754, "step": 251500 }, { "epoch": 2.8260393319964274, "grad_norm": 1.159993052482605, "learning_rate": 8.699958432103898e-06, "loss": 3.6259, "step": 251550 }, { "epoch": 2.8266010571667706, "grad_norm": 1.2689379453659058, "learning_rate": 8.697711518801048e-06, "loss": 3.6686, "step": 251600 }, { "epoch": 2.8271627823371137, "grad_norm": 1.140148639678955, "learning_rate": 8.695464605498197e-06, "loss": 3.6117, "step": 251650 }, { "epoch": 2.827724507507457, "grad_norm": 1.244386911392212, "learning_rate": 8.693217692195346e-06, "loss": 3.6526, "step": 251700 }, { "epoch": 2.8282862326778, "grad_norm": 1.2680684328079224, "learning_rate": 8.690970778892497e-06, "loss": 3.6854, "step": 251750 }, { "epoch": 2.828847957848143, "grad_norm": 1.140735149383545, "learning_rate": 8.688723865589647e-06, "loss": 3.6391, "step": 251800 }, { "epoch": 2.8294096830184863, "grad_norm": 1.1599724292755127, "learning_rate": 8.686476952286796e-06, "loss": 3.6935, "step": 251850 }, { "epoch": 2.8299714081888294, "grad_norm": 1.357729196548462, "learning_rate": 8.684230038983947e-06, "loss": 3.6596, "step": 251900 }, { "epoch": 2.8305331333591726, "grad_norm": 1.1147730350494385, "learning_rate": 8.681983125681096e-06, "loss": 3.7163, "step": 251950 }, { "epoch": 2.8310948585295157, "grad_norm": 1.0270379781723022, "learning_rate": 8.679736212378246e-06, "loss": 3.658, "step": 252000 }, { "epoch": 2.831656583699859, "grad_norm": 1.3949124813079834, "learning_rate": 8.677489299075395e-06, "loss": 3.6602, "step": 252050 }, { "epoch": 2.8322183088702024, "grad_norm": 1.193929672241211, "learning_rate": 8.675242385772546e-06, "loss": 3.6682, "step": 252100 }, { "epoch": 2.832780034040545, "grad_norm": 1.161643624305725, "learning_rate": 8.672995472469695e-06, "loss": 3.7082, "step": 252150 }, { "epoch": 2.8333417592108887, "grad_norm": 1.1729676723480225, "learning_rate": 8.670748559166845e-06, "loss": 3.6377, "step": 252200 }, { "epoch": 2.8339034843812314, "grad_norm": 1.2947314977645874, "learning_rate": 8.668501645863994e-06, "loss": 3.6678, "step": 252250 }, { "epoch": 2.834465209551575, "grad_norm": 1.1691313982009888, "learning_rate": 8.666254732561145e-06, "loss": 3.6623, "step": 252300 }, { "epoch": 2.8350269347219177, "grad_norm": 1.1190428733825684, "learning_rate": 8.664007819258294e-06, "loss": 3.6095, "step": 252350 }, { "epoch": 2.8355886598922613, "grad_norm": 1.2951792478561401, "learning_rate": 8.661760905955444e-06, "loss": 3.6125, "step": 252400 }, { "epoch": 2.8361503850626044, "grad_norm": 1.2349745035171509, "learning_rate": 8.659513992652595e-06, "loss": 3.6809, "step": 252450 }, { "epoch": 2.8367121102329476, "grad_norm": 1.2198148965835571, "learning_rate": 8.657267079349744e-06, "loss": 3.6674, "step": 252500 }, { "epoch": 2.8372738354032907, "grad_norm": 1.2092562913894653, "learning_rate": 8.655020166046893e-06, "loss": 3.6697, "step": 252550 }, { "epoch": 2.837835560573634, "grad_norm": 1.0556259155273438, "learning_rate": 8.652773252744043e-06, "loss": 3.7152, "step": 252600 }, { "epoch": 2.838397285743977, "grad_norm": 1.175010323524475, "learning_rate": 8.650526339441194e-06, "loss": 3.5873, "step": 252650 }, { "epoch": 2.83895901091432, "grad_norm": 1.234927773475647, "learning_rate": 8.648279426138343e-06, "loss": 3.5207, "step": 252700 }, { "epoch": 2.8395207360846633, "grad_norm": 1.3387160301208496, "learning_rate": 8.646032512835492e-06, "loss": 3.6267, "step": 252750 }, { "epoch": 2.8400824612550064, "grad_norm": 1.3436369895935059, "learning_rate": 8.643785599532643e-06, "loss": 3.6727, "step": 252800 }, { "epoch": 2.8406441864253495, "grad_norm": 1.1480870246887207, "learning_rate": 8.641538686229793e-06, "loss": 3.5908, "step": 252850 }, { "epoch": 2.8412059115956927, "grad_norm": 1.1611076593399048, "learning_rate": 8.639291772926942e-06, "loss": 3.5875, "step": 252900 }, { "epoch": 2.841767636766036, "grad_norm": 1.1544500589370728, "learning_rate": 8.637044859624091e-06, "loss": 3.6649, "step": 252950 }, { "epoch": 2.842329361936379, "grad_norm": 1.2998915910720825, "learning_rate": 8.634797946321242e-06, "loss": 3.6305, "step": 253000 }, { "epoch": 2.842891087106722, "grad_norm": 1.068569302558899, "learning_rate": 8.632551033018392e-06, "loss": 3.5911, "step": 253050 }, { "epoch": 2.8434528122770653, "grad_norm": 1.2081338167190552, "learning_rate": 8.63030411971554e-06, "loss": 3.7156, "step": 253100 }, { "epoch": 2.8440145374474084, "grad_norm": 1.0115337371826172, "learning_rate": 8.628057206412692e-06, "loss": 3.5947, "step": 253150 }, { "epoch": 2.8445762626177515, "grad_norm": 1.187606930732727, "learning_rate": 8.625810293109841e-06, "loss": 3.5828, "step": 253200 }, { "epoch": 2.8451379877880947, "grad_norm": 1.1698554754257202, "learning_rate": 8.62356337980699e-06, "loss": 3.6434, "step": 253250 }, { "epoch": 2.845699712958438, "grad_norm": 1.352535367012024, "learning_rate": 8.62131646650414e-06, "loss": 3.5652, "step": 253300 }, { "epoch": 2.846261438128781, "grad_norm": 1.1410608291625977, "learning_rate": 8.61906955320129e-06, "loss": 3.6455, "step": 253350 }, { "epoch": 2.846823163299124, "grad_norm": 1.1857903003692627, "learning_rate": 8.61682263989844e-06, "loss": 3.6229, "step": 253400 }, { "epoch": 2.8473848884694677, "grad_norm": 1.272344708442688, "learning_rate": 8.61457572659559e-06, "loss": 3.6278, "step": 253450 }, { "epoch": 2.8479466136398104, "grad_norm": 0.9948447346687317, "learning_rate": 8.61232881329274e-06, "loss": 3.6406, "step": 253500 }, { "epoch": 2.848508338810154, "grad_norm": 1.1672770977020264, "learning_rate": 8.61008189998989e-06, "loss": 3.6908, "step": 253550 }, { "epoch": 2.8490700639804967, "grad_norm": 1.1695061922073364, "learning_rate": 8.607834986687039e-06, "loss": 3.6661, "step": 253600 }, { "epoch": 2.8496317891508403, "grad_norm": 1.0867644548416138, "learning_rate": 8.605588073384188e-06, "loss": 3.7171, "step": 253650 }, { "epoch": 2.850193514321183, "grad_norm": 1.081315040588379, "learning_rate": 8.60334116008134e-06, "loss": 3.5856, "step": 253700 }, { "epoch": 2.8507552394915265, "grad_norm": 1.0130645036697388, "learning_rate": 8.601094246778489e-06, "loss": 3.6423, "step": 253750 }, { "epoch": 2.8513169646618697, "grad_norm": 1.2605386972427368, "learning_rate": 8.598847333475638e-06, "loss": 3.6737, "step": 253800 }, { "epoch": 2.851878689832213, "grad_norm": 1.3743705749511719, "learning_rate": 8.596600420172787e-06, "loss": 3.6046, "step": 253850 }, { "epoch": 2.852440415002556, "grad_norm": 1.1312140226364136, "learning_rate": 8.594353506869938e-06, "loss": 3.6586, "step": 253900 }, { "epoch": 2.853002140172899, "grad_norm": 1.0801674127578735, "learning_rate": 8.592106593567088e-06, "loss": 3.7009, "step": 253950 }, { "epoch": 2.8535638653432422, "grad_norm": 1.1705094575881958, "learning_rate": 8.589859680264237e-06, "loss": 3.6314, "step": 254000 }, { "epoch": 2.8541255905135854, "grad_norm": 1.0673092603683472, "learning_rate": 8.587612766961388e-06, "loss": 3.6911, "step": 254050 }, { "epoch": 2.8546873156839285, "grad_norm": 1.3116928339004517, "learning_rate": 8.585365853658537e-06, "loss": 3.7241, "step": 254100 }, { "epoch": 2.8552490408542717, "grad_norm": 1.0950393676757812, "learning_rate": 8.583118940355687e-06, "loss": 3.6792, "step": 254150 }, { "epoch": 2.855810766024615, "grad_norm": 1.0413347482681274, "learning_rate": 8.580872027052836e-06, "loss": 3.6527, "step": 254200 }, { "epoch": 2.856372491194958, "grad_norm": 1.2501533031463623, "learning_rate": 8.578625113749987e-06, "loss": 3.5793, "step": 254250 }, { "epoch": 2.856934216365301, "grad_norm": 1.2898277044296265, "learning_rate": 8.576378200447136e-06, "loss": 3.6846, "step": 254300 }, { "epoch": 2.8574959415356442, "grad_norm": 1.5056421756744385, "learning_rate": 8.574131287144286e-06, "loss": 3.7161, "step": 254350 }, { "epoch": 2.8580576667059874, "grad_norm": 1.0803170204162598, "learning_rate": 8.571884373841437e-06, "loss": 3.6196, "step": 254400 }, { "epoch": 2.8586193918763305, "grad_norm": 1.1864174604415894, "learning_rate": 8.569637460538586e-06, "loss": 3.7141, "step": 254450 }, { "epoch": 2.8591811170466737, "grad_norm": 1.1153154373168945, "learning_rate": 8.567390547235735e-06, "loss": 3.5161, "step": 254500 }, { "epoch": 2.859742842217017, "grad_norm": 1.189439296722412, "learning_rate": 8.565143633932885e-06, "loss": 3.5413, "step": 254550 }, { "epoch": 2.86030456738736, "grad_norm": 1.0375182628631592, "learning_rate": 8.562896720630036e-06, "loss": 3.6565, "step": 254600 }, { "epoch": 2.860866292557703, "grad_norm": 1.4673399925231934, "learning_rate": 8.560649807327185e-06, "loss": 3.643, "step": 254650 }, { "epoch": 2.8614280177280462, "grad_norm": 1.2392841577529907, "learning_rate": 8.558402894024334e-06, "loss": 3.6671, "step": 254700 }, { "epoch": 2.8619897428983894, "grad_norm": 1.0539594888687134, "learning_rate": 8.556155980721485e-06, "loss": 3.6985, "step": 254750 }, { "epoch": 2.862551468068733, "grad_norm": 1.079054832458496, "learning_rate": 8.553909067418635e-06, "loss": 3.5707, "step": 254800 }, { "epoch": 2.8631131932390756, "grad_norm": 1.0152620077133179, "learning_rate": 8.55170709238184e-06, "loss": 3.5897, "step": 254850 }, { "epoch": 2.8636749184094192, "grad_norm": 1.104443073272705, "learning_rate": 8.549460179078991e-06, "loss": 3.6845, "step": 254900 }, { "epoch": 2.864236643579762, "grad_norm": 1.1426451206207275, "learning_rate": 8.54721326577614e-06, "loss": 3.6354, "step": 254950 }, { "epoch": 2.8647983687501055, "grad_norm": 1.0804557800292969, "learning_rate": 8.54496635247329e-06, "loss": 3.6437, "step": 255000 }, { "epoch": 2.865360093920448, "grad_norm": 1.399170994758606, "learning_rate": 8.542719439170441e-06, "loss": 3.6506, "step": 255050 }, { "epoch": 2.865921819090792, "grad_norm": 1.1732996702194214, "learning_rate": 8.54047252586759e-06, "loss": 3.6017, "step": 255100 }, { "epoch": 2.866483544261135, "grad_norm": 1.1958513259887695, "learning_rate": 8.53822561256474e-06, "loss": 3.7021, "step": 255150 }, { "epoch": 2.867045269431478, "grad_norm": 1.0984779596328735, "learning_rate": 8.535978699261889e-06, "loss": 3.6368, "step": 255200 }, { "epoch": 2.8676069946018212, "grad_norm": 1.165334701538086, "learning_rate": 8.53373178595904e-06, "loss": 3.68, "step": 255250 }, { "epoch": 2.8681687197721644, "grad_norm": 1.0844663381576538, "learning_rate": 8.53148487265619e-06, "loss": 3.6386, "step": 255300 }, { "epoch": 2.8687304449425075, "grad_norm": 1.0539356470108032, "learning_rate": 8.529237959353339e-06, "loss": 3.6779, "step": 255350 }, { "epoch": 2.8692921701128506, "grad_norm": 1.1420127153396606, "learning_rate": 8.52699104605049e-06, "loss": 3.6267, "step": 255400 }, { "epoch": 2.869853895283194, "grad_norm": 1.3625901937484741, "learning_rate": 8.524744132747639e-06, "loss": 3.6376, "step": 255450 }, { "epoch": 2.870415620453537, "grad_norm": 1.2747001647949219, "learning_rate": 8.522497219444788e-06, "loss": 3.6448, "step": 255500 }, { "epoch": 2.87097734562388, "grad_norm": 1.2826663255691528, "learning_rate": 8.520250306141938e-06, "loss": 3.623, "step": 255550 }, { "epoch": 2.871539070794223, "grad_norm": 1.063071608543396, "learning_rate": 8.518003392839089e-06, "loss": 3.6853, "step": 255600 }, { "epoch": 2.8721007959645664, "grad_norm": 1.176928997039795, "learning_rate": 8.515756479536238e-06, "loss": 3.6428, "step": 255650 }, { "epoch": 2.8726625211349095, "grad_norm": 1.1733942031860352, "learning_rate": 8.513509566233387e-06, "loss": 3.6082, "step": 255700 }, { "epoch": 2.8732242463052526, "grad_norm": 1.1952406167984009, "learning_rate": 8.511262652930538e-06, "loss": 3.6099, "step": 255750 }, { "epoch": 2.873785971475596, "grad_norm": 1.1005576848983765, "learning_rate": 8.509015739627688e-06, "loss": 3.5694, "step": 255800 }, { "epoch": 2.874347696645939, "grad_norm": 1.929207682609558, "learning_rate": 8.506768826324837e-06, "loss": 3.6943, "step": 255850 }, { "epoch": 2.874909421816282, "grad_norm": 1.0577033758163452, "learning_rate": 8.504521913021986e-06, "loss": 3.6198, "step": 255900 }, { "epoch": 2.875471146986625, "grad_norm": 0.9591113924980164, "learning_rate": 8.502274999719137e-06, "loss": 3.6189, "step": 255950 }, { "epoch": 2.8760328721569683, "grad_norm": 1.1310296058654785, "learning_rate": 8.500028086416287e-06, "loss": 3.6907, "step": 256000 }, { "epoch": 2.8765945973273115, "grad_norm": 1.4007959365844727, "learning_rate": 8.497781173113436e-06, "loss": 3.6813, "step": 256050 }, { "epoch": 2.8771563224976546, "grad_norm": 1.0054186582565308, "learning_rate": 8.495534259810587e-06, "loss": 3.7063, "step": 256100 }, { "epoch": 2.877718047667998, "grad_norm": 1.243122935295105, "learning_rate": 8.493287346507736e-06, "loss": 3.6054, "step": 256150 }, { "epoch": 2.878279772838341, "grad_norm": 1.2411491870880127, "learning_rate": 8.491040433204886e-06, "loss": 3.7219, "step": 256200 }, { "epoch": 2.8788414980086845, "grad_norm": 0.8729803562164307, "learning_rate": 8.488793519902035e-06, "loss": 3.6231, "step": 256250 }, { "epoch": 2.879403223179027, "grad_norm": 1.0931246280670166, "learning_rate": 8.486546606599186e-06, "loss": 3.6765, "step": 256300 }, { "epoch": 2.879964948349371, "grad_norm": 1.240601897239685, "learning_rate": 8.484299693296335e-06, "loss": 3.6464, "step": 256350 }, { "epoch": 2.8805266735197135, "grad_norm": 1.2226136922836304, "learning_rate": 8.482052779993485e-06, "loss": 3.5675, "step": 256400 }, { "epoch": 2.881088398690057, "grad_norm": 1.1992837190628052, "learning_rate": 8.479805866690634e-06, "loss": 3.6351, "step": 256450 }, { "epoch": 2.8816501238604, "grad_norm": 1.183427095413208, "learning_rate": 8.477558953387785e-06, "loss": 3.6925, "step": 256500 }, { "epoch": 2.8822118490307433, "grad_norm": 1.1084449291229248, "learning_rate": 8.475312040084934e-06, "loss": 3.5992, "step": 256550 }, { "epoch": 2.8827735742010865, "grad_norm": 1.1920135021209717, "learning_rate": 8.473065126782083e-06, "loss": 3.5795, "step": 256600 }, { "epoch": 2.8833352993714296, "grad_norm": 1.4281495809555054, "learning_rate": 8.470818213479234e-06, "loss": 3.5584, "step": 256650 }, { "epoch": 2.8838970245417728, "grad_norm": 1.243395447731018, "learning_rate": 8.468571300176384e-06, "loss": 3.5865, "step": 256700 }, { "epoch": 2.884458749712116, "grad_norm": 1.2723886966705322, "learning_rate": 8.466324386873533e-06, "loss": 3.6345, "step": 256750 }, { "epoch": 2.885020474882459, "grad_norm": 1.2030029296875, "learning_rate": 8.464077473570682e-06, "loss": 3.7157, "step": 256800 }, { "epoch": 2.885582200052802, "grad_norm": 1.2293763160705566, "learning_rate": 8.461830560267833e-06, "loss": 3.5977, "step": 256850 }, { "epoch": 2.8861439252231453, "grad_norm": 1.2963725328445435, "learning_rate": 8.459583646964983e-06, "loss": 3.6438, "step": 256900 }, { "epoch": 2.8867056503934885, "grad_norm": 1.3075108528137207, "learning_rate": 8.457336733662132e-06, "loss": 3.572, "step": 256950 }, { "epoch": 2.8872673755638316, "grad_norm": 1.0276139974594116, "learning_rate": 8.455089820359283e-06, "loss": 3.6522, "step": 257000 }, { "epoch": 2.8878291007341748, "grad_norm": 1.1830617189407349, "learning_rate": 8.452842907056432e-06, "loss": 3.6152, "step": 257050 }, { "epoch": 2.888390825904518, "grad_norm": 1.220034122467041, "learning_rate": 8.450595993753582e-06, "loss": 3.6242, "step": 257100 }, { "epoch": 2.888952551074861, "grad_norm": 1.1185407638549805, "learning_rate": 8.448349080450731e-06, "loss": 3.6761, "step": 257150 }, { "epoch": 2.889514276245204, "grad_norm": 1.1763192415237427, "learning_rate": 8.446102167147882e-06, "loss": 3.7019, "step": 257200 }, { "epoch": 2.8900760014155473, "grad_norm": 1.1944607496261597, "learning_rate": 8.443900192111088e-06, "loss": 3.6723, "step": 257250 }, { "epoch": 2.8906377265858905, "grad_norm": 1.1896106004714966, "learning_rate": 8.441653278808239e-06, "loss": 3.6422, "step": 257300 }, { "epoch": 2.8911994517562336, "grad_norm": 1.2760024070739746, "learning_rate": 8.439406365505388e-06, "loss": 3.648, "step": 257350 }, { "epoch": 2.8917611769265767, "grad_norm": 1.2450604438781738, "learning_rate": 8.437159452202538e-06, "loss": 3.6805, "step": 257400 }, { "epoch": 2.89232290209692, "grad_norm": 1.4671880006790161, "learning_rate": 8.434912538899687e-06, "loss": 3.5994, "step": 257450 }, { "epoch": 2.8928846272672635, "grad_norm": 1.2130038738250732, "learning_rate": 8.432665625596838e-06, "loss": 3.675, "step": 257500 }, { "epoch": 2.893446352437606, "grad_norm": 1.521599531173706, "learning_rate": 8.430418712293987e-06, "loss": 3.613, "step": 257550 }, { "epoch": 2.8940080776079498, "grad_norm": 1.0279889106750488, "learning_rate": 8.428171798991137e-06, "loss": 3.6016, "step": 257600 }, { "epoch": 2.8945698027782925, "grad_norm": 1.2140361070632935, "learning_rate": 8.425924885688288e-06, "loss": 3.5809, "step": 257650 }, { "epoch": 2.895131527948636, "grad_norm": 1.276888370513916, "learning_rate": 8.423677972385437e-06, "loss": 3.6281, "step": 257700 }, { "epoch": 2.8956932531189787, "grad_norm": 1.1586765050888062, "learning_rate": 8.421431059082586e-06, "loss": 3.6154, "step": 257750 }, { "epoch": 2.8962549782893223, "grad_norm": 1.2609378099441528, "learning_rate": 8.419184145779736e-06, "loss": 3.5866, "step": 257800 }, { "epoch": 2.8968167034596655, "grad_norm": 1.2265291213989258, "learning_rate": 8.416937232476887e-06, "loss": 3.6479, "step": 257850 }, { "epoch": 2.8973784286300086, "grad_norm": 1.1802059412002563, "learning_rate": 8.414690319174036e-06, "loss": 3.656, "step": 257900 }, { "epoch": 2.8979401538003517, "grad_norm": 1.4281153678894043, "learning_rate": 8.412443405871185e-06, "loss": 3.6283, "step": 257950 }, { "epoch": 2.898501878970695, "grad_norm": 1.294285774230957, "learning_rate": 8.410196492568336e-06, "loss": 3.6803, "step": 258000 }, { "epoch": 2.899063604141038, "grad_norm": 1.129927158355713, "learning_rate": 8.407949579265485e-06, "loss": 3.717, "step": 258050 }, { "epoch": 2.899625329311381, "grad_norm": 1.1277838945388794, "learning_rate": 8.405702665962635e-06, "loss": 3.6004, "step": 258100 }, { "epoch": 2.9001870544817243, "grad_norm": 1.3431695699691772, "learning_rate": 8.403455752659784e-06, "loss": 3.7418, "step": 258150 }, { "epoch": 2.9007487796520675, "grad_norm": 1.0524933338165283, "learning_rate": 8.401208839356935e-06, "loss": 3.609, "step": 258200 }, { "epoch": 2.9013105048224106, "grad_norm": 1.154251217842102, "learning_rate": 8.398961926054083e-06, "loss": 3.6867, "step": 258250 }, { "epoch": 2.9018722299927537, "grad_norm": 1.2041724920272827, "learning_rate": 8.396715012751234e-06, "loss": 3.733, "step": 258300 }, { "epoch": 2.902433955163097, "grad_norm": 1.2063065767288208, "learning_rate": 8.394468099448383e-06, "loss": 3.653, "step": 258350 }, { "epoch": 2.90299568033344, "grad_norm": 1.105446219444275, "learning_rate": 8.392221186145532e-06, "loss": 3.6039, "step": 258400 }, { "epoch": 2.903557405503783, "grad_norm": 1.1139814853668213, "learning_rate": 8.389974272842682e-06, "loss": 3.6277, "step": 258450 }, { "epoch": 2.9041191306741263, "grad_norm": 1.2182289361953735, "learning_rate": 8.387727359539833e-06, "loss": 3.7213, "step": 258500 }, { "epoch": 2.9046808558444694, "grad_norm": 1.2544748783111572, "learning_rate": 8.385480446236982e-06, "loss": 3.5964, "step": 258550 }, { "epoch": 2.9052425810148126, "grad_norm": 1.1787946224212646, "learning_rate": 8.383233532934131e-06, "loss": 3.6215, "step": 258600 }, { "epoch": 2.9058043061851557, "grad_norm": 1.1869672536849976, "learning_rate": 8.380986619631282e-06, "loss": 3.6412, "step": 258650 }, { "epoch": 2.906366031355499, "grad_norm": 1.1628656387329102, "learning_rate": 8.378739706328432e-06, "loss": 3.6151, "step": 258700 }, { "epoch": 2.906927756525842, "grad_norm": 1.1098726987838745, "learning_rate": 8.376492793025581e-06, "loss": 3.6668, "step": 258750 }, { "epoch": 2.907489481696185, "grad_norm": 1.0621229410171509, "learning_rate": 8.37424587972273e-06, "loss": 3.7166, "step": 258800 }, { "epoch": 2.9080512068665287, "grad_norm": 1.275601863861084, "learning_rate": 8.371998966419881e-06, "loss": 3.6822, "step": 258850 }, { "epoch": 2.9086129320368714, "grad_norm": 1.1331881284713745, "learning_rate": 8.36975205311703e-06, "loss": 3.5805, "step": 258900 }, { "epoch": 2.909174657207215, "grad_norm": 1.4205636978149414, "learning_rate": 8.36750513981418e-06, "loss": 3.6829, "step": 258950 }, { "epoch": 2.9097363823775577, "grad_norm": 1.1129319667816162, "learning_rate": 8.365258226511331e-06, "loss": 3.723, "step": 259000 }, { "epoch": 2.9102981075479013, "grad_norm": 1.158905267715454, "learning_rate": 8.36301131320848e-06, "loss": 3.6381, "step": 259050 }, { "epoch": 2.910859832718244, "grad_norm": 0.996040403842926, "learning_rate": 8.36076439990563e-06, "loss": 3.6112, "step": 259100 }, { "epoch": 2.9114215578885876, "grad_norm": 1.0164822340011597, "learning_rate": 8.358517486602779e-06, "loss": 3.6798, "step": 259150 }, { "epoch": 2.9119832830589307, "grad_norm": 1.1888489723205566, "learning_rate": 8.35627057329993e-06, "loss": 3.6172, "step": 259200 }, { "epoch": 2.912545008229274, "grad_norm": 1.2077574729919434, "learning_rate": 8.35402365999708e-06, "loss": 3.5739, "step": 259250 }, { "epoch": 2.913106733399617, "grad_norm": 1.1407501697540283, "learning_rate": 8.351776746694229e-06, "loss": 3.6503, "step": 259300 }, { "epoch": 2.91366845856996, "grad_norm": 1.205299735069275, "learning_rate": 8.34952983339138e-06, "loss": 3.5773, "step": 259350 }, { "epoch": 2.9142301837403033, "grad_norm": 1.109458327293396, "learning_rate": 8.347282920088529e-06, "loss": 3.6248, "step": 259400 }, { "epoch": 2.9147919089106464, "grad_norm": 1.1124844551086426, "learning_rate": 8.345036006785678e-06, "loss": 3.6433, "step": 259450 }, { "epoch": 2.9153536340809896, "grad_norm": 1.4625402688980103, "learning_rate": 8.342789093482828e-06, "loss": 3.6646, "step": 259500 }, { "epoch": 2.9159153592513327, "grad_norm": 1.2352852821350098, "learning_rate": 8.340542180179979e-06, "loss": 3.6157, "step": 259550 }, { "epoch": 2.916477084421676, "grad_norm": 1.5141936540603638, "learning_rate": 8.338295266877128e-06, "loss": 3.641, "step": 259600 }, { "epoch": 2.917038809592019, "grad_norm": 1.446658968925476, "learning_rate": 8.336048353574277e-06, "loss": 3.6134, "step": 259650 }, { "epoch": 2.917600534762362, "grad_norm": 1.0663747787475586, "learning_rate": 8.333801440271428e-06, "loss": 3.6095, "step": 259700 }, { "epoch": 2.9181622599327053, "grad_norm": 1.14284348487854, "learning_rate": 8.331554526968578e-06, "loss": 3.5978, "step": 259750 }, { "epoch": 2.9187239851030484, "grad_norm": 1.254639744758606, "learning_rate": 8.329307613665727e-06, "loss": 3.669, "step": 259800 }, { "epoch": 2.9192857102733916, "grad_norm": 1.1343966722488403, "learning_rate": 8.327060700362876e-06, "loss": 3.6259, "step": 259850 }, { "epoch": 2.9198474354437347, "grad_norm": 0.9781042337417603, "learning_rate": 8.324813787060027e-06, "loss": 3.602, "step": 259900 }, { "epoch": 2.920409160614078, "grad_norm": 1.1170762777328491, "learning_rate": 8.322566873757176e-06, "loss": 3.6324, "step": 259950 }, { "epoch": 2.920970885784421, "grad_norm": 1.15255606174469, "learning_rate": 8.320319960454326e-06, "loss": 3.6607, "step": 260000 }, { "epoch": 2.921532610954764, "grad_norm": 1.0240029096603394, "learning_rate": 8.318073047151475e-06, "loss": 3.6105, "step": 260050 }, { "epoch": 2.9220943361251073, "grad_norm": 1.3248906135559082, "learning_rate": 8.315826133848626e-06, "loss": 3.6399, "step": 260100 }, { "epoch": 2.9226560612954504, "grad_norm": 1.2919012308120728, "learning_rate": 8.313579220545775e-06, "loss": 3.6315, "step": 260150 }, { "epoch": 2.923217786465794, "grad_norm": 1.1979879140853882, "learning_rate": 8.311332307242925e-06, "loss": 3.6296, "step": 260200 }, { "epoch": 2.9237795116361367, "grad_norm": 1.1804121732711792, "learning_rate": 8.309085393940076e-06, "loss": 3.6084, "step": 260250 }, { "epoch": 2.9243412368064803, "grad_norm": 1.197500467300415, "learning_rate": 8.306838480637225e-06, "loss": 3.6655, "step": 260300 }, { "epoch": 2.924902961976823, "grad_norm": 1.2215604782104492, "learning_rate": 8.304591567334374e-06, "loss": 3.657, "step": 260350 }, { "epoch": 2.9254646871471666, "grad_norm": 1.2738559246063232, "learning_rate": 8.302344654031524e-06, "loss": 3.657, "step": 260400 }, { "epoch": 2.9260264123175093, "grad_norm": 1.082836389541626, "learning_rate": 8.300097740728675e-06, "loss": 3.6958, "step": 260450 }, { "epoch": 2.926588137487853, "grad_norm": 1.264907717704773, "learning_rate": 8.297850827425824e-06, "loss": 3.6554, "step": 260500 }, { "epoch": 2.927149862658196, "grad_norm": 1.1699469089508057, "learning_rate": 8.295603914122973e-06, "loss": 3.6287, "step": 260550 }, { "epoch": 2.927711587828539, "grad_norm": 1.3663575649261475, "learning_rate": 8.293401939086181e-06, "loss": 3.6935, "step": 260600 }, { "epoch": 2.9282733129988823, "grad_norm": 1.222821831703186, "learning_rate": 8.29115502578333e-06, "loss": 3.6179, "step": 260650 }, { "epoch": 2.9288350381692254, "grad_norm": 1.0978554487228394, "learning_rate": 8.28890811248048e-06, "loss": 3.6874, "step": 260700 }, { "epoch": 2.9293967633395686, "grad_norm": 1.1071609258651733, "learning_rate": 8.28666119917763e-06, "loss": 3.6344, "step": 260750 }, { "epoch": 2.9299584885099117, "grad_norm": 1.1441104412078857, "learning_rate": 8.28441428587478e-06, "loss": 3.7294, "step": 260800 }, { "epoch": 2.930520213680255, "grad_norm": 1.0611757040023804, "learning_rate": 8.28216737257193e-06, "loss": 3.5771, "step": 260850 }, { "epoch": 2.931081938850598, "grad_norm": 1.0881798267364502, "learning_rate": 8.27992045926908e-06, "loss": 3.577, "step": 260900 }, { "epoch": 2.931643664020941, "grad_norm": 1.148647427558899, "learning_rate": 8.27767354596623e-06, "loss": 3.6592, "step": 260950 }, { "epoch": 2.9322053891912843, "grad_norm": 1.1440473794937134, "learning_rate": 8.275426632663379e-06, "loss": 3.6693, "step": 261000 }, { "epoch": 2.9327671143616274, "grad_norm": 1.1127500534057617, "learning_rate": 8.273179719360528e-06, "loss": 3.6086, "step": 261050 }, { "epoch": 2.9333288395319705, "grad_norm": 1.1406452655792236, "learning_rate": 8.27093280605768e-06, "loss": 3.6302, "step": 261100 }, { "epoch": 2.9338905647023137, "grad_norm": 1.1945470571517944, "learning_rate": 8.268730831020885e-06, "loss": 3.7131, "step": 261150 }, { "epoch": 2.934452289872657, "grad_norm": 1.1877835988998413, "learning_rate": 8.266483917718036e-06, "loss": 3.5216, "step": 261200 }, { "epoch": 2.935014015043, "grad_norm": 1.5802574157714844, "learning_rate": 8.264237004415185e-06, "loss": 3.6957, "step": 261250 }, { "epoch": 2.935575740213343, "grad_norm": 1.353175401687622, "learning_rate": 8.261990091112335e-06, "loss": 3.6262, "step": 261300 }, { "epoch": 2.9361374653836863, "grad_norm": 1.2118300199508667, "learning_rate": 8.259743177809486e-06, "loss": 3.7121, "step": 261350 }, { "epoch": 2.9366991905540294, "grad_norm": 1.281163215637207, "learning_rate": 8.257496264506635e-06, "loss": 3.6363, "step": 261400 }, { "epoch": 2.9372609157243725, "grad_norm": 1.1473280191421509, "learning_rate": 8.255249351203784e-06, "loss": 3.6721, "step": 261450 }, { "epoch": 2.9378226408947157, "grad_norm": 1.249420166015625, "learning_rate": 8.253002437900934e-06, "loss": 3.5846, "step": 261500 }, { "epoch": 2.9383843660650593, "grad_norm": 1.2632077932357788, "learning_rate": 8.250755524598085e-06, "loss": 3.6419, "step": 261550 }, { "epoch": 2.938946091235402, "grad_norm": 1.0302752256393433, "learning_rate": 8.248508611295234e-06, "loss": 3.5984, "step": 261600 }, { "epoch": 2.9395078164057455, "grad_norm": 1.2598966360092163, "learning_rate": 8.246261697992383e-06, "loss": 3.6639, "step": 261650 }, { "epoch": 2.9400695415760882, "grad_norm": 1.1481045484542847, "learning_rate": 8.244014784689533e-06, "loss": 3.6319, "step": 261700 }, { "epoch": 2.940631266746432, "grad_norm": 1.176296591758728, "learning_rate": 8.241767871386684e-06, "loss": 3.6737, "step": 261750 }, { "epoch": 2.9411929919167745, "grad_norm": 1.1154247522354126, "learning_rate": 8.239520958083833e-06, "loss": 3.5941, "step": 261800 }, { "epoch": 2.941754717087118, "grad_norm": 1.2286163568496704, "learning_rate": 8.237274044780982e-06, "loss": 3.6045, "step": 261850 }, { "epoch": 2.9423164422574613, "grad_norm": 1.1951974630355835, "learning_rate": 8.235027131478133e-06, "loss": 3.6135, "step": 261900 }, { "epoch": 2.9428781674278044, "grad_norm": 1.1343259811401367, "learning_rate": 8.232780218175283e-06, "loss": 3.6668, "step": 261950 }, { "epoch": 2.9434398925981475, "grad_norm": 1.4456795454025269, "learning_rate": 8.230533304872432e-06, "loss": 3.6074, "step": 262000 }, { "epoch": 2.9440016177684907, "grad_norm": 1.2335236072540283, "learning_rate": 8.228286391569581e-06, "loss": 3.5383, "step": 262050 }, { "epoch": 2.944563342938834, "grad_norm": 1.1757346391677856, "learning_rate": 8.226039478266732e-06, "loss": 3.6207, "step": 262100 }, { "epoch": 2.945125068109177, "grad_norm": 1.1784416437149048, "learning_rate": 8.223792564963882e-06, "loss": 3.7592, "step": 262150 }, { "epoch": 2.94568679327952, "grad_norm": 1.1022356748580933, "learning_rate": 8.221545651661031e-06, "loss": 3.5952, "step": 262200 }, { "epoch": 2.9462485184498632, "grad_norm": 1.3761985301971436, "learning_rate": 8.219298738358182e-06, "loss": 3.6662, "step": 262250 }, { "epoch": 2.9468102436202064, "grad_norm": 1.0535807609558105, "learning_rate": 8.217051825055331e-06, "loss": 3.608, "step": 262300 }, { "epoch": 2.9473719687905495, "grad_norm": 1.0880423784255981, "learning_rate": 8.21480491175248e-06, "loss": 3.6238, "step": 262350 }, { "epoch": 2.9479336939608927, "grad_norm": 0.931067705154419, "learning_rate": 8.21255799844963e-06, "loss": 3.6487, "step": 262400 }, { "epoch": 2.948495419131236, "grad_norm": 1.377273678779602, "learning_rate": 8.210311085146781e-06, "loss": 3.6544, "step": 262450 }, { "epoch": 2.949057144301579, "grad_norm": 1.1592832803726196, "learning_rate": 8.20806417184393e-06, "loss": 3.7145, "step": 262500 }, { "epoch": 2.949618869471922, "grad_norm": 1.6375360488891602, "learning_rate": 8.20581725854108e-06, "loss": 3.6314, "step": 262550 }, { "epoch": 2.9501805946422652, "grad_norm": 1.5016745328903198, "learning_rate": 8.20357034523823e-06, "loss": 3.6528, "step": 262600 }, { "epoch": 2.9507423198126084, "grad_norm": 1.2260786294937134, "learning_rate": 8.20132343193538e-06, "loss": 3.6993, "step": 262650 }, { "epoch": 2.9513040449829515, "grad_norm": 1.317764163017273, "learning_rate": 8.19907651863253e-06, "loss": 3.6427, "step": 262700 }, { "epoch": 2.9518657701532947, "grad_norm": 1.0656663179397583, "learning_rate": 8.196829605329679e-06, "loss": 3.7057, "step": 262750 }, { "epoch": 2.952427495323638, "grad_norm": 1.2142822742462158, "learning_rate": 8.19458269202683e-06, "loss": 3.7001, "step": 262800 }, { "epoch": 2.952989220493981, "grad_norm": 1.0361478328704834, "learning_rate": 8.192335778723979e-06, "loss": 3.585, "step": 262850 }, { "epoch": 2.9535509456643245, "grad_norm": 1.3071506023406982, "learning_rate": 8.190088865421128e-06, "loss": 3.6227, "step": 262900 }, { "epoch": 2.954112670834667, "grad_norm": 1.0098520517349243, "learning_rate": 8.18784195211828e-06, "loss": 3.6593, "step": 262950 }, { "epoch": 2.954674396005011, "grad_norm": 1.2244157791137695, "learning_rate": 8.185595038815428e-06, "loss": 3.661, "step": 263000 }, { "epoch": 2.9552361211753535, "grad_norm": 1.1741385459899902, "learning_rate": 8.183348125512578e-06, "loss": 3.625, "step": 263050 }, { "epoch": 2.955797846345697, "grad_norm": 1.1160554885864258, "learning_rate": 8.181101212209727e-06, "loss": 3.639, "step": 263100 }, { "epoch": 2.95635957151604, "grad_norm": 1.3024271726608276, "learning_rate": 8.178854298906878e-06, "loss": 3.6041, "step": 263150 }, { "epoch": 2.9569212966863834, "grad_norm": 1.1722395420074463, "learning_rate": 8.176607385604027e-06, "loss": 3.7005, "step": 263200 }, { "epoch": 2.9574830218567265, "grad_norm": 1.129657506942749, "learning_rate": 8.174360472301177e-06, "loss": 3.6384, "step": 263250 }, { "epoch": 2.9580447470270697, "grad_norm": 1.3723276853561401, "learning_rate": 8.172113558998326e-06, "loss": 3.6765, "step": 263300 }, { "epoch": 2.958606472197413, "grad_norm": 1.1832383871078491, "learning_rate": 8.169866645695477e-06, "loss": 3.6051, "step": 263350 }, { "epoch": 2.959168197367756, "grad_norm": 1.2455813884735107, "learning_rate": 8.167619732392626e-06, "loss": 3.6145, "step": 263400 }, { "epoch": 2.959729922538099, "grad_norm": 1.0869892835617065, "learning_rate": 8.165372819089776e-06, "loss": 3.625, "step": 263450 }, { "epoch": 2.960291647708442, "grad_norm": 0.987187922000885, "learning_rate": 8.163125905786927e-06, "loss": 3.6918, "step": 263500 }, { "epoch": 2.9608533728787854, "grad_norm": 1.2373844385147095, "learning_rate": 8.160878992484076e-06, "loss": 3.6743, "step": 263550 }, { "epoch": 2.9614150980491285, "grad_norm": 1.2048543691635132, "learning_rate": 8.158632079181225e-06, "loss": 3.6173, "step": 263600 }, { "epoch": 2.9619768232194716, "grad_norm": 1.2274523973464966, "learning_rate": 8.156385165878375e-06, "loss": 3.6511, "step": 263650 }, { "epoch": 2.962538548389815, "grad_norm": 1.2022584676742554, "learning_rate": 8.154138252575526e-06, "loss": 3.5976, "step": 263700 }, { "epoch": 2.963100273560158, "grad_norm": 1.2064024209976196, "learning_rate": 8.151891339272675e-06, "loss": 3.5609, "step": 263750 }, { "epoch": 2.963661998730501, "grad_norm": 1.0156909227371216, "learning_rate": 8.149644425969824e-06, "loss": 3.6531, "step": 263800 }, { "epoch": 2.964223723900844, "grad_norm": 1.1341277360916138, "learning_rate": 8.147397512666975e-06, "loss": 3.6602, "step": 263850 }, { "epoch": 2.9647854490711874, "grad_norm": 1.2537122964859009, "learning_rate": 8.145195537630181e-06, "loss": 3.6318, "step": 263900 }, { "epoch": 2.9653471742415305, "grad_norm": 1.1408644914627075, "learning_rate": 8.142948624327332e-06, "loss": 3.5669, "step": 263950 }, { "epoch": 2.9659088994118736, "grad_norm": 1.0433231592178345, "learning_rate": 8.140701711024482e-06, "loss": 3.6918, "step": 264000 }, { "epoch": 2.9664706245822168, "grad_norm": 1.2975996732711792, "learning_rate": 8.138454797721631e-06, "loss": 3.6188, "step": 264050 }, { "epoch": 2.96703234975256, "grad_norm": 1.2062848806381226, "learning_rate": 8.13620788441878e-06, "loss": 3.6281, "step": 264100 }, { "epoch": 2.967594074922903, "grad_norm": 1.1855512857437134, "learning_rate": 8.133960971115931e-06, "loss": 3.6488, "step": 264150 }, { "epoch": 2.968155800093246, "grad_norm": 1.1772892475128174, "learning_rate": 8.13171405781308e-06, "loss": 3.6896, "step": 264200 }, { "epoch": 2.96871752526359, "grad_norm": 1.1771161556243896, "learning_rate": 8.12946714451023e-06, "loss": 3.5914, "step": 264250 }, { "epoch": 2.9692792504339325, "grad_norm": 1.2375949621200562, "learning_rate": 8.12722023120738e-06, "loss": 3.5846, "step": 264300 }, { "epoch": 2.969840975604276, "grad_norm": 1.2629023790359497, "learning_rate": 8.12497331790453e-06, "loss": 3.6376, "step": 264350 }, { "epoch": 2.9704027007746188, "grad_norm": 1.022668480873108, "learning_rate": 8.12272640460168e-06, "loss": 3.6028, "step": 264400 }, { "epoch": 2.9709644259449624, "grad_norm": 1.0815987586975098, "learning_rate": 8.120479491298829e-06, "loss": 3.6658, "step": 264450 }, { "epoch": 2.9715261511153055, "grad_norm": 1.2520629167556763, "learning_rate": 8.11823257799598e-06, "loss": 3.6826, "step": 264500 }, { "epoch": 2.9720878762856486, "grad_norm": 1.1516664028167725, "learning_rate": 8.115985664693129e-06, "loss": 3.647, "step": 264550 }, { "epoch": 2.9726496014559918, "grad_norm": 1.0715162754058838, "learning_rate": 8.113738751390278e-06, "loss": 3.5649, "step": 264600 }, { "epoch": 2.973211326626335, "grad_norm": 1.186252236366272, "learning_rate": 8.111491838087428e-06, "loss": 3.6375, "step": 264650 }, { "epoch": 2.973773051796678, "grad_norm": 1.1471000909805298, "learning_rate": 8.109244924784579e-06, "loss": 3.634, "step": 264700 }, { "epoch": 2.974334776967021, "grad_norm": 1.2380133867263794, "learning_rate": 8.106998011481728e-06, "loss": 3.6771, "step": 264750 }, { "epoch": 2.9748965021373643, "grad_norm": 1.0482748746871948, "learning_rate": 8.104751098178877e-06, "loss": 3.628, "step": 264800 }, { "epoch": 2.9754582273077075, "grad_norm": 1.1627472639083862, "learning_rate": 8.102504184876028e-06, "loss": 3.665, "step": 264850 }, { "epoch": 2.9760199524780506, "grad_norm": 1.2711812257766724, "learning_rate": 8.100257271573178e-06, "loss": 3.6635, "step": 264900 }, { "epoch": 2.9765816776483938, "grad_norm": 1.2783652544021606, "learning_rate": 8.098010358270327e-06, "loss": 3.6356, "step": 264950 }, { "epoch": 2.977143402818737, "grad_norm": 1.17335045337677, "learning_rate": 8.095763444967476e-06, "loss": 3.596, "step": 265000 }, { "epoch": 2.97770512798908, "grad_norm": 1.1174033880233765, "learning_rate": 8.093516531664627e-06, "loss": 3.6033, "step": 265050 }, { "epoch": 2.978266853159423, "grad_norm": 1.2793207168579102, "learning_rate": 8.091269618361777e-06, "loss": 3.6049, "step": 265100 }, { "epoch": 2.9788285783297663, "grad_norm": 1.1536370515823364, "learning_rate": 8.089022705058926e-06, "loss": 3.6841, "step": 265150 }, { "epoch": 2.9793903035001095, "grad_norm": 1.2100703716278076, "learning_rate": 8.086775791756077e-06, "loss": 3.655, "step": 265200 }, { "epoch": 2.9799520286704526, "grad_norm": 1.0454168319702148, "learning_rate": 8.084528878453226e-06, "loss": 3.598, "step": 265250 }, { "epoch": 2.9805137538407958, "grad_norm": 1.0071556568145752, "learning_rate": 8.082281965150376e-06, "loss": 3.6403, "step": 265300 }, { "epoch": 2.981075479011139, "grad_norm": 1.1551620960235596, "learning_rate": 8.080035051847525e-06, "loss": 3.643, "step": 265350 }, { "epoch": 2.981637204181482, "grad_norm": 1.1440470218658447, "learning_rate": 8.077788138544674e-06, "loss": 3.6378, "step": 265400 }, { "epoch": 2.982198929351825, "grad_norm": 1.1203893423080444, "learning_rate": 8.075541225241824e-06, "loss": 3.5999, "step": 265450 }, { "epoch": 2.9827606545221683, "grad_norm": 1.146336317062378, "learning_rate": 8.073294311938975e-06, "loss": 3.6543, "step": 265500 }, { "epoch": 2.9833223796925115, "grad_norm": 1.1753056049346924, "learning_rate": 8.071047398636124e-06, "loss": 3.6093, "step": 265550 }, { "epoch": 2.983884104862855, "grad_norm": 1.254305124282837, "learning_rate": 8.068800485333273e-06, "loss": 3.6747, "step": 265600 }, { "epoch": 2.9844458300331977, "grad_norm": 1.2076882123947144, "learning_rate": 8.066553572030423e-06, "loss": 3.661, "step": 265650 }, { "epoch": 2.9850075552035413, "grad_norm": 1.1521852016448975, "learning_rate": 8.064306658727574e-06, "loss": 3.6043, "step": 265700 }, { "epoch": 2.985569280373884, "grad_norm": 1.5098720788955688, "learning_rate": 8.062059745424723e-06, "loss": 3.6875, "step": 265750 }, { "epoch": 2.9861310055442276, "grad_norm": 1.168294906616211, "learning_rate": 8.059812832121872e-06, "loss": 3.6478, "step": 265800 }, { "epoch": 2.9866927307145708, "grad_norm": 1.1249010562896729, "learning_rate": 8.057565918819023e-06, "loss": 3.6628, "step": 265850 }, { "epoch": 2.987254455884914, "grad_norm": 1.4948920011520386, "learning_rate": 8.055319005516173e-06, "loss": 3.5833, "step": 265900 }, { "epoch": 2.987816181055257, "grad_norm": 1.4017730951309204, "learning_rate": 8.053072092213322e-06, "loss": 3.6423, "step": 265950 }, { "epoch": 2.9883779062256, "grad_norm": 1.1269339323043823, "learning_rate": 8.050825178910471e-06, "loss": 3.6902, "step": 266000 }, { "epoch": 2.9889396313959433, "grad_norm": 1.5000494718551636, "learning_rate": 8.048578265607622e-06, "loss": 3.5804, "step": 266050 }, { "epoch": 2.9895013565662865, "grad_norm": 1.0375537872314453, "learning_rate": 8.046331352304772e-06, "loss": 3.6515, "step": 266100 }, { "epoch": 2.9900630817366296, "grad_norm": 1.1606764793395996, "learning_rate": 8.04408443900192e-06, "loss": 3.6638, "step": 266150 }, { "epoch": 2.9906248069069727, "grad_norm": 1.0183250904083252, "learning_rate": 8.041837525699072e-06, "loss": 3.6846, "step": 266200 }, { "epoch": 2.991186532077316, "grad_norm": 1.2175195217132568, "learning_rate": 8.039590612396221e-06, "loss": 3.6546, "step": 266250 }, { "epoch": 2.991748257247659, "grad_norm": 1.2877438068389893, "learning_rate": 8.03734369909337e-06, "loss": 3.6357, "step": 266300 }, { "epoch": 2.992309982418002, "grad_norm": 1.2034518718719482, "learning_rate": 8.03509678579052e-06, "loss": 3.6697, "step": 266350 }, { "epoch": 2.9928717075883453, "grad_norm": 1.2930511236190796, "learning_rate": 8.03284987248767e-06, "loss": 3.6098, "step": 266400 }, { "epoch": 2.9934334327586885, "grad_norm": 1.302494764328003, "learning_rate": 8.03060295918482e-06, "loss": 3.6286, "step": 266450 }, { "epoch": 2.9939951579290316, "grad_norm": 1.1157937049865723, "learning_rate": 8.02835604588197e-06, "loss": 3.5678, "step": 266500 }, { "epoch": 2.9945568830993747, "grad_norm": 1.2050288915634155, "learning_rate": 8.02610913257912e-06, "loss": 3.5746, "step": 266550 }, { "epoch": 2.995118608269718, "grad_norm": 1.3136342763900757, "learning_rate": 8.02386221927627e-06, "loss": 3.6396, "step": 266600 }, { "epoch": 2.995680333440061, "grad_norm": 1.1844359636306763, "learning_rate": 8.021615305973419e-06, "loss": 3.66, "step": 266650 }, { "epoch": 2.996242058610404, "grad_norm": 1.1723449230194092, "learning_rate": 8.019368392670568e-06, "loss": 3.5934, "step": 266700 }, { "epoch": 2.9968037837807473, "grad_norm": 1.1537070274353027, "learning_rate": 8.01712147936772e-06, "loss": 3.6387, "step": 266750 }, { "epoch": 2.9973655089510904, "grad_norm": 1.2313581705093384, "learning_rate": 8.014874566064869e-06, "loss": 3.6024, "step": 266800 }, { "epoch": 2.9979272341214336, "grad_norm": 1.2923564910888672, "learning_rate": 8.012627652762018e-06, "loss": 3.6989, "step": 266850 }, { "epoch": 2.9984889592917767, "grad_norm": 1.2889326810836792, "learning_rate": 8.010380739459167e-06, "loss": 3.7019, "step": 266900 }, { "epoch": 2.9990506844621203, "grad_norm": 1.8831459283828735, "learning_rate": 8.008133826156318e-06, "loss": 3.679, "step": 266950 }, { "epoch": 2.999612409632463, "grad_norm": 1.1464004516601562, "learning_rate": 8.005886912853468e-06, "loss": 3.6694, "step": 267000 }, { "epoch": 2.9999943827482967, "eval_bleu": 6.0595, "eval_gen_len": 21.5704, "eval_loss": 3.4992141723632812, "eval_runtime": 25714.3846, "eval_samples_per_second": 13.846, "eval_steps_per_second": 0.865, "step": 267034 } ], "logging_steps": 50, "max_steps": 445055, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 7.847598620639232e+16, "train_batch_size": 16, "trial_name": null, "trial_params": null }