diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,37443 @@ +{ + "best_metric": 3.4992141723632812, + "best_model_checkpoint": "en-to-lg-nllb-cc/checkpoint-267034", + "epoch": 2.9999943827482967, + "eval_steps": 500, + "global_step": 267034, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0005617251703431579, + "grad_norm": 1.403859257698059, + "learning_rate": 1.999775308669715e-05, + "loss": 4.8949, + "step": 50 + }, + { + "epoch": 0.0011234503406863158, + "grad_norm": 1.5137478113174438, + "learning_rate": 1.9995596049926417e-05, + "loss": 4.7613, + "step": 100 + }, + { + "epoch": 0.0016851755110294736, + "grad_norm": 0.933453381061554, + "learning_rate": 1.9993349136623566e-05, + "loss": 4.5466, + "step": 150 + }, + { + "epoch": 0.0022469006813726315, + "grad_norm": 0.7829797863960266, + "learning_rate": 1.9991102223320716e-05, + "loss": 4.3399, + "step": 200 + }, + { + "epoch": 0.0028086258517157894, + "grad_norm": 0.8315831422805786, + "learning_rate": 1.9988855310017865e-05, + "loss": 4.3509, + "step": 250 + }, + { + "epoch": 0.0033703510220589473, + "grad_norm": 0.8067842125892639, + "learning_rate": 1.9986608396715014e-05, + "loss": 4.3661, + "step": 300 + }, + { + "epoch": 0.003932076192402105, + "grad_norm": 0.5773376822471619, + "learning_rate": 1.9984361483412163e-05, + "loss": 4.1911, + "step": 350 + }, + { + "epoch": 0.004493801362745263, + "grad_norm": 0.6907451748847961, + "learning_rate": 1.9982114570109313e-05, + "loss": 4.2254, + "step": 400 + }, + { + "epoch": 0.005055526533088421, + "grad_norm": 0.6285920143127441, + "learning_rate": 1.9979867656806465e-05, + "loss": 4.211, + "step": 450 + }, + { + "epoch": 0.005617251703431579, + "grad_norm": 0.6995010375976562, + "learning_rate": 1.9977620743503615e-05, + "loss": 4.1618, + "step": 500 + }, + { + "epoch": 0.006178976873774737, + "grad_norm": 0.7596259117126465, + "learning_rate": 1.9975373830200764e-05, + "loss": 4.2133, + "step": 550 + }, + { + "epoch": 0.0067407020441178946, + "grad_norm": 0.6245077848434448, + "learning_rate": 1.9973126916897913e-05, + "loss": 4.1891, + "step": 600 + }, + { + "epoch": 0.0073024272144610524, + "grad_norm": 0.6988118886947632, + "learning_rate": 1.9970880003595063e-05, + "loss": 4.1388, + "step": 650 + }, + { + "epoch": 0.00786415238480421, + "grad_norm": 0.6634756922721863, + "learning_rate": 1.9968633090292212e-05, + "loss": 4.1258, + "step": 700 + }, + { + "epoch": 0.008425877555147369, + "grad_norm": 0.5972195863723755, + "learning_rate": 1.996638617698936e-05, + "loss": 4.1774, + "step": 750 + }, + { + "epoch": 0.008987602725490526, + "grad_norm": 0.7192025184631348, + "learning_rate": 1.9964139263686514e-05, + "loss": 4.1227, + "step": 800 + }, + { + "epoch": 0.009549327895833685, + "grad_norm": 0.6875787377357483, + "learning_rate": 1.9961892350383663e-05, + "loss": 4.076, + "step": 850 + }, + { + "epoch": 0.010111053066176842, + "grad_norm": 0.7681994438171387, + "learning_rate": 1.9959645437080813e-05, + "loss": 4.1251, + "step": 900 + }, + { + "epoch": 0.01067277823652, + "grad_norm": 0.7524916529655457, + "learning_rate": 1.9957398523777962e-05, + "loss": 4.0552, + "step": 950 + }, + { + "epoch": 0.011234503406863158, + "grad_norm": 0.7998524904251099, + "learning_rate": 1.995515161047511e-05, + "loss": 4.1645, + "step": 1000 + }, + { + "epoch": 0.011796228577206316, + "grad_norm": 0.8324928879737854, + "learning_rate": 1.995290469717226e-05, + "loss": 4.0986, + "step": 1050 + }, + { + "epoch": 0.012357953747549473, + "grad_norm": 0.9623536467552185, + "learning_rate": 1.995065778386941e-05, + "loss": 4.1097, + "step": 1100 + }, + { + "epoch": 0.012919678917892632, + "grad_norm": 0.6794806122779846, + "learning_rate": 1.9948410870566563e-05, + "loss": 4.0934, + "step": 1150 + }, + { + "epoch": 0.013481404088235789, + "grad_norm": 0.8683093786239624, + "learning_rate": 1.9946163957263712e-05, + "loss": 4.0484, + "step": 1200 + }, + { + "epoch": 0.014043129258578948, + "grad_norm": 0.8103470802307129, + "learning_rate": 1.994391704396086e-05, + "loss": 4.1611, + "step": 1250 + }, + { + "epoch": 0.014604854428922105, + "grad_norm": 0.779623806476593, + "learning_rate": 1.994167013065801e-05, + "loss": 4.0097, + "step": 1300 + }, + { + "epoch": 0.015166579599265264, + "grad_norm": 0.7256409525871277, + "learning_rate": 1.993942321735516e-05, + "loss": 4.0929, + "step": 1350 + }, + { + "epoch": 0.01572830476960842, + "grad_norm": 0.8892433643341064, + "learning_rate": 1.993717630405231e-05, + "loss": 4.0195, + "step": 1400 + }, + { + "epoch": 0.01629002993995158, + "grad_norm": 0.8380836248397827, + "learning_rate": 1.993492939074946e-05, + "loss": 4.0752, + "step": 1450 + }, + { + "epoch": 0.016851755110294738, + "grad_norm": 0.803108811378479, + "learning_rate": 1.993268247744661e-05, + "loss": 4.0432, + "step": 1500 + }, + { + "epoch": 0.017413480280637893, + "grad_norm": 0.8453735113143921, + "learning_rate": 1.993043556414376e-05, + "loss": 4.119, + "step": 1550 + }, + { + "epoch": 0.017975205450981052, + "grad_norm": 0.8968884348869324, + "learning_rate": 1.992818865084091e-05, + "loss": 4.0632, + "step": 1600 + }, + { + "epoch": 0.01853693062132421, + "grad_norm": 0.7678434252738953, + "learning_rate": 1.992594173753806e-05, + "loss": 3.9581, + "step": 1650 + }, + { + "epoch": 0.01909865579166737, + "grad_norm": 0.7526470422744751, + "learning_rate": 1.992369482423521e-05, + "loss": 4.0942, + "step": 1700 + }, + { + "epoch": 0.01966038096201053, + "grad_norm": 0.7269778847694397, + "learning_rate": 1.9921447910932358e-05, + "loss": 4.0031, + "step": 1750 + }, + { + "epoch": 0.020222106132353684, + "grad_norm": 0.8589780926704407, + "learning_rate": 1.9919200997629507e-05, + "loss": 4.0544, + "step": 1800 + }, + { + "epoch": 0.020783831302696842, + "grad_norm": 0.7709614038467407, + "learning_rate": 1.991695408432666e-05, + "loss": 4.0253, + "step": 1850 + }, + { + "epoch": 0.02134555647304, + "grad_norm": 1.058897614479065, + "learning_rate": 1.991470717102381e-05, + "loss": 4.0436, + "step": 1900 + }, + { + "epoch": 0.02190728164338316, + "grad_norm": 0.7720006704330444, + "learning_rate": 1.991246025772096e-05, + "loss": 4.0896, + "step": 1950 + }, + { + "epoch": 0.022469006813726315, + "grad_norm": 0.8622137904167175, + "learning_rate": 1.9910213344418108e-05, + "loss": 4.1142, + "step": 2000 + }, + { + "epoch": 0.023030731984069474, + "grad_norm": 0.7667924761772156, + "learning_rate": 1.9907966431115257e-05, + "loss": 4.0354, + "step": 2050 + }, + { + "epoch": 0.023592457154412633, + "grad_norm": 0.7575067281723022, + "learning_rate": 1.9905719517812406e-05, + "loss": 4.0697, + "step": 2100 + }, + { + "epoch": 0.02415418232475579, + "grad_norm": 0.6507667899131775, + "learning_rate": 1.9903472604509556e-05, + "loss": 4.0695, + "step": 2150 + }, + { + "epoch": 0.024715907495098947, + "grad_norm": 0.7584772109985352, + "learning_rate": 1.990122569120671e-05, + "loss": 4.0884, + "step": 2200 + }, + { + "epoch": 0.025277632665442105, + "grad_norm": 0.7901142835617065, + "learning_rate": 1.9898978777903858e-05, + "loss": 3.9693, + "step": 2250 + }, + { + "epoch": 0.025839357835785264, + "grad_norm": 0.8463560342788696, + "learning_rate": 1.9896731864601007e-05, + "loss": 4.0007, + "step": 2300 + }, + { + "epoch": 0.026401083006128423, + "grad_norm": 0.8583828806877136, + "learning_rate": 1.9894484951298156e-05, + "loss": 4.1057, + "step": 2350 + }, + { + "epoch": 0.026962808176471578, + "grad_norm": 0.8160439133644104, + "learning_rate": 1.9892238037995306e-05, + "loss": 4.0635, + "step": 2400 + }, + { + "epoch": 0.027524533346814737, + "grad_norm": 0.8757029175758362, + "learning_rate": 1.9889991124692455e-05, + "loss": 4.1076, + "step": 2450 + }, + { + "epoch": 0.028086258517157896, + "grad_norm": 0.9236558079719543, + "learning_rate": 1.9887744211389604e-05, + "loss": 3.9999, + "step": 2500 + }, + { + "epoch": 0.028647983687501054, + "grad_norm": 0.866594135761261, + "learning_rate": 1.9885497298086757e-05, + "loss": 3.9782, + "step": 2550 + }, + { + "epoch": 0.02920970885784421, + "grad_norm": 0.9537388682365417, + "learning_rate": 1.9883250384783906e-05, + "loss": 4.0354, + "step": 2600 + }, + { + "epoch": 0.02977143402818737, + "grad_norm": 0.7946441769599915, + "learning_rate": 1.9881003471481056e-05, + "loss": 4.0499, + "step": 2650 + }, + { + "epoch": 0.030333159198530527, + "grad_norm": 0.7472932934761047, + "learning_rate": 1.9878756558178205e-05, + "loss": 3.9899, + "step": 2700 + }, + { + "epoch": 0.030894884368873686, + "grad_norm": 0.7326238751411438, + "learning_rate": 1.9876509644875354e-05, + "loss": 4.0243, + "step": 2750 + }, + { + "epoch": 0.03145660953921684, + "grad_norm": 0.9543971419334412, + "learning_rate": 1.9874262731572504e-05, + "loss": 4.0569, + "step": 2800 + }, + { + "epoch": 0.032018334709560003, + "grad_norm": 0.7659034132957458, + "learning_rate": 1.9872015818269653e-05, + "loss": 3.9717, + "step": 2850 + }, + { + "epoch": 0.03258005987990316, + "grad_norm": 0.8442956209182739, + "learning_rate": 1.9869768904966806e-05, + "loss": 4.1692, + "step": 2900 + }, + { + "epoch": 0.033141785050246314, + "grad_norm": 0.7701689600944519, + "learning_rate": 1.9867521991663955e-05, + "loss": 3.9567, + "step": 2950 + }, + { + "epoch": 0.033703510220589476, + "grad_norm": 11.462121963500977, + "learning_rate": 1.9865275078361104e-05, + "loss": 4.0557, + "step": 3000 + }, + { + "epoch": 0.03426523539093263, + "grad_norm": 0.9673064947128296, + "learning_rate": 1.9863028165058254e-05, + "loss": 3.983, + "step": 3050 + }, + { + "epoch": 0.03482696056127579, + "grad_norm": 0.8910813927650452, + "learning_rate": 1.9860781251755403e-05, + "loss": 4.055, + "step": 3100 + }, + { + "epoch": 0.03538868573161895, + "grad_norm": 0.8741032481193542, + "learning_rate": 1.9858534338452552e-05, + "loss": 3.9814, + "step": 3150 + }, + { + "epoch": 0.035950410901962104, + "grad_norm": 0.789903461933136, + "learning_rate": 1.98562874251497e-05, + "loss": 3.9904, + "step": 3200 + }, + { + "epoch": 0.03651213607230527, + "grad_norm": 0.7773197293281555, + "learning_rate": 1.9854040511846854e-05, + "loss": 4.0051, + "step": 3250 + }, + { + "epoch": 0.03707386124264842, + "grad_norm": 0.8257423043251038, + "learning_rate": 1.9851793598544004e-05, + "loss": 4.0133, + "step": 3300 + }, + { + "epoch": 0.03763558641299158, + "grad_norm": 0.946387529373169, + "learning_rate": 1.9849546685241153e-05, + "loss": 4.0593, + "step": 3350 + }, + { + "epoch": 0.03819731158333474, + "grad_norm": 0.8175517916679382, + "learning_rate": 1.9847299771938302e-05, + "loss": 4.007, + "step": 3400 + }, + { + "epoch": 0.038759036753677895, + "grad_norm": 1.033706545829773, + "learning_rate": 1.984505285863545e-05, + "loss": 4.0251, + "step": 3450 + }, + { + "epoch": 0.03932076192402106, + "grad_norm": 0.8714016675949097, + "learning_rate": 1.98428059453326e-05, + "loss": 4.0367, + "step": 3500 + }, + { + "epoch": 0.03988248709436421, + "grad_norm": 1.0509799718856812, + "learning_rate": 1.984055903202975e-05, + "loss": 4.0318, + "step": 3550 + }, + { + "epoch": 0.04044421226470737, + "grad_norm": 0.8878036737442017, + "learning_rate": 1.98383121187269e-05, + "loss": 4.0543, + "step": 3600 + }, + { + "epoch": 0.04100593743505053, + "grad_norm": 0.7888545989990234, + "learning_rate": 1.9836065205424052e-05, + "loss": 3.9551, + "step": 3650 + }, + { + "epoch": 0.041567662605393685, + "grad_norm": 0.8576290011405945, + "learning_rate": 1.98338182921212e-05, + "loss": 4.0179, + "step": 3700 + }, + { + "epoch": 0.04212938777573684, + "grad_norm": 0.8358418345451355, + "learning_rate": 1.983157137881835e-05, + "loss": 4.0658, + "step": 3750 + }, + { + "epoch": 0.04269111294608, + "grad_norm": 0.9385461807250977, + "learning_rate": 1.98293244655155e-05, + "loss": 4.051, + "step": 3800 + }, + { + "epoch": 0.04325283811642316, + "grad_norm": 0.9508741497993469, + "learning_rate": 1.982707755221265e-05, + "loss": 4.0268, + "step": 3850 + }, + { + "epoch": 0.04381456328676632, + "grad_norm": 0.8164873719215393, + "learning_rate": 1.98248306389098e-05, + "loss": 4.009, + "step": 3900 + }, + { + "epoch": 0.044376288457109475, + "grad_norm": 0.986486554145813, + "learning_rate": 1.9822583725606948e-05, + "loss": 4.0036, + "step": 3950 + }, + { + "epoch": 0.04493801362745263, + "grad_norm": 0.9176609516143799, + "learning_rate": 1.98203368123041e-05, + "loss": 4.0901, + "step": 4000 + }, + { + "epoch": 0.04549973879779579, + "grad_norm": 0.9912000298500061, + "learning_rate": 1.981808989900125e-05, + "loss": 3.9454, + "step": 4050 + }, + { + "epoch": 0.04606146396813895, + "grad_norm": 1.068574070930481, + "learning_rate": 1.98158429856984e-05, + "loss": 4.1076, + "step": 4100 + }, + { + "epoch": 0.0466231891384821, + "grad_norm": 0.7717074751853943, + "learning_rate": 1.981359607239555e-05, + "loss": 4.0415, + "step": 4150 + }, + { + "epoch": 0.047184914308825265, + "grad_norm": 0.8814721703529358, + "learning_rate": 1.9811349159092698e-05, + "loss": 3.9491, + "step": 4200 + }, + { + "epoch": 0.04774663947916842, + "grad_norm": 0.8565316796302795, + "learning_rate": 1.9809102245789847e-05, + "loss": 4.0645, + "step": 4250 + }, + { + "epoch": 0.04830836464951158, + "grad_norm": 0.9432961940765381, + "learning_rate": 1.9806855332486997e-05, + "loss": 4.0238, + "step": 4300 + }, + { + "epoch": 0.04887008981985474, + "grad_norm": 1.158420205116272, + "learning_rate": 1.9804653357450206e-05, + "loss": 3.9297, + "step": 4350 + }, + { + "epoch": 0.04943181499019789, + "grad_norm": 0.9152107834815979, + "learning_rate": 1.9802406444147355e-05, + "loss": 3.8737, + "step": 4400 + }, + { + "epoch": 0.049993540160541056, + "grad_norm": 0.9017531871795654, + "learning_rate": 1.9800159530844505e-05, + "loss": 4.0031, + "step": 4450 + }, + { + "epoch": 0.05055526533088421, + "grad_norm": 3.250427007675171, + "learning_rate": 1.9797912617541654e-05, + "loss": 3.9004, + "step": 4500 + }, + { + "epoch": 0.051116990501227366, + "grad_norm": 1.2878793478012085, + "learning_rate": 1.9795665704238803e-05, + "loss": 3.9565, + "step": 4550 + }, + { + "epoch": 0.05167871567157053, + "grad_norm": 1.0191071033477783, + "learning_rate": 1.9793418790935953e-05, + "loss": 3.9954, + "step": 4600 + }, + { + "epoch": 0.052240440841913684, + "grad_norm": 0.8436045050621033, + "learning_rate": 1.9791171877633105e-05, + "loss": 4.0165, + "step": 4650 + }, + { + "epoch": 0.052802166012256846, + "grad_norm": 1.0180222988128662, + "learning_rate": 1.9788924964330255e-05, + "loss": 3.9563, + "step": 4700 + }, + { + "epoch": 0.0533638911826, + "grad_norm": 0.9640880227088928, + "learning_rate": 1.9786678051027404e-05, + "loss": 3.9195, + "step": 4750 + }, + { + "epoch": 0.053925616352943156, + "grad_norm": 1.0035834312438965, + "learning_rate": 1.9784431137724553e-05, + "loss": 3.9712, + "step": 4800 + }, + { + "epoch": 0.05448734152328632, + "grad_norm": 0.8614216446876526, + "learning_rate": 1.9782184224421703e-05, + "loss": 3.969, + "step": 4850 + }, + { + "epoch": 0.055049066693629474, + "grad_norm": 1.079421877861023, + "learning_rate": 1.9779937311118852e-05, + "loss": 3.9919, + "step": 4900 + }, + { + "epoch": 0.055610791863972636, + "grad_norm": 0.8971319198608398, + "learning_rate": 1.9777690397816e-05, + "loss": 4.1197, + "step": 4950 + }, + { + "epoch": 0.05617251703431579, + "grad_norm": 0.8105068802833557, + "learning_rate": 1.9775443484513154e-05, + "loss": 3.9572, + "step": 5000 + }, + { + "epoch": 0.05673424220465895, + "grad_norm": 1.0254584550857544, + "learning_rate": 1.9773196571210303e-05, + "loss": 3.9764, + "step": 5050 + }, + { + "epoch": 0.05729596737500211, + "grad_norm": 0.8980420231819153, + "learning_rate": 1.9770949657907453e-05, + "loss": 3.9426, + "step": 5100 + }, + { + "epoch": 0.057857692545345264, + "grad_norm": 1.1586838960647583, + "learning_rate": 1.9768702744604602e-05, + "loss": 3.9144, + "step": 5150 + }, + { + "epoch": 0.05841941771568842, + "grad_norm": 0.858310341835022, + "learning_rate": 1.976645583130175e-05, + "loss": 3.9721, + "step": 5200 + }, + { + "epoch": 0.05898114288603158, + "grad_norm": 0.8306008577346802, + "learning_rate": 1.97642089179989e-05, + "loss": 3.9499, + "step": 5250 + }, + { + "epoch": 0.05954286805637474, + "grad_norm": 0.9015403985977173, + "learning_rate": 1.976196200469605e-05, + "loss": 3.9362, + "step": 5300 + }, + { + "epoch": 0.0601045932267179, + "grad_norm": 0.7949343323707581, + "learning_rate": 1.9759715091393203e-05, + "loss": 3.9575, + "step": 5350 + }, + { + "epoch": 0.060666318397061055, + "grad_norm": 1.0842740535736084, + "learning_rate": 1.9757468178090352e-05, + "loss": 3.9215, + "step": 5400 + }, + { + "epoch": 0.06122804356740421, + "grad_norm": 0.8949707746505737, + "learning_rate": 1.97552212647875e-05, + "loss": 3.9472, + "step": 5450 + }, + { + "epoch": 0.06178976873774737, + "grad_norm": 1.0328564643859863, + "learning_rate": 1.975297435148465e-05, + "loss": 4.0258, + "step": 5500 + }, + { + "epoch": 0.06235149390809053, + "grad_norm": 0.9843789339065552, + "learning_rate": 1.97507274381818e-05, + "loss": 3.9164, + "step": 5550 + }, + { + "epoch": 0.06291321907843368, + "grad_norm": 1.3258886337280273, + "learning_rate": 1.974848052487895e-05, + "loss": 4.0296, + "step": 5600 + }, + { + "epoch": 0.06347494424877684, + "grad_norm": 0.9008415341377258, + "learning_rate": 1.97462336115761e-05, + "loss": 3.9413, + "step": 5650 + }, + { + "epoch": 0.06403666941912001, + "grad_norm": 0.7957488298416138, + "learning_rate": 1.974398669827325e-05, + "loss": 4.0003, + "step": 5700 + }, + { + "epoch": 0.06459839458946316, + "grad_norm": 0.9556325078010559, + "learning_rate": 1.97417397849704e-05, + "loss": 3.9204, + "step": 5750 + }, + { + "epoch": 0.06516011975980632, + "grad_norm": 0.8846645951271057, + "learning_rate": 1.973949287166755e-05, + "loss": 3.9414, + "step": 5800 + }, + { + "epoch": 0.06572184493014947, + "grad_norm": 0.9160873889923096, + "learning_rate": 1.97372459583647e-05, + "loss": 3.9641, + "step": 5850 + }, + { + "epoch": 0.06628357010049263, + "grad_norm": 0.9347891807556152, + "learning_rate": 1.973499904506185e-05, + "loss": 4.0136, + "step": 5900 + }, + { + "epoch": 0.0668452952708358, + "grad_norm": 0.8801557421684265, + "learning_rate": 1.9732752131758998e-05, + "loss": 4.0376, + "step": 5950 + }, + { + "epoch": 0.06740702044117895, + "grad_norm": 0.849269688129425, + "learning_rate": 1.9730505218456147e-05, + "loss": 3.9612, + "step": 6000 + }, + { + "epoch": 0.06796874561152211, + "grad_norm": 1.066200852394104, + "learning_rate": 1.97282583051533e-05, + "loss": 4.0119, + "step": 6050 + }, + { + "epoch": 0.06853047078186526, + "grad_norm": 0.9853302240371704, + "learning_rate": 1.972601139185045e-05, + "loss": 3.913, + "step": 6100 + }, + { + "epoch": 0.06909219595220842, + "grad_norm": 0.849815309047699, + "learning_rate": 1.97237644785476e-05, + "loss": 3.9944, + "step": 6150 + }, + { + "epoch": 0.06965392112255157, + "grad_norm": 0.7928768396377563, + "learning_rate": 1.9721517565244748e-05, + "loss": 3.925, + "step": 6200 + }, + { + "epoch": 0.07021564629289474, + "grad_norm": 1.0310895442962646, + "learning_rate": 1.9719270651941897e-05, + "loss": 3.9194, + "step": 6250 + }, + { + "epoch": 0.0707773714632379, + "grad_norm": 1.1710147857666016, + "learning_rate": 1.9717023738639046e-05, + "loss": 3.9733, + "step": 6300 + }, + { + "epoch": 0.07133909663358105, + "grad_norm": 0.8295989036560059, + "learning_rate": 1.9714776825336196e-05, + "loss": 3.9791, + "step": 6350 + }, + { + "epoch": 0.07190082180392421, + "grad_norm": 0.7764708995819092, + "learning_rate": 1.971252991203335e-05, + "loss": 3.933, + "step": 6400 + }, + { + "epoch": 0.07246254697426736, + "grad_norm": 0.9420117139816284, + "learning_rate": 1.9710282998730498e-05, + "loss": 3.9969, + "step": 6450 + }, + { + "epoch": 0.07302427214461053, + "grad_norm": 0.956198513507843, + "learning_rate": 1.9708036085427647e-05, + "loss": 3.9355, + "step": 6500 + }, + { + "epoch": 0.07358599731495369, + "grad_norm": 0.965861439704895, + "learning_rate": 1.9705789172124796e-05, + "loss": 3.8971, + "step": 6550 + }, + { + "epoch": 0.07414772248529684, + "grad_norm": 1.030989408493042, + "learning_rate": 1.9703542258821946e-05, + "loss": 3.994, + "step": 6600 + }, + { + "epoch": 0.07470944765564, + "grad_norm": 0.9188275933265686, + "learning_rate": 1.9701295345519095e-05, + "loss": 3.9572, + "step": 6650 + }, + { + "epoch": 0.07527117282598315, + "grad_norm": 0.9203095436096191, + "learning_rate": 1.9699048432216244e-05, + "loss": 3.9916, + "step": 6700 + }, + { + "epoch": 0.07583289799632632, + "grad_norm": 0.9306748509407043, + "learning_rate": 1.9696801518913397e-05, + "loss": 3.9701, + "step": 6750 + }, + { + "epoch": 0.07639462316666948, + "grad_norm": 0.7836359143257141, + "learning_rate": 1.9694554605610546e-05, + "loss": 3.9381, + "step": 6800 + }, + { + "epoch": 0.07695634833701263, + "grad_norm": 1.0823968648910522, + "learning_rate": 1.9692307692307696e-05, + "loss": 3.9308, + "step": 6850 + }, + { + "epoch": 0.07751807350735579, + "grad_norm": 1.1342588663101196, + "learning_rate": 1.9690060779004845e-05, + "loss": 3.958, + "step": 6900 + }, + { + "epoch": 0.07807979867769894, + "grad_norm": 0.9671859741210938, + "learning_rate": 1.9687813865701994e-05, + "loss": 3.9878, + "step": 6950 + }, + { + "epoch": 0.07864152384804211, + "grad_norm": 0.925604522228241, + "learning_rate": 1.9685566952399144e-05, + "loss": 3.9283, + "step": 7000 + }, + { + "epoch": 0.07920324901838527, + "grad_norm": 0.8741235136985779, + "learning_rate": 1.9683320039096293e-05, + "loss": 3.865, + "step": 7050 + }, + { + "epoch": 0.07976497418872842, + "grad_norm": 0.8538075089454651, + "learning_rate": 1.9681073125793446e-05, + "loss": 3.8876, + "step": 7100 + }, + { + "epoch": 0.08032669935907158, + "grad_norm": 0.948414146900177, + "learning_rate": 1.967882621249059e-05, + "loss": 3.9869, + "step": 7150 + }, + { + "epoch": 0.08088842452941473, + "grad_norm": 1.2384617328643799, + "learning_rate": 1.967657929918774e-05, + "loss": 3.971, + "step": 7200 + }, + { + "epoch": 0.08145014969975789, + "grad_norm": 0.9849569797515869, + "learning_rate": 1.967433238588489e-05, + "loss": 3.9783, + "step": 7250 + }, + { + "epoch": 0.08201187487010106, + "grad_norm": 0.923664927482605, + "learning_rate": 1.967208547258204e-05, + "loss": 3.9348, + "step": 7300 + }, + { + "epoch": 0.08257360004044421, + "grad_norm": 0.9964247345924377, + "learning_rate": 1.9669838559279192e-05, + "loss": 3.954, + "step": 7350 + }, + { + "epoch": 0.08313532521078737, + "grad_norm": 1.320538878440857, + "learning_rate": 1.966759164597634e-05, + "loss": 3.9864, + "step": 7400 + }, + { + "epoch": 0.08369705038113052, + "grad_norm": 0.930242657661438, + "learning_rate": 1.966534473267349e-05, + "loss": 3.9783, + "step": 7450 + }, + { + "epoch": 0.08425877555147368, + "grad_norm": 0.9589186906814575, + "learning_rate": 1.966309781937064e-05, + "loss": 4.0511, + "step": 7500 + }, + { + "epoch": 0.08482050072181685, + "grad_norm": 0.9792759418487549, + "learning_rate": 1.966085090606779e-05, + "loss": 3.9537, + "step": 7550 + }, + { + "epoch": 0.08538222589216, + "grad_norm": 1.03499436378479, + "learning_rate": 1.965860399276494e-05, + "loss": 3.9217, + "step": 7600 + }, + { + "epoch": 0.08594395106250316, + "grad_norm": 0.8685311675071716, + "learning_rate": 1.9656357079462088e-05, + "loss": 3.8727, + "step": 7650 + }, + { + "epoch": 0.08650567623284632, + "grad_norm": 0.9089112281799316, + "learning_rate": 1.9654155104425297e-05, + "loss": 3.9332, + "step": 7700 + }, + { + "epoch": 0.08706740140318947, + "grad_norm": 0.8686912655830383, + "learning_rate": 1.9651908191122447e-05, + "loss": 4.0604, + "step": 7750 + }, + { + "epoch": 0.08762912657353264, + "grad_norm": 1.0852463245391846, + "learning_rate": 1.9649661277819596e-05, + "loss": 3.9402, + "step": 7800 + }, + { + "epoch": 0.0881908517438758, + "grad_norm": 1.0768681764602661, + "learning_rate": 1.9647414364516745e-05, + "loss": 3.9484, + "step": 7850 + }, + { + "epoch": 0.08875257691421895, + "grad_norm": 0.9111924171447754, + "learning_rate": 1.9645167451213895e-05, + "loss": 3.897, + "step": 7900 + }, + { + "epoch": 0.0893143020845621, + "grad_norm": 1.0498020648956299, + "learning_rate": 1.9642920537911044e-05, + "loss": 3.9207, + "step": 7950 + }, + { + "epoch": 0.08987602725490526, + "grad_norm": 0.9853581190109253, + "learning_rate": 1.9640673624608197e-05, + "loss": 3.9408, + "step": 8000 + }, + { + "epoch": 0.09043775242524842, + "grad_norm": 1.053979754447937, + "learning_rate": 1.9638426711305346e-05, + "loss": 3.9448, + "step": 8050 + }, + { + "epoch": 0.09099947759559159, + "grad_norm": 0.9768669009208679, + "learning_rate": 1.9636179798002495e-05, + "loss": 3.9075, + "step": 8100 + }, + { + "epoch": 0.09156120276593474, + "grad_norm": 0.9274827241897583, + "learning_rate": 1.9633932884699645e-05, + "loss": 3.9015, + "step": 8150 + }, + { + "epoch": 0.0921229279362779, + "grad_norm": 0.9641353487968445, + "learning_rate": 1.9631685971396794e-05, + "loss": 3.9437, + "step": 8200 + }, + { + "epoch": 0.09268465310662105, + "grad_norm": 0.9485934972763062, + "learning_rate": 1.9629439058093943e-05, + "loss": 3.8863, + "step": 8250 + }, + { + "epoch": 0.0932463782769642, + "grad_norm": 0.880131185054779, + "learning_rate": 1.9627192144791093e-05, + "loss": 3.8816, + "step": 8300 + }, + { + "epoch": 0.09380810344730738, + "grad_norm": 1.05781888961792, + "learning_rate": 1.9624945231488245e-05, + "loss": 3.9312, + "step": 8350 + }, + { + "epoch": 0.09436982861765053, + "grad_norm": 0.8662311434745789, + "learning_rate": 1.9622698318185395e-05, + "loss": 3.9234, + "step": 8400 + }, + { + "epoch": 0.09493155378799369, + "grad_norm": 0.8740962147712708, + "learning_rate": 1.9620451404882544e-05, + "loss": 3.8708, + "step": 8450 + }, + { + "epoch": 0.09549327895833684, + "grad_norm": 0.8588781952857971, + "learning_rate": 1.9618204491579693e-05, + "loss": 3.9236, + "step": 8500 + }, + { + "epoch": 0.09605500412868, + "grad_norm": 1.1292428970336914, + "learning_rate": 1.9615957578276843e-05, + "loss": 3.938, + "step": 8550 + }, + { + "epoch": 0.09661672929902317, + "grad_norm": 0.9856300354003906, + "learning_rate": 1.9613710664973992e-05, + "loss": 3.8818, + "step": 8600 + }, + { + "epoch": 0.09717845446936632, + "grad_norm": 0.9480489492416382, + "learning_rate": 1.961146375167114e-05, + "loss": 3.9085, + "step": 8650 + }, + { + "epoch": 0.09774017963970948, + "grad_norm": 0.9648436903953552, + "learning_rate": 1.9609216838368294e-05, + "loss": 3.9123, + "step": 8700 + }, + { + "epoch": 0.09830190481005263, + "grad_norm": 1.042819857597351, + "learning_rate": 1.9606969925065443e-05, + "loss": 3.8794, + "step": 8750 + }, + { + "epoch": 0.09886362998039579, + "grad_norm": 0.9484936594963074, + "learning_rate": 1.9604723011762592e-05, + "loss": 3.9091, + "step": 8800 + }, + { + "epoch": 0.09942535515073896, + "grad_norm": 1.0150383710861206, + "learning_rate": 1.9602476098459742e-05, + "loss": 3.9898, + "step": 8850 + }, + { + "epoch": 0.09998708032108211, + "grad_norm": 1.0116498470306396, + "learning_rate": 1.960022918515689e-05, + "loss": 3.9227, + "step": 8900 + }, + { + "epoch": 0.10054880549142527, + "grad_norm": 1.054053783416748, + "learning_rate": 1.959798227185404e-05, + "loss": 3.9434, + "step": 8950 + }, + { + "epoch": 0.10111053066176842, + "grad_norm": 1.049973964691162, + "learning_rate": 1.959573535855119e-05, + "loss": 3.9419, + "step": 9000 + }, + { + "epoch": 0.10167225583211158, + "grad_norm": 0.8714705109596252, + "learning_rate": 1.9593488445248342e-05, + "loss": 3.8853, + "step": 9050 + }, + { + "epoch": 0.10223398100245473, + "grad_norm": 0.9340712428092957, + "learning_rate": 1.9591241531945492e-05, + "loss": 3.9283, + "step": 9100 + }, + { + "epoch": 0.1027957061727979, + "grad_norm": 0.8932335376739502, + "learning_rate": 1.958899461864264e-05, + "loss": 3.9738, + "step": 9150 + }, + { + "epoch": 0.10335743134314106, + "grad_norm": 0.9076048731803894, + "learning_rate": 1.958674770533979e-05, + "loss": 3.865, + "step": 9200 + }, + { + "epoch": 0.10391915651348421, + "grad_norm": 1.030010461807251, + "learning_rate": 1.958450079203694e-05, + "loss": 4.0057, + "step": 9250 + }, + { + "epoch": 0.10448088168382737, + "grad_norm": 0.978993833065033, + "learning_rate": 1.958225387873409e-05, + "loss": 3.891, + "step": 9300 + }, + { + "epoch": 0.10504260685417052, + "grad_norm": 1.023034691810608, + "learning_rate": 1.958000696543124e-05, + "loss": 3.9213, + "step": 9350 + }, + { + "epoch": 0.10560433202451369, + "grad_norm": 1.2275335788726807, + "learning_rate": 1.957776005212839e-05, + "loss": 3.9326, + "step": 9400 + }, + { + "epoch": 0.10616605719485685, + "grad_norm": 1.1626818180084229, + "learning_rate": 1.957551313882554e-05, + "loss": 3.958, + "step": 9450 + }, + { + "epoch": 0.1067277823652, + "grad_norm": 0.953780472278595, + "learning_rate": 1.957326622552269e-05, + "loss": 3.9185, + "step": 9500 + }, + { + "epoch": 0.10728950753554316, + "grad_norm": 1.0907609462738037, + "learning_rate": 1.957101931221984e-05, + "loss": 3.8956, + "step": 9550 + }, + { + "epoch": 0.10785123270588631, + "grad_norm": 0.8553085923194885, + "learning_rate": 1.956877239891699e-05, + "loss": 3.9707, + "step": 9600 + }, + { + "epoch": 0.10841295787622948, + "grad_norm": 1.1353096961975098, + "learning_rate": 1.9566525485614138e-05, + "loss": 3.8704, + "step": 9650 + }, + { + "epoch": 0.10897468304657264, + "grad_norm": 1.0312119722366333, + "learning_rate": 1.9564278572311287e-05, + "loss": 3.9181, + "step": 9700 + }, + { + "epoch": 0.10953640821691579, + "grad_norm": 2.1263670921325684, + "learning_rate": 1.9562076597274496e-05, + "loss": 3.8704, + "step": 9750 + }, + { + "epoch": 0.11009813338725895, + "grad_norm": 1.009627103805542, + "learning_rate": 1.9559829683971646e-05, + "loss": 3.8944, + "step": 9800 + }, + { + "epoch": 0.1106598585576021, + "grad_norm": 1.0444031953811646, + "learning_rate": 1.9557582770668795e-05, + "loss": 3.9716, + "step": 9850 + }, + { + "epoch": 0.11122158372794527, + "grad_norm": 1.0081613063812256, + "learning_rate": 1.9555335857365944e-05, + "loss": 3.9325, + "step": 9900 + }, + { + "epoch": 0.11178330889828843, + "grad_norm": 1.05756413936615, + "learning_rate": 1.9553088944063094e-05, + "loss": 3.9864, + "step": 9950 + }, + { + "epoch": 0.11234503406863158, + "grad_norm": 1.0573816299438477, + "learning_rate": 1.9550842030760243e-05, + "loss": 3.9455, + "step": 10000 + }, + { + "epoch": 0.11290675923897474, + "grad_norm": 1.108229637145996, + "learning_rate": 1.9548595117457396e-05, + "loss": 3.9542, + "step": 10050 + }, + { + "epoch": 0.1134684844093179, + "grad_norm": 0.9424736499786377, + "learning_rate": 1.9546348204154545e-05, + "loss": 3.9492, + "step": 10100 + }, + { + "epoch": 0.11403020957966105, + "grad_norm": 1.1272293329238892, + "learning_rate": 1.9544101290851694e-05, + "loss": 3.9152, + "step": 10150 + }, + { + "epoch": 0.11459193475000422, + "grad_norm": 0.9885505437850952, + "learning_rate": 1.9541854377548844e-05, + "loss": 3.9176, + "step": 10200 + }, + { + "epoch": 0.11515365992034737, + "grad_norm": 1.211566686630249, + "learning_rate": 1.9539607464245993e-05, + "loss": 3.9019, + "step": 10250 + }, + { + "epoch": 0.11571538509069053, + "grad_norm": 0.9763829112052917, + "learning_rate": 1.9537360550943142e-05, + "loss": 3.8857, + "step": 10300 + }, + { + "epoch": 0.11627711026103368, + "grad_norm": 1.1474283933639526, + "learning_rate": 1.953511363764029e-05, + "loss": 3.9197, + "step": 10350 + }, + { + "epoch": 0.11683883543137684, + "grad_norm": 0.9613032937049866, + "learning_rate": 1.9532866724337444e-05, + "loss": 3.9486, + "step": 10400 + }, + { + "epoch": 0.11740056060172001, + "grad_norm": 0.9909958839416504, + "learning_rate": 1.9530619811034593e-05, + "loss": 3.9196, + "step": 10450 + }, + { + "epoch": 0.11796228577206316, + "grad_norm": 1.0783482789993286, + "learning_rate": 1.9528372897731743e-05, + "loss": 3.9109, + "step": 10500 + }, + { + "epoch": 0.11852401094240632, + "grad_norm": 0.9794992804527283, + "learning_rate": 1.9526125984428892e-05, + "loss": 3.8683, + "step": 10550 + }, + { + "epoch": 0.11908573611274947, + "grad_norm": 1.1354711055755615, + "learning_rate": 1.952387907112604e-05, + "loss": 3.8885, + "step": 10600 + }, + { + "epoch": 0.11964746128309263, + "grad_norm": 1.0472307205200195, + "learning_rate": 1.952163215782319e-05, + "loss": 3.9329, + "step": 10650 + }, + { + "epoch": 0.1202091864534358, + "grad_norm": 0.9388378262519836, + "learning_rate": 1.951938524452034e-05, + "loss": 3.9027, + "step": 10700 + }, + { + "epoch": 0.12077091162377895, + "grad_norm": 0.9691028594970703, + "learning_rate": 1.9517138331217493e-05, + "loss": 3.8493, + "step": 10750 + }, + { + "epoch": 0.12133263679412211, + "grad_norm": 0.8358737826347351, + "learning_rate": 1.9514891417914642e-05, + "loss": 3.8618, + "step": 10800 + }, + { + "epoch": 0.12189436196446526, + "grad_norm": 1.1211429834365845, + "learning_rate": 1.951264450461179e-05, + "loss": 3.9208, + "step": 10850 + }, + { + "epoch": 0.12245608713480842, + "grad_norm": 0.9543302059173584, + "learning_rate": 1.951039759130894e-05, + "loss": 3.8512, + "step": 10900 + }, + { + "epoch": 0.12301781230515157, + "grad_norm": 1.0548458099365234, + "learning_rate": 1.950815067800609e-05, + "loss": 3.9096, + "step": 10950 + }, + { + "epoch": 0.12357953747549474, + "grad_norm": 0.8686736226081848, + "learning_rate": 1.950590376470324e-05, + "loss": 3.8479, + "step": 11000 + }, + { + "epoch": 0.1241412626458379, + "grad_norm": 1.0551873445510864, + "learning_rate": 1.950365685140039e-05, + "loss": 3.9508, + "step": 11050 + }, + { + "epoch": 0.12470298781618105, + "grad_norm": 1.0495314598083496, + "learning_rate": 1.950140993809754e-05, + "loss": 3.9477, + "step": 11100 + }, + { + "epoch": 0.12526471298652422, + "grad_norm": 0.885094165802002, + "learning_rate": 1.949916302479469e-05, + "loss": 3.8275, + "step": 11150 + }, + { + "epoch": 0.12582643815686737, + "grad_norm": 1.0710852146148682, + "learning_rate": 1.949691611149184e-05, + "loss": 3.8888, + "step": 11200 + }, + { + "epoch": 0.12638816332721053, + "grad_norm": 0.8449720144271851, + "learning_rate": 1.949466919818899e-05, + "loss": 3.8312, + "step": 11250 + }, + { + "epoch": 0.12694988849755368, + "grad_norm": 0.9350094795227051, + "learning_rate": 1.949242228488614e-05, + "loss": 3.8534, + "step": 11300 + }, + { + "epoch": 0.12751161366789684, + "grad_norm": 0.9171339869499207, + "learning_rate": 1.9490175371583288e-05, + "loss": 3.8905, + "step": 11350 + }, + { + "epoch": 0.12807333883824001, + "grad_norm": 0.8574774861335754, + "learning_rate": 1.9487928458280437e-05, + "loss": 3.9054, + "step": 11400 + }, + { + "epoch": 0.12863506400858316, + "grad_norm": 0.9054346084594727, + "learning_rate": 1.9485681544977587e-05, + "loss": 3.8086, + "step": 11450 + }, + { + "epoch": 0.12919678917892632, + "grad_norm": 0.9606080055236816, + "learning_rate": 1.948343463167474e-05, + "loss": 3.9055, + "step": 11500 + }, + { + "epoch": 0.12975851434926947, + "grad_norm": 1.0003029108047485, + "learning_rate": 1.948118771837189e-05, + "loss": 3.8765, + "step": 11550 + }, + { + "epoch": 0.13032023951961264, + "grad_norm": 1.1126985549926758, + "learning_rate": 1.9478940805069038e-05, + "loss": 3.8959, + "step": 11600 + }, + { + "epoch": 0.1308819646899558, + "grad_norm": 0.8803656101226807, + "learning_rate": 1.9476693891766187e-05, + "loss": 3.8938, + "step": 11650 + }, + { + "epoch": 0.13144368986029895, + "grad_norm": 0.9052019119262695, + "learning_rate": 1.9474446978463337e-05, + "loss": 3.9625, + "step": 11700 + }, + { + "epoch": 0.13200541503064211, + "grad_norm": 1.0454570055007935, + "learning_rate": 1.9472200065160486e-05, + "loss": 3.8853, + "step": 11750 + }, + { + "epoch": 0.13256714020098526, + "grad_norm": 1.0287830829620361, + "learning_rate": 1.9469953151857635e-05, + "loss": 3.8961, + "step": 11800 + }, + { + "epoch": 0.13312886537132843, + "grad_norm": 0.985406219959259, + "learning_rate": 1.9467706238554788e-05, + "loss": 3.8954, + "step": 11850 + }, + { + "epoch": 0.1336905905416716, + "grad_norm": 1.020377278327942, + "learning_rate": 1.9465459325251937e-05, + "loss": 3.9196, + "step": 11900 + }, + { + "epoch": 0.13425231571201474, + "grad_norm": 1.0363904237747192, + "learning_rate": 1.9463212411949087e-05, + "loss": 3.8959, + "step": 11950 + }, + { + "epoch": 0.1348140408823579, + "grad_norm": 0.8309126496315002, + "learning_rate": 1.9460965498646236e-05, + "loss": 3.8518, + "step": 12000 + }, + { + "epoch": 0.13537576605270105, + "grad_norm": 1.0168477296829224, + "learning_rate": 1.9458718585343385e-05, + "loss": 3.8139, + "step": 12050 + }, + { + "epoch": 0.13593749122304422, + "grad_norm": 0.9926967024803162, + "learning_rate": 1.9456471672040534e-05, + "loss": 3.9504, + "step": 12100 + }, + { + "epoch": 0.13649921639338738, + "grad_norm": 1.094438910484314, + "learning_rate": 1.9454224758737684e-05, + "loss": 3.8508, + "step": 12150 + }, + { + "epoch": 0.13706094156373053, + "grad_norm": 1.0631167888641357, + "learning_rate": 1.9451977845434836e-05, + "loss": 3.9198, + "step": 12200 + }, + { + "epoch": 0.1376226667340737, + "grad_norm": 0.9869749546051025, + "learning_rate": 1.9449730932131986e-05, + "loss": 3.9386, + "step": 12250 + }, + { + "epoch": 0.13818439190441684, + "grad_norm": 0.9479770660400391, + "learning_rate": 1.9447484018829135e-05, + "loss": 3.9549, + "step": 12300 + }, + { + "epoch": 0.13874611707476, + "grad_norm": 1.067071795463562, + "learning_rate": 1.9445237105526284e-05, + "loss": 3.9053, + "step": 12350 + }, + { + "epoch": 0.13930784224510315, + "grad_norm": 1.1760696172714233, + "learning_rate": 1.9442990192223434e-05, + "loss": 3.984, + "step": 12400 + }, + { + "epoch": 0.13986956741544632, + "grad_norm": 0.9269402623176575, + "learning_rate": 1.9440743278920583e-05, + "loss": 3.9179, + "step": 12450 + }, + { + "epoch": 0.14043129258578949, + "grad_norm": 0.839056134223938, + "learning_rate": 1.9438496365617732e-05, + "loss": 3.936, + "step": 12500 + }, + { + "epoch": 0.14099301775613263, + "grad_norm": 0.846478283405304, + "learning_rate": 1.9436249452314885e-05, + "loss": 3.8743, + "step": 12550 + }, + { + "epoch": 0.1415547429264758, + "grad_norm": 1.2440580129623413, + "learning_rate": 1.9434002539012034e-05, + "loss": 3.9124, + "step": 12600 + }, + { + "epoch": 0.14211646809681894, + "grad_norm": 0.9950169920921326, + "learning_rate": 1.9431755625709184e-05, + "loss": 3.8883, + "step": 12650 + }, + { + "epoch": 0.1426781932671621, + "grad_norm": 0.9739125967025757, + "learning_rate": 1.9429508712406333e-05, + "loss": 3.8438, + "step": 12700 + }, + { + "epoch": 0.14323991843750528, + "grad_norm": 0.9917782545089722, + "learning_rate": 1.9427261799103482e-05, + "loss": 3.8871, + "step": 12750 + }, + { + "epoch": 0.14380164360784842, + "grad_norm": 0.9435528516769409, + "learning_rate": 1.942501488580063e-05, + "loss": 3.8766, + "step": 12800 + }, + { + "epoch": 0.1443633687781916, + "grad_norm": 1.1224132776260376, + "learning_rate": 1.942276797249778e-05, + "loss": 3.9131, + "step": 12850 + }, + { + "epoch": 0.14492509394853473, + "grad_norm": 0.9452431201934814, + "learning_rate": 1.9420521059194934e-05, + "loss": 3.8974, + "step": 12900 + }, + { + "epoch": 0.1454868191188779, + "grad_norm": 0.9240500330924988, + "learning_rate": 1.9418274145892083e-05, + "loss": 3.8462, + "step": 12950 + }, + { + "epoch": 0.14604854428922107, + "grad_norm": 1.0114437341690063, + "learning_rate": 1.9416027232589232e-05, + "loss": 3.8693, + "step": 13000 + }, + { + "epoch": 0.1466102694595642, + "grad_norm": 0.9811971783638, + "learning_rate": 1.941378031928638e-05, + "loss": 3.9183, + "step": 13050 + }, + { + "epoch": 0.14717199462990738, + "grad_norm": 0.8882970213890076, + "learning_rate": 1.941153340598353e-05, + "loss": 3.9643, + "step": 13100 + }, + { + "epoch": 0.14773371980025052, + "grad_norm": 0.9701553583145142, + "learning_rate": 1.940928649268068e-05, + "loss": 3.8478, + "step": 13150 + }, + { + "epoch": 0.1482954449705937, + "grad_norm": 1.0480194091796875, + "learning_rate": 1.940703957937783e-05, + "loss": 3.8611, + "step": 13200 + }, + { + "epoch": 0.14885717014093686, + "grad_norm": 1.0310245752334595, + "learning_rate": 1.9404792666074982e-05, + "loss": 3.9213, + "step": 13250 + }, + { + "epoch": 0.14941889531128, + "grad_norm": 1.2251856327056885, + "learning_rate": 1.940254575277213e-05, + "loss": 3.8666, + "step": 13300 + }, + { + "epoch": 0.14998062048162317, + "grad_norm": 1.121207356452942, + "learning_rate": 1.940029883946928e-05, + "loss": 3.8986, + "step": 13350 + }, + { + "epoch": 0.1505423456519663, + "grad_norm": 1.0388009548187256, + "learning_rate": 1.939805192616643e-05, + "loss": 3.8735, + "step": 13400 + }, + { + "epoch": 0.15110407082230948, + "grad_norm": 1.0072270631790161, + "learning_rate": 1.939580501286358e-05, + "loss": 3.8796, + "step": 13450 + }, + { + "epoch": 0.15166579599265265, + "grad_norm": 1.1749372482299805, + "learning_rate": 1.939355809956073e-05, + "loss": 3.9541, + "step": 13500 + }, + { + "epoch": 0.1522275211629958, + "grad_norm": 0.996511697769165, + "learning_rate": 1.9391311186257878e-05, + "loss": 3.9101, + "step": 13550 + }, + { + "epoch": 0.15278924633333896, + "grad_norm": 0.8689504265785217, + "learning_rate": 1.938906427295503e-05, + "loss": 3.8309, + "step": 13600 + }, + { + "epoch": 0.1533509715036821, + "grad_norm": 0.8855342268943787, + "learning_rate": 1.938681735965218e-05, + "loss": 3.9087, + "step": 13650 + }, + { + "epoch": 0.15391269667402527, + "grad_norm": 0.8916892409324646, + "learning_rate": 1.938457044634933e-05, + "loss": 3.8328, + "step": 13700 + }, + { + "epoch": 0.15447442184436844, + "grad_norm": 0.9709575772285461, + "learning_rate": 1.938232353304648e-05, + "loss": 3.9131, + "step": 13750 + }, + { + "epoch": 0.15503614701471158, + "grad_norm": 1.1655782461166382, + "learning_rate": 1.9380121558009685e-05, + "loss": 3.9148, + "step": 13800 + }, + { + "epoch": 0.15559787218505475, + "grad_norm": 1.018314003944397, + "learning_rate": 1.9377874644706834e-05, + "loss": 3.9141, + "step": 13850 + }, + { + "epoch": 0.1561595973553979, + "grad_norm": 0.9596139788627625, + "learning_rate": 1.9375627731403987e-05, + "loss": 3.9603, + "step": 13900 + }, + { + "epoch": 0.15672132252574106, + "grad_norm": 0.993695080280304, + "learning_rate": 1.9373380818101136e-05, + "loss": 3.8694, + "step": 13950 + }, + { + "epoch": 0.15728304769608423, + "grad_norm": 0.9984117746353149, + "learning_rate": 1.9371133904798285e-05, + "loss": 3.9069, + "step": 14000 + }, + { + "epoch": 0.15784477286642737, + "grad_norm": 1.0261962413787842, + "learning_rate": 1.9368886991495435e-05, + "loss": 3.7683, + "step": 14050 + }, + { + "epoch": 0.15840649803677054, + "grad_norm": 0.886608362197876, + "learning_rate": 1.9366640078192584e-05, + "loss": 3.8899, + "step": 14100 + }, + { + "epoch": 0.15896822320711368, + "grad_norm": 0.9339202046394348, + "learning_rate": 1.9364393164889733e-05, + "loss": 3.7592, + "step": 14150 + }, + { + "epoch": 0.15952994837745685, + "grad_norm": 0.919339120388031, + "learning_rate": 1.9362146251586883e-05, + "loss": 3.9459, + "step": 14200 + }, + { + "epoch": 0.1600916735478, + "grad_norm": 1.0983084440231323, + "learning_rate": 1.9359899338284035e-05, + "loss": 3.9416, + "step": 14250 + }, + { + "epoch": 0.16065339871814316, + "grad_norm": 0.9921066164970398, + "learning_rate": 1.9357652424981185e-05, + "loss": 3.9033, + "step": 14300 + }, + { + "epoch": 0.16121512388848633, + "grad_norm": 0.9689832329750061, + "learning_rate": 1.9355405511678334e-05, + "loss": 3.9637, + "step": 14350 + }, + { + "epoch": 0.16177684905882947, + "grad_norm": 1.1760526895523071, + "learning_rate": 1.9353158598375483e-05, + "loss": 3.8931, + "step": 14400 + }, + { + "epoch": 0.16233857422917264, + "grad_norm": 2.066866636276245, + "learning_rate": 1.9350911685072633e-05, + "loss": 3.9355, + "step": 14450 + }, + { + "epoch": 0.16290029939951578, + "grad_norm": 1.0954084396362305, + "learning_rate": 1.9348664771769782e-05, + "loss": 3.9496, + "step": 14500 + }, + { + "epoch": 0.16346202456985895, + "grad_norm": 1.043190598487854, + "learning_rate": 1.934641785846693e-05, + "loss": 3.8599, + "step": 14550 + }, + { + "epoch": 0.16402374974020212, + "grad_norm": 1.1450411081314087, + "learning_rate": 1.9344170945164084e-05, + "loss": 3.9175, + "step": 14600 + }, + { + "epoch": 0.16458547491054526, + "grad_norm": 1.3400977849960327, + "learning_rate": 1.9341924031861233e-05, + "loss": 3.7876, + "step": 14650 + }, + { + "epoch": 0.16514720008088843, + "grad_norm": 1.030008316040039, + "learning_rate": 1.9339677118558383e-05, + "loss": 3.8666, + "step": 14700 + }, + { + "epoch": 0.16570892525123157, + "grad_norm": 0.9812994599342346, + "learning_rate": 1.9337430205255532e-05, + "loss": 3.8993, + "step": 14750 + }, + { + "epoch": 0.16627065042157474, + "grad_norm": 0.9178810119628906, + "learning_rate": 1.933518329195268e-05, + "loss": 3.9666, + "step": 14800 + }, + { + "epoch": 0.1668323755919179, + "grad_norm": 1.3065404891967773, + "learning_rate": 1.933293637864983e-05, + "loss": 3.8952, + "step": 14850 + }, + { + "epoch": 0.16739410076226105, + "grad_norm": 1.1173884868621826, + "learning_rate": 1.933068946534698e-05, + "loss": 3.8152, + "step": 14900 + }, + { + "epoch": 0.16795582593260422, + "grad_norm": 1.0459538698196411, + "learning_rate": 1.9328442552044133e-05, + "loss": 3.9137, + "step": 14950 + }, + { + "epoch": 0.16851755110294736, + "grad_norm": 1.1197466850280762, + "learning_rate": 1.9326195638741282e-05, + "loss": 3.8563, + "step": 15000 + }, + { + "epoch": 0.16907927627329053, + "grad_norm": 1.0688142776489258, + "learning_rate": 1.932394872543843e-05, + "loss": 3.8626, + "step": 15050 + }, + { + "epoch": 0.1696410014436337, + "grad_norm": 0.9067109823226929, + "learning_rate": 1.932170181213558e-05, + "loss": 3.8535, + "step": 15100 + }, + { + "epoch": 0.17020272661397684, + "grad_norm": 1.0777970552444458, + "learning_rate": 1.931945489883273e-05, + "loss": 3.9188, + "step": 15150 + }, + { + "epoch": 0.17076445178432, + "grad_norm": 0.9435402750968933, + "learning_rate": 1.931720798552988e-05, + "loss": 3.8346, + "step": 15200 + }, + { + "epoch": 0.17132617695466315, + "grad_norm": 1.010043978691101, + "learning_rate": 1.931496107222703e-05, + "loss": 3.8905, + "step": 15250 + }, + { + "epoch": 0.17188790212500632, + "grad_norm": 0.9614844918251038, + "learning_rate": 1.931271415892418e-05, + "loss": 3.8939, + "step": 15300 + }, + { + "epoch": 0.1724496272953495, + "grad_norm": 1.1249394416809082, + "learning_rate": 1.931046724562133e-05, + "loss": 3.9328, + "step": 15350 + }, + { + "epoch": 0.17301135246569263, + "grad_norm": 1.04349684715271, + "learning_rate": 1.930822033231848e-05, + "loss": 3.8523, + "step": 15400 + }, + { + "epoch": 0.1735730776360358, + "grad_norm": 9.898025512695312, + "learning_rate": 1.930597341901563e-05, + "loss": 3.8503, + "step": 15450 + }, + { + "epoch": 0.17413480280637894, + "grad_norm": 0.9678813815116882, + "learning_rate": 1.930372650571278e-05, + "loss": 3.8466, + "step": 15500 + }, + { + "epoch": 0.1746965279767221, + "grad_norm": 0.9452058672904968, + "learning_rate": 1.9301479592409928e-05, + "loss": 3.8458, + "step": 15550 + }, + { + "epoch": 0.17525825314706528, + "grad_norm": 1.0991216897964478, + "learning_rate": 1.9299232679107077e-05, + "loss": 3.9693, + "step": 15600 + }, + { + "epoch": 0.17581997831740842, + "grad_norm": 1.1074953079223633, + "learning_rate": 1.9296985765804226e-05, + "loss": 3.8205, + "step": 15650 + }, + { + "epoch": 0.1763817034877516, + "grad_norm": 0.8646476864814758, + "learning_rate": 1.929473885250138e-05, + "loss": 3.8831, + "step": 15700 + }, + { + "epoch": 0.17694342865809473, + "grad_norm": 1.0361870527267456, + "learning_rate": 1.929249193919853e-05, + "loss": 3.985, + "step": 15750 + }, + { + "epoch": 0.1775051538284379, + "grad_norm": 1.405374526977539, + "learning_rate": 1.9290245025895678e-05, + "loss": 3.8676, + "step": 15800 + }, + { + "epoch": 0.17806687899878107, + "grad_norm": 1.0663036108016968, + "learning_rate": 1.9287998112592827e-05, + "loss": 3.9483, + "step": 15850 + }, + { + "epoch": 0.1786286041691242, + "grad_norm": 1.0984610319137573, + "learning_rate": 1.9285751199289976e-05, + "loss": 3.7979, + "step": 15900 + }, + { + "epoch": 0.17919032933946738, + "grad_norm": 1.0702370405197144, + "learning_rate": 1.9283504285987126e-05, + "loss": 3.8597, + "step": 15950 + }, + { + "epoch": 0.17975205450981052, + "grad_norm": 1.00674307346344, + "learning_rate": 1.9281257372684275e-05, + "loss": 3.9925, + "step": 16000 + }, + { + "epoch": 0.1803137796801537, + "grad_norm": 1.7243638038635254, + "learning_rate": 1.9279010459381428e-05, + "loss": 3.9243, + "step": 16050 + }, + { + "epoch": 0.18087550485049683, + "grad_norm": 0.9632601141929626, + "learning_rate": 1.9276763546078577e-05, + "loss": 3.9168, + "step": 16100 + }, + { + "epoch": 0.18143723002084, + "grad_norm": 1.0414319038391113, + "learning_rate": 1.9274516632775726e-05, + "loss": 3.917, + "step": 16150 + }, + { + "epoch": 0.18199895519118317, + "grad_norm": 1.0618311166763306, + "learning_rate": 1.9272314657738932e-05, + "loss": 3.8461, + "step": 16200 + }, + { + "epoch": 0.1825606803615263, + "grad_norm": 0.9754511117935181, + "learning_rate": 1.927006774443608e-05, + "loss": 3.8397, + "step": 16250 + }, + { + "epoch": 0.18312240553186948, + "grad_norm": 1.0348036289215088, + "learning_rate": 1.9267820831133234e-05, + "loss": 3.9151, + "step": 16300 + }, + { + "epoch": 0.18368413070221262, + "grad_norm": 1.0387177467346191, + "learning_rate": 1.9265573917830384e-05, + "loss": 3.8342, + "step": 16350 + }, + { + "epoch": 0.1842458558725558, + "grad_norm": 1.0977225303649902, + "learning_rate": 1.9263327004527533e-05, + "loss": 3.9074, + "step": 16400 + }, + { + "epoch": 0.18480758104289896, + "grad_norm": 0.9499973058700562, + "learning_rate": 1.9261080091224682e-05, + "loss": 3.9021, + "step": 16450 + }, + { + "epoch": 0.1853693062132421, + "grad_norm": 0.9755128026008606, + "learning_rate": 1.925883317792183e-05, + "loss": 3.8882, + "step": 16500 + }, + { + "epoch": 0.18593103138358527, + "grad_norm": 0.958068311214447, + "learning_rate": 1.925658626461898e-05, + "loss": 3.916, + "step": 16550 + }, + { + "epoch": 0.1864927565539284, + "grad_norm": 1.119287133216858, + "learning_rate": 1.925433935131613e-05, + "loss": 3.8832, + "step": 16600 + }, + { + "epoch": 0.18705448172427158, + "grad_norm": 1.097261667251587, + "learning_rate": 1.925209243801328e-05, + "loss": 3.8584, + "step": 16650 + }, + { + "epoch": 0.18761620689461475, + "grad_norm": 1.1982165575027466, + "learning_rate": 1.9249845524710432e-05, + "loss": 3.7963, + "step": 16700 + }, + { + "epoch": 0.1881779320649579, + "grad_norm": 1.0916881561279297, + "learning_rate": 1.924759861140758e-05, + "loss": 3.9541, + "step": 16750 + }, + { + "epoch": 0.18873965723530106, + "grad_norm": 1.1338039636611938, + "learning_rate": 1.924535169810473e-05, + "loss": 3.947, + "step": 16800 + }, + { + "epoch": 0.1893013824056442, + "grad_norm": 1.116104006767273, + "learning_rate": 1.924310478480188e-05, + "loss": 3.912, + "step": 16850 + }, + { + "epoch": 0.18986310757598737, + "grad_norm": 1.2745022773742676, + "learning_rate": 1.924085787149903e-05, + "loss": 3.8085, + "step": 16900 + }, + { + "epoch": 0.19042483274633054, + "grad_norm": 1.0067696571350098, + "learning_rate": 1.923861095819618e-05, + "loss": 3.8644, + "step": 16950 + }, + { + "epoch": 0.19098655791667368, + "grad_norm": 0.9528264403343201, + "learning_rate": 1.9236364044893328e-05, + "loss": 3.8716, + "step": 17000 + }, + { + "epoch": 0.19154828308701685, + "grad_norm": 0.9192308783531189, + "learning_rate": 1.923411713159048e-05, + "loss": 3.8261, + "step": 17050 + }, + { + "epoch": 0.19211000825736, + "grad_norm": 0.8801952004432678, + "learning_rate": 1.923187021828763e-05, + "loss": 3.9287, + "step": 17100 + }, + { + "epoch": 0.19267173342770316, + "grad_norm": 1.1006280183792114, + "learning_rate": 1.922962330498478e-05, + "loss": 3.8829, + "step": 17150 + }, + { + "epoch": 0.19323345859804633, + "grad_norm": 0.9863724112510681, + "learning_rate": 1.922737639168193e-05, + "loss": 3.8876, + "step": 17200 + }, + { + "epoch": 0.19379518376838947, + "grad_norm": 0.9066800475120544, + "learning_rate": 1.9225129478379078e-05, + "loss": 3.7709, + "step": 17250 + }, + { + "epoch": 0.19435690893873264, + "grad_norm": 1.0071481466293335, + "learning_rate": 1.9222882565076227e-05, + "loss": 3.8305, + "step": 17300 + }, + { + "epoch": 0.19491863410907578, + "grad_norm": 0.962822437286377, + "learning_rate": 1.9220635651773377e-05, + "loss": 3.8863, + "step": 17350 + }, + { + "epoch": 0.19548035927941895, + "grad_norm": 0.9313477873802185, + "learning_rate": 1.921838873847053e-05, + "loss": 3.8798, + "step": 17400 + }, + { + "epoch": 0.19604208444976212, + "grad_norm": 0.8907088041305542, + "learning_rate": 1.921614182516768e-05, + "loss": 3.8138, + "step": 17450 + }, + { + "epoch": 0.19660380962010526, + "grad_norm": 0.8724405765533447, + "learning_rate": 1.9213894911864828e-05, + "loss": 3.8547, + "step": 17500 + }, + { + "epoch": 0.19716553479044843, + "grad_norm": 1.0886657238006592, + "learning_rate": 1.9211647998561977e-05, + "loss": 3.8818, + "step": 17550 + }, + { + "epoch": 0.19772725996079157, + "grad_norm": 1.1140875816345215, + "learning_rate": 1.9209446023525183e-05, + "loss": 3.8971, + "step": 17600 + }, + { + "epoch": 0.19828898513113474, + "grad_norm": 1.0241341590881348, + "learning_rate": 1.9207199110222333e-05, + "loss": 3.8416, + "step": 17650 + }, + { + "epoch": 0.1988507103014779, + "grad_norm": 0.9751248359680176, + "learning_rate": 1.9204952196919485e-05, + "loss": 3.7553, + "step": 17700 + }, + { + "epoch": 0.19941243547182105, + "grad_norm": 1.0477747917175293, + "learning_rate": 1.9202705283616635e-05, + "loss": 3.8363, + "step": 17750 + }, + { + "epoch": 0.19997416064216422, + "grad_norm": 1.0788570642471313, + "learning_rate": 1.9200458370313784e-05, + "loss": 3.7763, + "step": 17800 + }, + { + "epoch": 0.20053588581250736, + "grad_norm": 1.0035017728805542, + "learning_rate": 1.9198211457010933e-05, + "loss": 3.8727, + "step": 17850 + }, + { + "epoch": 0.20109761098285053, + "grad_norm": 0.9635750651359558, + "learning_rate": 1.9195964543708083e-05, + "loss": 3.8129, + "step": 17900 + }, + { + "epoch": 0.2016593361531937, + "grad_norm": 1.2292413711547852, + "learning_rate": 1.9193717630405232e-05, + "loss": 3.8897, + "step": 17950 + }, + { + "epoch": 0.20222106132353684, + "grad_norm": 1.3051143884658813, + "learning_rate": 1.919147071710238e-05, + "loss": 3.8518, + "step": 18000 + }, + { + "epoch": 0.20278278649388, + "grad_norm": 1.1229668855667114, + "learning_rate": 1.9189223803799534e-05, + "loss": 3.8051, + "step": 18050 + }, + { + "epoch": 0.20334451166422315, + "grad_norm": 0.9775665402412415, + "learning_rate": 1.9186976890496683e-05, + "loss": 3.879, + "step": 18100 + }, + { + "epoch": 0.20390623683456632, + "grad_norm": 1.1217602491378784, + "learning_rate": 1.9184729977193833e-05, + "loss": 3.8303, + "step": 18150 + }, + { + "epoch": 0.20446796200490946, + "grad_norm": 1.9044723510742188, + "learning_rate": 1.9182483063890982e-05, + "loss": 3.872, + "step": 18200 + }, + { + "epoch": 0.20502968717525263, + "grad_norm": 0.9213973879814148, + "learning_rate": 1.918023615058813e-05, + "loss": 3.8859, + "step": 18250 + }, + { + "epoch": 0.2055914123455958, + "grad_norm": 1.2292492389678955, + "learning_rate": 1.917798923728528e-05, + "loss": 3.8682, + "step": 18300 + }, + { + "epoch": 0.20615313751593894, + "grad_norm": 0.9231402277946472, + "learning_rate": 1.917574232398243e-05, + "loss": 3.833, + "step": 18350 + }, + { + "epoch": 0.2067148626862821, + "grad_norm": 1.0987199544906616, + "learning_rate": 1.9173495410679583e-05, + "loss": 3.8887, + "step": 18400 + }, + { + "epoch": 0.20727658785662526, + "grad_norm": 1.301283359527588, + "learning_rate": 1.9171248497376732e-05, + "loss": 3.962, + "step": 18450 + }, + { + "epoch": 0.20783831302696842, + "grad_norm": 0.8758549690246582, + "learning_rate": 1.916900158407388e-05, + "loss": 3.8472, + "step": 18500 + }, + { + "epoch": 0.2084000381973116, + "grad_norm": 1.331650733947754, + "learning_rate": 1.916675467077103e-05, + "loss": 3.9557, + "step": 18550 + }, + { + "epoch": 0.20896176336765473, + "grad_norm": 1.001512885093689, + "learning_rate": 1.916450775746818e-05, + "loss": 3.9078, + "step": 18600 + }, + { + "epoch": 0.2095234885379979, + "grad_norm": 1.0210082530975342, + "learning_rate": 1.916226084416533e-05, + "loss": 3.7859, + "step": 18650 + }, + { + "epoch": 0.21008521370834105, + "grad_norm": 1.0081088542938232, + "learning_rate": 1.916001393086248e-05, + "loss": 3.9527, + "step": 18700 + }, + { + "epoch": 0.21064693887868421, + "grad_norm": 1.1623823642730713, + "learning_rate": 1.915776701755963e-05, + "loss": 3.8055, + "step": 18750 + }, + { + "epoch": 0.21120866404902738, + "grad_norm": 1.0312272310256958, + "learning_rate": 1.915552010425678e-05, + "loss": 3.8481, + "step": 18800 + }, + { + "epoch": 0.21177038921937053, + "grad_norm": 1.018204927444458, + "learning_rate": 1.915327319095393e-05, + "loss": 3.8106, + "step": 18850 + }, + { + "epoch": 0.2123321143897137, + "grad_norm": 1.005672574043274, + "learning_rate": 1.915102627765108e-05, + "loss": 3.839, + "step": 18900 + }, + { + "epoch": 0.21289383956005684, + "grad_norm": 0.9410429000854492, + "learning_rate": 1.914877936434823e-05, + "loss": 3.8118, + "step": 18950 + }, + { + "epoch": 0.2134555647304, + "grad_norm": 1.0433132648468018, + "learning_rate": 1.9146532451045378e-05, + "loss": 3.8394, + "step": 19000 + }, + { + "epoch": 0.21401728990074317, + "grad_norm": 1.1158668994903564, + "learning_rate": 1.9144285537742527e-05, + "loss": 3.8788, + "step": 19050 + }, + { + "epoch": 0.21457901507108632, + "grad_norm": 0.9531320929527283, + "learning_rate": 1.914203862443968e-05, + "loss": 3.9105, + "step": 19100 + }, + { + "epoch": 0.21514074024142948, + "grad_norm": 0.8815164566040039, + "learning_rate": 1.913979171113683e-05, + "loss": 3.8621, + "step": 19150 + }, + { + "epoch": 0.21570246541177263, + "grad_norm": 1.0627539157867432, + "learning_rate": 1.913754479783398e-05, + "loss": 3.8282, + "step": 19200 + }, + { + "epoch": 0.2162641905821158, + "grad_norm": 0.9545466303825378, + "learning_rate": 1.9135297884531128e-05, + "loss": 3.8455, + "step": 19250 + }, + { + "epoch": 0.21682591575245896, + "grad_norm": 1.0652786493301392, + "learning_rate": 1.9133050971228277e-05, + "loss": 3.8802, + "step": 19300 + }, + { + "epoch": 0.2173876409228021, + "grad_norm": 1.087463617324829, + "learning_rate": 1.9130804057925426e-05, + "loss": 3.7916, + "step": 19350 + }, + { + "epoch": 0.21794936609314527, + "grad_norm": 0.9182972311973572, + "learning_rate": 1.9128557144622576e-05, + "loss": 3.8506, + "step": 19400 + }, + { + "epoch": 0.21851109126348842, + "grad_norm": 0.9740565419197083, + "learning_rate": 1.912631023131973e-05, + "loss": 3.8396, + "step": 19450 + }, + { + "epoch": 0.21907281643383159, + "grad_norm": 1.2423253059387207, + "learning_rate": 1.9124063318016878e-05, + "loss": 3.8392, + "step": 19500 + }, + { + "epoch": 0.21963454160417475, + "grad_norm": 0.9749589562416077, + "learning_rate": 1.9121816404714027e-05, + "loss": 3.8517, + "step": 19550 + }, + { + "epoch": 0.2201962667745179, + "grad_norm": 1.0176291465759277, + "learning_rate": 1.9119569491411176e-05, + "loss": 3.9056, + "step": 19600 + }, + { + "epoch": 0.22075799194486107, + "grad_norm": 1.1581487655639648, + "learning_rate": 1.9117322578108326e-05, + "loss": 3.8801, + "step": 19650 + }, + { + "epoch": 0.2213197171152042, + "grad_norm": 0.9785657525062561, + "learning_rate": 1.9115075664805475e-05, + "loss": 3.7482, + "step": 19700 + }, + { + "epoch": 0.22188144228554738, + "grad_norm": 1.1080032587051392, + "learning_rate": 1.9112828751502624e-05, + "loss": 3.8123, + "step": 19750 + }, + { + "epoch": 0.22244316745589054, + "grad_norm": 0.9258034825325012, + "learning_rate": 1.9110581838199777e-05, + "loss": 3.8657, + "step": 19800 + }, + { + "epoch": 0.2230048926262337, + "grad_norm": 0.9505318999290466, + "learning_rate": 1.9108334924896926e-05, + "loss": 3.898, + "step": 19850 + }, + { + "epoch": 0.22356661779657686, + "grad_norm": 1.0489028692245483, + "learning_rate": 1.9106088011594076e-05, + "loss": 3.7987, + "step": 19900 + }, + { + "epoch": 0.22412834296692, + "grad_norm": 1.0717949867248535, + "learning_rate": 1.9103841098291225e-05, + "loss": 3.9085, + "step": 19950 + }, + { + "epoch": 0.22469006813726317, + "grad_norm": 0.9286379814147949, + "learning_rate": 1.9101594184988374e-05, + "loss": 3.845, + "step": 20000 + }, + { + "epoch": 0.2252517933076063, + "grad_norm": 1.0841072797775269, + "learning_rate": 1.9099347271685524e-05, + "loss": 3.881, + "step": 20050 + }, + { + "epoch": 0.22581351847794948, + "grad_norm": 0.8188834190368652, + "learning_rate": 1.9097100358382673e-05, + "loss": 3.9063, + "step": 20100 + }, + { + "epoch": 0.22637524364829265, + "grad_norm": 0.9557147026062012, + "learning_rate": 1.9094853445079826e-05, + "loss": 3.8553, + "step": 20150 + }, + { + "epoch": 0.2269369688186358, + "grad_norm": 1.0040074586868286, + "learning_rate": 1.9092606531776975e-05, + "loss": 3.8604, + "step": 20200 + }, + { + "epoch": 0.22749869398897896, + "grad_norm": 0.9978244304656982, + "learning_rate": 1.9090359618474124e-05, + "loss": 3.9428, + "step": 20250 + }, + { + "epoch": 0.2280604191593221, + "grad_norm": 0.998877227306366, + "learning_rate": 1.9088112705171274e-05, + "loss": 3.8784, + "step": 20300 + }, + { + "epoch": 0.22862214432966527, + "grad_norm": 1.1113442182540894, + "learning_rate": 1.9085865791868423e-05, + "loss": 3.8211, + "step": 20350 + }, + { + "epoch": 0.22918386950000844, + "grad_norm": 0.9248872399330139, + "learning_rate": 1.9083618878565572e-05, + "loss": 3.8687, + "step": 20400 + }, + { + "epoch": 0.22974559467035158, + "grad_norm": 1.293752670288086, + "learning_rate": 1.908137196526272e-05, + "loss": 3.8408, + "step": 20450 + }, + { + "epoch": 0.23030731984069475, + "grad_norm": 1.0347455739974976, + "learning_rate": 1.9079125051959874e-05, + "loss": 3.854, + "step": 20500 + }, + { + "epoch": 0.2308690450110379, + "grad_norm": 1.0671968460083008, + "learning_rate": 1.9076878138657023e-05, + "loss": 3.8493, + "step": 20550 + }, + { + "epoch": 0.23143077018138106, + "grad_norm": 1.0472993850708008, + "learning_rate": 1.9074631225354173e-05, + "loss": 3.7963, + "step": 20600 + }, + { + "epoch": 0.23199249535172423, + "grad_norm": 1.1053742170333862, + "learning_rate": 1.9072384312051322e-05, + "loss": 3.8703, + "step": 20650 + }, + { + "epoch": 0.23255422052206737, + "grad_norm": 0.9656935930252075, + "learning_rate": 1.907013739874847e-05, + "loss": 3.8551, + "step": 20700 + }, + { + "epoch": 0.23311594569241054, + "grad_norm": 1.0903059244155884, + "learning_rate": 1.906789048544562e-05, + "loss": 3.8798, + "step": 20750 + }, + { + "epoch": 0.23367767086275368, + "grad_norm": 1.178858995437622, + "learning_rate": 1.906564357214277e-05, + "loss": 3.8763, + "step": 20800 + }, + { + "epoch": 0.23423939603309685, + "grad_norm": 1.2152944803237915, + "learning_rate": 1.906339665883992e-05, + "loss": 3.9042, + "step": 20850 + }, + { + "epoch": 0.23480112120344002, + "grad_norm": 1.1343653202056885, + "learning_rate": 1.9061149745537072e-05, + "loss": 3.8355, + "step": 20900 + }, + { + "epoch": 0.23536284637378316, + "grad_norm": 1.0135072469711304, + "learning_rate": 1.905890283223422e-05, + "loss": 3.886, + "step": 20950 + }, + { + "epoch": 0.23592457154412633, + "grad_norm": 1.7519757747650146, + "learning_rate": 1.905665591893137e-05, + "loss": 3.7514, + "step": 21000 + }, + { + "epoch": 0.23648629671446947, + "grad_norm": 1.1740312576293945, + "learning_rate": 1.905440900562852e-05, + "loss": 3.8605, + "step": 21050 + }, + { + "epoch": 0.23704802188481264, + "grad_norm": 1.0153892040252686, + "learning_rate": 1.905216209232567e-05, + "loss": 3.8826, + "step": 21100 + }, + { + "epoch": 0.2376097470551558, + "grad_norm": 1.056986927986145, + "learning_rate": 1.904991517902282e-05, + "loss": 3.7912, + "step": 21150 + }, + { + "epoch": 0.23817147222549895, + "grad_norm": 1.4163366556167603, + "learning_rate": 1.9047668265719968e-05, + "loss": 3.8459, + "step": 21200 + }, + { + "epoch": 0.23873319739584212, + "grad_norm": 1.0286545753479004, + "learning_rate": 1.904542135241712e-05, + "loss": 3.8383, + "step": 21250 + }, + { + "epoch": 0.23929492256618526, + "grad_norm": 1.12931227684021, + "learning_rate": 1.904317443911427e-05, + "loss": 3.8532, + "step": 21300 + }, + { + "epoch": 0.23985664773652843, + "grad_norm": 0.9231151938438416, + "learning_rate": 1.904092752581142e-05, + "loss": 3.8494, + "step": 21350 + }, + { + "epoch": 0.2404183729068716, + "grad_norm": 1.382602572441101, + "learning_rate": 1.903868061250857e-05, + "loss": 3.851, + "step": 21400 + }, + { + "epoch": 0.24098009807721474, + "grad_norm": 0.9720520973205566, + "learning_rate": 1.9036433699205718e-05, + "loss": 3.8763, + "step": 21450 + }, + { + "epoch": 0.2415418232475579, + "grad_norm": 1.3178774118423462, + "learning_rate": 1.9034186785902867e-05, + "loss": 3.8769, + "step": 21500 + }, + { + "epoch": 0.24210354841790105, + "grad_norm": 1.0083391666412354, + "learning_rate": 1.9031939872600017e-05, + "loss": 3.8123, + "step": 21550 + }, + { + "epoch": 0.24266527358824422, + "grad_norm": 0.8887113928794861, + "learning_rate": 1.902969295929717e-05, + "loss": 3.8488, + "step": 21600 + }, + { + "epoch": 0.2432269987585874, + "grad_norm": 1.0424660444259644, + "learning_rate": 1.902744604599432e-05, + "loss": 3.922, + "step": 21650 + }, + { + "epoch": 0.24378872392893053, + "grad_norm": 1.0282403230667114, + "learning_rate": 1.9025199132691468e-05, + "loss": 3.7764, + "step": 21700 + }, + { + "epoch": 0.2443504490992737, + "grad_norm": 0.9313274621963501, + "learning_rate": 1.9022952219388617e-05, + "loss": 3.8267, + "step": 21750 + }, + { + "epoch": 0.24491217426961684, + "grad_norm": 0.9676574468612671, + "learning_rate": 1.9020705306085767e-05, + "loss": 3.8855, + "step": 21800 + }, + { + "epoch": 0.24547389943996, + "grad_norm": 0.9758104681968689, + "learning_rate": 1.9018503331048972e-05, + "loss": 3.8527, + "step": 21850 + }, + { + "epoch": 0.24603562461030315, + "grad_norm": 1.0635255575180054, + "learning_rate": 1.9016256417746122e-05, + "loss": 3.8487, + "step": 21900 + }, + { + "epoch": 0.24659734978064632, + "grad_norm": 1.0568088293075562, + "learning_rate": 1.901400950444327e-05, + "loss": 3.826, + "step": 21950 + }, + { + "epoch": 0.2471590749509895, + "grad_norm": 0.9051997065544128, + "learning_rate": 1.901176259114042e-05, + "loss": 3.8514, + "step": 22000 + }, + { + "epoch": 0.24772080012133263, + "grad_norm": 1.0521992444992065, + "learning_rate": 1.900951567783757e-05, + "loss": 3.8023, + "step": 22050 + }, + { + "epoch": 0.2482825252916758, + "grad_norm": 1.194155216217041, + "learning_rate": 1.9007268764534722e-05, + "loss": 3.751, + "step": 22100 + }, + { + "epoch": 0.24884425046201894, + "grad_norm": 1.0270308256149292, + "learning_rate": 1.9005021851231872e-05, + "loss": 3.8069, + "step": 22150 + }, + { + "epoch": 0.2494059756323621, + "grad_norm": 1.0364912748336792, + "learning_rate": 1.900277493792902e-05, + "loss": 3.8077, + "step": 22200 + }, + { + "epoch": 0.24996770080270528, + "grad_norm": 1.1937390565872192, + "learning_rate": 1.900052802462617e-05, + "loss": 3.8628, + "step": 22250 + }, + { + "epoch": 0.25052942597304845, + "grad_norm": 1.036866307258606, + "learning_rate": 1.899828111132332e-05, + "loss": 3.8209, + "step": 22300 + }, + { + "epoch": 0.2510911511433916, + "grad_norm": 1.014030933380127, + "learning_rate": 1.899603419802047e-05, + "loss": 3.8304, + "step": 22350 + }, + { + "epoch": 0.25165287631373473, + "grad_norm": 1.2910370826721191, + "learning_rate": 1.899378728471762e-05, + "loss": 3.8726, + "step": 22400 + }, + { + "epoch": 0.2522146014840779, + "grad_norm": 0.9999923706054688, + "learning_rate": 1.899154037141477e-05, + "loss": 3.809, + "step": 22450 + }, + { + "epoch": 0.25277632665442107, + "grad_norm": 1.0789824724197388, + "learning_rate": 1.898929345811192e-05, + "loss": 3.8186, + "step": 22500 + }, + { + "epoch": 0.2533380518247642, + "grad_norm": 0.976632833480835, + "learning_rate": 1.898704654480907e-05, + "loss": 3.8389, + "step": 22550 + }, + { + "epoch": 0.25389977699510735, + "grad_norm": 1.0078363418579102, + "learning_rate": 1.8984844569772276e-05, + "loss": 3.7896, + "step": 22600 + }, + { + "epoch": 0.25446150216545055, + "grad_norm": 1.0185294151306152, + "learning_rate": 1.8982597656469425e-05, + "loss": 3.7948, + "step": 22650 + }, + { + "epoch": 0.2550232273357937, + "grad_norm": 1.1852279901504517, + "learning_rate": 1.8980350743166574e-05, + "loss": 3.8872, + "step": 22700 + }, + { + "epoch": 0.25558495250613683, + "grad_norm": 1.039040446281433, + "learning_rate": 1.8978103829863727e-05, + "loss": 3.8026, + "step": 22750 + }, + { + "epoch": 0.25614667767648003, + "grad_norm": 0.9940977692604065, + "learning_rate": 1.8975856916560876e-05, + "loss": 3.9207, + "step": 22800 + }, + { + "epoch": 0.25670840284682317, + "grad_norm": 0.9345098733901978, + "learning_rate": 1.8973610003258026e-05, + "loss": 3.8501, + "step": 22850 + }, + { + "epoch": 0.2572701280171663, + "grad_norm": 1.0778815746307373, + "learning_rate": 1.8971363089955175e-05, + "loss": 3.9176, + "step": 22900 + }, + { + "epoch": 0.25783185318750945, + "grad_norm": 1.1364803314208984, + "learning_rate": 1.8969116176652324e-05, + "loss": 3.8351, + "step": 22950 + }, + { + "epoch": 0.25839357835785265, + "grad_norm": 0.9966996908187866, + "learning_rate": 1.8966869263349474e-05, + "loss": 3.89, + "step": 23000 + }, + { + "epoch": 0.2589553035281958, + "grad_norm": 1.0334755182266235, + "learning_rate": 1.8964622350046623e-05, + "loss": 3.8796, + "step": 23050 + }, + { + "epoch": 0.25951702869853893, + "grad_norm": 1.1375792026519775, + "learning_rate": 1.8962375436743776e-05, + "loss": 3.8086, + "step": 23100 + }, + { + "epoch": 0.26007875386888213, + "grad_norm": 1.1250351667404175, + "learning_rate": 1.8960128523440925e-05, + "loss": 3.9187, + "step": 23150 + }, + { + "epoch": 0.26064047903922527, + "grad_norm": 0.9979472756385803, + "learning_rate": 1.8957881610138074e-05, + "loss": 3.8808, + "step": 23200 + }, + { + "epoch": 0.2612022042095684, + "grad_norm": 1.0042188167572021, + "learning_rate": 1.8955634696835224e-05, + "loss": 3.9456, + "step": 23250 + }, + { + "epoch": 0.2617639293799116, + "grad_norm": 1.0758593082427979, + "learning_rate": 1.8953387783532373e-05, + "loss": 3.7855, + "step": 23300 + }, + { + "epoch": 0.26232565455025475, + "grad_norm": 0.9572558999061584, + "learning_rate": 1.8951140870229522e-05, + "loss": 3.8474, + "step": 23350 + }, + { + "epoch": 0.2628873797205979, + "grad_norm": 1.0140413045883179, + "learning_rate": 1.894889395692667e-05, + "loss": 3.8706, + "step": 23400 + }, + { + "epoch": 0.26344910489094103, + "grad_norm": 1.0036652088165283, + "learning_rate": 1.8946647043623824e-05, + "loss": 3.8338, + "step": 23450 + }, + { + "epoch": 0.26401083006128423, + "grad_norm": 1.1257251501083374, + "learning_rate": 1.8944400130320973e-05, + "loss": 3.9037, + "step": 23500 + }, + { + "epoch": 0.26457255523162737, + "grad_norm": 1.1244933605194092, + "learning_rate": 1.8942153217018123e-05, + "loss": 3.8757, + "step": 23550 + }, + { + "epoch": 0.2651342804019705, + "grad_norm": 1.1385120153427124, + "learning_rate": 1.8939906303715272e-05, + "loss": 3.9324, + "step": 23600 + }, + { + "epoch": 0.2656960055723137, + "grad_norm": 1.2949111461639404, + "learning_rate": 1.893765939041242e-05, + "loss": 3.8725, + "step": 23650 + }, + { + "epoch": 0.26625773074265685, + "grad_norm": 1.1291465759277344, + "learning_rate": 1.893541247710957e-05, + "loss": 3.8566, + "step": 23700 + }, + { + "epoch": 0.266819455913, + "grad_norm": 1.0833088159561157, + "learning_rate": 1.893316556380672e-05, + "loss": 3.8305, + "step": 23750 + }, + { + "epoch": 0.2673811810833432, + "grad_norm": 1.0878947973251343, + "learning_rate": 1.8930918650503873e-05, + "loss": 3.8964, + "step": 23800 + }, + { + "epoch": 0.26794290625368633, + "grad_norm": 1.03005051612854, + "learning_rate": 1.8928671737201022e-05, + "loss": 3.9039, + "step": 23850 + }, + { + "epoch": 0.26850463142402947, + "grad_norm": 1.0353825092315674, + "learning_rate": 1.892642482389817e-05, + "loss": 3.7323, + "step": 23900 + }, + { + "epoch": 0.2690663565943726, + "grad_norm": 0.9908705353736877, + "learning_rate": 1.892417791059532e-05, + "loss": 3.8291, + "step": 23950 + }, + { + "epoch": 0.2696280817647158, + "grad_norm": 1.1991065740585327, + "learning_rate": 1.892193099729247e-05, + "loss": 3.8784, + "step": 24000 + }, + { + "epoch": 0.27018980693505895, + "grad_norm": 1.3117947578430176, + "learning_rate": 1.891968408398962e-05, + "loss": 3.8274, + "step": 24050 + }, + { + "epoch": 0.2707515321054021, + "grad_norm": 0.9964234828948975, + "learning_rate": 1.891743717068677e-05, + "loss": 3.8596, + "step": 24100 + }, + { + "epoch": 0.2713132572757453, + "grad_norm": 1.0427865982055664, + "learning_rate": 1.8915190257383918e-05, + "loss": 3.7689, + "step": 24150 + }, + { + "epoch": 0.27187498244608843, + "grad_norm": 1.1147197484970093, + "learning_rate": 1.891294334408107e-05, + "loss": 3.8092, + "step": 24200 + }, + { + "epoch": 0.2724367076164316, + "grad_norm": 1.0301438570022583, + "learning_rate": 1.891069643077822e-05, + "loss": 3.7719, + "step": 24250 + }, + { + "epoch": 0.27299843278677477, + "grad_norm": 1.2350659370422363, + "learning_rate": 1.890844951747537e-05, + "loss": 3.831, + "step": 24300 + }, + { + "epoch": 0.2735601579571179, + "grad_norm": 0.917402446269989, + "learning_rate": 1.890620260417252e-05, + "loss": 3.7391, + "step": 24350 + }, + { + "epoch": 0.27412188312746105, + "grad_norm": 1.1625767946243286, + "learning_rate": 1.8903955690869668e-05, + "loss": 3.859, + "step": 24400 + }, + { + "epoch": 0.2746836082978042, + "grad_norm": 1.1152362823486328, + "learning_rate": 1.8901708777566817e-05, + "loss": 3.8546, + "step": 24450 + }, + { + "epoch": 0.2752453334681474, + "grad_norm": 0.9273315668106079, + "learning_rate": 1.8899461864263967e-05, + "loss": 3.8469, + "step": 24500 + }, + { + "epoch": 0.27580705863849053, + "grad_norm": 1.0813549757003784, + "learning_rate": 1.889721495096112e-05, + "loss": 3.8394, + "step": 24550 + }, + { + "epoch": 0.2763687838088337, + "grad_norm": 0.9964554309844971, + "learning_rate": 1.889496803765827e-05, + "loss": 3.869, + "step": 24600 + }, + { + "epoch": 0.27693050897917687, + "grad_norm": 0.8538205623626709, + "learning_rate": 1.8892721124355418e-05, + "loss": 3.8306, + "step": 24650 + }, + { + "epoch": 0.27749223414952, + "grad_norm": 1.0182757377624512, + "learning_rate": 1.8890474211052567e-05, + "loss": 3.8614, + "step": 24700 + }, + { + "epoch": 0.27805395931986315, + "grad_norm": 1.0067983865737915, + "learning_rate": 1.8888227297749717e-05, + "loss": 3.8798, + "step": 24750 + }, + { + "epoch": 0.2786156844902063, + "grad_norm": 1.0838966369628906, + "learning_rate": 1.8885980384446866e-05, + "loss": 3.8004, + "step": 24800 + }, + { + "epoch": 0.2791774096605495, + "grad_norm": 1.0354911088943481, + "learning_rate": 1.8883733471144015e-05, + "loss": 3.7978, + "step": 24850 + }, + { + "epoch": 0.27973913483089263, + "grad_norm": 1.3074768781661987, + "learning_rate": 1.8881486557841168e-05, + "loss": 3.8454, + "step": 24900 + }, + { + "epoch": 0.2803008600012358, + "grad_norm": 0.9944579601287842, + "learning_rate": 1.8879239644538317e-05, + "loss": 3.9164, + "step": 24950 + }, + { + "epoch": 0.28086258517157897, + "grad_norm": 0.8879290223121643, + "learning_rate": 1.8876992731235467e-05, + "loss": 3.8754, + "step": 25000 + }, + { + "epoch": 0.2814243103419221, + "grad_norm": 1.0343937873840332, + "learning_rate": 1.8874745817932616e-05, + "loss": 3.8178, + "step": 25050 + }, + { + "epoch": 0.28198603551226525, + "grad_norm": 1.0482374429702759, + "learning_rate": 1.8872498904629765e-05, + "loss": 3.8456, + "step": 25100 + }, + { + "epoch": 0.28254776068260845, + "grad_norm": 1.1767076253890991, + "learning_rate": 1.8870251991326914e-05, + "loss": 3.8205, + "step": 25150 + }, + { + "epoch": 0.2831094858529516, + "grad_norm": 1.220213770866394, + "learning_rate": 1.8868005078024064e-05, + "loss": 3.8225, + "step": 25200 + }, + { + "epoch": 0.28367121102329473, + "grad_norm": 0.9741342067718506, + "learning_rate": 1.8865758164721217e-05, + "loss": 3.8642, + "step": 25250 + }, + { + "epoch": 0.2842329361936379, + "grad_norm": 1.030258297920227, + "learning_rate": 1.8863511251418366e-05, + "loss": 3.8519, + "step": 25300 + }, + { + "epoch": 0.28479466136398107, + "grad_norm": 1.2043633460998535, + "learning_rate": 1.8861264338115515e-05, + "loss": 3.8263, + "step": 25350 + }, + { + "epoch": 0.2853563865343242, + "grad_norm": 1.0785086154937744, + "learning_rate": 1.8859017424812664e-05, + "loss": 3.7636, + "step": 25400 + }, + { + "epoch": 0.28591811170466735, + "grad_norm": 0.9269720911979675, + "learning_rate": 1.8856770511509814e-05, + "loss": 3.8372, + "step": 25450 + }, + { + "epoch": 0.28647983687501055, + "grad_norm": 1.099853277206421, + "learning_rate": 1.8854523598206963e-05, + "loss": 3.8628, + "step": 25500 + }, + { + "epoch": 0.2870415620453537, + "grad_norm": 1.1465600728988647, + "learning_rate": 1.8852276684904112e-05, + "loss": 3.8568, + "step": 25550 + }, + { + "epoch": 0.28760328721569683, + "grad_norm": 0.9685181975364685, + "learning_rate": 1.8850029771601265e-05, + "loss": 3.8208, + "step": 25600 + }, + { + "epoch": 0.28816501238604003, + "grad_norm": 1.1846619844436646, + "learning_rate": 1.8847782858298414e-05, + "loss": 3.8492, + "step": 25650 + }, + { + "epoch": 0.2887267375563832, + "grad_norm": 1.008122205734253, + "learning_rate": 1.8845535944995564e-05, + "loss": 3.7525, + "step": 25700 + }, + { + "epoch": 0.2892884627267263, + "grad_norm": 1.1963845491409302, + "learning_rate": 1.8843289031692713e-05, + "loss": 3.8768, + "step": 25750 + }, + { + "epoch": 0.28985018789706946, + "grad_norm": 1.0384691953659058, + "learning_rate": 1.8841042118389862e-05, + "loss": 3.7988, + "step": 25800 + }, + { + "epoch": 0.29041191306741265, + "grad_norm": 0.917803168296814, + "learning_rate": 1.8838795205087012e-05, + "loss": 3.8486, + "step": 25850 + }, + { + "epoch": 0.2909736382377558, + "grad_norm": 1.0689321756362915, + "learning_rate": 1.883654829178416e-05, + "loss": 3.8321, + "step": 25900 + }, + { + "epoch": 0.29153536340809894, + "grad_norm": 0.9739123582839966, + "learning_rate": 1.8834301378481314e-05, + "loss": 3.8251, + "step": 25950 + }, + { + "epoch": 0.29209708857844213, + "grad_norm": 0.859523594379425, + "learning_rate": 1.8832054465178463e-05, + "loss": 3.8407, + "step": 26000 + }, + { + "epoch": 0.2926588137487853, + "grad_norm": 1.5353630781173706, + "learning_rate": 1.8829807551875612e-05, + "loss": 3.8114, + "step": 26050 + }, + { + "epoch": 0.2932205389191284, + "grad_norm": 1.1663600206375122, + "learning_rate": 1.882756063857276e-05, + "loss": 3.8922, + "step": 26100 + }, + { + "epoch": 0.2937822640894716, + "grad_norm": 1.1306334733963013, + "learning_rate": 1.882531372526991e-05, + "loss": 3.8349, + "step": 26150 + }, + { + "epoch": 0.29434398925981475, + "grad_norm": 1.1433279514312744, + "learning_rate": 1.882306681196706e-05, + "loss": 3.7827, + "step": 26200 + }, + { + "epoch": 0.2949057144301579, + "grad_norm": 0.9676677584648132, + "learning_rate": 1.882081989866421e-05, + "loss": 3.7992, + "step": 26250 + }, + { + "epoch": 0.29546743960050104, + "grad_norm": 1.019875168800354, + "learning_rate": 1.8818572985361362e-05, + "loss": 3.813, + "step": 26300 + }, + { + "epoch": 0.29602916477084423, + "grad_norm": 1.047001600265503, + "learning_rate": 1.881632607205851e-05, + "loss": 3.8601, + "step": 26350 + }, + { + "epoch": 0.2965908899411874, + "grad_norm": 1.2332990169525146, + "learning_rate": 1.881407915875566e-05, + "loss": 3.8068, + "step": 26400 + }, + { + "epoch": 0.2971526151115305, + "grad_norm": 0.885316789150238, + "learning_rate": 1.881183224545281e-05, + "loss": 3.8525, + "step": 26450 + }, + { + "epoch": 0.2977143402818737, + "grad_norm": 1.4388920068740845, + "learning_rate": 1.880958533214996e-05, + "loss": 3.8855, + "step": 26500 + }, + { + "epoch": 0.29827606545221685, + "grad_norm": 1.2492674589157104, + "learning_rate": 1.880733841884711e-05, + "loss": 3.8343, + "step": 26550 + }, + { + "epoch": 0.29883779062256, + "grad_norm": 0.9600359797477722, + "learning_rate": 1.8805091505544258e-05, + "loss": 3.8629, + "step": 26600 + }, + { + "epoch": 0.29939951579290314, + "grad_norm": 1.4856162071228027, + "learning_rate": 1.880284459224141e-05, + "loss": 3.8172, + "step": 26650 + }, + { + "epoch": 0.29996124096324633, + "grad_norm": 1.16232168674469, + "learning_rate": 1.880059767893856e-05, + "loss": 3.8039, + "step": 26700 + }, + { + "epoch": 0.3005229661335895, + "grad_norm": 1.2946114540100098, + "learning_rate": 1.879835076563571e-05, + "loss": 3.739, + "step": 26750 + }, + { + "epoch": 0.3010846913039326, + "grad_norm": 1.1929129362106323, + "learning_rate": 1.879610385233286e-05, + "loss": 3.8478, + "step": 26800 + }, + { + "epoch": 0.3016464164742758, + "grad_norm": 1.0640249252319336, + "learning_rate": 1.8793856939030008e-05, + "loss": 3.7763, + "step": 26850 + }, + { + "epoch": 0.30220814164461896, + "grad_norm": 1.3911046981811523, + "learning_rate": 1.8791610025727158e-05, + "loss": 3.8613, + "step": 26900 + }, + { + "epoch": 0.3027698668149621, + "grad_norm": 1.0830105543136597, + "learning_rate": 1.8789408050690367e-05, + "loss": 3.854, + "step": 26950 + }, + { + "epoch": 0.3033315919853053, + "grad_norm": 1.2694368362426758, + "learning_rate": 1.8787161137387516e-05, + "loss": 3.824, + "step": 27000 + }, + { + "epoch": 0.30389331715564843, + "grad_norm": 1.2115559577941895, + "learning_rate": 1.8784914224084665e-05, + "loss": 3.8349, + "step": 27050 + }, + { + "epoch": 0.3044550423259916, + "grad_norm": 1.0922532081604004, + "learning_rate": 1.8782667310781815e-05, + "loss": 3.8274, + "step": 27100 + }, + { + "epoch": 0.3050167674963347, + "grad_norm": 1.0742512941360474, + "learning_rate": 1.8780420397478964e-05, + "loss": 3.772, + "step": 27150 + }, + { + "epoch": 0.3055784926666779, + "grad_norm": 1.2460834980010986, + "learning_rate": 1.877821842244217e-05, + "loss": 3.8323, + "step": 27200 + }, + { + "epoch": 0.30614021783702106, + "grad_norm": 1.0943710803985596, + "learning_rate": 1.8775971509139323e-05, + "loss": 3.8507, + "step": 27250 + }, + { + "epoch": 0.3067019430073642, + "grad_norm": 1.0860657691955566, + "learning_rate": 1.8773724595836472e-05, + "loss": 3.7839, + "step": 27300 + }, + { + "epoch": 0.3072636681777074, + "grad_norm": 0.9497836828231812, + "learning_rate": 1.877147768253362e-05, + "loss": 3.8384, + "step": 27350 + }, + { + "epoch": 0.30782539334805054, + "grad_norm": 0.932269811630249, + "learning_rate": 1.876923076923077e-05, + "loss": 3.8193, + "step": 27400 + }, + { + "epoch": 0.3083871185183937, + "grad_norm": 0.9704727530479431, + "learning_rate": 1.876698385592792e-05, + "loss": 3.7925, + "step": 27450 + }, + { + "epoch": 0.3089488436887369, + "grad_norm": 0.8907328248023987, + "learning_rate": 1.876473694262507e-05, + "loss": 3.8849, + "step": 27500 + }, + { + "epoch": 0.30951056885908, + "grad_norm": 1.0381135940551758, + "learning_rate": 1.876249002932222e-05, + "loss": 3.845, + "step": 27550 + }, + { + "epoch": 0.31007229402942316, + "grad_norm": 0.9771307706832886, + "learning_rate": 1.876024311601937e-05, + "loss": 3.7812, + "step": 27600 + }, + { + "epoch": 0.3106340191997663, + "grad_norm": 1.2536208629608154, + "learning_rate": 1.875799620271652e-05, + "loss": 3.8203, + "step": 27650 + }, + { + "epoch": 0.3111957443701095, + "grad_norm": 1.2103676795959473, + "learning_rate": 1.875574928941367e-05, + "loss": 3.8483, + "step": 27700 + }, + { + "epoch": 0.31175746954045264, + "grad_norm": 1.2546497583389282, + "learning_rate": 1.875350237611082e-05, + "loss": 3.8319, + "step": 27750 + }, + { + "epoch": 0.3123191947107958, + "grad_norm": 1.0846409797668457, + "learning_rate": 1.875125546280797e-05, + "loss": 3.8015, + "step": 27800 + }, + { + "epoch": 0.312880919881139, + "grad_norm": 1.2220613956451416, + "learning_rate": 1.8749053487771175e-05, + "loss": 3.7921, + "step": 27850 + }, + { + "epoch": 0.3134426450514821, + "grad_norm": 1.0720977783203125, + "learning_rate": 1.8746806574468327e-05, + "loss": 3.8055, + "step": 27900 + }, + { + "epoch": 0.31400437022182526, + "grad_norm": 1.0860999822616577, + "learning_rate": 1.8744559661165477e-05, + "loss": 3.8301, + "step": 27950 + }, + { + "epoch": 0.31456609539216845, + "grad_norm": 1.1940875053405762, + "learning_rate": 1.8742312747862626e-05, + "loss": 3.793, + "step": 28000 + }, + { + "epoch": 0.3151278205625116, + "grad_norm": 1.1205674409866333, + "learning_rate": 1.8740065834559775e-05, + "loss": 3.781, + "step": 28050 + }, + { + "epoch": 0.31568954573285474, + "grad_norm": 1.1552340984344482, + "learning_rate": 1.8737818921256924e-05, + "loss": 3.8237, + "step": 28100 + }, + { + "epoch": 0.3162512709031979, + "grad_norm": 1.0858317613601685, + "learning_rate": 1.8735572007954074e-05, + "loss": 3.8862, + "step": 28150 + }, + { + "epoch": 0.3168129960735411, + "grad_norm": 1.2606555223464966, + "learning_rate": 1.8733325094651223e-05, + "loss": 3.8343, + "step": 28200 + }, + { + "epoch": 0.3173747212438842, + "grad_norm": 1.202958583831787, + "learning_rate": 1.8731078181348376e-05, + "loss": 3.7782, + "step": 28250 + }, + { + "epoch": 0.31793644641422736, + "grad_norm": 1.0832439661026, + "learning_rate": 1.8728831268045525e-05, + "loss": 3.8162, + "step": 28300 + }, + { + "epoch": 0.31849817158457056, + "grad_norm": 4.111958980560303, + "learning_rate": 1.8726584354742674e-05, + "loss": 3.8364, + "step": 28350 + }, + { + "epoch": 0.3190598967549137, + "grad_norm": 0.9650757312774658, + "learning_rate": 1.8724337441439824e-05, + "loss": 3.8207, + "step": 28400 + }, + { + "epoch": 0.31962162192525684, + "grad_norm": 1.0663647651672363, + "learning_rate": 1.8722090528136973e-05, + "loss": 3.8145, + "step": 28450 + }, + { + "epoch": 0.3201833470956, + "grad_norm": 0.9724984169006348, + "learning_rate": 1.8719843614834122e-05, + "loss": 3.7738, + "step": 28500 + }, + { + "epoch": 0.3207450722659432, + "grad_norm": 1.0874778032302856, + "learning_rate": 1.8717596701531272e-05, + "loss": 3.8575, + "step": 28550 + }, + { + "epoch": 0.3213067974362863, + "grad_norm": 1.0395227670669556, + "learning_rate": 1.8715349788228424e-05, + "loss": 3.8257, + "step": 28600 + }, + { + "epoch": 0.32186852260662946, + "grad_norm": 0.9556145668029785, + "learning_rate": 1.8713102874925574e-05, + "loss": 3.8482, + "step": 28650 + }, + { + "epoch": 0.32243024777697266, + "grad_norm": 0.9420982599258423, + "learning_rate": 1.8710855961622723e-05, + "loss": 3.7761, + "step": 28700 + }, + { + "epoch": 0.3229919729473158, + "grad_norm": 0.8945040106773376, + "learning_rate": 1.8708609048319872e-05, + "loss": 3.8195, + "step": 28750 + }, + { + "epoch": 0.32355369811765894, + "grad_norm": 1.645750641822815, + "learning_rate": 1.870636213501702e-05, + "loss": 3.7286, + "step": 28800 + }, + { + "epoch": 0.32411542328800214, + "grad_norm": 0.9189350605010986, + "learning_rate": 1.870411522171417e-05, + "loss": 3.8004, + "step": 28850 + }, + { + "epoch": 0.3246771484583453, + "grad_norm": 1.14378023147583, + "learning_rate": 1.870186830841132e-05, + "loss": 3.7989, + "step": 28900 + }, + { + "epoch": 0.3252388736286884, + "grad_norm": 1.208552360534668, + "learning_rate": 1.8699621395108473e-05, + "loss": 3.7862, + "step": 28950 + }, + { + "epoch": 0.32580059879903156, + "grad_norm": 1.0845381021499634, + "learning_rate": 1.8697374481805622e-05, + "loss": 3.8527, + "step": 29000 + }, + { + "epoch": 0.32636232396937476, + "grad_norm": 0.9852644205093384, + "learning_rate": 1.869512756850277e-05, + "loss": 3.8067, + "step": 29050 + }, + { + "epoch": 0.3269240491397179, + "grad_norm": 1.0806910991668701, + "learning_rate": 1.869288065519992e-05, + "loss": 3.7952, + "step": 29100 + }, + { + "epoch": 0.32748577431006104, + "grad_norm": 1.2093960046768188, + "learning_rate": 1.869063374189707e-05, + "loss": 3.9072, + "step": 29150 + }, + { + "epoch": 0.32804749948040424, + "grad_norm": 1.2655662298202515, + "learning_rate": 1.868838682859422e-05, + "loss": 3.8486, + "step": 29200 + }, + { + "epoch": 0.3286092246507474, + "grad_norm": 1.1227494478225708, + "learning_rate": 1.868613991529137e-05, + "loss": 3.8285, + "step": 29250 + }, + { + "epoch": 0.3291709498210905, + "grad_norm": 1.0291639566421509, + "learning_rate": 1.868389300198852e-05, + "loss": 3.8551, + "step": 29300 + }, + { + "epoch": 0.3297326749914337, + "grad_norm": 1.0368773937225342, + "learning_rate": 1.868164608868567e-05, + "loss": 3.8432, + "step": 29350 + }, + { + "epoch": 0.33029440016177686, + "grad_norm": 1.1037966012954712, + "learning_rate": 1.867939917538282e-05, + "loss": 3.8626, + "step": 29400 + }, + { + "epoch": 0.33085612533212, + "grad_norm": 0.9336369037628174, + "learning_rate": 1.867715226207997e-05, + "loss": 3.7892, + "step": 29450 + }, + { + "epoch": 0.33141785050246314, + "grad_norm": 1.087012767791748, + "learning_rate": 1.867490534877712e-05, + "loss": 3.8418, + "step": 29500 + }, + { + "epoch": 0.33197957567280634, + "grad_norm": 1.1471550464630127, + "learning_rate": 1.8672658435474268e-05, + "loss": 3.7594, + "step": 29550 + }, + { + "epoch": 0.3325413008431495, + "grad_norm": 1.2329821586608887, + "learning_rate": 1.8670411522171418e-05, + "loss": 3.8126, + "step": 29600 + }, + { + "epoch": 0.3331030260134926, + "grad_norm": 1.287199854850769, + "learning_rate": 1.866816460886857e-05, + "loss": 3.8232, + "step": 29650 + }, + { + "epoch": 0.3336647511838358, + "grad_norm": 1.009225606918335, + "learning_rate": 1.866591769556572e-05, + "loss": 3.9107, + "step": 29700 + }, + { + "epoch": 0.33422647635417896, + "grad_norm": 1.088283896446228, + "learning_rate": 1.866367078226287e-05, + "loss": 3.8091, + "step": 29750 + }, + { + "epoch": 0.3347882015245221, + "grad_norm": 1.095376968383789, + "learning_rate": 1.8661423868960018e-05, + "loss": 3.7958, + "step": 29800 + }, + { + "epoch": 0.3353499266948653, + "grad_norm": 1.1659440994262695, + "learning_rate": 1.8659176955657167e-05, + "loss": 3.7734, + "step": 29850 + }, + { + "epoch": 0.33591165186520844, + "grad_norm": 1.0501476526260376, + "learning_rate": 1.8656930042354317e-05, + "loss": 3.7342, + "step": 29900 + }, + { + "epoch": 0.3364733770355516, + "grad_norm": 1.17960786819458, + "learning_rate": 1.8654683129051466e-05, + "loss": 3.7517, + "step": 29950 + }, + { + "epoch": 0.3370351022058947, + "grad_norm": 1.0510057210922241, + "learning_rate": 1.865243621574862e-05, + "loss": 3.8809, + "step": 30000 + }, + { + "epoch": 0.3375968273762379, + "grad_norm": 0.9583242535591125, + "learning_rate": 1.8650189302445768e-05, + "loss": 3.7781, + "step": 30050 + }, + { + "epoch": 0.33815855254658106, + "grad_norm": 1.257326364517212, + "learning_rate": 1.8647942389142917e-05, + "loss": 3.7793, + "step": 30100 + }, + { + "epoch": 0.3387202777169242, + "grad_norm": 0.9945911169052124, + "learning_rate": 1.8645695475840067e-05, + "loss": 3.7762, + "step": 30150 + }, + { + "epoch": 0.3392820028872674, + "grad_norm": 1.1452234983444214, + "learning_rate": 1.8643448562537216e-05, + "loss": 3.8735, + "step": 30200 + }, + { + "epoch": 0.33984372805761054, + "grad_norm": 1.0031211376190186, + "learning_rate": 1.8641201649234365e-05, + "loss": 3.8606, + "step": 30250 + }, + { + "epoch": 0.3404054532279537, + "grad_norm": 1.2691720724105835, + "learning_rate": 1.8638954735931515e-05, + "loss": 3.7905, + "step": 30300 + }, + { + "epoch": 0.3409671783982968, + "grad_norm": 1.214430332183838, + "learning_rate": 1.8636707822628664e-05, + "loss": 3.8372, + "step": 30350 + }, + { + "epoch": 0.34152890356864, + "grad_norm": 1.068723201751709, + "learning_rate": 1.8634460909325817e-05, + "loss": 3.8458, + "step": 30400 + }, + { + "epoch": 0.34209062873898316, + "grad_norm": 1.2963320016860962, + "learning_rate": 1.8632213996022966e-05, + "loss": 3.843, + "step": 30450 + }, + { + "epoch": 0.3426523539093263, + "grad_norm": 0.9694681763648987, + "learning_rate": 1.8629967082720115e-05, + "loss": 3.784, + "step": 30500 + }, + { + "epoch": 0.3432140790796695, + "grad_norm": 0.9249500632286072, + "learning_rate": 1.8627720169417265e-05, + "loss": 3.7642, + "step": 30550 + }, + { + "epoch": 0.34377580425001264, + "grad_norm": 0.9498476982116699, + "learning_rate": 1.8625473256114414e-05, + "loss": 3.8571, + "step": 30600 + }, + { + "epoch": 0.3443375294203558, + "grad_norm": 1.084003210067749, + "learning_rate": 1.8623226342811563e-05, + "loss": 3.7847, + "step": 30650 + }, + { + "epoch": 0.344899254590699, + "grad_norm": 1.0181688070297241, + "learning_rate": 1.8620979429508713e-05, + "loss": 3.7549, + "step": 30700 + }, + { + "epoch": 0.3454609797610421, + "grad_norm": 1.1446342468261719, + "learning_rate": 1.8618732516205865e-05, + "loss": 3.8927, + "step": 30750 + }, + { + "epoch": 0.34602270493138526, + "grad_norm": 1.0579931735992432, + "learning_rate": 1.8616485602903015e-05, + "loss": 3.8483, + "step": 30800 + }, + { + "epoch": 0.3465844301017284, + "grad_norm": 1.0482735633850098, + "learning_rate": 1.8614238689600164e-05, + "loss": 3.909, + "step": 30850 + }, + { + "epoch": 0.3471461552720716, + "grad_norm": 1.0333802700042725, + "learning_rate": 1.8611991776297313e-05, + "loss": 3.8737, + "step": 30900 + }, + { + "epoch": 0.34770788044241474, + "grad_norm": 1.035236120223999, + "learning_rate": 1.8609744862994463e-05, + "loss": 3.7939, + "step": 30950 + }, + { + "epoch": 0.3482696056127579, + "grad_norm": 1.209030270576477, + "learning_rate": 1.8607497949691612e-05, + "loss": 3.9275, + "step": 31000 + }, + { + "epoch": 0.3488313307831011, + "grad_norm": 1.029096245765686, + "learning_rate": 1.860525103638876e-05, + "loss": 3.763, + "step": 31050 + }, + { + "epoch": 0.3493930559534442, + "grad_norm": 0.9945694208145142, + "learning_rate": 1.8603004123085914e-05, + "loss": 3.8589, + "step": 31100 + }, + { + "epoch": 0.34995478112378736, + "grad_norm": 1.048553228378296, + "learning_rate": 1.8600757209783063e-05, + "loss": 3.8801, + "step": 31150 + }, + { + "epoch": 0.35051650629413056, + "grad_norm": 1.2170648574829102, + "learning_rate": 1.8598510296480213e-05, + "loss": 3.8499, + "step": 31200 + }, + { + "epoch": 0.3510782314644737, + "grad_norm": 0.8793922662734985, + "learning_rate": 1.8596263383177362e-05, + "loss": 3.9009, + "step": 31250 + }, + { + "epoch": 0.35163995663481684, + "grad_norm": 0.9794766306877136, + "learning_rate": 1.859401646987451e-05, + "loss": 3.8049, + "step": 31300 + }, + { + "epoch": 0.35220168180516, + "grad_norm": 1.1491682529449463, + "learning_rate": 1.859176955657166e-05, + "loss": 3.7784, + "step": 31350 + }, + { + "epoch": 0.3527634069755032, + "grad_norm": 0.9648553133010864, + "learning_rate": 1.858952264326881e-05, + "loss": 3.8466, + "step": 31400 + }, + { + "epoch": 0.3533251321458463, + "grad_norm": 1.122705101966858, + "learning_rate": 1.8587275729965963e-05, + "loss": 3.8339, + "step": 31450 + }, + { + "epoch": 0.35388685731618946, + "grad_norm": 1.4351998567581177, + "learning_rate": 1.8585028816663112e-05, + "loss": 3.823, + "step": 31500 + }, + { + "epoch": 0.35444858248653266, + "grad_norm": 1.0422778129577637, + "learning_rate": 1.858278190336026e-05, + "loss": 3.851, + "step": 31550 + }, + { + "epoch": 0.3550103076568758, + "grad_norm": 1.0606135129928589, + "learning_rate": 1.858053499005741e-05, + "loss": 3.8408, + "step": 31600 + }, + { + "epoch": 0.35557203282721894, + "grad_norm": 1.278016448020935, + "learning_rate": 1.857828807675456e-05, + "loss": 3.8172, + "step": 31650 + }, + { + "epoch": 0.35613375799756214, + "grad_norm": 0.9278510808944702, + "learning_rate": 1.857604116345171e-05, + "loss": 3.7797, + "step": 31700 + }, + { + "epoch": 0.3566954831679053, + "grad_norm": 1.0965937376022339, + "learning_rate": 1.857379425014886e-05, + "loss": 3.7864, + "step": 31750 + }, + { + "epoch": 0.3572572083382484, + "grad_norm": 1.060134768486023, + "learning_rate": 1.857154733684601e-05, + "loss": 3.831, + "step": 31800 + }, + { + "epoch": 0.35781893350859156, + "grad_norm": 1.3302299976348877, + "learning_rate": 1.856930042354316e-05, + "loss": 3.8384, + "step": 31850 + }, + { + "epoch": 0.35838065867893476, + "grad_norm": 1.0352579355239868, + "learning_rate": 1.856705351024031e-05, + "loss": 3.7721, + "step": 31900 + }, + { + "epoch": 0.3589423838492779, + "grad_norm": 1.1207737922668457, + "learning_rate": 1.856480659693746e-05, + "loss": 3.8045, + "step": 31950 + }, + { + "epoch": 0.35950410901962104, + "grad_norm": 0.9285112619400024, + "learning_rate": 1.856255968363461e-05, + "loss": 3.751, + "step": 32000 + }, + { + "epoch": 0.36006583418996424, + "grad_norm": 1.2063809633255005, + "learning_rate": 1.8560312770331758e-05, + "loss": 3.799, + "step": 32050 + }, + { + "epoch": 0.3606275593603074, + "grad_norm": 1.1076829433441162, + "learning_rate": 1.8558065857028907e-05, + "loss": 3.8278, + "step": 32100 + }, + { + "epoch": 0.3611892845306505, + "grad_norm": 1.0293713808059692, + "learning_rate": 1.855581894372606e-05, + "loss": 3.8191, + "step": 32150 + }, + { + "epoch": 0.36175100970099366, + "grad_norm": 1.1451561450958252, + "learning_rate": 1.855357203042321e-05, + "loss": 3.8873, + "step": 32200 + }, + { + "epoch": 0.36231273487133686, + "grad_norm": 1.0879944562911987, + "learning_rate": 1.855132511712036e-05, + "loss": 3.7784, + "step": 32250 + }, + { + "epoch": 0.36287446004168, + "grad_norm": 1.052733302116394, + "learning_rate": 1.8549078203817508e-05, + "loss": 3.8414, + "step": 32300 + }, + { + "epoch": 0.36343618521202314, + "grad_norm": 1.3449853658676147, + "learning_rate": 1.8546831290514657e-05, + "loss": 3.7888, + "step": 32350 + }, + { + "epoch": 0.36399791038236634, + "grad_norm": 0.9949585199356079, + "learning_rate": 1.8544584377211806e-05, + "loss": 3.8237, + "step": 32400 + }, + { + "epoch": 0.3645596355527095, + "grad_norm": 0.9596227407455444, + "learning_rate": 1.8542337463908956e-05, + "loss": 3.8026, + "step": 32450 + }, + { + "epoch": 0.3651213607230526, + "grad_norm": 1.0283395051956177, + "learning_rate": 1.854009055060611e-05, + "loss": 3.8657, + "step": 32500 + }, + { + "epoch": 0.3656830858933958, + "grad_norm": 1.1715400218963623, + "learning_rate": 1.8537843637303258e-05, + "loss": 3.7421, + "step": 32550 + }, + { + "epoch": 0.36624481106373896, + "grad_norm": 0.9519177675247192, + "learning_rate": 1.8535596724000407e-05, + "loss": 3.7558, + "step": 32600 + }, + { + "epoch": 0.3668065362340821, + "grad_norm": 1.0189740657806396, + "learning_rate": 1.8533349810697556e-05, + "loss": 3.8338, + "step": 32650 + }, + { + "epoch": 0.36736826140442524, + "grad_norm": 1.0201365947723389, + "learning_rate": 1.8531102897394706e-05, + "loss": 3.7984, + "step": 32700 + }, + { + "epoch": 0.36792998657476844, + "grad_norm": 1.0041818618774414, + "learning_rate": 1.8528855984091855e-05, + "loss": 3.8174, + "step": 32750 + }, + { + "epoch": 0.3684917117451116, + "grad_norm": 0.9530069231987, + "learning_rate": 1.8526609070789004e-05, + "loss": 3.8309, + "step": 32800 + }, + { + "epoch": 0.3690534369154547, + "grad_norm": 1.1176979541778564, + "learning_rate": 1.8524362157486157e-05, + "loss": 3.8478, + "step": 32850 + }, + { + "epoch": 0.3696151620857979, + "grad_norm": 1.1033034324645996, + "learning_rate": 1.8522115244183306e-05, + "loss": 3.8247, + "step": 32900 + }, + { + "epoch": 0.37017688725614106, + "grad_norm": 1.0508520603179932, + "learning_rate": 1.8519868330880456e-05, + "loss": 3.8233, + "step": 32950 + }, + { + "epoch": 0.3707386124264842, + "grad_norm": 1.377721905708313, + "learning_rate": 1.8517621417577605e-05, + "loss": 3.7988, + "step": 33000 + }, + { + "epoch": 0.3713003375968274, + "grad_norm": 1.0736579895019531, + "learning_rate": 1.8515374504274754e-05, + "loss": 3.8424, + "step": 33050 + }, + { + "epoch": 0.37186206276717054, + "grad_norm": 1.0143682956695557, + "learning_rate": 1.8513127590971904e-05, + "loss": 3.8026, + "step": 33100 + }, + { + "epoch": 0.3724237879375137, + "grad_norm": 1.228042483329773, + "learning_rate": 1.8510880677669053e-05, + "loss": 3.7849, + "step": 33150 + }, + { + "epoch": 0.3729855131078568, + "grad_norm": 1.1448063850402832, + "learning_rate": 1.8508633764366206e-05, + "loss": 3.7735, + "step": 33200 + }, + { + "epoch": 0.3735472382782, + "grad_norm": 1.9350948333740234, + "learning_rate": 1.8506386851063355e-05, + "loss": 3.7989, + "step": 33250 + }, + { + "epoch": 0.37410896344854316, + "grad_norm": 0.9634103178977966, + "learning_rate": 1.8504139937760504e-05, + "loss": 3.8728, + "step": 33300 + }, + { + "epoch": 0.3746706886188863, + "grad_norm": 1.0275721549987793, + "learning_rate": 1.8501893024457654e-05, + "loss": 3.841, + "step": 33350 + }, + { + "epoch": 0.3752324137892295, + "grad_norm": 1.592858076095581, + "learning_rate": 1.8499646111154803e-05, + "loss": 3.8184, + "step": 33400 + }, + { + "epoch": 0.37579413895957264, + "grad_norm": 0.8911204934120178, + "learning_rate": 1.8497399197851952e-05, + "loss": 3.8351, + "step": 33450 + }, + { + "epoch": 0.3763558641299158, + "grad_norm": 1.016668438911438, + "learning_rate": 1.849519722281516e-05, + "loss": 3.789, + "step": 33500 + }, + { + "epoch": 0.376917589300259, + "grad_norm": 1.1443250179290771, + "learning_rate": 1.849295030951231e-05, + "loss": 3.8201, + "step": 33550 + }, + { + "epoch": 0.3774793144706021, + "grad_norm": 1.1348165273666382, + "learning_rate": 1.849070339620946e-05, + "loss": 3.775, + "step": 33600 + }, + { + "epoch": 0.37804103964094526, + "grad_norm": 0.9641530513763428, + "learning_rate": 1.848845648290661e-05, + "loss": 3.8072, + "step": 33650 + }, + { + "epoch": 0.3786027648112884, + "grad_norm": 1.157254934310913, + "learning_rate": 1.848620956960376e-05, + "loss": 3.8643, + "step": 33700 + }, + { + "epoch": 0.3791644899816316, + "grad_norm": 1.276639699935913, + "learning_rate": 1.8483962656300908e-05, + "loss": 3.8395, + "step": 33750 + }, + { + "epoch": 0.37972621515197474, + "grad_norm": 1.1866374015808105, + "learning_rate": 1.8481715742998057e-05, + "loss": 3.8799, + "step": 33800 + }, + { + "epoch": 0.3802879403223179, + "grad_norm": 1.016951322555542, + "learning_rate": 1.847946882969521e-05, + "loss": 3.8445, + "step": 33850 + }, + { + "epoch": 0.3808496654926611, + "grad_norm": 1.0595694780349731, + "learning_rate": 1.847722191639236e-05, + "loss": 3.7482, + "step": 33900 + }, + { + "epoch": 0.3814113906630042, + "grad_norm": 1.0442867279052734, + "learning_rate": 1.847497500308951e-05, + "loss": 3.9199, + "step": 33950 + }, + { + "epoch": 0.38197311583334737, + "grad_norm": 1.3628323078155518, + "learning_rate": 1.8472728089786658e-05, + "loss": 3.836, + "step": 34000 + }, + { + "epoch": 0.3825348410036905, + "grad_norm": 1.1050171852111816, + "learning_rate": 1.8470481176483807e-05, + "loss": 3.7957, + "step": 34050 + }, + { + "epoch": 0.3830965661740337, + "grad_norm": 1.0539454221725464, + "learning_rate": 1.8468234263180957e-05, + "loss": 3.7208, + "step": 34100 + }, + { + "epoch": 0.38365829134437685, + "grad_norm": 1.0863726139068604, + "learning_rate": 1.8465987349878106e-05, + "loss": 3.7781, + "step": 34150 + }, + { + "epoch": 0.38422001651472, + "grad_norm": 0.9803478121757507, + "learning_rate": 1.846374043657526e-05, + "loss": 3.8272, + "step": 34200 + }, + { + "epoch": 0.3847817416850632, + "grad_norm": 1.2689176797866821, + "learning_rate": 1.8461493523272408e-05, + "loss": 3.8052, + "step": 34250 + }, + { + "epoch": 0.3853434668554063, + "grad_norm": 0.9259783029556274, + "learning_rate": 1.8459246609969557e-05, + "loss": 3.8363, + "step": 34300 + }, + { + "epoch": 0.38590519202574947, + "grad_norm": 1.2102526426315308, + "learning_rate": 1.8456999696666707e-05, + "loss": 3.8209, + "step": 34350 + }, + { + "epoch": 0.38646691719609266, + "grad_norm": 1.1629327535629272, + "learning_rate": 1.8454752783363856e-05, + "loss": 3.7606, + "step": 34400 + }, + { + "epoch": 0.3870286423664358, + "grad_norm": 1.1801527738571167, + "learning_rate": 1.8452505870061005e-05, + "loss": 3.8129, + "step": 34450 + }, + { + "epoch": 0.38759036753677895, + "grad_norm": 1.0275895595550537, + "learning_rate": 1.8450258956758155e-05, + "loss": 3.7998, + "step": 34500 + }, + { + "epoch": 0.3881520927071221, + "grad_norm": 0.9174814820289612, + "learning_rate": 1.8448012043455307e-05, + "loss": 3.8167, + "step": 34550 + }, + { + "epoch": 0.3887138178774653, + "grad_norm": 0.9516996741294861, + "learning_rate": 1.8445765130152457e-05, + "loss": 3.7663, + "step": 34600 + }, + { + "epoch": 0.3892755430478084, + "grad_norm": 1.0816078186035156, + "learning_rate": 1.8443518216849606e-05, + "loss": 3.7428, + "step": 34650 + }, + { + "epoch": 0.38983726821815157, + "grad_norm": 0.9516173005104065, + "learning_rate": 1.8441271303546755e-05, + "loss": 3.7543, + "step": 34700 + }, + { + "epoch": 0.39039899338849476, + "grad_norm": 1.052913784980774, + "learning_rate": 1.8439024390243905e-05, + "loss": 3.8121, + "step": 34750 + }, + { + "epoch": 0.3909607185588379, + "grad_norm": 1.1837207078933716, + "learning_rate": 1.8436777476941054e-05, + "loss": 3.795, + "step": 34800 + }, + { + "epoch": 0.39152244372918105, + "grad_norm": 1.059316873550415, + "learning_rate": 1.8434530563638203e-05, + "loss": 3.8403, + "step": 34850 + }, + { + "epoch": 0.39208416889952424, + "grad_norm": 1.0185049772262573, + "learning_rate": 1.8432283650335353e-05, + "loss": 3.9025, + "step": 34900 + }, + { + "epoch": 0.3926458940698674, + "grad_norm": 0.9250991344451904, + "learning_rate": 1.8430036737032505e-05, + "loss": 3.8261, + "step": 34950 + }, + { + "epoch": 0.3932076192402105, + "grad_norm": 1.0445061922073364, + "learning_rate": 1.8427789823729655e-05, + "loss": 3.805, + "step": 35000 + }, + { + "epoch": 0.39376934441055367, + "grad_norm": 1.1061943769454956, + "learning_rate": 1.8425542910426804e-05, + "loss": 3.8592, + "step": 35050 + }, + { + "epoch": 0.39433106958089686, + "grad_norm": 1.0674915313720703, + "learning_rate": 1.8423295997123953e-05, + "loss": 3.8049, + "step": 35100 + }, + { + "epoch": 0.39489279475124, + "grad_norm": 1.020250678062439, + "learning_rate": 1.8421049083821102e-05, + "loss": 3.8215, + "step": 35150 + }, + { + "epoch": 0.39545451992158315, + "grad_norm": 1.0812455415725708, + "learning_rate": 1.8418802170518252e-05, + "loss": 3.8594, + "step": 35200 + }, + { + "epoch": 0.39601624509192634, + "grad_norm": 1.0903102159500122, + "learning_rate": 1.84165552572154e-05, + "loss": 3.7893, + "step": 35250 + }, + { + "epoch": 0.3965779702622695, + "grad_norm": 0.9832062125205994, + "learning_rate": 1.8414308343912554e-05, + "loss": 3.7683, + "step": 35300 + }, + { + "epoch": 0.3971396954326126, + "grad_norm": 1.0267248153686523, + "learning_rate": 1.8412061430609703e-05, + "loss": 3.8308, + "step": 35350 + }, + { + "epoch": 0.3977014206029558, + "grad_norm": 0.9348560571670532, + "learning_rate": 1.8409814517306852e-05, + "loss": 3.7979, + "step": 35400 + }, + { + "epoch": 0.39826314577329897, + "grad_norm": 1.093872308731079, + "learning_rate": 1.8407567604004002e-05, + "loss": 3.7905, + "step": 35450 + }, + { + "epoch": 0.3988248709436421, + "grad_norm": 1.1041386127471924, + "learning_rate": 1.840532069070115e-05, + "loss": 3.8041, + "step": 35500 + }, + { + "epoch": 0.39938659611398525, + "grad_norm": 1.947174072265625, + "learning_rate": 1.84030737773983e-05, + "loss": 3.8064, + "step": 35550 + }, + { + "epoch": 0.39994832128432845, + "grad_norm": 1.0973021984100342, + "learning_rate": 1.840082686409545e-05, + "loss": 3.8587, + "step": 35600 + }, + { + "epoch": 0.4005100464546716, + "grad_norm": 1.0880939960479736, + "learning_rate": 1.8398579950792602e-05, + "loss": 3.7737, + "step": 35650 + }, + { + "epoch": 0.40107177162501473, + "grad_norm": 1.2912418842315674, + "learning_rate": 1.8396333037489752e-05, + "loss": 3.8031, + "step": 35700 + }, + { + "epoch": 0.4016334967953579, + "grad_norm": 1.2398872375488281, + "learning_rate": 1.83940861241869e-05, + "loss": 3.7387, + "step": 35750 + }, + { + "epoch": 0.40219522196570107, + "grad_norm": 1.6796479225158691, + "learning_rate": 1.839183921088405e-05, + "loss": 3.8944, + "step": 35800 + }, + { + "epoch": 0.4027569471360442, + "grad_norm": 1.0580264329910278, + "learning_rate": 1.83895922975812e-05, + "loss": 3.7805, + "step": 35850 + }, + { + "epoch": 0.4033186723063874, + "grad_norm": 1.1395939588546753, + "learning_rate": 1.838734538427835e-05, + "loss": 3.7848, + "step": 35900 + }, + { + "epoch": 0.40388039747673055, + "grad_norm": 1.790340542793274, + "learning_rate": 1.83850984709755e-05, + "loss": 3.7637, + "step": 35950 + }, + { + "epoch": 0.4044421226470737, + "grad_norm": 1.0293999910354614, + "learning_rate": 1.838285155767265e-05, + "loss": 3.7648, + "step": 36000 + }, + { + "epoch": 0.40500384781741683, + "grad_norm": 1.108468770980835, + "learning_rate": 1.83806046443698e-05, + "loss": 3.6868, + "step": 36050 + }, + { + "epoch": 0.40556557298776, + "grad_norm": 0.9420512914657593, + "learning_rate": 1.837835773106695e-05, + "loss": 3.7962, + "step": 36100 + }, + { + "epoch": 0.40612729815810317, + "grad_norm": 1.288345456123352, + "learning_rate": 1.83761108177641e-05, + "loss": 3.8194, + "step": 36150 + }, + { + "epoch": 0.4066890233284463, + "grad_norm": 0.9674716591835022, + "learning_rate": 1.8373863904461248e-05, + "loss": 3.7895, + "step": 36200 + }, + { + "epoch": 0.4072507484987895, + "grad_norm": 1.0295203924179077, + "learning_rate": 1.8371616991158398e-05, + "loss": 3.758, + "step": 36250 + }, + { + "epoch": 0.40781247366913265, + "grad_norm": 1.009635329246521, + "learning_rate": 1.8369370077855547e-05, + "loss": 3.8189, + "step": 36300 + }, + { + "epoch": 0.4083741988394758, + "grad_norm": 1.264251947402954, + "learning_rate": 1.83671231645527e-05, + "loss": 3.7445, + "step": 36350 + }, + { + "epoch": 0.40893592400981893, + "grad_norm": 1.1355700492858887, + "learning_rate": 1.836487625124985e-05, + "loss": 3.7772, + "step": 36400 + }, + { + "epoch": 0.4094976491801621, + "grad_norm": 1.634529709815979, + "learning_rate": 1.8362629337946998e-05, + "loss": 3.8384, + "step": 36450 + }, + { + "epoch": 0.41005937435050527, + "grad_norm": 1.0426417589187622, + "learning_rate": 1.8360382424644148e-05, + "loss": 3.7586, + "step": 36500 + }, + { + "epoch": 0.4106210995208484, + "grad_norm": 1.0055906772613525, + "learning_rate": 1.8358135511341297e-05, + "loss": 3.7987, + "step": 36550 + }, + { + "epoch": 0.4111828246911916, + "grad_norm": 1.060428261756897, + "learning_rate": 1.8355888598038446e-05, + "loss": 3.8098, + "step": 36600 + }, + { + "epoch": 0.41174454986153475, + "grad_norm": 1.1187878847122192, + "learning_rate": 1.8353641684735596e-05, + "loss": 3.7981, + "step": 36650 + }, + { + "epoch": 0.4123062750318779, + "grad_norm": 1.1359686851501465, + "learning_rate": 1.8351394771432748e-05, + "loss": 3.8966, + "step": 36700 + }, + { + "epoch": 0.4128680002022211, + "grad_norm": 1.1621538400650024, + "learning_rate": 1.834919279639595e-05, + "loss": 3.8835, + "step": 36750 + }, + { + "epoch": 0.4134297253725642, + "grad_norm": 1.0647566318511963, + "learning_rate": 1.83469458830931e-05, + "loss": 3.8796, + "step": 36800 + }, + { + "epoch": 0.41399145054290737, + "grad_norm": 1.1750829219818115, + "learning_rate": 1.8344698969790253e-05, + "loss": 3.8571, + "step": 36850 + }, + { + "epoch": 0.4145531757132505, + "grad_norm": 1.250691294670105, + "learning_rate": 1.8342452056487402e-05, + "loss": 3.877, + "step": 36900 + }, + { + "epoch": 0.4151149008835937, + "grad_norm": 1.0842593908309937, + "learning_rate": 1.834020514318455e-05, + "loss": 3.738, + "step": 36950 + }, + { + "epoch": 0.41567662605393685, + "grad_norm": 1.1104592084884644, + "learning_rate": 1.83379582298817e-05, + "loss": 3.7505, + "step": 37000 + }, + { + "epoch": 0.41623835122428, + "grad_norm": 1.1149715185165405, + "learning_rate": 1.833571131657885e-05, + "loss": 3.8392, + "step": 37050 + }, + { + "epoch": 0.4168000763946232, + "grad_norm": 1.08084237575531, + "learning_rate": 1.8333464403276e-05, + "loss": 3.7977, + "step": 37100 + }, + { + "epoch": 0.41736180156496633, + "grad_norm": 1.0322520732879639, + "learning_rate": 1.833121748997315e-05, + "loss": 3.7871, + "step": 37150 + }, + { + "epoch": 0.41792352673530947, + "grad_norm": 1.0011364221572876, + "learning_rate": 1.8328970576670298e-05, + "loss": 3.8695, + "step": 37200 + }, + { + "epoch": 0.41848525190565267, + "grad_norm": 1.4675151109695435, + "learning_rate": 1.832672366336745e-05, + "loss": 3.8172, + "step": 37250 + }, + { + "epoch": 0.4190469770759958, + "grad_norm": 1.182085394859314, + "learning_rate": 1.83244767500646e-05, + "loss": 3.6947, + "step": 37300 + }, + { + "epoch": 0.41960870224633895, + "grad_norm": 1.2066044807434082, + "learning_rate": 1.832222983676175e-05, + "loss": 3.8281, + "step": 37350 + }, + { + "epoch": 0.4201704274166821, + "grad_norm": 1.1199288368225098, + "learning_rate": 1.83199829234589e-05, + "loss": 3.7712, + "step": 37400 + }, + { + "epoch": 0.4207321525870253, + "grad_norm": 1.0046926736831665, + "learning_rate": 1.8317736010156048e-05, + "loss": 3.8223, + "step": 37450 + }, + { + "epoch": 0.42129387775736843, + "grad_norm": 1.0446561574935913, + "learning_rate": 1.8315489096853197e-05, + "loss": 3.8042, + "step": 37500 + }, + { + "epoch": 0.42185560292771157, + "grad_norm": 1.0896804332733154, + "learning_rate": 1.8313242183550347e-05, + "loss": 3.7873, + "step": 37550 + }, + { + "epoch": 0.42241732809805477, + "grad_norm": 1.162729024887085, + "learning_rate": 1.83109952702475e-05, + "loss": 3.9176, + "step": 37600 + }, + { + "epoch": 0.4229790532683979, + "grad_norm": 1.2584238052368164, + "learning_rate": 1.830874835694465e-05, + "loss": 3.719, + "step": 37650 + }, + { + "epoch": 0.42354077843874105, + "grad_norm": 0.991865873336792, + "learning_rate": 1.8306501443641798e-05, + "loss": 3.786, + "step": 37700 + }, + { + "epoch": 0.42410250360908425, + "grad_norm": 1.1252045631408691, + "learning_rate": 1.8304254530338947e-05, + "loss": 3.8155, + "step": 37750 + }, + { + "epoch": 0.4246642287794274, + "grad_norm": 1.1655724048614502, + "learning_rate": 1.8302007617036097e-05, + "loss": 3.7323, + "step": 37800 + }, + { + "epoch": 0.42522595394977053, + "grad_norm": 1.1222635507583618, + "learning_rate": 1.8299760703733246e-05, + "loss": 3.8007, + "step": 37850 + }, + { + "epoch": 0.42578767912011367, + "grad_norm": 1.05940842628479, + "learning_rate": 1.8297513790430395e-05, + "loss": 3.7313, + "step": 37900 + }, + { + "epoch": 0.42634940429045687, + "grad_norm": 1.0890491008758545, + "learning_rate": 1.8295266877127548e-05, + "loss": 3.8056, + "step": 37950 + }, + { + "epoch": 0.4269111294608, + "grad_norm": 1.1040856838226318, + "learning_rate": 1.8293019963824697e-05, + "loss": 3.7458, + "step": 38000 + }, + { + "epoch": 0.42747285463114315, + "grad_norm": 1.5603495836257935, + "learning_rate": 1.8290773050521847e-05, + "loss": 3.7967, + "step": 38050 + }, + { + "epoch": 0.42803457980148635, + "grad_norm": 0.9499339461326599, + "learning_rate": 1.8288526137218996e-05, + "loss": 3.7156, + "step": 38100 + }, + { + "epoch": 0.4285963049718295, + "grad_norm": 0.9521780610084534, + "learning_rate": 1.8286279223916145e-05, + "loss": 3.7458, + "step": 38150 + }, + { + "epoch": 0.42915803014217263, + "grad_norm": 1.0303791761398315, + "learning_rate": 1.8284032310613295e-05, + "loss": 3.7431, + "step": 38200 + }, + { + "epoch": 0.42971975531251577, + "grad_norm": 1.2307860851287842, + "learning_rate": 1.8281785397310444e-05, + "loss": 3.7695, + "step": 38250 + }, + { + "epoch": 0.43028148048285897, + "grad_norm": 1.1721601486206055, + "learning_rate": 1.8279538484007597e-05, + "loss": 3.9299, + "step": 38300 + }, + { + "epoch": 0.4308432056532021, + "grad_norm": 1.0966248512268066, + "learning_rate": 1.8277291570704746e-05, + "loss": 3.7414, + "step": 38350 + }, + { + "epoch": 0.43140493082354525, + "grad_norm": 1.1516059637069702, + "learning_rate": 1.8275044657401895e-05, + "loss": 3.7757, + "step": 38400 + }, + { + "epoch": 0.43196665599388845, + "grad_norm": 1.0876235961914062, + "learning_rate": 1.8272797744099044e-05, + "loss": 3.7672, + "step": 38450 + }, + { + "epoch": 0.4325283811642316, + "grad_norm": 0.9670385718345642, + "learning_rate": 1.8270550830796194e-05, + "loss": 3.7899, + "step": 38500 + }, + { + "epoch": 0.43309010633457473, + "grad_norm": 1.2033318281173706, + "learning_rate": 1.8268303917493343e-05, + "loss": 3.8375, + "step": 38550 + }, + { + "epoch": 0.43365183150491793, + "grad_norm": 1.1501555442810059, + "learning_rate": 1.8266101942456552e-05, + "loss": 3.8231, + "step": 38600 + }, + { + "epoch": 0.43421355667526107, + "grad_norm": 1.027575969696045, + "learning_rate": 1.8263855029153702e-05, + "loss": 3.7934, + "step": 38650 + }, + { + "epoch": 0.4347752818456042, + "grad_norm": 1.1309833526611328, + "learning_rate": 1.826160811585085e-05, + "loss": 3.7897, + "step": 38700 + }, + { + "epoch": 0.43533700701594735, + "grad_norm": 1.0121960639953613, + "learning_rate": 1.8259361202548e-05, + "loss": 3.7848, + "step": 38750 + }, + { + "epoch": 0.43589873218629055, + "grad_norm": 1.1874136924743652, + "learning_rate": 1.825711428924515e-05, + "loss": 3.8107, + "step": 38800 + }, + { + "epoch": 0.4364604573566337, + "grad_norm": 1.5305979251861572, + "learning_rate": 1.82548673759423e-05, + "loss": 3.8421, + "step": 38850 + }, + { + "epoch": 0.43702218252697683, + "grad_norm": 1.0823510885238647, + "learning_rate": 1.825262046263945e-05, + "loss": 3.8185, + "step": 38900 + }, + { + "epoch": 0.43758390769732003, + "grad_norm": 0.98363196849823, + "learning_rate": 1.82503735493366e-05, + "loss": 3.7812, + "step": 38950 + }, + { + "epoch": 0.43814563286766317, + "grad_norm": 1.1372807025909424, + "learning_rate": 1.824812663603375e-05, + "loss": 3.8593, + "step": 39000 + }, + { + "epoch": 0.4387073580380063, + "grad_norm": 1.0772751569747925, + "learning_rate": 1.82458797227309e-05, + "loss": 3.8541, + "step": 39050 + }, + { + "epoch": 0.4392690832083495, + "grad_norm": 1.145782470703125, + "learning_rate": 1.824363280942805e-05, + "loss": 3.855, + "step": 39100 + }, + { + "epoch": 0.43983080837869265, + "grad_norm": 0.8724556565284729, + "learning_rate": 1.8241385896125198e-05, + "loss": 3.7936, + "step": 39150 + }, + { + "epoch": 0.4403925335490358, + "grad_norm": 2.4490976333618164, + "learning_rate": 1.8239138982822348e-05, + "loss": 3.7432, + "step": 39200 + }, + { + "epoch": 0.44095425871937893, + "grad_norm": 1.1330267190933228, + "learning_rate": 1.8236892069519497e-05, + "loss": 3.8154, + "step": 39250 + }, + { + "epoch": 0.44151598388972213, + "grad_norm": 1.1379812955856323, + "learning_rate": 1.823464515621665e-05, + "loss": 3.78, + "step": 39300 + }, + { + "epoch": 0.44207770906006527, + "grad_norm": 1.0772302150726318, + "learning_rate": 1.82323982429138e-05, + "loss": 3.7834, + "step": 39350 + }, + { + "epoch": 0.4426394342304084, + "grad_norm": 1.1633509397506714, + "learning_rate": 1.8230151329610948e-05, + "loss": 3.8166, + "step": 39400 + }, + { + "epoch": 0.4432011594007516, + "grad_norm": 1.1506922245025635, + "learning_rate": 1.8227904416308098e-05, + "loss": 3.7748, + "step": 39450 + }, + { + "epoch": 0.44376288457109475, + "grad_norm": 1.0494705438613892, + "learning_rate": 1.8225657503005247e-05, + "loss": 3.8452, + "step": 39500 + }, + { + "epoch": 0.4443246097414379, + "grad_norm": 1.2255929708480835, + "learning_rate": 1.8223410589702396e-05, + "loss": 3.7949, + "step": 39550 + }, + { + "epoch": 0.4448863349117811, + "grad_norm": 1.0744147300720215, + "learning_rate": 1.8221163676399546e-05, + "loss": 3.7633, + "step": 39600 + }, + { + "epoch": 0.44544806008212423, + "grad_norm": 1.0726063251495361, + "learning_rate": 1.8218916763096698e-05, + "loss": 3.7455, + "step": 39650 + }, + { + "epoch": 0.4460097852524674, + "grad_norm": 1.1668347120285034, + "learning_rate": 1.8216669849793848e-05, + "loss": 3.7894, + "step": 39700 + }, + { + "epoch": 0.4465715104228105, + "grad_norm": 1.0200697183609009, + "learning_rate": 1.8214422936490997e-05, + "loss": 3.7932, + "step": 39750 + }, + { + "epoch": 0.4471332355931537, + "grad_norm": 1.0065279006958008, + "learning_rate": 1.8212176023188146e-05, + "loss": 3.8165, + "step": 39800 + }, + { + "epoch": 0.44769496076349685, + "grad_norm": 0.9998236298561096, + "learning_rate": 1.8209929109885295e-05, + "loss": 3.7712, + "step": 39850 + }, + { + "epoch": 0.44825668593384, + "grad_norm": 1.0551574230194092, + "learning_rate": 1.8207682196582445e-05, + "loss": 3.8372, + "step": 39900 + }, + { + "epoch": 0.4488184111041832, + "grad_norm": 1.0093590021133423, + "learning_rate": 1.8205435283279594e-05, + "loss": 3.8506, + "step": 39950 + }, + { + "epoch": 0.44938013627452633, + "grad_norm": 1.2401299476623535, + "learning_rate": 1.8203188369976747e-05, + "loss": 3.7613, + "step": 40000 + }, + { + "epoch": 0.4499418614448695, + "grad_norm": 1.0292561054229736, + "learning_rate": 1.8200941456673896e-05, + "loss": 3.734, + "step": 40050 + }, + { + "epoch": 0.4505035866152126, + "grad_norm": 1.2993911504745483, + "learning_rate": 1.8198694543371045e-05, + "loss": 3.7656, + "step": 40100 + }, + { + "epoch": 0.4510653117855558, + "grad_norm": 1.1888697147369385, + "learning_rate": 1.8196447630068195e-05, + "loss": 3.7041, + "step": 40150 + }, + { + "epoch": 0.45162703695589895, + "grad_norm": 1.0635584592819214, + "learning_rate": 1.8194200716765344e-05, + "loss": 3.7893, + "step": 40200 + }, + { + "epoch": 0.4521887621262421, + "grad_norm": 1.0225001573562622, + "learning_rate": 1.8191953803462493e-05, + "loss": 3.834, + "step": 40250 + }, + { + "epoch": 0.4527504872965853, + "grad_norm": 1.1304889917373657, + "learning_rate": 1.8189706890159643e-05, + "loss": 3.7892, + "step": 40300 + }, + { + "epoch": 0.45331221246692843, + "grad_norm": 1.0044043064117432, + "learning_rate": 1.8187459976856795e-05, + "loss": 3.8025, + "step": 40350 + }, + { + "epoch": 0.4538739376372716, + "grad_norm": 1.046651005744934, + "learning_rate": 1.8185213063553945e-05, + "loss": 3.7732, + "step": 40400 + }, + { + "epoch": 0.45443566280761477, + "grad_norm": 1.2860900163650513, + "learning_rate": 1.8182966150251094e-05, + "loss": 3.779, + "step": 40450 + }, + { + "epoch": 0.4549973879779579, + "grad_norm": 1.0236151218414307, + "learning_rate": 1.8180719236948243e-05, + "loss": 3.8071, + "step": 40500 + }, + { + "epoch": 0.45555911314830105, + "grad_norm": 1.234114170074463, + "learning_rate": 1.8178472323645393e-05, + "loss": 3.7356, + "step": 40550 + }, + { + "epoch": 0.4561208383186442, + "grad_norm": 1.2366036176681519, + "learning_rate": 1.8176225410342542e-05, + "loss": 3.7333, + "step": 40600 + }, + { + "epoch": 0.4566825634889874, + "grad_norm": 1.0069408416748047, + "learning_rate": 1.817397849703969e-05, + "loss": 3.7977, + "step": 40650 + }, + { + "epoch": 0.45724428865933053, + "grad_norm": 1.099448561668396, + "learning_rate": 1.8171731583736844e-05, + "loss": 3.8713, + "step": 40700 + }, + { + "epoch": 0.4578060138296737, + "grad_norm": 0.9119434952735901, + "learning_rate": 1.8169484670433993e-05, + "loss": 3.7438, + "step": 40750 + }, + { + "epoch": 0.45836773900001687, + "grad_norm": 1.0150121450424194, + "learning_rate": 1.8167237757131143e-05, + "loss": 3.7982, + "step": 40800 + }, + { + "epoch": 0.45892946417036, + "grad_norm": 1.1290061473846436, + "learning_rate": 1.8164990843828292e-05, + "loss": 3.8241, + "step": 40850 + }, + { + "epoch": 0.45949118934070315, + "grad_norm": 1.3187247514724731, + "learning_rate": 1.816274393052544e-05, + "loss": 3.881, + "step": 40900 + }, + { + "epoch": 0.46005291451104635, + "grad_norm": 0.955877423286438, + "learning_rate": 1.816049701722259e-05, + "loss": 3.7177, + "step": 40950 + }, + { + "epoch": 0.4606146396813895, + "grad_norm": 1.0432859659194946, + "learning_rate": 1.815825010391974e-05, + "loss": 3.8475, + "step": 41000 + }, + { + "epoch": 0.46117636485173263, + "grad_norm": 1.1149173974990845, + "learning_rate": 1.8156003190616893e-05, + "loss": 3.8116, + "step": 41050 + }, + { + "epoch": 0.4617380900220758, + "grad_norm": 1.0391672849655151, + "learning_rate": 1.8153756277314042e-05, + "loss": 3.8382, + "step": 41100 + }, + { + "epoch": 0.462299815192419, + "grad_norm": 1.134096384048462, + "learning_rate": 1.815150936401119e-05, + "loss": 3.827, + "step": 41150 + }, + { + "epoch": 0.4628615403627621, + "grad_norm": 1.1240386962890625, + "learning_rate": 1.814926245070834e-05, + "loss": 3.8302, + "step": 41200 + }, + { + "epoch": 0.46342326553310526, + "grad_norm": 1.013810634613037, + "learning_rate": 1.814701553740549e-05, + "loss": 3.726, + "step": 41250 + }, + { + "epoch": 0.46398499070344845, + "grad_norm": 1.0930726528167725, + "learning_rate": 1.814476862410264e-05, + "loss": 3.7472, + "step": 41300 + }, + { + "epoch": 0.4645467158737916, + "grad_norm": 1.162300944328308, + "learning_rate": 1.814252171079979e-05, + "loss": 3.8044, + "step": 41350 + }, + { + "epoch": 0.46510844104413473, + "grad_norm": 1.0755993127822876, + "learning_rate": 1.8140274797496938e-05, + "loss": 3.8546, + "step": 41400 + }, + { + "epoch": 0.46567016621447793, + "grad_norm": 1.058703064918518, + "learning_rate": 1.813802788419409e-05, + "loss": 3.768, + "step": 41450 + }, + { + "epoch": 0.4662318913848211, + "grad_norm": 1.0064525604248047, + "learning_rate": 1.813578097089124e-05, + "loss": 3.7942, + "step": 41500 + }, + { + "epoch": 0.4667936165551642, + "grad_norm": 1.0544313192367554, + "learning_rate": 1.813353405758839e-05, + "loss": 3.7394, + "step": 41550 + }, + { + "epoch": 0.46735534172550736, + "grad_norm": 1.0749855041503906, + "learning_rate": 1.813128714428554e-05, + "loss": 3.7752, + "step": 41600 + }, + { + "epoch": 0.46791706689585055, + "grad_norm": 1.1062192916870117, + "learning_rate": 1.8129040230982688e-05, + "loss": 3.758, + "step": 41650 + }, + { + "epoch": 0.4684787920661937, + "grad_norm": 1.165113091468811, + "learning_rate": 1.8126793317679837e-05, + "loss": 3.8017, + "step": 41700 + }, + { + "epoch": 0.46904051723653684, + "grad_norm": 1.0622942447662354, + "learning_rate": 1.8124546404376986e-05, + "loss": 3.8276, + "step": 41750 + }, + { + "epoch": 0.46960224240688003, + "grad_norm": 1.0217134952545166, + "learning_rate": 1.812229949107414e-05, + "loss": 3.8924, + "step": 41800 + }, + { + "epoch": 0.4701639675772232, + "grad_norm": 1.0252145528793335, + "learning_rate": 1.812005257777129e-05, + "loss": 3.793, + "step": 41850 + }, + { + "epoch": 0.4707256927475663, + "grad_norm": 1.1097480058670044, + "learning_rate": 1.8117805664468438e-05, + "loss": 3.8127, + "step": 41900 + }, + { + "epoch": 0.47128741791790946, + "grad_norm": 1.2562998533248901, + "learning_rate": 1.8115558751165587e-05, + "loss": 3.7614, + "step": 41950 + }, + { + "epoch": 0.47184914308825265, + "grad_norm": 0.9875652194023132, + "learning_rate": 1.8113311837862736e-05, + "loss": 3.8162, + "step": 42000 + }, + { + "epoch": 0.4724108682585958, + "grad_norm": 1.0522080659866333, + "learning_rate": 1.8111064924559886e-05, + "loss": 3.7795, + "step": 42050 + }, + { + "epoch": 0.47297259342893894, + "grad_norm": 1.094172477722168, + "learning_rate": 1.8108818011257035e-05, + "loss": 3.7722, + "step": 42100 + }, + { + "epoch": 0.47353431859928213, + "grad_norm": 1.3399633169174194, + "learning_rate": 1.8106571097954188e-05, + "loss": 3.6965, + "step": 42150 + }, + { + "epoch": 0.4740960437696253, + "grad_norm": 1.1942780017852783, + "learning_rate": 1.8104324184651337e-05, + "loss": 3.7639, + "step": 42200 + }, + { + "epoch": 0.4746577689399684, + "grad_norm": 1.0598344802856445, + "learning_rate": 1.8102077271348486e-05, + "loss": 3.7796, + "step": 42250 + }, + { + "epoch": 0.4752194941103116, + "grad_norm": 1.0890198945999146, + "learning_rate": 1.8099830358045636e-05, + "loss": 3.7223, + "step": 42300 + }, + { + "epoch": 0.47578121928065475, + "grad_norm": 1.029687523841858, + "learning_rate": 1.809762838300884e-05, + "loss": 3.7558, + "step": 42350 + }, + { + "epoch": 0.4763429444509979, + "grad_norm": 0.9146989583969116, + "learning_rate": 1.809538146970599e-05, + "loss": 3.702, + "step": 42400 + }, + { + "epoch": 0.47690466962134104, + "grad_norm": 1.1156704425811768, + "learning_rate": 1.8093134556403144e-05, + "loss": 3.8005, + "step": 42450 + }, + { + "epoch": 0.47746639479168423, + "grad_norm": 1.1274996995925903, + "learning_rate": 1.8090887643100293e-05, + "loss": 3.8125, + "step": 42500 + }, + { + "epoch": 0.4780281199620274, + "grad_norm": 1.1434170007705688, + "learning_rate": 1.8088640729797442e-05, + "loss": 3.7812, + "step": 42550 + }, + { + "epoch": 0.4785898451323705, + "grad_norm": 1.032583236694336, + "learning_rate": 1.808639381649459e-05, + "loss": 3.7945, + "step": 42600 + }, + { + "epoch": 0.4791515703027137, + "grad_norm": 1.0314099788665771, + "learning_rate": 1.808414690319174e-05, + "loss": 3.7249, + "step": 42650 + }, + { + "epoch": 0.47971329547305686, + "grad_norm": 1.1356045007705688, + "learning_rate": 1.808189998988889e-05, + "loss": 3.8384, + "step": 42700 + }, + { + "epoch": 0.4802750206434, + "grad_norm": 1.1715441942214966, + "learning_rate": 1.807965307658604e-05, + "loss": 3.858, + "step": 42750 + }, + { + "epoch": 0.4808367458137432, + "grad_norm": 1.0327473878860474, + "learning_rate": 1.8077406163283192e-05, + "loss": 3.8273, + "step": 42800 + }, + { + "epoch": 0.48139847098408634, + "grad_norm": 1.1157915592193604, + "learning_rate": 1.807515924998034e-05, + "loss": 3.7637, + "step": 42850 + }, + { + "epoch": 0.4819601961544295, + "grad_norm": 1.0877025127410889, + "learning_rate": 1.807291233667749e-05, + "loss": 3.7895, + "step": 42900 + }, + { + "epoch": 0.4825219213247726, + "grad_norm": 1.0560516119003296, + "learning_rate": 1.807066542337464e-05, + "loss": 3.7441, + "step": 42950 + }, + { + "epoch": 0.4830836464951158, + "grad_norm": 1.0265017747879028, + "learning_rate": 1.806841851007179e-05, + "loss": 3.8135, + "step": 43000 + }, + { + "epoch": 0.48364537166545896, + "grad_norm": 1.2721837759017944, + "learning_rate": 1.806617159676894e-05, + "loss": 3.7467, + "step": 43050 + }, + { + "epoch": 0.4842070968358021, + "grad_norm": 1.0625399351119995, + "learning_rate": 1.8063924683466088e-05, + "loss": 3.694, + "step": 43100 + }, + { + "epoch": 0.4847688220061453, + "grad_norm": 1.0817184448242188, + "learning_rate": 1.806167777016324e-05, + "loss": 3.7768, + "step": 43150 + }, + { + "epoch": 0.48533054717648844, + "grad_norm": 1.1214500665664673, + "learning_rate": 1.805943085686039e-05, + "loss": 3.833, + "step": 43200 + }, + { + "epoch": 0.4858922723468316, + "grad_norm": 1.082459568977356, + "learning_rate": 1.805718394355754e-05, + "loss": 3.8078, + "step": 43250 + }, + { + "epoch": 0.4864539975171748, + "grad_norm": 1.3082919120788574, + "learning_rate": 1.805493703025469e-05, + "loss": 3.7968, + "step": 43300 + }, + { + "epoch": 0.4870157226875179, + "grad_norm": 1.1481602191925049, + "learning_rate": 1.8052690116951838e-05, + "loss": 3.7631, + "step": 43350 + }, + { + "epoch": 0.48757744785786106, + "grad_norm": 1.0289119482040405, + "learning_rate": 1.8050443203648987e-05, + "loss": 3.8111, + "step": 43400 + }, + { + "epoch": 0.4881391730282042, + "grad_norm": 1.1134016513824463, + "learning_rate": 1.8048196290346137e-05, + "loss": 3.7552, + "step": 43450 + }, + { + "epoch": 0.4887008981985474, + "grad_norm": 1.1744118928909302, + "learning_rate": 1.804594937704329e-05, + "loss": 3.9021, + "step": 43500 + }, + { + "epoch": 0.48926262336889054, + "grad_norm": 1.0428119897842407, + "learning_rate": 1.804370246374044e-05, + "loss": 3.7142, + "step": 43550 + }, + { + "epoch": 0.4898243485392337, + "grad_norm": 1.2769163846969604, + "learning_rate": 1.8041455550437588e-05, + "loss": 3.7732, + "step": 43600 + }, + { + "epoch": 0.4903860737095769, + "grad_norm": 1.1050018072128296, + "learning_rate": 1.8039208637134737e-05, + "loss": 3.8573, + "step": 43650 + }, + { + "epoch": 0.49094779887992, + "grad_norm": 1.1701105833053589, + "learning_rate": 1.8036961723831887e-05, + "loss": 3.7084, + "step": 43700 + }, + { + "epoch": 0.49150952405026316, + "grad_norm": 1.3127405643463135, + "learning_rate": 1.8034714810529036e-05, + "loss": 3.834, + "step": 43750 + }, + { + "epoch": 0.4920712492206063, + "grad_norm": 1.1261199712753296, + "learning_rate": 1.8032467897226185e-05, + "loss": 3.7859, + "step": 43800 + }, + { + "epoch": 0.4926329743909495, + "grad_norm": 0.8711647987365723, + "learning_rate": 1.8030220983923338e-05, + "loss": 3.7959, + "step": 43850 + }, + { + "epoch": 0.49319469956129264, + "grad_norm": 1.015077829360962, + "learning_rate": 1.8027974070620487e-05, + "loss": 3.78, + "step": 43900 + }, + { + "epoch": 0.4937564247316358, + "grad_norm": 1.1011428833007812, + "learning_rate": 1.8025727157317637e-05, + "loss": 3.8081, + "step": 43950 + }, + { + "epoch": 0.494318149901979, + "grad_norm": 1.3075041770935059, + "learning_rate": 1.8023480244014786e-05, + "loss": 3.7885, + "step": 44000 + }, + { + "epoch": 0.4948798750723221, + "grad_norm": 1.1464548110961914, + "learning_rate": 1.8021233330711935e-05, + "loss": 3.8184, + "step": 44050 + }, + { + "epoch": 0.49544160024266526, + "grad_norm": 1.1513046026229858, + "learning_rate": 1.8018986417409085e-05, + "loss": 3.7232, + "step": 44100 + }, + { + "epoch": 0.49600332541300846, + "grad_norm": 1.0444048643112183, + "learning_rate": 1.8016739504106234e-05, + "loss": 3.7568, + "step": 44150 + }, + { + "epoch": 0.4965650505833516, + "grad_norm": 1.0750826597213745, + "learning_rate": 1.8014492590803387e-05, + "loss": 3.7301, + "step": 44200 + }, + { + "epoch": 0.49712677575369474, + "grad_norm": 1.0664238929748535, + "learning_rate": 1.8012245677500536e-05, + "loss": 3.8694, + "step": 44250 + }, + { + "epoch": 0.4976885009240379, + "grad_norm": 0.9873380661010742, + "learning_rate": 1.8009998764197685e-05, + "loss": 3.8639, + "step": 44300 + }, + { + "epoch": 0.4982502260943811, + "grad_norm": 1.2076796293258667, + "learning_rate": 1.8007751850894835e-05, + "loss": 3.7719, + "step": 44350 + }, + { + "epoch": 0.4988119512647242, + "grad_norm": 1.0296562910079956, + "learning_rate": 1.8005504937591984e-05, + "loss": 3.7266, + "step": 44400 + }, + { + "epoch": 0.49937367643506736, + "grad_norm": 1.1904340982437134, + "learning_rate": 1.8003258024289133e-05, + "loss": 3.7871, + "step": 44450 + }, + { + "epoch": 0.49993540160541056, + "grad_norm": 1.0401132106781006, + "learning_rate": 1.8001011110986283e-05, + "loss": 3.7817, + "step": 44500 + }, + { + "epoch": 0.5004971267757536, + "grad_norm": 1.9463406801223755, + "learning_rate": 1.7998764197683435e-05, + "loss": 3.8001, + "step": 44550 + }, + { + "epoch": 0.5010588519460969, + "grad_norm": 1.1187658309936523, + "learning_rate": 1.7996517284380585e-05, + "loss": 3.7387, + "step": 44600 + }, + { + "epoch": 0.50162057711644, + "grad_norm": 1.1454840898513794, + "learning_rate": 1.7994270371077734e-05, + "loss": 3.8329, + "step": 44650 + }, + { + "epoch": 0.5021823022867832, + "grad_norm": 1.2378987073898315, + "learning_rate": 1.7992023457774883e-05, + "loss": 3.8162, + "step": 44700 + }, + { + "epoch": 0.5027440274571263, + "grad_norm": 2.6603729724884033, + "learning_rate": 1.7989776544472033e-05, + "loss": 3.7931, + "step": 44750 + }, + { + "epoch": 0.5033057526274695, + "grad_norm": 1.2083017826080322, + "learning_rate": 1.7987529631169182e-05, + "loss": 3.863, + "step": 44800 + }, + { + "epoch": 0.5038674777978126, + "grad_norm": 1.057800054550171, + "learning_rate": 1.798528271786633e-05, + "loss": 3.7743, + "step": 44850 + }, + { + "epoch": 0.5044292029681559, + "grad_norm": 1.142313838005066, + "learning_rate": 1.7983035804563484e-05, + "loss": 3.7603, + "step": 44900 + }, + { + "epoch": 0.504990928138499, + "grad_norm": 1.0015513896942139, + "learning_rate": 1.7980788891260633e-05, + "loss": 3.7382, + "step": 44950 + }, + { + "epoch": 0.5055526533088421, + "grad_norm": 1.2800809144973755, + "learning_rate": 1.7978541977957783e-05, + "loss": 3.7812, + "step": 45000 + }, + { + "epoch": 0.5061143784791853, + "grad_norm": 1.1067944765090942, + "learning_rate": 1.7976295064654932e-05, + "loss": 3.7519, + "step": 45050 + }, + { + "epoch": 0.5066761036495284, + "grad_norm": 0.9286227822303772, + "learning_rate": 1.797404815135208e-05, + "loss": 3.703, + "step": 45100 + }, + { + "epoch": 0.5072378288198716, + "grad_norm": 1.068844199180603, + "learning_rate": 1.797180123804923e-05, + "loss": 3.8061, + "step": 45150 + }, + { + "epoch": 0.5077995539902147, + "grad_norm": 1.1424062252044678, + "learning_rate": 1.796955432474638e-05, + "loss": 3.7488, + "step": 45200 + }, + { + "epoch": 0.508361279160558, + "grad_norm": 1.1277310848236084, + "learning_rate": 1.7967307411443532e-05, + "loss": 3.6984, + "step": 45250 + }, + { + "epoch": 0.5089230043309011, + "grad_norm": 1.0205165147781372, + "learning_rate": 1.7965060498140682e-05, + "loss": 3.7432, + "step": 45300 + }, + { + "epoch": 0.5094847295012442, + "grad_norm": 1.3567684888839722, + "learning_rate": 1.796281358483783e-05, + "loss": 3.8955, + "step": 45350 + }, + { + "epoch": 0.5100464546715874, + "grad_norm": 0.9939897656440735, + "learning_rate": 1.796056667153498e-05, + "loss": 3.8553, + "step": 45400 + }, + { + "epoch": 0.5106081798419305, + "grad_norm": 1.4429136514663696, + "learning_rate": 1.795831975823213e-05, + "loss": 3.7886, + "step": 45450 + }, + { + "epoch": 0.5111699050122737, + "grad_norm": 1.3088704347610474, + "learning_rate": 1.7956117783195336e-05, + "loss": 3.7935, + "step": 45500 + }, + { + "epoch": 0.5117316301826168, + "grad_norm": 1.0963215827941895, + "learning_rate": 1.795387086989249e-05, + "loss": 3.7274, + "step": 45550 + }, + { + "epoch": 0.5122933553529601, + "grad_norm": 1.1551778316497803, + "learning_rate": 1.7951623956589638e-05, + "loss": 3.7619, + "step": 45600 + }, + { + "epoch": 0.5128550805233032, + "grad_norm": 1.2554855346679688, + "learning_rate": 1.7949377043286787e-05, + "loss": 3.8272, + "step": 45650 + }, + { + "epoch": 0.5134168056936463, + "grad_norm": 1.3378087282180786, + "learning_rate": 1.7947130129983936e-05, + "loss": 3.7626, + "step": 45700 + }, + { + "epoch": 0.5139785308639895, + "grad_norm": 1.0493336915969849, + "learning_rate": 1.7944883216681086e-05, + "loss": 3.8167, + "step": 45750 + }, + { + "epoch": 0.5145402560343326, + "grad_norm": 1.1279903650283813, + "learning_rate": 1.7942636303378235e-05, + "loss": 3.7878, + "step": 45800 + }, + { + "epoch": 0.5151019812046758, + "grad_norm": 1.1540268659591675, + "learning_rate": 1.7940389390075384e-05, + "loss": 3.8052, + "step": 45850 + }, + { + "epoch": 0.5156637063750189, + "grad_norm": 1.316269040107727, + "learning_rate": 1.7938142476772537e-05, + "loss": 3.832, + "step": 45900 + }, + { + "epoch": 0.5162254315453622, + "grad_norm": 0.9910030364990234, + "learning_rate": 1.7935895563469686e-05, + "loss": 3.7452, + "step": 45950 + }, + { + "epoch": 0.5167871567157053, + "grad_norm": 1.1497697830200195, + "learning_rate": 1.7933648650166836e-05, + "loss": 3.7542, + "step": 46000 + }, + { + "epoch": 0.5173488818860484, + "grad_norm": 1.08818781375885, + "learning_rate": 1.7931401736863985e-05, + "loss": 3.8004, + "step": 46050 + }, + { + "epoch": 0.5179106070563916, + "grad_norm": 1.0237677097320557, + "learning_rate": 1.7929154823561134e-05, + "loss": 3.8064, + "step": 46100 + }, + { + "epoch": 0.5184723322267347, + "grad_norm": 1.1072641611099243, + "learning_rate": 1.7926907910258284e-05, + "loss": 3.7539, + "step": 46150 + }, + { + "epoch": 0.5190340573970779, + "grad_norm": 1.2030876874923706, + "learning_rate": 1.7924660996955433e-05, + "loss": 3.8923, + "step": 46200 + }, + { + "epoch": 0.5195957825674211, + "grad_norm": 1.003875732421875, + "learning_rate": 1.7922414083652586e-05, + "loss": 3.7836, + "step": 46250 + }, + { + "epoch": 0.5201575077377643, + "grad_norm": 1.058693766593933, + "learning_rate": 1.7920167170349735e-05, + "loss": 3.757, + "step": 46300 + }, + { + "epoch": 0.5207192329081074, + "grad_norm": 1.0645219087600708, + "learning_rate": 1.7917920257046884e-05, + "loss": 3.8424, + "step": 46350 + }, + { + "epoch": 0.5212809580784505, + "grad_norm": 1.0608023405075073, + "learning_rate": 1.7915673343744034e-05, + "loss": 3.7468, + "step": 46400 + }, + { + "epoch": 0.5218426832487937, + "grad_norm": 1.049337387084961, + "learning_rate": 1.7913426430441183e-05, + "loss": 3.8215, + "step": 46450 + }, + { + "epoch": 0.5224044084191368, + "grad_norm": 1.2163463830947876, + "learning_rate": 1.7911179517138332e-05, + "loss": 3.7684, + "step": 46500 + }, + { + "epoch": 0.52296613358948, + "grad_norm": 1.433159589767456, + "learning_rate": 1.790893260383548e-05, + "loss": 3.7887, + "step": 46550 + }, + { + "epoch": 0.5235278587598232, + "grad_norm": 2.4778525829315186, + "learning_rate": 1.790668569053263e-05, + "loss": 3.8918, + "step": 46600 + }, + { + "epoch": 0.5240895839301664, + "grad_norm": 1.0728590488433838, + "learning_rate": 1.7904438777229784e-05, + "loss": 3.7463, + "step": 46650 + }, + { + "epoch": 0.5246513091005095, + "grad_norm": 1.0942366123199463, + "learning_rate": 1.7902191863926933e-05, + "loss": 3.8691, + "step": 46700 + }, + { + "epoch": 0.5252130342708526, + "grad_norm": 1.2694611549377441, + "learning_rate": 1.7899944950624082e-05, + "loss": 3.7357, + "step": 46750 + }, + { + "epoch": 0.5257747594411958, + "grad_norm": 0.9541887044906616, + "learning_rate": 1.789769803732123e-05, + "loss": 3.8223, + "step": 46800 + }, + { + "epoch": 0.5263364846115389, + "grad_norm": 1.0459247827529907, + "learning_rate": 1.789545112401838e-05, + "loss": 3.811, + "step": 46850 + }, + { + "epoch": 0.5268982097818821, + "grad_norm": 1.215408444404602, + "learning_rate": 1.789320421071553e-05, + "loss": 3.8257, + "step": 46900 + }, + { + "epoch": 0.5274599349522253, + "grad_norm": 1.2091480493545532, + "learning_rate": 1.789095729741268e-05, + "loss": 3.7794, + "step": 46950 + }, + { + "epoch": 0.5280216601225685, + "grad_norm": 1.0477527379989624, + "learning_rate": 1.7888710384109832e-05, + "loss": 3.7567, + "step": 47000 + }, + { + "epoch": 0.5285833852929116, + "grad_norm": 1.0070216655731201, + "learning_rate": 1.788646347080698e-05, + "loss": 3.8107, + "step": 47050 + }, + { + "epoch": 0.5291451104632547, + "grad_norm": 0.9893909096717834, + "learning_rate": 1.788421655750413e-05, + "loss": 3.7764, + "step": 47100 + }, + { + "epoch": 0.5297068356335979, + "grad_norm": 0.9789925217628479, + "learning_rate": 1.788196964420128e-05, + "loss": 3.822, + "step": 47150 + }, + { + "epoch": 0.530268560803941, + "grad_norm": 0.9236642122268677, + "learning_rate": 1.787972273089843e-05, + "loss": 3.8091, + "step": 47200 + }, + { + "epoch": 0.5308302859742843, + "grad_norm": 0.9914894700050354, + "learning_rate": 1.787747581759558e-05, + "loss": 3.7204, + "step": 47250 + }, + { + "epoch": 0.5313920111446274, + "grad_norm": 1.0681378841400146, + "learning_rate": 1.7875228904292728e-05, + "loss": 3.778, + "step": 47300 + }, + { + "epoch": 0.5319537363149706, + "grad_norm": 0.9172595739364624, + "learning_rate": 1.787298199098988e-05, + "loss": 3.7245, + "step": 47350 + }, + { + "epoch": 0.5325154614853137, + "grad_norm": 1.0290454626083374, + "learning_rate": 1.787073507768703e-05, + "loss": 3.6851, + "step": 47400 + }, + { + "epoch": 0.5330771866556568, + "grad_norm": 1.009787678718567, + "learning_rate": 1.786848816438418e-05, + "loss": 3.7659, + "step": 47450 + }, + { + "epoch": 0.533638911826, + "grad_norm": 1.0187406539916992, + "learning_rate": 1.786624125108133e-05, + "loss": 3.8141, + "step": 47500 + }, + { + "epoch": 0.5342006369963431, + "grad_norm": 1.3476841449737549, + "learning_rate": 1.7863994337778478e-05, + "loss": 3.7867, + "step": 47550 + }, + { + "epoch": 0.5347623621666864, + "grad_norm": 0.9151012897491455, + "learning_rate": 1.7861747424475627e-05, + "loss": 3.7697, + "step": 47600 + }, + { + "epoch": 0.5353240873370295, + "grad_norm": 1.2090747356414795, + "learning_rate": 1.7859500511172777e-05, + "loss": 3.7745, + "step": 47650 + }, + { + "epoch": 0.5358858125073727, + "grad_norm": 1.0624113082885742, + "learning_rate": 1.785725359786993e-05, + "loss": 3.8045, + "step": 47700 + }, + { + "epoch": 0.5364475376777158, + "grad_norm": 0.9345954656600952, + "learning_rate": 1.785500668456708e-05, + "loss": 3.7731, + "step": 47750 + }, + { + "epoch": 0.5370092628480589, + "grad_norm": 1.0432896614074707, + "learning_rate": 1.7852759771264228e-05, + "loss": 3.7274, + "step": 47800 + }, + { + "epoch": 0.5375709880184021, + "grad_norm": 1.1865384578704834, + "learning_rate": 1.7850512857961377e-05, + "loss": 3.808, + "step": 47850 + }, + { + "epoch": 0.5381327131887452, + "grad_norm": 1.299694538116455, + "learning_rate": 1.7848265944658527e-05, + "loss": 3.8216, + "step": 47900 + }, + { + "epoch": 0.5386944383590885, + "grad_norm": 1.0396870374679565, + "learning_rate": 1.7846019031355676e-05, + "loss": 3.7653, + "step": 47950 + }, + { + "epoch": 0.5392561635294316, + "grad_norm": 1.0869770050048828, + "learning_rate": 1.7843772118052825e-05, + "loss": 3.7826, + "step": 48000 + }, + { + "epoch": 0.5398178886997748, + "grad_norm": 1.1290099620819092, + "learning_rate": 1.7841525204749978e-05, + "loss": 3.7563, + "step": 48050 + }, + { + "epoch": 0.5403796138701179, + "grad_norm": 1.1860477924346924, + "learning_rate": 1.7839278291447127e-05, + "loss": 3.7272, + "step": 48100 + }, + { + "epoch": 0.540941339040461, + "grad_norm": 0.928475022315979, + "learning_rate": 1.7837031378144277e-05, + "loss": 3.75, + "step": 48150 + }, + { + "epoch": 0.5415030642108042, + "grad_norm": 1.1937905550003052, + "learning_rate": 1.7834784464841426e-05, + "loss": 3.729, + "step": 48200 + }, + { + "epoch": 0.5420647893811473, + "grad_norm": 0.9828348159790039, + "learning_rate": 1.7832537551538575e-05, + "loss": 3.7127, + "step": 48250 + }, + { + "epoch": 0.5426265145514906, + "grad_norm": 1.0111702680587769, + "learning_rate": 1.7830290638235725e-05, + "loss": 3.7807, + "step": 48300 + }, + { + "epoch": 0.5431882397218337, + "grad_norm": 1.058048963546753, + "learning_rate": 1.7828043724932874e-05, + "loss": 3.8394, + "step": 48350 + }, + { + "epoch": 0.5437499648921769, + "grad_norm": 1.034980058670044, + "learning_rate": 1.7825796811630027e-05, + "loss": 3.8054, + "step": 48400 + }, + { + "epoch": 0.54431169006252, + "grad_norm": 1.2437156438827515, + "learning_rate": 1.7823549898327176e-05, + "loss": 3.7458, + "step": 48450 + }, + { + "epoch": 0.5448734152328631, + "grad_norm": 1.2050033807754517, + "learning_rate": 1.7821302985024325e-05, + "loss": 3.8883, + "step": 48500 + }, + { + "epoch": 0.5454351404032063, + "grad_norm": 1.2797744274139404, + "learning_rate": 1.7819056071721474e-05, + "loss": 3.8127, + "step": 48550 + }, + { + "epoch": 0.5459968655735495, + "grad_norm": 1.0819027423858643, + "learning_rate": 1.7816809158418624e-05, + "loss": 3.7676, + "step": 48600 + }, + { + "epoch": 0.5465585907438927, + "grad_norm": 1.1629184484481812, + "learning_rate": 1.7814562245115773e-05, + "loss": 3.7014, + "step": 48650 + }, + { + "epoch": 0.5471203159142358, + "grad_norm": 1.2764919996261597, + "learning_rate": 1.7812315331812922e-05, + "loss": 3.7961, + "step": 48700 + }, + { + "epoch": 0.547682041084579, + "grad_norm": 1.063462734222412, + "learning_rate": 1.7810068418510075e-05, + "loss": 3.79, + "step": 48750 + }, + { + "epoch": 0.5482437662549221, + "grad_norm": 1.0746026039123535, + "learning_rate": 1.7807821505207224e-05, + "loss": 3.7543, + "step": 48800 + }, + { + "epoch": 0.5488054914252652, + "grad_norm": 1.0190949440002441, + "learning_rate": 1.7805574591904374e-05, + "loss": 3.7683, + "step": 48850 + }, + { + "epoch": 0.5493672165956084, + "grad_norm": 1.0866377353668213, + "learning_rate": 1.7803327678601523e-05, + "loss": 3.8017, + "step": 48900 + }, + { + "epoch": 0.5499289417659516, + "grad_norm": 1.1085233688354492, + "learning_rate": 1.7801080765298672e-05, + "loss": 3.7577, + "step": 48950 + }, + { + "epoch": 0.5504906669362948, + "grad_norm": 1.2025694847106934, + "learning_rate": 1.7798833851995822e-05, + "loss": 3.7937, + "step": 49000 + }, + { + "epoch": 0.5510523921066379, + "grad_norm": 1.199409008026123, + "learning_rate": 1.779658693869297e-05, + "loss": 3.7602, + "step": 49050 + }, + { + "epoch": 0.5516141172769811, + "grad_norm": 1.1424399614334106, + "learning_rate": 1.7794340025390124e-05, + "loss": 3.7481, + "step": 49100 + }, + { + "epoch": 0.5521758424473242, + "grad_norm": 0.9761651754379272, + "learning_rate": 1.7792093112087273e-05, + "loss": 3.7907, + "step": 49150 + }, + { + "epoch": 0.5527375676176673, + "grad_norm": 1.0852487087249756, + "learning_rate": 1.7789846198784422e-05, + "loss": 3.7912, + "step": 49200 + }, + { + "epoch": 0.5532992927880105, + "grad_norm": 1.1241354942321777, + "learning_rate": 1.7787599285481572e-05, + "loss": 3.7338, + "step": 49250 + }, + { + "epoch": 0.5538610179583537, + "grad_norm": 1.0911507606506348, + "learning_rate": 1.778535237217872e-05, + "loss": 3.6668, + "step": 49300 + }, + { + "epoch": 0.5544227431286969, + "grad_norm": 1.608834981918335, + "learning_rate": 1.778310545887587e-05, + "loss": 3.7726, + "step": 49350 + }, + { + "epoch": 0.55498446829904, + "grad_norm": 1.217059850692749, + "learning_rate": 1.778085854557302e-05, + "loss": 3.7954, + "step": 49400 + }, + { + "epoch": 0.5555461934693832, + "grad_norm": 1.0991884469985962, + "learning_rate": 1.7778611632270172e-05, + "loss": 3.7368, + "step": 49450 + }, + { + "epoch": 0.5561079186397263, + "grad_norm": 0.9799400568008423, + "learning_rate": 1.777636471896732e-05, + "loss": 3.7249, + "step": 49500 + }, + { + "epoch": 0.5566696438100694, + "grad_norm": 1.1263720989227295, + "learning_rate": 1.777411780566447e-05, + "loss": 3.8127, + "step": 49550 + }, + { + "epoch": 0.5572313689804126, + "grad_norm": 1.0370090007781982, + "learning_rate": 1.777187089236162e-05, + "loss": 3.6818, + "step": 49600 + }, + { + "epoch": 0.5577930941507558, + "grad_norm": 2.1102616786956787, + "learning_rate": 1.7769668917324826e-05, + "loss": 3.7238, + "step": 49650 + }, + { + "epoch": 0.558354819321099, + "grad_norm": 1.13498055934906, + "learning_rate": 1.7767422004021976e-05, + "loss": 3.7812, + "step": 49700 + }, + { + "epoch": 0.5589165444914421, + "grad_norm": 0.9529072046279907, + "learning_rate": 1.7765175090719128e-05, + "loss": 3.7319, + "step": 49750 + }, + { + "epoch": 0.5594782696617853, + "grad_norm": 1.0441597700119019, + "learning_rate": 1.7762928177416278e-05, + "loss": 3.7676, + "step": 49800 + }, + { + "epoch": 0.5600399948321284, + "grad_norm": 1.1571311950683594, + "learning_rate": 1.7760681264113427e-05, + "loss": 3.7601, + "step": 49850 + }, + { + "epoch": 0.5606017200024715, + "grad_norm": 1.0752143859863281, + "learning_rate": 1.7758434350810576e-05, + "loss": 3.7677, + "step": 49900 + }, + { + "epoch": 0.5611634451728148, + "grad_norm": 0.9833121299743652, + "learning_rate": 1.7756187437507726e-05, + "loss": 3.8865, + "step": 49950 + }, + { + "epoch": 0.5617251703431579, + "grad_norm": 1.0245425701141357, + "learning_rate": 1.7753940524204875e-05, + "loss": 3.8149, + "step": 50000 + }, + { + "epoch": 0.5622868955135011, + "grad_norm": 1.0369062423706055, + "learning_rate": 1.7751693610902024e-05, + "loss": 3.7772, + "step": 50050 + }, + { + "epoch": 0.5628486206838442, + "grad_norm": 1.0579601526260376, + "learning_rate": 1.7749446697599177e-05, + "loss": 3.722, + "step": 50100 + }, + { + "epoch": 0.5634103458541874, + "grad_norm": 1.058449625968933, + "learning_rate": 1.7747199784296326e-05, + "loss": 3.8031, + "step": 50150 + }, + { + "epoch": 0.5639720710245305, + "grad_norm": 1.1619818210601807, + "learning_rate": 1.7744952870993475e-05, + "loss": 3.8059, + "step": 50200 + }, + { + "epoch": 0.5645337961948736, + "grad_norm": 1.1637179851531982, + "learning_rate": 1.7742705957690625e-05, + "loss": 3.7732, + "step": 50250 + }, + { + "epoch": 0.5650955213652169, + "grad_norm": 1.051743745803833, + "learning_rate": 1.7740459044387774e-05, + "loss": 3.7137, + "step": 50300 + }, + { + "epoch": 0.56565724653556, + "grad_norm": 1.0153290033340454, + "learning_rate": 1.7738212131084923e-05, + "loss": 3.7306, + "step": 50350 + }, + { + "epoch": 0.5662189717059032, + "grad_norm": 1.2448453903198242, + "learning_rate": 1.7735965217782073e-05, + "loss": 3.8135, + "step": 50400 + }, + { + "epoch": 0.5667806968762463, + "grad_norm": 1.077943205833435, + "learning_rate": 1.7733718304479225e-05, + "loss": 3.7605, + "step": 50450 + }, + { + "epoch": 0.5673424220465895, + "grad_norm": 1.080708384513855, + "learning_rate": 1.7731471391176375e-05, + "loss": 3.8126, + "step": 50500 + }, + { + "epoch": 0.5679041472169326, + "grad_norm": 1.1805051565170288, + "learning_rate": 1.7729224477873524e-05, + "loss": 3.7539, + "step": 50550 + }, + { + "epoch": 0.5684658723872758, + "grad_norm": 1.0887198448181152, + "learning_rate": 1.7726977564570673e-05, + "loss": 3.8005, + "step": 50600 + }, + { + "epoch": 0.569027597557619, + "grad_norm": 1.1345741748809814, + "learning_rate": 1.7724730651267823e-05, + "loss": 3.8272, + "step": 50650 + }, + { + "epoch": 0.5695893227279621, + "grad_norm": 0.9996209740638733, + "learning_rate": 1.7722483737964972e-05, + "loss": 3.7027, + "step": 50700 + }, + { + "epoch": 0.5701510478983053, + "grad_norm": 1.140913724899292, + "learning_rate": 1.772023682466212e-05, + "loss": 3.8024, + "step": 50750 + }, + { + "epoch": 0.5707127730686484, + "grad_norm": 1.1962040662765503, + "learning_rate": 1.7717989911359274e-05, + "loss": 3.8789, + "step": 50800 + }, + { + "epoch": 0.5712744982389916, + "grad_norm": 1.0278069972991943, + "learning_rate": 1.7715742998056423e-05, + "loss": 3.7437, + "step": 50850 + }, + { + "epoch": 0.5718362234093347, + "grad_norm": 1.2031197547912598, + "learning_rate": 1.7713496084753573e-05, + "loss": 3.7805, + "step": 50900 + }, + { + "epoch": 0.572397948579678, + "grad_norm": 1.1851876974105835, + "learning_rate": 1.7711249171450722e-05, + "loss": 3.8535, + "step": 50950 + }, + { + "epoch": 0.5729596737500211, + "grad_norm": 1.0673991441726685, + "learning_rate": 1.770900225814787e-05, + "loss": 3.7427, + "step": 51000 + }, + { + "epoch": 0.5735213989203642, + "grad_norm": 1.0498384237289429, + "learning_rate": 1.770675534484502e-05, + "loss": 3.7233, + "step": 51050 + }, + { + "epoch": 0.5740831240907074, + "grad_norm": 1.1012871265411377, + "learning_rate": 1.770450843154217e-05, + "loss": 3.7578, + "step": 51100 + }, + { + "epoch": 0.5746448492610505, + "grad_norm": 1.0736808776855469, + "learning_rate": 1.770226151823932e-05, + "loss": 3.8299, + "step": 51150 + }, + { + "epoch": 0.5752065744313937, + "grad_norm": 1.1782079935073853, + "learning_rate": 1.7700014604936472e-05, + "loss": 3.8076, + "step": 51200 + }, + { + "epoch": 0.5757682996017368, + "grad_norm": 1.1583595275878906, + "learning_rate": 1.769776769163362e-05, + "loss": 3.8288, + "step": 51250 + }, + { + "epoch": 0.5763300247720801, + "grad_norm": 0.9925355911254883, + "learning_rate": 1.769552077833077e-05, + "loss": 3.7869, + "step": 51300 + }, + { + "epoch": 0.5768917499424232, + "grad_norm": 1.0739439725875854, + "learning_rate": 1.769327386502792e-05, + "loss": 3.7328, + "step": 51350 + }, + { + "epoch": 0.5774534751127663, + "grad_norm": 1.141987681388855, + "learning_rate": 1.769102695172507e-05, + "loss": 3.7463, + "step": 51400 + }, + { + "epoch": 0.5780152002831095, + "grad_norm": 1.0860463380813599, + "learning_rate": 1.768878003842222e-05, + "loss": 3.7705, + "step": 51450 + }, + { + "epoch": 0.5785769254534526, + "grad_norm": 1.0678012371063232, + "learning_rate": 1.7686533125119368e-05, + "loss": 3.745, + "step": 51500 + }, + { + "epoch": 0.5791386506237958, + "grad_norm": 1.1007298231124878, + "learning_rate": 1.768428621181652e-05, + "loss": 3.7283, + "step": 51550 + }, + { + "epoch": 0.5797003757941389, + "grad_norm": 1.037177562713623, + "learning_rate": 1.768203929851367e-05, + "loss": 3.823, + "step": 51600 + }, + { + "epoch": 0.5802621009644822, + "grad_norm": 1.0051183700561523, + "learning_rate": 1.767979238521082e-05, + "loss": 3.7693, + "step": 51650 + }, + { + "epoch": 0.5808238261348253, + "grad_norm": 1.1470272541046143, + "learning_rate": 1.767754547190797e-05, + "loss": 3.7568, + "step": 51700 + }, + { + "epoch": 0.5813855513051684, + "grad_norm": 1.1250337362289429, + "learning_rate": 1.7675298558605118e-05, + "loss": 3.7366, + "step": 51750 + }, + { + "epoch": 0.5819472764755116, + "grad_norm": 0.9637058973312378, + "learning_rate": 1.7673051645302267e-05, + "loss": 3.7163, + "step": 51800 + }, + { + "epoch": 0.5825090016458547, + "grad_norm": 1.10556161403656, + "learning_rate": 1.7670804731999416e-05, + "loss": 3.8689, + "step": 51850 + }, + { + "epoch": 0.5830707268161979, + "grad_norm": 0.9998056292533875, + "learning_rate": 1.766855781869657e-05, + "loss": 3.8945, + "step": 51900 + }, + { + "epoch": 0.583632451986541, + "grad_norm": 1.2014474868774414, + "learning_rate": 1.766631090539372e-05, + "loss": 3.7499, + "step": 51950 + }, + { + "epoch": 0.5841941771568843, + "grad_norm": 1.1735973358154297, + "learning_rate": 1.7664063992090868e-05, + "loss": 3.825, + "step": 52000 + }, + { + "epoch": 0.5847559023272274, + "grad_norm": 1.0286104679107666, + "learning_rate": 1.7661817078788017e-05, + "loss": 3.7428, + "step": 52050 + }, + { + "epoch": 0.5853176274975705, + "grad_norm": 1.306618094444275, + "learning_rate": 1.7659570165485166e-05, + "loss": 3.7499, + "step": 52100 + }, + { + "epoch": 0.5858793526679137, + "grad_norm": 1.021728515625, + "learning_rate": 1.7657323252182316e-05, + "loss": 3.758, + "step": 52150 + }, + { + "epoch": 0.5864410778382568, + "grad_norm": 1.2071304321289062, + "learning_rate": 1.7655076338879465e-05, + "loss": 3.772, + "step": 52200 + }, + { + "epoch": 0.5870028030086, + "grad_norm": 1.4372752904891968, + "learning_rate": 1.7652829425576618e-05, + "loss": 3.8094, + "step": 52250 + }, + { + "epoch": 0.5875645281789432, + "grad_norm": 1.0974149703979492, + "learning_rate": 1.7650582512273767e-05, + "loss": 3.8042, + "step": 52300 + }, + { + "epoch": 0.5881262533492864, + "grad_norm": 1.1430492401123047, + "learning_rate": 1.7648335598970916e-05, + "loss": 3.7804, + "step": 52350 + }, + { + "epoch": 0.5886879785196295, + "grad_norm": 1.0837328433990479, + "learning_rate": 1.7646088685668066e-05, + "loss": 3.724, + "step": 52400 + }, + { + "epoch": 0.5892497036899726, + "grad_norm": 1.0210705995559692, + "learning_rate": 1.7643841772365215e-05, + "loss": 3.8235, + "step": 52450 + }, + { + "epoch": 0.5898114288603158, + "grad_norm": 1.0554842948913574, + "learning_rate": 1.7641594859062364e-05, + "loss": 3.7633, + "step": 52500 + }, + { + "epoch": 0.5903731540306589, + "grad_norm": 1.0610668659210205, + "learning_rate": 1.7639347945759514e-05, + "loss": 3.7446, + "step": 52550 + }, + { + "epoch": 0.5909348792010021, + "grad_norm": 1.0977150201797485, + "learning_rate": 1.7637101032456666e-05, + "loss": 3.648, + "step": 52600 + }, + { + "epoch": 0.5914966043713453, + "grad_norm": 1.229286789894104, + "learning_rate": 1.7634854119153816e-05, + "loss": 3.8027, + "step": 52650 + }, + { + "epoch": 0.5920583295416885, + "grad_norm": 1.0663516521453857, + "learning_rate": 1.7632607205850965e-05, + "loss": 3.7696, + "step": 52700 + }, + { + "epoch": 0.5926200547120316, + "grad_norm": 1.1510690450668335, + "learning_rate": 1.7630360292548114e-05, + "loss": 3.8121, + "step": 52750 + }, + { + "epoch": 0.5931817798823747, + "grad_norm": 1.0974013805389404, + "learning_rate": 1.7628113379245264e-05, + "loss": 3.8296, + "step": 52800 + }, + { + "epoch": 0.5937435050527179, + "grad_norm": 1.0291469097137451, + "learning_rate": 1.7625866465942413e-05, + "loss": 3.7295, + "step": 52850 + }, + { + "epoch": 0.594305230223061, + "grad_norm": 1.1655592918395996, + "learning_rate": 1.7623619552639562e-05, + "loss": 3.7481, + "step": 52900 + }, + { + "epoch": 0.5948669553934042, + "grad_norm": 1.3898149728775024, + "learning_rate": 1.7621372639336715e-05, + "loss": 3.8032, + "step": 52950 + }, + { + "epoch": 0.5954286805637474, + "grad_norm": 0.9948254823684692, + "learning_rate": 1.7619125726033864e-05, + "loss": 3.7734, + "step": 53000 + }, + { + "epoch": 0.5959904057340906, + "grad_norm": 1.1202493906021118, + "learning_rate": 1.7616878812731014e-05, + "loss": 3.7293, + "step": 53050 + }, + { + "epoch": 0.5965521309044337, + "grad_norm": 1.0994091033935547, + "learning_rate": 1.7614631899428163e-05, + "loss": 3.7558, + "step": 53100 + }, + { + "epoch": 0.5971138560747768, + "grad_norm": 1.0649958848953247, + "learning_rate": 1.7612384986125312e-05, + "loss": 3.7704, + "step": 53150 + }, + { + "epoch": 0.59767558124512, + "grad_norm": 1.0792980194091797, + "learning_rate": 1.761013807282246e-05, + "loss": 3.8017, + "step": 53200 + }, + { + "epoch": 0.5982373064154631, + "grad_norm": 1.0352792739868164, + "learning_rate": 1.760789115951961e-05, + "loss": 3.7618, + "step": 53250 + }, + { + "epoch": 0.5987990315858063, + "grad_norm": 1.2318084239959717, + "learning_rate": 1.7605644246216764e-05, + "loss": 3.7041, + "step": 53300 + }, + { + "epoch": 0.5993607567561495, + "grad_norm": 1.151821255683899, + "learning_rate": 1.7603397332913913e-05, + "loss": 3.792, + "step": 53350 + }, + { + "epoch": 0.5999224819264927, + "grad_norm": 1.0399080514907837, + "learning_rate": 1.7601150419611062e-05, + "loss": 3.8003, + "step": 53400 + }, + { + "epoch": 0.6004842070968358, + "grad_norm": 1.2682665586471558, + "learning_rate": 1.759890350630821e-05, + "loss": 3.7432, + "step": 53450 + }, + { + "epoch": 0.601045932267179, + "grad_norm": 1.3421366214752197, + "learning_rate": 1.759665659300536e-05, + "loss": 3.8493, + "step": 53500 + }, + { + "epoch": 0.6016076574375221, + "grad_norm": 1.2913321256637573, + "learning_rate": 1.759440967970251e-05, + "loss": 3.7462, + "step": 53550 + }, + { + "epoch": 0.6021693826078652, + "grad_norm": 1.2670401334762573, + "learning_rate": 1.759216276639966e-05, + "loss": 3.8123, + "step": 53600 + }, + { + "epoch": 0.6027311077782085, + "grad_norm": 0.8991925716400146, + "learning_rate": 1.7589915853096812e-05, + "loss": 3.767, + "step": 53650 + }, + { + "epoch": 0.6032928329485516, + "grad_norm": 1.1282976865768433, + "learning_rate": 1.758766893979396e-05, + "loss": 3.7155, + "step": 53700 + }, + { + "epoch": 0.6038545581188948, + "grad_norm": 1.3516185283660889, + "learning_rate": 1.7585466964757164e-05, + "loss": 3.7661, + "step": 53750 + }, + { + "epoch": 0.6044162832892379, + "grad_norm": 1.5720571279525757, + "learning_rate": 1.7583220051454313e-05, + "loss": 3.7899, + "step": 53800 + }, + { + "epoch": 0.604978008459581, + "grad_norm": 1.2103759050369263, + "learning_rate": 1.7580973138151466e-05, + "loss": 3.7905, + "step": 53850 + }, + { + "epoch": 0.6055397336299242, + "grad_norm": 1.0293008089065552, + "learning_rate": 1.7578726224848615e-05, + "loss": 3.7961, + "step": 53900 + }, + { + "epoch": 0.6061014588002673, + "grad_norm": 1.0778840780258179, + "learning_rate": 1.7576479311545765e-05, + "loss": 3.8131, + "step": 53950 + }, + { + "epoch": 0.6066631839706106, + "grad_norm": 1.012994647026062, + "learning_rate": 1.7574232398242914e-05, + "loss": 3.7049, + "step": 54000 + }, + { + "epoch": 0.6072249091409537, + "grad_norm": 1.2515206336975098, + "learning_rate": 1.7571985484940063e-05, + "loss": 3.7537, + "step": 54050 + }, + { + "epoch": 0.6077866343112969, + "grad_norm": 1.3033223152160645, + "learning_rate": 1.7569738571637213e-05, + "loss": 3.8329, + "step": 54100 + }, + { + "epoch": 0.60834835948164, + "grad_norm": 1.1931025981903076, + "learning_rate": 1.7567491658334362e-05, + "loss": 3.822, + "step": 54150 + }, + { + "epoch": 0.6089100846519832, + "grad_norm": 0.8393592238426208, + "learning_rate": 1.756528968329757e-05, + "loss": 3.6741, + "step": 54200 + }, + { + "epoch": 0.6094718098223263, + "grad_norm": 1.1479541063308716, + "learning_rate": 1.756304276999472e-05, + "loss": 3.6592, + "step": 54250 + }, + { + "epoch": 0.6100335349926694, + "grad_norm": 1.1731626987457275, + "learning_rate": 1.756079585669187e-05, + "loss": 3.7719, + "step": 54300 + }, + { + "epoch": 0.6105952601630127, + "grad_norm": 1.626962661743164, + "learning_rate": 1.755854894338902e-05, + "loss": 3.8338, + "step": 54350 + }, + { + "epoch": 0.6111569853333558, + "grad_norm": 0.9360986351966858, + "learning_rate": 1.755630203008617e-05, + "loss": 3.758, + "step": 54400 + }, + { + "epoch": 0.611718710503699, + "grad_norm": 1.112356185913086, + "learning_rate": 1.7554055116783318e-05, + "loss": 3.8045, + "step": 54450 + }, + { + "epoch": 0.6122804356740421, + "grad_norm": 1.0053960084915161, + "learning_rate": 1.755180820348047e-05, + "loss": 3.7159, + "step": 54500 + }, + { + "epoch": 0.6128421608443853, + "grad_norm": 1.3038718700408936, + "learning_rate": 1.754956129017762e-05, + "loss": 3.7571, + "step": 54550 + }, + { + "epoch": 0.6134038860147284, + "grad_norm": 1.116718053817749, + "learning_rate": 1.754731437687477e-05, + "loss": 3.7673, + "step": 54600 + }, + { + "epoch": 0.6139656111850716, + "grad_norm": 1.0109951496124268, + "learning_rate": 1.754506746357192e-05, + "loss": 3.8169, + "step": 54650 + }, + { + "epoch": 0.6145273363554148, + "grad_norm": 1.0003641843795776, + "learning_rate": 1.7542820550269068e-05, + "loss": 3.6633, + "step": 54700 + }, + { + "epoch": 0.6150890615257579, + "grad_norm": 1.0796260833740234, + "learning_rate": 1.7540573636966217e-05, + "loss": 3.7145, + "step": 54750 + }, + { + "epoch": 0.6156507866961011, + "grad_norm": 1.163537859916687, + "learning_rate": 1.7538326723663366e-05, + "loss": 3.8101, + "step": 54800 + }, + { + "epoch": 0.6162125118664442, + "grad_norm": 1.0152901411056519, + "learning_rate": 1.753607981036052e-05, + "loss": 3.7383, + "step": 54850 + }, + { + "epoch": 0.6167742370367874, + "grad_norm": 0.9968041777610779, + "learning_rate": 1.753383289705767e-05, + "loss": 3.7473, + "step": 54900 + }, + { + "epoch": 0.6173359622071305, + "grad_norm": 1.4523382186889648, + "learning_rate": 1.7531585983754818e-05, + "loss": 3.7098, + "step": 54950 + }, + { + "epoch": 0.6178976873774737, + "grad_norm": 1.306656002998352, + "learning_rate": 1.7529339070451967e-05, + "loss": 3.6992, + "step": 55000 + }, + { + "epoch": 0.6184594125478169, + "grad_norm": 1.0262917280197144, + "learning_rate": 1.7527092157149116e-05, + "loss": 3.826, + "step": 55050 + }, + { + "epoch": 0.61902113771816, + "grad_norm": 1.0945398807525635, + "learning_rate": 1.7524845243846266e-05, + "loss": 3.836, + "step": 55100 + }, + { + "epoch": 0.6195828628885032, + "grad_norm": 1.0034478902816772, + "learning_rate": 1.7522598330543415e-05, + "loss": 3.7401, + "step": 55150 + }, + { + "epoch": 0.6201445880588463, + "grad_norm": 1.073586106300354, + "learning_rate": 1.7520351417240568e-05, + "loss": 3.7507, + "step": 55200 + }, + { + "epoch": 0.6207063132291895, + "grad_norm": 1.35128653049469, + "learning_rate": 1.7518104503937717e-05, + "loss": 3.8347, + "step": 55250 + }, + { + "epoch": 0.6212680383995326, + "grad_norm": 1.1628330945968628, + "learning_rate": 1.7515857590634866e-05, + "loss": 3.7984, + "step": 55300 + }, + { + "epoch": 0.6218297635698758, + "grad_norm": 1.1050388813018799, + "learning_rate": 1.7513610677332016e-05, + "loss": 3.679, + "step": 55350 + }, + { + "epoch": 0.622391488740219, + "grad_norm": 1.391716718673706, + "learning_rate": 1.7511363764029165e-05, + "loss": 3.7539, + "step": 55400 + }, + { + "epoch": 0.6229532139105621, + "grad_norm": 1.0777093172073364, + "learning_rate": 1.7509116850726314e-05, + "loss": 3.8025, + "step": 55450 + }, + { + "epoch": 0.6235149390809053, + "grad_norm": 1.2782025337219238, + "learning_rate": 1.7506869937423464e-05, + "loss": 3.7464, + "step": 55500 + }, + { + "epoch": 0.6240766642512484, + "grad_norm": 0.9614410996437073, + "learning_rate": 1.7504623024120616e-05, + "loss": 3.7748, + "step": 55550 + }, + { + "epoch": 0.6246383894215916, + "grad_norm": 1.1084448099136353, + "learning_rate": 1.7502376110817766e-05, + "loss": 3.7589, + "step": 55600 + }, + { + "epoch": 0.6252001145919347, + "grad_norm": 1.0383925437927246, + "learning_rate": 1.7500129197514915e-05, + "loss": 3.7624, + "step": 55650 + }, + { + "epoch": 0.625761839762278, + "grad_norm": 1.2714895009994507, + "learning_rate": 1.7497882284212064e-05, + "loss": 3.8049, + "step": 55700 + }, + { + "epoch": 0.6263235649326211, + "grad_norm": 1.1721174716949463, + "learning_rate": 1.7495635370909214e-05, + "loss": 3.7391, + "step": 55750 + }, + { + "epoch": 0.6268852901029642, + "grad_norm": 1.386572241783142, + "learning_rate": 1.7493388457606363e-05, + "loss": 3.6544, + "step": 55800 + }, + { + "epoch": 0.6274470152733074, + "grad_norm": 1.186914086341858, + "learning_rate": 1.7491141544303512e-05, + "loss": 3.7147, + "step": 55850 + }, + { + "epoch": 0.6280087404436505, + "grad_norm": 1.1338549852371216, + "learning_rate": 1.7488894631000665e-05, + "loss": 3.7277, + "step": 55900 + }, + { + "epoch": 0.6285704656139937, + "grad_norm": 1.3115663528442383, + "learning_rate": 1.7486647717697814e-05, + "loss": 3.7401, + "step": 55950 + }, + { + "epoch": 0.6291321907843369, + "grad_norm": 1.172100305557251, + "learning_rate": 1.7484400804394964e-05, + "loss": 3.8122, + "step": 56000 + }, + { + "epoch": 0.62969391595468, + "grad_norm": 1.0340574979782104, + "learning_rate": 1.7482153891092113e-05, + "loss": 3.8087, + "step": 56050 + }, + { + "epoch": 0.6302556411250232, + "grad_norm": 1.0701868534088135, + "learning_rate": 1.7479906977789262e-05, + "loss": 3.7115, + "step": 56100 + }, + { + "epoch": 0.6308173662953663, + "grad_norm": 1.0869747400283813, + "learning_rate": 1.747766006448641e-05, + "loss": 3.7915, + "step": 56150 + }, + { + "epoch": 0.6313790914657095, + "grad_norm": 1.1002542972564697, + "learning_rate": 1.747541315118356e-05, + "loss": 3.6988, + "step": 56200 + }, + { + "epoch": 0.6319408166360526, + "grad_norm": 1.124596357345581, + "learning_rate": 1.7473166237880714e-05, + "loss": 3.78, + "step": 56250 + }, + { + "epoch": 0.6325025418063958, + "grad_norm": 1.2207372188568115, + "learning_rate": 1.7470919324577863e-05, + "loss": 3.7913, + "step": 56300 + }, + { + "epoch": 0.633064266976739, + "grad_norm": 1.2835536003112793, + "learning_rate": 1.7468672411275012e-05, + "loss": 3.7761, + "step": 56350 + }, + { + "epoch": 0.6336259921470822, + "grad_norm": 1.2327477931976318, + "learning_rate": 1.746642549797216e-05, + "loss": 3.8194, + "step": 56400 + }, + { + "epoch": 0.6341877173174253, + "grad_norm": 1.2116644382476807, + "learning_rate": 1.7464223522935367e-05, + "loss": 3.8478, + "step": 56450 + }, + { + "epoch": 0.6347494424877684, + "grad_norm": 1.1993951797485352, + "learning_rate": 1.7461976609632517e-05, + "loss": 3.7858, + "step": 56500 + }, + { + "epoch": 0.6353111676581116, + "grad_norm": 1.223915457725525, + "learning_rate": 1.745972969632967e-05, + "loss": 3.7529, + "step": 56550 + }, + { + "epoch": 0.6358728928284547, + "grad_norm": 1.1108373403549194, + "learning_rate": 1.745748278302682e-05, + "loss": 3.7742, + "step": 56600 + }, + { + "epoch": 0.6364346179987979, + "grad_norm": 1.0354841947555542, + "learning_rate": 1.7455235869723968e-05, + "loss": 3.7672, + "step": 56650 + }, + { + "epoch": 0.6369963431691411, + "grad_norm": 1.059706687927246, + "learning_rate": 1.7452988956421117e-05, + "loss": 3.7729, + "step": 56700 + }, + { + "epoch": 0.6375580683394843, + "grad_norm": 1.0827964544296265, + "learning_rate": 1.7450742043118267e-05, + "loss": 3.6867, + "step": 56750 + }, + { + "epoch": 0.6381197935098274, + "grad_norm": 1.078853964805603, + "learning_rate": 1.7448495129815416e-05, + "loss": 3.7249, + "step": 56800 + }, + { + "epoch": 0.6386815186801705, + "grad_norm": 1.1734760999679565, + "learning_rate": 1.7446248216512565e-05, + "loss": 3.8154, + "step": 56850 + }, + { + "epoch": 0.6392432438505137, + "grad_norm": 1.1998199224472046, + "learning_rate": 1.7444001303209718e-05, + "loss": 3.7453, + "step": 56900 + }, + { + "epoch": 0.6398049690208568, + "grad_norm": 1.1486512422561646, + "learning_rate": 1.7441754389906867e-05, + "loss": 3.7046, + "step": 56950 + }, + { + "epoch": 0.6403666941912, + "grad_norm": 2.4619200229644775, + "learning_rate": 1.7439507476604017e-05, + "loss": 3.7254, + "step": 57000 + }, + { + "epoch": 0.6409284193615432, + "grad_norm": 1.2601149082183838, + "learning_rate": 1.7437260563301166e-05, + "loss": 3.7013, + "step": 57050 + }, + { + "epoch": 0.6414901445318864, + "grad_norm": 1.07093346118927, + "learning_rate": 1.7435013649998315e-05, + "loss": 3.8067, + "step": 57100 + }, + { + "epoch": 0.6420518697022295, + "grad_norm": 1.187065839767456, + "learning_rate": 1.7432766736695465e-05, + "loss": 3.7688, + "step": 57150 + }, + { + "epoch": 0.6426135948725726, + "grad_norm": 1.0475225448608398, + "learning_rate": 1.7430519823392614e-05, + "loss": 3.7031, + "step": 57200 + }, + { + "epoch": 0.6431753200429158, + "grad_norm": 1.1916700601577759, + "learning_rate": 1.7428272910089767e-05, + "loss": 3.7313, + "step": 57250 + }, + { + "epoch": 0.6437370452132589, + "grad_norm": 1.2383705377578735, + "learning_rate": 1.7426025996786916e-05, + "loss": 3.8268, + "step": 57300 + }, + { + "epoch": 0.6442987703836022, + "grad_norm": 1.1119415760040283, + "learning_rate": 1.7423779083484065e-05, + "loss": 3.7676, + "step": 57350 + }, + { + "epoch": 0.6448604955539453, + "grad_norm": 1.2077629566192627, + "learning_rate": 1.7421532170181215e-05, + "loss": 3.8255, + "step": 57400 + }, + { + "epoch": 0.6454222207242885, + "grad_norm": 1.078245997428894, + "learning_rate": 1.7419285256878364e-05, + "loss": 3.7174, + "step": 57450 + }, + { + "epoch": 0.6459839458946316, + "grad_norm": 1.1462548971176147, + "learning_rate": 1.7417038343575513e-05, + "loss": 3.8094, + "step": 57500 + }, + { + "epoch": 0.6465456710649747, + "grad_norm": 1.1107356548309326, + "learning_rate": 1.7414791430272663e-05, + "loss": 3.8238, + "step": 57550 + }, + { + "epoch": 0.6471073962353179, + "grad_norm": 0.9890028834342957, + "learning_rate": 1.7412544516969815e-05, + "loss": 3.8116, + "step": 57600 + }, + { + "epoch": 0.647669121405661, + "grad_norm": 1.0611302852630615, + "learning_rate": 1.7410297603666965e-05, + "loss": 3.7203, + "step": 57650 + }, + { + "epoch": 0.6482308465760043, + "grad_norm": 1.0499742031097412, + "learning_rate": 1.7408050690364114e-05, + "loss": 3.7378, + "step": 57700 + }, + { + "epoch": 0.6487925717463474, + "grad_norm": 1.1752382516860962, + "learning_rate": 1.7405803777061263e-05, + "loss": 3.7246, + "step": 57750 + }, + { + "epoch": 0.6493542969166906, + "grad_norm": 1.149659514427185, + "learning_rate": 1.7403556863758413e-05, + "loss": 3.7695, + "step": 57800 + }, + { + "epoch": 0.6499160220870337, + "grad_norm": 0.9258421063423157, + "learning_rate": 1.740135488872162e-05, + "loss": 3.7976, + "step": 57850 + }, + { + "epoch": 0.6504777472573768, + "grad_norm": 1.2829738855361938, + "learning_rate": 1.739910797541877e-05, + "loss": 3.7877, + "step": 57900 + }, + { + "epoch": 0.65103947242772, + "grad_norm": 1.0089658498764038, + "learning_rate": 1.739686106211592e-05, + "loss": 3.7311, + "step": 57950 + }, + { + "epoch": 0.6516011975980631, + "grad_norm": 1.102944016456604, + "learning_rate": 1.739461414881307e-05, + "loss": 3.7644, + "step": 58000 + }, + { + "epoch": 0.6521629227684064, + "grad_norm": 1.213724136352539, + "learning_rate": 1.739236723551022e-05, + "loss": 3.7867, + "step": 58050 + }, + { + "epoch": 0.6527246479387495, + "grad_norm": 1.135790467262268, + "learning_rate": 1.739012032220737e-05, + "loss": 3.7394, + "step": 58100 + }, + { + "epoch": 0.6532863731090927, + "grad_norm": 1.1218922138214111, + "learning_rate": 1.7387873408904518e-05, + "loss": 3.8348, + "step": 58150 + }, + { + "epoch": 0.6538480982794358, + "grad_norm": 1.5928314924240112, + "learning_rate": 1.7385626495601667e-05, + "loss": 3.6883, + "step": 58200 + }, + { + "epoch": 0.6544098234497789, + "grad_norm": 1.0138171911239624, + "learning_rate": 1.738337958229882e-05, + "loss": 3.6912, + "step": 58250 + }, + { + "epoch": 0.6549715486201221, + "grad_norm": 1.1871720552444458, + "learning_rate": 1.738113266899597e-05, + "loss": 3.7674, + "step": 58300 + }, + { + "epoch": 0.6555332737904653, + "grad_norm": 1.1729800701141357, + "learning_rate": 1.737888575569312e-05, + "loss": 3.7707, + "step": 58350 + }, + { + "epoch": 0.6560949989608085, + "grad_norm": 0.9070509076118469, + "learning_rate": 1.7376638842390268e-05, + "loss": 3.8358, + "step": 58400 + }, + { + "epoch": 0.6566567241311516, + "grad_norm": 1.1585700511932373, + "learning_rate": 1.7374391929087417e-05, + "loss": 3.7481, + "step": 58450 + }, + { + "epoch": 0.6572184493014948, + "grad_norm": 1.192552924156189, + "learning_rate": 1.7372145015784566e-05, + "loss": 3.7324, + "step": 58500 + }, + { + "epoch": 0.6577801744718379, + "grad_norm": 1.0995116233825684, + "learning_rate": 1.7369898102481716e-05, + "loss": 3.814, + "step": 58550 + }, + { + "epoch": 0.658341899642181, + "grad_norm": 1.1598093509674072, + "learning_rate": 1.736765118917887e-05, + "loss": 3.8403, + "step": 58600 + }, + { + "epoch": 0.6589036248125242, + "grad_norm": 0.9769778847694397, + "learning_rate": 1.7365404275876018e-05, + "loss": 3.744, + "step": 58650 + }, + { + "epoch": 0.6594653499828674, + "grad_norm": 1.102225661277771, + "learning_rate": 1.7363157362573167e-05, + "loss": 3.8054, + "step": 58700 + }, + { + "epoch": 0.6600270751532106, + "grad_norm": 1.0175129175186157, + "learning_rate": 1.7360910449270316e-05, + "loss": 3.6721, + "step": 58750 + }, + { + "epoch": 0.6605888003235537, + "grad_norm": 1.1308871507644653, + "learning_rate": 1.7358663535967466e-05, + "loss": 3.7017, + "step": 58800 + }, + { + "epoch": 0.6611505254938969, + "grad_norm": 1.059259057044983, + "learning_rate": 1.7356416622664615e-05, + "loss": 3.7659, + "step": 58850 + }, + { + "epoch": 0.66171225066424, + "grad_norm": 0.8596951961517334, + "learning_rate": 1.7354169709361764e-05, + "loss": 3.6726, + "step": 58900 + }, + { + "epoch": 0.6622739758345831, + "grad_norm": 1.2974934577941895, + "learning_rate": 1.7351922796058917e-05, + "loss": 3.74, + "step": 58950 + }, + { + "epoch": 0.6628357010049263, + "grad_norm": 1.1300395727157593, + "learning_rate": 1.7349675882756066e-05, + "loss": 3.7653, + "step": 59000 + }, + { + "epoch": 0.6633974261752695, + "grad_norm": 1.097085952758789, + "learning_rate": 1.7347428969453216e-05, + "loss": 3.8094, + "step": 59050 + }, + { + "epoch": 0.6639591513456127, + "grad_norm": 1.1728578805923462, + "learning_rate": 1.7345182056150365e-05, + "loss": 3.794, + "step": 59100 + }, + { + "epoch": 0.6645208765159558, + "grad_norm": 1.0365574359893799, + "learning_rate": 1.7342935142847514e-05, + "loss": 3.7189, + "step": 59150 + }, + { + "epoch": 0.665082601686299, + "grad_norm": 1.1213483810424805, + "learning_rate": 1.7340688229544664e-05, + "loss": 3.7355, + "step": 59200 + }, + { + "epoch": 0.6656443268566421, + "grad_norm": 1.0311239957809448, + "learning_rate": 1.7338441316241813e-05, + "loss": 3.7611, + "step": 59250 + }, + { + "epoch": 0.6662060520269852, + "grad_norm": 1.113905906677246, + "learning_rate": 1.7336194402938966e-05, + "loss": 3.7474, + "step": 59300 + }, + { + "epoch": 0.6667677771973284, + "grad_norm": 0.9911109209060669, + "learning_rate": 1.7333947489636115e-05, + "loss": 3.751, + "step": 59350 + }, + { + "epoch": 0.6673295023676716, + "grad_norm": 1.1860027313232422, + "learning_rate": 1.7331700576333264e-05, + "loss": 3.7477, + "step": 59400 + }, + { + "epoch": 0.6678912275380148, + "grad_norm": 1.041394591331482, + "learning_rate": 1.7329453663030414e-05, + "loss": 3.6717, + "step": 59450 + }, + { + "epoch": 0.6684529527083579, + "grad_norm": 1.0377204418182373, + "learning_rate": 1.7327206749727563e-05, + "loss": 3.7745, + "step": 59500 + }, + { + "epoch": 0.6690146778787011, + "grad_norm": 1.1058553457260132, + "learning_rate": 1.7324959836424712e-05, + "loss": 3.7069, + "step": 59550 + }, + { + "epoch": 0.6695764030490442, + "grad_norm": 1.2549515962600708, + "learning_rate": 1.732275786138792e-05, + "loss": 3.7848, + "step": 59600 + }, + { + "epoch": 0.6701381282193873, + "grad_norm": 1.1881179809570312, + "learning_rate": 1.732051094808507e-05, + "loss": 3.8035, + "step": 59650 + }, + { + "epoch": 0.6706998533897306, + "grad_norm": 1.087872862815857, + "learning_rate": 1.731826403478222e-05, + "loss": 3.7718, + "step": 59700 + }, + { + "epoch": 0.6712615785600737, + "grad_norm": 1.0337735414505005, + "learning_rate": 1.731601712147937e-05, + "loss": 3.7516, + "step": 59750 + }, + { + "epoch": 0.6718233037304169, + "grad_norm": 1.0685954093933105, + "learning_rate": 1.731377020817652e-05, + "loss": 3.7406, + "step": 59800 + }, + { + "epoch": 0.67238502890076, + "grad_norm": 1.3235671520233154, + "learning_rate": 1.7311523294873668e-05, + "loss": 3.7622, + "step": 59850 + }, + { + "epoch": 0.6729467540711032, + "grad_norm": 1.0132572650909424, + "learning_rate": 1.7309276381570817e-05, + "loss": 3.7176, + "step": 59900 + }, + { + "epoch": 0.6735084792414463, + "grad_norm": 1.0859169960021973, + "learning_rate": 1.730702946826797e-05, + "loss": 3.7555, + "step": 59950 + }, + { + "epoch": 0.6740702044117894, + "grad_norm": 1.0595098733901978, + "learning_rate": 1.730478255496512e-05, + "loss": 3.7683, + "step": 60000 + }, + { + "epoch": 0.6746319295821327, + "grad_norm": 1.2074511051177979, + "learning_rate": 1.730253564166227e-05, + "loss": 3.7437, + "step": 60050 + }, + { + "epoch": 0.6751936547524758, + "grad_norm": 1.5620605945587158, + "learning_rate": 1.7300288728359418e-05, + "loss": 3.7324, + "step": 60100 + }, + { + "epoch": 0.675755379922819, + "grad_norm": 1.1325360536575317, + "learning_rate": 1.7298041815056567e-05, + "loss": 3.7337, + "step": 60150 + }, + { + "epoch": 0.6763171050931621, + "grad_norm": 1.1669329404830933, + "learning_rate": 1.7295794901753717e-05, + "loss": 3.7386, + "step": 60200 + }, + { + "epoch": 0.6768788302635053, + "grad_norm": 1.217720866203308, + "learning_rate": 1.7293547988450866e-05, + "loss": 3.7491, + "step": 60250 + }, + { + "epoch": 0.6774405554338484, + "grad_norm": 1.1027276515960693, + "learning_rate": 1.729130107514802e-05, + "loss": 3.828, + "step": 60300 + }, + { + "epoch": 0.6780022806041915, + "grad_norm": 1.3892951011657715, + "learning_rate": 1.7289054161845168e-05, + "loss": 3.8093, + "step": 60350 + }, + { + "epoch": 0.6785640057745348, + "grad_norm": 1.090294599533081, + "learning_rate": 1.7286807248542317e-05, + "loss": 3.8163, + "step": 60400 + }, + { + "epoch": 0.6791257309448779, + "grad_norm": 1.0623295307159424, + "learning_rate": 1.7284560335239467e-05, + "loss": 3.832, + "step": 60450 + }, + { + "epoch": 0.6796874561152211, + "grad_norm": 1.1373815536499023, + "learning_rate": 1.7282313421936616e-05, + "loss": 3.7288, + "step": 60500 + }, + { + "epoch": 0.6802491812855642, + "grad_norm": 1.2594444751739502, + "learning_rate": 1.7280066508633765e-05, + "loss": 3.8319, + "step": 60550 + }, + { + "epoch": 0.6808109064559074, + "grad_norm": 1.963243007659912, + "learning_rate": 1.7277819595330915e-05, + "loss": 3.6892, + "step": 60600 + }, + { + "epoch": 0.6813726316262505, + "grad_norm": 1.0691497325897217, + "learning_rate": 1.7275572682028064e-05, + "loss": 3.7616, + "step": 60650 + }, + { + "epoch": 0.6819343567965936, + "grad_norm": 1.0307921171188354, + "learning_rate": 1.7273325768725217e-05, + "loss": 3.7629, + "step": 60700 + }, + { + "epoch": 0.6824960819669369, + "grad_norm": 1.1926541328430176, + "learning_rate": 1.7271078855422366e-05, + "loss": 3.7875, + "step": 60750 + }, + { + "epoch": 0.68305780713728, + "grad_norm": 1.0767431259155273, + "learning_rate": 1.7268831942119515e-05, + "loss": 3.6975, + "step": 60800 + }, + { + "epoch": 0.6836195323076232, + "grad_norm": 1.0437264442443848, + "learning_rate": 1.7266585028816665e-05, + "loss": 3.7654, + "step": 60850 + }, + { + "epoch": 0.6841812574779663, + "grad_norm": 1.0138980150222778, + "learning_rate": 1.7264338115513814e-05, + "loss": 3.766, + "step": 60900 + }, + { + "epoch": 0.6847429826483095, + "grad_norm": 1.0664023160934448, + "learning_rate": 1.7262091202210963e-05, + "loss": 3.8365, + "step": 60950 + }, + { + "epoch": 0.6853047078186526, + "grad_norm": 1.0536384582519531, + "learning_rate": 1.7259844288908113e-05, + "loss": 3.8036, + "step": 61000 + }, + { + "epoch": 0.6858664329889959, + "grad_norm": 1.0200104713439941, + "learning_rate": 1.7257597375605265e-05, + "loss": 3.8067, + "step": 61050 + }, + { + "epoch": 0.686428158159339, + "grad_norm": 1.2039310932159424, + "learning_rate": 1.7255350462302415e-05, + "loss": 3.8287, + "step": 61100 + }, + { + "epoch": 0.6869898833296821, + "grad_norm": 1.0688422918319702, + "learning_rate": 1.7253103548999564e-05, + "loss": 3.6781, + "step": 61150 + }, + { + "epoch": 0.6875516085000253, + "grad_norm": 1.2913188934326172, + "learning_rate": 1.7250856635696713e-05, + "loss": 3.7841, + "step": 61200 + }, + { + "epoch": 0.6881133336703684, + "grad_norm": 0.9989462494850159, + "learning_rate": 1.7248609722393863e-05, + "loss": 3.7842, + "step": 61250 + }, + { + "epoch": 0.6886750588407116, + "grad_norm": 1.056571125984192, + "learning_rate": 1.7246362809091012e-05, + "loss": 3.7838, + "step": 61300 + }, + { + "epoch": 0.6892367840110547, + "grad_norm": 1.1829471588134766, + "learning_rate": 1.724411589578816e-05, + "loss": 3.7119, + "step": 61350 + }, + { + "epoch": 0.689798509181398, + "grad_norm": 1.0671151876449585, + "learning_rate": 1.7241868982485314e-05, + "loss": 3.7321, + "step": 61400 + }, + { + "epoch": 0.6903602343517411, + "grad_norm": 1.1066913604736328, + "learning_rate": 1.7239622069182463e-05, + "loss": 3.7692, + "step": 61450 + }, + { + "epoch": 0.6909219595220842, + "grad_norm": 1.166175127029419, + "learning_rate": 1.7237375155879612e-05, + "loss": 3.736, + "step": 61500 + }, + { + "epoch": 0.6914836846924274, + "grad_norm": 1.100232481956482, + "learning_rate": 1.7235128242576762e-05, + "loss": 3.8028, + "step": 61550 + }, + { + "epoch": 0.6920454098627705, + "grad_norm": 1.2112360000610352, + "learning_rate": 1.723288132927391e-05, + "loss": 3.7698, + "step": 61600 + }, + { + "epoch": 0.6926071350331137, + "grad_norm": 1.086358904838562, + "learning_rate": 1.723063441597106e-05, + "loss": 3.7637, + "step": 61650 + }, + { + "epoch": 0.6931688602034568, + "grad_norm": 1.1099216938018799, + "learning_rate": 1.722838750266821e-05, + "loss": 3.7888, + "step": 61700 + }, + { + "epoch": 0.6937305853738001, + "grad_norm": 0.988438606262207, + "learning_rate": 1.7226140589365362e-05, + "loss": 3.8313, + "step": 61750 + }, + { + "epoch": 0.6942923105441432, + "grad_norm": 1.0678561925888062, + "learning_rate": 1.7223893676062512e-05, + "loss": 3.7451, + "step": 61800 + }, + { + "epoch": 0.6948540357144863, + "grad_norm": 1.293940782546997, + "learning_rate": 1.722164676275966e-05, + "loss": 3.7087, + "step": 61850 + }, + { + "epoch": 0.6954157608848295, + "grad_norm": 1.0884907245635986, + "learning_rate": 1.721939984945681e-05, + "loss": 3.7037, + "step": 61900 + }, + { + "epoch": 0.6959774860551726, + "grad_norm": 1.0982439517974854, + "learning_rate": 1.721715293615396e-05, + "loss": 3.6668, + "step": 61950 + }, + { + "epoch": 0.6965392112255158, + "grad_norm": 1.0753116607666016, + "learning_rate": 1.721490602285111e-05, + "loss": 3.6831, + "step": 62000 + }, + { + "epoch": 0.697100936395859, + "grad_norm": 1.009289026260376, + "learning_rate": 1.721265910954826e-05, + "loss": 3.7106, + "step": 62050 + }, + { + "epoch": 0.6976626615662022, + "grad_norm": 1.2232242822647095, + "learning_rate": 1.721041219624541e-05, + "loss": 3.7235, + "step": 62100 + }, + { + "epoch": 0.6982243867365453, + "grad_norm": 1.4713053703308105, + "learning_rate": 1.720816528294256e-05, + "loss": 3.7499, + "step": 62150 + }, + { + "epoch": 0.6987861119068884, + "grad_norm": 1.074111819267273, + "learning_rate": 1.720591836963971e-05, + "loss": 3.8075, + "step": 62200 + }, + { + "epoch": 0.6993478370772316, + "grad_norm": 1.2166283130645752, + "learning_rate": 1.720367145633686e-05, + "loss": 3.6972, + "step": 62250 + }, + { + "epoch": 0.6999095622475747, + "grad_norm": 1.212631344795227, + "learning_rate": 1.720142454303401e-05, + "loss": 3.7652, + "step": 62300 + }, + { + "epoch": 0.7004712874179179, + "grad_norm": 0.9701681137084961, + "learning_rate": 1.7199177629731158e-05, + "loss": 3.7186, + "step": 62350 + }, + { + "epoch": 0.7010330125882611, + "grad_norm": 1.0456335544586182, + "learning_rate": 1.7196930716428307e-05, + "loss": 3.8562, + "step": 62400 + }, + { + "epoch": 0.7015947377586043, + "grad_norm": 1.0489026308059692, + "learning_rate": 1.719468380312546e-05, + "loss": 3.6795, + "step": 62450 + }, + { + "epoch": 0.7021564629289474, + "grad_norm": 1.1351799964904785, + "learning_rate": 1.719243688982261e-05, + "loss": 3.766, + "step": 62500 + }, + { + "epoch": 0.7027181880992905, + "grad_norm": 1.2965370416641235, + "learning_rate": 1.7190189976519758e-05, + "loss": 3.8151, + "step": 62550 + }, + { + "epoch": 0.7032799132696337, + "grad_norm": 0.9872881174087524, + "learning_rate": 1.7187943063216908e-05, + "loss": 3.7155, + "step": 62600 + }, + { + "epoch": 0.7038416384399768, + "grad_norm": 1.045161247253418, + "learning_rate": 1.7185696149914057e-05, + "loss": 3.8091, + "step": 62650 + }, + { + "epoch": 0.70440336361032, + "grad_norm": 1.2162364721298218, + "learning_rate": 1.7183449236611206e-05, + "loss": 3.8024, + "step": 62700 + }, + { + "epoch": 0.7049650887806632, + "grad_norm": 1.1977596282958984, + "learning_rate": 1.7181202323308356e-05, + "loss": 3.8153, + "step": 62750 + }, + { + "epoch": 0.7055268139510064, + "grad_norm": 1.1208105087280273, + "learning_rate": 1.7178955410005508e-05, + "loss": 3.778, + "step": 62800 + }, + { + "epoch": 0.7060885391213495, + "grad_norm": 1.4856350421905518, + "learning_rate": 1.7176708496702658e-05, + "loss": 3.7663, + "step": 62850 + }, + { + "epoch": 0.7066502642916926, + "grad_norm": 1.2106389999389648, + "learning_rate": 1.7174461583399807e-05, + "loss": 3.7157, + "step": 62900 + }, + { + "epoch": 0.7072119894620358, + "grad_norm": 1.0710458755493164, + "learning_rate": 1.7172214670096956e-05, + "loss": 3.7688, + "step": 62950 + }, + { + "epoch": 0.7077737146323789, + "grad_norm": 1.059740424156189, + "learning_rate": 1.7169967756794106e-05, + "loss": 3.6703, + "step": 63000 + }, + { + "epoch": 0.7083354398027221, + "grad_norm": 1.0456651449203491, + "learning_rate": 1.7167720843491255e-05, + "loss": 3.7715, + "step": 63050 + }, + { + "epoch": 0.7088971649730653, + "grad_norm": 1.1711243391036987, + "learning_rate": 1.7165473930188404e-05, + "loss": 3.7545, + "step": 63100 + }, + { + "epoch": 0.7094588901434085, + "grad_norm": 1.1514251232147217, + "learning_rate": 1.7163227016885557e-05, + "loss": 3.7125, + "step": 63150 + }, + { + "epoch": 0.7100206153137516, + "grad_norm": 1.2115432024002075, + "learning_rate": 1.7160980103582706e-05, + "loss": 3.7972, + "step": 63200 + }, + { + "epoch": 0.7105823404840947, + "grad_norm": 1.1310622692108154, + "learning_rate": 1.7158733190279855e-05, + "loss": 3.7626, + "step": 63250 + }, + { + "epoch": 0.7111440656544379, + "grad_norm": 1.085196614265442, + "learning_rate": 1.7156486276977005e-05, + "loss": 3.8308, + "step": 63300 + }, + { + "epoch": 0.711705790824781, + "grad_norm": 1.0386687517166138, + "learning_rate": 1.7154239363674154e-05, + "loss": 3.8163, + "step": 63350 + }, + { + "epoch": 0.7122675159951243, + "grad_norm": 1.1654423475265503, + "learning_rate": 1.7151992450371303e-05, + "loss": 3.6759, + "step": 63400 + }, + { + "epoch": 0.7128292411654674, + "grad_norm": 1.0651174783706665, + "learning_rate": 1.7149745537068453e-05, + "loss": 3.717, + "step": 63450 + }, + { + "epoch": 0.7133909663358106, + "grad_norm": 1.1307576894760132, + "learning_rate": 1.7147498623765605e-05, + "loss": 3.8315, + "step": 63500 + }, + { + "epoch": 0.7139526915061537, + "grad_norm": 1.1386432647705078, + "learning_rate": 1.7145251710462755e-05, + "loss": 3.7392, + "step": 63550 + }, + { + "epoch": 0.7145144166764968, + "grad_norm": 1.1393775939941406, + "learning_rate": 1.7143004797159904e-05, + "loss": 3.705, + "step": 63600 + }, + { + "epoch": 0.71507614184684, + "grad_norm": 1.118025779724121, + "learning_rate": 1.7140757883857053e-05, + "loss": 3.699, + "step": 63650 + }, + { + "epoch": 0.7156378670171831, + "grad_norm": 0.9614808559417725, + "learning_rate": 1.7138510970554203e-05, + "loss": 3.7437, + "step": 63700 + }, + { + "epoch": 0.7161995921875264, + "grad_norm": 1.0394705533981323, + "learning_rate": 1.7136264057251352e-05, + "loss": 3.7325, + "step": 63750 + }, + { + "epoch": 0.7167613173578695, + "grad_norm": 1.0388416051864624, + "learning_rate": 1.71340171439485e-05, + "loss": 3.7791, + "step": 63800 + }, + { + "epoch": 0.7173230425282127, + "grad_norm": 1.1777559518814087, + "learning_rate": 1.713177023064565e-05, + "loss": 3.7322, + "step": 63850 + }, + { + "epoch": 0.7178847676985558, + "grad_norm": 1.1553236246109009, + "learning_rate": 1.7129523317342803e-05, + "loss": 3.7408, + "step": 63900 + }, + { + "epoch": 0.718446492868899, + "grad_norm": 1.136378526687622, + "learning_rate": 1.7127276404039953e-05, + "loss": 3.8013, + "step": 63950 + }, + { + "epoch": 0.7190082180392421, + "grad_norm": 1.2489324808120728, + "learning_rate": 1.7125029490737102e-05, + "loss": 3.8406, + "step": 64000 + }, + { + "epoch": 0.7195699432095852, + "grad_norm": 1.0770765542984009, + "learning_rate": 1.712278257743425e-05, + "loss": 3.7281, + "step": 64050 + }, + { + "epoch": 0.7201316683799285, + "grad_norm": 0.9865077137947083, + "learning_rate": 1.71205356641314e-05, + "loss": 3.7968, + "step": 64100 + }, + { + "epoch": 0.7206933935502716, + "grad_norm": 1.0182303190231323, + "learning_rate": 1.711828875082855e-05, + "loss": 3.6896, + "step": 64150 + }, + { + "epoch": 0.7212551187206148, + "grad_norm": 0.9671421051025391, + "learning_rate": 1.71160418375257e-05, + "loss": 3.8275, + "step": 64200 + }, + { + "epoch": 0.7218168438909579, + "grad_norm": 1.1231694221496582, + "learning_rate": 1.7113794924222852e-05, + "loss": 3.6555, + "step": 64250 + }, + { + "epoch": 0.722378569061301, + "grad_norm": 1.1389505863189697, + "learning_rate": 1.711154801092e-05, + "loss": 3.7869, + "step": 64300 + }, + { + "epoch": 0.7229402942316442, + "grad_norm": 1.0590343475341797, + "learning_rate": 1.710930109761715e-05, + "loss": 3.6597, + "step": 64350 + }, + { + "epoch": 0.7235020194019873, + "grad_norm": 1.0033551454544067, + "learning_rate": 1.71070541843143e-05, + "loss": 3.8088, + "step": 64400 + }, + { + "epoch": 0.7240637445723306, + "grad_norm": 1.258637547492981, + "learning_rate": 1.710480727101145e-05, + "loss": 3.734, + "step": 64450 + }, + { + "epoch": 0.7246254697426737, + "grad_norm": 1.2087608575820923, + "learning_rate": 1.71025603577086e-05, + "loss": 3.7397, + "step": 64500 + }, + { + "epoch": 0.7251871949130169, + "grad_norm": 1.1399991512298584, + "learning_rate": 1.7100313444405748e-05, + "loss": 3.762, + "step": 64550 + }, + { + "epoch": 0.72574892008336, + "grad_norm": 0.8821276426315308, + "learning_rate": 1.70980665311029e-05, + "loss": 3.6886, + "step": 64600 + }, + { + "epoch": 0.7263106452537031, + "grad_norm": 1.2095016241073608, + "learning_rate": 1.709581961780005e-05, + "loss": 3.7076, + "step": 64650 + }, + { + "epoch": 0.7268723704240463, + "grad_norm": 1.2419086694717407, + "learning_rate": 1.70935727044972e-05, + "loss": 3.8058, + "step": 64700 + }, + { + "epoch": 0.7274340955943895, + "grad_norm": 1.1642667055130005, + "learning_rate": 1.709132579119435e-05, + "loss": 3.6804, + "step": 64750 + }, + { + "epoch": 0.7279958207647327, + "grad_norm": 1.1226531267166138, + "learning_rate": 1.7089078877891498e-05, + "loss": 3.693, + "step": 64800 + }, + { + "epoch": 0.7285575459350758, + "grad_norm": 0.9891392588615417, + "learning_rate": 1.7086831964588647e-05, + "loss": 3.7193, + "step": 64850 + }, + { + "epoch": 0.729119271105419, + "grad_norm": 1.0887752771377563, + "learning_rate": 1.7084585051285796e-05, + "loss": 3.765, + "step": 64900 + }, + { + "epoch": 0.7296809962757621, + "grad_norm": 1.1410701274871826, + "learning_rate": 1.708233813798295e-05, + "loss": 3.7251, + "step": 64950 + }, + { + "epoch": 0.7302427214461052, + "grad_norm": 1.1968011856079102, + "learning_rate": 1.70800912246801e-05, + "loss": 3.6613, + "step": 65000 + }, + { + "epoch": 0.7308044466164484, + "grad_norm": 1.0664294958114624, + "learning_rate": 1.7077844311377248e-05, + "loss": 3.7821, + "step": 65050 + }, + { + "epoch": 0.7313661717867916, + "grad_norm": 1.238824725151062, + "learning_rate": 1.7075597398074397e-05, + "loss": 3.7638, + "step": 65100 + }, + { + "epoch": 0.7319278969571348, + "grad_norm": 1.074196457862854, + "learning_rate": 1.7073350484771546e-05, + "loss": 3.7331, + "step": 65150 + }, + { + "epoch": 0.7324896221274779, + "grad_norm": 1.1788239479064941, + "learning_rate": 1.7071103571468696e-05, + "loss": 3.6592, + "step": 65200 + }, + { + "epoch": 0.7330513472978211, + "grad_norm": 1.2634332180023193, + "learning_rate": 1.7068856658165845e-05, + "loss": 3.7664, + "step": 65250 + }, + { + "epoch": 0.7336130724681642, + "grad_norm": 0.9647150635719299, + "learning_rate": 1.7066609744862998e-05, + "loss": 3.7154, + "step": 65300 + }, + { + "epoch": 0.7341747976385073, + "grad_norm": 1.1323342323303223, + "learning_rate": 1.7064362831560147e-05, + "loss": 3.7684, + "step": 65350 + }, + { + "epoch": 0.7347365228088505, + "grad_norm": 1.0603891611099243, + "learning_rate": 1.7062115918257296e-05, + "loss": 3.8007, + "step": 65400 + }, + { + "epoch": 0.7352982479791937, + "grad_norm": 1.0535929203033447, + "learning_rate": 1.7059869004954446e-05, + "loss": 3.7819, + "step": 65450 + }, + { + "epoch": 0.7358599731495369, + "grad_norm": 1.1961077451705933, + "learning_rate": 1.7057622091651595e-05, + "loss": 3.8177, + "step": 65500 + }, + { + "epoch": 0.73642169831988, + "grad_norm": 1.076874017715454, + "learning_rate": 1.7055375178348744e-05, + "loss": 3.6888, + "step": 65550 + }, + { + "epoch": 0.7369834234902232, + "grad_norm": 1.0798044204711914, + "learning_rate": 1.7053128265045894e-05, + "loss": 3.8633, + "step": 65600 + }, + { + "epoch": 0.7375451486605663, + "grad_norm": 1.0618581771850586, + "learning_rate": 1.7050881351743046e-05, + "loss": 3.7193, + "step": 65650 + }, + { + "epoch": 0.7381068738309094, + "grad_norm": 1.2281302213668823, + "learning_rate": 1.7048634438440196e-05, + "loss": 3.6743, + "step": 65700 + }, + { + "epoch": 0.7386685990012527, + "grad_norm": 1.03543221950531, + "learning_rate": 1.7046387525137345e-05, + "loss": 3.8051, + "step": 65750 + }, + { + "epoch": 0.7392303241715958, + "grad_norm": 0.8722898364067078, + "learning_rate": 1.7044140611834494e-05, + "loss": 3.7792, + "step": 65800 + }, + { + "epoch": 0.739792049341939, + "grad_norm": 0.9191735982894897, + "learning_rate": 1.7041893698531644e-05, + "loss": 3.8145, + "step": 65850 + }, + { + "epoch": 0.7403537745122821, + "grad_norm": 0.9913333058357239, + "learning_rate": 1.7039646785228793e-05, + "loss": 3.7993, + "step": 65900 + }, + { + "epoch": 0.7409154996826253, + "grad_norm": 1.242875576019287, + "learning_rate": 1.7037399871925942e-05, + "loss": 3.6952, + "step": 65950 + }, + { + "epoch": 0.7414772248529684, + "grad_norm": 1.1154929399490356, + "learning_rate": 1.7035152958623095e-05, + "loss": 3.7458, + "step": 66000 + }, + { + "epoch": 0.7420389500233115, + "grad_norm": 1.2310267686843872, + "learning_rate": 1.7032906045320244e-05, + "loss": 3.7183, + "step": 66050 + }, + { + "epoch": 0.7426006751936548, + "grad_norm": 1.2243191003799438, + "learning_rate": 1.7030659132017394e-05, + "loss": 3.662, + "step": 66100 + }, + { + "epoch": 0.7431624003639979, + "grad_norm": 0.9692927002906799, + "learning_rate": 1.7028412218714543e-05, + "loss": 3.6818, + "step": 66150 + }, + { + "epoch": 0.7437241255343411, + "grad_norm": 1.0961863994598389, + "learning_rate": 1.7026165305411692e-05, + "loss": 3.7774, + "step": 66200 + }, + { + "epoch": 0.7442858507046842, + "grad_norm": 1.1476986408233643, + "learning_rate": 1.702391839210884e-05, + "loss": 3.779, + "step": 66250 + }, + { + "epoch": 0.7448475758750274, + "grad_norm": 1.0727260112762451, + "learning_rate": 1.702167147880599e-05, + "loss": 3.7399, + "step": 66300 + }, + { + "epoch": 0.7454093010453705, + "grad_norm": 1.2328996658325195, + "learning_rate": 1.7019424565503144e-05, + "loss": 3.681, + "step": 66350 + }, + { + "epoch": 0.7459710262157137, + "grad_norm": 1.0276544094085693, + "learning_rate": 1.7017177652200293e-05, + "loss": 3.7449, + "step": 66400 + }, + { + "epoch": 0.7465327513860569, + "grad_norm": 1.0378201007843018, + "learning_rate": 1.7014930738897442e-05, + "loss": 3.7523, + "step": 66450 + }, + { + "epoch": 0.7470944765564, + "grad_norm": 0.9312450885772705, + "learning_rate": 1.701268382559459e-05, + "loss": 3.7095, + "step": 66500 + }, + { + "epoch": 0.7476562017267432, + "grad_norm": 1.17537522315979, + "learning_rate": 1.701043691229174e-05, + "loss": 3.7707, + "step": 66550 + }, + { + "epoch": 0.7482179268970863, + "grad_norm": 0.9856059551239014, + "learning_rate": 1.700818999898889e-05, + "loss": 3.7125, + "step": 66600 + }, + { + "epoch": 0.7487796520674295, + "grad_norm": 1.0982396602630615, + "learning_rate": 1.700594308568604e-05, + "loss": 3.6904, + "step": 66650 + }, + { + "epoch": 0.7493413772377726, + "grad_norm": 1.0933263301849365, + "learning_rate": 1.7003696172383192e-05, + "loss": 3.8068, + "step": 66700 + }, + { + "epoch": 0.7499031024081158, + "grad_norm": 1.0702201128005981, + "learning_rate": 1.700144925908034e-05, + "loss": 3.7596, + "step": 66750 + }, + { + "epoch": 0.750464827578459, + "grad_norm": 1.0075763463974, + "learning_rate": 1.699920234577749e-05, + "loss": 3.7165, + "step": 66800 + }, + { + "epoch": 0.7510265527488021, + "grad_norm": 1.0466583967208862, + "learning_rate": 1.699695543247464e-05, + "loss": 3.727, + "step": 66850 + }, + { + "epoch": 0.7515882779191453, + "grad_norm": 1.0358221530914307, + "learning_rate": 1.699470851917179e-05, + "loss": 3.78, + "step": 66900 + }, + { + "epoch": 0.7521500030894884, + "grad_norm": 1.0360231399536133, + "learning_rate": 1.699246160586894e-05, + "loss": 3.781, + "step": 66950 + }, + { + "epoch": 0.7527117282598316, + "grad_norm": 1.3516156673431396, + "learning_rate": 1.6990214692566088e-05, + "loss": 3.7346, + "step": 67000 + }, + { + "epoch": 0.7532734534301747, + "grad_norm": 1.1824400424957275, + "learning_rate": 1.698796777926324e-05, + "loss": 3.7438, + "step": 67050 + }, + { + "epoch": 0.753835178600518, + "grad_norm": 1.1349762678146362, + "learning_rate": 1.698572086596039e-05, + "loss": 3.7134, + "step": 67100 + }, + { + "epoch": 0.7543969037708611, + "grad_norm": 1.0137215852737427, + "learning_rate": 1.698347395265754e-05, + "loss": 3.7641, + "step": 67150 + }, + { + "epoch": 0.7549586289412042, + "grad_norm": 1.0727646350860596, + "learning_rate": 1.698122703935469e-05, + "loss": 3.7501, + "step": 67200 + }, + { + "epoch": 0.7555203541115474, + "grad_norm": 1.266788125038147, + "learning_rate": 1.6979025064317895e-05, + "loss": 3.7392, + "step": 67250 + }, + { + "epoch": 0.7560820792818905, + "grad_norm": 1.1174174547195435, + "learning_rate": 1.6976778151015044e-05, + "loss": 3.7231, + "step": 67300 + }, + { + "epoch": 0.7566438044522337, + "grad_norm": 1.0874195098876953, + "learning_rate": 1.6974531237712197e-05, + "loss": 3.7797, + "step": 67350 + }, + { + "epoch": 0.7572055296225768, + "grad_norm": 1.1917657852172852, + "learning_rate": 1.6972284324409346e-05, + "loss": 3.6537, + "step": 67400 + }, + { + "epoch": 0.7577672547929201, + "grad_norm": 1.0290495157241821, + "learning_rate": 1.6970037411106495e-05, + "loss": 3.6701, + "step": 67450 + }, + { + "epoch": 0.7583289799632632, + "grad_norm": 1.1170862913131714, + "learning_rate": 1.6967790497803645e-05, + "loss": 3.7691, + "step": 67500 + }, + { + "epoch": 0.7588907051336063, + "grad_norm": 1.0585973262786865, + "learning_rate": 1.6965543584500794e-05, + "loss": 3.7705, + "step": 67550 + }, + { + "epoch": 0.7594524303039495, + "grad_norm": 1.1106197834014893, + "learning_rate": 1.6963296671197943e-05, + "loss": 3.7491, + "step": 67600 + }, + { + "epoch": 0.7600141554742926, + "grad_norm": 1.280634880065918, + "learning_rate": 1.6961049757895093e-05, + "loss": 3.7504, + "step": 67650 + }, + { + "epoch": 0.7605758806446358, + "grad_norm": 1.1722393035888672, + "learning_rate": 1.6958802844592245e-05, + "loss": 3.8211, + "step": 67700 + }, + { + "epoch": 0.7611376058149789, + "grad_norm": 1.1163361072540283, + "learning_rate": 1.6956555931289395e-05, + "loss": 3.6904, + "step": 67750 + }, + { + "epoch": 0.7616993309853222, + "grad_norm": 1.0139960050582886, + "learning_rate": 1.6954309017986544e-05, + "loss": 3.7935, + "step": 67800 + }, + { + "epoch": 0.7622610561556653, + "grad_norm": 1.1248722076416016, + "learning_rate": 1.6952062104683693e-05, + "loss": 3.6819, + "step": 67850 + }, + { + "epoch": 0.7628227813260084, + "grad_norm": 1.1061235666275024, + "learning_rate": 1.6949815191380843e-05, + "loss": 3.6705, + "step": 67900 + }, + { + "epoch": 0.7633845064963516, + "grad_norm": 1.1143214702606201, + "learning_rate": 1.6947568278077992e-05, + "loss": 3.7279, + "step": 67950 + }, + { + "epoch": 0.7639462316666947, + "grad_norm": 1.142398476600647, + "learning_rate": 1.694532136477514e-05, + "loss": 3.6923, + "step": 68000 + }, + { + "epoch": 0.7645079568370379, + "grad_norm": 1.0227819681167603, + "learning_rate": 1.6943074451472294e-05, + "loss": 3.7693, + "step": 68050 + }, + { + "epoch": 0.765069682007381, + "grad_norm": 1.331196665763855, + "learning_rate": 1.6940827538169443e-05, + "loss": 3.773, + "step": 68100 + }, + { + "epoch": 0.7656314071777243, + "grad_norm": 0.9705712199211121, + "learning_rate": 1.6938580624866593e-05, + "loss": 3.7086, + "step": 68150 + }, + { + "epoch": 0.7661931323480674, + "grad_norm": 1.227713942527771, + "learning_rate": 1.6936333711563742e-05, + "loss": 3.7594, + "step": 68200 + }, + { + "epoch": 0.7667548575184105, + "grad_norm": 1.119600534439087, + "learning_rate": 1.693408679826089e-05, + "loss": 3.8176, + "step": 68250 + }, + { + "epoch": 0.7673165826887537, + "grad_norm": 1.046105146408081, + "learning_rate": 1.693183988495804e-05, + "loss": 3.7069, + "step": 68300 + }, + { + "epoch": 0.7678783078590968, + "grad_norm": 1.4448554515838623, + "learning_rate": 1.692959297165519e-05, + "loss": 3.6717, + "step": 68350 + }, + { + "epoch": 0.76844003302944, + "grad_norm": 1.1133235692977905, + "learning_rate": 1.692734605835234e-05, + "loss": 3.709, + "step": 68400 + }, + { + "epoch": 0.7690017581997832, + "grad_norm": 1.3116734027862549, + "learning_rate": 1.6925099145049492e-05, + "loss": 3.7271, + "step": 68450 + }, + { + "epoch": 0.7695634833701264, + "grad_norm": 1.1115169525146484, + "learning_rate": 1.692285223174664e-05, + "loss": 3.7566, + "step": 68500 + }, + { + "epoch": 0.7701252085404695, + "grad_norm": 1.179801106452942, + "learning_rate": 1.692060531844379e-05, + "loss": 3.786, + "step": 68550 + }, + { + "epoch": 0.7706869337108126, + "grad_norm": 0.9484862089157104, + "learning_rate": 1.691835840514094e-05, + "loss": 3.7471, + "step": 68600 + }, + { + "epoch": 0.7712486588811558, + "grad_norm": 1.0878593921661377, + "learning_rate": 1.691611149183809e-05, + "loss": 3.7765, + "step": 68650 + }, + { + "epoch": 0.7718103840514989, + "grad_norm": 1.1715834140777588, + "learning_rate": 1.691386457853524e-05, + "loss": 3.7477, + "step": 68700 + }, + { + "epoch": 0.7723721092218421, + "grad_norm": 1.1169553995132446, + "learning_rate": 1.6911617665232388e-05, + "loss": 3.7611, + "step": 68750 + }, + { + "epoch": 0.7729338343921853, + "grad_norm": 1.0869131088256836, + "learning_rate": 1.690937075192954e-05, + "loss": 3.7331, + "step": 68800 + }, + { + "epoch": 0.7734955595625285, + "grad_norm": 1.1528838872909546, + "learning_rate": 1.690712383862669e-05, + "loss": 3.7113, + "step": 68850 + }, + { + "epoch": 0.7740572847328716, + "grad_norm": 1.2826603651046753, + "learning_rate": 1.690487692532384e-05, + "loss": 3.7586, + "step": 68900 + }, + { + "epoch": 0.7746190099032148, + "grad_norm": 1.2303800582885742, + "learning_rate": 1.690263001202099e-05, + "loss": 3.7164, + "step": 68950 + }, + { + "epoch": 0.7751807350735579, + "grad_norm": 0.8959895968437195, + "learning_rate": 1.6900428036984194e-05, + "loss": 3.7172, + "step": 69000 + }, + { + "epoch": 0.775742460243901, + "grad_norm": 0.96584552526474, + "learning_rate": 1.6898181123681344e-05, + "loss": 3.6761, + "step": 69050 + }, + { + "epoch": 0.7763041854142442, + "grad_norm": 1.37136971950531, + "learning_rate": 1.6895934210378493e-05, + "loss": 3.755, + "step": 69100 + }, + { + "epoch": 0.7768659105845874, + "grad_norm": 2.480153799057007, + "learning_rate": 1.6893687297075642e-05, + "loss": 3.7863, + "step": 69150 + }, + { + "epoch": 0.7774276357549306, + "grad_norm": 1.323038101196289, + "learning_rate": 1.689144038377279e-05, + "loss": 3.7252, + "step": 69200 + }, + { + "epoch": 0.7779893609252737, + "grad_norm": 1.1072900295257568, + "learning_rate": 1.688919347046994e-05, + "loss": 3.6942, + "step": 69250 + }, + { + "epoch": 0.7785510860956169, + "grad_norm": 1.2455112934112549, + "learning_rate": 1.6886946557167094e-05, + "loss": 3.7065, + "step": 69300 + }, + { + "epoch": 0.77911281126596, + "grad_norm": 1.305863618850708, + "learning_rate": 1.6884699643864243e-05, + "loss": 3.7533, + "step": 69350 + }, + { + "epoch": 0.7796745364363031, + "grad_norm": 1.0050400495529175, + "learning_rate": 1.6882452730561392e-05, + "loss": 3.7948, + "step": 69400 + }, + { + "epoch": 0.7802362616066464, + "grad_norm": 0.9906814098358154, + "learning_rate": 1.688020581725854e-05, + "loss": 3.7531, + "step": 69450 + }, + { + "epoch": 0.7807979867769895, + "grad_norm": 1.166979432106018, + "learning_rate": 1.687795890395569e-05, + "loss": 3.721, + "step": 69500 + }, + { + "epoch": 0.7813597119473327, + "grad_norm": 1.264424443244934, + "learning_rate": 1.687571199065284e-05, + "loss": 3.7448, + "step": 69550 + }, + { + "epoch": 0.7819214371176758, + "grad_norm": 1.1750078201293945, + "learning_rate": 1.687346507734999e-05, + "loss": 3.7879, + "step": 69600 + }, + { + "epoch": 0.782483162288019, + "grad_norm": 1.0952597856521606, + "learning_rate": 1.6871218164047142e-05, + "loss": 3.7952, + "step": 69650 + }, + { + "epoch": 0.7830448874583621, + "grad_norm": 1.0783863067626953, + "learning_rate": 1.686897125074429e-05, + "loss": 3.7517, + "step": 69700 + }, + { + "epoch": 0.7836066126287052, + "grad_norm": 1.161342740058899, + "learning_rate": 1.686672433744144e-05, + "loss": 3.753, + "step": 69750 + }, + { + "epoch": 0.7841683377990485, + "grad_norm": 1.1949788331985474, + "learning_rate": 1.686447742413859e-05, + "loss": 3.7667, + "step": 69800 + }, + { + "epoch": 0.7847300629693916, + "grad_norm": 1.2240240573883057, + "learning_rate": 1.686223051083574e-05, + "loss": 3.7418, + "step": 69850 + }, + { + "epoch": 0.7852917881397348, + "grad_norm": 1.1848735809326172, + "learning_rate": 1.685998359753289e-05, + "loss": 3.7769, + "step": 69900 + }, + { + "epoch": 0.7858535133100779, + "grad_norm": 0.9923997521400452, + "learning_rate": 1.6857736684230038e-05, + "loss": 3.7339, + "step": 69950 + }, + { + "epoch": 0.786415238480421, + "grad_norm": 1.0285292863845825, + "learning_rate": 1.685548977092719e-05, + "loss": 3.6968, + "step": 70000 + }, + { + "epoch": 0.7869769636507642, + "grad_norm": 1.2908881902694702, + "learning_rate": 1.685324285762434e-05, + "loss": 3.7946, + "step": 70050 + }, + { + "epoch": 0.7875386888211073, + "grad_norm": 1.19662344455719, + "learning_rate": 1.685099594432149e-05, + "loss": 3.7436, + "step": 70100 + }, + { + "epoch": 0.7881004139914506, + "grad_norm": 1.0358531475067139, + "learning_rate": 1.684874903101864e-05, + "loss": 3.7474, + "step": 70150 + }, + { + "epoch": 0.7886621391617937, + "grad_norm": 1.1050183773040771, + "learning_rate": 1.6846502117715788e-05, + "loss": 3.6748, + "step": 70200 + }, + { + "epoch": 0.7892238643321369, + "grad_norm": 1.2210628986358643, + "learning_rate": 1.6844255204412937e-05, + "loss": 3.6993, + "step": 70250 + }, + { + "epoch": 0.78978558950248, + "grad_norm": 1.192805528640747, + "learning_rate": 1.6842008291110087e-05, + "loss": 3.7257, + "step": 70300 + }, + { + "epoch": 0.7903473146728232, + "grad_norm": 1.0933042764663696, + "learning_rate": 1.683976137780724e-05, + "loss": 3.6633, + "step": 70350 + }, + { + "epoch": 0.7909090398431663, + "grad_norm": 1.2619465589523315, + "learning_rate": 1.683751446450439e-05, + "loss": 3.8322, + "step": 70400 + }, + { + "epoch": 0.7914707650135094, + "grad_norm": 1.2451398372650146, + "learning_rate": 1.6835267551201538e-05, + "loss": 3.7935, + "step": 70450 + }, + { + "epoch": 0.7920324901838527, + "grad_norm": 1.280688762664795, + "learning_rate": 1.6833020637898687e-05, + "loss": 3.7188, + "step": 70500 + }, + { + "epoch": 0.7925942153541958, + "grad_norm": 1.074777364730835, + "learning_rate": 1.6830773724595837e-05, + "loss": 3.8526, + "step": 70550 + }, + { + "epoch": 0.793155940524539, + "grad_norm": 1.1776268482208252, + "learning_rate": 1.6828526811292986e-05, + "loss": 3.7506, + "step": 70600 + }, + { + "epoch": 0.7937176656948821, + "grad_norm": 0.954605758190155, + "learning_rate": 1.6826279897990135e-05, + "loss": 3.749, + "step": 70650 + }, + { + "epoch": 0.7942793908652253, + "grad_norm": 1.364682674407959, + "learning_rate": 1.6824032984687285e-05, + "loss": 3.7188, + "step": 70700 + }, + { + "epoch": 0.7948411160355684, + "grad_norm": 1.1577703952789307, + "learning_rate": 1.6821786071384437e-05, + "loss": 3.6976, + "step": 70750 + }, + { + "epoch": 0.7954028412059116, + "grad_norm": 1.2413133382797241, + "learning_rate": 1.6819539158081587e-05, + "loss": 3.8732, + "step": 70800 + }, + { + "epoch": 0.7959645663762548, + "grad_norm": 0.9502367973327637, + "learning_rate": 1.6817292244778736e-05, + "loss": 3.7627, + "step": 70850 + }, + { + "epoch": 0.7965262915465979, + "grad_norm": 1.2539441585540771, + "learning_rate": 1.6815045331475885e-05, + "loss": 3.796, + "step": 70900 + }, + { + "epoch": 0.7970880167169411, + "grad_norm": 3.9363365173339844, + "learning_rate": 1.6812798418173035e-05, + "loss": 3.7875, + "step": 70950 + }, + { + "epoch": 0.7976497418872842, + "grad_norm": 1.077473759651184, + "learning_rate": 1.6810551504870184e-05, + "loss": 3.7428, + "step": 71000 + }, + { + "epoch": 0.7982114670576274, + "grad_norm": 1.1866124868392944, + "learning_rate": 1.6808304591567333e-05, + "loss": 3.7312, + "step": 71050 + }, + { + "epoch": 0.7987731922279705, + "grad_norm": 1.7925058603286743, + "learning_rate": 1.6806057678264486e-05, + "loss": 3.794, + "step": 71100 + }, + { + "epoch": 0.7993349173983137, + "grad_norm": 1.7285633087158203, + "learning_rate": 1.6803810764961635e-05, + "loss": 3.8283, + "step": 71150 + }, + { + "epoch": 0.7998966425686569, + "grad_norm": 1.1912788152694702, + "learning_rate": 1.6801563851658785e-05, + "loss": 3.7616, + "step": 71200 + }, + { + "epoch": 0.800458367739, + "grad_norm": 0.989124059677124, + "learning_rate": 1.6799316938355934e-05, + "loss": 3.7482, + "step": 71250 + }, + { + "epoch": 0.8010200929093432, + "grad_norm": 1.122393250465393, + "learning_rate": 1.6797070025053083e-05, + "loss": 3.7222, + "step": 71300 + }, + { + "epoch": 0.8015818180796863, + "grad_norm": 0.9863786101341248, + "learning_rate": 1.6794823111750233e-05, + "loss": 3.6859, + "step": 71350 + }, + { + "epoch": 0.8021435432500295, + "grad_norm": 1.1153448820114136, + "learning_rate": 1.6792576198447382e-05, + "loss": 3.7667, + "step": 71400 + }, + { + "epoch": 0.8027052684203726, + "grad_norm": 1.0489387512207031, + "learning_rate": 1.6790329285144535e-05, + "loss": 3.6631, + "step": 71450 + }, + { + "epoch": 0.8032669935907158, + "grad_norm": 1.1601120233535767, + "learning_rate": 1.6788082371841684e-05, + "loss": 3.7535, + "step": 71500 + }, + { + "epoch": 0.803828718761059, + "grad_norm": 1.1417239904403687, + "learning_rate": 1.6785835458538833e-05, + "loss": 3.6421, + "step": 71550 + }, + { + "epoch": 0.8043904439314021, + "grad_norm": 1.6279891729354858, + "learning_rate": 1.6783588545235983e-05, + "loss": 3.8245, + "step": 71600 + }, + { + "epoch": 0.8049521691017453, + "grad_norm": 0.9623803496360779, + "learning_rate": 1.6781341631933132e-05, + "loss": 3.7142, + "step": 71650 + }, + { + "epoch": 0.8055138942720884, + "grad_norm": 1.1247252225875854, + "learning_rate": 1.677909471863028e-05, + "loss": 3.7255, + "step": 71700 + }, + { + "epoch": 0.8060756194424316, + "grad_norm": 1.153218388557434, + "learning_rate": 1.677684780532743e-05, + "loss": 3.691, + "step": 71750 + }, + { + "epoch": 0.8066373446127748, + "grad_norm": 1.188510775566101, + "learning_rate": 1.6774600892024583e-05, + "loss": 3.7395, + "step": 71800 + }, + { + "epoch": 0.807199069783118, + "grad_norm": 1.0337402820587158, + "learning_rate": 1.6772353978721732e-05, + "loss": 3.7563, + "step": 71850 + }, + { + "epoch": 0.8077607949534611, + "grad_norm": 1.3426690101623535, + "learning_rate": 1.6770107065418882e-05, + "loss": 3.7542, + "step": 71900 + }, + { + "epoch": 0.8083225201238042, + "grad_norm": 1.0045791864395142, + "learning_rate": 1.676786015211603e-05, + "loss": 3.7913, + "step": 71950 + }, + { + "epoch": 0.8088842452941474, + "grad_norm": 1.1135921478271484, + "learning_rate": 1.676561323881318e-05, + "loss": 3.7338, + "step": 72000 + }, + { + "epoch": 0.8094459704644905, + "grad_norm": 0.9275521039962769, + "learning_rate": 1.676336632551033e-05, + "loss": 3.6551, + "step": 72050 + }, + { + "epoch": 0.8100076956348337, + "grad_norm": 1.1273943185806274, + "learning_rate": 1.676111941220748e-05, + "loss": 3.6895, + "step": 72100 + }, + { + "epoch": 0.8105694208051769, + "grad_norm": 1.1047978401184082, + "learning_rate": 1.6758872498904632e-05, + "loss": 3.6553, + "step": 72150 + }, + { + "epoch": 0.81113114597552, + "grad_norm": 1.4752873182296753, + "learning_rate": 1.675662558560178e-05, + "loss": 3.7694, + "step": 72200 + }, + { + "epoch": 0.8116928711458632, + "grad_norm": 1.2518274784088135, + "learning_rate": 1.675437867229893e-05, + "loss": 3.6904, + "step": 72250 + }, + { + "epoch": 0.8122545963162063, + "grad_norm": 1.0234527587890625, + "learning_rate": 1.675213175899608e-05, + "loss": 3.7434, + "step": 72300 + }, + { + "epoch": 0.8128163214865495, + "grad_norm": 1.005708932876587, + "learning_rate": 1.674988484569323e-05, + "loss": 3.7529, + "step": 72350 + }, + { + "epoch": 0.8133780466568926, + "grad_norm": 1.0410258769989014, + "learning_rate": 1.674763793239038e-05, + "loss": 3.7049, + "step": 72400 + }, + { + "epoch": 0.8139397718272358, + "grad_norm": 1.0871193408966064, + "learning_rate": 1.6745391019087528e-05, + "loss": 3.7236, + "step": 72450 + }, + { + "epoch": 0.814501496997579, + "grad_norm": 1.1557812690734863, + "learning_rate": 1.674314410578468e-05, + "loss": 3.7443, + "step": 72500 + }, + { + "epoch": 0.8150632221679222, + "grad_norm": 1.1970670223236084, + "learning_rate": 1.674089719248183e-05, + "loss": 3.7485, + "step": 72550 + }, + { + "epoch": 0.8156249473382653, + "grad_norm": 1.3610422611236572, + "learning_rate": 1.673865027917898e-05, + "loss": 3.7038, + "step": 72600 + }, + { + "epoch": 0.8161866725086084, + "grad_norm": 1.1171669960021973, + "learning_rate": 1.673640336587613e-05, + "loss": 3.7223, + "step": 72650 + }, + { + "epoch": 0.8167483976789516, + "grad_norm": 1.2198951244354248, + "learning_rate": 1.6734156452573278e-05, + "loss": 3.7455, + "step": 72700 + }, + { + "epoch": 0.8173101228492947, + "grad_norm": 1.0705602169036865, + "learning_rate": 1.6731909539270427e-05, + "loss": 3.7316, + "step": 72750 + }, + { + "epoch": 0.8178718480196379, + "grad_norm": 1.0684411525726318, + "learning_rate": 1.6729662625967576e-05, + "loss": 3.74, + "step": 72800 + }, + { + "epoch": 0.8184335731899811, + "grad_norm": 1.1701427698135376, + "learning_rate": 1.672741571266473e-05, + "loss": 3.7103, + "step": 72850 + }, + { + "epoch": 0.8189952983603243, + "grad_norm": 1.0043283700942993, + "learning_rate": 1.6725168799361878e-05, + "loss": 3.7895, + "step": 72900 + }, + { + "epoch": 0.8195570235306674, + "grad_norm": 1.2557294368743896, + "learning_rate": 1.6722921886059028e-05, + "loss": 3.7005, + "step": 72950 + }, + { + "epoch": 0.8201187487010105, + "grad_norm": 1.1772818565368652, + "learning_rate": 1.6720674972756177e-05, + "loss": 3.6787, + "step": 73000 + }, + { + "epoch": 0.8206804738713537, + "grad_norm": 1.0652531385421753, + "learning_rate": 1.6718428059453326e-05, + "loss": 3.6785, + "step": 73050 + }, + { + "epoch": 0.8212421990416968, + "grad_norm": 1.3456592559814453, + "learning_rate": 1.6716181146150476e-05, + "loss": 3.8048, + "step": 73100 + }, + { + "epoch": 0.8218039242120401, + "grad_norm": 1.377454161643982, + "learning_rate": 1.6713934232847625e-05, + "loss": 3.7803, + "step": 73150 + }, + { + "epoch": 0.8223656493823832, + "grad_norm": 1.0337969064712524, + "learning_rate": 1.6711687319544778e-05, + "loss": 3.778, + "step": 73200 + }, + { + "epoch": 0.8229273745527264, + "grad_norm": 1.2524175643920898, + "learning_rate": 1.6709440406241927e-05, + "loss": 3.6722, + "step": 73250 + }, + { + "epoch": 0.8234890997230695, + "grad_norm": 1.074506163597107, + "learning_rate": 1.6707193492939076e-05, + "loss": 3.602, + "step": 73300 + }, + { + "epoch": 0.8240508248934126, + "grad_norm": 1.0358381271362305, + "learning_rate": 1.6704946579636226e-05, + "loss": 3.7684, + "step": 73350 + }, + { + "epoch": 0.8246125500637558, + "grad_norm": 1.223362922668457, + "learning_rate": 1.6702699666333375e-05, + "loss": 3.7429, + "step": 73400 + }, + { + "epoch": 0.8251742752340989, + "grad_norm": 1.2116211652755737, + "learning_rate": 1.6700452753030524e-05, + "loss": 3.7784, + "step": 73450 + }, + { + "epoch": 0.8257360004044422, + "grad_norm": 1.03231942653656, + "learning_rate": 1.6698205839727673e-05, + "loss": 3.7228, + "step": 73500 + }, + { + "epoch": 0.8262977255747853, + "grad_norm": 1.0249361991882324, + "learning_rate": 1.6695958926424826e-05, + "loss": 3.6967, + "step": 73550 + }, + { + "epoch": 0.8268594507451285, + "grad_norm": 1.1133832931518555, + "learning_rate": 1.6693756951388032e-05, + "loss": 3.7659, + "step": 73600 + }, + { + "epoch": 0.8274211759154716, + "grad_norm": 1.1546299457550049, + "learning_rate": 1.669151003808518e-05, + "loss": 3.6815, + "step": 73650 + }, + { + "epoch": 0.8279829010858147, + "grad_norm": 1.14987313747406, + "learning_rate": 1.668926312478233e-05, + "loss": 3.7479, + "step": 73700 + }, + { + "epoch": 0.8285446262561579, + "grad_norm": 1.496004343032837, + "learning_rate": 1.668701621147948e-05, + "loss": 3.7435, + "step": 73750 + }, + { + "epoch": 0.829106351426501, + "grad_norm": 1.0407507419586182, + "learning_rate": 1.668476929817663e-05, + "loss": 3.8376, + "step": 73800 + }, + { + "epoch": 0.8296680765968443, + "grad_norm": 1.1707595586776733, + "learning_rate": 1.6682522384873782e-05, + "loss": 3.6745, + "step": 73850 + }, + { + "epoch": 0.8302298017671874, + "grad_norm": 1.1068360805511475, + "learning_rate": 1.668027547157093e-05, + "loss": 3.6911, + "step": 73900 + }, + { + "epoch": 0.8307915269375306, + "grad_norm": 0.9708052277565002, + "learning_rate": 1.667802855826808e-05, + "loss": 3.7387, + "step": 73950 + }, + { + "epoch": 0.8313532521078737, + "grad_norm": 1.101502537727356, + "learning_rate": 1.667578164496523e-05, + "loss": 3.7328, + "step": 74000 + }, + { + "epoch": 0.8319149772782168, + "grad_norm": 1.345021367073059, + "learning_rate": 1.667353473166238e-05, + "loss": 3.7497, + "step": 74050 + }, + { + "epoch": 0.83247670244856, + "grad_norm": 1.2573903799057007, + "learning_rate": 1.667128781835953e-05, + "loss": 3.8261, + "step": 74100 + }, + { + "epoch": 0.8330384276189031, + "grad_norm": 1.2525533437728882, + "learning_rate": 1.6669085843322738e-05, + "loss": 3.7412, + "step": 74150 + }, + { + "epoch": 0.8336001527892464, + "grad_norm": 1.1022635698318481, + "learning_rate": 1.6666838930019887e-05, + "loss": 3.7042, + "step": 74200 + }, + { + "epoch": 0.8341618779595895, + "grad_norm": 1.1093451976776123, + "learning_rate": 1.6664592016717037e-05, + "loss": 3.714, + "step": 74250 + }, + { + "epoch": 0.8347236031299327, + "grad_norm": 1.1967084407806396, + "learning_rate": 1.6662345103414186e-05, + "loss": 3.7316, + "step": 74300 + }, + { + "epoch": 0.8352853283002758, + "grad_norm": 1.1166884899139404, + "learning_rate": 1.6660098190111335e-05, + "loss": 3.7497, + "step": 74350 + }, + { + "epoch": 0.8358470534706189, + "grad_norm": 1.2759777307510376, + "learning_rate": 1.6657851276808485e-05, + "loss": 3.7377, + "step": 74400 + }, + { + "epoch": 0.8364087786409621, + "grad_norm": 1.3579269647598267, + "learning_rate": 1.6655604363505634e-05, + "loss": 3.7868, + "step": 74450 + }, + { + "epoch": 0.8369705038113053, + "grad_norm": 1.010847568511963, + "learning_rate": 1.6653357450202787e-05, + "loss": 3.7852, + "step": 74500 + }, + { + "epoch": 0.8375322289816485, + "grad_norm": 1.3050994873046875, + "learning_rate": 1.6651110536899936e-05, + "loss": 3.7719, + "step": 74550 + }, + { + "epoch": 0.8380939541519916, + "grad_norm": 1.1435143947601318, + "learning_rate": 1.6648863623597085e-05, + "loss": 3.7672, + "step": 74600 + }, + { + "epoch": 0.8386556793223348, + "grad_norm": 1.064864993095398, + "learning_rate": 1.6646616710294235e-05, + "loss": 3.6735, + "step": 74650 + }, + { + "epoch": 0.8392174044926779, + "grad_norm": 1.2209954261779785, + "learning_rate": 1.6644369796991384e-05, + "loss": 3.6891, + "step": 74700 + }, + { + "epoch": 0.839779129663021, + "grad_norm": 1.390276312828064, + "learning_rate": 1.6642122883688533e-05, + "loss": 3.7265, + "step": 74750 + }, + { + "epoch": 0.8403408548333642, + "grad_norm": 1.1255176067352295, + "learning_rate": 1.6639875970385682e-05, + "loss": 3.7504, + "step": 74800 + }, + { + "epoch": 0.8409025800037074, + "grad_norm": 1.1659913063049316, + "learning_rate": 1.6637629057082835e-05, + "loss": 3.7539, + "step": 74850 + }, + { + "epoch": 0.8414643051740506, + "grad_norm": 1.150168776512146, + "learning_rate": 1.6635382143779984e-05, + "loss": 3.7142, + "step": 74900 + }, + { + "epoch": 0.8420260303443937, + "grad_norm": 1.237899899482727, + "learning_rate": 1.6633135230477134e-05, + "loss": 3.7639, + "step": 74950 + }, + { + "epoch": 0.8425877555147369, + "grad_norm": 1.293717384338379, + "learning_rate": 1.6630888317174283e-05, + "loss": 3.7224, + "step": 75000 + }, + { + "epoch": 0.84314948068508, + "grad_norm": 0.9956775903701782, + "learning_rate": 1.6628641403871432e-05, + "loss": 3.7572, + "step": 75050 + }, + { + "epoch": 0.8437112058554231, + "grad_norm": 1.0118999481201172, + "learning_rate": 1.6626394490568582e-05, + "loss": 3.7923, + "step": 75100 + }, + { + "epoch": 0.8442729310257663, + "grad_norm": 1.2486600875854492, + "learning_rate": 1.662414757726573e-05, + "loss": 3.7677, + "step": 75150 + }, + { + "epoch": 0.8448346561961095, + "grad_norm": 1.0034174919128418, + "learning_rate": 1.6621900663962884e-05, + "loss": 3.7479, + "step": 75200 + }, + { + "epoch": 0.8453963813664527, + "grad_norm": 1.0463954210281372, + "learning_rate": 1.6619653750660033e-05, + "loss": 3.6765, + "step": 75250 + }, + { + "epoch": 0.8459581065367958, + "grad_norm": 1.0026686191558838, + "learning_rate": 1.6617406837357182e-05, + "loss": 3.6909, + "step": 75300 + }, + { + "epoch": 0.846519831707139, + "grad_norm": 1.2389321327209473, + "learning_rate": 1.6615159924054332e-05, + "loss": 3.8334, + "step": 75350 + }, + { + "epoch": 0.8470815568774821, + "grad_norm": 1.0024304389953613, + "learning_rate": 1.661291301075148e-05, + "loss": 3.6617, + "step": 75400 + }, + { + "epoch": 0.8476432820478252, + "grad_norm": 1.1129672527313232, + "learning_rate": 1.661066609744863e-05, + "loss": 3.7321, + "step": 75450 + }, + { + "epoch": 0.8482050072181685, + "grad_norm": 1.1076396703720093, + "learning_rate": 1.660841918414578e-05, + "loss": 3.6988, + "step": 75500 + }, + { + "epoch": 0.8487667323885116, + "grad_norm": 1.0923964977264404, + "learning_rate": 1.6606172270842932e-05, + "loss": 3.7346, + "step": 75550 + }, + { + "epoch": 0.8493284575588548, + "grad_norm": 1.1150474548339844, + "learning_rate": 1.6603925357540082e-05, + "loss": 3.7394, + "step": 75600 + }, + { + "epoch": 0.8498901827291979, + "grad_norm": 1.1458702087402344, + "learning_rate": 1.660167844423723e-05, + "loss": 3.7441, + "step": 75650 + }, + { + "epoch": 0.8504519078995411, + "grad_norm": 1.1019251346588135, + "learning_rate": 1.659943153093438e-05, + "loss": 3.7684, + "step": 75700 + }, + { + "epoch": 0.8510136330698842, + "grad_norm": 1.418558120727539, + "learning_rate": 1.659718461763153e-05, + "loss": 3.7951, + "step": 75750 + }, + { + "epoch": 0.8515753582402273, + "grad_norm": 1.058002233505249, + "learning_rate": 1.659493770432868e-05, + "loss": 3.7021, + "step": 75800 + }, + { + "epoch": 0.8521370834105706, + "grad_norm": 1.3031290769577026, + "learning_rate": 1.6592690791025828e-05, + "loss": 3.7322, + "step": 75850 + }, + { + "epoch": 0.8526988085809137, + "grad_norm": 1.255540132522583, + "learning_rate": 1.6590443877722978e-05, + "loss": 3.7647, + "step": 75900 + }, + { + "epoch": 0.8532605337512569, + "grad_norm": 1.2593520879745483, + "learning_rate": 1.658819696442013e-05, + "loss": 3.6771, + "step": 75950 + }, + { + "epoch": 0.8538222589216, + "grad_norm": 1.0368095636367798, + "learning_rate": 1.658595005111728e-05, + "loss": 3.6938, + "step": 76000 + }, + { + "epoch": 0.8543839840919432, + "grad_norm": 1.133135199546814, + "learning_rate": 1.658370313781443e-05, + "loss": 3.7244, + "step": 76050 + }, + { + "epoch": 0.8549457092622863, + "grad_norm": 1.1056394577026367, + "learning_rate": 1.6581456224511578e-05, + "loss": 3.6896, + "step": 76100 + }, + { + "epoch": 0.8555074344326294, + "grad_norm": 1.148077368736267, + "learning_rate": 1.6579209311208728e-05, + "loss": 3.7582, + "step": 76150 + }, + { + "epoch": 0.8560691596029727, + "grad_norm": 1.1761956214904785, + "learning_rate": 1.6576962397905877e-05, + "loss": 3.738, + "step": 76200 + }, + { + "epoch": 0.8566308847733158, + "grad_norm": 1.1236742734909058, + "learning_rate": 1.6574715484603026e-05, + "loss": 3.7053, + "step": 76250 + }, + { + "epoch": 0.857192609943659, + "grad_norm": 1.1957411766052246, + "learning_rate": 1.657246857130018e-05, + "loss": 3.8153, + "step": 76300 + }, + { + "epoch": 0.8577543351140021, + "grad_norm": 1.0819793939590454, + "learning_rate": 1.6570221657997328e-05, + "loss": 3.7302, + "step": 76350 + }, + { + "epoch": 0.8583160602843453, + "grad_norm": 1.1566054821014404, + "learning_rate": 1.6567974744694478e-05, + "loss": 3.7149, + "step": 76400 + }, + { + "epoch": 0.8588777854546884, + "grad_norm": 1.377386450767517, + "learning_rate": 1.6565727831391627e-05, + "loss": 3.6913, + "step": 76450 + }, + { + "epoch": 0.8594395106250315, + "grad_norm": 1.092625617980957, + "learning_rate": 1.6563480918088776e-05, + "loss": 3.6786, + "step": 76500 + }, + { + "epoch": 0.8600012357953748, + "grad_norm": 1.1253376007080078, + "learning_rate": 1.6561234004785925e-05, + "loss": 3.7482, + "step": 76550 + }, + { + "epoch": 0.8605629609657179, + "grad_norm": 1.1006630659103394, + "learning_rate": 1.6558987091483075e-05, + "loss": 3.7111, + "step": 76600 + }, + { + "epoch": 0.8611246861360611, + "grad_norm": 1.213733196258545, + "learning_rate": 1.6556740178180228e-05, + "loss": 3.6594, + "step": 76650 + }, + { + "epoch": 0.8616864113064042, + "grad_norm": 1.0177408456802368, + "learning_rate": 1.6554493264877377e-05, + "loss": 3.7734, + "step": 76700 + }, + { + "epoch": 0.8622481364767474, + "grad_norm": 1.1692383289337158, + "learning_rate": 1.6552246351574526e-05, + "loss": 3.6968, + "step": 76750 + }, + { + "epoch": 0.8628098616470905, + "grad_norm": 1.2537058591842651, + "learning_rate": 1.6549999438271675e-05, + "loss": 3.7714, + "step": 76800 + }, + { + "epoch": 0.8633715868174338, + "grad_norm": 1.1537922620773315, + "learning_rate": 1.6547752524968825e-05, + "loss": 3.7202, + "step": 76850 + }, + { + "epoch": 0.8639333119877769, + "grad_norm": 1.0910826921463013, + "learning_rate": 1.6545505611665974e-05, + "loss": 3.7392, + "step": 76900 + }, + { + "epoch": 0.86449503715812, + "grad_norm": 1.245719075202942, + "learning_rate": 1.6543258698363123e-05, + "loss": 3.7756, + "step": 76950 + }, + { + "epoch": 0.8650567623284632, + "grad_norm": 1.1607590913772583, + "learning_rate": 1.6541011785060276e-05, + "loss": 3.7172, + "step": 77000 + }, + { + "epoch": 0.8656184874988063, + "grad_norm": 1.1951005458831787, + "learning_rate": 1.6538764871757425e-05, + "loss": 3.7168, + "step": 77050 + }, + { + "epoch": 0.8661802126691495, + "grad_norm": 1.0030970573425293, + "learning_rate": 1.6536517958454575e-05, + "loss": 3.7622, + "step": 77100 + }, + { + "epoch": 0.8667419378394926, + "grad_norm": 1.0266594886779785, + "learning_rate": 1.6534271045151724e-05, + "loss": 3.7146, + "step": 77150 + }, + { + "epoch": 0.8673036630098359, + "grad_norm": 1.2694624662399292, + "learning_rate": 1.6532024131848873e-05, + "loss": 3.7644, + "step": 77200 + }, + { + "epoch": 0.867865388180179, + "grad_norm": 1.3341777324676514, + "learning_rate": 1.6529777218546023e-05, + "loss": 3.7883, + "step": 77250 + }, + { + "epoch": 0.8684271133505221, + "grad_norm": 1.3228938579559326, + "learning_rate": 1.6527530305243172e-05, + "loss": 3.7568, + "step": 77300 + }, + { + "epoch": 0.8689888385208653, + "grad_norm": 1.1019830703735352, + "learning_rate": 1.6525283391940325e-05, + "loss": 3.7358, + "step": 77350 + }, + { + "epoch": 0.8695505636912084, + "grad_norm": 1.214781641960144, + "learning_rate": 1.6523036478637474e-05, + "loss": 3.6864, + "step": 77400 + }, + { + "epoch": 0.8701122888615516, + "grad_norm": 1.3680715560913086, + "learning_rate": 1.6520789565334623e-05, + "loss": 3.7394, + "step": 77450 + }, + { + "epoch": 0.8706740140318947, + "grad_norm": 1.1794883012771606, + "learning_rate": 1.6518542652031773e-05, + "loss": 3.7502, + "step": 77500 + }, + { + "epoch": 0.871235739202238, + "grad_norm": 1.2064045667648315, + "learning_rate": 1.6516295738728922e-05, + "loss": 3.7202, + "step": 77550 + }, + { + "epoch": 0.8717974643725811, + "grad_norm": 1.033135175704956, + "learning_rate": 1.651404882542607e-05, + "loss": 3.6429, + "step": 77600 + }, + { + "epoch": 0.8723591895429242, + "grad_norm": 1.030489206314087, + "learning_rate": 1.651180191212322e-05, + "loss": 3.7168, + "step": 77650 + }, + { + "epoch": 0.8729209147132674, + "grad_norm": 1.1649249792099, + "learning_rate": 1.6509554998820373e-05, + "loss": 3.7608, + "step": 77700 + }, + { + "epoch": 0.8734826398836105, + "grad_norm": 1.425357460975647, + "learning_rate": 1.6507308085517523e-05, + "loss": 3.857, + "step": 77750 + }, + { + "epoch": 0.8740443650539537, + "grad_norm": 1.0895726680755615, + "learning_rate": 1.6505061172214672e-05, + "loss": 3.7146, + "step": 77800 + }, + { + "epoch": 0.8746060902242968, + "grad_norm": 1.160264253616333, + "learning_rate": 1.650281425891182e-05, + "loss": 3.7892, + "step": 77850 + }, + { + "epoch": 0.8751678153946401, + "grad_norm": 1.16293203830719, + "learning_rate": 1.650056734560897e-05, + "loss": 3.7371, + "step": 77900 + }, + { + "epoch": 0.8757295405649832, + "grad_norm": 1.2008315324783325, + "learning_rate": 1.649832043230612e-05, + "loss": 3.6677, + "step": 77950 + }, + { + "epoch": 0.8762912657353263, + "grad_norm": 1.1021394729614258, + "learning_rate": 1.649607351900327e-05, + "loss": 3.7296, + "step": 78000 + }, + { + "epoch": 0.8768529909056695, + "grad_norm": 1.0809792280197144, + "learning_rate": 1.6493826605700422e-05, + "loss": 3.7592, + "step": 78050 + }, + { + "epoch": 0.8774147160760126, + "grad_norm": 1.251360535621643, + "learning_rate": 1.649157969239757e-05, + "loss": 3.7237, + "step": 78100 + }, + { + "epoch": 0.8779764412463558, + "grad_norm": 1.233972430229187, + "learning_rate": 1.648933277909472e-05, + "loss": 3.6951, + "step": 78150 + }, + { + "epoch": 0.878538166416699, + "grad_norm": 1.0440303087234497, + "learning_rate": 1.648708586579187e-05, + "loss": 3.7149, + "step": 78200 + }, + { + "epoch": 0.8790998915870422, + "grad_norm": 1.253597378730774, + "learning_rate": 1.648483895248902e-05, + "loss": 3.7138, + "step": 78250 + }, + { + "epoch": 0.8796616167573853, + "grad_norm": 1.2321425676345825, + "learning_rate": 1.648259203918617e-05, + "loss": 3.714, + "step": 78300 + }, + { + "epoch": 0.8802233419277284, + "grad_norm": 1.1274776458740234, + "learning_rate": 1.6480345125883318e-05, + "loss": 3.7227, + "step": 78350 + }, + { + "epoch": 0.8807850670980716, + "grad_norm": 1.1239445209503174, + "learning_rate": 1.647809821258047e-05, + "loss": 3.6846, + "step": 78400 + }, + { + "epoch": 0.8813467922684147, + "grad_norm": 0.9932996034622192, + "learning_rate": 1.647585129927762e-05, + "loss": 3.7678, + "step": 78450 + }, + { + "epoch": 0.8819085174387579, + "grad_norm": 1.1654386520385742, + "learning_rate": 1.647360438597477e-05, + "loss": 3.7241, + "step": 78500 + }, + { + "epoch": 0.8824702426091011, + "grad_norm": 1.1453014612197876, + "learning_rate": 1.647135747267192e-05, + "loss": 3.7153, + "step": 78550 + }, + { + "epoch": 0.8830319677794443, + "grad_norm": 1.1624643802642822, + "learning_rate": 1.6469110559369068e-05, + "loss": 3.8312, + "step": 78600 + }, + { + "epoch": 0.8835936929497874, + "grad_norm": 1.084373116493225, + "learning_rate": 1.6466863646066217e-05, + "loss": 3.7449, + "step": 78650 + }, + { + "epoch": 0.8841554181201305, + "grad_norm": 0.9288941025733948, + "learning_rate": 1.6464616732763366e-05, + "loss": 3.7183, + "step": 78700 + }, + { + "epoch": 0.8847171432904737, + "grad_norm": 0.990007221698761, + "learning_rate": 1.646236981946052e-05, + "loss": 3.6809, + "step": 78750 + }, + { + "epoch": 0.8852788684608168, + "grad_norm": 1.068748116493225, + "learning_rate": 1.646012290615767e-05, + "loss": 3.8225, + "step": 78800 + }, + { + "epoch": 0.88584059363116, + "grad_norm": 1.0568158626556396, + "learning_rate": 1.6457875992854818e-05, + "loss": 3.7485, + "step": 78850 + }, + { + "epoch": 0.8864023188015032, + "grad_norm": 1.3895962238311768, + "learning_rate": 1.6455629079551967e-05, + "loss": 3.6925, + "step": 78900 + }, + { + "epoch": 0.8869640439718464, + "grad_norm": 1.1887255907058716, + "learning_rate": 1.6453382166249116e-05, + "loss": 3.7772, + "step": 78950 + }, + { + "epoch": 0.8875257691421895, + "grad_norm": 1.065321445465088, + "learning_rate": 1.6451135252946266e-05, + "loss": 3.702, + "step": 79000 + }, + { + "epoch": 0.8880874943125326, + "grad_norm": 1.207621455192566, + "learning_rate": 1.6448888339643415e-05, + "loss": 3.7194, + "step": 79050 + }, + { + "epoch": 0.8886492194828758, + "grad_norm": 2.230903148651123, + "learning_rate": 1.6446641426340564e-05, + "loss": 3.7189, + "step": 79100 + }, + { + "epoch": 0.8892109446532189, + "grad_norm": 1.2009443044662476, + "learning_rate": 1.6444394513037717e-05, + "loss": 3.7664, + "step": 79150 + }, + { + "epoch": 0.8897726698235622, + "grad_norm": 1.0513954162597656, + "learning_rate": 1.6442192538000923e-05, + "loss": 3.7869, + "step": 79200 + }, + { + "epoch": 0.8903343949939053, + "grad_norm": 1.2765376567840576, + "learning_rate": 1.6439945624698072e-05, + "loss": 3.6946, + "step": 79250 + }, + { + "epoch": 0.8908961201642485, + "grad_norm": 1.2688416242599487, + "learning_rate": 1.643769871139522e-05, + "loss": 3.7589, + "step": 79300 + }, + { + "epoch": 0.8914578453345916, + "grad_norm": 0.9216362237930298, + "learning_rate": 1.643545179809237e-05, + "loss": 3.7159, + "step": 79350 + }, + { + "epoch": 0.8920195705049347, + "grad_norm": 1.1441956758499146, + "learning_rate": 1.6433204884789524e-05, + "loss": 3.7159, + "step": 79400 + }, + { + "epoch": 0.8925812956752779, + "grad_norm": 1.182159662246704, + "learning_rate": 1.6430957971486673e-05, + "loss": 3.8107, + "step": 79450 + }, + { + "epoch": 0.893143020845621, + "grad_norm": 1.177890419960022, + "learning_rate": 1.6428711058183822e-05, + "loss": 3.7341, + "step": 79500 + }, + { + "epoch": 0.8937047460159643, + "grad_norm": 1.2310272455215454, + "learning_rate": 1.642646414488097e-05, + "loss": 3.7716, + "step": 79550 + }, + { + "epoch": 0.8942664711863074, + "grad_norm": 1.9204705953598022, + "learning_rate": 1.642421723157812e-05, + "loss": 3.7327, + "step": 79600 + }, + { + "epoch": 0.8948281963566506, + "grad_norm": 1.085312843322754, + "learning_rate": 1.642197031827527e-05, + "loss": 3.7847, + "step": 79650 + }, + { + "epoch": 0.8953899215269937, + "grad_norm": 1.079974889755249, + "learning_rate": 1.641972340497242e-05, + "loss": 3.7471, + "step": 79700 + }, + { + "epoch": 0.8959516466973368, + "grad_norm": 1.1707812547683716, + "learning_rate": 1.6417476491669572e-05, + "loss": 3.7397, + "step": 79750 + }, + { + "epoch": 0.89651337186768, + "grad_norm": 0.9927946329116821, + "learning_rate": 1.641522957836672e-05, + "loss": 3.7296, + "step": 79800 + }, + { + "epoch": 0.8970750970380231, + "grad_norm": 1.0839712619781494, + "learning_rate": 1.641298266506387e-05, + "loss": 3.7817, + "step": 79850 + }, + { + "epoch": 0.8976368222083664, + "grad_norm": 1.1929811239242554, + "learning_rate": 1.641073575176102e-05, + "loss": 3.7673, + "step": 79900 + }, + { + "epoch": 0.8981985473787095, + "grad_norm": 1.1976163387298584, + "learning_rate": 1.640848883845817e-05, + "loss": 3.737, + "step": 79950 + }, + { + "epoch": 0.8987602725490527, + "grad_norm": 1.006429672241211, + "learning_rate": 1.640624192515532e-05, + "loss": 3.7814, + "step": 80000 + }, + { + "epoch": 0.8993219977193958, + "grad_norm": 1.101015567779541, + "learning_rate": 1.6403995011852468e-05, + "loss": 3.8, + "step": 80050 + }, + { + "epoch": 0.899883722889739, + "grad_norm": 1.1051702499389648, + "learning_rate": 1.6401748098549617e-05, + "loss": 3.7417, + "step": 80100 + }, + { + "epoch": 0.9004454480600821, + "grad_norm": 1.4870634078979492, + "learning_rate": 1.639950118524677e-05, + "loss": 3.6325, + "step": 80150 + }, + { + "epoch": 0.9010071732304252, + "grad_norm": 1.0054829120635986, + "learning_rate": 1.639725427194392e-05, + "loss": 3.6966, + "step": 80200 + }, + { + "epoch": 0.9015688984007685, + "grad_norm": 1.2589412927627563, + "learning_rate": 1.639500735864107e-05, + "loss": 3.7446, + "step": 80250 + }, + { + "epoch": 0.9021306235711116, + "grad_norm": 1.1754649877548218, + "learning_rate": 1.6392760445338218e-05, + "loss": 3.6498, + "step": 80300 + }, + { + "epoch": 0.9026923487414548, + "grad_norm": 0.9390884637832642, + "learning_rate": 1.6390513532035367e-05, + "loss": 3.7443, + "step": 80350 + }, + { + "epoch": 0.9032540739117979, + "grad_norm": 0.9853662848472595, + "learning_rate": 1.6388266618732517e-05, + "loss": 3.7775, + "step": 80400 + }, + { + "epoch": 0.903815799082141, + "grad_norm": 1.0741585493087769, + "learning_rate": 1.6386019705429666e-05, + "loss": 3.7435, + "step": 80450 + }, + { + "epoch": 0.9043775242524842, + "grad_norm": 1.0210248231887817, + "learning_rate": 1.638377279212682e-05, + "loss": 3.696, + "step": 80500 + }, + { + "epoch": 0.9049392494228274, + "grad_norm": 1.2705175876617432, + "learning_rate": 1.6381525878823968e-05, + "loss": 3.7158, + "step": 80550 + }, + { + "epoch": 0.9055009745931706, + "grad_norm": 1.2171190977096558, + "learning_rate": 1.6379278965521117e-05, + "loss": 3.7859, + "step": 80600 + }, + { + "epoch": 0.9060626997635137, + "grad_norm": 0.9464622735977173, + "learning_rate": 1.6377032052218267e-05, + "loss": 3.7325, + "step": 80650 + }, + { + "epoch": 0.9066244249338569, + "grad_norm": 1.2099945545196533, + "learning_rate": 1.6374785138915416e-05, + "loss": 3.6913, + "step": 80700 + }, + { + "epoch": 0.9071861501042, + "grad_norm": 1.1484709978103638, + "learning_rate": 1.6372538225612565e-05, + "loss": 3.7389, + "step": 80750 + }, + { + "epoch": 0.9077478752745431, + "grad_norm": 1.155481219291687, + "learning_rate": 1.6370291312309715e-05, + "loss": 3.7722, + "step": 80800 + }, + { + "epoch": 0.9083096004448863, + "grad_norm": 1.1051075458526611, + "learning_rate": 1.6368044399006867e-05, + "loss": 3.7531, + "step": 80850 + }, + { + "epoch": 0.9088713256152295, + "grad_norm": 1.0399267673492432, + "learning_rate": 1.6365797485704017e-05, + "loss": 3.6485, + "step": 80900 + }, + { + "epoch": 0.9094330507855727, + "grad_norm": 1.14091157913208, + "learning_rate": 1.6363550572401166e-05, + "loss": 3.6963, + "step": 80950 + }, + { + "epoch": 0.9099947759559158, + "grad_norm": 2.3374216556549072, + "learning_rate": 1.6361303659098315e-05, + "loss": 3.7764, + "step": 81000 + }, + { + "epoch": 0.910556501126259, + "grad_norm": 1.1379660367965698, + "learning_rate": 1.635910168406152e-05, + "loss": 3.6665, + "step": 81050 + }, + { + "epoch": 0.9111182262966021, + "grad_norm": 1.128388524055481, + "learning_rate": 1.635685477075867e-05, + "loss": 3.8047, + "step": 81100 + }, + { + "epoch": 0.9116799514669452, + "grad_norm": 1.1033473014831543, + "learning_rate": 1.6354607857455823e-05, + "loss": 3.7733, + "step": 81150 + }, + { + "epoch": 0.9122416766372884, + "grad_norm": 1.1441519260406494, + "learning_rate": 1.6352360944152973e-05, + "loss": 3.7884, + "step": 81200 + }, + { + "epoch": 0.9128034018076316, + "grad_norm": 1.1598007678985596, + "learning_rate": 1.6350114030850122e-05, + "loss": 3.7307, + "step": 81250 + }, + { + "epoch": 0.9133651269779748, + "grad_norm": 1.1273629665374756, + "learning_rate": 1.634786711754727e-05, + "loss": 3.7872, + "step": 81300 + }, + { + "epoch": 0.9139268521483179, + "grad_norm": 0.9979346394538879, + "learning_rate": 1.634562020424442e-05, + "loss": 3.7104, + "step": 81350 + }, + { + "epoch": 0.9144885773186611, + "grad_norm": 0.9176252484321594, + "learning_rate": 1.634337329094157e-05, + "loss": 3.693, + "step": 81400 + }, + { + "epoch": 0.9150503024890042, + "grad_norm": 1.0392706394195557, + "learning_rate": 1.634112637763872e-05, + "loss": 3.7285, + "step": 81450 + }, + { + "epoch": 0.9156120276593473, + "grad_norm": 1.0022395849227905, + "learning_rate": 1.6338879464335872e-05, + "loss": 3.6837, + "step": 81500 + }, + { + "epoch": 0.9161737528296905, + "grad_norm": 1.1621699333190918, + "learning_rate": 1.633663255103302e-05, + "loss": 3.6972, + "step": 81550 + }, + { + "epoch": 0.9167354780000337, + "grad_norm": 1.1367850303649902, + "learning_rate": 1.633438563773017e-05, + "loss": 3.6952, + "step": 81600 + }, + { + "epoch": 0.9172972031703769, + "grad_norm": 1.3023808002471924, + "learning_rate": 1.633213872442732e-05, + "loss": 3.7635, + "step": 81650 + }, + { + "epoch": 0.91785892834072, + "grad_norm": 1.0885933637619019, + "learning_rate": 1.632989181112447e-05, + "loss": 3.7296, + "step": 81700 + }, + { + "epoch": 0.9184206535110632, + "grad_norm": 1.1319997310638428, + "learning_rate": 1.632764489782162e-05, + "loss": 3.7595, + "step": 81750 + }, + { + "epoch": 0.9189823786814063, + "grad_norm": 1.052115797996521, + "learning_rate": 1.6325397984518768e-05, + "loss": 3.7607, + "step": 81800 + }, + { + "epoch": 0.9195441038517495, + "grad_norm": 1.0686609745025635, + "learning_rate": 1.632315107121592e-05, + "loss": 3.7502, + "step": 81850 + }, + { + "epoch": 0.9201058290220927, + "grad_norm": 1.1252344846725464, + "learning_rate": 1.632090415791307e-05, + "loss": 3.6367, + "step": 81900 + }, + { + "epoch": 0.9206675541924358, + "grad_norm": 1.3658348321914673, + "learning_rate": 1.631865724461022e-05, + "loss": 3.7372, + "step": 81950 + }, + { + "epoch": 0.921229279362779, + "grad_norm": 1.0420020818710327, + "learning_rate": 1.631641033130737e-05, + "loss": 3.7205, + "step": 82000 + }, + { + "epoch": 0.9217910045331221, + "grad_norm": 1.2236835956573486, + "learning_rate": 1.6314163418004518e-05, + "loss": 3.7831, + "step": 82050 + }, + { + "epoch": 0.9223527297034653, + "grad_norm": 1.1498215198516846, + "learning_rate": 1.6311916504701667e-05, + "loss": 3.7857, + "step": 82100 + }, + { + "epoch": 0.9229144548738084, + "grad_norm": 1.0487319231033325, + "learning_rate": 1.6309669591398816e-05, + "loss": 3.7117, + "step": 82150 + }, + { + "epoch": 0.9234761800441516, + "grad_norm": 0.9913235902786255, + "learning_rate": 1.630742267809597e-05, + "loss": 3.7913, + "step": 82200 + }, + { + "epoch": 0.9240379052144948, + "grad_norm": 1.0620747804641724, + "learning_rate": 1.630517576479312e-05, + "loss": 3.7876, + "step": 82250 + }, + { + "epoch": 0.924599630384838, + "grad_norm": 1.0348880290985107, + "learning_rate": 1.6302928851490268e-05, + "loss": 3.8007, + "step": 82300 + }, + { + "epoch": 0.9251613555551811, + "grad_norm": 1.0195879936218262, + "learning_rate": 1.6300681938187417e-05, + "loss": 3.7364, + "step": 82350 + }, + { + "epoch": 0.9257230807255242, + "grad_norm": 1.4597419500350952, + "learning_rate": 1.6298435024884566e-05, + "loss": 3.7365, + "step": 82400 + }, + { + "epoch": 0.9262848058958674, + "grad_norm": 1.0478322505950928, + "learning_rate": 1.6296188111581716e-05, + "loss": 3.6908, + "step": 82450 + }, + { + "epoch": 0.9268465310662105, + "grad_norm": 1.2558234930038452, + "learning_rate": 1.6293941198278865e-05, + "loss": 3.7282, + "step": 82500 + }, + { + "epoch": 0.9274082562365537, + "grad_norm": 1.243863821029663, + "learning_rate": 1.6291694284976018e-05, + "loss": 3.7268, + "step": 82550 + }, + { + "epoch": 0.9279699814068969, + "grad_norm": 1.3925235271453857, + "learning_rate": 1.6289447371673167e-05, + "loss": 3.7183, + "step": 82600 + }, + { + "epoch": 0.92853170657724, + "grad_norm": 1.1192623376846313, + "learning_rate": 1.6287200458370316e-05, + "loss": 3.6893, + "step": 82650 + }, + { + "epoch": 0.9290934317475832, + "grad_norm": 1.2902791500091553, + "learning_rate": 1.6284953545067466e-05, + "loss": 3.764, + "step": 82700 + }, + { + "epoch": 0.9296551569179263, + "grad_norm": 1.401732325553894, + "learning_rate": 1.6282706631764615e-05, + "loss": 3.7047, + "step": 82750 + }, + { + "epoch": 0.9302168820882695, + "grad_norm": 1.1082121133804321, + "learning_rate": 1.6280459718461764e-05, + "loss": 3.7185, + "step": 82800 + }, + { + "epoch": 0.9307786072586126, + "grad_norm": 1.0722180604934692, + "learning_rate": 1.6278212805158914e-05, + "loss": 3.7609, + "step": 82850 + }, + { + "epoch": 0.9313403324289559, + "grad_norm": 1.0038912296295166, + "learning_rate": 1.6275965891856066e-05, + "loss": 3.7067, + "step": 82900 + }, + { + "epoch": 0.931902057599299, + "grad_norm": 1.0864737033843994, + "learning_rate": 1.6273718978553216e-05, + "loss": 3.7562, + "step": 82950 + }, + { + "epoch": 0.9324637827696421, + "grad_norm": 1.1519027948379517, + "learning_rate": 1.6271472065250365e-05, + "loss": 3.7136, + "step": 83000 + }, + { + "epoch": 0.9330255079399853, + "grad_norm": 1.0338419675827026, + "learning_rate": 1.6269225151947514e-05, + "loss": 3.7762, + "step": 83050 + }, + { + "epoch": 0.9335872331103284, + "grad_norm": 1.1340597867965698, + "learning_rate": 1.6266978238644664e-05, + "loss": 3.7631, + "step": 83100 + }, + { + "epoch": 0.9341489582806716, + "grad_norm": 1.0290595293045044, + "learning_rate": 1.6264731325341813e-05, + "loss": 3.6794, + "step": 83150 + }, + { + "epoch": 0.9347106834510147, + "grad_norm": 1.1557437181472778, + "learning_rate": 1.6262484412038962e-05, + "loss": 3.7217, + "step": 83200 + }, + { + "epoch": 0.935272408621358, + "grad_norm": 1.2205008268356323, + "learning_rate": 1.6260237498736115e-05, + "loss": 3.7248, + "step": 83250 + }, + { + "epoch": 0.9358341337917011, + "grad_norm": 1.0577982664108276, + "learning_rate": 1.6257990585433264e-05, + "loss": 3.7641, + "step": 83300 + }, + { + "epoch": 0.9363958589620442, + "grad_norm": 1.2078427076339722, + "learning_rate": 1.6255743672130414e-05, + "loss": 3.7042, + "step": 83350 + }, + { + "epoch": 0.9369575841323874, + "grad_norm": 1.1405781507492065, + "learning_rate": 1.6253496758827563e-05, + "loss": 3.7604, + "step": 83400 + }, + { + "epoch": 0.9375193093027305, + "grad_norm": 1.126896619796753, + "learning_rate": 1.6251249845524712e-05, + "loss": 3.7297, + "step": 83450 + }, + { + "epoch": 0.9380810344730737, + "grad_norm": 2.030507802963257, + "learning_rate": 1.624900293222186e-05, + "loss": 3.7266, + "step": 83500 + }, + { + "epoch": 0.9386427596434168, + "grad_norm": 0.8830623626708984, + "learning_rate": 1.624675601891901e-05, + "loss": 3.6499, + "step": 83550 + }, + { + "epoch": 0.9392044848137601, + "grad_norm": 1.0999085903167725, + "learning_rate": 1.6244509105616163e-05, + "loss": 3.8124, + "step": 83600 + }, + { + "epoch": 0.9397662099841032, + "grad_norm": 1.0144646167755127, + "learning_rate": 1.6242262192313313e-05, + "loss": 3.693, + "step": 83650 + }, + { + "epoch": 0.9403279351544463, + "grad_norm": 1.2930108308792114, + "learning_rate": 1.6240015279010462e-05, + "loss": 3.7773, + "step": 83700 + }, + { + "epoch": 0.9408896603247895, + "grad_norm": 1.0382893085479736, + "learning_rate": 1.623776836570761e-05, + "loss": 3.7728, + "step": 83750 + }, + { + "epoch": 0.9414513854951326, + "grad_norm": 1.1282553672790527, + "learning_rate": 1.623552145240476e-05, + "loss": 3.745, + "step": 83800 + }, + { + "epoch": 0.9420131106654758, + "grad_norm": 1.1230621337890625, + "learning_rate": 1.623327453910191e-05, + "loss": 3.6773, + "step": 83850 + }, + { + "epoch": 0.9425748358358189, + "grad_norm": 1.0702792406082153, + "learning_rate": 1.623102762579906e-05, + "loss": 3.6582, + "step": 83900 + }, + { + "epoch": 0.9431365610061622, + "grad_norm": 1.0770357847213745, + "learning_rate": 1.6228780712496212e-05, + "loss": 3.7332, + "step": 83950 + }, + { + "epoch": 0.9436982861765053, + "grad_norm": 1.0595725774765015, + "learning_rate": 1.622653379919336e-05, + "loss": 3.664, + "step": 84000 + }, + { + "epoch": 0.9442600113468484, + "grad_norm": 1.3068342208862305, + "learning_rate": 1.622428688589051e-05, + "loss": 3.7584, + "step": 84050 + }, + { + "epoch": 0.9448217365171916, + "grad_norm": 1.1843857765197754, + "learning_rate": 1.622203997258766e-05, + "loss": 3.7288, + "step": 84100 + }, + { + "epoch": 0.9453834616875347, + "grad_norm": 1.2081139087677002, + "learning_rate": 1.621979305928481e-05, + "loss": 3.7039, + "step": 84150 + }, + { + "epoch": 0.9459451868578779, + "grad_norm": 1.5447063446044922, + "learning_rate": 1.621754614598196e-05, + "loss": 3.7121, + "step": 84200 + }, + { + "epoch": 0.9465069120282211, + "grad_norm": 1.1818842887878418, + "learning_rate": 1.6215299232679108e-05, + "loss": 3.731, + "step": 84250 + }, + { + "epoch": 0.9470686371985643, + "grad_norm": 1.17287278175354, + "learning_rate": 1.621305231937626e-05, + "loss": 3.6707, + "step": 84300 + }, + { + "epoch": 0.9476303623689074, + "grad_norm": 1.114976167678833, + "learning_rate": 1.621080540607341e-05, + "loss": 3.724, + "step": 84350 + }, + { + "epoch": 0.9481920875392505, + "grad_norm": 1.4023295640945435, + "learning_rate": 1.620855849277056e-05, + "loss": 3.837, + "step": 84400 + }, + { + "epoch": 0.9487538127095937, + "grad_norm": 1.0715686082839966, + "learning_rate": 1.620631157946771e-05, + "loss": 3.7689, + "step": 84450 + }, + { + "epoch": 0.9493155378799368, + "grad_norm": 1.0490070581436157, + "learning_rate": 1.6204064666164858e-05, + "loss": 3.7099, + "step": 84500 + }, + { + "epoch": 0.94987726305028, + "grad_norm": 1.0359930992126465, + "learning_rate": 1.6201817752862007e-05, + "loss": 3.6603, + "step": 84550 + }, + { + "epoch": 0.9504389882206232, + "grad_norm": 1.259287714958191, + "learning_rate": 1.6199570839559157e-05, + "loss": 3.8236, + "step": 84600 + }, + { + "epoch": 0.9510007133909664, + "grad_norm": 1.167106032371521, + "learning_rate": 1.6197323926256306e-05, + "loss": 3.7038, + "step": 84650 + }, + { + "epoch": 0.9515624385613095, + "grad_norm": 1.2514828443527222, + "learning_rate": 1.619507701295346e-05, + "loss": 3.8219, + "step": 84700 + }, + { + "epoch": 0.9521241637316527, + "grad_norm": 1.179477572441101, + "learning_rate": 1.6192830099650608e-05, + "loss": 3.6342, + "step": 84750 + }, + { + "epoch": 0.9526858889019958, + "grad_norm": 1.272379994392395, + "learning_rate": 1.6190583186347757e-05, + "loss": 3.6968, + "step": 84800 + }, + { + "epoch": 0.9532476140723389, + "grad_norm": 1.0392554998397827, + "learning_rate": 1.6188336273044907e-05, + "loss": 3.741, + "step": 84850 + }, + { + "epoch": 0.9538093392426821, + "grad_norm": 1.019022822380066, + "learning_rate": 1.6186089359742056e-05, + "loss": 3.7318, + "step": 84900 + }, + { + "epoch": 0.9543710644130253, + "grad_norm": 1.1417900323867798, + "learning_rate": 1.6183842446439205e-05, + "loss": 3.6845, + "step": 84950 + }, + { + "epoch": 0.9549327895833685, + "grad_norm": 1.342798113822937, + "learning_rate": 1.6181595533136355e-05, + "loss": 3.7322, + "step": 85000 + }, + { + "epoch": 0.9554945147537116, + "grad_norm": 1.011507511138916, + "learning_rate": 1.6179348619833507e-05, + "loss": 3.7205, + "step": 85050 + }, + { + "epoch": 0.9560562399240548, + "grad_norm": 1.0429478883743286, + "learning_rate": 1.6177101706530657e-05, + "loss": 3.7183, + "step": 85100 + }, + { + "epoch": 0.9566179650943979, + "grad_norm": 1.3029417991638184, + "learning_rate": 1.6174854793227806e-05, + "loss": 3.7278, + "step": 85150 + }, + { + "epoch": 0.957179690264741, + "grad_norm": 1.1098041534423828, + "learning_rate": 1.6172607879924955e-05, + "loss": 3.7224, + "step": 85200 + }, + { + "epoch": 0.9577414154350842, + "grad_norm": 1.0785588026046753, + "learning_rate": 1.6170360966622104e-05, + "loss": 3.7441, + "step": 85250 + }, + { + "epoch": 0.9583031406054274, + "grad_norm": 1.1768405437469482, + "learning_rate": 1.6168114053319254e-05, + "loss": 3.7158, + "step": 85300 + }, + { + "epoch": 0.9588648657757706, + "grad_norm": 0.9101014137268066, + "learning_rate": 1.6165867140016403e-05, + "loss": 3.684, + "step": 85350 + }, + { + "epoch": 0.9594265909461137, + "grad_norm": 1.017488956451416, + "learning_rate": 1.6163620226713556e-05, + "loss": 3.7257, + "step": 85400 + }, + { + "epoch": 0.9599883161164569, + "grad_norm": 1.0188761949539185, + "learning_rate": 1.6161373313410705e-05, + "loss": 3.7245, + "step": 85450 + }, + { + "epoch": 0.9605500412868, + "grad_norm": 1.5345085859298706, + "learning_rate": 1.6159126400107854e-05, + "loss": 3.8303, + "step": 85500 + }, + { + "epoch": 0.9611117664571431, + "grad_norm": 0.9981528520584106, + "learning_rate": 1.6156879486805004e-05, + "loss": 3.7071, + "step": 85550 + }, + { + "epoch": 0.9616734916274864, + "grad_norm": 1.1540861129760742, + "learning_rate": 1.6154632573502153e-05, + "loss": 3.7012, + "step": 85600 + }, + { + "epoch": 0.9622352167978295, + "grad_norm": 1.0878418684005737, + "learning_rate": 1.6152385660199302e-05, + "loss": 3.7355, + "step": 85650 + }, + { + "epoch": 0.9627969419681727, + "grad_norm": 1.034525752067566, + "learning_rate": 1.6150138746896452e-05, + "loss": 3.7422, + "step": 85700 + }, + { + "epoch": 0.9633586671385158, + "grad_norm": 1.1636149883270264, + "learning_rate": 1.6147891833593604e-05, + "loss": 3.7703, + "step": 85750 + }, + { + "epoch": 0.963920392308859, + "grad_norm": 1.3120676279067993, + "learning_rate": 1.6145689858556807e-05, + "loss": 3.6778, + "step": 85800 + }, + { + "epoch": 0.9644821174792021, + "grad_norm": 1.1310409307479858, + "learning_rate": 1.6143442945253956e-05, + "loss": 3.7145, + "step": 85850 + }, + { + "epoch": 0.9650438426495452, + "grad_norm": 1.0255100727081299, + "learning_rate": 1.614119603195111e-05, + "loss": 3.8163, + "step": 85900 + }, + { + "epoch": 0.9656055678198885, + "grad_norm": 0.9281632304191589, + "learning_rate": 1.6138949118648258e-05, + "loss": 3.7115, + "step": 85950 + }, + { + "epoch": 0.9661672929902316, + "grad_norm": 1.1930404901504517, + "learning_rate": 1.6136702205345408e-05, + "loss": 3.8068, + "step": 86000 + }, + { + "epoch": 0.9667290181605748, + "grad_norm": 1.111885666847229, + "learning_rate": 1.6134455292042557e-05, + "loss": 3.7737, + "step": 86050 + }, + { + "epoch": 0.9672907433309179, + "grad_norm": 1.0258991718292236, + "learning_rate": 1.6132208378739706e-05, + "loss": 3.8056, + "step": 86100 + }, + { + "epoch": 0.967852468501261, + "grad_norm": 1.169977068901062, + "learning_rate": 1.6129961465436856e-05, + "loss": 3.7443, + "step": 86150 + }, + { + "epoch": 0.9684141936716042, + "grad_norm": 1.0789185762405396, + "learning_rate": 1.6127714552134005e-05, + "loss": 3.7529, + "step": 86200 + }, + { + "epoch": 0.9689759188419473, + "grad_norm": 1.2293628454208374, + "learning_rate": 1.6125467638831158e-05, + "loss": 3.722, + "step": 86250 + }, + { + "epoch": 0.9695376440122906, + "grad_norm": 1.1016701459884644, + "learning_rate": 1.6123220725528307e-05, + "loss": 3.7634, + "step": 86300 + }, + { + "epoch": 0.9700993691826337, + "grad_norm": 1.0682452917099, + "learning_rate": 1.6120973812225456e-05, + "loss": 3.6436, + "step": 86350 + }, + { + "epoch": 0.9706610943529769, + "grad_norm": 1.1063158512115479, + "learning_rate": 1.6118726898922606e-05, + "loss": 3.6602, + "step": 86400 + }, + { + "epoch": 0.97122281952332, + "grad_norm": 1.049170732498169, + "learning_rate": 1.6116479985619755e-05, + "loss": 3.7842, + "step": 86450 + }, + { + "epoch": 0.9717845446936632, + "grad_norm": 1.108414888381958, + "learning_rate": 1.611427801058296e-05, + "loss": 3.7661, + "step": 86500 + }, + { + "epoch": 0.9723462698640063, + "grad_norm": 1.3780474662780762, + "learning_rate": 1.6112031097280113e-05, + "loss": 3.7555, + "step": 86550 + }, + { + "epoch": 0.9729079950343495, + "grad_norm": 1.1071162223815918, + "learning_rate": 1.6109784183977263e-05, + "loss": 3.7328, + "step": 86600 + }, + { + "epoch": 0.9734697202046927, + "grad_norm": 1.064674973487854, + "learning_rate": 1.6107537270674412e-05, + "loss": 3.6932, + "step": 86650 + }, + { + "epoch": 0.9740314453750358, + "grad_norm": 1.2865511178970337, + "learning_rate": 1.610529035737156e-05, + "loss": 3.6596, + "step": 86700 + }, + { + "epoch": 0.974593170545379, + "grad_norm": 1.1827284097671509, + "learning_rate": 1.610304344406871e-05, + "loss": 3.7418, + "step": 86750 + }, + { + "epoch": 0.9751548957157221, + "grad_norm": 1.0069527626037598, + "learning_rate": 1.610079653076586e-05, + "loss": 3.74, + "step": 86800 + }, + { + "epoch": 0.9757166208860653, + "grad_norm": 0.9784291386604309, + "learning_rate": 1.609854961746301e-05, + "loss": 3.7252, + "step": 86850 + }, + { + "epoch": 0.9762783460564084, + "grad_norm": 1.0610507726669312, + "learning_rate": 1.6096302704160162e-05, + "loss": 3.7043, + "step": 86900 + }, + { + "epoch": 0.9768400712267516, + "grad_norm": 1.321476936340332, + "learning_rate": 1.609405579085731e-05, + "loss": 3.7311, + "step": 86950 + }, + { + "epoch": 0.9774017963970948, + "grad_norm": 1.2549364566802979, + "learning_rate": 1.609180887755446e-05, + "loss": 3.7484, + "step": 87000 + }, + { + "epoch": 0.9779635215674379, + "grad_norm": 1.0170053243637085, + "learning_rate": 1.608956196425161e-05, + "loss": 3.6022, + "step": 87050 + }, + { + "epoch": 0.9785252467377811, + "grad_norm": 1.1055805683135986, + "learning_rate": 1.608731505094876e-05, + "loss": 3.7055, + "step": 87100 + }, + { + "epoch": 0.9790869719081242, + "grad_norm": 1.103136420249939, + "learning_rate": 1.608506813764591e-05, + "loss": 3.7579, + "step": 87150 + }, + { + "epoch": 0.9796486970784674, + "grad_norm": 1.2995116710662842, + "learning_rate": 1.6082821224343058e-05, + "loss": 3.7244, + "step": 87200 + }, + { + "epoch": 0.9802104222488105, + "grad_norm": 1.2343697547912598, + "learning_rate": 1.608057431104021e-05, + "loss": 3.7127, + "step": 87250 + }, + { + "epoch": 0.9807721474191538, + "grad_norm": 1.2577475309371948, + "learning_rate": 1.607832739773736e-05, + "loss": 3.8002, + "step": 87300 + }, + { + "epoch": 0.9813338725894969, + "grad_norm": 1.1043496131896973, + "learning_rate": 1.607608048443451e-05, + "loss": 3.7542, + "step": 87350 + }, + { + "epoch": 0.98189559775984, + "grad_norm": 1.2996944189071655, + "learning_rate": 1.607383357113166e-05, + "loss": 3.7628, + "step": 87400 + }, + { + "epoch": 0.9824573229301832, + "grad_norm": 1.0178141593933105, + "learning_rate": 1.6071586657828808e-05, + "loss": 3.7823, + "step": 87450 + }, + { + "epoch": 0.9830190481005263, + "grad_norm": 1.2404489517211914, + "learning_rate": 1.6069339744525957e-05, + "loss": 3.6674, + "step": 87500 + }, + { + "epoch": 0.9835807732708695, + "grad_norm": 1.1711283922195435, + "learning_rate": 1.6067092831223107e-05, + "loss": 3.7598, + "step": 87550 + }, + { + "epoch": 0.9841424984412126, + "grad_norm": 1.113303542137146, + "learning_rate": 1.606484591792026e-05, + "loss": 3.7011, + "step": 87600 + }, + { + "epoch": 0.9847042236115559, + "grad_norm": 1.2303305864334106, + "learning_rate": 1.606259900461741e-05, + "loss": 3.7933, + "step": 87650 + }, + { + "epoch": 0.985265948781899, + "grad_norm": 1.190563440322876, + "learning_rate": 1.6060352091314558e-05, + "loss": 3.7997, + "step": 87700 + }, + { + "epoch": 0.9858276739522421, + "grad_norm": 1.2915029525756836, + "learning_rate": 1.6058105178011707e-05, + "loss": 3.6851, + "step": 87750 + }, + { + "epoch": 0.9863893991225853, + "grad_norm": 0.9873474836349487, + "learning_rate": 1.6055858264708857e-05, + "loss": 3.7403, + "step": 87800 + }, + { + "epoch": 0.9869511242929284, + "grad_norm": 1.1766434907913208, + "learning_rate": 1.6053611351406006e-05, + "loss": 3.7529, + "step": 87850 + }, + { + "epoch": 0.9875128494632716, + "grad_norm": 1.0172233581542969, + "learning_rate": 1.6051364438103155e-05, + "loss": 3.766, + "step": 87900 + }, + { + "epoch": 0.9880745746336148, + "grad_norm": 1.124140739440918, + "learning_rate": 1.6049117524800305e-05, + "loss": 3.7028, + "step": 87950 + }, + { + "epoch": 0.988636299803958, + "grad_norm": 1.1819666624069214, + "learning_rate": 1.6046870611497457e-05, + "loss": 3.729, + "step": 88000 + }, + { + "epoch": 0.9891980249743011, + "grad_norm": 1.2545924186706543, + "learning_rate": 1.6044623698194607e-05, + "loss": 3.6809, + "step": 88050 + }, + { + "epoch": 0.9897597501446442, + "grad_norm": 1.0095438957214355, + "learning_rate": 1.6042376784891756e-05, + "loss": 3.7897, + "step": 88100 + }, + { + "epoch": 0.9903214753149874, + "grad_norm": 1.0697457790374756, + "learning_rate": 1.6040129871588905e-05, + "loss": 3.7674, + "step": 88150 + }, + { + "epoch": 0.9908832004853305, + "grad_norm": 1.0836268663406372, + "learning_rate": 1.6037882958286054e-05, + "loss": 3.6559, + "step": 88200 + }, + { + "epoch": 0.9914449256556737, + "grad_norm": 1.0364786386489868, + "learning_rate": 1.6035636044983204e-05, + "loss": 3.6434, + "step": 88250 + }, + { + "epoch": 0.9920066508260169, + "grad_norm": 1.1839394569396973, + "learning_rate": 1.6033389131680353e-05, + "loss": 3.6774, + "step": 88300 + }, + { + "epoch": 0.99256837599636, + "grad_norm": 1.2167590856552124, + "learning_rate": 1.6031142218377506e-05, + "loss": 3.7092, + "step": 88350 + }, + { + "epoch": 0.9931301011667032, + "grad_norm": 1.0059152841567993, + "learning_rate": 1.6028895305074655e-05, + "loss": 3.6941, + "step": 88400 + }, + { + "epoch": 0.9936918263370463, + "grad_norm": 1.1846799850463867, + "learning_rate": 1.6026648391771804e-05, + "loss": 3.7267, + "step": 88450 + }, + { + "epoch": 0.9942535515073895, + "grad_norm": 1.0491869449615479, + "learning_rate": 1.6024401478468954e-05, + "loss": 3.7686, + "step": 88500 + }, + { + "epoch": 0.9948152766777326, + "grad_norm": 1.0813937187194824, + "learning_rate": 1.6022154565166103e-05, + "loss": 3.696, + "step": 88550 + }, + { + "epoch": 0.9953770018480758, + "grad_norm": 1.1521028280258179, + "learning_rate": 1.6019907651863252e-05, + "loss": 3.6952, + "step": 88600 + }, + { + "epoch": 0.995938727018419, + "grad_norm": 1.0822917222976685, + "learning_rate": 1.6017705676826462e-05, + "loss": 3.7029, + "step": 88650 + }, + { + "epoch": 0.9965004521887622, + "grad_norm": 1.2565577030181885, + "learning_rate": 1.601545876352361e-05, + "loss": 3.7513, + "step": 88700 + }, + { + "epoch": 0.9970621773591053, + "grad_norm": 1.0764832496643066, + "learning_rate": 1.601321185022076e-05, + "loss": 3.7125, + "step": 88750 + }, + { + "epoch": 0.9976239025294484, + "grad_norm": 1.0988057851791382, + "learning_rate": 1.601096493691791e-05, + "loss": 3.7222, + "step": 88800 + }, + { + "epoch": 0.9981856276997916, + "grad_norm": 1.1058125495910645, + "learning_rate": 1.600871802361506e-05, + "loss": 3.7659, + "step": 88850 + }, + { + "epoch": 0.9987473528701347, + "grad_norm": 2.2041337490081787, + "learning_rate": 1.6006471110312208e-05, + "loss": 3.7178, + "step": 88900 + }, + { + "epoch": 0.9993090780404779, + "grad_norm": 1.0233266353607178, + "learning_rate": 1.6004224197009358e-05, + "loss": 3.8086, + "step": 88950 + }, + { + "epoch": 0.9998708032108211, + "grad_norm": 1.1076833009719849, + "learning_rate": 1.600197728370651e-05, + "loss": 3.709, + "step": 89000 + }, + { + "epoch": 0.9999943827482966, + "eval_bleu": 5.7847, + "eval_gen_len": 21.1148, + "eval_loss": 3.5926201343536377, + "eval_runtime": 23818.0921, + "eval_samples_per_second": 14.949, + "eval_steps_per_second": 0.934, + "step": 89011 + }, + { + "epoch": 1.0004325283811641, + "grad_norm": 1.1763598918914795, + "learning_rate": 1.599973037040366e-05, + "loss": 3.6897, + "step": 89050 + }, + { + "epoch": 1.0009942535515073, + "grad_norm": 1.2548326253890991, + "learning_rate": 1.599748345710081e-05, + "loss": 3.658, + "step": 89100 + }, + { + "epoch": 1.0015559787218506, + "grad_norm": 1.1703482866287231, + "learning_rate": 1.5995236543797958e-05, + "loss": 3.6923, + "step": 89150 + }, + { + "epoch": 1.0021177038921938, + "grad_norm": 0.9730241298675537, + "learning_rate": 1.5992989630495108e-05, + "loss": 3.6708, + "step": 89200 + }, + { + "epoch": 1.002679429062537, + "grad_norm": 1.160300850868225, + "learning_rate": 1.5990742717192257e-05, + "loss": 3.7001, + "step": 89250 + }, + { + "epoch": 1.00324115423288, + "grad_norm": 1.2727965116500854, + "learning_rate": 1.5988495803889406e-05, + "loss": 3.7369, + "step": 89300 + }, + { + "epoch": 1.0038028794032232, + "grad_norm": 1.0181578397750854, + "learning_rate": 1.598624889058656e-05, + "loss": 3.7247, + "step": 89350 + }, + { + "epoch": 1.0043646045735664, + "grad_norm": 1.1146141290664673, + "learning_rate": 1.5984001977283708e-05, + "loss": 3.7194, + "step": 89400 + }, + { + "epoch": 1.0049263297439095, + "grad_norm": 1.2422945499420166, + "learning_rate": 1.5981755063980858e-05, + "loss": 3.7583, + "step": 89450 + }, + { + "epoch": 1.0054880549142526, + "grad_norm": 1.0126484632492065, + "learning_rate": 1.5979508150678007e-05, + "loss": 3.6875, + "step": 89500 + }, + { + "epoch": 1.0060497800845958, + "grad_norm": 1.0848461389541626, + "learning_rate": 1.5977261237375156e-05, + "loss": 3.7259, + "step": 89550 + }, + { + "epoch": 1.006611505254939, + "grad_norm": 1.1333413124084473, + "learning_rate": 1.5975014324072306e-05, + "loss": 3.6812, + "step": 89600 + }, + { + "epoch": 1.007173230425282, + "grad_norm": 1.2641425132751465, + "learning_rate": 1.5972767410769455e-05, + "loss": 3.6745, + "step": 89650 + }, + { + "epoch": 1.0077349555956252, + "grad_norm": 1.11570143699646, + "learning_rate": 1.5970520497466608e-05, + "loss": 3.7447, + "step": 89700 + }, + { + "epoch": 1.0082966807659683, + "grad_norm": 1.2924162149429321, + "learning_rate": 1.5968273584163757e-05, + "loss": 3.795, + "step": 89750 + }, + { + "epoch": 1.0088584059363117, + "grad_norm": 1.1690458059310913, + "learning_rate": 1.5966026670860906e-05, + "loss": 3.6647, + "step": 89800 + }, + { + "epoch": 1.0094201311066548, + "grad_norm": 1.1844213008880615, + "learning_rate": 1.5963779757558055e-05, + "loss": 3.6823, + "step": 89850 + }, + { + "epoch": 1.009981856276998, + "grad_norm": 1.0748149156570435, + "learning_rate": 1.5961532844255205e-05, + "loss": 3.7264, + "step": 89900 + }, + { + "epoch": 1.0105435814473411, + "grad_norm": 1.9574732780456543, + "learning_rate": 1.5959285930952354e-05, + "loss": 3.7581, + "step": 89950 + }, + { + "epoch": 1.0111053066176843, + "grad_norm": 1.1267497539520264, + "learning_rate": 1.5957039017649503e-05, + "loss": 3.7218, + "step": 90000 + }, + { + "epoch": 1.0116670317880274, + "grad_norm": 1.1397321224212646, + "learning_rate": 1.5954792104346656e-05, + "loss": 3.7047, + "step": 90050 + }, + { + "epoch": 1.0122287569583706, + "grad_norm": 1.2505513429641724, + "learning_rate": 1.5952545191043805e-05, + "loss": 3.6746, + "step": 90100 + }, + { + "epoch": 1.0127904821287137, + "grad_norm": 1.0324023962020874, + "learning_rate": 1.5950298277740955e-05, + "loss": 3.8013, + "step": 90150 + }, + { + "epoch": 1.0133522072990568, + "grad_norm": 1.238370418548584, + "learning_rate": 1.5948051364438104e-05, + "loss": 3.764, + "step": 90200 + }, + { + "epoch": 1.0139139324694, + "grad_norm": 1.2217398881912231, + "learning_rate": 1.5945804451135253e-05, + "loss": 3.7278, + "step": 90250 + }, + { + "epoch": 1.0144756576397431, + "grad_norm": 1.2465476989746094, + "learning_rate": 1.5943557537832403e-05, + "loss": 3.6837, + "step": 90300 + }, + { + "epoch": 1.0150373828100863, + "grad_norm": 1.1466426849365234, + "learning_rate": 1.5941310624529552e-05, + "loss": 3.8206, + "step": 90350 + }, + { + "epoch": 1.0155991079804294, + "grad_norm": 1.092600703239441, + "learning_rate": 1.5939063711226705e-05, + "loss": 3.6569, + "step": 90400 + }, + { + "epoch": 1.0161608331507725, + "grad_norm": 1.0314606428146362, + "learning_rate": 1.5936816797923854e-05, + "loss": 3.7336, + "step": 90450 + }, + { + "epoch": 1.016722558321116, + "grad_norm": 1.1504980325698853, + "learning_rate": 1.5934569884621003e-05, + "loss": 3.6885, + "step": 90500 + }, + { + "epoch": 1.017284283491459, + "grad_norm": 1.2484686374664307, + "learning_rate": 1.5932322971318153e-05, + "loss": 3.7622, + "step": 90550 + }, + { + "epoch": 1.0178460086618022, + "grad_norm": 1.1624548435211182, + "learning_rate": 1.5930076058015302e-05, + "loss": 3.7382, + "step": 90600 + }, + { + "epoch": 1.0184077338321453, + "grad_norm": 1.129113793373108, + "learning_rate": 1.592782914471245e-05, + "loss": 3.7568, + "step": 90650 + }, + { + "epoch": 1.0189694590024885, + "grad_norm": 0.9480738043785095, + "learning_rate": 1.59255822314096e-05, + "loss": 3.7107, + "step": 90700 + }, + { + "epoch": 1.0195311841728316, + "grad_norm": 0.9821470379829407, + "learning_rate": 1.5923335318106753e-05, + "loss": 3.7356, + "step": 90750 + }, + { + "epoch": 1.0200929093431748, + "grad_norm": 1.2413480281829834, + "learning_rate": 1.5921088404803903e-05, + "loss": 3.7106, + "step": 90800 + }, + { + "epoch": 1.020654634513518, + "grad_norm": 0.9626780152320862, + "learning_rate": 1.5918841491501052e-05, + "loss": 3.728, + "step": 90850 + }, + { + "epoch": 1.021216359683861, + "grad_norm": 1.038010835647583, + "learning_rate": 1.59165945781982e-05, + "loss": 3.7322, + "step": 90900 + }, + { + "epoch": 1.0217780848542042, + "grad_norm": 1.1623151302337646, + "learning_rate": 1.591434766489535e-05, + "loss": 3.7519, + "step": 90950 + }, + { + "epoch": 1.0223398100245473, + "grad_norm": 1.1785417795181274, + "learning_rate": 1.59121007515925e-05, + "loss": 3.7059, + "step": 91000 + }, + { + "epoch": 1.0229015351948905, + "grad_norm": 1.0354048013687134, + "learning_rate": 1.590985383828965e-05, + "loss": 3.6771, + "step": 91050 + }, + { + "epoch": 1.0234632603652336, + "grad_norm": 1.1689685583114624, + "learning_rate": 1.5907606924986802e-05, + "loss": 3.7238, + "step": 91100 + }, + { + "epoch": 1.024024985535577, + "grad_norm": 1.102995753288269, + "learning_rate": 1.590536001168395e-05, + "loss": 3.721, + "step": 91150 + }, + { + "epoch": 1.0245867107059201, + "grad_norm": 1.0524019002914429, + "learning_rate": 1.59031130983811e-05, + "loss": 3.6991, + "step": 91200 + }, + { + "epoch": 1.0251484358762633, + "grad_norm": 1.2444039583206177, + "learning_rate": 1.590086618507825e-05, + "loss": 3.7336, + "step": 91250 + }, + { + "epoch": 1.0257101610466064, + "grad_norm": 1.0841864347457886, + "learning_rate": 1.58986192717754e-05, + "loss": 3.8199, + "step": 91300 + }, + { + "epoch": 1.0262718862169495, + "grad_norm": 1.1749396324157715, + "learning_rate": 1.589637235847255e-05, + "loss": 3.6557, + "step": 91350 + }, + { + "epoch": 1.0268336113872927, + "grad_norm": 1.1323601007461548, + "learning_rate": 1.5894125445169698e-05, + "loss": 3.7425, + "step": 91400 + }, + { + "epoch": 1.0273953365576358, + "grad_norm": 1.066879153251648, + "learning_rate": 1.589187853186685e-05, + "loss": 3.6942, + "step": 91450 + }, + { + "epoch": 1.027957061727979, + "grad_norm": 0.929542601108551, + "learning_rate": 1.5889631618564e-05, + "loss": 3.6912, + "step": 91500 + }, + { + "epoch": 1.028518786898322, + "grad_norm": 1.0617517232894897, + "learning_rate": 1.588738470526115e-05, + "loss": 3.6954, + "step": 91550 + }, + { + "epoch": 1.0290805120686652, + "grad_norm": 0.9447341561317444, + "learning_rate": 1.58851377919583e-05, + "loss": 3.685, + "step": 91600 + }, + { + "epoch": 1.0296422372390084, + "grad_norm": 1.0214871168136597, + "learning_rate": 1.5882890878655448e-05, + "loss": 3.581, + "step": 91650 + }, + { + "epoch": 1.0302039624093515, + "grad_norm": 1.0436146259307861, + "learning_rate": 1.5880643965352597e-05, + "loss": 3.7354, + "step": 91700 + }, + { + "epoch": 1.0307656875796947, + "grad_norm": 1.1642152070999146, + "learning_rate": 1.5878397052049746e-05, + "loss": 3.7397, + "step": 91750 + }, + { + "epoch": 1.0313274127500378, + "grad_norm": 1.1631594896316528, + "learning_rate": 1.58761501387469e-05, + "loss": 3.7611, + "step": 91800 + }, + { + "epoch": 1.0318891379203812, + "grad_norm": 1.2050690650939941, + "learning_rate": 1.587390322544405e-05, + "loss": 3.6778, + "step": 91850 + }, + { + "epoch": 1.0324508630907243, + "grad_norm": 1.268229365348816, + "learning_rate": 1.5871656312141198e-05, + "loss": 3.6934, + "step": 91900 + }, + { + "epoch": 1.0330125882610675, + "grad_norm": 1.3370764255523682, + "learning_rate": 1.5869409398838347e-05, + "loss": 3.7889, + "step": 91950 + }, + { + "epoch": 1.0335743134314106, + "grad_norm": 1.07756769657135, + "learning_rate": 1.5867162485535496e-05, + "loss": 3.6994, + "step": 92000 + }, + { + "epoch": 1.0341360386017537, + "grad_norm": 1.1191749572753906, + "learning_rate": 1.5864915572232646e-05, + "loss": 3.7082, + "step": 92050 + }, + { + "epoch": 1.0346977637720969, + "grad_norm": 1.1558022499084473, + "learning_rate": 1.5862668658929795e-05, + "loss": 3.729, + "step": 92100 + }, + { + "epoch": 1.03525948894244, + "grad_norm": 1.1125577688217163, + "learning_rate": 1.5860421745626944e-05, + "loss": 3.7722, + "step": 92150 + }, + { + "epoch": 1.0358212141127832, + "grad_norm": 1.2622685432434082, + "learning_rate": 1.5858174832324097e-05, + "loss": 3.7259, + "step": 92200 + }, + { + "epoch": 1.0363829392831263, + "grad_norm": 1.3767738342285156, + "learning_rate": 1.5855927919021246e-05, + "loss": 3.7607, + "step": 92250 + }, + { + "epoch": 1.0369446644534694, + "grad_norm": 0.9928566813468933, + "learning_rate": 1.5853681005718396e-05, + "loss": 3.7339, + "step": 92300 + }, + { + "epoch": 1.0375063896238126, + "grad_norm": 1.0747640132904053, + "learning_rate": 1.5851434092415545e-05, + "loss": 3.8035, + "step": 92350 + }, + { + "epoch": 1.0380681147941557, + "grad_norm": 1.2698166370391846, + "learning_rate": 1.5849187179112694e-05, + "loss": 3.7692, + "step": 92400 + }, + { + "epoch": 1.0386298399644989, + "grad_norm": 1.2842738628387451, + "learning_rate": 1.5846940265809844e-05, + "loss": 3.8173, + "step": 92450 + }, + { + "epoch": 1.0391915651348422, + "grad_norm": 1.1700290441513062, + "learning_rate": 1.5844693352506993e-05, + "loss": 3.7536, + "step": 92500 + }, + { + "epoch": 1.0397532903051854, + "grad_norm": 1.02366304397583, + "learning_rate": 1.5842446439204146e-05, + "loss": 3.7264, + "step": 92550 + }, + { + "epoch": 1.0403150154755285, + "grad_norm": 1.1328957080841064, + "learning_rate": 1.5840199525901295e-05, + "loss": 3.6595, + "step": 92600 + }, + { + "epoch": 1.0408767406458717, + "grad_norm": 1.1032906770706177, + "learning_rate": 1.5837952612598444e-05, + "loss": 3.7593, + "step": 92650 + }, + { + "epoch": 1.0414384658162148, + "grad_norm": 1.2312544584274292, + "learning_rate": 1.5835705699295594e-05, + "loss": 3.7379, + "step": 92700 + }, + { + "epoch": 1.042000190986558, + "grad_norm": 1.1248546838760376, + "learning_rate": 1.5833458785992743e-05, + "loss": 3.669, + "step": 92750 + }, + { + "epoch": 1.042561916156901, + "grad_norm": 1.175838828086853, + "learning_rate": 1.5831211872689892e-05, + "loss": 3.7091, + "step": 92800 + }, + { + "epoch": 1.0431236413272442, + "grad_norm": 1.394729733467102, + "learning_rate": 1.582896495938704e-05, + "loss": 3.7474, + "step": 92850 + }, + { + "epoch": 1.0436853664975874, + "grad_norm": 1.2979602813720703, + "learning_rate": 1.5826718046084194e-05, + "loss": 3.6861, + "step": 92900 + }, + { + "epoch": 1.0442470916679305, + "grad_norm": 1.091827392578125, + "learning_rate": 1.5824471132781344e-05, + "loss": 3.6909, + "step": 92950 + }, + { + "epoch": 1.0448088168382736, + "grad_norm": 1.2266827821731567, + "learning_rate": 1.5822224219478493e-05, + "loss": 3.7223, + "step": 93000 + }, + { + "epoch": 1.0453705420086168, + "grad_norm": 1.0776731967926025, + "learning_rate": 1.5819977306175642e-05, + "loss": 3.7334, + "step": 93050 + }, + { + "epoch": 1.04593226717896, + "grad_norm": 1.119864583015442, + "learning_rate": 1.581773039287279e-05, + "loss": 3.7641, + "step": 93100 + }, + { + "epoch": 1.046493992349303, + "grad_norm": 1.124421238899231, + "learning_rate": 1.581548347956994e-05, + "loss": 3.6658, + "step": 93150 + }, + { + "epoch": 1.0470557175196464, + "grad_norm": 1.103226900100708, + "learning_rate": 1.581323656626709e-05, + "loss": 3.6809, + "step": 93200 + }, + { + "epoch": 1.0476174426899896, + "grad_norm": 0.9896043539047241, + "learning_rate": 1.5810989652964243e-05, + "loss": 3.6122, + "step": 93250 + }, + { + "epoch": 1.0481791678603327, + "grad_norm": 1.226040244102478, + "learning_rate": 1.5808742739661392e-05, + "loss": 3.6669, + "step": 93300 + }, + { + "epoch": 1.0487408930306759, + "grad_norm": 1.0165162086486816, + "learning_rate": 1.580649582635854e-05, + "loss": 3.6696, + "step": 93350 + }, + { + "epoch": 1.049302618201019, + "grad_norm": 1.2807351350784302, + "learning_rate": 1.580424891305569e-05, + "loss": 3.7671, + "step": 93400 + }, + { + "epoch": 1.0498643433713621, + "grad_norm": 1.1978585720062256, + "learning_rate": 1.580200199975284e-05, + "loss": 3.77, + "step": 93450 + }, + { + "epoch": 1.0504260685417053, + "grad_norm": 1.1764203310012817, + "learning_rate": 1.579975508644999e-05, + "loss": 3.8034, + "step": 93500 + }, + { + "epoch": 1.0509877937120484, + "grad_norm": 1.1233216524124146, + "learning_rate": 1.579750817314714e-05, + "loss": 3.7798, + "step": 93550 + }, + { + "epoch": 1.0515495188823916, + "grad_norm": 1.6650383472442627, + "learning_rate": 1.579526125984429e-05, + "loss": 3.7086, + "step": 93600 + }, + { + "epoch": 1.0521112440527347, + "grad_norm": 1.1246583461761475, + "learning_rate": 1.579301434654144e-05, + "loss": 3.8124, + "step": 93650 + }, + { + "epoch": 1.0526729692230778, + "grad_norm": 1.054193377494812, + "learning_rate": 1.579076743323859e-05, + "loss": 3.7576, + "step": 93700 + }, + { + "epoch": 1.053234694393421, + "grad_norm": 1.1520317792892456, + "learning_rate": 1.578852051993574e-05, + "loss": 3.7045, + "step": 93750 + }, + { + "epoch": 1.0537964195637641, + "grad_norm": 1.1220756769180298, + "learning_rate": 1.578627360663289e-05, + "loss": 3.7785, + "step": 93800 + }, + { + "epoch": 1.0543581447341075, + "grad_norm": 1.1155728101730347, + "learning_rate": 1.5784026693330038e-05, + "loss": 3.7105, + "step": 93850 + }, + { + "epoch": 1.0549198699044506, + "grad_norm": 1.0775452852249146, + "learning_rate": 1.5781824718293247e-05, + "loss": 3.7609, + "step": 93900 + }, + { + "epoch": 1.0554815950747938, + "grad_norm": 1.1409608125686646, + "learning_rate": 1.5779577804990397e-05, + "loss": 3.7306, + "step": 93950 + }, + { + "epoch": 1.056043320245137, + "grad_norm": 1.1192364692687988, + "learning_rate": 1.5777330891687546e-05, + "loss": 3.6886, + "step": 94000 + }, + { + "epoch": 1.05660504541548, + "grad_norm": 1.3993706703186035, + "learning_rate": 1.5775083978384695e-05, + "loss": 3.6814, + "step": 94050 + }, + { + "epoch": 1.0571667705858232, + "grad_norm": 1.189325213432312, + "learning_rate": 1.5772837065081845e-05, + "loss": 3.7275, + "step": 94100 + }, + { + "epoch": 1.0577284957561663, + "grad_norm": 1.2303613424301147, + "learning_rate": 1.5770590151778994e-05, + "loss": 3.7163, + "step": 94150 + }, + { + "epoch": 1.0582902209265095, + "grad_norm": 1.0467129945755005, + "learning_rate": 1.5768343238476143e-05, + "loss": 3.6759, + "step": 94200 + }, + { + "epoch": 1.0588519460968526, + "grad_norm": 1.060932993888855, + "learning_rate": 1.5766096325173296e-05, + "loss": 3.6594, + "step": 94250 + }, + { + "epoch": 1.0594136712671958, + "grad_norm": 1.2679494619369507, + "learning_rate": 1.5763849411870445e-05, + "loss": 3.7668, + "step": 94300 + }, + { + "epoch": 1.059975396437539, + "grad_norm": 1.0520027875900269, + "learning_rate": 1.5761602498567595e-05, + "loss": 3.6711, + "step": 94350 + }, + { + "epoch": 1.060537121607882, + "grad_norm": 1.1975017786026, + "learning_rate": 1.5759355585264744e-05, + "loss": 3.7632, + "step": 94400 + }, + { + "epoch": 1.0610988467782252, + "grad_norm": 1.310758352279663, + "learning_rate": 1.5757108671961893e-05, + "loss": 3.736, + "step": 94450 + }, + { + "epoch": 1.0616605719485683, + "grad_norm": 1.2200254201889038, + "learning_rate": 1.5754861758659043e-05, + "loss": 3.766, + "step": 94500 + }, + { + "epoch": 1.0622222971189117, + "grad_norm": 1.3379442691802979, + "learning_rate": 1.5752614845356192e-05, + "loss": 3.7193, + "step": 94550 + }, + { + "epoch": 1.0627840222892548, + "grad_norm": 1.1302127838134766, + "learning_rate": 1.5750367932053345e-05, + "loss": 3.7355, + "step": 94600 + }, + { + "epoch": 1.063345747459598, + "grad_norm": 1.0796693563461304, + "learning_rate": 1.5748121018750494e-05, + "loss": 3.7235, + "step": 94650 + }, + { + "epoch": 1.0639074726299411, + "grad_norm": 1.236109972000122, + "learning_rate": 1.5745874105447643e-05, + "loss": 3.7059, + "step": 94700 + }, + { + "epoch": 1.0644691978002843, + "grad_norm": 1.2734140157699585, + "learning_rate": 1.5743627192144793e-05, + "loss": 3.7032, + "step": 94750 + }, + { + "epoch": 1.0650309229706274, + "grad_norm": 1.3345608711242676, + "learning_rate": 1.5741380278841942e-05, + "loss": 3.7339, + "step": 94800 + }, + { + "epoch": 1.0655926481409705, + "grad_norm": 1.1744345426559448, + "learning_rate": 1.573913336553909e-05, + "loss": 3.6792, + "step": 94850 + }, + { + "epoch": 1.0661543733113137, + "grad_norm": 1.0625845193862915, + "learning_rate": 1.573688645223624e-05, + "loss": 3.718, + "step": 94900 + }, + { + "epoch": 1.0667160984816568, + "grad_norm": 1.1035292148590088, + "learning_rate": 1.5734639538933393e-05, + "loss": 3.731, + "step": 94950 + }, + { + "epoch": 1.067277823652, + "grad_norm": 1.2014189958572388, + "learning_rate": 1.5732392625630543e-05, + "loss": 3.686, + "step": 95000 + }, + { + "epoch": 1.067839548822343, + "grad_norm": 1.1344666481018066, + "learning_rate": 1.5730145712327692e-05, + "loss": 3.7133, + "step": 95050 + }, + { + "epoch": 1.0684012739926863, + "grad_norm": 1.1860787868499756, + "learning_rate": 1.572789879902484e-05, + "loss": 3.6945, + "step": 95100 + }, + { + "epoch": 1.0689629991630294, + "grad_norm": 1.129529356956482, + "learning_rate": 1.572565188572199e-05, + "loss": 3.6903, + "step": 95150 + }, + { + "epoch": 1.0695247243333728, + "grad_norm": 0.9301499128341675, + "learning_rate": 1.572340497241914e-05, + "loss": 3.6702, + "step": 95200 + }, + { + "epoch": 1.070086449503716, + "grad_norm": 1.171042799949646, + "learning_rate": 1.572115805911629e-05, + "loss": 3.6832, + "step": 95250 + }, + { + "epoch": 1.070648174674059, + "grad_norm": 1.0683860778808594, + "learning_rate": 1.5718911145813442e-05, + "loss": 3.7034, + "step": 95300 + }, + { + "epoch": 1.0712098998444022, + "grad_norm": 1.2346833944320679, + "learning_rate": 1.571666423251059e-05, + "loss": 3.7396, + "step": 95350 + }, + { + "epoch": 1.0717716250147453, + "grad_norm": 1.0209951400756836, + "learning_rate": 1.571441731920774e-05, + "loss": 3.7822, + "step": 95400 + }, + { + "epoch": 1.0723333501850885, + "grad_norm": 1.2905466556549072, + "learning_rate": 1.571217040590489e-05, + "loss": 3.7594, + "step": 95450 + }, + { + "epoch": 1.0728950753554316, + "grad_norm": 1.3073861598968506, + "learning_rate": 1.570992349260204e-05, + "loss": 3.7016, + "step": 95500 + }, + { + "epoch": 1.0734568005257747, + "grad_norm": 1.2471473217010498, + "learning_rate": 1.570767657929919e-05, + "loss": 3.6912, + "step": 95550 + }, + { + "epoch": 1.0740185256961179, + "grad_norm": 1.2150790691375732, + "learning_rate": 1.5705429665996338e-05, + "loss": 3.7118, + "step": 95600 + }, + { + "epoch": 1.074580250866461, + "grad_norm": 1.1582289934158325, + "learning_rate": 1.570318275269349e-05, + "loss": 3.7255, + "step": 95650 + }, + { + "epoch": 1.0751419760368042, + "grad_norm": 1.164430022239685, + "learning_rate": 1.570093583939064e-05, + "loss": 3.7233, + "step": 95700 + }, + { + "epoch": 1.0757037012071473, + "grad_norm": 1.222058892250061, + "learning_rate": 1.569868892608779e-05, + "loss": 3.774, + "step": 95750 + }, + { + "epoch": 1.0762654263774905, + "grad_norm": 1.12148118019104, + "learning_rate": 1.569644201278494e-05, + "loss": 3.7889, + "step": 95800 + }, + { + "epoch": 1.0768271515478336, + "grad_norm": 0.9750103950500488, + "learning_rate": 1.5694195099482088e-05, + "loss": 3.728, + "step": 95850 + }, + { + "epoch": 1.077388876718177, + "grad_norm": 1.0557671785354614, + "learning_rate": 1.5691948186179237e-05, + "loss": 3.7635, + "step": 95900 + }, + { + "epoch": 1.07795060188852, + "grad_norm": 1.0505820512771606, + "learning_rate": 1.5689701272876386e-05, + "loss": 3.7686, + "step": 95950 + }, + { + "epoch": 1.0785123270588632, + "grad_norm": 1.1918795108795166, + "learning_rate": 1.568745435957354e-05, + "loss": 3.6623, + "step": 96000 + }, + { + "epoch": 1.0790740522292064, + "grad_norm": 1.1036181449890137, + "learning_rate": 1.568520744627069e-05, + "loss": 3.624, + "step": 96050 + }, + { + "epoch": 1.0796357773995495, + "grad_norm": 1.1361887454986572, + "learning_rate": 1.5683005471233894e-05, + "loss": 3.7652, + "step": 96100 + }, + { + "epoch": 1.0801975025698927, + "grad_norm": 1.04814612865448, + "learning_rate": 1.5680758557931044e-05, + "loss": 3.6505, + "step": 96150 + }, + { + "epoch": 1.0807592277402358, + "grad_norm": 1.2747437953948975, + "learning_rate": 1.5678511644628193e-05, + "loss": 3.6925, + "step": 96200 + }, + { + "epoch": 1.081320952910579, + "grad_norm": 0.9882274866104126, + "learning_rate": 1.5676264731325342e-05, + "loss": 3.7606, + "step": 96250 + }, + { + "epoch": 1.081882678080922, + "grad_norm": 1.2443149089813232, + "learning_rate": 1.5674017818022495e-05, + "loss": 3.7596, + "step": 96300 + }, + { + "epoch": 1.0824444032512652, + "grad_norm": 1.0911964178085327, + "learning_rate": 1.5671770904719644e-05, + "loss": 3.6851, + "step": 96350 + }, + { + "epoch": 1.0830061284216084, + "grad_norm": 1.175838828086853, + "learning_rate": 1.5669523991416794e-05, + "loss": 3.7198, + "step": 96400 + }, + { + "epoch": 1.0835678535919515, + "grad_norm": 1.2125855684280396, + "learning_rate": 1.5667277078113943e-05, + "loss": 3.7662, + "step": 96450 + }, + { + "epoch": 1.0841295787622949, + "grad_norm": 1.248186707496643, + "learning_rate": 1.5665030164811092e-05, + "loss": 3.6982, + "step": 96500 + }, + { + "epoch": 1.084691303932638, + "grad_norm": 1.143099308013916, + "learning_rate": 1.566278325150824e-05, + "loss": 3.8005, + "step": 96550 + }, + { + "epoch": 1.0852530291029812, + "grad_norm": 1.080655574798584, + "learning_rate": 1.566053633820539e-05, + "loss": 3.7345, + "step": 96600 + }, + { + "epoch": 1.0858147542733243, + "grad_norm": 1.2310138940811157, + "learning_rate": 1.5658289424902544e-05, + "loss": 3.6589, + "step": 96650 + }, + { + "epoch": 1.0863764794436674, + "grad_norm": 1.3286763429641724, + "learning_rate": 1.5656042511599693e-05, + "loss": 3.7821, + "step": 96700 + }, + { + "epoch": 1.0869382046140106, + "grad_norm": 1.1189380884170532, + "learning_rate": 1.5653795598296842e-05, + "loss": 3.6979, + "step": 96750 + }, + { + "epoch": 1.0874999297843537, + "grad_norm": 0.9450409412384033, + "learning_rate": 1.565154868499399e-05, + "loss": 3.7887, + "step": 96800 + }, + { + "epoch": 1.0880616549546969, + "grad_norm": 1.258392572402954, + "learning_rate": 1.564930177169114e-05, + "loss": 3.7496, + "step": 96850 + }, + { + "epoch": 1.08862338012504, + "grad_norm": 1.073094367980957, + "learning_rate": 1.564705485838829e-05, + "loss": 3.7592, + "step": 96900 + }, + { + "epoch": 1.0891851052953831, + "grad_norm": 1.1871005296707153, + "learning_rate": 1.564480794508544e-05, + "loss": 3.6373, + "step": 96950 + }, + { + "epoch": 1.0897468304657263, + "grad_norm": 1.2509921789169312, + "learning_rate": 1.5642561031782592e-05, + "loss": 3.8089, + "step": 97000 + }, + { + "epoch": 1.0903085556360694, + "grad_norm": 1.1163924932479858, + "learning_rate": 1.564031411847974e-05, + "loss": 3.7584, + "step": 97050 + }, + { + "epoch": 1.0908702808064126, + "grad_norm": 1.2207306623458862, + "learning_rate": 1.563806720517689e-05, + "loss": 3.7147, + "step": 97100 + }, + { + "epoch": 1.0914320059767557, + "grad_norm": 1.2841033935546875, + "learning_rate": 1.563582029187404e-05, + "loss": 3.8064, + "step": 97150 + }, + { + "epoch": 1.0919937311470989, + "grad_norm": 1.4015285968780518, + "learning_rate": 1.563357337857119e-05, + "loss": 3.6844, + "step": 97200 + }, + { + "epoch": 1.0925554563174422, + "grad_norm": 1.2208726406097412, + "learning_rate": 1.563132646526834e-05, + "loss": 3.7054, + "step": 97250 + }, + { + "epoch": 1.0931171814877854, + "grad_norm": 1.1583671569824219, + "learning_rate": 1.5629079551965488e-05, + "loss": 3.644, + "step": 97300 + }, + { + "epoch": 1.0936789066581285, + "grad_norm": 1.1001847982406616, + "learning_rate": 1.5626832638662637e-05, + "loss": 3.8188, + "step": 97350 + }, + { + "epoch": 1.0942406318284716, + "grad_norm": 1.0487697124481201, + "learning_rate": 1.562458572535979e-05, + "loss": 3.7028, + "step": 97400 + }, + { + "epoch": 1.0948023569988148, + "grad_norm": 0.9969462752342224, + "learning_rate": 1.562233881205694e-05, + "loss": 3.7467, + "step": 97450 + }, + { + "epoch": 1.095364082169158, + "grad_norm": 1.1875361204147339, + "learning_rate": 1.562009189875409e-05, + "loss": 3.6715, + "step": 97500 + }, + { + "epoch": 1.095925807339501, + "grad_norm": 1.092423915863037, + "learning_rate": 1.5617844985451238e-05, + "loss": 3.7414, + "step": 97550 + }, + { + "epoch": 1.0964875325098442, + "grad_norm": 1.3828538656234741, + "learning_rate": 1.5615598072148387e-05, + "loss": 3.7184, + "step": 97600 + }, + { + "epoch": 1.0970492576801874, + "grad_norm": 1.3633829355239868, + "learning_rate": 1.5613351158845537e-05, + "loss": 3.7591, + "step": 97650 + }, + { + "epoch": 1.0976109828505305, + "grad_norm": 0.983437716960907, + "learning_rate": 1.5611104245542686e-05, + "loss": 3.6628, + "step": 97700 + }, + { + "epoch": 1.0981727080208736, + "grad_norm": 1.1300578117370605, + "learning_rate": 1.560885733223984e-05, + "loss": 3.7409, + "step": 97750 + }, + { + "epoch": 1.0987344331912168, + "grad_norm": 1.1857285499572754, + "learning_rate": 1.5606610418936988e-05, + "loss": 3.6844, + "step": 97800 + }, + { + "epoch": 1.0992961583615601, + "grad_norm": 1.1955124139785767, + "learning_rate": 1.5604363505634137e-05, + "loss": 3.7202, + "step": 97850 + }, + { + "epoch": 1.0998578835319033, + "grad_norm": 0.9688912630081177, + "learning_rate": 1.5602116592331287e-05, + "loss": 3.7509, + "step": 97900 + }, + { + "epoch": 1.1004196087022464, + "grad_norm": 1.1819149255752563, + "learning_rate": 1.5599869679028436e-05, + "loss": 3.7001, + "step": 97950 + }, + { + "epoch": 1.1009813338725896, + "grad_norm": 1.1008272171020508, + "learning_rate": 1.5597622765725585e-05, + "loss": 3.7148, + "step": 98000 + }, + { + "epoch": 1.1015430590429327, + "grad_norm": 1.1886111497879028, + "learning_rate": 1.5595375852422735e-05, + "loss": 3.7624, + "step": 98050 + }, + { + "epoch": 1.1021047842132758, + "grad_norm": 0.9735630750656128, + "learning_rate": 1.5593128939119887e-05, + "loss": 3.7987, + "step": 98100 + }, + { + "epoch": 1.102666509383619, + "grad_norm": 0.9906808733940125, + "learning_rate": 1.5590882025817037e-05, + "loss": 3.7065, + "step": 98150 + }, + { + "epoch": 1.1032282345539621, + "grad_norm": 1.0516225099563599, + "learning_rate": 1.5588635112514186e-05, + "loss": 3.7193, + "step": 98200 + }, + { + "epoch": 1.1037899597243053, + "grad_norm": 1.4463839530944824, + "learning_rate": 1.5586388199211335e-05, + "loss": 3.7413, + "step": 98250 + }, + { + "epoch": 1.1043516848946484, + "grad_norm": 1.1022162437438965, + "learning_rate": 1.5584141285908484e-05, + "loss": 3.671, + "step": 98300 + }, + { + "epoch": 1.1049134100649916, + "grad_norm": 1.0157989263534546, + "learning_rate": 1.5581894372605634e-05, + "loss": 3.671, + "step": 98350 + }, + { + "epoch": 1.1054751352353347, + "grad_norm": 1.3087482452392578, + "learning_rate": 1.5579647459302783e-05, + "loss": 3.7227, + "step": 98400 + }, + { + "epoch": 1.1060368604056778, + "grad_norm": 1.2353326082229614, + "learning_rate": 1.5577400545999936e-05, + "loss": 3.6886, + "step": 98450 + }, + { + "epoch": 1.106598585576021, + "grad_norm": 1.1072086095809937, + "learning_rate": 1.5575153632697085e-05, + "loss": 3.7556, + "step": 98500 + }, + { + "epoch": 1.1071603107463641, + "grad_norm": 1.0533146858215332, + "learning_rate": 1.5572906719394234e-05, + "loss": 3.6718, + "step": 98550 + }, + { + "epoch": 1.1077220359167075, + "grad_norm": 1.1676750183105469, + "learning_rate": 1.5570659806091384e-05, + "loss": 3.7657, + "step": 98600 + }, + { + "epoch": 1.1082837610870506, + "grad_norm": 1.3429557085037231, + "learning_rate": 1.5568412892788533e-05, + "loss": 3.7055, + "step": 98650 + }, + { + "epoch": 1.1088454862573938, + "grad_norm": 1.054917573928833, + "learning_rate": 1.5566165979485682e-05, + "loss": 3.7481, + "step": 98700 + }, + { + "epoch": 1.109407211427737, + "grad_norm": 1.1266285181045532, + "learning_rate": 1.5563919066182832e-05, + "loss": 3.6911, + "step": 98750 + }, + { + "epoch": 1.10996893659808, + "grad_norm": 1.0789703130722046, + "learning_rate": 1.5561672152879984e-05, + "loss": 3.762, + "step": 98800 + }, + { + "epoch": 1.1105306617684232, + "grad_norm": 1.295448660850525, + "learning_rate": 1.5559425239577134e-05, + "loss": 3.7771, + "step": 98850 + }, + { + "epoch": 1.1110923869387663, + "grad_norm": 1.0544154644012451, + "learning_rate": 1.5557178326274283e-05, + "loss": 3.6607, + "step": 98900 + }, + { + "epoch": 1.1116541121091095, + "grad_norm": 1.0492252111434937, + "learning_rate": 1.5554931412971432e-05, + "loss": 3.7173, + "step": 98950 + }, + { + "epoch": 1.1122158372794526, + "grad_norm": 1.2639553546905518, + "learning_rate": 1.5552684499668582e-05, + "loss": 3.7465, + "step": 99000 + }, + { + "epoch": 1.1127775624497958, + "grad_norm": 1.0499306917190552, + "learning_rate": 1.555043758636573e-05, + "loss": 3.6379, + "step": 99050 + }, + { + "epoch": 1.113339287620139, + "grad_norm": 1.1774410009384155, + "learning_rate": 1.554819067306288e-05, + "loss": 3.7211, + "step": 99100 + }, + { + "epoch": 1.113901012790482, + "grad_norm": 1.1847046613693237, + "learning_rate": 1.5545943759760033e-05, + "loss": 3.7337, + "step": 99150 + }, + { + "epoch": 1.1144627379608254, + "grad_norm": 0.9903504252433777, + "learning_rate": 1.5543696846457182e-05, + "loss": 3.7566, + "step": 99200 + }, + { + "epoch": 1.1150244631311685, + "grad_norm": 1.2131892442703247, + "learning_rate": 1.554144993315433e-05, + "loss": 3.8094, + "step": 99250 + }, + { + "epoch": 1.1155861883015117, + "grad_norm": 1.0200257301330566, + "learning_rate": 1.553920301985148e-05, + "loss": 3.7359, + "step": 99300 + }, + { + "epoch": 1.1161479134718548, + "grad_norm": 1.1790155172348022, + "learning_rate": 1.553695610654863e-05, + "loss": 3.6969, + "step": 99350 + }, + { + "epoch": 1.116709638642198, + "grad_norm": 1.2871062755584717, + "learning_rate": 1.553470919324578e-05, + "loss": 3.7804, + "step": 99400 + }, + { + "epoch": 1.117271363812541, + "grad_norm": 1.1888140439987183, + "learning_rate": 1.553246227994293e-05, + "loss": 3.7267, + "step": 99450 + }, + { + "epoch": 1.1178330889828842, + "grad_norm": 1.1498525142669678, + "learning_rate": 1.553021536664008e-05, + "loss": 3.7379, + "step": 99500 + }, + { + "epoch": 1.1183948141532274, + "grad_norm": 1.5620925426483154, + "learning_rate": 1.552796845333723e-05, + "loss": 3.6527, + "step": 99550 + }, + { + "epoch": 1.1189565393235705, + "grad_norm": 1.0870367288589478, + "learning_rate": 1.552572154003438e-05, + "loss": 3.7163, + "step": 99600 + }, + { + "epoch": 1.1195182644939137, + "grad_norm": 1.0825954675674438, + "learning_rate": 1.552347462673153e-05, + "loss": 3.667, + "step": 99650 + }, + { + "epoch": 1.1200799896642568, + "grad_norm": 1.2793874740600586, + "learning_rate": 1.552122771342868e-05, + "loss": 3.6803, + "step": 99700 + }, + { + "epoch": 1.1206417148346, + "grad_norm": 1.1869399547576904, + "learning_rate": 1.5518980800125828e-05, + "loss": 3.7173, + "step": 99750 + }, + { + "epoch": 1.121203440004943, + "grad_norm": 1.2226241827011108, + "learning_rate": 1.5516733886822978e-05, + "loss": 3.6696, + "step": 99800 + }, + { + "epoch": 1.1217651651752862, + "grad_norm": 1.0593713521957397, + "learning_rate": 1.551448697352013e-05, + "loss": 3.7269, + "step": 99850 + }, + { + "epoch": 1.1223268903456294, + "grad_norm": 1.244052767753601, + "learning_rate": 1.551224006021728e-05, + "loss": 3.749, + "step": 99900 + }, + { + "epoch": 1.1228886155159727, + "grad_norm": 0.9682693481445312, + "learning_rate": 1.550999314691443e-05, + "loss": 3.7029, + "step": 99950 + }, + { + "epoch": 1.1234503406863159, + "grad_norm": 1.2320550680160522, + "learning_rate": 1.5507746233611578e-05, + "loss": 3.6564, + "step": 100000 + }, + { + "epoch": 1.124012065856659, + "grad_norm": 1.021205186843872, + "learning_rate": 1.5505499320308728e-05, + "loss": 3.6825, + "step": 100050 + }, + { + "epoch": 1.1245737910270022, + "grad_norm": 1.2643808126449585, + "learning_rate": 1.5503252407005877e-05, + "loss": 3.75, + "step": 100100 + }, + { + "epoch": 1.1251355161973453, + "grad_norm": 1.2180043458938599, + "learning_rate": 1.5501005493703026e-05, + "loss": 3.7175, + "step": 100150 + }, + { + "epoch": 1.1256972413676885, + "grad_norm": 1.088374137878418, + "learning_rate": 1.549875858040018e-05, + "loss": 3.7444, + "step": 100200 + }, + { + "epoch": 1.1262589665380316, + "grad_norm": 1.0124397277832031, + "learning_rate": 1.5496556605363385e-05, + "loss": 3.633, + "step": 100250 + }, + { + "epoch": 1.1268206917083747, + "grad_norm": 1.0882328748703003, + "learning_rate": 1.5494309692060534e-05, + "loss": 3.699, + "step": 100300 + }, + { + "epoch": 1.1273824168787179, + "grad_norm": 1.055010199546814, + "learning_rate": 1.5492062778757683e-05, + "loss": 3.6626, + "step": 100350 + }, + { + "epoch": 1.127944142049061, + "grad_norm": 1.2953052520751953, + "learning_rate": 1.5489815865454833e-05, + "loss": 3.75, + "step": 100400 + }, + { + "epoch": 1.1285058672194042, + "grad_norm": 1.1922369003295898, + "learning_rate": 1.5487568952151982e-05, + "loss": 3.6626, + "step": 100450 + }, + { + "epoch": 1.1290675923897473, + "grad_norm": 1.0497500896453857, + "learning_rate": 1.5485322038849135e-05, + "loss": 3.7183, + "step": 100500 + }, + { + "epoch": 1.1296293175600907, + "grad_norm": 1.1612207889556885, + "learning_rate": 1.5483075125546284e-05, + "loss": 3.7178, + "step": 100550 + }, + { + "epoch": 1.1301910427304338, + "grad_norm": 1.0290371179580688, + "learning_rate": 1.5480828212243433e-05, + "loss": 3.783, + "step": 100600 + }, + { + "epoch": 1.130752767900777, + "grad_norm": 1.1671088933944702, + "learning_rate": 1.5478581298940583e-05, + "loss": 3.7467, + "step": 100650 + }, + { + "epoch": 1.13131449307112, + "grad_norm": 1.2563223838806152, + "learning_rate": 1.5476334385637732e-05, + "loss": 3.6544, + "step": 100700 + }, + { + "epoch": 1.1318762182414632, + "grad_norm": 1.2090439796447754, + "learning_rate": 1.547408747233488e-05, + "loss": 3.7014, + "step": 100750 + }, + { + "epoch": 1.1324379434118064, + "grad_norm": 1.234808325767517, + "learning_rate": 1.547184055903203e-05, + "loss": 3.6535, + "step": 100800 + }, + { + "epoch": 1.1329996685821495, + "grad_norm": 1.047635793685913, + "learning_rate": 1.5469593645729183e-05, + "loss": 3.6579, + "step": 100850 + }, + { + "epoch": 1.1335613937524927, + "grad_norm": 1.0435529947280884, + "learning_rate": 1.5467346732426333e-05, + "loss": 3.6524, + "step": 100900 + }, + { + "epoch": 1.1341231189228358, + "grad_norm": 1.3722610473632812, + "learning_rate": 1.5465099819123482e-05, + "loss": 3.7106, + "step": 100950 + }, + { + "epoch": 1.134684844093179, + "grad_norm": 1.0594531297683716, + "learning_rate": 1.546285290582063e-05, + "loss": 3.6417, + "step": 101000 + }, + { + "epoch": 1.135246569263522, + "grad_norm": 0.9762507677078247, + "learning_rate": 1.546060599251778e-05, + "loss": 3.7199, + "step": 101050 + }, + { + "epoch": 1.1358082944338652, + "grad_norm": 1.0277429819107056, + "learning_rate": 1.545835907921493e-05, + "loss": 3.6726, + "step": 101100 + }, + { + "epoch": 1.1363700196042084, + "grad_norm": 1.122733473777771, + "learning_rate": 1.545611216591208e-05, + "loss": 3.6344, + "step": 101150 + }, + { + "epoch": 1.1369317447745515, + "grad_norm": 1.0020467042922974, + "learning_rate": 1.5453910190875285e-05, + "loss": 3.6257, + "step": 101200 + }, + { + "epoch": 1.1374934699448946, + "grad_norm": 1.2245641946792603, + "learning_rate": 1.5451663277572434e-05, + "loss": 3.7534, + "step": 101250 + }, + { + "epoch": 1.138055195115238, + "grad_norm": 1.1989482641220093, + "learning_rate": 1.5449416364269584e-05, + "loss": 3.68, + "step": 101300 + }, + { + "epoch": 1.1386169202855811, + "grad_norm": 1.059341311454773, + "learning_rate": 1.5447169450966733e-05, + "loss": 3.7051, + "step": 101350 + }, + { + "epoch": 1.1391786454559243, + "grad_norm": 1.4291677474975586, + "learning_rate": 1.5444922537663886e-05, + "loss": 3.7544, + "step": 101400 + }, + { + "epoch": 1.1397403706262674, + "grad_norm": 1.1562519073486328, + "learning_rate": 1.5442675624361035e-05, + "loss": 3.7279, + "step": 101450 + }, + { + "epoch": 1.1403020957966106, + "grad_norm": 1.0684752464294434, + "learning_rate": 1.5440428711058184e-05, + "loss": 3.7251, + "step": 101500 + }, + { + "epoch": 1.1408638209669537, + "grad_norm": 1.074190616607666, + "learning_rate": 1.5438181797755334e-05, + "loss": 3.7438, + "step": 101550 + }, + { + "epoch": 1.1414255461372969, + "grad_norm": 1.1465007066726685, + "learning_rate": 1.5435934884452483e-05, + "loss": 3.7327, + "step": 101600 + }, + { + "epoch": 1.14198727130764, + "grad_norm": 1.063234567642212, + "learning_rate": 1.5433687971149632e-05, + "loss": 3.69, + "step": 101650 + }, + { + "epoch": 1.1425489964779831, + "grad_norm": 1.1495132446289062, + "learning_rate": 1.5431441057846782e-05, + "loss": 3.7358, + "step": 101700 + }, + { + "epoch": 1.1431107216483263, + "grad_norm": 1.066072702407837, + "learning_rate": 1.5429194144543934e-05, + "loss": 3.6587, + "step": 101750 + }, + { + "epoch": 1.1436724468186694, + "grad_norm": 1.6146560907363892, + "learning_rate": 1.5426947231241084e-05, + "loss": 3.7906, + "step": 101800 + }, + { + "epoch": 1.1442341719890126, + "grad_norm": 1.0587708950042725, + "learning_rate": 1.5424700317938233e-05, + "loss": 3.6794, + "step": 101850 + }, + { + "epoch": 1.144795897159356, + "grad_norm": 1.157436728477478, + "learning_rate": 1.5422453404635382e-05, + "loss": 3.7041, + "step": 101900 + }, + { + "epoch": 1.145357622329699, + "grad_norm": 1.051474690437317, + "learning_rate": 1.5420206491332532e-05, + "loss": 3.6958, + "step": 101950 + }, + { + "epoch": 1.1459193475000422, + "grad_norm": 1.2667258977890015, + "learning_rate": 1.541795957802968e-05, + "loss": 3.7161, + "step": 102000 + }, + { + "epoch": 1.1464810726703853, + "grad_norm": 1.2712479829788208, + "learning_rate": 1.541571266472683e-05, + "loss": 3.7091, + "step": 102050 + }, + { + "epoch": 1.1470427978407285, + "grad_norm": 1.0357364416122437, + "learning_rate": 1.5413465751423983e-05, + "loss": 3.7355, + "step": 102100 + }, + { + "epoch": 1.1476045230110716, + "grad_norm": 1.2836031913757324, + "learning_rate": 1.5411218838121132e-05, + "loss": 3.7377, + "step": 102150 + }, + { + "epoch": 1.1481662481814148, + "grad_norm": 1.155813217163086, + "learning_rate": 1.540897192481828e-05, + "loss": 3.7245, + "step": 102200 + }, + { + "epoch": 1.148727973351758, + "grad_norm": 1.5667455196380615, + "learning_rate": 1.540672501151543e-05, + "loss": 3.6165, + "step": 102250 + }, + { + "epoch": 1.149289698522101, + "grad_norm": 1.0747267007827759, + "learning_rate": 1.540447809821258e-05, + "loss": 3.6887, + "step": 102300 + }, + { + "epoch": 1.1498514236924442, + "grad_norm": 1.7879501581192017, + "learning_rate": 1.540223118490973e-05, + "loss": 3.6892, + "step": 102350 + }, + { + "epoch": 1.1504131488627873, + "grad_norm": 1.2933627367019653, + "learning_rate": 1.539998427160688e-05, + "loss": 3.7902, + "step": 102400 + }, + { + "epoch": 1.1509748740331305, + "grad_norm": 1.0486758947372437, + "learning_rate": 1.539773735830403e-05, + "loss": 3.6981, + "step": 102450 + }, + { + "epoch": 1.1515365992034736, + "grad_norm": 1.2376704216003418, + "learning_rate": 1.539549044500118e-05, + "loss": 3.7477, + "step": 102500 + }, + { + "epoch": 1.1520983243738168, + "grad_norm": 1.3571228981018066, + "learning_rate": 1.539324353169833e-05, + "loss": 3.7425, + "step": 102550 + }, + { + "epoch": 1.15266004954416, + "grad_norm": 1.027738094329834, + "learning_rate": 1.539099661839548e-05, + "loss": 3.7008, + "step": 102600 + }, + { + "epoch": 1.1532217747145033, + "grad_norm": 1.1306037902832031, + "learning_rate": 1.538874970509263e-05, + "loss": 3.6775, + "step": 102650 + }, + { + "epoch": 1.1537834998848464, + "grad_norm": 1.0368131399154663, + "learning_rate": 1.5386502791789778e-05, + "loss": 3.6883, + "step": 102700 + }, + { + "epoch": 1.1543452250551895, + "grad_norm": 1.1494563817977905, + "learning_rate": 1.5384255878486928e-05, + "loss": 3.6778, + "step": 102750 + }, + { + "epoch": 1.1549069502255327, + "grad_norm": 1.192533254623413, + "learning_rate": 1.538200896518408e-05, + "loss": 3.6954, + "step": 102800 + }, + { + "epoch": 1.1554686753958758, + "grad_norm": 1.1109049320220947, + "learning_rate": 1.537976205188123e-05, + "loss": 3.7542, + "step": 102850 + }, + { + "epoch": 1.156030400566219, + "grad_norm": 1.0070918798446655, + "learning_rate": 1.537751513857838e-05, + "loss": 3.8029, + "step": 102900 + }, + { + "epoch": 1.1565921257365621, + "grad_norm": 0.9579331874847412, + "learning_rate": 1.5375268225275528e-05, + "loss": 3.7035, + "step": 102950 + }, + { + "epoch": 1.1571538509069053, + "grad_norm": 1.4610426425933838, + "learning_rate": 1.5373021311972678e-05, + "loss": 3.6512, + "step": 103000 + }, + { + "epoch": 1.1577155760772484, + "grad_norm": 1.1504992246627808, + "learning_rate": 1.5370774398669827e-05, + "loss": 3.6739, + "step": 103050 + }, + { + "epoch": 1.1582773012475915, + "grad_norm": 2.7192490100860596, + "learning_rate": 1.5368527485366976e-05, + "loss": 3.7345, + "step": 103100 + }, + { + "epoch": 1.1588390264179347, + "grad_norm": 1.1754087209701538, + "learning_rate": 1.536628057206413e-05, + "loss": 3.7255, + "step": 103150 + }, + { + "epoch": 1.1594007515882778, + "grad_norm": 1.2351810932159424, + "learning_rate": 1.5364033658761278e-05, + "loss": 3.6805, + "step": 103200 + }, + { + "epoch": 1.1599624767586212, + "grad_norm": 1.058080792427063, + "learning_rate": 1.5361786745458427e-05, + "loss": 3.6916, + "step": 103250 + }, + { + "epoch": 1.1605242019289643, + "grad_norm": 1.3002619743347168, + "learning_rate": 1.5359539832155577e-05, + "loss": 3.7551, + "step": 103300 + }, + { + "epoch": 1.1610859270993075, + "grad_norm": 1.1462116241455078, + "learning_rate": 1.5357292918852726e-05, + "loss": 3.5689, + "step": 103350 + }, + { + "epoch": 1.1616476522696506, + "grad_norm": 1.3079191446304321, + "learning_rate": 1.5355046005549875e-05, + "loss": 3.7718, + "step": 103400 + }, + { + "epoch": 1.1622093774399938, + "grad_norm": 1.3035516738891602, + "learning_rate": 1.5352799092247025e-05, + "loss": 3.632, + "step": 103450 + }, + { + "epoch": 1.162771102610337, + "grad_norm": 1.0845363140106201, + "learning_rate": 1.5350552178944177e-05, + "loss": 3.5912, + "step": 103500 + }, + { + "epoch": 1.16333282778068, + "grad_norm": 1.0748491287231445, + "learning_rate": 1.5348305265641327e-05, + "loss": 3.6705, + "step": 103550 + }, + { + "epoch": 1.1638945529510232, + "grad_norm": 1.2455357313156128, + "learning_rate": 1.5346058352338476e-05, + "loss": 3.7515, + "step": 103600 + }, + { + "epoch": 1.1644562781213663, + "grad_norm": 1.014829397201538, + "learning_rate": 1.5343811439035625e-05, + "loss": 3.6542, + "step": 103650 + }, + { + "epoch": 1.1650180032917095, + "grad_norm": 1.1986323595046997, + "learning_rate": 1.5341564525732775e-05, + "loss": 3.7361, + "step": 103700 + }, + { + "epoch": 1.1655797284620526, + "grad_norm": 1.2283251285552979, + "learning_rate": 1.5339317612429924e-05, + "loss": 3.7181, + "step": 103750 + }, + { + "epoch": 1.1661414536323957, + "grad_norm": 1.078841209411621, + "learning_rate": 1.5337070699127073e-05, + "loss": 3.6782, + "step": 103800 + }, + { + "epoch": 1.1667031788027389, + "grad_norm": 1.1124110221862793, + "learning_rate": 1.5334823785824226e-05, + "loss": 3.7446, + "step": 103850 + }, + { + "epoch": 1.167264903973082, + "grad_norm": 1.2355866432189941, + "learning_rate": 1.5332621810787432e-05, + "loss": 3.6567, + "step": 103900 + }, + { + "epoch": 1.1678266291434252, + "grad_norm": 1.1304376125335693, + "learning_rate": 1.533037489748458e-05, + "loss": 3.7139, + "step": 103950 + }, + { + "epoch": 1.1683883543137685, + "grad_norm": 1.1909832954406738, + "learning_rate": 1.532812798418173e-05, + "loss": 3.6523, + "step": 104000 + }, + { + "epoch": 1.1689500794841117, + "grad_norm": 1.1495308876037598, + "learning_rate": 1.532588107087888e-05, + "loss": 3.6911, + "step": 104050 + }, + { + "epoch": 1.1695118046544548, + "grad_norm": 1.2217122316360474, + "learning_rate": 1.532363415757603e-05, + "loss": 3.7879, + "step": 104100 + }, + { + "epoch": 1.170073529824798, + "grad_norm": 1.1243988275527954, + "learning_rate": 1.5321387244273182e-05, + "loss": 3.7405, + "step": 104150 + }, + { + "epoch": 1.170635254995141, + "grad_norm": 1.328789234161377, + "learning_rate": 1.531914033097033e-05, + "loss": 3.7426, + "step": 104200 + }, + { + "epoch": 1.1711969801654842, + "grad_norm": 1.166298270225525, + "learning_rate": 1.531689341766748e-05, + "loss": 3.6431, + "step": 104250 + }, + { + "epoch": 1.1717587053358274, + "grad_norm": 1.3083101511001587, + "learning_rate": 1.531464650436463e-05, + "loss": 3.7106, + "step": 104300 + }, + { + "epoch": 1.1723204305061705, + "grad_norm": 1.0733588933944702, + "learning_rate": 1.531239959106178e-05, + "loss": 3.6376, + "step": 104350 + }, + { + "epoch": 1.1728821556765137, + "grad_norm": 1.035312533378601, + "learning_rate": 1.531015267775893e-05, + "loss": 3.7374, + "step": 104400 + }, + { + "epoch": 1.1734438808468568, + "grad_norm": 0.9976434707641602, + "learning_rate": 1.5307905764456078e-05, + "loss": 3.7363, + "step": 104450 + }, + { + "epoch": 1.1740056060172, + "grad_norm": 1.200663685798645, + "learning_rate": 1.530565885115323e-05, + "loss": 3.8055, + "step": 104500 + }, + { + "epoch": 1.1745673311875433, + "grad_norm": 1.2299480438232422, + "learning_rate": 1.5303456876116436e-05, + "loss": 3.743, + "step": 104550 + }, + { + "epoch": 1.1751290563578864, + "grad_norm": 1.0745593309402466, + "learning_rate": 1.5301209962813586e-05, + "loss": 3.6963, + "step": 104600 + }, + { + "epoch": 1.1756907815282296, + "grad_norm": 1.1543883085250854, + "learning_rate": 1.5298963049510735e-05, + "loss": 3.7487, + "step": 104650 + }, + { + "epoch": 1.1762525066985727, + "grad_norm": 1.074861764907837, + "learning_rate": 1.5296716136207884e-05, + "loss": 3.7323, + "step": 104700 + }, + { + "epoch": 1.1768142318689159, + "grad_norm": 1.0769885778427124, + "learning_rate": 1.5294469222905034e-05, + "loss": 3.7242, + "step": 104750 + }, + { + "epoch": 1.177375957039259, + "grad_norm": 1.5533466339111328, + "learning_rate": 1.5292222309602186e-05, + "loss": 3.7333, + "step": 104800 + }, + { + "epoch": 1.1779376822096022, + "grad_norm": 1.3578367233276367, + "learning_rate": 1.5289975396299336e-05, + "loss": 3.6818, + "step": 104850 + }, + { + "epoch": 1.1784994073799453, + "grad_norm": 1.153058648109436, + "learning_rate": 1.5287728482996485e-05, + "loss": 3.6829, + "step": 104900 + }, + { + "epoch": 1.1790611325502884, + "grad_norm": 1.275603175163269, + "learning_rate": 1.5285481569693634e-05, + "loss": 3.6589, + "step": 104950 + }, + { + "epoch": 1.1796228577206316, + "grad_norm": 1.2551240921020508, + "learning_rate": 1.5283234656390784e-05, + "loss": 3.7048, + "step": 105000 + }, + { + "epoch": 1.1801845828909747, + "grad_norm": 1.1533915996551514, + "learning_rate": 1.5280987743087933e-05, + "loss": 3.7098, + "step": 105050 + }, + { + "epoch": 1.1807463080613179, + "grad_norm": 1.2769936323165894, + "learning_rate": 1.5278740829785082e-05, + "loss": 3.6012, + "step": 105100 + }, + { + "epoch": 1.181308033231661, + "grad_norm": 1.0044163465499878, + "learning_rate": 1.5276493916482235e-05, + "loss": 3.7165, + "step": 105150 + }, + { + "epoch": 1.1818697584020041, + "grad_norm": 1.1639456748962402, + "learning_rate": 1.5274247003179384e-05, + "loss": 3.7855, + "step": 105200 + }, + { + "epoch": 1.1824314835723473, + "grad_norm": 1.2102593183517456, + "learning_rate": 1.5272000089876534e-05, + "loss": 3.6921, + "step": 105250 + }, + { + "epoch": 1.1829932087426904, + "grad_norm": 1.1959712505340576, + "learning_rate": 1.5269753176573683e-05, + "loss": 3.714, + "step": 105300 + }, + { + "epoch": 1.1835549339130338, + "grad_norm": 0.9928191304206848, + "learning_rate": 1.5267506263270832e-05, + "loss": 3.725, + "step": 105350 + }, + { + "epoch": 1.184116659083377, + "grad_norm": 1.190778374671936, + "learning_rate": 1.526525934996798e-05, + "loss": 3.7029, + "step": 105400 + }, + { + "epoch": 1.18467838425372, + "grad_norm": 1.3059523105621338, + "learning_rate": 1.526301243666513e-05, + "loss": 3.6104, + "step": 105450 + }, + { + "epoch": 1.1852401094240632, + "grad_norm": 1.4401763677597046, + "learning_rate": 1.5260765523362284e-05, + "loss": 3.7813, + "step": 105500 + }, + { + "epoch": 1.1858018345944064, + "grad_norm": 1.1502825021743774, + "learning_rate": 1.5258518610059431e-05, + "loss": 3.7401, + "step": 105550 + }, + { + "epoch": 1.1863635597647495, + "grad_norm": 1.2435052394866943, + "learning_rate": 1.5256271696756582e-05, + "loss": 3.7393, + "step": 105600 + }, + { + "epoch": 1.1869252849350926, + "grad_norm": 2.0304958820343018, + "learning_rate": 1.5254024783453732e-05, + "loss": 3.7324, + "step": 105650 + }, + { + "epoch": 1.1874870101054358, + "grad_norm": 1.141411542892456, + "learning_rate": 1.5251777870150881e-05, + "loss": 3.655, + "step": 105700 + }, + { + "epoch": 1.188048735275779, + "grad_norm": 1.2367336750030518, + "learning_rate": 1.524953095684803e-05, + "loss": 3.6525, + "step": 105750 + }, + { + "epoch": 1.188610460446122, + "grad_norm": 1.052817463874817, + "learning_rate": 1.5247284043545181e-05, + "loss": 3.7261, + "step": 105800 + }, + { + "epoch": 1.1891721856164652, + "grad_norm": 1.1793750524520874, + "learning_rate": 1.524503713024233e-05, + "loss": 3.7451, + "step": 105850 + }, + { + "epoch": 1.1897339107868086, + "grad_norm": 1.0441009998321533, + "learning_rate": 1.524279021693948e-05, + "loss": 3.6667, + "step": 105900 + }, + { + "epoch": 1.1902956359571517, + "grad_norm": 1.2185910940170288, + "learning_rate": 1.5240543303636631e-05, + "loss": 3.7966, + "step": 105950 + }, + { + "epoch": 1.1908573611274949, + "grad_norm": 1.138037919998169, + "learning_rate": 1.523829639033378e-05, + "loss": 3.6983, + "step": 106000 + }, + { + "epoch": 1.191419086297838, + "grad_norm": 1.0619055032730103, + "learning_rate": 1.523604947703093e-05, + "loss": 3.7142, + "step": 106050 + }, + { + "epoch": 1.1919808114681811, + "grad_norm": 1.0743881464004517, + "learning_rate": 1.5233802563728079e-05, + "loss": 3.7762, + "step": 106100 + }, + { + "epoch": 1.1925425366385243, + "grad_norm": 1.375406265258789, + "learning_rate": 1.523155565042523e-05, + "loss": 3.736, + "step": 106150 + }, + { + "epoch": 1.1931042618088674, + "grad_norm": 1.1640788316726685, + "learning_rate": 1.522930873712238e-05, + "loss": 3.7453, + "step": 106200 + }, + { + "epoch": 1.1936659869792106, + "grad_norm": 1.2602263689041138, + "learning_rate": 1.5227061823819528e-05, + "loss": 3.6928, + "step": 106250 + }, + { + "epoch": 1.1942277121495537, + "grad_norm": 1.1936029195785522, + "learning_rate": 1.522481491051668e-05, + "loss": 3.6771, + "step": 106300 + }, + { + "epoch": 1.1947894373198968, + "grad_norm": 1.2278850078582764, + "learning_rate": 1.5222567997213829e-05, + "loss": 3.7069, + "step": 106350 + }, + { + "epoch": 1.19535116249024, + "grad_norm": 1.2584307193756104, + "learning_rate": 1.5220321083910978e-05, + "loss": 3.6803, + "step": 106400 + }, + { + "epoch": 1.1959128876605831, + "grad_norm": 1.0643031597137451, + "learning_rate": 1.5218074170608127e-05, + "loss": 3.6955, + "step": 106450 + }, + { + "epoch": 1.1964746128309263, + "grad_norm": 1.1142960786819458, + "learning_rate": 1.5215827257305278e-05, + "loss": 3.7017, + "step": 106500 + }, + { + "epoch": 1.1970363380012694, + "grad_norm": 1.0555589199066162, + "learning_rate": 1.5213580344002428e-05, + "loss": 3.6561, + "step": 106550 + }, + { + "epoch": 1.1975980631716125, + "grad_norm": 0.980929434299469, + "learning_rate": 1.5211333430699577e-05, + "loss": 3.7023, + "step": 106600 + }, + { + "epoch": 1.198159788341956, + "grad_norm": 1.0431848764419556, + "learning_rate": 1.5209086517396726e-05, + "loss": 3.6619, + "step": 106650 + }, + { + "epoch": 1.198721513512299, + "grad_norm": 0.95896315574646, + "learning_rate": 1.5206839604093877e-05, + "loss": 3.7181, + "step": 106700 + }, + { + "epoch": 1.1992832386826422, + "grad_norm": 1.2175319194793701, + "learning_rate": 1.5204592690791027e-05, + "loss": 3.6748, + "step": 106750 + }, + { + "epoch": 1.1998449638529853, + "grad_norm": 1.2603718042373657, + "learning_rate": 1.5202345777488176e-05, + "loss": 3.7236, + "step": 106800 + }, + { + "epoch": 1.2004066890233285, + "grad_norm": 1.4466156959533691, + "learning_rate": 1.5200098864185327e-05, + "loss": 3.6938, + "step": 106850 + }, + { + "epoch": 1.2009684141936716, + "grad_norm": 1.2379579544067383, + "learning_rate": 1.5197851950882476e-05, + "loss": 3.7017, + "step": 106900 + }, + { + "epoch": 1.2015301393640148, + "grad_norm": 1.505018711090088, + "learning_rate": 1.5195605037579626e-05, + "loss": 3.6715, + "step": 106950 + }, + { + "epoch": 1.202091864534358, + "grad_norm": 1.1268073320388794, + "learning_rate": 1.5193358124276775e-05, + "loss": 3.6606, + "step": 107000 + }, + { + "epoch": 1.202653589704701, + "grad_norm": 1.1279513835906982, + "learning_rate": 1.5191111210973926e-05, + "loss": 3.6994, + "step": 107050 + }, + { + "epoch": 1.2032153148750442, + "grad_norm": 1.2397137880325317, + "learning_rate": 1.5188864297671075e-05, + "loss": 3.7105, + "step": 107100 + }, + { + "epoch": 1.2037770400453873, + "grad_norm": 1.1287885904312134, + "learning_rate": 1.5186617384368225e-05, + "loss": 3.6917, + "step": 107150 + }, + { + "epoch": 1.2043387652157305, + "grad_norm": 1.0952703952789307, + "learning_rate": 1.5184370471065376e-05, + "loss": 3.7041, + "step": 107200 + }, + { + "epoch": 1.2049004903860738, + "grad_norm": 1.8485311269760132, + "learning_rate": 1.5182123557762525e-05, + "loss": 3.7, + "step": 107250 + }, + { + "epoch": 1.205462215556417, + "grad_norm": 1.0265604257583618, + "learning_rate": 1.5179876644459674e-05, + "loss": 3.7924, + "step": 107300 + }, + { + "epoch": 1.2060239407267601, + "grad_norm": 1.2580769062042236, + "learning_rate": 1.5177629731156824e-05, + "loss": 3.7509, + "step": 107350 + }, + { + "epoch": 1.2065856658971033, + "grad_norm": 1.2376199960708618, + "learning_rate": 1.5175382817853975e-05, + "loss": 3.7154, + "step": 107400 + }, + { + "epoch": 1.2071473910674464, + "grad_norm": 1.0325915813446045, + "learning_rate": 1.5173135904551124e-05, + "loss": 3.7247, + "step": 107450 + }, + { + "epoch": 1.2077091162377895, + "grad_norm": 1.171482801437378, + "learning_rate": 1.5170888991248273e-05, + "loss": 3.6708, + "step": 107500 + }, + { + "epoch": 1.2082708414081327, + "grad_norm": 1.3903834819793701, + "learning_rate": 1.5168642077945424e-05, + "loss": 3.6801, + "step": 107550 + }, + { + "epoch": 1.2088325665784758, + "grad_norm": 1.2108118534088135, + "learning_rate": 1.5166395164642574e-05, + "loss": 3.7365, + "step": 107600 + }, + { + "epoch": 1.209394291748819, + "grad_norm": 1.1592521667480469, + "learning_rate": 1.5164148251339723e-05, + "loss": 3.6927, + "step": 107650 + }, + { + "epoch": 1.209956016919162, + "grad_norm": 1.0881690979003906, + "learning_rate": 1.5161901338036872e-05, + "loss": 3.682, + "step": 107700 + }, + { + "epoch": 1.2105177420895052, + "grad_norm": 1.1194993257522583, + "learning_rate": 1.5159654424734023e-05, + "loss": 3.6634, + "step": 107750 + }, + { + "epoch": 1.2110794672598484, + "grad_norm": 1.1075609922409058, + "learning_rate": 1.5157407511431173e-05, + "loss": 3.7438, + "step": 107800 + }, + { + "epoch": 1.2116411924301915, + "grad_norm": 1.0043065547943115, + "learning_rate": 1.5155160598128322e-05, + "loss": 3.7448, + "step": 107850 + }, + { + "epoch": 1.2122029176005347, + "grad_norm": 1.1683133840560913, + "learning_rate": 1.5152913684825473e-05, + "loss": 3.6245, + "step": 107900 + }, + { + "epoch": 1.2127646427708778, + "grad_norm": 1.2724297046661377, + "learning_rate": 1.5150666771522622e-05, + "loss": 3.6912, + "step": 107950 + }, + { + "epoch": 1.2133263679412212, + "grad_norm": 1.1075135469436646, + "learning_rate": 1.5148419858219772e-05, + "loss": 3.7175, + "step": 108000 + }, + { + "epoch": 1.2138880931115643, + "grad_norm": 1.2676115036010742, + "learning_rate": 1.514617294491692e-05, + "loss": 3.7054, + "step": 108050 + }, + { + "epoch": 1.2144498182819075, + "grad_norm": 1.432173490524292, + "learning_rate": 1.5143926031614072e-05, + "loss": 3.67, + "step": 108100 + }, + { + "epoch": 1.2150115434522506, + "grad_norm": 1.1345535516738892, + "learning_rate": 1.5141679118311221e-05, + "loss": 3.704, + "step": 108150 + }, + { + "epoch": 1.2155732686225937, + "grad_norm": 1.0536555051803589, + "learning_rate": 1.513943220500837e-05, + "loss": 3.7215, + "step": 108200 + }, + { + "epoch": 1.2161349937929369, + "grad_norm": 1.1665971279144287, + "learning_rate": 1.5137185291705521e-05, + "loss": 3.7347, + "step": 108250 + }, + { + "epoch": 1.21669671896328, + "grad_norm": 1.1783932447433472, + "learning_rate": 1.513493837840267e-05, + "loss": 3.6768, + "step": 108300 + }, + { + "epoch": 1.2172584441336232, + "grad_norm": 0.9993147253990173, + "learning_rate": 1.513269146509982e-05, + "loss": 3.7377, + "step": 108350 + }, + { + "epoch": 1.2178201693039663, + "grad_norm": 1.104299783706665, + "learning_rate": 1.513044455179697e-05, + "loss": 3.7368, + "step": 108400 + }, + { + "epoch": 1.2183818944743094, + "grad_norm": 1.1289198398590088, + "learning_rate": 1.512819763849412e-05, + "loss": 3.6997, + "step": 108450 + }, + { + "epoch": 1.2189436196446526, + "grad_norm": 1.1112141609191895, + "learning_rate": 1.512595072519127e-05, + "loss": 3.7008, + "step": 108500 + }, + { + "epoch": 1.2195053448149957, + "grad_norm": 1.1366983652114868, + "learning_rate": 1.5123703811888419e-05, + "loss": 3.6892, + "step": 108550 + }, + { + "epoch": 1.220067069985339, + "grad_norm": 1.1454689502716064, + "learning_rate": 1.5121456898585568e-05, + "loss": 3.767, + "step": 108600 + }, + { + "epoch": 1.2206287951556822, + "grad_norm": 1.21815824508667, + "learning_rate": 1.511920998528272e-05, + "loss": 3.724, + "step": 108650 + }, + { + "epoch": 1.2211905203260254, + "grad_norm": 1.1037623882293701, + "learning_rate": 1.5116963071979869e-05, + "loss": 3.6244, + "step": 108700 + }, + { + "epoch": 1.2217522454963685, + "grad_norm": 1.2103321552276611, + "learning_rate": 1.5114716158677018e-05, + "loss": 3.753, + "step": 108750 + }, + { + "epoch": 1.2223139706667117, + "grad_norm": 1.3215216398239136, + "learning_rate": 1.5112469245374169e-05, + "loss": 3.6574, + "step": 108800 + }, + { + "epoch": 1.2228756958370548, + "grad_norm": 1.1635072231292725, + "learning_rate": 1.5110222332071318e-05, + "loss": 3.6211, + "step": 108850 + }, + { + "epoch": 1.223437421007398, + "grad_norm": 1.132358193397522, + "learning_rate": 1.5107975418768468e-05, + "loss": 3.6865, + "step": 108900 + }, + { + "epoch": 1.223999146177741, + "grad_norm": 1.190798282623291, + "learning_rate": 1.5105728505465617e-05, + "loss": 3.7067, + "step": 108950 + }, + { + "epoch": 1.2245608713480842, + "grad_norm": 1.2921621799468994, + "learning_rate": 1.5103481592162768e-05, + "loss": 3.7109, + "step": 109000 + }, + { + "epoch": 1.2251225965184274, + "grad_norm": 1.2048287391662598, + "learning_rate": 1.5101234678859917e-05, + "loss": 3.6675, + "step": 109050 + }, + { + "epoch": 1.2256843216887705, + "grad_norm": 1.2336902618408203, + "learning_rate": 1.5098987765557067e-05, + "loss": 3.6948, + "step": 109100 + }, + { + "epoch": 1.2262460468591136, + "grad_norm": 1.1223293542861938, + "learning_rate": 1.5096740852254218e-05, + "loss": 3.6968, + "step": 109150 + }, + { + "epoch": 1.2268077720294568, + "grad_norm": 1.248460054397583, + "learning_rate": 1.5094493938951367e-05, + "loss": 3.6594, + "step": 109200 + }, + { + "epoch": 1.2273694971998, + "grad_norm": 1.082493782043457, + "learning_rate": 1.5092291963914575e-05, + "loss": 3.7333, + "step": 109250 + }, + { + "epoch": 1.227931222370143, + "grad_norm": 1.0764968395233154, + "learning_rate": 1.5090045050611724e-05, + "loss": 3.7282, + "step": 109300 + }, + { + "epoch": 1.2284929475404864, + "grad_norm": 1.0554648637771606, + "learning_rate": 1.5087798137308873e-05, + "loss": 3.7384, + "step": 109350 + }, + { + "epoch": 1.2290546727108296, + "grad_norm": 1.110594630241394, + "learning_rate": 1.5085551224006023e-05, + "loss": 3.7397, + "step": 109400 + }, + { + "epoch": 1.2296163978811727, + "grad_norm": 1.0039312839508057, + "learning_rate": 1.5083304310703174e-05, + "loss": 3.6706, + "step": 109450 + }, + { + "epoch": 1.2301781230515159, + "grad_norm": 1.1047683954238892, + "learning_rate": 1.5081057397400323e-05, + "loss": 3.7168, + "step": 109500 + }, + { + "epoch": 1.230739848221859, + "grad_norm": 1.0532225370407104, + "learning_rate": 1.5078810484097472e-05, + "loss": 3.6769, + "step": 109550 + }, + { + "epoch": 1.2313015733922021, + "grad_norm": 1.2127481698989868, + "learning_rate": 1.5076563570794621e-05, + "loss": 3.7215, + "step": 109600 + }, + { + "epoch": 1.2318632985625453, + "grad_norm": 1.2573966979980469, + "learning_rate": 1.5074316657491773e-05, + "loss": 3.6805, + "step": 109650 + }, + { + "epoch": 1.2324250237328884, + "grad_norm": 1.2704925537109375, + "learning_rate": 1.5072069744188922e-05, + "loss": 3.7492, + "step": 109700 + }, + { + "epoch": 1.2329867489032316, + "grad_norm": 1.2403591871261597, + "learning_rate": 1.5069822830886071e-05, + "loss": 3.7066, + "step": 109750 + }, + { + "epoch": 1.2335484740735747, + "grad_norm": 1.384395718574524, + "learning_rate": 1.5067575917583222e-05, + "loss": 3.6857, + "step": 109800 + }, + { + "epoch": 1.2341101992439178, + "grad_norm": 1.0223932266235352, + "learning_rate": 1.5065329004280371e-05, + "loss": 3.6618, + "step": 109850 + }, + { + "epoch": 1.234671924414261, + "grad_norm": 1.2047230005264282, + "learning_rate": 1.506308209097752e-05, + "loss": 3.7276, + "step": 109900 + }, + { + "epoch": 1.2352336495846044, + "grad_norm": 0.9568697810173035, + "learning_rate": 1.506083517767467e-05, + "loss": 3.6579, + "step": 109950 + }, + { + "epoch": 1.2357953747549475, + "grad_norm": 1.0862027406692505, + "learning_rate": 1.5058588264371821e-05, + "loss": 3.7278, + "step": 110000 + }, + { + "epoch": 1.2363570999252906, + "grad_norm": 1.1824320554733276, + "learning_rate": 1.505634135106897e-05, + "loss": 3.7534, + "step": 110050 + }, + { + "epoch": 1.2369188250956338, + "grad_norm": 1.0982342958450317, + "learning_rate": 1.505409443776612e-05, + "loss": 3.7264, + "step": 110100 + }, + { + "epoch": 1.237480550265977, + "grad_norm": 0.9904276728630066, + "learning_rate": 1.505184752446327e-05, + "loss": 3.7695, + "step": 110150 + }, + { + "epoch": 1.23804227543632, + "grad_norm": 0.9733591079711914, + "learning_rate": 1.504960061116042e-05, + "loss": 3.7002, + "step": 110200 + }, + { + "epoch": 1.2386040006066632, + "grad_norm": 1.2126539945602417, + "learning_rate": 1.504735369785757e-05, + "loss": 3.7215, + "step": 110250 + }, + { + "epoch": 1.2391657257770063, + "grad_norm": 1.3110642433166504, + "learning_rate": 1.5045106784554719e-05, + "loss": 3.6464, + "step": 110300 + }, + { + "epoch": 1.2397274509473495, + "grad_norm": 1.0849037170410156, + "learning_rate": 1.504285987125187e-05, + "loss": 3.6448, + "step": 110350 + }, + { + "epoch": 1.2402891761176926, + "grad_norm": 1.155277967453003, + "learning_rate": 1.5040612957949019e-05, + "loss": 3.6634, + "step": 110400 + }, + { + "epoch": 1.2408509012880358, + "grad_norm": 1.2127948999404907, + "learning_rate": 1.5038366044646168e-05, + "loss": 3.7335, + "step": 110450 + }, + { + "epoch": 1.241412626458379, + "grad_norm": 1.2872679233551025, + "learning_rate": 1.503611913134332e-05, + "loss": 3.6575, + "step": 110500 + }, + { + "epoch": 1.241974351628722, + "grad_norm": 1.0627508163452148, + "learning_rate": 1.5033872218040469e-05, + "loss": 3.7264, + "step": 110550 + }, + { + "epoch": 1.2425360767990652, + "grad_norm": 1.176890254020691, + "learning_rate": 1.5031625304737618e-05, + "loss": 3.6949, + "step": 110600 + }, + { + "epoch": 1.2430978019694083, + "grad_norm": 1.2959498167037964, + "learning_rate": 1.5029378391434767e-05, + "loss": 3.7166, + "step": 110650 + }, + { + "epoch": 1.2436595271397517, + "grad_norm": 1.1180050373077393, + "learning_rate": 1.5027131478131918e-05, + "loss": 3.7013, + "step": 110700 + }, + { + "epoch": 1.2442212523100948, + "grad_norm": 1.2117652893066406, + "learning_rate": 1.5024884564829068e-05, + "loss": 3.6865, + "step": 110750 + }, + { + "epoch": 1.244782977480438, + "grad_norm": 1.0430887937545776, + "learning_rate": 1.5022637651526217e-05, + "loss": 3.658, + "step": 110800 + }, + { + "epoch": 1.2453447026507811, + "grad_norm": 1.1503196954727173, + "learning_rate": 1.5020390738223368e-05, + "loss": 3.7274, + "step": 110850 + }, + { + "epoch": 1.2459064278211243, + "grad_norm": 1.1620378494262695, + "learning_rate": 1.5018143824920517e-05, + "loss": 3.6921, + "step": 110900 + }, + { + "epoch": 1.2464681529914674, + "grad_norm": 1.1004583835601807, + "learning_rate": 1.5015896911617667e-05, + "loss": 3.6466, + "step": 110950 + }, + { + "epoch": 1.2470298781618105, + "grad_norm": 1.1144731044769287, + "learning_rate": 1.5013649998314816e-05, + "loss": 3.6323, + "step": 111000 + }, + { + "epoch": 1.2475916033321537, + "grad_norm": 1.1629180908203125, + "learning_rate": 1.5011403085011967e-05, + "loss": 3.5816, + "step": 111050 + }, + { + "epoch": 1.2481533285024968, + "grad_norm": 1.1811234951019287, + "learning_rate": 1.5009156171709116e-05, + "loss": 3.7079, + "step": 111100 + }, + { + "epoch": 1.24871505367284, + "grad_norm": 1.2378838062286377, + "learning_rate": 1.5006909258406266e-05, + "loss": 3.6832, + "step": 111150 + }, + { + "epoch": 1.2492767788431831, + "grad_norm": 1.2465800046920776, + "learning_rate": 1.5004662345103415e-05, + "loss": 3.6852, + "step": 111200 + }, + { + "epoch": 1.2498385040135263, + "grad_norm": 1.2350941896438599, + "learning_rate": 1.5002415431800566e-05, + "loss": 3.7366, + "step": 111250 + }, + { + "epoch": 1.2504002291838696, + "grad_norm": 1.2513706684112549, + "learning_rate": 1.5000213456763772e-05, + "loss": 3.7345, + "step": 111300 + }, + { + "epoch": 1.2509619543542128, + "grad_norm": 1.3040194511413574, + "learning_rate": 1.4997966543460923e-05, + "loss": 3.7018, + "step": 111350 + }, + { + "epoch": 1.251523679524556, + "grad_norm": 1.1939553022384644, + "learning_rate": 1.4995719630158072e-05, + "loss": 3.7485, + "step": 111400 + }, + { + "epoch": 1.252085404694899, + "grad_norm": 1.2655975818634033, + "learning_rate": 1.4993472716855221e-05, + "loss": 3.7011, + "step": 111450 + }, + { + "epoch": 1.2526471298652422, + "grad_norm": 1.149431824684143, + "learning_rate": 1.4991225803552372e-05, + "loss": 3.6237, + "step": 111500 + }, + { + "epoch": 1.2532088550355853, + "grad_norm": 1.2270399332046509, + "learning_rate": 1.4988978890249522e-05, + "loss": 3.6909, + "step": 111550 + }, + { + "epoch": 1.2537705802059285, + "grad_norm": 1.2324990034103394, + "learning_rate": 1.4986731976946671e-05, + "loss": 3.7021, + "step": 111600 + }, + { + "epoch": 1.2543323053762716, + "grad_norm": 1.0944278240203857, + "learning_rate": 1.498448506364382e-05, + "loss": 3.7399, + "step": 111650 + }, + { + "epoch": 1.2548940305466147, + "grad_norm": 1.1698552370071411, + "learning_rate": 1.4982238150340971e-05, + "loss": 3.7134, + "step": 111700 + }, + { + "epoch": 1.2554557557169579, + "grad_norm": 1.2568777799606323, + "learning_rate": 1.497999123703812e-05, + "loss": 3.6943, + "step": 111750 + }, + { + "epoch": 1.256017480887301, + "grad_norm": 1.315553069114685, + "learning_rate": 1.497774432373527e-05, + "loss": 3.7308, + "step": 111800 + }, + { + "epoch": 1.2565792060576442, + "grad_norm": 1.1367578506469727, + "learning_rate": 1.4975497410432421e-05, + "loss": 3.6847, + "step": 111850 + }, + { + "epoch": 1.2571409312279873, + "grad_norm": 1.129918098449707, + "learning_rate": 1.497325049712957e-05, + "loss": 3.6771, + "step": 111900 + }, + { + "epoch": 1.2577026563983305, + "grad_norm": 1.112339735031128, + "learning_rate": 1.497100358382672e-05, + "loss": 3.6944, + "step": 111950 + }, + { + "epoch": 1.2582643815686736, + "grad_norm": 1.0623482465744019, + "learning_rate": 1.4968756670523869e-05, + "loss": 3.7297, + "step": 112000 + }, + { + "epoch": 1.2588261067390167, + "grad_norm": 1.1499406099319458, + "learning_rate": 1.496650975722102e-05, + "loss": 3.6691, + "step": 112050 + }, + { + "epoch": 1.25938783190936, + "grad_norm": 1.095597505569458, + "learning_rate": 1.496426284391817e-05, + "loss": 3.6609, + "step": 112100 + }, + { + "epoch": 1.2599495570797032, + "grad_norm": 1.269573450088501, + "learning_rate": 1.4962015930615319e-05, + "loss": 3.659, + "step": 112150 + }, + { + "epoch": 1.2605112822500464, + "grad_norm": 1.1126152276992798, + "learning_rate": 1.4959769017312468e-05, + "loss": 3.6726, + "step": 112200 + }, + { + "epoch": 1.2610730074203895, + "grad_norm": 1.1701678037643433, + "learning_rate": 1.4957522104009619e-05, + "loss": 3.6442, + "step": 112250 + }, + { + "epoch": 1.2616347325907327, + "grad_norm": 1.4589455127716064, + "learning_rate": 1.4955275190706768e-05, + "loss": 3.7182, + "step": 112300 + }, + { + "epoch": 1.2621964577610758, + "grad_norm": 1.2170416116714478, + "learning_rate": 1.4953028277403918e-05, + "loss": 3.6852, + "step": 112350 + }, + { + "epoch": 1.262758182931419, + "grad_norm": 0.9736093282699585, + "learning_rate": 1.4950781364101069e-05, + "loss": 3.6293, + "step": 112400 + }, + { + "epoch": 1.263319908101762, + "grad_norm": 1.2244106531143188, + "learning_rate": 1.4948534450798218e-05, + "loss": 3.669, + "step": 112450 + }, + { + "epoch": 1.2638816332721052, + "grad_norm": 1.2522999048233032, + "learning_rate": 1.4946287537495367e-05, + "loss": 3.7332, + "step": 112500 + }, + { + "epoch": 1.2644433584424484, + "grad_norm": 1.381110429763794, + "learning_rate": 1.4944040624192517e-05, + "loss": 3.6822, + "step": 112550 + }, + { + "epoch": 1.2650050836127917, + "grad_norm": 1.1141459941864014, + "learning_rate": 1.4941793710889668e-05, + "loss": 3.7492, + "step": 112600 + }, + { + "epoch": 1.2655668087831349, + "grad_norm": 0.9814095497131348, + "learning_rate": 1.4939546797586817e-05, + "loss": 3.7012, + "step": 112650 + }, + { + "epoch": 1.266128533953478, + "grad_norm": 1.2120487689971924, + "learning_rate": 1.4937299884283966e-05, + "loss": 3.6702, + "step": 112700 + }, + { + "epoch": 1.2666902591238212, + "grad_norm": 1.2875162363052368, + "learning_rate": 1.4935052970981117e-05, + "loss": 3.6214, + "step": 112750 + }, + { + "epoch": 1.2672519842941643, + "grad_norm": 1.1989301443099976, + "learning_rate": 1.4932806057678267e-05, + "loss": 3.6967, + "step": 112800 + }, + { + "epoch": 1.2678137094645074, + "grad_norm": 1.0975011587142944, + "learning_rate": 1.4930559144375416e-05, + "loss": 3.71, + "step": 112850 + }, + { + "epoch": 1.2683754346348506, + "grad_norm": 1.4478185176849365, + "learning_rate": 1.4928312231072565e-05, + "loss": 3.6802, + "step": 112900 + }, + { + "epoch": 1.2689371598051937, + "grad_norm": 1.5076407194137573, + "learning_rate": 1.4926065317769716e-05, + "loss": 3.7218, + "step": 112950 + }, + { + "epoch": 1.2694988849755369, + "grad_norm": 1.088319182395935, + "learning_rate": 1.4923818404466866e-05, + "loss": 3.7645, + "step": 113000 + }, + { + "epoch": 1.27006061014588, + "grad_norm": 1.0324077606201172, + "learning_rate": 1.4921571491164015e-05, + "loss": 3.6841, + "step": 113050 + }, + { + "epoch": 1.2706223353162232, + "grad_norm": 1.0946040153503418, + "learning_rate": 1.4919324577861166e-05, + "loss": 3.7037, + "step": 113100 + }, + { + "epoch": 1.2711840604865663, + "grad_norm": 1.1757967472076416, + "learning_rate": 1.4917077664558315e-05, + "loss": 3.8264, + "step": 113150 + }, + { + "epoch": 1.2717457856569094, + "grad_norm": 1.2199627161026, + "learning_rate": 1.4914830751255464e-05, + "loss": 3.7236, + "step": 113200 + }, + { + "epoch": 1.2723075108272526, + "grad_norm": 1.1769957542419434, + "learning_rate": 1.4912583837952614e-05, + "loss": 3.7198, + "step": 113250 + }, + { + "epoch": 1.2728692359975957, + "grad_norm": 1.118649959564209, + "learning_rate": 1.4910336924649765e-05, + "loss": 3.7062, + "step": 113300 + }, + { + "epoch": 1.2734309611679389, + "grad_norm": 1.1054401397705078, + "learning_rate": 1.4908090011346914e-05, + "loss": 3.7428, + "step": 113350 + }, + { + "epoch": 1.273992686338282, + "grad_norm": 1.1560803651809692, + "learning_rate": 1.4905843098044063e-05, + "loss": 3.7675, + "step": 113400 + }, + { + "epoch": 1.2745544115086254, + "grad_norm": 1.1944975852966309, + "learning_rate": 1.4903596184741214e-05, + "loss": 3.693, + "step": 113450 + }, + { + "epoch": 1.2751161366789685, + "grad_norm": 1.030148983001709, + "learning_rate": 1.4901349271438364e-05, + "loss": 3.713, + "step": 113500 + }, + { + "epoch": 1.2756778618493116, + "grad_norm": 1.2132092714309692, + "learning_rate": 1.4899102358135513e-05, + "loss": 3.6853, + "step": 113550 + }, + { + "epoch": 1.2762395870196548, + "grad_norm": 1.1267964839935303, + "learning_rate": 1.4896855444832662e-05, + "loss": 3.8054, + "step": 113600 + }, + { + "epoch": 1.276801312189998, + "grad_norm": 1.0857360363006592, + "learning_rate": 1.4894608531529813e-05, + "loss": 3.7694, + "step": 113650 + }, + { + "epoch": 1.277363037360341, + "grad_norm": 1.1156096458435059, + "learning_rate": 1.4892361618226963e-05, + "loss": 3.6494, + "step": 113700 + }, + { + "epoch": 1.2779247625306842, + "grad_norm": 1.1470805406570435, + "learning_rate": 1.4890114704924112e-05, + "loss": 3.7046, + "step": 113750 + }, + { + "epoch": 1.2784864877010274, + "grad_norm": 1.1375935077667236, + "learning_rate": 1.4887867791621261e-05, + "loss": 3.655, + "step": 113800 + }, + { + "epoch": 1.2790482128713705, + "grad_norm": 1.490735411643982, + "learning_rate": 1.4885620878318412e-05, + "loss": 3.7503, + "step": 113850 + }, + { + "epoch": 1.2796099380417136, + "grad_norm": 1.081573486328125, + "learning_rate": 1.4883373965015562e-05, + "loss": 3.6072, + "step": 113900 + }, + { + "epoch": 1.280171663212057, + "grad_norm": 1.0996750593185425, + "learning_rate": 1.4881127051712711e-05, + "loss": 3.6634, + "step": 113950 + }, + { + "epoch": 1.2807333883824001, + "grad_norm": 1.2041651010513306, + "learning_rate": 1.4878880138409862e-05, + "loss": 3.6514, + "step": 114000 + }, + { + "epoch": 1.2812951135527433, + "grad_norm": 1.0212005376815796, + "learning_rate": 1.4876633225107011e-05, + "loss": 3.7077, + "step": 114050 + }, + { + "epoch": 1.2818568387230864, + "grad_norm": 1.223238229751587, + "learning_rate": 1.487438631180416e-05, + "loss": 3.6684, + "step": 114100 + }, + { + "epoch": 1.2824185638934296, + "grad_norm": 1.146500587463379, + "learning_rate": 1.487213939850131e-05, + "loss": 3.6591, + "step": 114150 + }, + { + "epoch": 1.2829802890637727, + "grad_norm": 1.1014022827148438, + "learning_rate": 1.4869892485198461e-05, + "loss": 3.7355, + "step": 114200 + }, + { + "epoch": 1.2835420142341158, + "grad_norm": 1.3315798044204712, + "learning_rate": 1.486764557189561e-05, + "loss": 3.7325, + "step": 114250 + }, + { + "epoch": 1.284103739404459, + "grad_norm": 1.263845682144165, + "learning_rate": 1.486539865859276e-05, + "loss": 3.6997, + "step": 114300 + }, + { + "epoch": 1.2846654645748021, + "grad_norm": 1.3485934734344482, + "learning_rate": 1.4863196683555967e-05, + "loss": 3.7422, + "step": 114350 + }, + { + "epoch": 1.2852271897451453, + "grad_norm": 1.086905598640442, + "learning_rate": 1.4860949770253117e-05, + "loss": 3.7186, + "step": 114400 + }, + { + "epoch": 1.2857889149154884, + "grad_norm": 1.1574716567993164, + "learning_rate": 1.4858702856950268e-05, + "loss": 3.698, + "step": 114450 + }, + { + "epoch": 1.2863506400858316, + "grad_norm": 5.0376482009887695, + "learning_rate": 1.4856455943647417e-05, + "loss": 3.6421, + "step": 114500 + }, + { + "epoch": 1.2869123652561747, + "grad_norm": 1.0967539548873901, + "learning_rate": 1.4854209030344566e-05, + "loss": 3.7413, + "step": 114550 + }, + { + "epoch": 1.2874740904265178, + "grad_norm": 1.210237741470337, + "learning_rate": 1.4851962117041715e-05, + "loss": 3.7403, + "step": 114600 + }, + { + "epoch": 1.288035815596861, + "grad_norm": 1.0677835941314697, + "learning_rate": 1.4849715203738867e-05, + "loss": 3.7031, + "step": 114650 + }, + { + "epoch": 1.2885975407672041, + "grad_norm": 1.0944105386734009, + "learning_rate": 1.4847468290436016e-05, + "loss": 3.6474, + "step": 114700 + }, + { + "epoch": 1.2891592659375473, + "grad_norm": 1.1223304271697998, + "learning_rate": 1.4845221377133165e-05, + "loss": 3.6784, + "step": 114750 + }, + { + "epoch": 1.2897209911078906, + "grad_norm": 1.1143394708633423, + "learning_rate": 1.4842974463830314e-05, + "loss": 3.7114, + "step": 114800 + }, + { + "epoch": 1.2902827162782338, + "grad_norm": 1.536241888999939, + "learning_rate": 1.4840727550527465e-05, + "loss": 3.6856, + "step": 114850 + }, + { + "epoch": 1.290844441448577, + "grad_norm": 0.9196121096611023, + "learning_rate": 1.4838480637224615e-05, + "loss": 3.7058, + "step": 114900 + }, + { + "epoch": 1.29140616661892, + "grad_norm": 1.2177156209945679, + "learning_rate": 1.4836233723921764e-05, + "loss": 3.6414, + "step": 114950 + }, + { + "epoch": 1.2919678917892632, + "grad_norm": 1.143907904624939, + "learning_rate": 1.4833986810618915e-05, + "loss": 3.7765, + "step": 115000 + }, + { + "epoch": 1.2925296169596063, + "grad_norm": 1.2396646738052368, + "learning_rate": 1.4831739897316064e-05, + "loss": 3.7571, + "step": 115050 + }, + { + "epoch": 1.2930913421299495, + "grad_norm": 1.1673306226730347, + "learning_rate": 1.4829492984013214e-05, + "loss": 3.704, + "step": 115100 + }, + { + "epoch": 1.2936530673002926, + "grad_norm": 1.3346099853515625, + "learning_rate": 1.4827246070710363e-05, + "loss": 3.66, + "step": 115150 + }, + { + "epoch": 1.2942147924706358, + "grad_norm": 1.3197540044784546, + "learning_rate": 1.4824999157407514e-05, + "loss": 3.6604, + "step": 115200 + }, + { + "epoch": 1.294776517640979, + "grad_norm": 1.0987499952316284, + "learning_rate": 1.4822752244104663e-05, + "loss": 3.6903, + "step": 115250 + }, + { + "epoch": 1.2953382428113223, + "grad_norm": 1.1738516092300415, + "learning_rate": 1.4820505330801813e-05, + "loss": 3.7153, + "step": 115300 + }, + { + "epoch": 1.2958999679816654, + "grad_norm": 1.0726885795593262, + "learning_rate": 1.4818258417498964e-05, + "loss": 3.7386, + "step": 115350 + }, + { + "epoch": 1.2964616931520085, + "grad_norm": 1.132994294166565, + "learning_rate": 1.4816011504196113e-05, + "loss": 3.5907, + "step": 115400 + }, + { + "epoch": 1.2970234183223517, + "grad_norm": 0.9900727868080139, + "learning_rate": 1.4813764590893262e-05, + "loss": 3.7105, + "step": 115450 + }, + { + "epoch": 1.2975851434926948, + "grad_norm": 1.0517420768737793, + "learning_rate": 1.4811517677590412e-05, + "loss": 3.7059, + "step": 115500 + }, + { + "epoch": 1.298146868663038, + "grad_norm": 1.2015070915222168, + "learning_rate": 1.4809270764287563e-05, + "loss": 3.7225, + "step": 115550 + }, + { + "epoch": 1.298708593833381, + "grad_norm": 1.1541849374771118, + "learning_rate": 1.4807023850984712e-05, + "loss": 3.6733, + "step": 115600 + }, + { + "epoch": 1.2992703190037242, + "grad_norm": 1.1331740617752075, + "learning_rate": 1.4804776937681861e-05, + "loss": 3.66, + "step": 115650 + }, + { + "epoch": 1.2998320441740674, + "grad_norm": 1.0337045192718506, + "learning_rate": 1.4802530024379012e-05, + "loss": 3.7144, + "step": 115700 + }, + { + "epoch": 1.3003937693444105, + "grad_norm": 1.3465909957885742, + "learning_rate": 1.4800283111076162e-05, + "loss": 3.7363, + "step": 115750 + }, + { + "epoch": 1.3009554945147537, + "grad_norm": 1.2244303226470947, + "learning_rate": 1.4798036197773311e-05, + "loss": 3.7488, + "step": 115800 + }, + { + "epoch": 1.3015172196850968, + "grad_norm": 1.0829503536224365, + "learning_rate": 1.479578928447046e-05, + "loss": 3.7574, + "step": 115850 + }, + { + "epoch": 1.30207894485544, + "grad_norm": 1.0618500709533691, + "learning_rate": 1.4793542371167611e-05, + "loss": 3.7326, + "step": 115900 + }, + { + "epoch": 1.302640670025783, + "grad_norm": 1.1596113443374634, + "learning_rate": 1.479129545786476e-05, + "loss": 3.6051, + "step": 115950 + }, + { + "epoch": 1.3032023951961262, + "grad_norm": 1.2710692882537842, + "learning_rate": 1.478904854456191e-05, + "loss": 3.6918, + "step": 116000 + }, + { + "epoch": 1.3037641203664694, + "grad_norm": 1.3042854070663452, + "learning_rate": 1.4786801631259061e-05, + "loss": 3.6836, + "step": 116050 + }, + { + "epoch": 1.3043258455368125, + "grad_norm": 1.170735478401184, + "learning_rate": 1.478455471795621e-05, + "loss": 3.7783, + "step": 116100 + }, + { + "epoch": 1.3048875707071559, + "grad_norm": 1.127317190170288, + "learning_rate": 1.478230780465336e-05, + "loss": 3.7091, + "step": 116150 + }, + { + "epoch": 1.305449295877499, + "grad_norm": 1.0851610898971558, + "learning_rate": 1.4780060891350509e-05, + "loss": 3.7688, + "step": 116200 + }, + { + "epoch": 1.3060110210478422, + "grad_norm": 1.5472756624221802, + "learning_rate": 1.477781397804766e-05, + "loss": 3.6622, + "step": 116250 + }, + { + "epoch": 1.3065727462181853, + "grad_norm": 1.0288634300231934, + "learning_rate": 1.477556706474481e-05, + "loss": 3.7608, + "step": 116300 + }, + { + "epoch": 1.3071344713885285, + "grad_norm": 1.2104113101959229, + "learning_rate": 1.4773320151441959e-05, + "loss": 3.7222, + "step": 116350 + }, + { + "epoch": 1.3076961965588716, + "grad_norm": 1.206800103187561, + "learning_rate": 1.4771073238139108e-05, + "loss": 3.672, + "step": 116400 + }, + { + "epoch": 1.3082579217292147, + "grad_norm": 1.1641842126846313, + "learning_rate": 1.4768826324836259e-05, + "loss": 3.7283, + "step": 116450 + }, + { + "epoch": 1.3088196468995579, + "grad_norm": 1.1273703575134277, + "learning_rate": 1.4766579411533408e-05, + "loss": 3.716, + "step": 116500 + }, + { + "epoch": 1.309381372069901, + "grad_norm": 1.2490730285644531, + "learning_rate": 1.4764332498230557e-05, + "loss": 3.735, + "step": 116550 + }, + { + "epoch": 1.3099430972402442, + "grad_norm": 1.1901054382324219, + "learning_rate": 1.4762085584927708e-05, + "loss": 3.6776, + "step": 116600 + }, + { + "epoch": 1.3105048224105875, + "grad_norm": 1.13470458984375, + "learning_rate": 1.4759838671624858e-05, + "loss": 3.6582, + "step": 116650 + }, + { + "epoch": 1.3110665475809307, + "grad_norm": 1.322670817375183, + "learning_rate": 1.4757591758322007e-05, + "loss": 3.7346, + "step": 116700 + }, + { + "epoch": 1.3116282727512738, + "grad_norm": 0.9900648593902588, + "learning_rate": 1.4755344845019156e-05, + "loss": 3.6795, + "step": 116750 + }, + { + "epoch": 1.312189997921617, + "grad_norm": 1.1879056692123413, + "learning_rate": 1.4753097931716307e-05, + "loss": 3.7087, + "step": 116800 + }, + { + "epoch": 1.31275172309196, + "grad_norm": 0.9856444597244263, + "learning_rate": 1.4750851018413457e-05, + "loss": 3.7151, + "step": 116850 + }, + { + "epoch": 1.3133134482623032, + "grad_norm": 1.1498546600341797, + "learning_rate": 1.4748604105110606e-05, + "loss": 3.6927, + "step": 116900 + }, + { + "epoch": 1.3138751734326464, + "grad_norm": 1.1568413972854614, + "learning_rate": 1.4746357191807757e-05, + "loss": 3.6835, + "step": 116950 + }, + { + "epoch": 1.3144368986029895, + "grad_norm": 1.1762259006500244, + "learning_rate": 1.4744110278504906e-05, + "loss": 3.7163, + "step": 117000 + }, + { + "epoch": 1.3149986237733327, + "grad_norm": 1.0447081327438354, + "learning_rate": 1.4741863365202056e-05, + "loss": 3.6836, + "step": 117050 + }, + { + "epoch": 1.3155603489436758, + "grad_norm": 1.3564331531524658, + "learning_rate": 1.4739616451899205e-05, + "loss": 3.7426, + "step": 117100 + }, + { + "epoch": 1.316122074114019, + "grad_norm": 1.373300313949585, + "learning_rate": 1.4737369538596356e-05, + "loss": 3.7362, + "step": 117150 + }, + { + "epoch": 1.316683799284362, + "grad_norm": 1.2786941528320312, + "learning_rate": 1.4735122625293505e-05, + "loss": 3.6725, + "step": 117200 + }, + { + "epoch": 1.3172455244547052, + "grad_norm": 1.2043522596359253, + "learning_rate": 1.4732875711990655e-05, + "loss": 3.6457, + "step": 117250 + }, + { + "epoch": 1.3178072496250484, + "grad_norm": 1.0595104694366455, + "learning_rate": 1.4730628798687806e-05, + "loss": 3.6185, + "step": 117300 + }, + { + "epoch": 1.3183689747953915, + "grad_norm": 1.272301435470581, + "learning_rate": 1.4728381885384955e-05, + "loss": 3.7125, + "step": 117350 + }, + { + "epoch": 1.3189306999657346, + "grad_norm": 1.266560435295105, + "learning_rate": 1.4726134972082104e-05, + "loss": 3.7035, + "step": 117400 + }, + { + "epoch": 1.3194924251360778, + "grad_norm": 1.0973302125930786, + "learning_rate": 1.4723888058779254e-05, + "loss": 3.7357, + "step": 117450 + }, + { + "epoch": 1.3200541503064211, + "grad_norm": 1.0538254976272583, + "learning_rate": 1.4721641145476405e-05, + "loss": 3.7578, + "step": 117500 + }, + { + "epoch": 1.3206158754767643, + "grad_norm": 1.1436649560928345, + "learning_rate": 1.4719394232173552e-05, + "loss": 3.6825, + "step": 117550 + }, + { + "epoch": 1.3211776006471074, + "grad_norm": 1.0678046941757202, + "learning_rate": 1.4717147318870702e-05, + "loss": 3.7167, + "step": 117600 + }, + { + "epoch": 1.3217393258174506, + "grad_norm": 1.199366569519043, + "learning_rate": 1.4714900405567851e-05, + "loss": 3.7284, + "step": 117650 + }, + { + "epoch": 1.3223010509877937, + "grad_norm": 1.2246750593185425, + "learning_rate": 1.4712653492265e-05, + "loss": 3.7814, + "step": 117700 + }, + { + "epoch": 1.3228627761581369, + "grad_norm": 1.0405752658843994, + "learning_rate": 1.4710406578962151e-05, + "loss": 3.7945, + "step": 117750 + }, + { + "epoch": 1.32342450132848, + "grad_norm": 1.1649138927459717, + "learning_rate": 1.47081596656593e-05, + "loss": 3.7168, + "step": 117800 + }, + { + "epoch": 1.3239862264988231, + "grad_norm": 1.1335917711257935, + "learning_rate": 1.470591275235645e-05, + "loss": 3.6777, + "step": 117850 + }, + { + "epoch": 1.3245479516691663, + "grad_norm": 1.229051113128662, + "learning_rate": 1.4703665839053601e-05, + "loss": 3.6618, + "step": 117900 + }, + { + "epoch": 1.3251096768395094, + "grad_norm": 1.164178729057312, + "learning_rate": 1.470141892575075e-05, + "loss": 3.6298, + "step": 117950 + }, + { + "epoch": 1.3256714020098528, + "grad_norm": 1.0775868892669678, + "learning_rate": 1.46991720124479e-05, + "loss": 3.7261, + "step": 118000 + }, + { + "epoch": 1.326233127180196, + "grad_norm": 1.1860066652297974, + "learning_rate": 1.4696925099145049e-05, + "loss": 3.7335, + "step": 118050 + }, + { + "epoch": 1.326794852350539, + "grad_norm": 1.2197784185409546, + "learning_rate": 1.46946781858422e-05, + "loss": 3.7328, + "step": 118100 + }, + { + "epoch": 1.3273565775208822, + "grad_norm": 1.1759244203567505, + "learning_rate": 1.469243127253935e-05, + "loss": 3.6612, + "step": 118150 + }, + { + "epoch": 1.3279183026912253, + "grad_norm": 1.2106807231903076, + "learning_rate": 1.4690184359236498e-05, + "loss": 3.6147, + "step": 118200 + }, + { + "epoch": 1.3284800278615685, + "grad_norm": 1.2478992938995361, + "learning_rate": 1.468793744593365e-05, + "loss": 3.6817, + "step": 118250 + }, + { + "epoch": 1.3290417530319116, + "grad_norm": 1.6169747114181519, + "learning_rate": 1.4685690532630799e-05, + "loss": 3.6878, + "step": 118300 + }, + { + "epoch": 1.3296034782022548, + "grad_norm": 1.3529175519943237, + "learning_rate": 1.4683488557594006e-05, + "loss": 3.8043, + "step": 118350 + }, + { + "epoch": 1.330165203372598, + "grad_norm": 0.9039379358291626, + "learning_rate": 1.4681241644291156e-05, + "loss": 3.7257, + "step": 118400 + }, + { + "epoch": 1.330726928542941, + "grad_norm": 1.121334433555603, + "learning_rate": 1.4678994730988305e-05, + "loss": 3.7019, + "step": 118450 + }, + { + "epoch": 1.3312886537132842, + "grad_norm": 1.0345427989959717, + "learning_rate": 1.4676747817685454e-05, + "loss": 3.6703, + "step": 118500 + }, + { + "epoch": 1.3318503788836273, + "grad_norm": 1.2660282850265503, + "learning_rate": 1.4674500904382605e-05, + "loss": 3.6788, + "step": 118550 + }, + { + "epoch": 1.3324121040539705, + "grad_norm": 1.0917772054672241, + "learning_rate": 1.4672253991079755e-05, + "loss": 3.68, + "step": 118600 + }, + { + "epoch": 1.3329738292243136, + "grad_norm": 1.1843641996383667, + "learning_rate": 1.4670007077776904e-05, + "loss": 3.6719, + "step": 118650 + }, + { + "epoch": 1.3335355543946568, + "grad_norm": 1.2436368465423584, + "learning_rate": 1.4667760164474053e-05, + "loss": 3.7044, + "step": 118700 + }, + { + "epoch": 1.334097279565, + "grad_norm": 1.1449768543243408, + "learning_rate": 1.4665513251171204e-05, + "loss": 3.7275, + "step": 118750 + }, + { + "epoch": 1.3346590047353433, + "grad_norm": 1.2294001579284668, + "learning_rate": 1.4663266337868354e-05, + "loss": 3.6284, + "step": 118800 + }, + { + "epoch": 1.3352207299056864, + "grad_norm": 1.0378721952438354, + "learning_rate": 1.4661019424565503e-05, + "loss": 3.7391, + "step": 118850 + }, + { + "epoch": 1.3357824550760296, + "grad_norm": 0.9750446081161499, + "learning_rate": 1.4658772511262654e-05, + "loss": 3.7817, + "step": 118900 + }, + { + "epoch": 1.3363441802463727, + "grad_norm": 1.2073137760162354, + "learning_rate": 1.4656525597959803e-05, + "loss": 3.7439, + "step": 118950 + }, + { + "epoch": 1.3369059054167158, + "grad_norm": 1.1098533868789673, + "learning_rate": 1.4654278684656953e-05, + "loss": 3.7667, + "step": 119000 + }, + { + "epoch": 1.337467630587059, + "grad_norm": 1.1253433227539062, + "learning_rate": 1.4652031771354102e-05, + "loss": 3.673, + "step": 119050 + }, + { + "epoch": 1.3380293557574021, + "grad_norm": 1.296921968460083, + "learning_rate": 1.4649784858051253e-05, + "loss": 3.7251, + "step": 119100 + }, + { + "epoch": 1.3385910809277453, + "grad_norm": 0.9098636507987976, + "learning_rate": 1.4647537944748402e-05, + "loss": 3.6904, + "step": 119150 + }, + { + "epoch": 1.3391528060980884, + "grad_norm": 1.0598982572555542, + "learning_rate": 1.4645291031445552e-05, + "loss": 3.65, + "step": 119200 + }, + { + "epoch": 1.3397145312684315, + "grad_norm": 0.9995057582855225, + "learning_rate": 1.4643044118142703e-05, + "loss": 3.6988, + "step": 119250 + }, + { + "epoch": 1.3402762564387747, + "grad_norm": 1.265258550643921, + "learning_rate": 1.4640797204839852e-05, + "loss": 3.6759, + "step": 119300 + }, + { + "epoch": 1.340837981609118, + "grad_norm": 1.3248783349990845, + "learning_rate": 1.4638550291537001e-05, + "loss": 3.5958, + "step": 119350 + }, + { + "epoch": 1.3413997067794612, + "grad_norm": 1.2543925046920776, + "learning_rate": 1.463630337823415e-05, + "loss": 3.6777, + "step": 119400 + }, + { + "epoch": 1.3419614319498043, + "grad_norm": 1.18267023563385, + "learning_rate": 1.4634056464931302e-05, + "loss": 3.6751, + "step": 119450 + }, + { + "epoch": 1.3425231571201475, + "grad_norm": 1.44830322265625, + "learning_rate": 1.4631809551628451e-05, + "loss": 3.6802, + "step": 119500 + }, + { + "epoch": 1.3430848822904906, + "grad_norm": 1.311497449874878, + "learning_rate": 1.46295626383256e-05, + "loss": 3.7528, + "step": 119550 + }, + { + "epoch": 1.3436466074608338, + "grad_norm": 1.0765646696090698, + "learning_rate": 1.4627315725022751e-05, + "loss": 3.7024, + "step": 119600 + }, + { + "epoch": 1.344208332631177, + "grad_norm": 1.1783418655395508, + "learning_rate": 1.46250688117199e-05, + "loss": 3.7267, + "step": 119650 + }, + { + "epoch": 1.34477005780152, + "grad_norm": 1.1875866651535034, + "learning_rate": 1.462282189841705e-05, + "loss": 3.6513, + "step": 119700 + }, + { + "epoch": 1.3453317829718632, + "grad_norm": 1.1184399127960205, + "learning_rate": 1.4620574985114199e-05, + "loss": 3.75, + "step": 119750 + }, + { + "epoch": 1.3458935081422063, + "grad_norm": 1.2593485116958618, + "learning_rate": 1.461832807181135e-05, + "loss": 3.6672, + "step": 119800 + }, + { + "epoch": 1.3464552333125495, + "grad_norm": 1.0915449857711792, + "learning_rate": 1.46160811585085e-05, + "loss": 3.656, + "step": 119850 + }, + { + "epoch": 1.3470169584828926, + "grad_norm": 1.3104792833328247, + "learning_rate": 1.4613834245205649e-05, + "loss": 3.686, + "step": 119900 + }, + { + "epoch": 1.3475786836532357, + "grad_norm": 1.1535950899124146, + "learning_rate": 1.46115873319028e-05, + "loss": 3.7439, + "step": 119950 + }, + { + "epoch": 1.3481404088235789, + "grad_norm": 1.2658485174179077, + "learning_rate": 1.4609340418599949e-05, + "loss": 3.7157, + "step": 120000 + }, + { + "epoch": 1.348702133993922, + "grad_norm": 1.2296286821365356, + "learning_rate": 1.4607093505297098e-05, + "loss": 3.6688, + "step": 120050 + }, + { + "epoch": 1.3492638591642652, + "grad_norm": 1.166917085647583, + "learning_rate": 1.4604846591994248e-05, + "loss": 3.6548, + "step": 120100 + }, + { + "epoch": 1.3498255843346085, + "grad_norm": 1.2615032196044922, + "learning_rate": 1.4602599678691399e-05, + "loss": 3.6502, + "step": 120150 + }, + { + "epoch": 1.3503873095049517, + "grad_norm": 1.0976001024246216, + "learning_rate": 1.4600352765388548e-05, + "loss": 3.6865, + "step": 120200 + }, + { + "epoch": 1.3509490346752948, + "grad_norm": 1.0847792625427246, + "learning_rate": 1.4598105852085697e-05, + "loss": 3.677, + "step": 120250 + }, + { + "epoch": 1.351510759845638, + "grad_norm": 1.1527408361434937, + "learning_rate": 1.4595858938782847e-05, + "loss": 3.6671, + "step": 120300 + }, + { + "epoch": 1.352072485015981, + "grad_norm": 1.0550224781036377, + "learning_rate": 1.4593612025479998e-05, + "loss": 3.6829, + "step": 120350 + }, + { + "epoch": 1.3526342101863242, + "grad_norm": 1.3266491889953613, + "learning_rate": 1.4591365112177147e-05, + "loss": 3.6975, + "step": 120400 + }, + { + "epoch": 1.3531959353566674, + "grad_norm": 1.0050033330917358, + "learning_rate": 1.4589118198874296e-05, + "loss": 3.6553, + "step": 120450 + }, + { + "epoch": 1.3537576605270105, + "grad_norm": 1.0242691040039062, + "learning_rate": 1.4586916223837504e-05, + "loss": 3.7015, + "step": 120500 + }, + { + "epoch": 1.3543193856973537, + "grad_norm": 1.1202960014343262, + "learning_rate": 1.4584669310534653e-05, + "loss": 3.6946, + "step": 120550 + }, + { + "epoch": 1.3548811108676968, + "grad_norm": 0.9728813767433167, + "learning_rate": 1.4582422397231804e-05, + "loss": 3.6664, + "step": 120600 + }, + { + "epoch": 1.3554428360380402, + "grad_norm": 1.2763217687606812, + "learning_rate": 1.4580175483928954e-05, + "loss": 3.7196, + "step": 120650 + }, + { + "epoch": 1.3560045612083833, + "grad_norm": 0.9913472533226013, + "learning_rate": 1.4577928570626103e-05, + "loss": 3.68, + "step": 120700 + }, + { + "epoch": 1.3565662863787264, + "grad_norm": 1.1339112520217896, + "learning_rate": 1.4575681657323252e-05, + "loss": 3.6267, + "step": 120750 + }, + { + "epoch": 1.3571280115490696, + "grad_norm": 1.1435879468917847, + "learning_rate": 1.4573434744020403e-05, + "loss": 3.6202, + "step": 120800 + }, + { + "epoch": 1.3576897367194127, + "grad_norm": 1.1710400581359863, + "learning_rate": 1.4571187830717553e-05, + "loss": 3.7836, + "step": 120850 + }, + { + "epoch": 1.3582514618897559, + "grad_norm": 1.1049959659576416, + "learning_rate": 1.4568940917414702e-05, + "loss": 3.6587, + "step": 120900 + }, + { + "epoch": 1.358813187060099, + "grad_norm": 0.995599091053009, + "learning_rate": 1.4566694004111853e-05, + "loss": 3.6778, + "step": 120950 + }, + { + "epoch": 1.3593749122304422, + "grad_norm": 1.1417511701583862, + "learning_rate": 1.4564536967341117e-05, + "loss": 3.7144, + "step": 121000 + }, + { + "epoch": 1.3599366374007853, + "grad_norm": 1.1375656127929688, + "learning_rate": 1.4562290054038266e-05, + "loss": 3.7073, + "step": 121050 + }, + { + "epoch": 1.3604983625711284, + "grad_norm": 1.0259642601013184, + "learning_rate": 1.4560043140735416e-05, + "loss": 3.6656, + "step": 121100 + }, + { + "epoch": 1.3610600877414716, + "grad_norm": 1.1820317506790161, + "learning_rate": 1.4557796227432565e-05, + "loss": 3.6812, + "step": 121150 + }, + { + "epoch": 1.3616218129118147, + "grad_norm": 1.1481682062149048, + "learning_rate": 1.4555549314129716e-05, + "loss": 3.7106, + "step": 121200 + }, + { + "epoch": 1.3621835380821579, + "grad_norm": 1.3173041343688965, + "learning_rate": 1.4553302400826865e-05, + "loss": 3.6936, + "step": 121250 + }, + { + "epoch": 1.362745263252501, + "grad_norm": 1.0805565118789673, + "learning_rate": 1.4551055487524015e-05, + "loss": 3.6706, + "step": 121300 + }, + { + "epoch": 1.3633069884228441, + "grad_norm": 1.17776620388031, + "learning_rate": 1.4548808574221166e-05, + "loss": 3.7006, + "step": 121350 + }, + { + "epoch": 1.3638687135931873, + "grad_norm": 1.0657659769058228, + "learning_rate": 1.4546561660918315e-05, + "loss": 3.748, + "step": 121400 + }, + { + "epoch": 1.3644304387635304, + "grad_norm": 1.0460538864135742, + "learning_rate": 1.4544314747615464e-05, + "loss": 3.704, + "step": 121450 + }, + { + "epoch": 1.3649921639338738, + "grad_norm": 1.2596518993377686, + "learning_rate": 1.4542067834312614e-05, + "loss": 3.6839, + "step": 121500 + }, + { + "epoch": 1.365553889104217, + "grad_norm": 1.397570252418518, + "learning_rate": 1.4539820921009765e-05, + "loss": 3.7181, + "step": 121550 + }, + { + "epoch": 1.36611561427456, + "grad_norm": 1.280348300933838, + "learning_rate": 1.4537574007706914e-05, + "loss": 3.7099, + "step": 121600 + }, + { + "epoch": 1.3666773394449032, + "grad_norm": 1.201006531715393, + "learning_rate": 1.4535327094404063e-05, + "loss": 3.7533, + "step": 121650 + }, + { + "epoch": 1.3672390646152464, + "grad_norm": 1.2775695323944092, + "learning_rate": 1.4533080181101213e-05, + "loss": 3.6657, + "step": 121700 + }, + { + "epoch": 1.3678007897855895, + "grad_norm": 1.1215566396713257, + "learning_rate": 1.4530833267798364e-05, + "loss": 3.6661, + "step": 121750 + }, + { + "epoch": 1.3683625149559326, + "grad_norm": 1.0805236101150513, + "learning_rate": 1.4528586354495513e-05, + "loss": 3.7707, + "step": 121800 + }, + { + "epoch": 1.3689242401262758, + "grad_norm": 1.1486221551895142, + "learning_rate": 1.4526339441192662e-05, + "loss": 3.7298, + "step": 121850 + }, + { + "epoch": 1.369485965296619, + "grad_norm": 1.1970341205596924, + "learning_rate": 1.4524092527889813e-05, + "loss": 3.7511, + "step": 121900 + }, + { + "epoch": 1.370047690466962, + "grad_norm": 1.014717936515808, + "learning_rate": 1.4521845614586963e-05, + "loss": 3.6929, + "step": 121950 + }, + { + "epoch": 1.3706094156373054, + "grad_norm": 1.0722380876541138, + "learning_rate": 1.4519598701284112e-05, + "loss": 3.6561, + "step": 122000 + }, + { + "epoch": 1.3711711408076486, + "grad_norm": 1.07088041305542, + "learning_rate": 1.4517351787981261e-05, + "loss": 3.8093, + "step": 122050 + }, + { + "epoch": 1.3717328659779917, + "grad_norm": 1.2771564722061157, + "learning_rate": 1.4515104874678412e-05, + "loss": 3.67, + "step": 122100 + }, + { + "epoch": 1.3722945911483349, + "grad_norm": 1.1219714879989624, + "learning_rate": 1.4512857961375562e-05, + "loss": 3.6666, + "step": 122150 + }, + { + "epoch": 1.372856316318678, + "grad_norm": 1.1300429105758667, + "learning_rate": 1.4510611048072711e-05, + "loss": 3.6496, + "step": 122200 + }, + { + "epoch": 1.3734180414890211, + "grad_norm": 2.0728774070739746, + "learning_rate": 1.4508364134769862e-05, + "loss": 3.775, + "step": 122250 + }, + { + "epoch": 1.3739797666593643, + "grad_norm": 1.1185555458068848, + "learning_rate": 1.4506117221467011e-05, + "loss": 3.629, + "step": 122300 + }, + { + "epoch": 1.3745414918297074, + "grad_norm": 1.1981322765350342, + "learning_rate": 1.450387030816416e-05, + "loss": 3.6022, + "step": 122350 + }, + { + "epoch": 1.3751032170000506, + "grad_norm": 1.2505836486816406, + "learning_rate": 1.450162339486131e-05, + "loss": 3.671, + "step": 122400 + }, + { + "epoch": 1.3756649421703937, + "grad_norm": 1.1908420324325562, + "learning_rate": 1.4499421419824517e-05, + "loss": 3.6314, + "step": 122450 + }, + { + "epoch": 1.3762266673407368, + "grad_norm": 1.2761627435684204, + "learning_rate": 1.4497174506521667e-05, + "loss": 3.7494, + "step": 122500 + }, + { + "epoch": 1.37678839251108, + "grad_norm": 1.2110997438430786, + "learning_rate": 1.4494927593218818e-05, + "loss": 3.6841, + "step": 122550 + }, + { + "epoch": 1.3773501176814231, + "grad_norm": 1.1500811576843262, + "learning_rate": 1.4492680679915967e-05, + "loss": 3.6617, + "step": 122600 + }, + { + "epoch": 1.3779118428517663, + "grad_norm": 1.080535888671875, + "learning_rate": 1.4490433766613116e-05, + "loss": 3.7118, + "step": 122650 + }, + { + "epoch": 1.3784735680221094, + "grad_norm": 1.1847634315490723, + "learning_rate": 1.4488186853310266e-05, + "loss": 3.6507, + "step": 122700 + }, + { + "epoch": 1.3790352931924525, + "grad_norm": 1.1334607601165771, + "learning_rate": 1.4485939940007417e-05, + "loss": 3.6664, + "step": 122750 + }, + { + "epoch": 1.3795970183627957, + "grad_norm": 1.0726972818374634, + "learning_rate": 1.4483693026704566e-05, + "loss": 3.7032, + "step": 122800 + }, + { + "epoch": 1.380158743533139, + "grad_norm": 1.2592869997024536, + "learning_rate": 1.4481446113401715e-05, + "loss": 3.6406, + "step": 122850 + }, + { + "epoch": 1.3807204687034822, + "grad_norm": 1.1625162363052368, + "learning_rate": 1.4479199200098866e-05, + "loss": 3.8139, + "step": 122900 + }, + { + "epoch": 1.3812821938738253, + "grad_norm": 1.1738134622573853, + "learning_rate": 1.4476952286796016e-05, + "loss": 3.7118, + "step": 122950 + }, + { + "epoch": 1.3818439190441685, + "grad_norm": 0.9934506416320801, + "learning_rate": 1.4474705373493165e-05, + "loss": 3.7588, + "step": 123000 + }, + { + "epoch": 1.3824056442145116, + "grad_norm": 1.1412187814712524, + "learning_rate": 1.4472458460190314e-05, + "loss": 3.7243, + "step": 123050 + }, + { + "epoch": 1.3829673693848548, + "grad_norm": 1.0966845750808716, + "learning_rate": 1.4470211546887465e-05, + "loss": 3.6771, + "step": 123100 + }, + { + "epoch": 1.383529094555198, + "grad_norm": 1.2062981128692627, + "learning_rate": 1.4467964633584615e-05, + "loss": 3.7103, + "step": 123150 + }, + { + "epoch": 1.384090819725541, + "grad_norm": 0.9867963194847107, + "learning_rate": 1.4465717720281764e-05, + "loss": 3.7591, + "step": 123200 + }, + { + "epoch": 1.3846525448958842, + "grad_norm": 1.0068353414535522, + "learning_rate": 1.4463470806978915e-05, + "loss": 3.6298, + "step": 123250 + }, + { + "epoch": 1.3852142700662273, + "grad_norm": 1.4179811477661133, + "learning_rate": 1.4461223893676064e-05, + "loss": 3.6966, + "step": 123300 + }, + { + "epoch": 1.3857759952365707, + "grad_norm": 1.0487943887710571, + "learning_rate": 1.4458976980373214e-05, + "loss": 3.736, + "step": 123350 + }, + { + "epoch": 1.3863377204069138, + "grad_norm": 1.301311731338501, + "learning_rate": 1.4456730067070363e-05, + "loss": 3.6252, + "step": 123400 + }, + { + "epoch": 1.386899445577257, + "grad_norm": 1.1093934774398804, + "learning_rate": 1.4454483153767514e-05, + "loss": 3.7113, + "step": 123450 + }, + { + "epoch": 1.3874611707476001, + "grad_norm": 1.0409111976623535, + "learning_rate": 1.4452236240464663e-05, + "loss": 3.6539, + "step": 123500 + }, + { + "epoch": 1.3880228959179433, + "grad_norm": 1.1051313877105713, + "learning_rate": 1.4449989327161813e-05, + "loss": 3.7126, + "step": 123550 + }, + { + "epoch": 1.3885846210882864, + "grad_norm": 1.1085457801818848, + "learning_rate": 1.4447742413858964e-05, + "loss": 3.692, + "step": 123600 + }, + { + "epoch": 1.3891463462586295, + "grad_norm": 1.0587685108184814, + "learning_rate": 1.4445495500556113e-05, + "loss": 3.7159, + "step": 123650 + }, + { + "epoch": 1.3897080714289727, + "grad_norm": 1.2765042781829834, + "learning_rate": 1.4443248587253262e-05, + "loss": 3.6629, + "step": 123700 + }, + { + "epoch": 1.3902697965993158, + "grad_norm": 1.0963952541351318, + "learning_rate": 1.4441001673950412e-05, + "loss": 3.6346, + "step": 123750 + }, + { + "epoch": 1.390831521769659, + "grad_norm": 1.1037063598632812, + "learning_rate": 1.4438754760647563e-05, + "loss": 3.6834, + "step": 123800 + }, + { + "epoch": 1.391393246940002, + "grad_norm": 1.283635139465332, + "learning_rate": 1.4436507847344712e-05, + "loss": 3.6974, + "step": 123850 + }, + { + "epoch": 1.3919549721103452, + "grad_norm": 1.2471383810043335, + "learning_rate": 1.4434260934041861e-05, + "loss": 3.7738, + "step": 123900 + }, + { + "epoch": 1.3925166972806884, + "grad_norm": 1.1486215591430664, + "learning_rate": 1.4432014020739012e-05, + "loss": 3.6628, + "step": 123950 + }, + { + "epoch": 1.3930784224510315, + "grad_norm": 1.244869351387024, + "learning_rate": 1.4429767107436162e-05, + "loss": 3.6546, + "step": 124000 + }, + { + "epoch": 1.3936401476213747, + "grad_norm": 1.3446252346038818, + "learning_rate": 1.442752019413331e-05, + "loss": 3.7014, + "step": 124050 + }, + { + "epoch": 1.3942018727917178, + "grad_norm": 1.0424251556396484, + "learning_rate": 1.442527328083046e-05, + "loss": 3.7377, + "step": 124100 + }, + { + "epoch": 1.394763597962061, + "grad_norm": 1.0342272520065308, + "learning_rate": 1.4423026367527611e-05, + "loss": 3.7084, + "step": 124150 + }, + { + "epoch": 1.3953253231324043, + "grad_norm": 1.1792279481887817, + "learning_rate": 1.442077945422476e-05, + "loss": 3.6867, + "step": 124200 + }, + { + "epoch": 1.3958870483027475, + "grad_norm": 1.029192328453064, + "learning_rate": 1.441853254092191e-05, + "loss": 3.6794, + "step": 124250 + }, + { + "epoch": 1.3964487734730906, + "grad_norm": 1.1158950328826904, + "learning_rate": 1.4416285627619059e-05, + "loss": 3.6644, + "step": 124300 + }, + { + "epoch": 1.3970104986434337, + "grad_norm": 1.1530020236968994, + "learning_rate": 1.441403871431621e-05, + "loss": 3.6921, + "step": 124350 + }, + { + "epoch": 1.3975722238137769, + "grad_norm": 1.100235939025879, + "learning_rate": 1.441179180101336e-05, + "loss": 3.6874, + "step": 124400 + }, + { + "epoch": 1.39813394898412, + "grad_norm": 1.1759610176086426, + "learning_rate": 1.4409544887710509e-05, + "loss": 3.7239, + "step": 124450 + }, + { + "epoch": 1.3986956741544632, + "grad_norm": 1.1375986337661743, + "learning_rate": 1.440729797440766e-05, + "loss": 3.693, + "step": 124500 + }, + { + "epoch": 1.3992573993248063, + "grad_norm": 1.2448017597198486, + "learning_rate": 1.4405051061104809e-05, + "loss": 3.649, + "step": 124550 + }, + { + "epoch": 1.3998191244951494, + "grad_norm": 1.1365571022033691, + "learning_rate": 1.4402804147801958e-05, + "loss": 3.6916, + "step": 124600 + }, + { + "epoch": 1.4003808496654926, + "grad_norm": 1.1279441118240356, + "learning_rate": 1.4400557234499108e-05, + "loss": 3.6958, + "step": 124650 + }, + { + "epoch": 1.400942574835836, + "grad_norm": 1.1293400526046753, + "learning_rate": 1.4398310321196259e-05, + "loss": 3.6368, + "step": 124700 + }, + { + "epoch": 1.401504300006179, + "grad_norm": 1.1918554306030273, + "learning_rate": 1.4396063407893408e-05, + "loss": 3.7152, + "step": 124750 + }, + { + "epoch": 1.4020660251765222, + "grad_norm": 1.128774881362915, + "learning_rate": 1.4393816494590557e-05, + "loss": 3.7738, + "step": 124800 + }, + { + "epoch": 1.4026277503468654, + "grad_norm": 1.1045053005218506, + "learning_rate": 1.4391569581287708e-05, + "loss": 3.6975, + "step": 124850 + }, + { + "epoch": 1.4031894755172085, + "grad_norm": 1.0207412242889404, + "learning_rate": 1.4389322667984858e-05, + "loss": 3.7152, + "step": 124900 + }, + { + "epoch": 1.4037512006875517, + "grad_norm": 1.2564656734466553, + "learning_rate": 1.4387075754682007e-05, + "loss": 3.7022, + "step": 124950 + }, + { + "epoch": 1.4043129258578948, + "grad_norm": 1.2463675737380981, + "learning_rate": 1.4384828841379156e-05, + "loss": 3.6474, + "step": 125000 + }, + { + "epoch": 1.404874651028238, + "grad_norm": 1.220080852508545, + "learning_rate": 1.4382581928076307e-05, + "loss": 3.6833, + "step": 125050 + }, + { + "epoch": 1.405436376198581, + "grad_norm": 1.3084275722503662, + "learning_rate": 1.4380335014773457e-05, + "loss": 3.7149, + "step": 125100 + }, + { + "epoch": 1.4059981013689242, + "grad_norm": 1.0215991735458374, + "learning_rate": 1.4378088101470606e-05, + "loss": 3.68, + "step": 125150 + }, + { + "epoch": 1.4065598265392674, + "grad_norm": 1.144219994544983, + "learning_rate": 1.4375841188167757e-05, + "loss": 3.6669, + "step": 125200 + }, + { + "epoch": 1.4071215517096105, + "grad_norm": 1.0633809566497803, + "learning_rate": 1.4373594274864906e-05, + "loss": 3.679, + "step": 125250 + }, + { + "epoch": 1.4076832768799536, + "grad_norm": 1.0320639610290527, + "learning_rate": 1.4371347361562056e-05, + "loss": 3.7242, + "step": 125300 + }, + { + "epoch": 1.4082450020502968, + "grad_norm": 1.2658512592315674, + "learning_rate": 1.4369100448259205e-05, + "loss": 3.6426, + "step": 125350 + }, + { + "epoch": 1.40880672722064, + "grad_norm": 1.3241974115371704, + "learning_rate": 1.4366853534956356e-05, + "loss": 3.6962, + "step": 125400 + }, + { + "epoch": 1.409368452390983, + "grad_norm": 1.142940878868103, + "learning_rate": 1.4364606621653505e-05, + "loss": 3.7261, + "step": 125450 + }, + { + "epoch": 1.4099301775613262, + "grad_norm": 1.0860401391983032, + "learning_rate": 1.4362359708350655e-05, + "loss": 3.6748, + "step": 125500 + }, + { + "epoch": 1.4104919027316696, + "grad_norm": 1.0803844928741455, + "learning_rate": 1.4360112795047806e-05, + "loss": 3.6706, + "step": 125550 + }, + { + "epoch": 1.4110536279020127, + "grad_norm": 1.0579431056976318, + "learning_rate": 1.4357865881744955e-05, + "loss": 3.6161, + "step": 125600 + }, + { + "epoch": 1.4116153530723559, + "grad_norm": 1.117509126663208, + "learning_rate": 1.4355618968442104e-05, + "loss": 3.6908, + "step": 125650 + }, + { + "epoch": 1.412177078242699, + "grad_norm": 0.9918485283851624, + "learning_rate": 1.4353372055139254e-05, + "loss": 3.6771, + "step": 125700 + }, + { + "epoch": 1.4127388034130421, + "grad_norm": 1.1350773572921753, + "learning_rate": 1.4351125141836405e-05, + "loss": 3.6657, + "step": 125750 + }, + { + "epoch": 1.4133005285833853, + "grad_norm": 1.097361445426941, + "learning_rate": 1.4348878228533554e-05, + "loss": 3.6585, + "step": 125800 + }, + { + "epoch": 1.4138622537537284, + "grad_norm": 1.077668309211731, + "learning_rate": 1.4346631315230703e-05, + "loss": 3.6636, + "step": 125850 + }, + { + "epoch": 1.4144239789240716, + "grad_norm": 1.076488971710205, + "learning_rate": 1.4344384401927852e-05, + "loss": 3.7048, + "step": 125900 + }, + { + "epoch": 1.4149857040944147, + "grad_norm": 1.0363222360610962, + "learning_rate": 1.4342137488625003e-05, + "loss": 3.7138, + "step": 125950 + }, + { + "epoch": 1.4155474292647579, + "grad_norm": 1.0795608758926392, + "learning_rate": 1.4339890575322153e-05, + "loss": 3.7081, + "step": 126000 + }, + { + "epoch": 1.4161091544351012, + "grad_norm": 1.0292444229125977, + "learning_rate": 1.4337643662019302e-05, + "loss": 3.698, + "step": 126050 + }, + { + "epoch": 1.4166708796054444, + "grad_norm": 1.1306686401367188, + "learning_rate": 1.4335396748716453e-05, + "loss": 3.7045, + "step": 126100 + }, + { + "epoch": 1.4172326047757875, + "grad_norm": 1.0027133226394653, + "learning_rate": 1.4333149835413602e-05, + "loss": 3.7866, + "step": 126150 + }, + { + "epoch": 1.4177943299461306, + "grad_norm": 1.2080094814300537, + "learning_rate": 1.4330902922110752e-05, + "loss": 3.7534, + "step": 126200 + }, + { + "epoch": 1.4183560551164738, + "grad_norm": 1.169230341911316, + "learning_rate": 1.4328656008807901e-05, + "loss": 3.6231, + "step": 126250 + }, + { + "epoch": 1.418917780286817, + "grad_norm": 1.258392572402954, + "learning_rate": 1.4326409095505052e-05, + "loss": 3.6981, + "step": 126300 + }, + { + "epoch": 1.41947950545716, + "grad_norm": 1.195525050163269, + "learning_rate": 1.4324162182202201e-05, + "loss": 3.7301, + "step": 126350 + }, + { + "epoch": 1.4200412306275032, + "grad_norm": 1.0613311529159546, + "learning_rate": 1.432191526889935e-05, + "loss": 3.6765, + "step": 126400 + }, + { + "epoch": 1.4206029557978463, + "grad_norm": 1.3959957361221313, + "learning_rate": 1.4319668355596502e-05, + "loss": 3.7244, + "step": 126450 + }, + { + "epoch": 1.4211646809681895, + "grad_norm": 1.214530348777771, + "learning_rate": 1.4317421442293651e-05, + "loss": 3.7079, + "step": 126500 + }, + { + "epoch": 1.4217264061385326, + "grad_norm": 1.0872678756713867, + "learning_rate": 1.43151745289908e-05, + "loss": 3.7104, + "step": 126550 + }, + { + "epoch": 1.4222881313088758, + "grad_norm": 1.122354507446289, + "learning_rate": 1.431292761568795e-05, + "loss": 3.6687, + "step": 126600 + }, + { + "epoch": 1.422849856479219, + "grad_norm": 1.29712975025177, + "learning_rate": 1.43106807023851e-05, + "loss": 3.6975, + "step": 126650 + }, + { + "epoch": 1.423411581649562, + "grad_norm": 1.2663835287094116, + "learning_rate": 1.430843378908225e-05, + "loss": 3.5613, + "step": 126700 + }, + { + "epoch": 1.4239733068199052, + "grad_norm": 1.0217410326004028, + "learning_rate": 1.43061868757794e-05, + "loss": 3.7152, + "step": 126750 + }, + { + "epoch": 1.4245350319902483, + "grad_norm": 1.2005146741867065, + "learning_rate": 1.430393996247655e-05, + "loss": 3.6889, + "step": 126800 + }, + { + "epoch": 1.4250967571605915, + "grad_norm": 1.1882201433181763, + "learning_rate": 1.43016930491737e-05, + "loss": 3.617, + "step": 126850 + }, + { + "epoch": 1.4256584823309348, + "grad_norm": 1.1847543716430664, + "learning_rate": 1.4299446135870849e-05, + "loss": 3.6783, + "step": 126900 + }, + { + "epoch": 1.426220207501278, + "grad_norm": 1.4459447860717773, + "learning_rate": 1.4297199222567998e-05, + "loss": 3.7479, + "step": 126950 + }, + { + "epoch": 1.4267819326716211, + "grad_norm": 1.077094554901123, + "learning_rate": 1.429495230926515e-05, + "loss": 3.6993, + "step": 127000 + }, + { + "epoch": 1.4273436578419643, + "grad_norm": 0.9693183302879333, + "learning_rate": 1.4292705395962299e-05, + "loss": 3.5469, + "step": 127050 + }, + { + "epoch": 1.4279053830123074, + "grad_norm": 0.9766404032707214, + "learning_rate": 1.4290458482659448e-05, + "loss": 3.721, + "step": 127100 + }, + { + "epoch": 1.4284671081826505, + "grad_norm": 1.2357624769210815, + "learning_rate": 1.4288211569356599e-05, + "loss": 3.6714, + "step": 127150 + }, + { + "epoch": 1.4290288333529937, + "grad_norm": 1.0561624765396118, + "learning_rate": 1.4285964656053748e-05, + "loss": 3.743, + "step": 127200 + }, + { + "epoch": 1.4295905585233368, + "grad_norm": 1.4372925758361816, + "learning_rate": 1.4283717742750898e-05, + "loss": 3.6237, + "step": 127250 + }, + { + "epoch": 1.43015228369368, + "grad_norm": 1.37134850025177, + "learning_rate": 1.4281470829448047e-05, + "loss": 3.7034, + "step": 127300 + }, + { + "epoch": 1.4307140088640231, + "grad_norm": 1.3327314853668213, + "learning_rate": 1.4279223916145198e-05, + "loss": 3.6874, + "step": 127350 + }, + { + "epoch": 1.4312757340343665, + "grad_norm": 1.2093868255615234, + "learning_rate": 1.4276977002842347e-05, + "loss": 3.6656, + "step": 127400 + }, + { + "epoch": 1.4318374592047096, + "grad_norm": 1.1644083261489868, + "learning_rate": 1.4274730089539497e-05, + "loss": 3.6415, + "step": 127450 + }, + { + "epoch": 1.4323991843750528, + "grad_norm": 1.0986287593841553, + "learning_rate": 1.4272483176236646e-05, + "loss": 3.7425, + "step": 127500 + }, + { + "epoch": 1.432960909545396, + "grad_norm": 1.1844991445541382, + "learning_rate": 1.4270236262933797e-05, + "loss": 3.6582, + "step": 127550 + }, + { + "epoch": 1.433522634715739, + "grad_norm": 1.1148576736450195, + "learning_rate": 1.4267989349630946e-05, + "loss": 3.637, + "step": 127600 + }, + { + "epoch": 1.4340843598860822, + "grad_norm": 1.2863951921463013, + "learning_rate": 1.4265787374594154e-05, + "loss": 3.6883, + "step": 127650 + }, + { + "epoch": 1.4346460850564253, + "grad_norm": 1.1279274225234985, + "learning_rate": 1.4263540461291303e-05, + "loss": 3.6854, + "step": 127700 + }, + { + "epoch": 1.4352078102267685, + "grad_norm": 1.2286314964294434, + "learning_rate": 1.4261293547988452e-05, + "loss": 3.6296, + "step": 127750 + }, + { + "epoch": 1.4357695353971116, + "grad_norm": 1.189313530921936, + "learning_rate": 1.4259046634685603e-05, + "loss": 3.7342, + "step": 127800 + }, + { + "epoch": 1.4363312605674547, + "grad_norm": 1.071767807006836, + "learning_rate": 1.4256799721382753e-05, + "loss": 3.7375, + "step": 127850 + }, + { + "epoch": 1.436892985737798, + "grad_norm": 1.6580266952514648, + "learning_rate": 1.4254552808079902e-05, + "loss": 3.6849, + "step": 127900 + }, + { + "epoch": 1.437454710908141, + "grad_norm": 1.3203271627426147, + "learning_rate": 1.4252305894777051e-05, + "loss": 3.7026, + "step": 127950 + }, + { + "epoch": 1.4380164360784842, + "grad_norm": 1.336090087890625, + "learning_rate": 1.4250058981474202e-05, + "loss": 3.7083, + "step": 128000 + }, + { + "epoch": 1.4385781612488273, + "grad_norm": 1.0259944200515747, + "learning_rate": 1.4247812068171352e-05, + "loss": 3.7734, + "step": 128050 + }, + { + "epoch": 1.4391398864191705, + "grad_norm": 1.3454315662384033, + "learning_rate": 1.4245565154868501e-05, + "loss": 3.6751, + "step": 128100 + }, + { + "epoch": 1.4397016115895136, + "grad_norm": 1.3078041076660156, + "learning_rate": 1.4243318241565652e-05, + "loss": 3.6879, + "step": 128150 + }, + { + "epoch": 1.4402633367598567, + "grad_norm": 1.1114927530288696, + "learning_rate": 1.4241071328262801e-05, + "loss": 3.6901, + "step": 128200 + }, + { + "epoch": 1.4408250619302, + "grad_norm": 1.1944338083267212, + "learning_rate": 1.423882441495995e-05, + "loss": 3.6592, + "step": 128250 + }, + { + "epoch": 1.4413867871005432, + "grad_norm": 1.1394671201705933, + "learning_rate": 1.42365775016571e-05, + "loss": 3.7205, + "step": 128300 + }, + { + "epoch": 1.4419485122708864, + "grad_norm": 1.316195011138916, + "learning_rate": 1.4234330588354251e-05, + "loss": 3.6758, + "step": 128350 + }, + { + "epoch": 1.4425102374412295, + "grad_norm": 1.1554863452911377, + "learning_rate": 1.42320836750514e-05, + "loss": 3.7103, + "step": 128400 + }, + { + "epoch": 1.4430719626115727, + "grad_norm": 1.152806043624878, + "learning_rate": 1.422983676174855e-05, + "loss": 3.7541, + "step": 128450 + }, + { + "epoch": 1.4436336877819158, + "grad_norm": 1.1429660320281982, + "learning_rate": 1.4227589848445699e-05, + "loss": 3.7786, + "step": 128500 + }, + { + "epoch": 1.444195412952259, + "grad_norm": 1.2185338735580444, + "learning_rate": 1.422534293514285e-05, + "loss": 3.6484, + "step": 128550 + }, + { + "epoch": 1.444757138122602, + "grad_norm": 1.334708333015442, + "learning_rate": 1.422309602184e-05, + "loss": 3.6916, + "step": 128600 + }, + { + "epoch": 1.4453188632929452, + "grad_norm": 1.0479364395141602, + "learning_rate": 1.4220849108537149e-05, + "loss": 3.7259, + "step": 128650 + }, + { + "epoch": 1.4458805884632884, + "grad_norm": 1.2427948713302612, + "learning_rate": 1.42186021952343e-05, + "loss": 3.7372, + "step": 128700 + }, + { + "epoch": 1.4464423136336317, + "grad_norm": 1.1542428731918335, + "learning_rate": 1.4216355281931449e-05, + "loss": 3.6836, + "step": 128750 + }, + { + "epoch": 1.4470040388039749, + "grad_norm": 1.1932610273361206, + "learning_rate": 1.4214108368628598e-05, + "loss": 3.7243, + "step": 128800 + }, + { + "epoch": 1.447565763974318, + "grad_norm": 1.0314894914627075, + "learning_rate": 1.4211861455325748e-05, + "loss": 3.6924, + "step": 128850 + }, + { + "epoch": 1.4481274891446612, + "grad_norm": 1.1394891738891602, + "learning_rate": 1.4209614542022899e-05, + "loss": 3.6632, + "step": 128900 + }, + { + "epoch": 1.4486892143150043, + "grad_norm": 1.2204251289367676, + "learning_rate": 1.4207367628720048e-05, + "loss": 3.7129, + "step": 128950 + }, + { + "epoch": 1.4492509394853474, + "grad_norm": 1.5714738368988037, + "learning_rate": 1.4205120715417197e-05, + "loss": 3.6579, + "step": 129000 + }, + { + "epoch": 1.4498126646556906, + "grad_norm": 1.2154221534729004, + "learning_rate": 1.4202873802114348e-05, + "loss": 3.6245, + "step": 129050 + }, + { + "epoch": 1.4503743898260337, + "grad_norm": 1.0103734731674194, + "learning_rate": 1.4200626888811498e-05, + "loss": 3.6631, + "step": 129100 + }, + { + "epoch": 1.4509361149963769, + "grad_norm": 1.0002745389938354, + "learning_rate": 1.4198379975508647e-05, + "loss": 3.7259, + "step": 129150 + }, + { + "epoch": 1.45149784016672, + "grad_norm": 1.2197633981704712, + "learning_rate": 1.4196133062205796e-05, + "loss": 3.6225, + "step": 129200 + }, + { + "epoch": 1.4520595653370632, + "grad_norm": 1.1183518171310425, + "learning_rate": 1.4193886148902947e-05, + "loss": 3.6419, + "step": 129250 + }, + { + "epoch": 1.4526212905074063, + "grad_norm": 1.2335314750671387, + "learning_rate": 1.4191639235600096e-05, + "loss": 3.5827, + "step": 129300 + }, + { + "epoch": 1.4531830156777494, + "grad_norm": 1.0736817121505737, + "learning_rate": 1.41894372605633e-05, + "loss": 3.6258, + "step": 129350 + }, + { + "epoch": 1.4537447408480926, + "grad_norm": 1.283150315284729, + "learning_rate": 1.4187190347260452e-05, + "loss": 3.7434, + "step": 129400 + }, + { + "epoch": 1.4543064660184357, + "grad_norm": 1.1960936784744263, + "learning_rate": 1.4184943433957601e-05, + "loss": 3.7223, + "step": 129450 + }, + { + "epoch": 1.4548681911887789, + "grad_norm": 1.3931291103363037, + "learning_rate": 1.418269652065475e-05, + "loss": 3.6696, + "step": 129500 + }, + { + "epoch": 1.455429916359122, + "grad_norm": 1.104906439781189, + "learning_rate": 1.41804496073519e-05, + "loss": 3.7111, + "step": 129550 + }, + { + "epoch": 1.4559916415294654, + "grad_norm": 1.100758671760559, + "learning_rate": 1.417820269404905e-05, + "loss": 3.6306, + "step": 129600 + }, + { + "epoch": 1.4565533666998085, + "grad_norm": 1.404305100440979, + "learning_rate": 1.41759557807462e-05, + "loss": 3.6803, + "step": 129650 + }, + { + "epoch": 1.4571150918701516, + "grad_norm": 1.3029145002365112, + "learning_rate": 1.417370886744335e-05, + "loss": 3.6495, + "step": 129700 + }, + { + "epoch": 1.4576768170404948, + "grad_norm": 1.038677453994751, + "learning_rate": 1.41714619541405e-05, + "loss": 3.6658, + "step": 129750 + }, + { + "epoch": 1.458238542210838, + "grad_norm": 1.1654647588729858, + "learning_rate": 1.416921504083765e-05, + "loss": 3.7818, + "step": 129800 + }, + { + "epoch": 1.458800267381181, + "grad_norm": 1.0105221271514893, + "learning_rate": 1.4166968127534799e-05, + "loss": 3.7315, + "step": 129850 + }, + { + "epoch": 1.4593619925515242, + "grad_norm": 1.0998730659484863, + "learning_rate": 1.4164721214231948e-05, + "loss": 3.7785, + "step": 129900 + }, + { + "epoch": 1.4599237177218674, + "grad_norm": 1.0934780836105347, + "learning_rate": 1.41624743009291e-05, + "loss": 3.6549, + "step": 129950 + }, + { + "epoch": 1.4604854428922105, + "grad_norm": 1.2375984191894531, + "learning_rate": 1.4160227387626249e-05, + "loss": 3.7077, + "step": 130000 + }, + { + "epoch": 1.4610471680625536, + "grad_norm": 1.1644408702850342, + "learning_rate": 1.4157980474323398e-05, + "loss": 3.6903, + "step": 130050 + }, + { + "epoch": 1.461608893232897, + "grad_norm": 1.0978128910064697, + "learning_rate": 1.4155733561020549e-05, + "loss": 3.6279, + "step": 130100 + }, + { + "epoch": 1.4621706184032401, + "grad_norm": 1.1605157852172852, + "learning_rate": 1.4153486647717698e-05, + "loss": 3.7902, + "step": 130150 + }, + { + "epoch": 1.4627323435735833, + "grad_norm": 1.271424412727356, + "learning_rate": 1.4151239734414848e-05, + "loss": 3.7077, + "step": 130200 + }, + { + "epoch": 1.4632940687439264, + "grad_norm": 1.1201791763305664, + "learning_rate": 1.4148992821111997e-05, + "loss": 3.6611, + "step": 130250 + }, + { + "epoch": 1.4638557939142696, + "grad_norm": 1.347887396812439, + "learning_rate": 1.4146745907809148e-05, + "loss": 3.697, + "step": 130300 + }, + { + "epoch": 1.4644175190846127, + "grad_norm": 1.1798382997512817, + "learning_rate": 1.4144498994506297e-05, + "loss": 3.626, + "step": 130350 + }, + { + "epoch": 1.4649792442549558, + "grad_norm": 3.724522829055786, + "learning_rate": 1.4142252081203447e-05, + "loss": 3.6137, + "step": 130400 + }, + { + "epoch": 1.465540969425299, + "grad_norm": 1.0803759098052979, + "learning_rate": 1.4140005167900598e-05, + "loss": 3.6533, + "step": 130450 + }, + { + "epoch": 1.4661026945956421, + "grad_norm": 1.1397240161895752, + "learning_rate": 1.4137758254597747e-05, + "loss": 3.7668, + "step": 130500 + }, + { + "epoch": 1.4666644197659853, + "grad_norm": 1.0932416915893555, + "learning_rate": 1.4135511341294896e-05, + "loss": 3.7927, + "step": 130550 + }, + { + "epoch": 1.4672261449363284, + "grad_norm": 1.1760075092315674, + "learning_rate": 1.4133264427992046e-05, + "loss": 3.6757, + "step": 130600 + }, + { + "epoch": 1.4677878701066716, + "grad_norm": 1.1441518068313599, + "learning_rate": 1.4131017514689197e-05, + "loss": 3.7135, + "step": 130650 + }, + { + "epoch": 1.4683495952770147, + "grad_norm": 1.0260703563690186, + "learning_rate": 1.4128770601386346e-05, + "loss": 3.644, + "step": 130700 + }, + { + "epoch": 1.4689113204473578, + "grad_norm": 1.1579915285110474, + "learning_rate": 1.4126523688083495e-05, + "loss": 3.6791, + "step": 130750 + }, + { + "epoch": 1.469473045617701, + "grad_norm": 1.1837948560714722, + "learning_rate": 1.4124276774780644e-05, + "loss": 3.6832, + "step": 130800 + }, + { + "epoch": 1.4700347707880441, + "grad_norm": 1.105905294418335, + "learning_rate": 1.4122029861477795e-05, + "loss": 3.6534, + "step": 130850 + }, + { + "epoch": 1.4705964959583873, + "grad_norm": 1.1882402896881104, + "learning_rate": 1.4119782948174945e-05, + "loss": 3.5552, + "step": 130900 + }, + { + "epoch": 1.4711582211287306, + "grad_norm": 1.0834487676620483, + "learning_rate": 1.4117536034872094e-05, + "loss": 3.6657, + "step": 130950 + }, + { + "epoch": 1.4717199462990738, + "grad_norm": 1.0879491567611694, + "learning_rate": 1.4115289121569245e-05, + "loss": 3.6612, + "step": 131000 + }, + { + "epoch": 1.472281671469417, + "grad_norm": 1.14711332321167, + "learning_rate": 1.4113042208266394e-05, + "loss": 3.6724, + "step": 131050 + }, + { + "epoch": 1.47284339663976, + "grad_norm": 1.1432496309280396, + "learning_rate": 1.4110795294963544e-05, + "loss": 3.69, + "step": 131100 + }, + { + "epoch": 1.4734051218101032, + "grad_norm": 1.0857371091842651, + "learning_rate": 1.4108548381660693e-05, + "loss": 3.6368, + "step": 131150 + }, + { + "epoch": 1.4739668469804463, + "grad_norm": 1.097549557685852, + "learning_rate": 1.4106301468357844e-05, + "loss": 3.6635, + "step": 131200 + }, + { + "epoch": 1.4745285721507895, + "grad_norm": 1.0512332916259766, + "learning_rate": 1.4104054555054993e-05, + "loss": 3.7048, + "step": 131250 + }, + { + "epoch": 1.4750902973211326, + "grad_norm": 1.1470379829406738, + "learning_rate": 1.4101807641752143e-05, + "loss": 3.7138, + "step": 131300 + }, + { + "epoch": 1.4756520224914758, + "grad_norm": 1.0969492197036743, + "learning_rate": 1.4099560728449294e-05, + "loss": 3.7089, + "step": 131350 + }, + { + "epoch": 1.476213747661819, + "grad_norm": 1.1652694940567017, + "learning_rate": 1.4097313815146443e-05, + "loss": 3.6553, + "step": 131400 + }, + { + "epoch": 1.4767754728321623, + "grad_norm": 1.1556792259216309, + "learning_rate": 1.4095066901843592e-05, + "loss": 3.582, + "step": 131450 + }, + { + "epoch": 1.4773371980025054, + "grad_norm": 1.1985116004943848, + "learning_rate": 1.4092819988540742e-05, + "loss": 3.635, + "step": 131500 + }, + { + "epoch": 1.4778989231728485, + "grad_norm": 1.3924946784973145, + "learning_rate": 1.4090573075237893e-05, + "loss": 3.7242, + "step": 131550 + }, + { + "epoch": 1.4784606483431917, + "grad_norm": 1.2415711879730225, + "learning_rate": 1.4088326161935042e-05, + "loss": 3.752, + "step": 131600 + }, + { + "epoch": 1.4790223735135348, + "grad_norm": 1.3485480546951294, + "learning_rate": 1.4086079248632191e-05, + "loss": 3.6784, + "step": 131650 + }, + { + "epoch": 1.479584098683878, + "grad_norm": 1.1684798002243042, + "learning_rate": 1.4083832335329342e-05, + "loss": 3.6775, + "step": 131700 + }, + { + "epoch": 1.480145823854221, + "grad_norm": 1.3369027376174927, + "learning_rate": 1.4081585422026492e-05, + "loss": 3.7093, + "step": 131750 + }, + { + "epoch": 1.4807075490245643, + "grad_norm": 1.2231554985046387, + "learning_rate": 1.4079338508723641e-05, + "loss": 3.7172, + "step": 131800 + }, + { + "epoch": 1.4812692741949074, + "grad_norm": 1.0951569080352783, + "learning_rate": 1.407709159542079e-05, + "loss": 3.7198, + "step": 131850 + }, + { + "epoch": 1.4818309993652505, + "grad_norm": 1.049019455909729, + "learning_rate": 1.4074844682117941e-05, + "loss": 3.7794, + "step": 131900 + }, + { + "epoch": 1.4823927245355937, + "grad_norm": 1.3006795644760132, + "learning_rate": 1.407259776881509e-05, + "loss": 3.6084, + "step": 131950 + }, + { + "epoch": 1.4829544497059368, + "grad_norm": 1.2521191835403442, + "learning_rate": 1.407035085551224e-05, + "loss": 3.7331, + "step": 132000 + }, + { + "epoch": 1.48351617487628, + "grad_norm": 1.125417709350586, + "learning_rate": 1.4068103942209391e-05, + "loss": 3.6835, + "step": 132050 + }, + { + "epoch": 1.484077900046623, + "grad_norm": 1.1457781791687012, + "learning_rate": 1.406585702890654e-05, + "loss": 3.7033, + "step": 132100 + }, + { + "epoch": 1.4846396252169662, + "grad_norm": 1.1376407146453857, + "learning_rate": 1.406361011560369e-05, + "loss": 3.7074, + "step": 132150 + }, + { + "epoch": 1.4852013503873094, + "grad_norm": 1.1445938348770142, + "learning_rate": 1.4061363202300839e-05, + "loss": 3.6893, + "step": 132200 + }, + { + "epoch": 1.4857630755576527, + "grad_norm": 1.0046522617340088, + "learning_rate": 1.405911628899799e-05, + "loss": 3.6741, + "step": 132250 + }, + { + "epoch": 1.4863248007279959, + "grad_norm": 1.265594482421875, + "learning_rate": 1.405686937569514e-05, + "loss": 3.6365, + "step": 132300 + }, + { + "epoch": 1.486886525898339, + "grad_norm": 1.1877540349960327, + "learning_rate": 1.4054622462392289e-05, + "loss": 3.6737, + "step": 132350 + }, + { + "epoch": 1.4874482510686822, + "grad_norm": 1.0919491052627563, + "learning_rate": 1.4052375549089438e-05, + "loss": 3.7282, + "step": 132400 + }, + { + "epoch": 1.4880099762390253, + "grad_norm": 1.1396973133087158, + "learning_rate": 1.4050128635786589e-05, + "loss": 3.6833, + "step": 132450 + }, + { + "epoch": 1.4885717014093685, + "grad_norm": 1.05573570728302, + "learning_rate": 1.4047881722483738e-05, + "loss": 3.7685, + "step": 132500 + }, + { + "epoch": 1.4891334265797116, + "grad_norm": 0.9932793974876404, + "learning_rate": 1.4045634809180887e-05, + "loss": 3.6699, + "step": 132550 + }, + { + "epoch": 1.4896951517500547, + "grad_norm": 1.2047851085662842, + "learning_rate": 1.4043387895878038e-05, + "loss": 3.6439, + "step": 132600 + }, + { + "epoch": 1.4902568769203979, + "grad_norm": 2.069572687149048, + "learning_rate": 1.4041140982575188e-05, + "loss": 3.6966, + "step": 132650 + }, + { + "epoch": 1.490818602090741, + "grad_norm": 1.0004225969314575, + "learning_rate": 1.4038894069272337e-05, + "loss": 3.7124, + "step": 132700 + }, + { + "epoch": 1.4913803272610842, + "grad_norm": 1.1484732627868652, + "learning_rate": 1.4036647155969486e-05, + "loss": 3.7173, + "step": 132750 + }, + { + "epoch": 1.4919420524314275, + "grad_norm": 1.0647058486938477, + "learning_rate": 1.4034400242666637e-05, + "loss": 3.7109, + "step": 132800 + }, + { + "epoch": 1.4925037776017707, + "grad_norm": 1.1456990242004395, + "learning_rate": 1.4032153329363787e-05, + "loss": 3.6422, + "step": 132850 + }, + { + "epoch": 1.4930655027721138, + "grad_norm": 1.2056528329849243, + "learning_rate": 1.4029906416060936e-05, + "loss": 3.6598, + "step": 132900 + }, + { + "epoch": 1.493627227942457, + "grad_norm": 1.307550072669983, + "learning_rate": 1.4027659502758087e-05, + "loss": 3.6904, + "step": 132950 + }, + { + "epoch": 1.4941889531128, + "grad_norm": 1.2377219200134277, + "learning_rate": 1.4025412589455236e-05, + "loss": 3.6669, + "step": 133000 + }, + { + "epoch": 1.4947506782831432, + "grad_norm": 1.174829363822937, + "learning_rate": 1.4023165676152386e-05, + "loss": 3.6073, + "step": 133050 + }, + { + "epoch": 1.4953124034534864, + "grad_norm": 1.102279782295227, + "learning_rate": 1.4020918762849535e-05, + "loss": 3.701, + "step": 133100 + }, + { + "epoch": 1.4958741286238295, + "grad_norm": 1.1181694269180298, + "learning_rate": 1.4018671849546686e-05, + "loss": 3.6779, + "step": 133150 + }, + { + "epoch": 1.4964358537941727, + "grad_norm": 1.0357697010040283, + "learning_rate": 1.4016424936243835e-05, + "loss": 3.6923, + "step": 133200 + }, + { + "epoch": 1.4969975789645158, + "grad_norm": 0.9432259202003479, + "learning_rate": 1.4014178022940985e-05, + "loss": 3.7048, + "step": 133250 + }, + { + "epoch": 1.497559304134859, + "grad_norm": 1.0554338693618774, + "learning_rate": 1.4011931109638136e-05, + "loss": 3.7422, + "step": 133300 + }, + { + "epoch": 1.498121029305202, + "grad_norm": 1.1480789184570312, + "learning_rate": 1.4009684196335285e-05, + "loss": 3.7092, + "step": 133350 + }, + { + "epoch": 1.4986827544755452, + "grad_norm": 1.0501972436904907, + "learning_rate": 1.4007437283032434e-05, + "loss": 3.7159, + "step": 133400 + }, + { + "epoch": 1.4992444796458884, + "grad_norm": 1.296221375465393, + "learning_rate": 1.4005190369729584e-05, + "loss": 3.6433, + "step": 133450 + }, + { + "epoch": 1.4998062048162315, + "grad_norm": 1.0956120491027832, + "learning_rate": 1.4002943456426735e-05, + "loss": 3.6807, + "step": 133500 + }, + { + "epoch": 1.5003679299865746, + "grad_norm": 1.1169849634170532, + "learning_rate": 1.4000696543123884e-05, + "loss": 3.6143, + "step": 133550 + }, + { + "epoch": 1.5009296551569178, + "grad_norm": 1.1566176414489746, + "learning_rate": 1.3998449629821033e-05, + "loss": 3.7322, + "step": 133600 + }, + { + "epoch": 1.501491380327261, + "grad_norm": 1.040257215499878, + "learning_rate": 1.3996202716518184e-05, + "loss": 3.6666, + "step": 133650 + }, + { + "epoch": 1.5020531054976043, + "grad_norm": 1.1366089582443237, + "learning_rate": 1.3993955803215334e-05, + "loss": 3.7142, + "step": 133700 + }, + { + "epoch": 1.5026148306679474, + "grad_norm": 1.131287693977356, + "learning_rate": 1.3991708889912483e-05, + "loss": 3.7191, + "step": 133750 + }, + { + "epoch": 1.5031765558382906, + "grad_norm": 1.1907302141189575, + "learning_rate": 1.3989461976609632e-05, + "loss": 3.6475, + "step": 133800 + }, + { + "epoch": 1.5037382810086337, + "grad_norm": 1.2153143882751465, + "learning_rate": 1.3987215063306783e-05, + "loss": 3.7205, + "step": 133850 + }, + { + "epoch": 1.5043000061789769, + "grad_norm": 1.1694841384887695, + "learning_rate": 1.3984968150003933e-05, + "loss": 3.6523, + "step": 133900 + }, + { + "epoch": 1.50486173134932, + "grad_norm": 1.141675591468811, + "learning_rate": 1.3982721236701082e-05, + "loss": 3.6907, + "step": 133950 + }, + { + "epoch": 1.5054234565196631, + "grad_norm": 1.1201070547103882, + "learning_rate": 1.3980474323398233e-05, + "loss": 3.752, + "step": 134000 + }, + { + "epoch": 1.5059851816900065, + "grad_norm": 1.5073152780532837, + "learning_rate": 1.3978227410095382e-05, + "loss": 3.6264, + "step": 134050 + }, + { + "epoch": 1.5065469068603496, + "grad_norm": 1.4463489055633545, + "learning_rate": 1.3975980496792532e-05, + "loss": 3.636, + "step": 134100 + }, + { + "epoch": 1.5071086320306928, + "grad_norm": 1.1376241445541382, + "learning_rate": 1.3973733583489681e-05, + "loss": 3.6875, + "step": 134150 + }, + { + "epoch": 1.507670357201036, + "grad_norm": 1.2015286684036255, + "learning_rate": 1.3971486670186832e-05, + "loss": 3.6775, + "step": 134200 + }, + { + "epoch": 1.508232082371379, + "grad_norm": 1.2428511381149292, + "learning_rate": 1.3969239756883981e-05, + "loss": 3.7064, + "step": 134250 + }, + { + "epoch": 1.5087938075417222, + "grad_norm": 1.1819654703140259, + "learning_rate": 1.396699284358113e-05, + "loss": 3.6566, + "step": 134300 + }, + { + "epoch": 1.5093555327120654, + "grad_norm": 1.1704370975494385, + "learning_rate": 1.396474593027828e-05, + "loss": 3.6326, + "step": 134350 + }, + { + "epoch": 1.5099172578824085, + "grad_norm": 1.0891886949539185, + "learning_rate": 1.396249901697543e-05, + "loss": 3.7003, + "step": 134400 + }, + { + "epoch": 1.5104789830527516, + "grad_norm": 1.1705445051193237, + "learning_rate": 1.396025210367258e-05, + "loss": 3.669, + "step": 134450 + }, + { + "epoch": 1.5110407082230948, + "grad_norm": 1.3058868646621704, + "learning_rate": 1.395800519036973e-05, + "loss": 3.6446, + "step": 134500 + }, + { + "epoch": 1.511602433393438, + "grad_norm": 1.1677734851837158, + "learning_rate": 1.395575827706688e-05, + "loss": 3.6687, + "step": 134550 + }, + { + "epoch": 1.512164158563781, + "grad_norm": 1.088308334350586, + "learning_rate": 1.3953556302030086e-05, + "loss": 3.6781, + "step": 134600 + }, + { + "epoch": 1.5127258837341242, + "grad_norm": 1.1753487586975098, + "learning_rate": 1.3951309388727237e-05, + "loss": 3.7159, + "step": 134650 + }, + { + "epoch": 1.5132876089044673, + "grad_norm": 1.3727972507476807, + "learning_rate": 1.3949062475424387e-05, + "loss": 3.6943, + "step": 134700 + }, + { + "epoch": 1.5138493340748105, + "grad_norm": 1.4732756614685059, + "learning_rate": 1.3946815562121536e-05, + "loss": 3.717, + "step": 134750 + }, + { + "epoch": 1.5144110592451536, + "grad_norm": 1.0472688674926758, + "learning_rate": 1.3944568648818685e-05, + "loss": 3.6123, + "step": 134800 + }, + { + "epoch": 1.5149727844154968, + "grad_norm": 1.0159178972244263, + "learning_rate": 1.3942321735515836e-05, + "loss": 3.6726, + "step": 134850 + }, + { + "epoch": 1.51553450958584, + "grad_norm": 0.938441812992096, + "learning_rate": 1.3940074822212986e-05, + "loss": 3.6414, + "step": 134900 + }, + { + "epoch": 1.516096234756183, + "grad_norm": 1.1761131286621094, + "learning_rate": 1.3937827908910135e-05, + "loss": 3.6857, + "step": 134950 + }, + { + "epoch": 1.5166579599265262, + "grad_norm": 1.54920494556427, + "learning_rate": 1.3935580995607286e-05, + "loss": 3.6467, + "step": 135000 + }, + { + "epoch": 1.5172196850968696, + "grad_norm": 1.1692527532577515, + "learning_rate": 1.3933334082304435e-05, + "loss": 3.71, + "step": 135050 + }, + { + "epoch": 1.5177814102672127, + "grad_norm": 1.1783121824264526, + "learning_rate": 1.3931087169001585e-05, + "loss": 3.6882, + "step": 135100 + }, + { + "epoch": 1.5183431354375558, + "grad_norm": 1.5568922758102417, + "learning_rate": 1.3928840255698734e-05, + "loss": 3.7029, + "step": 135150 + }, + { + "epoch": 1.518904860607899, + "grad_norm": 1.1709709167480469, + "learning_rate": 1.3926593342395885e-05, + "loss": 3.6893, + "step": 135200 + }, + { + "epoch": 1.5194665857782421, + "grad_norm": 1.2699800729751587, + "learning_rate": 1.3924346429093034e-05, + "loss": 3.6815, + "step": 135250 + }, + { + "epoch": 1.5200283109485853, + "grad_norm": 1.2915416955947876, + "learning_rate": 1.3922099515790184e-05, + "loss": 3.6813, + "step": 135300 + }, + { + "epoch": 1.5205900361189284, + "grad_norm": 1.2163134813308716, + "learning_rate": 1.3919852602487333e-05, + "loss": 3.7156, + "step": 135350 + }, + { + "epoch": 1.5211517612892718, + "grad_norm": 1.2914841175079346, + "learning_rate": 1.3917605689184484e-05, + "loss": 3.6713, + "step": 135400 + }, + { + "epoch": 1.521713486459615, + "grad_norm": 1.1673322916030884, + "learning_rate": 1.3915358775881633e-05, + "loss": 3.7357, + "step": 135450 + }, + { + "epoch": 1.522275211629958, + "grad_norm": 1.0428760051727295, + "learning_rate": 1.3913111862578783e-05, + "loss": 3.7632, + "step": 135500 + }, + { + "epoch": 1.5228369368003012, + "grad_norm": 1.1311368942260742, + "learning_rate": 1.3910864949275934e-05, + "loss": 3.6775, + "step": 135550 + }, + { + "epoch": 1.5233986619706443, + "grad_norm": 1.162413477897644, + "learning_rate": 1.3908618035973083e-05, + "loss": 3.7379, + "step": 135600 + }, + { + "epoch": 1.5239603871409875, + "grad_norm": 1.0958778858184814, + "learning_rate": 1.3906371122670232e-05, + "loss": 3.7454, + "step": 135650 + }, + { + "epoch": 1.5245221123113306, + "grad_norm": 1.0307263135910034, + "learning_rate": 1.3904124209367382e-05, + "loss": 3.7756, + "step": 135700 + }, + { + "epoch": 1.5250838374816738, + "grad_norm": 1.3076707124710083, + "learning_rate": 1.3901877296064533e-05, + "loss": 3.7433, + "step": 135750 + }, + { + "epoch": 1.525645562652017, + "grad_norm": 1.247527837753296, + "learning_rate": 1.3899630382761682e-05, + "loss": 3.6535, + "step": 135800 + }, + { + "epoch": 1.52620728782236, + "grad_norm": 1.0686296224594116, + "learning_rate": 1.3897383469458831e-05, + "loss": 3.6597, + "step": 135850 + }, + { + "epoch": 1.5267690129927032, + "grad_norm": 1.087568998336792, + "learning_rate": 1.3895136556155982e-05, + "loss": 3.6908, + "step": 135900 + }, + { + "epoch": 1.5273307381630463, + "grad_norm": 0.9681762456893921, + "learning_rate": 1.3892889642853131e-05, + "loss": 3.6816, + "step": 135950 + }, + { + "epoch": 1.5278924633333895, + "grad_norm": 1.1612462997436523, + "learning_rate": 1.389064272955028e-05, + "loss": 3.7534, + "step": 136000 + }, + { + "epoch": 1.5284541885037326, + "grad_norm": 1.172121286392212, + "learning_rate": 1.388839581624743e-05, + "loss": 3.7198, + "step": 136050 + }, + { + "epoch": 1.5290159136740757, + "grad_norm": 1.2644882202148438, + "learning_rate": 1.3886148902944581e-05, + "loss": 3.5908, + "step": 136100 + }, + { + "epoch": 1.5295776388444189, + "grad_norm": 3.783116340637207, + "learning_rate": 1.388390198964173e-05, + "loss": 3.7086, + "step": 136150 + }, + { + "epoch": 1.530139364014762, + "grad_norm": 1.1201610565185547, + "learning_rate": 1.388165507633888e-05, + "loss": 3.716, + "step": 136200 + }, + { + "epoch": 1.5307010891851052, + "grad_norm": 1.1066677570343018, + "learning_rate": 1.387940816303603e-05, + "loss": 3.6825, + "step": 136250 + }, + { + "epoch": 1.5312628143554483, + "grad_norm": 1.046709418296814, + "learning_rate": 1.387716124973318e-05, + "loss": 3.6776, + "step": 136300 + }, + { + "epoch": 1.5318245395257915, + "grad_norm": 1.0653623342514038, + "learning_rate": 1.387491433643033e-05, + "loss": 3.6368, + "step": 136350 + }, + { + "epoch": 1.5323862646961348, + "grad_norm": 1.281095027923584, + "learning_rate": 1.3872667423127479e-05, + "loss": 3.6537, + "step": 136400 + }, + { + "epoch": 1.532947989866478, + "grad_norm": 1.406171441078186, + "learning_rate": 1.387042050982463e-05, + "loss": 3.6694, + "step": 136450 + }, + { + "epoch": 1.533509715036821, + "grad_norm": 1.2122948169708252, + "learning_rate": 1.3868173596521779e-05, + "loss": 3.7041, + "step": 136500 + }, + { + "epoch": 1.5340714402071642, + "grad_norm": 1.148941159248352, + "learning_rate": 1.3865926683218928e-05, + "loss": 3.7345, + "step": 136550 + }, + { + "epoch": 1.5346331653775074, + "grad_norm": 1.2091238498687744, + "learning_rate": 1.386367976991608e-05, + "loss": 3.7257, + "step": 136600 + }, + { + "epoch": 1.5351948905478505, + "grad_norm": 1.314763069152832, + "learning_rate": 1.3861432856613229e-05, + "loss": 3.7493, + "step": 136650 + }, + { + "epoch": 1.5357566157181937, + "grad_norm": 1.0040805339813232, + "learning_rate": 1.3859185943310378e-05, + "loss": 3.6682, + "step": 136700 + }, + { + "epoch": 1.536318340888537, + "grad_norm": 1.187330722808838, + "learning_rate": 1.3856939030007527e-05, + "loss": 3.6458, + "step": 136750 + }, + { + "epoch": 1.5368800660588802, + "grad_norm": 1.1977338790893555, + "learning_rate": 1.3854692116704678e-05, + "loss": 3.7833, + "step": 136800 + }, + { + "epoch": 1.5374417912292233, + "grad_norm": 1.2390897274017334, + "learning_rate": 1.3852445203401828e-05, + "loss": 3.705, + "step": 136850 + }, + { + "epoch": 1.5380035163995665, + "grad_norm": 1.155185341835022, + "learning_rate": 1.3850198290098977e-05, + "loss": 3.6979, + "step": 136900 + }, + { + "epoch": 1.5385652415699096, + "grad_norm": 1.2628538608551025, + "learning_rate": 1.3847951376796126e-05, + "loss": 3.6829, + "step": 136950 + }, + { + "epoch": 1.5391269667402527, + "grad_norm": 1.5286192893981934, + "learning_rate": 1.3845704463493277e-05, + "loss": 3.7026, + "step": 137000 + }, + { + "epoch": 1.5396886919105959, + "grad_norm": 1.2452843189239502, + "learning_rate": 1.3843457550190427e-05, + "loss": 3.6803, + "step": 137050 + }, + { + "epoch": 1.540250417080939, + "grad_norm": 1.0073351860046387, + "learning_rate": 1.3841210636887576e-05, + "loss": 3.6946, + "step": 137100 + }, + { + "epoch": 1.5408121422512822, + "grad_norm": 1.2091684341430664, + "learning_rate": 1.3838963723584727e-05, + "loss": 3.5651, + "step": 137150 + }, + { + "epoch": 1.5413738674216253, + "grad_norm": 1.0364580154418945, + "learning_rate": 1.3836716810281876e-05, + "loss": 3.7116, + "step": 137200 + }, + { + "epoch": 1.5419355925919684, + "grad_norm": 1.0992523431777954, + "learning_rate": 1.3834469896979026e-05, + "loss": 3.606, + "step": 137250 + }, + { + "epoch": 1.5424973177623116, + "grad_norm": 1.0807006359100342, + "learning_rate": 1.3832222983676175e-05, + "loss": 3.6177, + "step": 137300 + }, + { + "epoch": 1.5430590429326547, + "grad_norm": 1.2157042026519775, + "learning_rate": 1.3829976070373326e-05, + "loss": 3.6924, + "step": 137350 + }, + { + "epoch": 1.5436207681029979, + "grad_norm": 1.1488664150238037, + "learning_rate": 1.3827729157070475e-05, + "loss": 3.7056, + "step": 137400 + }, + { + "epoch": 1.544182493273341, + "grad_norm": 1.1668779850006104, + "learning_rate": 1.3825482243767625e-05, + "loss": 3.7133, + "step": 137450 + }, + { + "epoch": 1.5447442184436841, + "grad_norm": 1.0572956800460815, + "learning_rate": 1.3823235330464776e-05, + "loss": 3.6649, + "step": 137500 + }, + { + "epoch": 1.5453059436140273, + "grad_norm": 1.1351521015167236, + "learning_rate": 1.3820988417161925e-05, + "loss": 3.6284, + "step": 137550 + }, + { + "epoch": 1.5458676687843704, + "grad_norm": 1.3236806392669678, + "learning_rate": 1.3818741503859074e-05, + "loss": 3.6578, + "step": 137600 + }, + { + "epoch": 1.5464293939547136, + "grad_norm": 1.1036694049835205, + "learning_rate": 1.3816494590556224e-05, + "loss": 3.6787, + "step": 137650 + }, + { + "epoch": 1.5469911191250567, + "grad_norm": 1.151566743850708, + "learning_rate": 1.3814247677253375e-05, + "loss": 3.7113, + "step": 137700 + }, + { + "epoch": 1.5475528442954, + "grad_norm": 0.9375407695770264, + "learning_rate": 1.3812000763950524e-05, + "loss": 3.6449, + "step": 137750 + }, + { + "epoch": 1.5481145694657432, + "grad_norm": 1.1477205753326416, + "learning_rate": 1.3809753850647673e-05, + "loss": 3.6459, + "step": 137800 + }, + { + "epoch": 1.5486762946360864, + "grad_norm": 1.1529345512390137, + "learning_rate": 1.3807506937344824e-05, + "loss": 3.715, + "step": 137850 + }, + { + "epoch": 1.5492380198064295, + "grad_norm": 1.256088137626648, + "learning_rate": 1.3805260024041973e-05, + "loss": 3.8204, + "step": 137900 + }, + { + "epoch": 1.5497997449767726, + "grad_norm": 1.1161493062973022, + "learning_rate": 1.3803013110739123e-05, + "loss": 3.6332, + "step": 137950 + }, + { + "epoch": 1.5503614701471158, + "grad_norm": 1.0601900815963745, + "learning_rate": 1.3800766197436272e-05, + "loss": 3.6654, + "step": 138000 + }, + { + "epoch": 1.550923195317459, + "grad_norm": 1.0600312948226929, + "learning_rate": 1.3798519284133423e-05, + "loss": 3.627, + "step": 138050 + }, + { + "epoch": 1.5514849204878023, + "grad_norm": 1.2241134643554688, + "learning_rate": 1.3796272370830572e-05, + "loss": 3.668, + "step": 138100 + }, + { + "epoch": 1.5520466456581454, + "grad_norm": 1.3569682836532593, + "learning_rate": 1.3794025457527722e-05, + "loss": 3.6421, + "step": 138150 + }, + { + "epoch": 1.5526083708284886, + "grad_norm": 1.1638623476028442, + "learning_rate": 1.3791778544224873e-05, + "loss": 3.6432, + "step": 138200 + }, + { + "epoch": 1.5531700959988317, + "grad_norm": 1.006299614906311, + "learning_rate": 1.3789531630922022e-05, + "loss": 3.6522, + "step": 138250 + }, + { + "epoch": 1.5537318211691749, + "grad_norm": 1.0800831317901611, + "learning_rate": 1.3787284717619171e-05, + "loss": 3.6606, + "step": 138300 + }, + { + "epoch": 1.554293546339518, + "grad_norm": 1.030090093612671, + "learning_rate": 1.378503780431632e-05, + "loss": 3.6854, + "step": 138350 + }, + { + "epoch": 1.5548552715098611, + "grad_norm": 1.1533602476119995, + "learning_rate": 1.3782790891013472e-05, + "loss": 3.6155, + "step": 138400 + }, + { + "epoch": 1.5554169966802043, + "grad_norm": 1.198823094367981, + "learning_rate": 1.3780543977710621e-05, + "loss": 3.8225, + "step": 138450 + }, + { + "epoch": 1.5559787218505474, + "grad_norm": 3.5411183834075928, + "learning_rate": 1.377829706440777e-05, + "loss": 3.6767, + "step": 138500 + }, + { + "epoch": 1.5565404470208906, + "grad_norm": 1.1365792751312256, + "learning_rate": 1.377605015110492e-05, + "loss": 3.6306, + "step": 138550 + }, + { + "epoch": 1.5571021721912337, + "grad_norm": 1.0482217073440552, + "learning_rate": 1.377380323780207e-05, + "loss": 3.6779, + "step": 138600 + }, + { + "epoch": 1.5576638973615768, + "grad_norm": 1.256150484085083, + "learning_rate": 1.377155632449922e-05, + "loss": 3.6576, + "step": 138650 + }, + { + "epoch": 1.55822562253192, + "grad_norm": 1.284239649772644, + "learning_rate": 1.3769354349462428e-05, + "loss": 3.602, + "step": 138700 + }, + { + "epoch": 1.5587873477022631, + "grad_norm": 1.1476000547409058, + "learning_rate": 1.3767107436159577e-05, + "loss": 3.7112, + "step": 138750 + }, + { + "epoch": 1.5593490728726063, + "grad_norm": 0.9629308581352234, + "learning_rate": 1.3764860522856726e-05, + "loss": 3.6444, + "step": 138800 + }, + { + "epoch": 1.5599107980429494, + "grad_norm": 1.180854082107544, + "learning_rate": 1.3762613609553877e-05, + "loss": 3.6665, + "step": 138850 + }, + { + "epoch": 1.5604725232132926, + "grad_norm": 1.1548141241073608, + "learning_rate": 1.3760366696251027e-05, + "loss": 3.6188, + "step": 138900 + }, + { + "epoch": 1.5610342483836357, + "grad_norm": 1.1530333757400513, + "learning_rate": 1.3758119782948176e-05, + "loss": 3.6888, + "step": 138950 + }, + { + "epoch": 1.5615959735539788, + "grad_norm": 1.10718834400177, + "learning_rate": 1.3755872869645325e-05, + "loss": 3.6498, + "step": 139000 + }, + { + "epoch": 1.562157698724322, + "grad_norm": 1.4435906410217285, + "learning_rate": 1.3753625956342476e-05, + "loss": 3.7191, + "step": 139050 + }, + { + "epoch": 1.5627194238946653, + "grad_norm": 1.1973544359207153, + "learning_rate": 1.3751379043039626e-05, + "loss": 3.674, + "step": 139100 + }, + { + "epoch": 1.5632811490650085, + "grad_norm": 1.0264248847961426, + "learning_rate": 1.3749132129736775e-05, + "loss": 3.6218, + "step": 139150 + }, + { + "epoch": 1.5638428742353516, + "grad_norm": 1.1173083782196045, + "learning_rate": 1.3746885216433926e-05, + "loss": 3.6692, + "step": 139200 + }, + { + "epoch": 1.5644045994056948, + "grad_norm": 1.0052393674850464, + "learning_rate": 1.3744638303131075e-05, + "loss": 3.7372, + "step": 139250 + }, + { + "epoch": 1.564966324576038, + "grad_norm": 1.0546742677688599, + "learning_rate": 1.3742391389828224e-05, + "loss": 3.7064, + "step": 139300 + }, + { + "epoch": 1.565528049746381, + "grad_norm": 1.112864375114441, + "learning_rate": 1.3740144476525374e-05, + "loss": 3.6565, + "step": 139350 + }, + { + "epoch": 1.5660897749167244, + "grad_norm": 1.0232577323913574, + "learning_rate": 1.3737897563222525e-05, + "loss": 3.7068, + "step": 139400 + }, + { + "epoch": 1.5666515000870676, + "grad_norm": 1.1858417987823486, + "learning_rate": 1.3735650649919674e-05, + "loss": 3.7506, + "step": 139450 + }, + { + "epoch": 1.5672132252574107, + "grad_norm": 1.325637936592102, + "learning_rate": 1.3733403736616823e-05, + "loss": 3.6607, + "step": 139500 + }, + { + "epoch": 1.5677749504277538, + "grad_norm": 1.0840829610824585, + "learning_rate": 1.3731156823313973e-05, + "loss": 3.7168, + "step": 139550 + }, + { + "epoch": 1.568336675598097, + "grad_norm": 1.2827153205871582, + "learning_rate": 1.3728909910011124e-05, + "loss": 3.6482, + "step": 139600 + }, + { + "epoch": 1.5688984007684401, + "grad_norm": 1.183478593826294, + "learning_rate": 1.3726662996708273e-05, + "loss": 3.6704, + "step": 139650 + }, + { + "epoch": 1.5694601259387833, + "grad_norm": 1.265271544456482, + "learning_rate": 1.3724416083405422e-05, + "loss": 3.7252, + "step": 139700 + }, + { + "epoch": 1.5700218511091264, + "grad_norm": 1.3925718069076538, + "learning_rate": 1.3722169170102573e-05, + "loss": 3.698, + "step": 139750 + }, + { + "epoch": 1.5705835762794695, + "grad_norm": 1.0602258443832397, + "learning_rate": 1.3719922256799723e-05, + "loss": 3.6162, + "step": 139800 + }, + { + "epoch": 1.5711453014498127, + "grad_norm": 1.487443447113037, + "learning_rate": 1.3717675343496872e-05, + "loss": 3.6364, + "step": 139850 + }, + { + "epoch": 1.5717070266201558, + "grad_norm": 1.189277172088623, + "learning_rate": 1.3715428430194021e-05, + "loss": 3.618, + "step": 139900 + }, + { + "epoch": 1.572268751790499, + "grad_norm": 1.0855302810668945, + "learning_rate": 1.3713181516891172e-05, + "loss": 3.6734, + "step": 139950 + }, + { + "epoch": 1.572830476960842, + "grad_norm": 1.330960988998413, + "learning_rate": 1.3710934603588322e-05, + "loss": 3.6967, + "step": 140000 + }, + { + "epoch": 1.5733922021311852, + "grad_norm": 1.1523526906967163, + "learning_rate": 1.3708687690285471e-05, + "loss": 3.6636, + "step": 140050 + }, + { + "epoch": 1.5739539273015284, + "grad_norm": 1.225419044494629, + "learning_rate": 1.3706440776982622e-05, + "loss": 3.7668, + "step": 140100 + }, + { + "epoch": 1.5745156524718715, + "grad_norm": 1.1650224924087524, + "learning_rate": 1.3704193863679771e-05, + "loss": 3.6304, + "step": 140150 + }, + { + "epoch": 1.5750773776422147, + "grad_norm": 1.2172008752822876, + "learning_rate": 1.370194695037692e-05, + "loss": 3.7296, + "step": 140200 + }, + { + "epoch": 1.5756391028125578, + "grad_norm": 1.1196541786193848, + "learning_rate": 1.369970003707407e-05, + "loss": 3.69, + "step": 140250 + }, + { + "epoch": 1.576200827982901, + "grad_norm": 1.1689870357513428, + "learning_rate": 1.3697453123771221e-05, + "loss": 3.7265, + "step": 140300 + }, + { + "epoch": 1.576762553153244, + "grad_norm": 1.0626927614212036, + "learning_rate": 1.369520621046837e-05, + "loss": 3.6962, + "step": 140350 + }, + { + "epoch": 1.5773242783235872, + "grad_norm": 1.0944437980651855, + "learning_rate": 1.369295929716552e-05, + "loss": 3.7403, + "step": 140400 + }, + { + "epoch": 1.5778860034939306, + "grad_norm": 1.046596646308899, + "learning_rate": 1.369071238386267e-05, + "loss": 3.6516, + "step": 140450 + }, + { + "epoch": 1.5784477286642737, + "grad_norm": 1.0734217166900635, + "learning_rate": 1.368846547055982e-05, + "loss": 3.756, + "step": 140500 + }, + { + "epoch": 1.5790094538346169, + "grad_norm": 1.1881511211395264, + "learning_rate": 1.368621855725697e-05, + "loss": 3.6859, + "step": 140550 + }, + { + "epoch": 1.57957117900496, + "grad_norm": 1.231651782989502, + "learning_rate": 1.3683971643954119e-05, + "loss": 3.6863, + "step": 140600 + }, + { + "epoch": 1.5801329041753032, + "grad_norm": 1.2965707778930664, + "learning_rate": 1.368172473065127e-05, + "loss": 3.6765, + "step": 140650 + }, + { + "epoch": 1.5806946293456463, + "grad_norm": 1.2120448350906372, + "learning_rate": 1.3679477817348419e-05, + "loss": 3.6023, + "step": 140700 + }, + { + "epoch": 1.5812563545159897, + "grad_norm": 1.1119015216827393, + "learning_rate": 1.3677230904045568e-05, + "loss": 3.7126, + "step": 140750 + }, + { + "epoch": 1.5818180796863328, + "grad_norm": 1.1401653289794922, + "learning_rate": 1.367498399074272e-05, + "loss": 3.6261, + "step": 140800 + }, + { + "epoch": 1.582379804856676, + "grad_norm": 1.1538084745407104, + "learning_rate": 1.3672737077439869e-05, + "loss": 3.7259, + "step": 140850 + }, + { + "epoch": 1.582941530027019, + "grad_norm": 1.398902177810669, + "learning_rate": 1.3670490164137018e-05, + "loss": 3.7078, + "step": 140900 + }, + { + "epoch": 1.5835032551973622, + "grad_norm": 1.173081398010254, + "learning_rate": 1.3668243250834167e-05, + "loss": 3.701, + "step": 140950 + }, + { + "epoch": 1.5840649803677054, + "grad_norm": 1.103658676147461, + "learning_rate": 1.3665996337531318e-05, + "loss": 3.6483, + "step": 141000 + }, + { + "epoch": 1.5846267055380485, + "grad_norm": 1.2359012365341187, + "learning_rate": 1.3663749424228468e-05, + "loss": 3.6826, + "step": 141050 + }, + { + "epoch": 1.5851884307083917, + "grad_norm": 1.2011862993240356, + "learning_rate": 1.3661502510925617e-05, + "loss": 3.7179, + "step": 141100 + }, + { + "epoch": 1.5857501558787348, + "grad_norm": 1.0387769937515259, + "learning_rate": 1.3659255597622766e-05, + "loss": 3.7604, + "step": 141150 + }, + { + "epoch": 1.586311881049078, + "grad_norm": 1.2249939441680908, + "learning_rate": 1.3657008684319917e-05, + "loss": 3.6343, + "step": 141200 + }, + { + "epoch": 1.586873606219421, + "grad_norm": 1.3379638195037842, + "learning_rate": 1.3654761771017066e-05, + "loss": 3.7245, + "step": 141250 + }, + { + "epoch": 1.5874353313897642, + "grad_norm": 1.1127783060073853, + "learning_rate": 1.3652559795980274e-05, + "loss": 3.6998, + "step": 141300 + }, + { + "epoch": 1.5879970565601074, + "grad_norm": 1.1300557851791382, + "learning_rate": 1.3650312882677423e-05, + "loss": 3.5868, + "step": 141350 + }, + { + "epoch": 1.5885587817304505, + "grad_norm": 1.0959243774414062, + "learning_rate": 1.3648065969374573e-05, + "loss": 3.6953, + "step": 141400 + }, + { + "epoch": 1.5891205069007937, + "grad_norm": 1.0239161252975464, + "learning_rate": 1.3645819056071724e-05, + "loss": 3.6807, + "step": 141450 + }, + { + "epoch": 1.5896822320711368, + "grad_norm": 1.115818977355957, + "learning_rate": 1.3643572142768873e-05, + "loss": 3.6034, + "step": 141500 + }, + { + "epoch": 1.59024395724148, + "grad_norm": 1.1584968566894531, + "learning_rate": 1.3641325229466022e-05, + "loss": 3.6404, + "step": 141550 + }, + { + "epoch": 1.590805682411823, + "grad_norm": 1.4609769582748413, + "learning_rate": 1.3639078316163172e-05, + "loss": 3.7024, + "step": 141600 + }, + { + "epoch": 1.5913674075821662, + "grad_norm": 1.1476662158966064, + "learning_rate": 1.3636831402860323e-05, + "loss": 3.6814, + "step": 141650 + }, + { + "epoch": 1.5919291327525094, + "grad_norm": 1.1682904958724976, + "learning_rate": 1.3634584489557472e-05, + "loss": 3.7362, + "step": 141700 + }, + { + "epoch": 1.5924908579228527, + "grad_norm": 1.1838876008987427, + "learning_rate": 1.3632337576254621e-05, + "loss": 3.6828, + "step": 141750 + }, + { + "epoch": 1.5930525830931959, + "grad_norm": 1.0836089849472046, + "learning_rate": 1.3630090662951772e-05, + "loss": 3.727, + "step": 141800 + }, + { + "epoch": 1.593614308263539, + "grad_norm": 1.0461901426315308, + "learning_rate": 1.3627843749648922e-05, + "loss": 3.6681, + "step": 141850 + }, + { + "epoch": 1.5941760334338821, + "grad_norm": 1.067426323890686, + "learning_rate": 1.3625596836346071e-05, + "loss": 3.6756, + "step": 141900 + }, + { + "epoch": 1.5947377586042253, + "grad_norm": 1.1616348028182983, + "learning_rate": 1.362334992304322e-05, + "loss": 3.6586, + "step": 141950 + }, + { + "epoch": 1.5952994837745684, + "grad_norm": 1.3039299249649048, + "learning_rate": 1.3621103009740371e-05, + "loss": 3.5232, + "step": 142000 + }, + { + "epoch": 1.5958612089449116, + "grad_norm": 1.2461251020431519, + "learning_rate": 1.361885609643752e-05, + "loss": 3.6559, + "step": 142050 + }, + { + "epoch": 1.596422934115255, + "grad_norm": 1.3545992374420166, + "learning_rate": 1.361660918313467e-05, + "loss": 3.7266, + "step": 142100 + }, + { + "epoch": 1.596984659285598, + "grad_norm": 2.3668854236602783, + "learning_rate": 1.361436226983182e-05, + "loss": 3.692, + "step": 142150 + }, + { + "epoch": 1.5975463844559412, + "grad_norm": 1.298037052154541, + "learning_rate": 1.361211535652897e-05, + "loss": 3.6174, + "step": 142200 + }, + { + "epoch": 1.5981081096262844, + "grad_norm": 1.0800282955169678, + "learning_rate": 1.360986844322612e-05, + "loss": 3.6606, + "step": 142250 + }, + { + "epoch": 1.5986698347966275, + "grad_norm": 1.1888720989227295, + "learning_rate": 1.3607621529923269e-05, + "loss": 3.6515, + "step": 142300 + }, + { + "epoch": 1.5992315599669706, + "grad_norm": 1.1780089139938354, + "learning_rate": 1.360537461662042e-05, + "loss": 3.7173, + "step": 142350 + }, + { + "epoch": 1.5997932851373138, + "grad_norm": 1.1620633602142334, + "learning_rate": 1.360312770331757e-05, + "loss": 3.6857, + "step": 142400 + }, + { + "epoch": 1.600355010307657, + "grad_norm": 1.112202525138855, + "learning_rate": 1.3600880790014719e-05, + "loss": 3.6677, + "step": 142450 + }, + { + "epoch": 1.600916735478, + "grad_norm": 1.1175018548965454, + "learning_rate": 1.3598633876711868e-05, + "loss": 3.6737, + "step": 142500 + }, + { + "epoch": 1.6014784606483432, + "grad_norm": 1.0645023584365845, + "learning_rate": 1.3596386963409019e-05, + "loss": 3.6471, + "step": 142550 + }, + { + "epoch": 1.6020401858186863, + "grad_norm": 1.084029197692871, + "learning_rate": 1.3594140050106168e-05, + "loss": 3.6971, + "step": 142600 + }, + { + "epoch": 1.6026019109890295, + "grad_norm": 1.1282392740249634, + "learning_rate": 1.3591893136803318e-05, + "loss": 3.7043, + "step": 142650 + }, + { + "epoch": 1.6031636361593726, + "grad_norm": 1.0407531261444092, + "learning_rate": 1.3589646223500469e-05, + "loss": 3.6152, + "step": 142700 + }, + { + "epoch": 1.6037253613297158, + "grad_norm": 1.0244865417480469, + "learning_rate": 1.3587399310197618e-05, + "loss": 3.7487, + "step": 142750 + }, + { + "epoch": 1.604287086500059, + "grad_norm": 1.1283087730407715, + "learning_rate": 1.3585152396894767e-05, + "loss": 3.6107, + "step": 142800 + }, + { + "epoch": 1.604848811670402, + "grad_norm": 1.0483598709106445, + "learning_rate": 1.3582905483591916e-05, + "loss": 3.6978, + "step": 142850 + }, + { + "epoch": 1.6054105368407452, + "grad_norm": 1.1567394733428955, + "learning_rate": 1.3580658570289067e-05, + "loss": 3.6641, + "step": 142900 + }, + { + "epoch": 1.6059722620110883, + "grad_norm": 1.137740969657898, + "learning_rate": 1.3578456595252273e-05, + "loss": 3.6434, + "step": 142950 + }, + { + "epoch": 1.6065339871814315, + "grad_norm": 1.0882683992385864, + "learning_rate": 1.3576209681949424e-05, + "loss": 3.7948, + "step": 143000 + }, + { + "epoch": 1.6070957123517746, + "grad_norm": 1.1528105735778809, + "learning_rate": 1.3573962768646574e-05, + "loss": 3.6705, + "step": 143050 + }, + { + "epoch": 1.607657437522118, + "grad_norm": 1.1751548051834106, + "learning_rate": 1.3571715855343723e-05, + "loss": 3.636, + "step": 143100 + }, + { + "epoch": 1.6082191626924611, + "grad_norm": 1.5588665008544922, + "learning_rate": 1.3569468942040872e-05, + "loss": 3.7752, + "step": 143150 + }, + { + "epoch": 1.6087808878628043, + "grad_norm": 1.3791288137435913, + "learning_rate": 1.3567222028738023e-05, + "loss": 3.6663, + "step": 143200 + }, + { + "epoch": 1.6093426130331474, + "grad_norm": 1.1276522874832153, + "learning_rate": 1.3564975115435173e-05, + "loss": 3.593, + "step": 143250 + }, + { + "epoch": 1.6099043382034905, + "grad_norm": 1.1951262950897217, + "learning_rate": 1.3562728202132322e-05, + "loss": 3.6767, + "step": 143300 + }, + { + "epoch": 1.6104660633738337, + "grad_norm": 1.2520290613174438, + "learning_rate": 1.3560481288829473e-05, + "loss": 3.6266, + "step": 143350 + }, + { + "epoch": 1.6110277885441768, + "grad_norm": 1.223700761795044, + "learning_rate": 1.3558234375526622e-05, + "loss": 3.6326, + "step": 143400 + }, + { + "epoch": 1.6115895137145202, + "grad_norm": 1.3785969018936157, + "learning_rate": 1.3555987462223772e-05, + "loss": 3.6384, + "step": 143450 + }, + { + "epoch": 1.6121512388848633, + "grad_norm": 1.168166995048523, + "learning_rate": 1.3553740548920921e-05, + "loss": 3.672, + "step": 143500 + }, + { + "epoch": 1.6127129640552065, + "grad_norm": 1.169132113456726, + "learning_rate": 1.3551493635618072e-05, + "loss": 3.6733, + "step": 143550 + }, + { + "epoch": 1.6132746892255496, + "grad_norm": 1.4326705932617188, + "learning_rate": 1.3549246722315221e-05, + "loss": 3.7418, + "step": 143600 + }, + { + "epoch": 1.6138364143958928, + "grad_norm": 1.1875486373901367, + "learning_rate": 1.354699980901237e-05, + "loss": 3.6875, + "step": 143650 + }, + { + "epoch": 1.614398139566236, + "grad_norm": 0.9022040367126465, + "learning_rate": 1.3544752895709522e-05, + "loss": 3.7225, + "step": 143700 + }, + { + "epoch": 1.614959864736579, + "grad_norm": 1.1078884601593018, + "learning_rate": 1.3542505982406671e-05, + "loss": 3.6703, + "step": 143750 + }, + { + "epoch": 1.6155215899069222, + "grad_norm": 1.1090278625488281, + "learning_rate": 1.354025906910382e-05, + "loss": 3.6247, + "step": 143800 + }, + { + "epoch": 1.6160833150772653, + "grad_norm": 1.1982507705688477, + "learning_rate": 1.353801215580097e-05, + "loss": 3.6545, + "step": 143850 + }, + { + "epoch": 1.6166450402476085, + "grad_norm": 1.2535675764083862, + "learning_rate": 1.353576524249812e-05, + "loss": 3.7138, + "step": 143900 + }, + { + "epoch": 1.6172067654179516, + "grad_norm": 1.0719192028045654, + "learning_rate": 1.353351832919527e-05, + "loss": 3.6688, + "step": 143950 + }, + { + "epoch": 1.6177684905882947, + "grad_norm": 0.994109034538269, + "learning_rate": 1.353127141589242e-05, + "loss": 3.7049, + "step": 144000 + }, + { + "epoch": 1.618330215758638, + "grad_norm": 1.3408006429672241, + "learning_rate": 1.352902450258957e-05, + "loss": 3.6536, + "step": 144050 + }, + { + "epoch": 1.618891940928981, + "grad_norm": 1.1289949417114258, + "learning_rate": 1.352677758928672e-05, + "loss": 3.6695, + "step": 144100 + }, + { + "epoch": 1.6194536660993242, + "grad_norm": 1.5051370859146118, + "learning_rate": 1.3524530675983869e-05, + "loss": 3.6928, + "step": 144150 + }, + { + "epoch": 1.6200153912696673, + "grad_norm": 1.1607725620269775, + "learning_rate": 1.3522283762681018e-05, + "loss": 3.6913, + "step": 144200 + }, + { + "epoch": 1.6205771164400105, + "grad_norm": 1.049977421760559, + "learning_rate": 1.352003684937817e-05, + "loss": 3.6361, + "step": 144250 + }, + { + "epoch": 1.6211388416103536, + "grad_norm": 0.9868359565734863, + "learning_rate": 1.3517789936075318e-05, + "loss": 3.7028, + "step": 144300 + }, + { + "epoch": 1.6217005667806967, + "grad_norm": 1.1460314989089966, + "learning_rate": 1.3515543022772468e-05, + "loss": 3.7152, + "step": 144350 + }, + { + "epoch": 1.6222622919510399, + "grad_norm": 1.1550520658493042, + "learning_rate": 1.3513296109469619e-05, + "loss": 3.7497, + "step": 144400 + }, + { + "epoch": 1.6228240171213832, + "grad_norm": 1.182363510131836, + "learning_rate": 1.3511049196166768e-05, + "loss": 3.6572, + "step": 144450 + }, + { + "epoch": 1.6233857422917264, + "grad_norm": 1.1829805374145508, + "learning_rate": 1.3508802282863917e-05, + "loss": 3.633, + "step": 144500 + }, + { + "epoch": 1.6239474674620695, + "grad_norm": 1.1781424283981323, + "learning_rate": 1.3506555369561067e-05, + "loss": 3.7259, + "step": 144550 + }, + { + "epoch": 1.6245091926324127, + "grad_norm": 1.1826103925704956, + "learning_rate": 1.3504308456258218e-05, + "loss": 3.6989, + "step": 144600 + }, + { + "epoch": 1.6250709178027558, + "grad_norm": 1.1167058944702148, + "learning_rate": 1.3502061542955367e-05, + "loss": 3.7153, + "step": 144650 + }, + { + "epoch": 1.625632642973099, + "grad_norm": 1.5774495601654053, + "learning_rate": 1.3499814629652516e-05, + "loss": 3.6014, + "step": 144700 + }, + { + "epoch": 1.626194368143442, + "grad_norm": 1.1393110752105713, + "learning_rate": 1.3497567716349666e-05, + "loss": 3.7552, + "step": 144750 + }, + { + "epoch": 1.6267560933137855, + "grad_norm": 1.0390650033950806, + "learning_rate": 1.3495320803046817e-05, + "loss": 3.6646, + "step": 144800 + }, + { + "epoch": 1.6273178184841286, + "grad_norm": 1.3771274089813232, + "learning_rate": 1.3493073889743966e-05, + "loss": 3.7414, + "step": 144850 + }, + { + "epoch": 1.6278795436544717, + "grad_norm": 1.2277251482009888, + "learning_rate": 1.3490826976441115e-05, + "loss": 3.7624, + "step": 144900 + }, + { + "epoch": 1.6284412688248149, + "grad_norm": 1.057714581489563, + "learning_rate": 1.3488580063138266e-05, + "loss": 3.747, + "step": 144950 + }, + { + "epoch": 1.629002993995158, + "grad_norm": 1.2887024879455566, + "learning_rate": 1.3486333149835416e-05, + "loss": 3.6581, + "step": 145000 + }, + { + "epoch": 1.6295647191655012, + "grad_norm": 1.094120740890503, + "learning_rate": 1.3484086236532565e-05, + "loss": 3.6541, + "step": 145050 + }, + { + "epoch": 1.6301264443358443, + "grad_norm": 1.047278881072998, + "learning_rate": 1.3481839323229714e-05, + "loss": 3.6114, + "step": 145100 + }, + { + "epoch": 1.6306881695061874, + "grad_norm": 1.2887330055236816, + "learning_rate": 1.3479592409926865e-05, + "loss": 3.7136, + "step": 145150 + }, + { + "epoch": 1.6312498946765306, + "grad_norm": 1.118496298789978, + "learning_rate": 1.3477345496624015e-05, + "loss": 3.7317, + "step": 145200 + }, + { + "epoch": 1.6318116198468737, + "grad_norm": 1.2175357341766357, + "learning_rate": 1.3475098583321164e-05, + "loss": 3.6609, + "step": 145250 + }, + { + "epoch": 1.6323733450172169, + "grad_norm": 1.1313681602478027, + "learning_rate": 1.3472851670018315e-05, + "loss": 3.6874, + "step": 145300 + }, + { + "epoch": 1.63293507018756, + "grad_norm": 1.1100246906280518, + "learning_rate": 1.3470604756715464e-05, + "loss": 3.6398, + "step": 145350 + }, + { + "epoch": 1.6334967953579032, + "grad_norm": 1.0380334854125977, + "learning_rate": 1.3468357843412614e-05, + "loss": 3.6961, + "step": 145400 + }, + { + "epoch": 1.6340585205282463, + "grad_norm": 1.1742682456970215, + "learning_rate": 1.3466110930109763e-05, + "loss": 3.6451, + "step": 145450 + }, + { + "epoch": 1.6346202456985894, + "grad_norm": 1.14945650100708, + "learning_rate": 1.3463864016806914e-05, + "loss": 3.6867, + "step": 145500 + }, + { + "epoch": 1.6351819708689326, + "grad_norm": 1.1547523736953735, + "learning_rate": 1.3461617103504063e-05, + "loss": 3.7471, + "step": 145550 + }, + { + "epoch": 1.6357436960392757, + "grad_norm": 1.1640186309814453, + "learning_rate": 1.3459370190201213e-05, + "loss": 3.6456, + "step": 145600 + }, + { + "epoch": 1.6363054212096189, + "grad_norm": 1.2586885690689087, + "learning_rate": 1.3457123276898364e-05, + "loss": 3.721, + "step": 145650 + }, + { + "epoch": 1.636867146379962, + "grad_norm": 1.0116692781448364, + "learning_rate": 1.3454876363595513e-05, + "loss": 3.6787, + "step": 145700 + }, + { + "epoch": 1.6374288715503051, + "grad_norm": 1.1636160612106323, + "learning_rate": 1.3452629450292662e-05, + "loss": 3.6964, + "step": 145750 + }, + { + "epoch": 1.6379905967206485, + "grad_norm": 1.192798376083374, + "learning_rate": 1.3450382536989812e-05, + "loss": 3.6242, + "step": 145800 + }, + { + "epoch": 1.6385523218909916, + "grad_norm": 1.0549873113632202, + "learning_rate": 1.3448135623686963e-05, + "loss": 3.7125, + "step": 145850 + }, + { + "epoch": 1.6391140470613348, + "grad_norm": 1.1213629245758057, + "learning_rate": 1.3445888710384112e-05, + "loss": 3.6865, + "step": 145900 + }, + { + "epoch": 1.639675772231678, + "grad_norm": 1.334212303161621, + "learning_rate": 1.3443641797081261e-05, + "loss": 3.6126, + "step": 145950 + }, + { + "epoch": 1.640237497402021, + "grad_norm": 1.2307395935058594, + "learning_rate": 1.3441394883778412e-05, + "loss": 3.6608, + "step": 146000 + }, + { + "epoch": 1.6407992225723642, + "grad_norm": 1.046105980873108, + "learning_rate": 1.3439147970475562e-05, + "loss": 3.6599, + "step": 146050 + }, + { + "epoch": 1.6413609477427074, + "grad_norm": 1.116716980934143, + "learning_rate": 1.343690105717271e-05, + "loss": 3.6798, + "step": 146100 + }, + { + "epoch": 1.6419226729130507, + "grad_norm": 1.1969423294067383, + "learning_rate": 1.343465414386986e-05, + "loss": 3.612, + "step": 146150 + }, + { + "epoch": 1.6424843980833939, + "grad_norm": 1.1656075716018677, + "learning_rate": 1.3432407230567011e-05, + "loss": 3.6648, + "step": 146200 + }, + { + "epoch": 1.643046123253737, + "grad_norm": 1.1140774488449097, + "learning_rate": 1.343016031726416e-05, + "loss": 3.6866, + "step": 146250 + }, + { + "epoch": 1.6436078484240801, + "grad_norm": 1.15651273727417, + "learning_rate": 1.342791340396131e-05, + "loss": 3.6857, + "step": 146300 + }, + { + "epoch": 1.6441695735944233, + "grad_norm": 1.2643494606018066, + "learning_rate": 1.3425666490658459e-05, + "loss": 3.6299, + "step": 146350 + }, + { + "epoch": 1.6447312987647664, + "grad_norm": 1.1954561471939087, + "learning_rate": 1.342341957735561e-05, + "loss": 3.5863, + "step": 146400 + }, + { + "epoch": 1.6452930239351096, + "grad_norm": 1.0443463325500488, + "learning_rate": 1.342117266405276e-05, + "loss": 3.6834, + "step": 146450 + }, + { + "epoch": 1.6458547491054527, + "grad_norm": 1.190408706665039, + "learning_rate": 1.3418925750749909e-05, + "loss": 3.7262, + "step": 146500 + }, + { + "epoch": 1.6464164742757958, + "grad_norm": 1.1993037462234497, + "learning_rate": 1.341667883744706e-05, + "loss": 3.6069, + "step": 146550 + }, + { + "epoch": 1.646978199446139, + "grad_norm": 1.260898470878601, + "learning_rate": 1.3414431924144209e-05, + "loss": 3.6386, + "step": 146600 + }, + { + "epoch": 1.6475399246164821, + "grad_norm": 1.1906691789627075, + "learning_rate": 1.3412185010841358e-05, + "loss": 3.7299, + "step": 146650 + }, + { + "epoch": 1.6481016497868253, + "grad_norm": 0.9800334572792053, + "learning_rate": 1.3409938097538508e-05, + "loss": 3.6911, + "step": 146700 + }, + { + "epoch": 1.6486633749571684, + "grad_norm": 1.2064788341522217, + "learning_rate": 1.3407691184235659e-05, + "loss": 3.6661, + "step": 146750 + }, + { + "epoch": 1.6492251001275116, + "grad_norm": 1.5114845037460327, + "learning_rate": 1.3405444270932808e-05, + "loss": 3.67, + "step": 146800 + }, + { + "epoch": 1.6497868252978547, + "grad_norm": 1.1322213411331177, + "learning_rate": 1.3403197357629957e-05, + "loss": 3.6547, + "step": 146850 + }, + { + "epoch": 1.6503485504681978, + "grad_norm": 1.2612017393112183, + "learning_rate": 1.3400950444327108e-05, + "loss": 3.7193, + "step": 146900 + }, + { + "epoch": 1.650910275638541, + "grad_norm": 1.2148751020431519, + "learning_rate": 1.3398703531024258e-05, + "loss": 3.6716, + "step": 146950 + }, + { + "epoch": 1.6514720008088841, + "grad_norm": 1.3447052240371704, + "learning_rate": 1.3396456617721407e-05, + "loss": 3.6779, + "step": 147000 + }, + { + "epoch": 1.6520337259792273, + "grad_norm": 1.0935733318328857, + "learning_rate": 1.3394209704418556e-05, + "loss": 3.6086, + "step": 147050 + }, + { + "epoch": 1.6525954511495704, + "grad_norm": 1.6140910387039185, + "learning_rate": 1.3392007729381762e-05, + "loss": 3.7257, + "step": 147100 + }, + { + "epoch": 1.6531571763199138, + "grad_norm": 1.2870099544525146, + "learning_rate": 1.3389760816078912e-05, + "loss": 3.6534, + "step": 147150 + }, + { + "epoch": 1.653718901490257, + "grad_norm": 1.1969412565231323, + "learning_rate": 1.3387513902776061e-05, + "loss": 3.608, + "step": 147200 + }, + { + "epoch": 1.6542806266606, + "grad_norm": 1.3302042484283447, + "learning_rate": 1.3385266989473212e-05, + "loss": 3.6281, + "step": 147250 + }, + { + "epoch": 1.6548423518309432, + "grad_norm": 1.2988100051879883, + "learning_rate": 1.3383020076170361e-05, + "loss": 3.6913, + "step": 147300 + }, + { + "epoch": 1.6554040770012863, + "grad_norm": 1.3460708856582642, + "learning_rate": 1.338077316286751e-05, + "loss": 3.659, + "step": 147350 + }, + { + "epoch": 1.6559658021716295, + "grad_norm": 1.122812271118164, + "learning_rate": 1.337852624956466e-05, + "loss": 3.722, + "step": 147400 + }, + { + "epoch": 1.6565275273419726, + "grad_norm": 1.312804937362671, + "learning_rate": 1.337627933626181e-05, + "loss": 3.6989, + "step": 147450 + }, + { + "epoch": 1.657089252512316, + "grad_norm": 1.2745600938796997, + "learning_rate": 1.337403242295896e-05, + "loss": 3.7124, + "step": 147500 + }, + { + "epoch": 1.6576509776826591, + "grad_norm": 1.1472702026367188, + "learning_rate": 1.337178550965611e-05, + "loss": 3.7748, + "step": 147550 + }, + { + "epoch": 1.6582127028530023, + "grad_norm": 1.2738385200500488, + "learning_rate": 1.336953859635326e-05, + "loss": 3.7381, + "step": 147600 + }, + { + "epoch": 1.6587744280233454, + "grad_norm": 1.1442394256591797, + "learning_rate": 1.336729168305041e-05, + "loss": 3.777, + "step": 147650 + }, + { + "epoch": 1.6593361531936885, + "grad_norm": 1.043246865272522, + "learning_rate": 1.3365044769747559e-05, + "loss": 3.7133, + "step": 147700 + }, + { + "epoch": 1.6598978783640317, + "grad_norm": 1.2269457578659058, + "learning_rate": 1.3362797856444708e-05, + "loss": 3.7473, + "step": 147750 + }, + { + "epoch": 1.6604596035343748, + "grad_norm": 1.331426739692688, + "learning_rate": 1.336055094314186e-05, + "loss": 3.7658, + "step": 147800 + }, + { + "epoch": 1.661021328704718, + "grad_norm": 1.048108458518982, + "learning_rate": 1.3358304029839009e-05, + "loss": 3.6972, + "step": 147850 + }, + { + "epoch": 1.6615830538750611, + "grad_norm": 1.126671314239502, + "learning_rate": 1.3356057116536158e-05, + "loss": 3.643, + "step": 147900 + }, + { + "epoch": 1.6621447790454043, + "grad_norm": 1.2158805131912231, + "learning_rate": 1.3353810203233309e-05, + "loss": 3.6418, + "step": 147950 + }, + { + "epoch": 1.6627065042157474, + "grad_norm": 1.0428258180618286, + "learning_rate": 1.3351563289930458e-05, + "loss": 3.6805, + "step": 148000 + }, + { + "epoch": 1.6632682293860905, + "grad_norm": 1.3096319437026978, + "learning_rate": 1.3349316376627608e-05, + "loss": 3.6689, + "step": 148050 + }, + { + "epoch": 1.6638299545564337, + "grad_norm": 1.0405086278915405, + "learning_rate": 1.3347069463324757e-05, + "loss": 3.7214, + "step": 148100 + }, + { + "epoch": 1.6643916797267768, + "grad_norm": 1.9383026361465454, + "learning_rate": 1.3344822550021908e-05, + "loss": 3.6636, + "step": 148150 + }, + { + "epoch": 1.66495340489712, + "grad_norm": 1.1549087762832642, + "learning_rate": 1.3342575636719057e-05, + "loss": 3.6556, + "step": 148200 + }, + { + "epoch": 1.665515130067463, + "grad_norm": 1.1671701669692993, + "learning_rate": 1.3340328723416207e-05, + "loss": 3.6536, + "step": 148250 + }, + { + "epoch": 1.6660768552378062, + "grad_norm": 1.178333044052124, + "learning_rate": 1.3338081810113358e-05, + "loss": 3.6632, + "step": 148300 + }, + { + "epoch": 1.6666385804081494, + "grad_norm": 1.2279329299926758, + "learning_rate": 1.3335834896810507e-05, + "loss": 3.7197, + "step": 148350 + }, + { + "epoch": 1.6672003055784925, + "grad_norm": 1.4624500274658203, + "learning_rate": 1.3333587983507656e-05, + "loss": 3.6903, + "step": 148400 + }, + { + "epoch": 1.6677620307488357, + "grad_norm": 1.1612122058868408, + "learning_rate": 1.3331341070204806e-05, + "loss": 3.6885, + "step": 148450 + }, + { + "epoch": 1.668323755919179, + "grad_norm": 1.1029609441757202, + "learning_rate": 1.3329094156901957e-05, + "loss": 3.6618, + "step": 148500 + }, + { + "epoch": 1.6688854810895222, + "grad_norm": 1.0497239828109741, + "learning_rate": 1.3326847243599106e-05, + "loss": 3.6376, + "step": 148550 + }, + { + "epoch": 1.6694472062598653, + "grad_norm": 1.1928558349609375, + "learning_rate": 1.3324600330296255e-05, + "loss": 3.6673, + "step": 148600 + }, + { + "epoch": 1.6700089314302085, + "grad_norm": 1.1223574876785278, + "learning_rate": 1.3322353416993406e-05, + "loss": 3.6929, + "step": 148650 + }, + { + "epoch": 1.6705706566005516, + "grad_norm": 1.122258186340332, + "learning_rate": 1.3320106503690556e-05, + "loss": 3.6343, + "step": 148700 + }, + { + "epoch": 1.6711323817708947, + "grad_norm": 1.1371307373046875, + "learning_rate": 1.3317859590387705e-05, + "loss": 3.5773, + "step": 148750 + }, + { + "epoch": 1.6716941069412379, + "grad_norm": 1.0262322425842285, + "learning_rate": 1.3315612677084854e-05, + "loss": 3.7727, + "step": 148800 + }, + { + "epoch": 1.6722558321115812, + "grad_norm": 1.0024133920669556, + "learning_rate": 1.3313365763782005e-05, + "loss": 3.7303, + "step": 148850 + }, + { + "epoch": 1.6728175572819244, + "grad_norm": 1.2706538438796997, + "learning_rate": 1.3311118850479155e-05, + "loss": 3.7602, + "step": 148900 + }, + { + "epoch": 1.6733792824522675, + "grad_norm": 1.0769323110580444, + "learning_rate": 1.3308871937176304e-05, + "loss": 3.6039, + "step": 148950 + }, + { + "epoch": 1.6739410076226107, + "grad_norm": 0.9288907051086426, + "learning_rate": 1.3306625023873453e-05, + "loss": 3.6582, + "step": 149000 + }, + { + "epoch": 1.6745027327929538, + "grad_norm": 1.0951144695281982, + "learning_rate": 1.3304378110570604e-05, + "loss": 3.6777, + "step": 149050 + }, + { + "epoch": 1.675064457963297, + "grad_norm": 1.2050976753234863, + "learning_rate": 1.3302131197267754e-05, + "loss": 3.6267, + "step": 149100 + }, + { + "epoch": 1.67562618313364, + "grad_norm": 1.0500702857971191, + "learning_rate": 1.3299884283964903e-05, + "loss": 3.6249, + "step": 149150 + }, + { + "epoch": 1.6761879083039832, + "grad_norm": 1.0992419719696045, + "learning_rate": 1.3297637370662054e-05, + "loss": 3.7171, + "step": 149200 + }, + { + "epoch": 1.6767496334743264, + "grad_norm": 1.1110466718673706, + "learning_rate": 1.3295390457359203e-05, + "loss": 3.6624, + "step": 149250 + }, + { + "epoch": 1.6773113586446695, + "grad_norm": 1.0994672775268555, + "learning_rate": 1.3293143544056352e-05, + "loss": 3.6212, + "step": 149300 + }, + { + "epoch": 1.6778730838150127, + "grad_norm": 1.1601837873458862, + "learning_rate": 1.3290896630753502e-05, + "loss": 3.5943, + "step": 149350 + }, + { + "epoch": 1.6784348089853558, + "grad_norm": 1.0538126230239868, + "learning_rate": 1.3288649717450653e-05, + "loss": 3.7741, + "step": 149400 + }, + { + "epoch": 1.678996534155699, + "grad_norm": 1.2923500537872314, + "learning_rate": 1.3286402804147802e-05, + "loss": 3.6315, + "step": 149450 + }, + { + "epoch": 1.679558259326042, + "grad_norm": 1.4426004886627197, + "learning_rate": 1.3284155890844951e-05, + "loss": 3.6831, + "step": 149500 + }, + { + "epoch": 1.6801199844963852, + "grad_norm": 1.1323589086532593, + "learning_rate": 1.3281908977542102e-05, + "loss": 3.6512, + "step": 149550 + }, + { + "epoch": 1.6806817096667284, + "grad_norm": 1.238003134727478, + "learning_rate": 1.3279662064239252e-05, + "loss": 3.711, + "step": 149600 + }, + { + "epoch": 1.6812434348370715, + "grad_norm": 1.192116379737854, + "learning_rate": 1.3277415150936401e-05, + "loss": 3.7677, + "step": 149650 + }, + { + "epoch": 1.6818051600074146, + "grad_norm": 1.178747296333313, + "learning_rate": 1.327516823763355e-05, + "loss": 3.6192, + "step": 149700 + }, + { + "epoch": 1.6823668851777578, + "grad_norm": 1.0546813011169434, + "learning_rate": 1.3272921324330701e-05, + "loss": 3.6233, + "step": 149750 + }, + { + "epoch": 1.682928610348101, + "grad_norm": 1.1599066257476807, + "learning_rate": 1.327067441102785e-05, + "loss": 3.7189, + "step": 149800 + }, + { + "epoch": 1.6834903355184443, + "grad_norm": 1.2025222778320312, + "learning_rate": 1.3268427497725e-05, + "loss": 3.6428, + "step": 149850 + }, + { + "epoch": 1.6840520606887874, + "grad_norm": 1.1144520044326782, + "learning_rate": 1.3266180584422151e-05, + "loss": 3.6962, + "step": 149900 + }, + { + "epoch": 1.6846137858591306, + "grad_norm": 1.1810729503631592, + "learning_rate": 1.32639336711193e-05, + "loss": 3.6183, + "step": 149950 + }, + { + "epoch": 1.6851755110294737, + "grad_norm": 3.7173421382904053, + "learning_rate": 1.326168675781645e-05, + "loss": 3.6547, + "step": 150000 + }, + { + "epoch": 1.6857372361998169, + "grad_norm": 1.180238962173462, + "learning_rate": 1.3259439844513599e-05, + "loss": 3.69, + "step": 150050 + }, + { + "epoch": 1.68629896137016, + "grad_norm": 1.4270219802856445, + "learning_rate": 1.325719293121075e-05, + "loss": 3.6633, + "step": 150100 + }, + { + "epoch": 1.6868606865405031, + "grad_norm": 1.2268694639205933, + "learning_rate": 1.32549460179079e-05, + "loss": 3.6613, + "step": 150150 + }, + { + "epoch": 1.6874224117108465, + "grad_norm": 1.1365429162979126, + "learning_rate": 1.3252699104605049e-05, + "loss": 3.6612, + "step": 150200 + }, + { + "epoch": 1.6879841368811896, + "grad_norm": 1.1073280572891235, + "learning_rate": 1.32504521913022e-05, + "loss": 3.5813, + "step": 150250 + }, + { + "epoch": 1.6885458620515328, + "grad_norm": 1.258264422416687, + "learning_rate": 1.3248205277999349e-05, + "loss": 3.6542, + "step": 150300 + }, + { + "epoch": 1.689107587221876, + "grad_norm": 1.1778461933135986, + "learning_rate": 1.3245958364696498e-05, + "loss": 3.6699, + "step": 150350 + }, + { + "epoch": 1.689669312392219, + "grad_norm": 1.2182080745697021, + "learning_rate": 1.3243711451393648e-05, + "loss": 3.6747, + "step": 150400 + }, + { + "epoch": 1.6902310375625622, + "grad_norm": 1.2940605878829956, + "learning_rate": 1.3241464538090799e-05, + "loss": 3.6862, + "step": 150450 + }, + { + "epoch": 1.6907927627329054, + "grad_norm": 1.1755141019821167, + "learning_rate": 1.3239217624787948e-05, + "loss": 3.7816, + "step": 150500 + }, + { + "epoch": 1.6913544879032485, + "grad_norm": 1.064827799797058, + "learning_rate": 1.3236970711485097e-05, + "loss": 3.6505, + "step": 150550 + }, + { + "epoch": 1.6919162130735916, + "grad_norm": 1.185843825340271, + "learning_rate": 1.3234723798182247e-05, + "loss": 3.6531, + "step": 150600 + }, + { + "epoch": 1.6924779382439348, + "grad_norm": 1.4173972606658936, + "learning_rate": 1.3232476884879398e-05, + "loss": 3.747, + "step": 150650 + }, + { + "epoch": 1.693039663414278, + "grad_norm": 1.2260687351226807, + "learning_rate": 1.3230229971576547e-05, + "loss": 3.6408, + "step": 150700 + }, + { + "epoch": 1.693601388584621, + "grad_norm": 1.0594072341918945, + "learning_rate": 1.3227983058273696e-05, + "loss": 3.6188, + "step": 150750 + }, + { + "epoch": 1.6941631137549642, + "grad_norm": 1.0781843662261963, + "learning_rate": 1.3225736144970847e-05, + "loss": 3.6484, + "step": 150800 + }, + { + "epoch": 1.6947248389253073, + "grad_norm": 1.171908974647522, + "learning_rate": 1.3223489231667997e-05, + "loss": 3.7537, + "step": 150850 + }, + { + "epoch": 1.6952865640956505, + "grad_norm": 1.2834751605987549, + "learning_rate": 1.3221242318365146e-05, + "loss": 3.65, + "step": 150900 + }, + { + "epoch": 1.6958482892659936, + "grad_norm": 1.113500952720642, + "learning_rate": 1.3218995405062295e-05, + "loss": 3.6374, + "step": 150950 + }, + { + "epoch": 1.6964100144363368, + "grad_norm": 1.1537258625030518, + "learning_rate": 1.3216748491759446e-05, + "loss": 3.6043, + "step": 151000 + }, + { + "epoch": 1.69697173960668, + "grad_norm": 1.1523864269256592, + "learning_rate": 1.3214501578456596e-05, + "loss": 3.6814, + "step": 151050 + }, + { + "epoch": 1.697533464777023, + "grad_norm": 1.1361587047576904, + "learning_rate": 1.3212254665153745e-05, + "loss": 3.6928, + "step": 151100 + }, + { + "epoch": 1.6980951899473662, + "grad_norm": 1.2203460931777954, + "learning_rate": 1.3210007751850896e-05, + "loss": 3.7051, + "step": 151150 + }, + { + "epoch": 1.6986569151177096, + "grad_norm": 1.1210293769836426, + "learning_rate": 1.3207760838548045e-05, + "loss": 3.7301, + "step": 151200 + }, + { + "epoch": 1.6992186402880527, + "grad_norm": 1.1691347360610962, + "learning_rate": 1.3205513925245194e-05, + "loss": 3.6296, + "step": 151250 + }, + { + "epoch": 1.6997803654583958, + "grad_norm": 1.1081463098526, + "learning_rate": 1.3203267011942344e-05, + "loss": 3.7019, + "step": 151300 + }, + { + "epoch": 1.700342090628739, + "grad_norm": 1.0898524522781372, + "learning_rate": 1.3201020098639495e-05, + "loss": 3.7002, + "step": 151350 + }, + { + "epoch": 1.7009038157990821, + "grad_norm": 1.0776554346084595, + "learning_rate": 1.3198773185336644e-05, + "loss": 3.6315, + "step": 151400 + }, + { + "epoch": 1.7014655409694253, + "grad_norm": 1.0310733318328857, + "learning_rate": 1.3196526272033793e-05, + "loss": 3.6213, + "step": 151450 + }, + { + "epoch": 1.7020272661397684, + "grad_norm": 1.294753074645996, + "learning_rate": 1.3194324296997001e-05, + "loss": 3.7238, + "step": 151500 + }, + { + "epoch": 1.7025889913101118, + "grad_norm": 1.391838788986206, + "learning_rate": 1.319207738369415e-05, + "loss": 3.7018, + "step": 151550 + }, + { + "epoch": 1.703150716480455, + "grad_norm": 1.1529507637023926, + "learning_rate": 1.31898304703913e-05, + "loss": 3.6026, + "step": 151600 + }, + { + "epoch": 1.703712441650798, + "grad_norm": 1.0914785861968994, + "learning_rate": 1.318758355708845e-05, + "loss": 3.7089, + "step": 151650 + }, + { + "epoch": 1.7042741668211412, + "grad_norm": 1.33309805393219, + "learning_rate": 1.31853366437856e-05, + "loss": 3.6505, + "step": 151700 + }, + { + "epoch": 1.7048358919914843, + "grad_norm": 1.1748844385147095, + "learning_rate": 1.318308973048275e-05, + "loss": 3.6378, + "step": 151750 + }, + { + "epoch": 1.7053976171618275, + "grad_norm": 1.3727186918258667, + "learning_rate": 1.31808428171799e-05, + "loss": 3.7217, + "step": 151800 + }, + { + "epoch": 1.7059593423321706, + "grad_norm": 1.2234445810317993, + "learning_rate": 1.317859590387705e-05, + "loss": 3.7484, + "step": 151850 + }, + { + "epoch": 1.7065210675025138, + "grad_norm": 1.3959254026412964, + "learning_rate": 1.3176348990574199e-05, + "loss": 3.7239, + "step": 151900 + }, + { + "epoch": 1.707082792672857, + "grad_norm": 1.5000845193862915, + "learning_rate": 1.3174102077271348e-05, + "loss": 3.6651, + "step": 151950 + }, + { + "epoch": 1.7076445178432, + "grad_norm": 1.2455960512161255, + "learning_rate": 1.31718551639685e-05, + "loss": 3.6371, + "step": 152000 + }, + { + "epoch": 1.7082062430135432, + "grad_norm": 1.1509108543395996, + "learning_rate": 1.3169608250665649e-05, + "loss": 3.695, + "step": 152050 + }, + { + "epoch": 1.7087679681838863, + "grad_norm": 1.1868137121200562, + "learning_rate": 1.3167361337362798e-05, + "loss": 3.6248, + "step": 152100 + }, + { + "epoch": 1.7093296933542295, + "grad_norm": 1.268827199935913, + "learning_rate": 1.3165114424059949e-05, + "loss": 3.6964, + "step": 152150 + }, + { + "epoch": 1.7098914185245726, + "grad_norm": 1.41181480884552, + "learning_rate": 1.3162867510757098e-05, + "loss": 3.6524, + "step": 152200 + }, + { + "epoch": 1.7104531436949157, + "grad_norm": 1.2418874502182007, + "learning_rate": 1.3160620597454248e-05, + "loss": 3.6359, + "step": 152250 + }, + { + "epoch": 1.7110148688652589, + "grad_norm": 1.390811562538147, + "learning_rate": 1.3158373684151397e-05, + "loss": 3.6681, + "step": 152300 + }, + { + "epoch": 1.711576594035602, + "grad_norm": 1.1755359172821045, + "learning_rate": 1.3156126770848548e-05, + "loss": 3.7523, + "step": 152350 + }, + { + "epoch": 1.7121383192059452, + "grad_norm": 1.3521742820739746, + "learning_rate": 1.3153879857545697e-05, + "loss": 3.6319, + "step": 152400 + }, + { + "epoch": 1.7127000443762883, + "grad_norm": 1.1433719396591187, + "learning_rate": 1.3151632944242847e-05, + "loss": 3.6684, + "step": 152450 + }, + { + "epoch": 1.7132617695466315, + "grad_norm": 1.052834391593933, + "learning_rate": 1.3149386030939998e-05, + "loss": 3.63, + "step": 152500 + }, + { + "epoch": 1.7138234947169748, + "grad_norm": 1.0887142419815063, + "learning_rate": 1.3147184055903203e-05, + "loss": 3.6899, + "step": 152550 + }, + { + "epoch": 1.714385219887318, + "grad_norm": 1.1405411958694458, + "learning_rate": 1.3144937142600353e-05, + "loss": 3.7286, + "step": 152600 + }, + { + "epoch": 1.714946945057661, + "grad_norm": 0.9536702632904053, + "learning_rate": 1.3142690229297504e-05, + "loss": 3.7611, + "step": 152650 + }, + { + "epoch": 1.7155086702280042, + "grad_norm": 1.643053650856018, + "learning_rate": 1.3140443315994653e-05, + "loss": 3.6665, + "step": 152700 + }, + { + "epoch": 1.7160703953983474, + "grad_norm": 1.278144121170044, + "learning_rate": 1.3138196402691802e-05, + "loss": 3.6664, + "step": 152750 + }, + { + "epoch": 1.7166321205686905, + "grad_norm": 1.1933716535568237, + "learning_rate": 1.3135949489388953e-05, + "loss": 3.6371, + "step": 152800 + }, + { + "epoch": 1.7171938457390339, + "grad_norm": 1.2203391790390015, + "learning_rate": 1.3133702576086103e-05, + "loss": 3.656, + "step": 152850 + }, + { + "epoch": 1.717755570909377, + "grad_norm": 1.1443779468536377, + "learning_rate": 1.3131455662783252e-05, + "loss": 3.7061, + "step": 152900 + }, + { + "epoch": 1.7183172960797202, + "grad_norm": 1.0043830871582031, + "learning_rate": 1.3129208749480401e-05, + "loss": 3.6858, + "step": 152950 + }, + { + "epoch": 1.7188790212500633, + "grad_norm": 1.0541163682937622, + "learning_rate": 1.3126961836177552e-05, + "loss": 3.697, + "step": 153000 + }, + { + "epoch": 1.7194407464204065, + "grad_norm": 1.1946184635162354, + "learning_rate": 1.3124714922874702e-05, + "loss": 3.7139, + "step": 153050 + }, + { + "epoch": 1.7200024715907496, + "grad_norm": 1.1679470539093018, + "learning_rate": 1.3122468009571851e-05, + "loss": 3.6856, + "step": 153100 + }, + { + "epoch": 1.7205641967610927, + "grad_norm": 1.140401005744934, + "learning_rate": 1.3120221096269002e-05, + "loss": 3.6471, + "step": 153150 + }, + { + "epoch": 1.7211259219314359, + "grad_norm": 1.2564250230789185, + "learning_rate": 1.3117974182966151e-05, + "loss": 3.6139, + "step": 153200 + }, + { + "epoch": 1.721687647101779, + "grad_norm": 1.1056748628616333, + "learning_rate": 1.31157272696633e-05, + "loss": 3.7062, + "step": 153250 + }, + { + "epoch": 1.7222493722721222, + "grad_norm": 1.2699270248413086, + "learning_rate": 1.311348035636045e-05, + "loss": 3.6892, + "step": 153300 + }, + { + "epoch": 1.7228110974424653, + "grad_norm": 1.2607805728912354, + "learning_rate": 1.3111233443057601e-05, + "loss": 3.6134, + "step": 153350 + }, + { + "epoch": 1.7233728226128084, + "grad_norm": 1.2254953384399414, + "learning_rate": 1.310898652975475e-05, + "loss": 3.721, + "step": 153400 + }, + { + "epoch": 1.7239345477831516, + "grad_norm": 1.3641060590744019, + "learning_rate": 1.31067396164519e-05, + "loss": 3.6985, + "step": 153450 + }, + { + "epoch": 1.7244962729534947, + "grad_norm": 1.052706003189087, + "learning_rate": 1.310449270314905e-05, + "loss": 3.7281, + "step": 153500 + }, + { + "epoch": 1.7250579981238379, + "grad_norm": 1.120184063911438, + "learning_rate": 1.31022457898462e-05, + "loss": 3.6994, + "step": 153550 + }, + { + "epoch": 1.725619723294181, + "grad_norm": 1.2567594051361084, + "learning_rate": 1.309999887654335e-05, + "loss": 3.6809, + "step": 153600 + }, + { + "epoch": 1.7261814484645241, + "grad_norm": 1.1006348133087158, + "learning_rate": 1.3097751963240499e-05, + "loss": 3.689, + "step": 153650 + }, + { + "epoch": 1.7267431736348673, + "grad_norm": 1.1645714044570923, + "learning_rate": 1.309550504993765e-05, + "loss": 3.6636, + "step": 153700 + }, + { + "epoch": 1.7273048988052104, + "grad_norm": 1.1966338157653809, + "learning_rate": 1.3093258136634799e-05, + "loss": 3.6868, + "step": 153750 + }, + { + "epoch": 1.7278666239755536, + "grad_norm": 1.1202161312103271, + "learning_rate": 1.3091011223331948e-05, + "loss": 3.6664, + "step": 153800 + }, + { + "epoch": 1.7284283491458967, + "grad_norm": 1.236364483833313, + "learning_rate": 1.30887643100291e-05, + "loss": 3.5549, + "step": 153850 + }, + { + "epoch": 1.72899007431624, + "grad_norm": 1.3125675916671753, + "learning_rate": 1.3086517396726249e-05, + "loss": 3.6765, + "step": 153900 + }, + { + "epoch": 1.7295517994865832, + "grad_norm": 1.2397410869598389, + "learning_rate": 1.3084270483423398e-05, + "loss": 3.6706, + "step": 153950 + }, + { + "epoch": 1.7301135246569264, + "grad_norm": 1.2448431253433228, + "learning_rate": 1.3082023570120547e-05, + "loss": 3.7243, + "step": 154000 + }, + { + "epoch": 1.7306752498272695, + "grad_norm": 1.0740070343017578, + "learning_rate": 1.3079776656817698e-05, + "loss": 3.6927, + "step": 154050 + }, + { + "epoch": 1.7312369749976126, + "grad_norm": 1.0397754907608032, + "learning_rate": 1.3077529743514848e-05, + "loss": 3.6253, + "step": 154100 + }, + { + "epoch": 1.7317987001679558, + "grad_norm": 0.9703950881958008, + "learning_rate": 1.3075282830211997e-05, + "loss": 3.6631, + "step": 154150 + }, + { + "epoch": 1.7323604253382991, + "grad_norm": 1.2280876636505127, + "learning_rate": 1.3073035916909146e-05, + "loss": 3.6468, + "step": 154200 + }, + { + "epoch": 1.7329221505086423, + "grad_norm": 1.0809295177459717, + "learning_rate": 1.3070789003606297e-05, + "loss": 3.6998, + "step": 154250 + }, + { + "epoch": 1.7334838756789854, + "grad_norm": 1.2361363172531128, + "learning_rate": 1.3068542090303446e-05, + "loss": 3.6474, + "step": 154300 + }, + { + "epoch": 1.7340456008493286, + "grad_norm": 1.0839402675628662, + "learning_rate": 1.3066295177000596e-05, + "loss": 3.6696, + "step": 154350 + }, + { + "epoch": 1.7346073260196717, + "grad_norm": 1.2606266736984253, + "learning_rate": 1.3064048263697747e-05, + "loss": 3.6083, + "step": 154400 + }, + { + "epoch": 1.7351690511900149, + "grad_norm": 1.069074034690857, + "learning_rate": 1.3061801350394896e-05, + "loss": 3.625, + "step": 154450 + }, + { + "epoch": 1.735730776360358, + "grad_norm": 1.063657283782959, + "learning_rate": 1.3059554437092045e-05, + "loss": 3.6952, + "step": 154500 + }, + { + "epoch": 1.7362925015307011, + "grad_norm": 1.1664756536483765, + "learning_rate": 1.3057307523789195e-05, + "loss": 3.6318, + "step": 154550 + }, + { + "epoch": 1.7368542267010443, + "grad_norm": 1.219751238822937, + "learning_rate": 1.3055060610486346e-05, + "loss": 3.727, + "step": 154600 + }, + { + "epoch": 1.7374159518713874, + "grad_norm": 1.2343968152999878, + "learning_rate": 1.3052813697183495e-05, + "loss": 3.6172, + "step": 154650 + }, + { + "epoch": 1.7379776770417306, + "grad_norm": 1.1121562719345093, + "learning_rate": 1.3050566783880644e-05, + "loss": 3.6709, + "step": 154700 + }, + { + "epoch": 1.7385394022120737, + "grad_norm": 1.5850425958633423, + "learning_rate": 1.3048319870577795e-05, + "loss": 3.7197, + "step": 154750 + }, + { + "epoch": 1.7391011273824168, + "grad_norm": 1.3235431909561157, + "learning_rate": 1.3046072957274945e-05, + "loss": 3.6213, + "step": 154800 + }, + { + "epoch": 1.73966285255276, + "grad_norm": 1.1664890050888062, + "learning_rate": 1.3043826043972094e-05, + "loss": 3.6559, + "step": 154850 + }, + { + "epoch": 1.7402245777231031, + "grad_norm": 1.647775411605835, + "learning_rate": 1.3041579130669243e-05, + "loss": 3.64, + "step": 154900 + }, + { + "epoch": 1.7407863028934463, + "grad_norm": 0.9884029626846313, + "learning_rate": 1.3039332217366394e-05, + "loss": 3.6411, + "step": 154950 + }, + { + "epoch": 1.7413480280637894, + "grad_norm": 1.123098611831665, + "learning_rate": 1.3037085304063544e-05, + "loss": 3.7255, + "step": 155000 + }, + { + "epoch": 1.7419097532341326, + "grad_norm": 1.176184892654419, + "learning_rate": 1.3034883329026751e-05, + "loss": 3.7112, + "step": 155050 + }, + { + "epoch": 1.7424714784044757, + "grad_norm": 1.0983113050460815, + "learning_rate": 1.30326364157239e-05, + "loss": 3.6281, + "step": 155100 + }, + { + "epoch": 1.7430332035748188, + "grad_norm": 1.1853564977645874, + "learning_rate": 1.303038950242105e-05, + "loss": 3.6697, + "step": 155150 + }, + { + "epoch": 1.743594928745162, + "grad_norm": 1.1369297504425049, + "learning_rate": 1.30281425891182e-05, + "loss": 3.6517, + "step": 155200 + }, + { + "epoch": 1.7441566539155053, + "grad_norm": 1.1019175052642822, + "learning_rate": 1.302589567581535e-05, + "loss": 3.6663, + "step": 155250 + }, + { + "epoch": 1.7447183790858485, + "grad_norm": 1.0535612106323242, + "learning_rate": 1.30236487625125e-05, + "loss": 3.6439, + "step": 155300 + }, + { + "epoch": 1.7452801042561916, + "grad_norm": 1.1677604913711548, + "learning_rate": 1.3021401849209649e-05, + "loss": 3.707, + "step": 155350 + }, + { + "epoch": 1.7458418294265348, + "grad_norm": 1.1271241903305054, + "learning_rate": 1.30191549359068e-05, + "loss": 3.7345, + "step": 155400 + }, + { + "epoch": 1.746403554596878, + "grad_norm": 1.1681445837020874, + "learning_rate": 1.301690802260395e-05, + "loss": 3.6706, + "step": 155450 + }, + { + "epoch": 1.746965279767221, + "grad_norm": 1.0617426633834839, + "learning_rate": 1.3014661109301099e-05, + "loss": 3.6912, + "step": 155500 + }, + { + "epoch": 1.7475270049375644, + "grad_norm": 1.1394972801208496, + "learning_rate": 1.3012414195998248e-05, + "loss": 3.5982, + "step": 155550 + }, + { + "epoch": 1.7480887301079076, + "grad_norm": 1.078129529953003, + "learning_rate": 1.3010167282695399e-05, + "loss": 3.6903, + "step": 155600 + }, + { + "epoch": 1.7486504552782507, + "grad_norm": 1.2922347784042358, + "learning_rate": 1.3007920369392548e-05, + "loss": 3.6904, + "step": 155650 + }, + { + "epoch": 1.7492121804485938, + "grad_norm": 1.561508297920227, + "learning_rate": 1.3005673456089698e-05, + "loss": 3.6604, + "step": 155700 + }, + { + "epoch": 1.749773905618937, + "grad_norm": 1.1499099731445312, + "learning_rate": 1.3003426542786849e-05, + "loss": 3.6863, + "step": 155750 + }, + { + "epoch": 1.7503356307892801, + "grad_norm": 1.0464122295379639, + "learning_rate": 1.3001179629483998e-05, + "loss": 3.6086, + "step": 155800 + }, + { + "epoch": 1.7508973559596233, + "grad_norm": 1.1842803955078125, + "learning_rate": 1.2998932716181147e-05, + "loss": 3.653, + "step": 155850 + }, + { + "epoch": 1.7514590811299664, + "grad_norm": 1.3405252695083618, + "learning_rate": 1.2996685802878296e-05, + "loss": 3.699, + "step": 155900 + }, + { + "epoch": 1.7520208063003095, + "grad_norm": 1.129227876663208, + "learning_rate": 1.2994438889575447e-05, + "loss": 3.6578, + "step": 155950 + }, + { + "epoch": 1.7525825314706527, + "grad_norm": 1.137797236442566, + "learning_rate": 1.2992191976272597e-05, + "loss": 3.7331, + "step": 156000 + }, + { + "epoch": 1.7531442566409958, + "grad_norm": 1.041616439819336, + "learning_rate": 1.2989945062969746e-05, + "loss": 3.6837, + "step": 156050 + }, + { + "epoch": 1.753705981811339, + "grad_norm": 1.1781810522079468, + "learning_rate": 1.2987698149666897e-05, + "loss": 3.635, + "step": 156100 + }, + { + "epoch": 1.754267706981682, + "grad_norm": 1.1395310163497925, + "learning_rate": 1.2985451236364046e-05, + "loss": 3.7524, + "step": 156150 + }, + { + "epoch": 1.7548294321520252, + "grad_norm": 1.4573463201522827, + "learning_rate": 1.2983204323061196e-05, + "loss": 3.7249, + "step": 156200 + }, + { + "epoch": 1.7553911573223684, + "grad_norm": 1.1571025848388672, + "learning_rate": 1.2980957409758345e-05, + "loss": 3.7074, + "step": 156250 + }, + { + "epoch": 1.7559528824927115, + "grad_norm": 1.0512663125991821, + "learning_rate": 1.2978710496455496e-05, + "loss": 3.6184, + "step": 156300 + }, + { + "epoch": 1.7565146076630547, + "grad_norm": 1.0732377767562866, + "learning_rate": 1.2976463583152645e-05, + "loss": 3.6676, + "step": 156350 + }, + { + "epoch": 1.7570763328333978, + "grad_norm": 1.127360463142395, + "learning_rate": 1.2974216669849795e-05, + "loss": 3.6764, + "step": 156400 + }, + { + "epoch": 1.757638058003741, + "grad_norm": 1.111916184425354, + "learning_rate": 1.2971969756546946e-05, + "loss": 3.6875, + "step": 156450 + }, + { + "epoch": 1.758199783174084, + "grad_norm": 1.1683317422866821, + "learning_rate": 1.2969722843244095e-05, + "loss": 3.7127, + "step": 156500 + }, + { + "epoch": 1.7587615083444275, + "grad_norm": 1.1425973176956177, + "learning_rate": 1.2967520868207301e-05, + "loss": 3.638, + "step": 156550 + }, + { + "epoch": 1.7593232335147706, + "grad_norm": 1.2834800481796265, + "learning_rate": 1.2965273954904452e-05, + "loss": 3.6476, + "step": 156600 + }, + { + "epoch": 1.7598849586851137, + "grad_norm": 1.1782361268997192, + "learning_rate": 1.2963027041601601e-05, + "loss": 3.6163, + "step": 156650 + }, + { + "epoch": 1.7604466838554569, + "grad_norm": 1.069303035736084, + "learning_rate": 1.296078012829875e-05, + "loss": 3.7065, + "step": 156700 + }, + { + "epoch": 1.7610084090258, + "grad_norm": 1.0960201025009155, + "learning_rate": 1.2958533214995902e-05, + "loss": 3.678, + "step": 156750 + }, + { + "epoch": 1.7615701341961432, + "grad_norm": 1.3260595798492432, + "learning_rate": 1.2956286301693051e-05, + "loss": 3.6791, + "step": 156800 + }, + { + "epoch": 1.7621318593664863, + "grad_norm": 1.1478679180145264, + "learning_rate": 1.29540393883902e-05, + "loss": 3.6592, + "step": 156850 + }, + { + "epoch": 1.7626935845368297, + "grad_norm": 1.0272244215011597, + "learning_rate": 1.295179247508735e-05, + "loss": 3.6611, + "step": 156900 + }, + { + "epoch": 1.7632553097071728, + "grad_norm": 1.1543618440628052, + "learning_rate": 1.29495455617845e-05, + "loss": 3.6443, + "step": 156950 + }, + { + "epoch": 1.763817034877516, + "grad_norm": 1.275540828704834, + "learning_rate": 1.294729864848165e-05, + "loss": 3.7771, + "step": 157000 + }, + { + "epoch": 1.764378760047859, + "grad_norm": 1.168523907661438, + "learning_rate": 1.29450517351788e-05, + "loss": 3.6167, + "step": 157050 + }, + { + "epoch": 1.7649404852182022, + "grad_norm": 0.9482042193412781, + "learning_rate": 1.294280482187595e-05, + "loss": 3.6669, + "step": 157100 + }, + { + "epoch": 1.7655022103885454, + "grad_norm": 1.0952187776565552, + "learning_rate": 1.29405579085731e-05, + "loss": 3.682, + "step": 157150 + }, + { + "epoch": 1.7660639355588885, + "grad_norm": 1.2059918642044067, + "learning_rate": 1.2938310995270249e-05, + "loss": 3.5579, + "step": 157200 + }, + { + "epoch": 1.7666256607292317, + "grad_norm": 1.1145679950714111, + "learning_rate": 1.2936064081967398e-05, + "loss": 3.6652, + "step": 157250 + }, + { + "epoch": 1.7671873858995748, + "grad_norm": 1.1019964218139648, + "learning_rate": 1.293381716866455e-05, + "loss": 3.6714, + "step": 157300 + }, + { + "epoch": 1.767749111069918, + "grad_norm": 1.5606224536895752, + "learning_rate": 1.2931570255361699e-05, + "loss": 3.6421, + "step": 157350 + }, + { + "epoch": 1.768310836240261, + "grad_norm": 1.1179559230804443, + "learning_rate": 1.2929323342058848e-05, + "loss": 3.6624, + "step": 157400 + }, + { + "epoch": 1.7688725614106042, + "grad_norm": 1.2087434530258179, + "learning_rate": 1.2927076428755999e-05, + "loss": 3.6901, + "step": 157450 + }, + { + "epoch": 1.7694342865809474, + "grad_norm": 1.0996601581573486, + "learning_rate": 1.2924829515453148e-05, + "loss": 3.6827, + "step": 157500 + }, + { + "epoch": 1.7699960117512905, + "grad_norm": 1.1604089736938477, + "learning_rate": 1.2922582602150297e-05, + "loss": 3.6265, + "step": 157550 + }, + { + "epoch": 1.7705577369216337, + "grad_norm": 1.0635164976119995, + "learning_rate": 1.2920335688847447e-05, + "loss": 3.7781, + "step": 157600 + }, + { + "epoch": 1.7711194620919768, + "grad_norm": 1.26372230052948, + "learning_rate": 1.2918088775544598e-05, + "loss": 3.6933, + "step": 157650 + }, + { + "epoch": 1.77168118726232, + "grad_norm": 1.1672794818878174, + "learning_rate": 1.2915841862241747e-05, + "loss": 3.6697, + "step": 157700 + }, + { + "epoch": 1.772242912432663, + "grad_norm": 1.0433988571166992, + "learning_rate": 1.2913594948938896e-05, + "loss": 3.6325, + "step": 157750 + }, + { + "epoch": 1.7728046376030062, + "grad_norm": 1.1019052267074585, + "learning_rate": 1.2911348035636046e-05, + "loss": 3.6513, + "step": 157800 + }, + { + "epoch": 1.7733663627733494, + "grad_norm": 1.1907793283462524, + "learning_rate": 1.2909101122333197e-05, + "loss": 3.7005, + "step": 157850 + }, + { + "epoch": 1.7739280879436927, + "grad_norm": 1.1211351156234741, + "learning_rate": 1.2906854209030346e-05, + "loss": 3.7068, + "step": 157900 + }, + { + "epoch": 1.7744898131140359, + "grad_norm": 1.1437180042266846, + "learning_rate": 1.2904607295727495e-05, + "loss": 3.693, + "step": 157950 + }, + { + "epoch": 1.775051538284379, + "grad_norm": 1.1743581295013428, + "learning_rate": 1.2902360382424646e-05, + "loss": 3.6189, + "step": 158000 + }, + { + "epoch": 1.7756132634547221, + "grad_norm": 1.1614916324615479, + "learning_rate": 1.2900113469121796e-05, + "loss": 3.7102, + "step": 158050 + }, + { + "epoch": 1.7761749886250653, + "grad_norm": 1.1317543983459473, + "learning_rate": 1.2897866555818945e-05, + "loss": 3.6501, + "step": 158100 + }, + { + "epoch": 1.7767367137954084, + "grad_norm": 1.259522557258606, + "learning_rate": 1.2895664580782153e-05, + "loss": 3.7159, + "step": 158150 + }, + { + "epoch": 1.7772984389657516, + "grad_norm": 1.4388420581817627, + "learning_rate": 1.2893417667479302e-05, + "loss": 3.6617, + "step": 158200 + }, + { + "epoch": 1.777860164136095, + "grad_norm": 1.279236912727356, + "learning_rate": 1.2891170754176451e-05, + "loss": 3.751, + "step": 158250 + }, + { + "epoch": 1.778421889306438, + "grad_norm": 1.0801838636398315, + "learning_rate": 1.2888923840873602e-05, + "loss": 3.6969, + "step": 158300 + }, + { + "epoch": 1.7789836144767812, + "grad_norm": 1.2110668420791626, + "learning_rate": 1.2886676927570752e-05, + "loss": 3.6743, + "step": 158350 + }, + { + "epoch": 1.7795453396471244, + "grad_norm": 1.1318589448928833, + "learning_rate": 1.2884430014267901e-05, + "loss": 3.6516, + "step": 158400 + }, + { + "epoch": 1.7801070648174675, + "grad_norm": 1.0363426208496094, + "learning_rate": 1.2882183100965052e-05, + "loss": 3.6595, + "step": 158450 + }, + { + "epoch": 1.7806687899878106, + "grad_norm": 1.2623485326766968, + "learning_rate": 1.2879936187662201e-05, + "loss": 3.6201, + "step": 158500 + }, + { + "epoch": 1.7812305151581538, + "grad_norm": 1.0081738233566284, + "learning_rate": 1.287768927435935e-05, + "loss": 3.7067, + "step": 158550 + }, + { + "epoch": 1.781792240328497, + "grad_norm": 1.1923474073410034, + "learning_rate": 1.28754423610565e-05, + "loss": 3.6912, + "step": 158600 + }, + { + "epoch": 1.78235396549884, + "grad_norm": 1.139208436012268, + "learning_rate": 1.2873195447753651e-05, + "loss": 3.6359, + "step": 158650 + }, + { + "epoch": 1.7829156906691832, + "grad_norm": 1.2712637186050415, + "learning_rate": 1.28709485344508e-05, + "loss": 3.7443, + "step": 158700 + }, + { + "epoch": 1.7834774158395263, + "grad_norm": 1.1234378814697266, + "learning_rate": 1.286870162114795e-05, + "loss": 3.6251, + "step": 158750 + }, + { + "epoch": 1.7840391410098695, + "grad_norm": 1.275105595588684, + "learning_rate": 1.2866454707845099e-05, + "loss": 3.6907, + "step": 158800 + }, + { + "epoch": 1.7846008661802126, + "grad_norm": 1.3843269348144531, + "learning_rate": 1.286420779454225e-05, + "loss": 3.6619, + "step": 158850 + }, + { + "epoch": 1.7851625913505558, + "grad_norm": 1.107105016708374, + "learning_rate": 1.28619608812394e-05, + "loss": 3.7125, + "step": 158900 + }, + { + "epoch": 1.785724316520899, + "grad_norm": 1.0981186628341675, + "learning_rate": 1.2859713967936548e-05, + "loss": 3.6826, + "step": 158950 + }, + { + "epoch": 1.786286041691242, + "grad_norm": 1.2366962432861328, + "learning_rate": 1.28574670546337e-05, + "loss": 3.6565, + "step": 159000 + }, + { + "epoch": 1.7868477668615852, + "grad_norm": 1.1007949113845825, + "learning_rate": 1.2855220141330849e-05, + "loss": 3.7028, + "step": 159050 + }, + { + "epoch": 1.7874094920319283, + "grad_norm": 1.1660207509994507, + "learning_rate": 1.2852973228027998e-05, + "loss": 3.6584, + "step": 159100 + }, + { + "epoch": 1.7879712172022715, + "grad_norm": 1.135389804840088, + "learning_rate": 1.2850726314725147e-05, + "loss": 3.6954, + "step": 159150 + }, + { + "epoch": 1.7885329423726146, + "grad_norm": 1.2505366802215576, + "learning_rate": 1.2848479401422298e-05, + "loss": 3.7322, + "step": 159200 + }, + { + "epoch": 1.789094667542958, + "grad_norm": 1.289759874343872, + "learning_rate": 1.2846232488119448e-05, + "loss": 3.7352, + "step": 159250 + }, + { + "epoch": 1.7896563927133011, + "grad_norm": 1.5705466270446777, + "learning_rate": 1.2843985574816597e-05, + "loss": 3.6966, + "step": 159300 + }, + { + "epoch": 1.7902181178836443, + "grad_norm": 1.1748906373977661, + "learning_rate": 1.2841738661513748e-05, + "loss": 3.6485, + "step": 159350 + }, + { + "epoch": 1.7907798430539874, + "grad_norm": 1.2288804054260254, + "learning_rate": 1.2839491748210897e-05, + "loss": 3.7492, + "step": 159400 + }, + { + "epoch": 1.7913415682243305, + "grad_norm": 1.230589747428894, + "learning_rate": 1.2837244834908047e-05, + "loss": 3.7004, + "step": 159450 + }, + { + "epoch": 1.7919032933946737, + "grad_norm": 1.1219701766967773, + "learning_rate": 1.2834997921605196e-05, + "loss": 3.6086, + "step": 159500 + }, + { + "epoch": 1.7924650185650168, + "grad_norm": 1.1636079549789429, + "learning_rate": 1.2832751008302347e-05, + "loss": 3.6157, + "step": 159550 + }, + { + "epoch": 1.7930267437353602, + "grad_norm": 1.3621783256530762, + "learning_rate": 1.2830504094999496e-05, + "loss": 3.6461, + "step": 159600 + }, + { + "epoch": 1.7935884689057033, + "grad_norm": 1.0143169164657593, + "learning_rate": 1.2828257181696646e-05, + "loss": 3.7079, + "step": 159650 + }, + { + "epoch": 1.7941501940760465, + "grad_norm": 1.0718719959259033, + "learning_rate": 1.2826010268393797e-05, + "loss": 3.6971, + "step": 159700 + }, + { + "epoch": 1.7947119192463896, + "grad_norm": 0.9802167415618896, + "learning_rate": 1.2823763355090946e-05, + "loss": 3.6349, + "step": 159750 + }, + { + "epoch": 1.7952736444167328, + "grad_norm": 1.4045599699020386, + "learning_rate": 1.2821516441788095e-05, + "loss": 3.7373, + "step": 159800 + }, + { + "epoch": 1.795835369587076, + "grad_norm": 1.0568664073944092, + "learning_rate": 1.2819269528485245e-05, + "loss": 3.6243, + "step": 159850 + }, + { + "epoch": 1.796397094757419, + "grad_norm": 1.1222189664840698, + "learning_rate": 1.2817022615182396e-05, + "loss": 3.6887, + "step": 159900 + }, + { + "epoch": 1.7969588199277622, + "grad_norm": 1.0660656690597534, + "learning_rate": 1.2814775701879545e-05, + "loss": 3.6987, + "step": 159950 + }, + { + "epoch": 1.7975205450981053, + "grad_norm": 1.1978665590286255, + "learning_rate": 1.2812528788576694e-05, + "loss": 3.7301, + "step": 160000 + }, + { + "epoch": 1.7980822702684485, + "grad_norm": 1.1811360120773315, + "learning_rate": 1.2810281875273845e-05, + "loss": 3.6331, + "step": 160050 + }, + { + "epoch": 1.7986439954387916, + "grad_norm": 1.0400501489639282, + "learning_rate": 1.2808034961970995e-05, + "loss": 3.6449, + "step": 160100 + }, + { + "epoch": 1.7992057206091348, + "grad_norm": 1.3442927598953247, + "learning_rate": 1.2805788048668144e-05, + "loss": 3.6649, + "step": 160150 + }, + { + "epoch": 1.799767445779478, + "grad_norm": 1.1856480836868286, + "learning_rate": 1.2803541135365293e-05, + "loss": 3.7003, + "step": 160200 + }, + { + "epoch": 1.800329170949821, + "grad_norm": 1.2278929948806763, + "learning_rate": 1.2801294222062444e-05, + "loss": 3.6596, + "step": 160250 + }, + { + "epoch": 1.8008908961201642, + "grad_norm": 1.3022801876068115, + "learning_rate": 1.2799047308759594e-05, + "loss": 3.6942, + "step": 160300 + }, + { + "epoch": 1.8014526212905073, + "grad_norm": 1.1130034923553467, + "learning_rate": 1.2796800395456743e-05, + "loss": 3.6764, + "step": 160350 + }, + { + "epoch": 1.8020143464608505, + "grad_norm": 1.1749629974365234, + "learning_rate": 1.2794553482153892e-05, + "loss": 3.6354, + "step": 160400 + }, + { + "epoch": 1.8025760716311936, + "grad_norm": 1.0336318016052246, + "learning_rate": 1.2792306568851043e-05, + "loss": 3.6024, + "step": 160450 + }, + { + "epoch": 1.8031377968015367, + "grad_norm": 1.0810723304748535, + "learning_rate": 1.2790059655548193e-05, + "loss": 3.602, + "step": 160500 + }, + { + "epoch": 1.8036995219718799, + "grad_norm": 1.1231781244277954, + "learning_rate": 1.2787812742245342e-05, + "loss": 3.6974, + "step": 160550 + }, + { + "epoch": 1.8042612471422232, + "grad_norm": 1.1039625406265259, + "learning_rate": 1.2785565828942493e-05, + "loss": 3.6196, + "step": 160600 + }, + { + "epoch": 1.8048229723125664, + "grad_norm": 1.094585657119751, + "learning_rate": 1.2783318915639642e-05, + "loss": 3.6405, + "step": 160650 + }, + { + "epoch": 1.8053846974829095, + "grad_norm": 1.246232271194458, + "learning_rate": 1.2781072002336792e-05, + "loss": 3.6121, + "step": 160700 + }, + { + "epoch": 1.8059464226532527, + "grad_norm": 1.2481576204299927, + "learning_rate": 1.277882508903394e-05, + "loss": 3.5926, + "step": 160750 + }, + { + "epoch": 1.8065081478235958, + "grad_norm": 1.3390722274780273, + "learning_rate": 1.2776578175731092e-05, + "loss": 3.7399, + "step": 160800 + }, + { + "epoch": 1.807069872993939, + "grad_norm": 1.299284815788269, + "learning_rate": 1.2774331262428241e-05, + "loss": 3.6092, + "step": 160850 + }, + { + "epoch": 1.807631598164282, + "grad_norm": 1.2629300355911255, + "learning_rate": 1.277208434912539e-05, + "loss": 3.6756, + "step": 160900 + }, + { + "epoch": 1.8081933233346255, + "grad_norm": 1.2889231443405151, + "learning_rate": 1.2769837435822541e-05, + "loss": 3.6702, + "step": 160950 + }, + { + "epoch": 1.8087550485049686, + "grad_norm": 1.0715359449386597, + "learning_rate": 1.276759052251969e-05, + "loss": 3.6998, + "step": 161000 + }, + { + "epoch": 1.8093167736753117, + "grad_norm": 1.08761465549469, + "learning_rate": 1.276534360921684e-05, + "loss": 3.667, + "step": 161050 + }, + { + "epoch": 1.8098784988456549, + "grad_norm": 1.1700549125671387, + "learning_rate": 1.276309669591399e-05, + "loss": 3.659, + "step": 161100 + }, + { + "epoch": 1.810440224015998, + "grad_norm": 1.126656174659729, + "learning_rate": 1.276084978261114e-05, + "loss": 3.7039, + "step": 161150 + }, + { + "epoch": 1.8110019491863412, + "grad_norm": 1.1791270971298218, + "learning_rate": 1.275860286930829e-05, + "loss": 3.6929, + "step": 161200 + }, + { + "epoch": 1.8115636743566843, + "grad_norm": 1.0086944103240967, + "learning_rate": 1.2756355956005439e-05, + "loss": 3.7241, + "step": 161250 + }, + { + "epoch": 1.8121253995270274, + "grad_norm": 1.1907039880752563, + "learning_rate": 1.275410904270259e-05, + "loss": 3.6694, + "step": 161300 + }, + { + "epoch": 1.8126871246973706, + "grad_norm": 1.202735185623169, + "learning_rate": 1.275186212939974e-05, + "loss": 3.6575, + "step": 161350 + }, + { + "epoch": 1.8132488498677137, + "grad_norm": 1.0789495706558228, + "learning_rate": 1.2749615216096889e-05, + "loss": 3.621, + "step": 161400 + }, + { + "epoch": 1.8138105750380569, + "grad_norm": 1.6504883766174316, + "learning_rate": 1.2747368302794038e-05, + "loss": 3.5736, + "step": 161450 + }, + { + "epoch": 1.8143723002084, + "grad_norm": 1.1836166381835938, + "learning_rate": 1.2745121389491189e-05, + "loss": 3.7147, + "step": 161500 + }, + { + "epoch": 1.8149340253787432, + "grad_norm": 1.2792590856552124, + "learning_rate": 1.2742874476188338e-05, + "loss": 3.6766, + "step": 161550 + }, + { + "epoch": 1.8154957505490863, + "grad_norm": 1.2648366689682007, + "learning_rate": 1.2740627562885488e-05, + "loss": 3.714, + "step": 161600 + }, + { + "epoch": 1.8160574757194294, + "grad_norm": 1.0378963947296143, + "learning_rate": 1.2738380649582639e-05, + "loss": 3.6357, + "step": 161650 + }, + { + "epoch": 1.8166192008897726, + "grad_norm": 1.0107951164245605, + "learning_rate": 1.2736133736279788e-05, + "loss": 3.7196, + "step": 161700 + }, + { + "epoch": 1.8171809260601157, + "grad_norm": 1.324353814125061, + "learning_rate": 1.2733886822976937e-05, + "loss": 3.6357, + "step": 161750 + }, + { + "epoch": 1.8177426512304589, + "grad_norm": 1.1754910945892334, + "learning_rate": 1.2731639909674087e-05, + "loss": 3.6433, + "step": 161800 + }, + { + "epoch": 1.818304376400802, + "grad_norm": 1.1803934574127197, + "learning_rate": 1.2729392996371238e-05, + "loss": 3.6711, + "step": 161850 + }, + { + "epoch": 1.8188661015711451, + "grad_norm": 1.0327410697937012, + "learning_rate": 1.2727146083068387e-05, + "loss": 3.7195, + "step": 161900 + }, + { + "epoch": 1.8194278267414885, + "grad_norm": 1.205920934677124, + "learning_rate": 1.2724899169765536e-05, + "loss": 3.6471, + "step": 161950 + }, + { + "epoch": 1.8199895519118316, + "grad_norm": 0.9720444083213806, + "learning_rate": 1.2722652256462686e-05, + "loss": 3.6269, + "step": 162000 + }, + { + "epoch": 1.8205512770821748, + "grad_norm": 1.2071832418441772, + "learning_rate": 1.2720405343159837e-05, + "loss": 3.6919, + "step": 162050 + }, + { + "epoch": 1.821113002252518, + "grad_norm": 1.2657171487808228, + "learning_rate": 1.2718158429856986e-05, + "loss": 3.6709, + "step": 162100 + }, + { + "epoch": 1.821674727422861, + "grad_norm": 1.1440403461456299, + "learning_rate": 1.2715911516554135e-05, + "loss": 3.705, + "step": 162150 + }, + { + "epoch": 1.8222364525932042, + "grad_norm": 1.1335704326629639, + "learning_rate": 1.2713664603251286e-05, + "loss": 3.6568, + "step": 162200 + }, + { + "epoch": 1.8227981777635474, + "grad_norm": 1.2695765495300293, + "learning_rate": 1.2711417689948436e-05, + "loss": 3.6917, + "step": 162250 + }, + { + "epoch": 1.8233599029338907, + "grad_norm": 1.1169294118881226, + "learning_rate": 1.2709170776645585e-05, + "loss": 3.6422, + "step": 162300 + }, + { + "epoch": 1.8239216281042339, + "grad_norm": 1.1687711477279663, + "learning_rate": 1.2706923863342734e-05, + "loss": 3.7103, + "step": 162350 + }, + { + "epoch": 1.824483353274577, + "grad_norm": 1.162584662437439, + "learning_rate": 1.2704676950039885e-05, + "loss": 3.6889, + "step": 162400 + }, + { + "epoch": 1.8250450784449201, + "grad_norm": 1.252883791923523, + "learning_rate": 1.2702430036737035e-05, + "loss": 3.6662, + "step": 162450 + }, + { + "epoch": 1.8256068036152633, + "grad_norm": 1.131441593170166, + "learning_rate": 1.2700183123434184e-05, + "loss": 3.7217, + "step": 162500 + }, + { + "epoch": 1.8261685287856064, + "grad_norm": 1.153075098991394, + "learning_rate": 1.2697936210131335e-05, + "loss": 3.6241, + "step": 162550 + }, + { + "epoch": 1.8267302539559496, + "grad_norm": 1.070625901222229, + "learning_rate": 1.2695689296828484e-05, + "loss": 3.6846, + "step": 162600 + }, + { + "epoch": 1.8272919791262927, + "grad_norm": 1.016002893447876, + "learning_rate": 1.2693442383525633e-05, + "loss": 3.6648, + "step": 162650 + }, + { + "epoch": 1.8278537042966359, + "grad_norm": 1.0782279968261719, + "learning_rate": 1.2691195470222783e-05, + "loss": 3.6122, + "step": 162700 + }, + { + "epoch": 1.828415429466979, + "grad_norm": 1.3411065340042114, + "learning_rate": 1.2688948556919934e-05, + "loss": 3.6722, + "step": 162750 + }, + { + "epoch": 1.8289771546373221, + "grad_norm": 0.9754913449287415, + "learning_rate": 1.2686701643617083e-05, + "loss": 3.6901, + "step": 162800 + }, + { + "epoch": 1.8295388798076653, + "grad_norm": 1.3152989149093628, + "learning_rate": 1.2684454730314232e-05, + "loss": 3.7504, + "step": 162850 + }, + { + "epoch": 1.8301006049780084, + "grad_norm": 1.2586301565170288, + "learning_rate": 1.2682207817011383e-05, + "loss": 3.6349, + "step": 162900 + }, + { + "epoch": 1.8306623301483516, + "grad_norm": 1.1225285530090332, + "learning_rate": 1.2679960903708533e-05, + "loss": 3.7205, + "step": 162950 + }, + { + "epoch": 1.8312240553186947, + "grad_norm": 1.2042137384414673, + "learning_rate": 1.2677713990405682e-05, + "loss": 3.613, + "step": 163000 + }, + { + "epoch": 1.8317857804890378, + "grad_norm": 1.0179029703140259, + "learning_rate": 1.2675467077102831e-05, + "loss": 3.7395, + "step": 163050 + }, + { + "epoch": 1.832347505659381, + "grad_norm": 1.2871730327606201, + "learning_rate": 1.2673220163799982e-05, + "loss": 3.6364, + "step": 163100 + }, + { + "epoch": 1.8329092308297241, + "grad_norm": 0.9877760410308838, + "learning_rate": 1.2670973250497132e-05, + "loss": 3.6152, + "step": 163150 + }, + { + "epoch": 1.8334709560000673, + "grad_norm": 1.0872858762741089, + "learning_rate": 1.2668726337194281e-05, + "loss": 3.6376, + "step": 163200 + }, + { + "epoch": 1.8340326811704104, + "grad_norm": 1.5374056100845337, + "learning_rate": 1.2666479423891432e-05, + "loss": 3.6909, + "step": 163250 + }, + { + "epoch": 1.8345944063407538, + "grad_norm": 0.9771916270256042, + "learning_rate": 1.2664232510588581e-05, + "loss": 3.6638, + "step": 163300 + }, + { + "epoch": 1.835156131511097, + "grad_norm": 1.1272979974746704, + "learning_rate": 1.266198559728573e-05, + "loss": 3.6806, + "step": 163350 + }, + { + "epoch": 1.83571785668144, + "grad_norm": 1.0883663892745972, + "learning_rate": 1.2659783622248935e-05, + "loss": 3.596, + "step": 163400 + }, + { + "epoch": 1.8362795818517832, + "grad_norm": 1.067847490310669, + "learning_rate": 1.2657536708946086e-05, + "loss": 3.7046, + "step": 163450 + }, + { + "epoch": 1.8368413070221263, + "grad_norm": 2.010690212249756, + "learning_rate": 1.2655289795643235e-05, + "loss": 3.6551, + "step": 163500 + }, + { + "epoch": 1.8374030321924695, + "grad_norm": 1.235400915145874, + "learning_rate": 1.2653042882340385e-05, + "loss": 3.6273, + "step": 163550 + }, + { + "epoch": 1.8379647573628126, + "grad_norm": 1.1057569980621338, + "learning_rate": 1.2650795969037536e-05, + "loss": 3.5601, + "step": 163600 + }, + { + "epoch": 1.838526482533156, + "grad_norm": 1.0582396984100342, + "learning_rate": 1.2648549055734685e-05, + "loss": 3.69, + "step": 163650 + }, + { + "epoch": 1.8390882077034991, + "grad_norm": 1.1918880939483643, + "learning_rate": 1.2646302142431834e-05, + "loss": 3.7185, + "step": 163700 + }, + { + "epoch": 1.8396499328738423, + "grad_norm": 1.1677442789077759, + "learning_rate": 1.2644055229128984e-05, + "loss": 3.6586, + "step": 163750 + }, + { + "epoch": 1.8402116580441854, + "grad_norm": 1.2134548425674438, + "learning_rate": 1.2641808315826135e-05, + "loss": 3.8218, + "step": 163800 + }, + { + "epoch": 1.8407733832145285, + "grad_norm": 1.1581451892852783, + "learning_rate": 1.2639561402523284e-05, + "loss": 3.7075, + "step": 163850 + }, + { + "epoch": 1.8413351083848717, + "grad_norm": 1.1051182746887207, + "learning_rate": 1.2637314489220433e-05, + "loss": 3.7662, + "step": 163900 + }, + { + "epoch": 1.8418968335552148, + "grad_norm": 1.2703778743743896, + "learning_rate": 1.2635067575917584e-05, + "loss": 3.6097, + "step": 163950 + }, + { + "epoch": 1.842458558725558, + "grad_norm": 1.1580830812454224, + "learning_rate": 1.2632820662614734e-05, + "loss": 3.7451, + "step": 164000 + }, + { + "epoch": 1.8430202838959011, + "grad_norm": 1.3673245906829834, + "learning_rate": 1.2630573749311883e-05, + "loss": 3.7244, + "step": 164050 + }, + { + "epoch": 1.8435820090662443, + "grad_norm": 1.2850085496902466, + "learning_rate": 1.2628326836009032e-05, + "loss": 3.669, + "step": 164100 + }, + { + "epoch": 1.8441437342365874, + "grad_norm": 1.1741001605987549, + "learning_rate": 1.2626079922706183e-05, + "loss": 3.7257, + "step": 164150 + }, + { + "epoch": 1.8447054594069305, + "grad_norm": 1.191554069519043, + "learning_rate": 1.2623833009403332e-05, + "loss": 3.6091, + "step": 164200 + }, + { + "epoch": 1.8452671845772737, + "grad_norm": 1.0448939800262451, + "learning_rate": 1.2621586096100482e-05, + "loss": 3.6468, + "step": 164250 + }, + { + "epoch": 1.8458289097476168, + "grad_norm": 1.1517430543899536, + "learning_rate": 1.2619339182797631e-05, + "loss": 3.7041, + "step": 164300 + }, + { + "epoch": 1.84639063491796, + "grad_norm": 1.2172197103500366, + "learning_rate": 1.2617092269494782e-05, + "loss": 3.6273, + "step": 164350 + }, + { + "epoch": 1.846952360088303, + "grad_norm": 1.2285085916519165, + "learning_rate": 1.2614845356191931e-05, + "loss": 3.5998, + "step": 164400 + }, + { + "epoch": 1.8475140852586462, + "grad_norm": 1.0099074840545654, + "learning_rate": 1.261259844288908e-05, + "loss": 3.6736, + "step": 164450 + }, + { + "epoch": 1.8480758104289894, + "grad_norm": 1.1768802404403687, + "learning_rate": 1.2610351529586232e-05, + "loss": 3.6699, + "step": 164500 + }, + { + "epoch": 1.8486375355993325, + "grad_norm": 1.1922498941421509, + "learning_rate": 1.2608104616283381e-05, + "loss": 3.7537, + "step": 164550 + }, + { + "epoch": 1.8491992607696757, + "grad_norm": 1.1091722249984741, + "learning_rate": 1.260585770298053e-05, + "loss": 3.6485, + "step": 164600 + }, + { + "epoch": 1.849760985940019, + "grad_norm": 1.1923208236694336, + "learning_rate": 1.260361078967768e-05, + "loss": 3.706, + "step": 164650 + }, + { + "epoch": 1.8503227111103622, + "grad_norm": 1.2572157382965088, + "learning_rate": 1.260136387637483e-05, + "loss": 3.6803, + "step": 164700 + }, + { + "epoch": 1.8508844362807053, + "grad_norm": 1.1056233644485474, + "learning_rate": 1.259911696307198e-05, + "loss": 3.6433, + "step": 164750 + }, + { + "epoch": 1.8514461614510485, + "grad_norm": 1.1612497568130493, + "learning_rate": 1.259687004976913e-05, + "loss": 3.6575, + "step": 164800 + }, + { + "epoch": 1.8520078866213916, + "grad_norm": 1.1800098419189453, + "learning_rate": 1.259462313646628e-05, + "loss": 3.6447, + "step": 164850 + }, + { + "epoch": 1.8525696117917347, + "grad_norm": 1.2071192264556885, + "learning_rate": 1.259237622316343e-05, + "loss": 3.6733, + "step": 164900 + }, + { + "epoch": 1.8531313369620779, + "grad_norm": 1.476125717163086, + "learning_rate": 1.2590129309860579e-05, + "loss": 3.6673, + "step": 164950 + }, + { + "epoch": 1.8536930621324212, + "grad_norm": 1.2621724605560303, + "learning_rate": 1.2587882396557728e-05, + "loss": 3.6506, + "step": 165000 + }, + { + "epoch": 1.8542547873027644, + "grad_norm": 1.1475720405578613, + "learning_rate": 1.258563548325488e-05, + "loss": 3.6903, + "step": 165050 + }, + { + "epoch": 1.8548165124731075, + "grad_norm": 1.089660406112671, + "learning_rate": 1.2583388569952029e-05, + "loss": 3.6402, + "step": 165100 + }, + { + "epoch": 1.8553782376434507, + "grad_norm": 1.3196265697479248, + "learning_rate": 1.2581141656649178e-05, + "loss": 3.6875, + "step": 165150 + }, + { + "epoch": 1.8559399628137938, + "grad_norm": 1.2380402088165283, + "learning_rate": 1.2578894743346329e-05, + "loss": 3.6614, + "step": 165200 + }, + { + "epoch": 1.856501687984137, + "grad_norm": 1.473062515258789, + "learning_rate": 1.2576647830043478e-05, + "loss": 3.7173, + "step": 165250 + }, + { + "epoch": 1.85706341315448, + "grad_norm": 1.1862760782241821, + "learning_rate": 1.2574400916740628e-05, + "loss": 3.6043, + "step": 165300 + }, + { + "epoch": 1.8576251383248232, + "grad_norm": 1.079772710800171, + "learning_rate": 1.2572154003437777e-05, + "loss": 3.7339, + "step": 165350 + }, + { + "epoch": 1.8581868634951664, + "grad_norm": 1.1612881422042847, + "learning_rate": 1.2569907090134928e-05, + "loss": 3.6729, + "step": 165400 + }, + { + "epoch": 1.8587485886655095, + "grad_norm": 1.1970138549804688, + "learning_rate": 1.2567660176832077e-05, + "loss": 3.671, + "step": 165450 + }, + { + "epoch": 1.8593103138358527, + "grad_norm": 1.266577124595642, + "learning_rate": 1.2565413263529227e-05, + "loss": 3.6625, + "step": 165500 + }, + { + "epoch": 1.8598720390061958, + "grad_norm": 1.0181468725204468, + "learning_rate": 1.2563211288492434e-05, + "loss": 3.6748, + "step": 165550 + }, + { + "epoch": 1.860433764176539, + "grad_norm": 1.2418451309204102, + "learning_rate": 1.2560964375189583e-05, + "loss": 3.6395, + "step": 165600 + }, + { + "epoch": 1.860995489346882, + "grad_norm": 2.6148860454559326, + "learning_rate": 1.2558717461886733e-05, + "loss": 3.6926, + "step": 165650 + }, + { + "epoch": 1.8615572145172252, + "grad_norm": 1.2383267879486084, + "learning_rate": 1.2556470548583884e-05, + "loss": 3.708, + "step": 165700 + }, + { + "epoch": 1.8621189396875684, + "grad_norm": 1.0923413038253784, + "learning_rate": 1.2554223635281033e-05, + "loss": 3.6836, + "step": 165750 + }, + { + "epoch": 1.8626806648579115, + "grad_norm": 1.1667050123214722, + "learning_rate": 1.2551976721978182e-05, + "loss": 3.6588, + "step": 165800 + }, + { + "epoch": 1.8632423900282546, + "grad_norm": 1.1292152404785156, + "learning_rate": 1.2549729808675333e-05, + "loss": 3.6692, + "step": 165850 + }, + { + "epoch": 1.8638041151985978, + "grad_norm": 1.0999101400375366, + "learning_rate": 1.2547482895372483e-05, + "loss": 3.6453, + "step": 165900 + }, + { + "epoch": 1.864365840368941, + "grad_norm": 1.339193344116211, + "learning_rate": 1.2545235982069632e-05, + "loss": 3.6964, + "step": 165950 + }, + { + "epoch": 1.8649275655392843, + "grad_norm": 1.148620843887329, + "learning_rate": 1.2542989068766781e-05, + "loss": 3.7522, + "step": 166000 + }, + { + "epoch": 1.8654892907096274, + "grad_norm": 1.2183562517166138, + "learning_rate": 1.2540742155463932e-05, + "loss": 3.6897, + "step": 166050 + }, + { + "epoch": 1.8660510158799706, + "grad_norm": 1.0874114036560059, + "learning_rate": 1.2538495242161082e-05, + "loss": 3.7102, + "step": 166100 + }, + { + "epoch": 1.8666127410503137, + "grad_norm": 1.3484716415405273, + "learning_rate": 1.2536248328858231e-05, + "loss": 3.6323, + "step": 166150 + }, + { + "epoch": 1.8671744662206569, + "grad_norm": 1.0507627725601196, + "learning_rate": 1.2534001415555382e-05, + "loss": 3.6477, + "step": 166200 + }, + { + "epoch": 1.867736191391, + "grad_norm": 0.9727892279624939, + "learning_rate": 1.2531754502252531e-05, + "loss": 3.7327, + "step": 166250 + }, + { + "epoch": 1.8682979165613434, + "grad_norm": 1.1271682977676392, + "learning_rate": 1.252950758894968e-05, + "loss": 3.6228, + "step": 166300 + }, + { + "epoch": 1.8688596417316865, + "grad_norm": 1.0812020301818848, + "learning_rate": 1.252726067564683e-05, + "loss": 3.7272, + "step": 166350 + }, + { + "epoch": 1.8694213669020296, + "grad_norm": 1.0943026542663574, + "learning_rate": 1.2525013762343981e-05, + "loss": 3.6819, + "step": 166400 + }, + { + "epoch": 1.8699830920723728, + "grad_norm": 1.1824041604995728, + "learning_rate": 1.252276684904113e-05, + "loss": 3.6363, + "step": 166450 + }, + { + "epoch": 1.870544817242716, + "grad_norm": 1.389378547668457, + "learning_rate": 1.252051993573828e-05, + "loss": 3.659, + "step": 166500 + }, + { + "epoch": 1.871106542413059, + "grad_norm": 1.2918018102645874, + "learning_rate": 1.251827302243543e-05, + "loss": 3.604, + "step": 166550 + }, + { + "epoch": 1.8716682675834022, + "grad_norm": 1.1217737197875977, + "learning_rate": 1.251602610913258e-05, + "loss": 3.7202, + "step": 166600 + }, + { + "epoch": 1.8722299927537454, + "grad_norm": 1.286453366279602, + "learning_rate": 1.251377919582973e-05, + "loss": 3.6355, + "step": 166650 + }, + { + "epoch": 1.8727917179240885, + "grad_norm": 1.2219747304916382, + "learning_rate": 1.2511532282526879e-05, + "loss": 3.7565, + "step": 166700 + }, + { + "epoch": 1.8733534430944316, + "grad_norm": 1.135813593864441, + "learning_rate": 1.250928536922403e-05, + "loss": 3.6647, + "step": 166750 + }, + { + "epoch": 1.8739151682647748, + "grad_norm": 1.8363410234451294, + "learning_rate": 1.2507038455921179e-05, + "loss": 3.6416, + "step": 166800 + }, + { + "epoch": 1.874476893435118, + "grad_norm": 1.0600069761276245, + "learning_rate": 1.2504791542618328e-05, + "loss": 3.7299, + "step": 166850 + }, + { + "epoch": 1.875038618605461, + "grad_norm": 0.9839280247688293, + "learning_rate": 1.2502544629315478e-05, + "loss": 3.7078, + "step": 166900 + }, + { + "epoch": 1.8756003437758042, + "grad_norm": 1.0196044445037842, + "learning_rate": 1.2500297716012629e-05, + "loss": 3.6207, + "step": 166950 + }, + { + "epoch": 1.8761620689461473, + "grad_norm": 1.2017905712127686, + "learning_rate": 1.2498050802709778e-05, + "loss": 3.6814, + "step": 167000 + }, + { + "epoch": 1.8767237941164905, + "grad_norm": 1.1838090419769287, + "learning_rate": 1.2495803889406927e-05, + "loss": 3.6566, + "step": 167050 + }, + { + "epoch": 1.8772855192868336, + "grad_norm": 1.3739231824874878, + "learning_rate": 1.2493556976104078e-05, + "loss": 3.6778, + "step": 167100 + }, + { + "epoch": 1.8778472444571768, + "grad_norm": 1.1968698501586914, + "learning_rate": 1.2491310062801228e-05, + "loss": 3.6274, + "step": 167150 + }, + { + "epoch": 1.87840896962752, + "grad_norm": 1.2168124914169312, + "learning_rate": 1.2489063149498377e-05, + "loss": 3.7561, + "step": 167200 + }, + { + "epoch": 1.878970694797863, + "grad_norm": 0.9981439709663391, + "learning_rate": 1.2486816236195526e-05, + "loss": 3.6272, + "step": 167250 + }, + { + "epoch": 1.8795324199682062, + "grad_norm": 1.1597669124603271, + "learning_rate": 1.2484569322892677e-05, + "loss": 3.6901, + "step": 167300 + }, + { + "epoch": 1.8800941451385496, + "grad_norm": 1.1445658206939697, + "learning_rate": 1.2482322409589827e-05, + "loss": 3.6727, + "step": 167350 + }, + { + "epoch": 1.8806558703088927, + "grad_norm": 1.12496817111969, + "learning_rate": 1.2480075496286976e-05, + "loss": 3.6599, + "step": 167400 + }, + { + "epoch": 1.8812175954792358, + "grad_norm": 0.900623619556427, + "learning_rate": 1.2477828582984127e-05, + "loss": 3.6356, + "step": 167450 + }, + { + "epoch": 1.881779320649579, + "grad_norm": 1.2634716033935547, + "learning_rate": 1.2475581669681276e-05, + "loss": 3.6375, + "step": 167500 + }, + { + "epoch": 1.8823410458199221, + "grad_norm": 1.1005879640579224, + "learning_rate": 1.2473334756378425e-05, + "loss": 3.665, + "step": 167550 + }, + { + "epoch": 1.8829027709902653, + "grad_norm": 1.0364843606948853, + "learning_rate": 1.2471087843075575e-05, + "loss": 3.6104, + "step": 167600 + }, + { + "epoch": 1.8834644961606086, + "grad_norm": 1.0718927383422852, + "learning_rate": 1.2468840929772726e-05, + "loss": 3.6778, + "step": 167650 + }, + { + "epoch": 1.8840262213309518, + "grad_norm": 1.133893609046936, + "learning_rate": 1.2466594016469875e-05, + "loss": 3.6997, + "step": 167700 + }, + { + "epoch": 1.884587946501295, + "grad_norm": 1.0615676641464233, + "learning_rate": 1.2464347103167024e-05, + "loss": 3.6485, + "step": 167750 + }, + { + "epoch": 1.885149671671638, + "grad_norm": 1.2539339065551758, + "learning_rate": 1.2462100189864175e-05, + "loss": 3.6845, + "step": 167800 + }, + { + "epoch": 1.8857113968419812, + "grad_norm": 1.1269609928131104, + "learning_rate": 1.2459853276561325e-05, + "loss": 3.6504, + "step": 167850 + }, + { + "epoch": 1.8862731220123243, + "grad_norm": 1.220188856124878, + "learning_rate": 1.2457606363258474e-05, + "loss": 3.6054, + "step": 167900 + }, + { + "epoch": 1.8868348471826675, + "grad_norm": 1.1640559434890747, + "learning_rate": 1.2455359449955623e-05, + "loss": 3.6642, + "step": 167950 + }, + { + "epoch": 1.8873965723530106, + "grad_norm": 1.1580772399902344, + "learning_rate": 1.2453112536652774e-05, + "loss": 3.683, + "step": 168000 + }, + { + "epoch": 1.8879582975233538, + "grad_norm": 1.1342999935150146, + "learning_rate": 1.2450865623349924e-05, + "loss": 3.6598, + "step": 168050 + }, + { + "epoch": 1.888520022693697, + "grad_norm": 1.1601992845535278, + "learning_rate": 1.2448618710047073e-05, + "loss": 3.6451, + "step": 168100 + }, + { + "epoch": 1.88908174786404, + "grad_norm": 1.161476969718933, + "learning_rate": 1.2446371796744224e-05, + "loss": 3.6168, + "step": 168150 + }, + { + "epoch": 1.8896434730343832, + "grad_norm": 1.0610288381576538, + "learning_rate": 1.2444124883441373e-05, + "loss": 3.7246, + "step": 168200 + }, + { + "epoch": 1.8902051982047263, + "grad_norm": 1.1727944612503052, + "learning_rate": 1.2441877970138523e-05, + "loss": 3.7116, + "step": 168250 + }, + { + "epoch": 1.8907669233750695, + "grad_norm": 1.1720994710922241, + "learning_rate": 1.2439631056835672e-05, + "loss": 3.6688, + "step": 168300 + }, + { + "epoch": 1.8913286485454126, + "grad_norm": 1.2003391981124878, + "learning_rate": 1.2437384143532823e-05, + "loss": 3.5629, + "step": 168350 + }, + { + "epoch": 1.8918903737157557, + "grad_norm": 1.2217501401901245, + "learning_rate": 1.2435137230229972e-05, + "loss": 3.6789, + "step": 168400 + }, + { + "epoch": 1.8924520988860989, + "grad_norm": 1.2102724313735962, + "learning_rate": 1.2432890316927122e-05, + "loss": 3.698, + "step": 168450 + }, + { + "epoch": 1.893013824056442, + "grad_norm": 1.123992919921875, + "learning_rate": 1.2430643403624273e-05, + "loss": 3.6312, + "step": 168500 + }, + { + "epoch": 1.8935755492267852, + "grad_norm": 1.128458023071289, + "learning_rate": 1.2428396490321422e-05, + "loss": 3.7031, + "step": 168550 + }, + { + "epoch": 1.8941372743971283, + "grad_norm": 1.181666612625122, + "learning_rate": 1.2426149577018571e-05, + "loss": 3.5914, + "step": 168600 + }, + { + "epoch": 1.8946989995674715, + "grad_norm": 1.11942458152771, + "learning_rate": 1.242390266371572e-05, + "loss": 3.7215, + "step": 168650 + }, + { + "epoch": 1.8952607247378148, + "grad_norm": 1.2772291898727417, + "learning_rate": 1.2421655750412872e-05, + "loss": 3.6783, + "step": 168700 + }, + { + "epoch": 1.895822449908158, + "grad_norm": 1.3013349771499634, + "learning_rate": 1.2419408837110021e-05, + "loss": 3.686, + "step": 168750 + }, + { + "epoch": 1.896384175078501, + "grad_norm": 0.983788251876831, + "learning_rate": 1.241716192380717e-05, + "loss": 3.6762, + "step": 168800 + }, + { + "epoch": 1.8969459002488442, + "grad_norm": 1.247036099433899, + "learning_rate": 1.241491501050432e-05, + "loss": 3.6746, + "step": 168850 + }, + { + "epoch": 1.8975076254191874, + "grad_norm": 1.238627314567566, + "learning_rate": 1.241266809720147e-05, + "loss": 3.6322, + "step": 168900 + }, + { + "epoch": 1.8980693505895305, + "grad_norm": 1.1986321210861206, + "learning_rate": 1.241042118389862e-05, + "loss": 3.7118, + "step": 168950 + }, + { + "epoch": 1.8986310757598739, + "grad_norm": 1.5572178363800049, + "learning_rate": 1.240817427059577e-05, + "loss": 3.6297, + "step": 169000 + }, + { + "epoch": 1.899192800930217, + "grad_norm": 1.256276249885559, + "learning_rate": 1.240592735729292e-05, + "loss": 3.6844, + "step": 169050 + }, + { + "epoch": 1.8997545261005602, + "grad_norm": 1.120082974433899, + "learning_rate": 1.240368044399007e-05, + "loss": 3.6543, + "step": 169100 + }, + { + "epoch": 1.9003162512709033, + "grad_norm": 1.2404358386993408, + "learning_rate": 1.2401433530687219e-05, + "loss": 3.6648, + "step": 169150 + }, + { + "epoch": 1.9008779764412465, + "grad_norm": 1.2647347450256348, + "learning_rate": 1.2399186617384368e-05, + "loss": 3.7098, + "step": 169200 + }, + { + "epoch": 1.9014397016115896, + "grad_norm": 1.0600517988204956, + "learning_rate": 1.239693970408152e-05, + "loss": 3.6644, + "step": 169250 + }, + { + "epoch": 1.9020014267819327, + "grad_norm": 1.066779613494873, + "learning_rate": 1.2394692790778668e-05, + "loss": 3.7638, + "step": 169300 + }, + { + "epoch": 1.9025631519522759, + "grad_norm": 1.0992931127548218, + "learning_rate": 1.2392445877475818e-05, + "loss": 3.6887, + "step": 169350 + }, + { + "epoch": 1.903124877122619, + "grad_norm": 1.0564545392990112, + "learning_rate": 1.2390198964172969e-05, + "loss": 3.6604, + "step": 169400 + }, + { + "epoch": 1.9036866022929622, + "grad_norm": 1.122517704963684, + "learning_rate": 1.2387952050870118e-05, + "loss": 3.621, + "step": 169450 + }, + { + "epoch": 1.9042483274633053, + "grad_norm": 1.1531624794006348, + "learning_rate": 1.2385705137567267e-05, + "loss": 3.6562, + "step": 169500 + }, + { + "epoch": 1.9048100526336484, + "grad_norm": 1.1910247802734375, + "learning_rate": 1.2383458224264417e-05, + "loss": 3.6425, + "step": 169550 + }, + { + "epoch": 1.9053717778039916, + "grad_norm": 1.2163705825805664, + "learning_rate": 1.2381211310961568e-05, + "loss": 3.601, + "step": 169600 + }, + { + "epoch": 1.9059335029743347, + "grad_norm": 1.324793815612793, + "learning_rate": 1.2378964397658717e-05, + "loss": 3.6598, + "step": 169650 + }, + { + "epoch": 1.9064952281446779, + "grad_norm": 1.1938973665237427, + "learning_rate": 1.2376717484355866e-05, + "loss": 3.7127, + "step": 169700 + }, + { + "epoch": 1.907056953315021, + "grad_norm": 1.1128785610198975, + "learning_rate": 1.2374470571053017e-05, + "loss": 3.6492, + "step": 169750 + }, + { + "epoch": 1.9076186784853641, + "grad_norm": 1.3163511753082275, + "learning_rate": 1.2372223657750167e-05, + "loss": 3.6423, + "step": 169800 + }, + { + "epoch": 1.9081804036557073, + "grad_norm": 1.1273045539855957, + "learning_rate": 1.2369976744447316e-05, + "loss": 3.6753, + "step": 169850 + }, + { + "epoch": 1.9087421288260504, + "grad_norm": 1.101928472518921, + "learning_rate": 1.2367729831144465e-05, + "loss": 3.7705, + "step": 169900 + }, + { + "epoch": 1.9093038539963936, + "grad_norm": 1.085390567779541, + "learning_rate": 1.2365482917841616e-05, + "loss": 3.6355, + "step": 169950 + }, + { + "epoch": 1.909865579166737, + "grad_norm": 1.0475810766220093, + "learning_rate": 1.2363236004538766e-05, + "loss": 3.6887, + "step": 170000 + }, + { + "epoch": 1.91042730433708, + "grad_norm": 1.0383238792419434, + "learning_rate": 1.2360989091235915e-05, + "loss": 3.6665, + "step": 170050 + }, + { + "epoch": 1.9109890295074232, + "grad_norm": 1.354358196258545, + "learning_rate": 1.2358742177933066e-05, + "loss": 3.6792, + "step": 170100 + }, + { + "epoch": 1.9115507546777664, + "grad_norm": 1.1882213354110718, + "learning_rate": 1.2356495264630215e-05, + "loss": 3.8028, + "step": 170150 + }, + { + "epoch": 1.9121124798481095, + "grad_norm": 0.9827480912208557, + "learning_rate": 1.2354248351327365e-05, + "loss": 3.6659, + "step": 170200 + }, + { + "epoch": 1.9126742050184526, + "grad_norm": 1.2606571912765503, + "learning_rate": 1.2352001438024514e-05, + "loss": 3.7601, + "step": 170250 + }, + { + "epoch": 1.9132359301887958, + "grad_norm": 1.0390775203704834, + "learning_rate": 1.2349754524721665e-05, + "loss": 3.6812, + "step": 170300 + }, + { + "epoch": 1.9137976553591391, + "grad_norm": 1.2463587522506714, + "learning_rate": 1.2347507611418814e-05, + "loss": 3.6176, + "step": 170350 + }, + { + "epoch": 1.9143593805294823, + "grad_norm": 1.0027610063552856, + "learning_rate": 1.2345305636382022e-05, + "loss": 3.6706, + "step": 170400 + }, + { + "epoch": 1.9149211056998254, + "grad_norm": 1.1805909872055054, + "learning_rate": 1.2343058723079171e-05, + "loss": 3.6551, + "step": 170450 + }, + { + "epoch": 1.9154828308701686, + "grad_norm": 1.3694796562194824, + "learning_rate": 1.234081180977632e-05, + "loss": 3.7188, + "step": 170500 + }, + { + "epoch": 1.9160445560405117, + "grad_norm": 1.2324082851409912, + "learning_rate": 1.233856489647347e-05, + "loss": 3.646, + "step": 170550 + }, + { + "epoch": 1.9166062812108549, + "grad_norm": 0.9831752777099609, + "learning_rate": 1.2336317983170621e-05, + "loss": 3.637, + "step": 170600 + }, + { + "epoch": 1.917168006381198, + "grad_norm": 1.0351693630218506, + "learning_rate": 1.233407106986777e-05, + "loss": 3.7079, + "step": 170650 + }, + { + "epoch": 1.9177297315515411, + "grad_norm": 1.325768232345581, + "learning_rate": 1.233182415656492e-05, + "loss": 3.752, + "step": 170700 + }, + { + "epoch": 1.9182914567218843, + "grad_norm": 1.097913146018982, + "learning_rate": 1.232957724326207e-05, + "loss": 3.6707, + "step": 170750 + }, + { + "epoch": 1.9188531818922274, + "grad_norm": 1.0953683853149414, + "learning_rate": 1.232733032995922e-05, + "loss": 3.6086, + "step": 170800 + }, + { + "epoch": 1.9194149070625706, + "grad_norm": 1.2423237562179565, + "learning_rate": 1.2325083416656369e-05, + "loss": 3.6191, + "step": 170850 + }, + { + "epoch": 1.9199766322329137, + "grad_norm": 1.1714104413986206, + "learning_rate": 1.2322836503353518e-05, + "loss": 3.7158, + "step": 170900 + }, + { + "epoch": 1.9205383574032568, + "grad_norm": 1.0403722524642944, + "learning_rate": 1.2320634528316726e-05, + "loss": 3.6497, + "step": 170950 + }, + { + "epoch": 1.9211000825736, + "grad_norm": 1.111100196838379, + "learning_rate": 1.2318387615013875e-05, + "loss": 3.6101, + "step": 171000 + }, + { + "epoch": 1.9216618077439431, + "grad_norm": 1.2852191925048828, + "learning_rate": 1.2316140701711026e-05, + "loss": 3.6072, + "step": 171050 + }, + { + "epoch": 1.9222235329142863, + "grad_norm": 1.4478634595870972, + "learning_rate": 1.2313893788408176e-05, + "loss": 3.702, + "step": 171100 + }, + { + "epoch": 1.9227852580846294, + "grad_norm": 1.2640856504440308, + "learning_rate": 1.2311646875105325e-05, + "loss": 3.5434, + "step": 171150 + }, + { + "epoch": 1.9233469832549726, + "grad_norm": 1.191200852394104, + "learning_rate": 1.2309444900068533e-05, + "loss": 3.5903, + "step": 171200 + }, + { + "epoch": 1.9239087084253157, + "grad_norm": 1.1147801876068115, + "learning_rate": 1.2307197986765682e-05, + "loss": 3.676, + "step": 171250 + }, + { + "epoch": 1.9244704335956588, + "grad_norm": 1.3612926006317139, + "learning_rate": 1.2304951073462831e-05, + "loss": 3.7212, + "step": 171300 + }, + { + "epoch": 1.9250321587660022, + "grad_norm": 1.052769422531128, + "learning_rate": 1.2302704160159982e-05, + "loss": 3.6766, + "step": 171350 + }, + { + "epoch": 1.9255938839363453, + "grad_norm": 1.0386183261871338, + "learning_rate": 1.2300457246857132e-05, + "loss": 3.6387, + "step": 171400 + }, + { + "epoch": 1.9261556091066885, + "grad_norm": 1.2325382232666016, + "learning_rate": 1.2298210333554281e-05, + "loss": 3.6958, + "step": 171450 + }, + { + "epoch": 1.9267173342770316, + "grad_norm": 1.1987472772598267, + "learning_rate": 1.229596342025143e-05, + "loss": 3.7288, + "step": 171500 + }, + { + "epoch": 1.9272790594473748, + "grad_norm": 1.0316622257232666, + "learning_rate": 1.2293716506948581e-05, + "loss": 3.6099, + "step": 171550 + }, + { + "epoch": 1.927840784617718, + "grad_norm": 1.1573472023010254, + "learning_rate": 1.229146959364573e-05, + "loss": 3.6843, + "step": 171600 + }, + { + "epoch": 1.928402509788061, + "grad_norm": 1.351535677909851, + "learning_rate": 1.228922268034288e-05, + "loss": 3.6412, + "step": 171650 + }, + { + "epoch": 1.9289642349584044, + "grad_norm": 1.1215472221374512, + "learning_rate": 1.2286975767040031e-05, + "loss": 3.6187, + "step": 171700 + }, + { + "epoch": 1.9295259601287476, + "grad_norm": 1.1648633480072021, + "learning_rate": 1.228472885373718e-05, + "loss": 3.6905, + "step": 171750 + }, + { + "epoch": 1.9300876852990907, + "grad_norm": 1.2470959424972534, + "learning_rate": 1.228248194043433e-05, + "loss": 3.6249, + "step": 171800 + }, + { + "epoch": 1.9306494104694338, + "grad_norm": 2.624309778213501, + "learning_rate": 1.2280235027131479e-05, + "loss": 3.6018, + "step": 171850 + }, + { + "epoch": 1.931211135639777, + "grad_norm": 1.2481865882873535, + "learning_rate": 1.227798811382863e-05, + "loss": 3.653, + "step": 171900 + }, + { + "epoch": 1.9317728608101201, + "grad_norm": 1.2186285257339478, + "learning_rate": 1.227574120052578e-05, + "loss": 3.666, + "step": 171950 + }, + { + "epoch": 1.9323345859804633, + "grad_norm": 1.1497619152069092, + "learning_rate": 1.2273494287222929e-05, + "loss": 3.676, + "step": 172000 + }, + { + "epoch": 1.9328963111508064, + "grad_norm": 1.1843317747116089, + "learning_rate": 1.227124737392008e-05, + "loss": 3.6265, + "step": 172050 + }, + { + "epoch": 1.9334580363211495, + "grad_norm": 1.1359236240386963, + "learning_rate": 1.2269000460617229e-05, + "loss": 3.6598, + "step": 172100 + }, + { + "epoch": 1.9340197614914927, + "grad_norm": 1.17789888381958, + "learning_rate": 1.2266753547314378e-05, + "loss": 3.6281, + "step": 172150 + }, + { + "epoch": 1.9345814866618358, + "grad_norm": 1.2440000772476196, + "learning_rate": 1.2264506634011527e-05, + "loss": 3.6616, + "step": 172200 + }, + { + "epoch": 1.935143211832179, + "grad_norm": 1.3088947534561157, + "learning_rate": 1.2262259720708678e-05, + "loss": 3.7083, + "step": 172250 + }, + { + "epoch": 1.935704937002522, + "grad_norm": 1.0753567218780518, + "learning_rate": 1.2260012807405828e-05, + "loss": 3.5984, + "step": 172300 + }, + { + "epoch": 1.9362666621728652, + "grad_norm": 1.1660813093185425, + "learning_rate": 1.2257765894102977e-05, + "loss": 3.649, + "step": 172350 + }, + { + "epoch": 1.9368283873432084, + "grad_norm": 1.3445658683776855, + "learning_rate": 1.2255518980800128e-05, + "loss": 3.7311, + "step": 172400 + }, + { + "epoch": 1.9373901125135515, + "grad_norm": 1.0266178846359253, + "learning_rate": 1.2253272067497277e-05, + "loss": 3.5824, + "step": 172450 + }, + { + "epoch": 1.9379518376838947, + "grad_norm": 1.0390028953552246, + "learning_rate": 1.2251025154194427e-05, + "loss": 3.6594, + "step": 172500 + }, + { + "epoch": 1.9385135628542378, + "grad_norm": 1.387713074684143, + "learning_rate": 1.2248778240891576e-05, + "loss": 3.6387, + "step": 172550 + }, + { + "epoch": 1.939075288024581, + "grad_norm": 1.2674925327301025, + "learning_rate": 1.2246531327588727e-05, + "loss": 3.6946, + "step": 172600 + }, + { + "epoch": 1.939637013194924, + "grad_norm": 1.0736058950424194, + "learning_rate": 1.2244284414285876e-05, + "loss": 3.7241, + "step": 172650 + }, + { + "epoch": 1.9401987383652675, + "grad_norm": 1.0840920209884644, + "learning_rate": 1.2242037500983026e-05, + "loss": 3.643, + "step": 172700 + }, + { + "epoch": 1.9407604635356106, + "grad_norm": 1.102241039276123, + "learning_rate": 1.2239790587680177e-05, + "loss": 3.7588, + "step": 172750 + }, + { + "epoch": 1.9413221887059537, + "grad_norm": 1.0428377389907837, + "learning_rate": 1.2237543674377326e-05, + "loss": 3.5957, + "step": 172800 + }, + { + "epoch": 1.9418839138762969, + "grad_norm": 1.10206937789917, + "learning_rate": 1.2235296761074475e-05, + "loss": 3.6582, + "step": 172850 + }, + { + "epoch": 1.94244563904664, + "grad_norm": 1.1507540941238403, + "learning_rate": 1.2233049847771625e-05, + "loss": 3.7045, + "step": 172900 + }, + { + "epoch": 1.9430073642169832, + "grad_norm": 2.4136881828308105, + "learning_rate": 1.2230802934468776e-05, + "loss": 3.6613, + "step": 172950 + }, + { + "epoch": 1.9435690893873263, + "grad_norm": 1.395912528038025, + "learning_rate": 1.2228556021165925e-05, + "loss": 3.6723, + "step": 173000 + }, + { + "epoch": 1.9441308145576697, + "grad_norm": 1.20100998878479, + "learning_rate": 1.2226309107863074e-05, + "loss": 3.6252, + "step": 173050 + }, + { + "epoch": 1.9446925397280128, + "grad_norm": 1.0831165313720703, + "learning_rate": 1.2224062194560224e-05, + "loss": 3.6134, + "step": 173100 + }, + { + "epoch": 1.945254264898356, + "grad_norm": 1.2886075973510742, + "learning_rate": 1.2221815281257375e-05, + "loss": 3.673, + "step": 173150 + }, + { + "epoch": 1.945815990068699, + "grad_norm": 1.2429815530776978, + "learning_rate": 1.2219568367954524e-05, + "loss": 3.6499, + "step": 173200 + }, + { + "epoch": 1.9463777152390422, + "grad_norm": 1.1157153844833374, + "learning_rate": 1.2217321454651673e-05, + "loss": 3.6351, + "step": 173250 + }, + { + "epoch": 1.9469394404093854, + "grad_norm": 1.1003978252410889, + "learning_rate": 1.2215074541348824e-05, + "loss": 3.6049, + "step": 173300 + }, + { + "epoch": 1.9475011655797285, + "grad_norm": 1.092509150505066, + "learning_rate": 1.2212827628045974e-05, + "loss": 3.6381, + "step": 173350 + }, + { + "epoch": 1.9480628907500717, + "grad_norm": 1.0967395305633545, + "learning_rate": 1.2210580714743123e-05, + "loss": 3.6851, + "step": 173400 + }, + { + "epoch": 1.9486246159204148, + "grad_norm": 1.0878044366836548, + "learning_rate": 1.2208333801440272e-05, + "loss": 3.6377, + "step": 173450 + }, + { + "epoch": 1.949186341090758, + "grad_norm": 1.120462417602539, + "learning_rate": 1.2206086888137423e-05, + "loss": 3.709, + "step": 173500 + }, + { + "epoch": 1.949748066261101, + "grad_norm": 1.0878864526748657, + "learning_rate": 1.2203839974834573e-05, + "loss": 3.6988, + "step": 173550 + }, + { + "epoch": 1.9503097914314442, + "grad_norm": 1.0665409564971924, + "learning_rate": 1.2201593061531722e-05, + "loss": 3.6464, + "step": 173600 + }, + { + "epoch": 1.9508715166017874, + "grad_norm": 1.1327898502349854, + "learning_rate": 1.2199346148228873e-05, + "loss": 3.6786, + "step": 173650 + }, + { + "epoch": 1.9514332417721305, + "grad_norm": 1.3542909622192383, + "learning_rate": 1.2197099234926022e-05, + "loss": 3.6545, + "step": 173700 + }, + { + "epoch": 1.9519949669424737, + "grad_norm": 1.3699864149093628, + "learning_rate": 1.2194852321623172e-05, + "loss": 3.6166, + "step": 173750 + }, + { + "epoch": 1.9525566921128168, + "grad_norm": 1.2797566652297974, + "learning_rate": 1.219260540832032e-05, + "loss": 3.6561, + "step": 173800 + }, + { + "epoch": 1.95311841728316, + "grad_norm": 1.0340110063552856, + "learning_rate": 1.2190358495017472e-05, + "loss": 3.7516, + "step": 173850 + }, + { + "epoch": 1.953680142453503, + "grad_norm": 1.2984986305236816, + "learning_rate": 1.2188111581714621e-05, + "loss": 3.726, + "step": 173900 + }, + { + "epoch": 1.9542418676238462, + "grad_norm": 1.1449670791625977, + "learning_rate": 1.218586466841177e-05, + "loss": 3.6731, + "step": 173950 + }, + { + "epoch": 1.9548035927941894, + "grad_norm": 1.189428448677063, + "learning_rate": 1.2183617755108921e-05, + "loss": 3.655, + "step": 174000 + }, + { + "epoch": 1.9553653179645327, + "grad_norm": 1.0728976726531982, + "learning_rate": 1.218137084180607e-05, + "loss": 3.6844, + "step": 174050 + }, + { + "epoch": 1.9559270431348759, + "grad_norm": 1.316543698310852, + "learning_rate": 1.217912392850322e-05, + "loss": 3.685, + "step": 174100 + }, + { + "epoch": 1.956488768305219, + "grad_norm": 1.0091776847839355, + "learning_rate": 1.217687701520037e-05, + "loss": 3.7103, + "step": 174150 + }, + { + "epoch": 1.9570504934755621, + "grad_norm": 1.1065725088119507, + "learning_rate": 1.217463010189752e-05, + "loss": 3.7193, + "step": 174200 + }, + { + "epoch": 1.9576122186459053, + "grad_norm": 1.032600998878479, + "learning_rate": 1.217238318859467e-05, + "loss": 3.6122, + "step": 174250 + }, + { + "epoch": 1.9581739438162484, + "grad_norm": 1.183090090751648, + "learning_rate": 1.2170136275291819e-05, + "loss": 3.6051, + "step": 174300 + }, + { + "epoch": 1.9587356689865916, + "grad_norm": 1.437174677848816, + "learning_rate": 1.216788936198897e-05, + "loss": 3.6526, + "step": 174350 + }, + { + "epoch": 1.959297394156935, + "grad_norm": 1.0554615259170532, + "learning_rate": 1.216564244868612e-05, + "loss": 3.6749, + "step": 174400 + }, + { + "epoch": 1.959859119327278, + "grad_norm": 1.0573745965957642, + "learning_rate": 1.2163395535383269e-05, + "loss": 3.6799, + "step": 174450 + }, + { + "epoch": 1.9604208444976212, + "grad_norm": 1.2770339250564575, + "learning_rate": 1.2161148622080418e-05, + "loss": 3.6972, + "step": 174500 + }, + { + "epoch": 1.9609825696679644, + "grad_norm": 1.0049799680709839, + "learning_rate": 1.2158901708777569e-05, + "loss": 3.6382, + "step": 174550 + }, + { + "epoch": 1.9615442948383075, + "grad_norm": 1.2158585786819458, + "learning_rate": 1.2156654795474718e-05, + "loss": 3.6784, + "step": 174600 + }, + { + "epoch": 1.9621060200086506, + "grad_norm": 1.1403707265853882, + "learning_rate": 1.2154407882171868e-05, + "loss": 3.6615, + "step": 174650 + }, + { + "epoch": 1.9626677451789938, + "grad_norm": 1.1854654550552368, + "learning_rate": 1.2152160968869019e-05, + "loss": 3.6604, + "step": 174700 + }, + { + "epoch": 1.963229470349337, + "grad_norm": 1.1664259433746338, + "learning_rate": 1.2149914055566168e-05, + "loss": 3.6374, + "step": 174750 + }, + { + "epoch": 1.96379119551968, + "grad_norm": 1.2385070323944092, + "learning_rate": 1.2147667142263317e-05, + "loss": 3.6751, + "step": 174800 + }, + { + "epoch": 1.9643529206900232, + "grad_norm": 1.2153853178024292, + "learning_rate": 1.2145420228960467e-05, + "loss": 3.6107, + "step": 174850 + }, + { + "epoch": 1.9649146458603663, + "grad_norm": 1.2236731052398682, + "learning_rate": 1.2143173315657618e-05, + "loss": 3.6285, + "step": 174900 + }, + { + "epoch": 1.9654763710307095, + "grad_norm": 1.0807764530181885, + "learning_rate": 1.2140926402354767e-05, + "loss": 3.6047, + "step": 174950 + }, + { + "epoch": 1.9660380962010526, + "grad_norm": 1.1282566785812378, + "learning_rate": 1.2138679489051916e-05, + "loss": 3.6077, + "step": 175000 + }, + { + "epoch": 1.9665998213713958, + "grad_norm": 1.238545536994934, + "learning_rate": 1.2136432575749066e-05, + "loss": 3.7162, + "step": 175050 + }, + { + "epoch": 1.967161546541739, + "grad_norm": 1.094677448272705, + "learning_rate": 1.2134185662446217e-05, + "loss": 3.699, + "step": 175100 + }, + { + "epoch": 1.967723271712082, + "grad_norm": 1.2476550340652466, + "learning_rate": 1.2131938749143366e-05, + "loss": 3.6716, + "step": 175150 + }, + { + "epoch": 1.9682849968824252, + "grad_norm": 1.1629317998886108, + "learning_rate": 1.2129691835840515e-05, + "loss": 3.6337, + "step": 175200 + }, + { + "epoch": 1.9688467220527683, + "grad_norm": 1.261732816696167, + "learning_rate": 1.2127444922537666e-05, + "loss": 3.6534, + "step": 175250 + }, + { + "epoch": 1.9694084472231115, + "grad_norm": 1.1375350952148438, + "learning_rate": 1.2125198009234816e-05, + "loss": 3.7596, + "step": 175300 + }, + { + "epoch": 1.9699701723934546, + "grad_norm": 1.0128322839736938, + "learning_rate": 1.2122951095931965e-05, + "loss": 3.5842, + "step": 175350 + }, + { + "epoch": 1.970531897563798, + "grad_norm": 1.4134347438812256, + "learning_rate": 1.2120704182629114e-05, + "loss": 3.7659, + "step": 175400 + }, + { + "epoch": 1.9710936227341411, + "grad_norm": 1.1948753595352173, + "learning_rate": 1.2118457269326265e-05, + "loss": 3.6362, + "step": 175450 + }, + { + "epoch": 1.9716553479044843, + "grad_norm": 1.2572717666625977, + "learning_rate": 1.2116210356023415e-05, + "loss": 3.7703, + "step": 175500 + }, + { + "epoch": 1.9722170730748274, + "grad_norm": 1.0441925525665283, + "learning_rate": 1.2113963442720564e-05, + "loss": 3.6649, + "step": 175550 + }, + { + "epoch": 1.9727787982451706, + "grad_norm": 1.2285887002944946, + "learning_rate": 1.2111716529417715e-05, + "loss": 3.6652, + "step": 175600 + }, + { + "epoch": 1.9733405234155137, + "grad_norm": 1.2192641496658325, + "learning_rate": 1.2109469616114864e-05, + "loss": 3.6174, + "step": 175650 + }, + { + "epoch": 1.9739022485858568, + "grad_norm": 1.2564349174499512, + "learning_rate": 1.2107222702812014e-05, + "loss": 3.6524, + "step": 175700 + }, + { + "epoch": 1.9744639737562002, + "grad_norm": 1.3048930168151855, + "learning_rate": 1.2104975789509163e-05, + "loss": 3.7103, + "step": 175750 + }, + { + "epoch": 1.9750256989265433, + "grad_norm": 1.180362582206726, + "learning_rate": 1.2102728876206314e-05, + "loss": 3.6536, + "step": 175800 + }, + { + "epoch": 1.9755874240968865, + "grad_norm": 1.1488772630691528, + "learning_rate": 1.210052690116952e-05, + "loss": 3.7154, + "step": 175850 + }, + { + "epoch": 1.9761491492672296, + "grad_norm": 1.1649621725082397, + "learning_rate": 1.209827998786667e-05, + "loss": 3.5611, + "step": 175900 + }, + { + "epoch": 1.9767108744375728, + "grad_norm": 1.1073241233825684, + "learning_rate": 1.209603307456382e-05, + "loss": 3.6528, + "step": 175950 + }, + { + "epoch": 1.977272599607916, + "grad_norm": 1.1920549869537354, + "learning_rate": 1.209378616126097e-05, + "loss": 3.6264, + "step": 176000 + }, + { + "epoch": 1.977834324778259, + "grad_norm": 1.4693737030029297, + "learning_rate": 1.2091539247958119e-05, + "loss": 3.6645, + "step": 176050 + }, + { + "epoch": 1.9783960499486022, + "grad_norm": 2.4251062870025635, + "learning_rate": 1.208929233465527e-05, + "loss": 3.6509, + "step": 176100 + }, + { + "epoch": 1.9789577751189453, + "grad_norm": 1.1257350444793701, + "learning_rate": 1.2087045421352419e-05, + "loss": 3.6939, + "step": 176150 + }, + { + "epoch": 1.9795195002892885, + "grad_norm": 1.2718253135681152, + "learning_rate": 1.2084798508049568e-05, + "loss": 3.6891, + "step": 176200 + }, + { + "epoch": 1.9800812254596316, + "grad_norm": 1.1150072813034058, + "learning_rate": 1.208255159474672e-05, + "loss": 3.6344, + "step": 176250 + }, + { + "epoch": 1.9806429506299748, + "grad_norm": 1.1545917987823486, + "learning_rate": 1.2080304681443869e-05, + "loss": 3.5648, + "step": 176300 + }, + { + "epoch": 1.981204675800318, + "grad_norm": 1.1448496580123901, + "learning_rate": 1.2078057768141018e-05, + "loss": 3.6085, + "step": 176350 + }, + { + "epoch": 1.981766400970661, + "grad_norm": 1.113794207572937, + "learning_rate": 1.2075810854838167e-05, + "loss": 3.5622, + "step": 176400 + }, + { + "epoch": 1.9823281261410042, + "grad_norm": 1.2571277618408203, + "learning_rate": 1.2073563941535318e-05, + "loss": 3.5989, + "step": 176450 + }, + { + "epoch": 1.9828898513113473, + "grad_norm": 1.4230687618255615, + "learning_rate": 1.2071317028232468e-05, + "loss": 3.6226, + "step": 176500 + }, + { + "epoch": 1.9834515764816905, + "grad_norm": 1.2205623388290405, + "learning_rate": 1.2069070114929617e-05, + "loss": 3.7403, + "step": 176550 + }, + { + "epoch": 1.9840133016520336, + "grad_norm": 1.0938069820404053, + "learning_rate": 1.2066823201626768e-05, + "loss": 3.7403, + "step": 176600 + }, + { + "epoch": 1.9845750268223767, + "grad_norm": 1.2703936100006104, + "learning_rate": 1.2064576288323917e-05, + "loss": 3.6401, + "step": 176650 + }, + { + "epoch": 1.9851367519927199, + "grad_norm": 1.0708928108215332, + "learning_rate": 1.2062329375021067e-05, + "loss": 3.6527, + "step": 176700 + }, + { + "epoch": 1.9856984771630632, + "grad_norm": 1.0272630453109741, + "learning_rate": 1.2060082461718216e-05, + "loss": 3.7237, + "step": 176750 + }, + { + "epoch": 1.9862602023334064, + "grad_norm": 1.2990872859954834, + "learning_rate": 1.2057835548415367e-05, + "loss": 3.5784, + "step": 176800 + }, + { + "epoch": 1.9868219275037495, + "grad_norm": 1.2626888751983643, + "learning_rate": 1.2055588635112516e-05, + "loss": 3.6648, + "step": 176850 + }, + { + "epoch": 1.9873836526740927, + "grad_norm": 1.298158049583435, + "learning_rate": 1.2053341721809666e-05, + "loss": 3.6625, + "step": 176900 + }, + { + "epoch": 1.9879453778444358, + "grad_norm": 1.1682379245758057, + "learning_rate": 1.2051094808506817e-05, + "loss": 3.6812, + "step": 176950 + }, + { + "epoch": 1.988507103014779, + "grad_norm": 1.0595457553863525, + "learning_rate": 1.2048847895203966e-05, + "loss": 3.7265, + "step": 177000 + }, + { + "epoch": 1.989068828185122, + "grad_norm": 1.2153105735778809, + "learning_rate": 1.2046600981901115e-05, + "loss": 3.6573, + "step": 177050 + }, + { + "epoch": 1.9896305533554655, + "grad_norm": 1.3306924104690552, + "learning_rate": 1.2044354068598265e-05, + "loss": 3.6213, + "step": 177100 + }, + { + "epoch": 1.9901922785258086, + "grad_norm": 1.3015691041946411, + "learning_rate": 1.2042107155295416e-05, + "loss": 3.6733, + "step": 177150 + }, + { + "epoch": 1.9907540036961517, + "grad_norm": 1.3520183563232422, + "learning_rate": 1.2039860241992565e-05, + "loss": 3.5986, + "step": 177200 + }, + { + "epoch": 1.9913157288664949, + "grad_norm": 1.1858617067337036, + "learning_rate": 1.2037613328689714e-05, + "loss": 3.6491, + "step": 177250 + }, + { + "epoch": 1.991877454036838, + "grad_norm": 1.3016369342803955, + "learning_rate": 1.2035366415386865e-05, + "loss": 3.7812, + "step": 177300 + }, + { + "epoch": 1.9924391792071812, + "grad_norm": 1.259308099746704, + "learning_rate": 1.2033119502084014e-05, + "loss": 3.663, + "step": 177350 + }, + { + "epoch": 1.9930009043775243, + "grad_norm": 0.9826324582099915, + "learning_rate": 1.2030872588781164e-05, + "loss": 3.6208, + "step": 177400 + }, + { + "epoch": 1.9935626295478674, + "grad_norm": 1.122003436088562, + "learning_rate": 1.2028625675478313e-05, + "loss": 3.5831, + "step": 177450 + }, + { + "epoch": 1.9941243547182106, + "grad_norm": 1.0870612859725952, + "learning_rate": 1.2026378762175464e-05, + "loss": 3.6434, + "step": 177500 + }, + { + "epoch": 1.9946860798885537, + "grad_norm": 1.2947250604629517, + "learning_rate": 1.2024131848872613e-05, + "loss": 3.6164, + "step": 177550 + }, + { + "epoch": 1.9952478050588969, + "grad_norm": 1.055662751197815, + "learning_rate": 1.2021884935569763e-05, + "loss": 3.6337, + "step": 177600 + }, + { + "epoch": 1.99580953022924, + "grad_norm": 1.093037724494934, + "learning_rate": 1.2019638022266912e-05, + "loss": 3.6574, + "step": 177650 + }, + { + "epoch": 1.9963712553995832, + "grad_norm": 1.1708205938339233, + "learning_rate": 1.2017391108964063e-05, + "loss": 3.6793, + "step": 177700 + }, + { + "epoch": 1.9969329805699263, + "grad_norm": 1.315970778465271, + "learning_rate": 1.2015144195661212e-05, + "loss": 3.6725, + "step": 177750 + }, + { + "epoch": 1.9974947057402694, + "grad_norm": 1.1181354522705078, + "learning_rate": 1.2012897282358362e-05, + "loss": 3.6046, + "step": 177800 + }, + { + "epoch": 1.9980564309106126, + "grad_norm": 1.1091622114181519, + "learning_rate": 1.2010650369055513e-05, + "loss": 3.7024, + "step": 177850 + }, + { + "epoch": 1.9986181560809557, + "grad_norm": 1.0360990762710571, + "learning_rate": 1.2008403455752662e-05, + "loss": 3.6336, + "step": 177900 + }, + { + "epoch": 1.9991798812512989, + "grad_norm": 1.18849515914917, + "learning_rate": 1.2006156542449811e-05, + "loss": 3.6415, + "step": 177950 + }, + { + "epoch": 1.999741606421642, + "grad_norm": 1.1326818466186523, + "learning_rate": 1.200390962914696e-05, + "loss": 3.7134, + "step": 178000 + }, + { + "epoch": 2.0, + "eval_bleu": 5.9547, + "eval_gen_len": 21.5378, + "eval_loss": 3.5279541015625, + "eval_runtime": 25257.1404, + "eval_samples_per_second": 14.097, + "eval_steps_per_second": 0.881, + "step": 178023 + }, + { + "epoch": 2.000303331591985, + "grad_norm": 1.4539496898651123, + "learning_rate": 1.2001662715844112e-05, + "loss": 3.6639, + "step": 178050 + }, + { + "epoch": 2.0008650567623283, + "grad_norm": 1.0738657712936401, + "learning_rate": 1.199941580254126e-05, + "loss": 3.6008, + "step": 178100 + }, + { + "epoch": 2.0014267819326714, + "grad_norm": 1.2506818771362305, + "learning_rate": 1.1997168889238409e-05, + "loss": 3.6954, + "step": 178150 + }, + { + "epoch": 2.0019885071030146, + "grad_norm": 1.0013850927352905, + "learning_rate": 1.1994921975935558e-05, + "loss": 3.6818, + "step": 178200 + }, + { + "epoch": 2.002550232273358, + "grad_norm": 1.1331974267959595, + "learning_rate": 1.1992675062632709e-05, + "loss": 3.6265, + "step": 178250 + }, + { + "epoch": 2.0031119574437013, + "grad_norm": 1.3222802877426147, + "learning_rate": 1.1990428149329858e-05, + "loss": 3.6171, + "step": 178300 + }, + { + "epoch": 2.0036736826140444, + "grad_norm": 1.0162914991378784, + "learning_rate": 1.1988181236027008e-05, + "loss": 3.6243, + "step": 178350 + }, + { + "epoch": 2.0042354077843876, + "grad_norm": 1.3031641244888306, + "learning_rate": 1.1985934322724157e-05, + "loss": 3.7056, + "step": 178400 + }, + { + "epoch": 2.0047971329547307, + "grad_norm": 1.261915922164917, + "learning_rate": 1.1983687409421308e-05, + "loss": 3.6977, + "step": 178450 + }, + { + "epoch": 2.005358858125074, + "grad_norm": 1.172046422958374, + "learning_rate": 1.1981440496118457e-05, + "loss": 3.5842, + "step": 178500 + }, + { + "epoch": 2.005920583295417, + "grad_norm": 1.1569795608520508, + "learning_rate": 1.1979193582815607e-05, + "loss": 3.6705, + "step": 178550 + }, + { + "epoch": 2.00648230846576, + "grad_norm": 1.1903021335601807, + "learning_rate": 1.1976946669512758e-05, + "loss": 3.682, + "step": 178600 + }, + { + "epoch": 2.0070440336361033, + "grad_norm": 1.133834958076477, + "learning_rate": 1.1974699756209907e-05, + "loss": 3.6267, + "step": 178650 + }, + { + "epoch": 2.0076057588064464, + "grad_norm": 0.9919807314872742, + "learning_rate": 1.1972452842907056e-05, + "loss": 3.6693, + "step": 178700 + }, + { + "epoch": 2.0081674839767896, + "grad_norm": 1.1761101484298706, + "learning_rate": 1.1970205929604206e-05, + "loss": 3.6217, + "step": 178750 + }, + { + "epoch": 2.0087292091471327, + "grad_norm": 1.1369194984436035, + "learning_rate": 1.1968003954567413e-05, + "loss": 3.6111, + "step": 178800 + }, + { + "epoch": 2.009290934317476, + "grad_norm": 1.0703468322753906, + "learning_rate": 1.1965757041264562e-05, + "loss": 3.677, + "step": 178850 + }, + { + "epoch": 2.009852659487819, + "grad_norm": 1.1398569345474243, + "learning_rate": 1.1963510127961713e-05, + "loss": 3.5947, + "step": 178900 + }, + { + "epoch": 2.010414384658162, + "grad_norm": 1.134208083152771, + "learning_rate": 1.1961263214658863e-05, + "loss": 3.6926, + "step": 178950 + }, + { + "epoch": 2.0109761098285053, + "grad_norm": 1.1300833225250244, + "learning_rate": 1.1959016301356012e-05, + "loss": 3.6604, + "step": 179000 + }, + { + "epoch": 2.0115378349988484, + "grad_norm": 1.3578588962554932, + "learning_rate": 1.1956769388053161e-05, + "loss": 3.7019, + "step": 179050 + }, + { + "epoch": 2.0120995601691916, + "grad_norm": 1.3306846618652344, + "learning_rate": 1.1954522474750312e-05, + "loss": 3.7365, + "step": 179100 + }, + { + "epoch": 2.0126612853395347, + "grad_norm": 1.1825461387634277, + "learning_rate": 1.1952275561447462e-05, + "loss": 3.6839, + "step": 179150 + }, + { + "epoch": 2.013223010509878, + "grad_norm": 1.1960430145263672, + "learning_rate": 1.1950028648144611e-05, + "loss": 3.6459, + "step": 179200 + }, + { + "epoch": 2.013784735680221, + "grad_norm": 1.111112117767334, + "learning_rate": 1.1947781734841762e-05, + "loss": 3.5988, + "step": 179250 + }, + { + "epoch": 2.014346460850564, + "grad_norm": 1.2758218050003052, + "learning_rate": 1.1945534821538911e-05, + "loss": 3.6658, + "step": 179300 + }, + { + "epoch": 2.0149081860209073, + "grad_norm": 1.2596542835235596, + "learning_rate": 1.194328790823606e-05, + "loss": 3.6152, + "step": 179350 + }, + { + "epoch": 2.0154699111912504, + "grad_norm": 1.0755528211593628, + "learning_rate": 1.194104099493321e-05, + "loss": 3.5839, + "step": 179400 + }, + { + "epoch": 2.0160316363615935, + "grad_norm": 1.2503666877746582, + "learning_rate": 1.1938794081630361e-05, + "loss": 3.6438, + "step": 179450 + }, + { + "epoch": 2.0165933615319367, + "grad_norm": 1.22162663936615, + "learning_rate": 1.193654716832751e-05, + "loss": 3.6185, + "step": 179500 + }, + { + "epoch": 2.01715508670228, + "grad_norm": 1.2688690423965454, + "learning_rate": 1.193430025502466e-05, + "loss": 3.6753, + "step": 179550 + }, + { + "epoch": 2.0177168118726234, + "grad_norm": 1.0360270738601685, + "learning_rate": 1.193205334172181e-05, + "loss": 3.6716, + "step": 179600 + }, + { + "epoch": 2.0182785370429666, + "grad_norm": 1.278264045715332, + "learning_rate": 1.192980642841896e-05, + "loss": 3.6364, + "step": 179650 + }, + { + "epoch": 2.0188402622133097, + "grad_norm": 1.4267991781234741, + "learning_rate": 1.192755951511611e-05, + "loss": 3.6839, + "step": 179700 + }, + { + "epoch": 2.019401987383653, + "grad_norm": 1.0959867238998413, + "learning_rate": 1.1925312601813259e-05, + "loss": 3.6806, + "step": 179750 + }, + { + "epoch": 2.019963712553996, + "grad_norm": 1.009420394897461, + "learning_rate": 1.192306568851041e-05, + "loss": 3.6266, + "step": 179800 + }, + { + "epoch": 2.020525437724339, + "grad_norm": 1.0204261541366577, + "learning_rate": 1.1920818775207559e-05, + "loss": 3.6664, + "step": 179850 + }, + { + "epoch": 2.0210871628946823, + "grad_norm": 1.2288864850997925, + "learning_rate": 1.1918571861904708e-05, + "loss": 3.6486, + "step": 179900 + }, + { + "epoch": 2.0216488880650254, + "grad_norm": 1.0590049028396606, + "learning_rate": 1.1916324948601858e-05, + "loss": 3.651, + "step": 179950 + }, + { + "epoch": 2.0222106132353685, + "grad_norm": 1.2189126014709473, + "learning_rate": 1.1914078035299009e-05, + "loss": 3.6959, + "step": 180000 + }, + { + "epoch": 2.0227723384057117, + "grad_norm": 1.2482175827026367, + "learning_rate": 1.1911831121996158e-05, + "loss": 3.715, + "step": 180050 + }, + { + "epoch": 2.023334063576055, + "grad_norm": 1.0789780616760254, + "learning_rate": 1.1909584208693307e-05, + "loss": 3.6599, + "step": 180100 + }, + { + "epoch": 2.023895788746398, + "grad_norm": 1.0925908088684082, + "learning_rate": 1.1907337295390458e-05, + "loss": 3.6499, + "step": 180150 + }, + { + "epoch": 2.024457513916741, + "grad_norm": 1.1479415893554688, + "learning_rate": 1.1905090382087608e-05, + "loss": 3.7215, + "step": 180200 + }, + { + "epoch": 2.0250192390870843, + "grad_norm": 1.3037455081939697, + "learning_rate": 1.1902843468784757e-05, + "loss": 3.6362, + "step": 180250 + }, + { + "epoch": 2.0255809642574274, + "grad_norm": 1.1210711002349854, + "learning_rate": 1.1900596555481906e-05, + "loss": 3.6479, + "step": 180300 + }, + { + "epoch": 2.0261426894277705, + "grad_norm": 1.0629558563232422, + "learning_rate": 1.1898349642179057e-05, + "loss": 3.6371, + "step": 180350 + }, + { + "epoch": 2.0267044145981137, + "grad_norm": 1.052786946296692, + "learning_rate": 1.1896102728876207e-05, + "loss": 3.712, + "step": 180400 + }, + { + "epoch": 2.027266139768457, + "grad_norm": 1.3238290548324585, + "learning_rate": 1.1893855815573356e-05, + "loss": 3.6417, + "step": 180450 + }, + { + "epoch": 2.0278278649388, + "grad_norm": 0.9874387383460999, + "learning_rate": 1.1891608902270507e-05, + "loss": 3.6863, + "step": 180500 + }, + { + "epoch": 2.028389590109143, + "grad_norm": 1.063574194908142, + "learning_rate": 1.1889361988967656e-05, + "loss": 3.6966, + "step": 180550 + }, + { + "epoch": 2.0289513152794862, + "grad_norm": 1.119516372680664, + "learning_rate": 1.1887115075664805e-05, + "loss": 3.6497, + "step": 180600 + }, + { + "epoch": 2.0295130404498294, + "grad_norm": 1.3257489204406738, + "learning_rate": 1.1884868162361955e-05, + "loss": 3.5998, + "step": 180650 + }, + { + "epoch": 2.0300747656201725, + "grad_norm": 1.4712659120559692, + "learning_rate": 1.1882621249059106e-05, + "loss": 3.7032, + "step": 180700 + }, + { + "epoch": 2.0306364907905157, + "grad_norm": 1.154805302619934, + "learning_rate": 1.1880374335756255e-05, + "loss": 3.6574, + "step": 180750 + }, + { + "epoch": 2.031198215960859, + "grad_norm": 1.1678481101989746, + "learning_rate": 1.1878127422453404e-05, + "loss": 3.7243, + "step": 180800 + }, + { + "epoch": 2.031759941131202, + "grad_norm": 1.2385306358337402, + "learning_rate": 1.1875880509150555e-05, + "loss": 3.6573, + "step": 180850 + }, + { + "epoch": 2.032321666301545, + "grad_norm": 1.1208651065826416, + "learning_rate": 1.1873633595847705e-05, + "loss": 3.6187, + "step": 180900 + }, + { + "epoch": 2.0328833914718887, + "grad_norm": 1.2408610582351685, + "learning_rate": 1.1871386682544854e-05, + "loss": 3.7124, + "step": 180950 + }, + { + "epoch": 2.033445116642232, + "grad_norm": 1.1662001609802246, + "learning_rate": 1.1869139769242003e-05, + "loss": 3.6918, + "step": 181000 + }, + { + "epoch": 2.034006841812575, + "grad_norm": 1.667796015739441, + "learning_rate": 1.1866892855939154e-05, + "loss": 3.6008, + "step": 181050 + }, + { + "epoch": 2.034568566982918, + "grad_norm": 1.2891185283660889, + "learning_rate": 1.1864645942636304e-05, + "loss": 3.692, + "step": 181100 + }, + { + "epoch": 2.0351302921532612, + "grad_norm": 1.0154170989990234, + "learning_rate": 1.1862399029333453e-05, + "loss": 3.7017, + "step": 181150 + }, + { + "epoch": 2.0356920173236044, + "grad_norm": 1.350826382637024, + "learning_rate": 1.1860152116030604e-05, + "loss": 3.6805, + "step": 181200 + }, + { + "epoch": 2.0362537424939475, + "grad_norm": 1.127942442893982, + "learning_rate": 1.185795014099381e-05, + "loss": 3.6208, + "step": 181250 + }, + { + "epoch": 2.0368154676642907, + "grad_norm": 1.4027514457702637, + "learning_rate": 1.185570322769096e-05, + "loss": 3.612, + "step": 181300 + }, + { + "epoch": 2.037377192834634, + "grad_norm": 1.2916492223739624, + "learning_rate": 1.185345631438811e-05, + "loss": 3.6641, + "step": 181350 + }, + { + "epoch": 2.037938918004977, + "grad_norm": 1.3983794450759888, + "learning_rate": 1.185120940108526e-05, + "loss": 3.6643, + "step": 181400 + }, + { + "epoch": 2.03850064317532, + "grad_norm": 1.2258174419403076, + "learning_rate": 1.1848962487782409e-05, + "loss": 3.6637, + "step": 181450 + }, + { + "epoch": 2.0390623683456632, + "grad_norm": 1.2498546838760376, + "learning_rate": 1.184671557447956e-05, + "loss": 3.6297, + "step": 181500 + }, + { + "epoch": 2.0396240935160064, + "grad_norm": 1.1554557085037231, + "learning_rate": 1.184446866117671e-05, + "loss": 3.7559, + "step": 181550 + }, + { + "epoch": 2.0401858186863495, + "grad_norm": 1.154598355293274, + "learning_rate": 1.1842221747873859e-05, + "loss": 3.6876, + "step": 181600 + }, + { + "epoch": 2.0407475438566927, + "grad_norm": 1.2266391515731812, + "learning_rate": 1.1839974834571008e-05, + "loss": 3.6902, + "step": 181650 + }, + { + "epoch": 2.041309269027036, + "grad_norm": 1.2957992553710938, + "learning_rate": 1.1837727921268159e-05, + "loss": 3.6062, + "step": 181700 + }, + { + "epoch": 2.041870994197379, + "grad_norm": 0.9922723770141602, + "learning_rate": 1.1835481007965308e-05, + "loss": 3.5309, + "step": 181750 + }, + { + "epoch": 2.042432719367722, + "grad_norm": 1.1541030406951904, + "learning_rate": 1.1833234094662458e-05, + "loss": 3.6341, + "step": 181800 + }, + { + "epoch": 2.0429944445380652, + "grad_norm": 1.2921535968780518, + "learning_rate": 1.1830987181359609e-05, + "loss": 3.6493, + "step": 181850 + }, + { + "epoch": 2.0435561697084084, + "grad_norm": 1.2229830026626587, + "learning_rate": 1.1828740268056758e-05, + "loss": 3.7006, + "step": 181900 + }, + { + "epoch": 2.0441178948787515, + "grad_norm": 1.2174954414367676, + "learning_rate": 1.1826493354753907e-05, + "loss": 3.6276, + "step": 181950 + }, + { + "epoch": 2.0446796200490946, + "grad_norm": 1.0949270725250244, + "learning_rate": 1.1824246441451057e-05, + "loss": 3.6625, + "step": 182000 + }, + { + "epoch": 2.045241345219438, + "grad_norm": 1.0617477893829346, + "learning_rate": 1.1821999528148208e-05, + "loss": 3.6732, + "step": 182050 + }, + { + "epoch": 2.045803070389781, + "grad_norm": 1.1221001148223877, + "learning_rate": 1.1819752614845357e-05, + "loss": 3.609, + "step": 182100 + }, + { + "epoch": 2.046364795560124, + "grad_norm": 1.1541255712509155, + "learning_rate": 1.1817505701542506e-05, + "loss": 3.683, + "step": 182150 + }, + { + "epoch": 2.046926520730467, + "grad_norm": 1.1322633028030396, + "learning_rate": 1.1815258788239657e-05, + "loss": 3.6044, + "step": 182200 + }, + { + "epoch": 2.0474882459008104, + "grad_norm": 1.0582863092422485, + "learning_rate": 1.1813011874936806e-05, + "loss": 3.6516, + "step": 182250 + }, + { + "epoch": 2.048049971071154, + "grad_norm": 1.0736987590789795, + "learning_rate": 1.1810764961633956e-05, + "loss": 3.6536, + "step": 182300 + }, + { + "epoch": 2.048611696241497, + "grad_norm": 1.0526435375213623, + "learning_rate": 1.1808518048331105e-05, + "loss": 3.6763, + "step": 182350 + }, + { + "epoch": 2.0491734214118402, + "grad_norm": 1.4919220209121704, + "learning_rate": 1.1806271135028256e-05, + "loss": 3.6844, + "step": 182400 + }, + { + "epoch": 2.0497351465821834, + "grad_norm": 1.0388952493667603, + "learning_rate": 1.1804024221725405e-05, + "loss": 3.6663, + "step": 182450 + }, + { + "epoch": 2.0502968717525265, + "grad_norm": 1.1224077939987183, + "learning_rate": 1.1801777308422555e-05, + "loss": 3.6662, + "step": 182500 + }, + { + "epoch": 2.0508585969228696, + "grad_norm": 1.2052985429763794, + "learning_rate": 1.1799530395119704e-05, + "loss": 3.589, + "step": 182550 + }, + { + "epoch": 2.051420322093213, + "grad_norm": 1.1244508028030396, + "learning_rate": 1.1797283481816855e-05, + "loss": 3.7222, + "step": 182600 + }, + { + "epoch": 2.051982047263556, + "grad_norm": 1.211931824684143, + "learning_rate": 1.1795036568514004e-05, + "loss": 3.6968, + "step": 182650 + }, + { + "epoch": 2.052543772433899, + "grad_norm": 1.1259714365005493, + "learning_rate": 1.1792789655211154e-05, + "loss": 3.7227, + "step": 182700 + }, + { + "epoch": 2.053105497604242, + "grad_norm": 1.1324070692062378, + "learning_rate": 1.1790542741908305e-05, + "loss": 3.6529, + "step": 182750 + }, + { + "epoch": 2.0536672227745854, + "grad_norm": 1.025359869003296, + "learning_rate": 1.1788295828605454e-05, + "loss": 3.6157, + "step": 182800 + }, + { + "epoch": 2.0542289479449285, + "grad_norm": 1.016520380973816, + "learning_rate": 1.1786048915302603e-05, + "loss": 3.6116, + "step": 182850 + }, + { + "epoch": 2.0547906731152716, + "grad_norm": 1.1791082620620728, + "learning_rate": 1.1783802001999753e-05, + "loss": 3.6682, + "step": 182900 + }, + { + "epoch": 2.055352398285615, + "grad_norm": 1.1866360902786255, + "learning_rate": 1.1781555088696904e-05, + "loss": 3.6468, + "step": 182950 + }, + { + "epoch": 2.055914123455958, + "grad_norm": 1.1281777620315552, + "learning_rate": 1.1779308175394053e-05, + "loss": 3.6572, + "step": 183000 + }, + { + "epoch": 2.056475848626301, + "grad_norm": 1.286525845527649, + "learning_rate": 1.1777061262091202e-05, + "loss": 3.6769, + "step": 183050 + }, + { + "epoch": 2.057037573796644, + "grad_norm": 1.0334420204162598, + "learning_rate": 1.177485928705441e-05, + "loss": 3.6316, + "step": 183100 + }, + { + "epoch": 2.0575992989669873, + "grad_norm": 1.4133697748184204, + "learning_rate": 1.177261237375156e-05, + "loss": 3.6962, + "step": 183150 + }, + { + "epoch": 2.0581610241373305, + "grad_norm": 1.1942658424377441, + "learning_rate": 1.177036546044871e-05, + "loss": 3.7566, + "step": 183200 + }, + { + "epoch": 2.0587227493076736, + "grad_norm": 1.0915184020996094, + "learning_rate": 1.176811854714586e-05, + "loss": 3.5812, + "step": 183250 + }, + { + "epoch": 2.0592844744780168, + "grad_norm": 1.1224225759506226, + "learning_rate": 1.1765871633843009e-05, + "loss": 3.6077, + "step": 183300 + }, + { + "epoch": 2.05984619964836, + "grad_norm": 1.21129310131073, + "learning_rate": 1.1763624720540158e-05, + "loss": 3.6295, + "step": 183350 + }, + { + "epoch": 2.060407924818703, + "grad_norm": 1.2197438478469849, + "learning_rate": 1.176137780723731e-05, + "loss": 3.6664, + "step": 183400 + }, + { + "epoch": 2.060969649989046, + "grad_norm": 1.0733387470245361, + "learning_rate": 1.1759130893934459e-05, + "loss": 3.6526, + "step": 183450 + }, + { + "epoch": 2.0615313751593893, + "grad_norm": 1.2364966869354248, + "learning_rate": 1.1756883980631608e-05, + "loss": 3.6675, + "step": 183500 + }, + { + "epoch": 2.0620931003297325, + "grad_norm": 1.0612620115280151, + "learning_rate": 1.1754637067328757e-05, + "loss": 3.6113, + "step": 183550 + }, + { + "epoch": 2.0626548255000756, + "grad_norm": 1.0921076536178589, + "learning_rate": 1.1752390154025908e-05, + "loss": 3.6378, + "step": 183600 + }, + { + "epoch": 2.063216550670419, + "grad_norm": 1.0724772214889526, + "learning_rate": 1.1750143240723057e-05, + "loss": 3.6179, + "step": 183650 + }, + { + "epoch": 2.0637782758407623, + "grad_norm": 1.191307783126831, + "learning_rate": 1.1747896327420207e-05, + "loss": 3.6389, + "step": 183700 + }, + { + "epoch": 2.0643400010111055, + "grad_norm": 1.1037898063659668, + "learning_rate": 1.1745649414117358e-05, + "loss": 3.6549, + "step": 183750 + }, + { + "epoch": 2.0649017261814486, + "grad_norm": 1.1357640027999878, + "learning_rate": 1.1743402500814507e-05, + "loss": 3.73, + "step": 183800 + }, + { + "epoch": 2.0654634513517918, + "grad_norm": 1.1400688886642456, + "learning_rate": 1.1741155587511656e-05, + "loss": 3.6958, + "step": 183850 + }, + { + "epoch": 2.066025176522135, + "grad_norm": 1.2211834192276, + "learning_rate": 1.1738908674208806e-05, + "loss": 3.6394, + "step": 183900 + }, + { + "epoch": 2.066586901692478, + "grad_norm": 1.3305988311767578, + "learning_rate": 1.1736661760905957e-05, + "loss": 3.6558, + "step": 183950 + }, + { + "epoch": 2.067148626862821, + "grad_norm": 1.3925904035568237, + "learning_rate": 1.1734414847603106e-05, + "loss": 3.5734, + "step": 184000 + }, + { + "epoch": 2.0677103520331643, + "grad_norm": 1.1970486640930176, + "learning_rate": 1.1732167934300255e-05, + "loss": 3.6224, + "step": 184050 + }, + { + "epoch": 2.0682720772035075, + "grad_norm": 1.1737490892410278, + "learning_rate": 1.1729921020997406e-05, + "loss": 3.7166, + "step": 184100 + }, + { + "epoch": 2.0688338023738506, + "grad_norm": 1.2897924184799194, + "learning_rate": 1.1727674107694556e-05, + "loss": 3.7132, + "step": 184150 + }, + { + "epoch": 2.0693955275441938, + "grad_norm": 1.4397519826889038, + "learning_rate": 1.1725427194391705e-05, + "loss": 3.6388, + "step": 184200 + }, + { + "epoch": 2.069957252714537, + "grad_norm": 1.2425822019577026, + "learning_rate": 1.1723180281088854e-05, + "loss": 3.6275, + "step": 184250 + }, + { + "epoch": 2.07051897788488, + "grad_norm": 1.143621802330017, + "learning_rate": 1.1720933367786005e-05, + "loss": 3.6994, + "step": 184300 + }, + { + "epoch": 2.071080703055223, + "grad_norm": 1.2500793933868408, + "learning_rate": 1.1718686454483155e-05, + "loss": 3.6829, + "step": 184350 + }, + { + "epoch": 2.0716424282255663, + "grad_norm": 1.2162973880767822, + "learning_rate": 1.1716439541180304e-05, + "loss": 3.6092, + "step": 184400 + }, + { + "epoch": 2.0722041533959095, + "grad_norm": 1.1496464014053345, + "learning_rate": 1.1714192627877455e-05, + "loss": 3.6196, + "step": 184450 + }, + { + "epoch": 2.0727658785662526, + "grad_norm": 1.0511910915374756, + "learning_rate": 1.1711945714574604e-05, + "loss": 3.6539, + "step": 184500 + }, + { + "epoch": 2.0733276037365957, + "grad_norm": 1.103660225868225, + "learning_rate": 1.1709698801271754e-05, + "loss": 3.6648, + "step": 184550 + }, + { + "epoch": 2.073889328906939, + "grad_norm": 1.0108928680419922, + "learning_rate": 1.1707451887968903e-05, + "loss": 3.7308, + "step": 184600 + }, + { + "epoch": 2.074451054077282, + "grad_norm": 1.1093605756759644, + "learning_rate": 1.1705204974666054e-05, + "loss": 3.6693, + "step": 184650 + }, + { + "epoch": 2.075012779247625, + "grad_norm": 1.1762808561325073, + "learning_rate": 1.1702958061363203e-05, + "loss": 3.6235, + "step": 184700 + }, + { + "epoch": 2.0755745044179683, + "grad_norm": 1.3621156215667725, + "learning_rate": 1.1700711148060353e-05, + "loss": 3.5564, + "step": 184750 + }, + { + "epoch": 2.0761362295883115, + "grad_norm": 1.2420388460159302, + "learning_rate": 1.1698464234757504e-05, + "loss": 3.6166, + "step": 184800 + }, + { + "epoch": 2.0766979547586546, + "grad_norm": 1.2055463790893555, + "learning_rate": 1.1696217321454653e-05, + "loss": 3.7241, + "step": 184850 + }, + { + "epoch": 2.0772596799289977, + "grad_norm": 1.192975401878357, + "learning_rate": 1.1693970408151802e-05, + "loss": 3.7106, + "step": 184900 + }, + { + "epoch": 2.077821405099341, + "grad_norm": 1.2383537292480469, + "learning_rate": 1.1691723494848952e-05, + "loss": 3.6599, + "step": 184950 + }, + { + "epoch": 2.0783831302696845, + "grad_norm": 1.301489233970642, + "learning_rate": 1.1689476581546103e-05, + "loss": 3.6764, + "step": 185000 + }, + { + "epoch": 2.0789448554400276, + "grad_norm": 1.3114230632781982, + "learning_rate": 1.1687229668243252e-05, + "loss": 3.6454, + "step": 185050 + }, + { + "epoch": 2.0795065806103707, + "grad_norm": 1.1451566219329834, + "learning_rate": 1.1684982754940401e-05, + "loss": 3.6942, + "step": 185100 + }, + { + "epoch": 2.080068305780714, + "grad_norm": 1.0022562742233276, + "learning_rate": 1.168273584163755e-05, + "loss": 3.6398, + "step": 185150 + }, + { + "epoch": 2.080630030951057, + "grad_norm": 1.2197297811508179, + "learning_rate": 1.1680488928334702e-05, + "loss": 3.6824, + "step": 185200 + }, + { + "epoch": 2.0811917561214, + "grad_norm": 1.25179123878479, + "learning_rate": 1.1678242015031851e-05, + "loss": 3.6963, + "step": 185250 + }, + { + "epoch": 2.0817534812917433, + "grad_norm": 1.1346317529678345, + "learning_rate": 1.1675995101729e-05, + "loss": 3.731, + "step": 185300 + }, + { + "epoch": 2.0823152064620865, + "grad_norm": 1.211344838142395, + "learning_rate": 1.1673748188426151e-05, + "loss": 3.6652, + "step": 185350 + }, + { + "epoch": 2.0828769316324296, + "grad_norm": 1.2521421909332275, + "learning_rate": 1.16715012751233e-05, + "loss": 3.7182, + "step": 185400 + }, + { + "epoch": 2.0834386568027727, + "grad_norm": 1.1842113733291626, + "learning_rate": 1.166925436182045e-05, + "loss": 3.6045, + "step": 185450 + }, + { + "epoch": 2.084000381973116, + "grad_norm": 1.0520801544189453, + "learning_rate": 1.1667007448517599e-05, + "loss": 3.6785, + "step": 185500 + }, + { + "epoch": 2.084562107143459, + "grad_norm": 1.1933637857437134, + "learning_rate": 1.166476053521475e-05, + "loss": 3.5994, + "step": 185550 + }, + { + "epoch": 2.085123832313802, + "grad_norm": 1.1843327283859253, + "learning_rate": 1.16625136219119e-05, + "loss": 3.688, + "step": 185600 + }, + { + "epoch": 2.0856855574841453, + "grad_norm": 1.059524655342102, + "learning_rate": 1.1660266708609049e-05, + "loss": 3.6427, + "step": 185650 + }, + { + "epoch": 2.0862472826544884, + "grad_norm": 1.3903207778930664, + "learning_rate": 1.16580197953062e-05, + "loss": 3.6132, + "step": 185700 + }, + { + "epoch": 2.0868090078248316, + "grad_norm": 1.3105016946792603, + "learning_rate": 1.1655772882003349e-05, + "loss": 3.6907, + "step": 185750 + }, + { + "epoch": 2.0873707329951747, + "grad_norm": 1.014356255531311, + "learning_rate": 1.1653525968700498e-05, + "loss": 3.6845, + "step": 185800 + }, + { + "epoch": 2.087932458165518, + "grad_norm": 1.4541751146316528, + "learning_rate": 1.1651279055397648e-05, + "loss": 3.6286, + "step": 185850 + }, + { + "epoch": 2.088494183335861, + "grad_norm": 1.0511747598648071, + "learning_rate": 1.1649032142094799e-05, + "loss": 3.7113, + "step": 185900 + }, + { + "epoch": 2.089055908506204, + "grad_norm": 1.2117619514465332, + "learning_rate": 1.1646785228791948e-05, + "loss": 3.6283, + "step": 185950 + }, + { + "epoch": 2.0896176336765473, + "grad_norm": 1.050269365310669, + "learning_rate": 1.1644538315489097e-05, + "loss": 3.6544, + "step": 186000 + }, + { + "epoch": 2.0901793588468904, + "grad_norm": 1.167660117149353, + "learning_rate": 1.1642291402186248e-05, + "loss": 3.6782, + "step": 186050 + }, + { + "epoch": 2.0907410840172336, + "grad_norm": 1.171392560005188, + "learning_rate": 1.1640044488883398e-05, + "loss": 3.6661, + "step": 186100 + }, + { + "epoch": 2.0913028091875767, + "grad_norm": 1.2858030796051025, + "learning_rate": 1.1637797575580547e-05, + "loss": 3.6055, + "step": 186150 + }, + { + "epoch": 2.09186453435792, + "grad_norm": 1.2579023838043213, + "learning_rate": 1.1635550662277696e-05, + "loss": 3.6953, + "step": 186200 + }, + { + "epoch": 2.092426259528263, + "grad_norm": 1.1274867057800293, + "learning_rate": 1.1633303748974847e-05, + "loss": 3.5906, + "step": 186250 + }, + { + "epoch": 2.092987984698606, + "grad_norm": 1.3431787490844727, + "learning_rate": 1.1631056835671997e-05, + "loss": 3.6835, + "step": 186300 + }, + { + "epoch": 2.0935497098689497, + "grad_norm": 1.0384411811828613, + "learning_rate": 1.1628809922369146e-05, + "loss": 3.6966, + "step": 186350 + }, + { + "epoch": 2.094111435039293, + "grad_norm": 1.216713547706604, + "learning_rate": 1.1626563009066297e-05, + "loss": 3.6826, + "step": 186400 + }, + { + "epoch": 2.094673160209636, + "grad_norm": 1.3694875240325928, + "learning_rate": 1.1624316095763446e-05, + "loss": 3.6066, + "step": 186450 + }, + { + "epoch": 2.095234885379979, + "grad_norm": 1.1893595457077026, + "learning_rate": 1.1622069182460596e-05, + "loss": 3.6106, + "step": 186500 + }, + { + "epoch": 2.0957966105503223, + "grad_norm": 1.1873623132705688, + "learning_rate": 1.1619822269157745e-05, + "loss": 3.7164, + "step": 186550 + }, + { + "epoch": 2.0963583357206654, + "grad_norm": 1.196428656578064, + "learning_rate": 1.1617575355854896e-05, + "loss": 3.5973, + "step": 186600 + }, + { + "epoch": 2.0969200608910086, + "grad_norm": 1.1118874549865723, + "learning_rate": 1.1615328442552045e-05, + "loss": 3.6804, + "step": 186650 + }, + { + "epoch": 2.0974817860613517, + "grad_norm": 1.3892502784729004, + "learning_rate": 1.1613081529249195e-05, + "loss": 3.7129, + "step": 186700 + }, + { + "epoch": 2.098043511231695, + "grad_norm": 1.0746665000915527, + "learning_rate": 1.1610879554212402e-05, + "loss": 3.6837, + "step": 186750 + }, + { + "epoch": 2.098605236402038, + "grad_norm": 1.3109077215194702, + "learning_rate": 1.1608632640909552e-05, + "loss": 3.6782, + "step": 186800 + }, + { + "epoch": 2.099166961572381, + "grad_norm": 1.3797475099563599, + "learning_rate": 1.1606385727606701e-05, + "loss": 3.7002, + "step": 186850 + }, + { + "epoch": 2.0997286867427243, + "grad_norm": 1.1863510608673096, + "learning_rate": 1.1604138814303852e-05, + "loss": 3.6765, + "step": 186900 + }, + { + "epoch": 2.1002904119130674, + "grad_norm": 1.1817659139633179, + "learning_rate": 1.1601891901001001e-05, + "loss": 3.698, + "step": 186950 + }, + { + "epoch": 2.1008521370834106, + "grad_norm": 1.0148324966430664, + "learning_rate": 1.159964498769815e-05, + "loss": 3.6419, + "step": 187000 + }, + { + "epoch": 2.1014138622537537, + "grad_norm": 1.1663976907730103, + "learning_rate": 1.1597398074395302e-05, + "loss": 3.6567, + "step": 187050 + }, + { + "epoch": 2.101975587424097, + "grad_norm": 1.1736410856246948, + "learning_rate": 1.159515116109245e-05, + "loss": 3.6098, + "step": 187100 + }, + { + "epoch": 2.10253731259444, + "grad_norm": 1.2112184762954712, + "learning_rate": 1.15929042477896e-05, + "loss": 3.5847, + "step": 187150 + }, + { + "epoch": 2.103099037764783, + "grad_norm": 1.570095419883728, + "learning_rate": 1.159065733448675e-05, + "loss": 3.6716, + "step": 187200 + }, + { + "epoch": 2.1036607629351263, + "grad_norm": 1.0069478750228882, + "learning_rate": 1.15884104211839e-05, + "loss": 3.6609, + "step": 187250 + }, + { + "epoch": 2.1042224881054694, + "grad_norm": 0.9928694367408752, + "learning_rate": 1.158616350788105e-05, + "loss": 3.6824, + "step": 187300 + }, + { + "epoch": 2.1047842132758126, + "grad_norm": 1.1236019134521484, + "learning_rate": 1.1583916594578199e-05, + "loss": 3.6956, + "step": 187350 + }, + { + "epoch": 2.1053459384461557, + "grad_norm": 1.145443081855774, + "learning_rate": 1.158166968127535e-05, + "loss": 3.5721, + "step": 187400 + }, + { + "epoch": 2.105907663616499, + "grad_norm": 1.148145318031311, + "learning_rate": 1.15794227679725e-05, + "loss": 3.5823, + "step": 187450 + }, + { + "epoch": 2.106469388786842, + "grad_norm": 1.2535157203674316, + "learning_rate": 1.1577175854669649e-05, + "loss": 3.6493, + "step": 187500 + }, + { + "epoch": 2.107031113957185, + "grad_norm": 2.8420257568359375, + "learning_rate": 1.1574928941366798e-05, + "loss": 3.6648, + "step": 187550 + }, + { + "epoch": 2.1075928391275283, + "grad_norm": 1.2739931344985962, + "learning_rate": 1.1572682028063949e-05, + "loss": 3.5955, + "step": 187600 + }, + { + "epoch": 2.1081545642978714, + "grad_norm": 1.2763255834579468, + "learning_rate": 1.1570435114761098e-05, + "loss": 3.6826, + "step": 187650 + }, + { + "epoch": 2.108716289468215, + "grad_norm": 1.135329246520996, + "learning_rate": 1.1568188201458248e-05, + "loss": 3.5985, + "step": 187700 + }, + { + "epoch": 2.109278014638558, + "grad_norm": 1.0852805376052856, + "learning_rate": 1.1565941288155397e-05, + "loss": 3.7168, + "step": 187750 + }, + { + "epoch": 2.1098397398089013, + "grad_norm": 1.18838632106781, + "learning_rate": 1.1563694374852548e-05, + "loss": 3.5967, + "step": 187800 + }, + { + "epoch": 2.1104014649792444, + "grad_norm": 1.0903185606002808, + "learning_rate": 1.1561447461549697e-05, + "loss": 3.6357, + "step": 187850 + }, + { + "epoch": 2.1109631901495876, + "grad_norm": 1.1050207614898682, + "learning_rate": 1.1559200548246847e-05, + "loss": 3.6751, + "step": 187900 + }, + { + "epoch": 2.1115249153199307, + "grad_norm": 1.1145532131195068, + "learning_rate": 1.1556953634943998e-05, + "loss": 3.6524, + "step": 187950 + }, + { + "epoch": 2.112086640490274, + "grad_norm": 1.222307562828064, + "learning_rate": 1.1554706721641147e-05, + "loss": 3.667, + "step": 188000 + }, + { + "epoch": 2.112648365660617, + "grad_norm": 1.4427876472473145, + "learning_rate": 1.1552459808338296e-05, + "loss": 3.6596, + "step": 188050 + }, + { + "epoch": 2.11321009083096, + "grad_norm": 1.021989107131958, + "learning_rate": 1.1550212895035446e-05, + "loss": 3.6318, + "step": 188100 + }, + { + "epoch": 2.1137718160013033, + "grad_norm": 1.2242029905319214, + "learning_rate": 1.1547965981732597e-05, + "loss": 3.594, + "step": 188150 + }, + { + "epoch": 2.1143335411716464, + "grad_norm": 1.1787387132644653, + "learning_rate": 1.1545719068429746e-05, + "loss": 3.6721, + "step": 188200 + }, + { + "epoch": 2.1148952663419895, + "grad_norm": 1.0349905490875244, + "learning_rate": 1.1543472155126895e-05, + "loss": 3.7216, + "step": 188250 + }, + { + "epoch": 2.1154569915123327, + "grad_norm": 1.1641631126403809, + "learning_rate": 1.1541225241824046e-05, + "loss": 3.6352, + "step": 188300 + }, + { + "epoch": 2.116018716682676, + "grad_norm": 1.2305693626403809, + "learning_rate": 1.1538978328521196e-05, + "loss": 3.6903, + "step": 188350 + }, + { + "epoch": 2.116580441853019, + "grad_norm": 1.3482508659362793, + "learning_rate": 1.1536731415218345e-05, + "loss": 3.6506, + "step": 188400 + }, + { + "epoch": 2.117142167023362, + "grad_norm": 1.0359673500061035, + "learning_rate": 1.1534484501915494e-05, + "loss": 3.6182, + "step": 188450 + }, + { + "epoch": 2.1177038921937053, + "grad_norm": 1.2034860849380493, + "learning_rate": 1.1532237588612645e-05, + "loss": 3.6963, + "step": 188500 + }, + { + "epoch": 2.1182656173640484, + "grad_norm": 1.143668293952942, + "learning_rate": 1.1529990675309795e-05, + "loss": 3.7738, + "step": 188550 + }, + { + "epoch": 2.1188273425343915, + "grad_norm": 1.3331931829452515, + "learning_rate": 1.1527743762006944e-05, + "loss": 3.5772, + "step": 188600 + }, + { + "epoch": 2.1193890677047347, + "grad_norm": 1.3241474628448486, + "learning_rate": 1.1525496848704095e-05, + "loss": 3.653, + "step": 188650 + }, + { + "epoch": 2.119950792875078, + "grad_norm": 1.1467458009719849, + "learning_rate": 1.1523249935401244e-05, + "loss": 3.7084, + "step": 188700 + }, + { + "epoch": 2.120512518045421, + "grad_norm": 1.1851401329040527, + "learning_rate": 1.1521003022098394e-05, + "loss": 3.6817, + "step": 188750 + }, + { + "epoch": 2.121074243215764, + "grad_norm": 1.1613062620162964, + "learning_rate": 1.1518756108795543e-05, + "loss": 3.6369, + "step": 188800 + }, + { + "epoch": 2.1216359683861072, + "grad_norm": 1.0273969173431396, + "learning_rate": 1.1516509195492694e-05, + "loss": 3.7027, + "step": 188850 + }, + { + "epoch": 2.1221976935564504, + "grad_norm": 1.080709457397461, + "learning_rate": 1.1514262282189843e-05, + "loss": 3.6969, + "step": 188900 + }, + { + "epoch": 2.1227594187267935, + "grad_norm": 1.1772719621658325, + "learning_rate": 1.1512015368886992e-05, + "loss": 3.6234, + "step": 188950 + }, + { + "epoch": 2.1233211438971367, + "grad_norm": 1.0060964822769165, + "learning_rate": 1.1509768455584143e-05, + "loss": 3.6852, + "step": 189000 + }, + { + "epoch": 2.1238828690674803, + "grad_norm": 1.2354568243026733, + "learning_rate": 1.1507521542281293e-05, + "loss": 3.6297, + "step": 189050 + }, + { + "epoch": 2.1244445942378234, + "grad_norm": 1.106758952140808, + "learning_rate": 1.1505274628978442e-05, + "loss": 3.6663, + "step": 189100 + }, + { + "epoch": 2.1250063194081665, + "grad_norm": 1.177374243736267, + "learning_rate": 1.1503027715675591e-05, + "loss": 3.7331, + "step": 189150 + }, + { + "epoch": 2.1255680445785097, + "grad_norm": 1.298965334892273, + "learning_rate": 1.1500780802372742e-05, + "loss": 3.5916, + "step": 189200 + }, + { + "epoch": 2.126129769748853, + "grad_norm": 1.2143000364303589, + "learning_rate": 1.1498533889069892e-05, + "loss": 3.7088, + "step": 189250 + }, + { + "epoch": 2.126691494919196, + "grad_norm": 1.0102218389511108, + "learning_rate": 1.1496286975767041e-05, + "loss": 3.6115, + "step": 189300 + }, + { + "epoch": 2.127253220089539, + "grad_norm": 1.1801633834838867, + "learning_rate": 1.149404006246419e-05, + "loss": 3.6492, + "step": 189350 + }, + { + "epoch": 2.1278149452598822, + "grad_norm": 1.0461994409561157, + "learning_rate": 1.1491793149161341e-05, + "loss": 3.6076, + "step": 189400 + }, + { + "epoch": 2.1283766704302254, + "grad_norm": 1.0831447839736938, + "learning_rate": 1.148954623585849e-05, + "loss": 3.6448, + "step": 189450 + }, + { + "epoch": 2.1289383956005685, + "grad_norm": 1.1427998542785645, + "learning_rate": 1.148729932255564e-05, + "loss": 3.5827, + "step": 189500 + }, + { + "epoch": 2.1295001207709117, + "grad_norm": 1.3421205282211304, + "learning_rate": 1.1485052409252791e-05, + "loss": 3.6507, + "step": 189550 + }, + { + "epoch": 2.130061845941255, + "grad_norm": 1.284712553024292, + "learning_rate": 1.148280549594994e-05, + "loss": 3.6311, + "step": 189600 + }, + { + "epoch": 2.130623571111598, + "grad_norm": 1.7064614295959473, + "learning_rate": 1.148055858264709e-05, + "loss": 3.6555, + "step": 189650 + }, + { + "epoch": 2.131185296281941, + "grad_norm": 1.0825332403182983, + "learning_rate": 1.1478311669344239e-05, + "loss": 3.6579, + "step": 189700 + }, + { + "epoch": 2.1317470214522842, + "grad_norm": 1.303828477859497, + "learning_rate": 1.1476109694307447e-05, + "loss": 3.6228, + "step": 189750 + }, + { + "epoch": 2.1323087466226274, + "grad_norm": 1.2739299535751343, + "learning_rate": 1.1473862781004596e-05, + "loss": 3.7096, + "step": 189800 + }, + { + "epoch": 2.1328704717929705, + "grad_norm": 1.215766191482544, + "learning_rate": 1.1471615867701747e-05, + "loss": 3.7074, + "step": 189850 + }, + { + "epoch": 2.1334321969633137, + "grad_norm": 1.0622475147247314, + "learning_rate": 1.1469368954398896e-05, + "loss": 3.7047, + "step": 189900 + }, + { + "epoch": 2.133993922133657, + "grad_norm": 1.1313506364822388, + "learning_rate": 1.1467122041096046e-05, + "loss": 3.7185, + "step": 189950 + }, + { + "epoch": 2.134555647304, + "grad_norm": 1.1207643747329712, + "learning_rate": 1.1464875127793197e-05, + "loss": 3.6292, + "step": 190000 + }, + { + "epoch": 2.135117372474343, + "grad_norm": 1.1433231830596924, + "learning_rate": 1.1462628214490346e-05, + "loss": 3.6774, + "step": 190050 + }, + { + "epoch": 2.135679097644686, + "grad_norm": 1.3061046600341797, + "learning_rate": 1.1460381301187495e-05, + "loss": 3.6434, + "step": 190100 + }, + { + "epoch": 2.1362408228150294, + "grad_norm": 1.1008479595184326, + "learning_rate": 1.1458134387884645e-05, + "loss": 3.6726, + "step": 190150 + }, + { + "epoch": 2.1368025479853725, + "grad_norm": 1.1142629384994507, + "learning_rate": 1.1455887474581796e-05, + "loss": 3.6476, + "step": 190200 + }, + { + "epoch": 2.1373642731557156, + "grad_norm": 1.2122528553009033, + "learning_rate": 1.1453640561278945e-05, + "loss": 3.7287, + "step": 190250 + }, + { + "epoch": 2.137925998326059, + "grad_norm": 1.0298256874084473, + "learning_rate": 1.1451393647976094e-05, + "loss": 3.6936, + "step": 190300 + }, + { + "epoch": 2.138487723496402, + "grad_norm": 1.2197506427764893, + "learning_rate": 1.1449146734673244e-05, + "loss": 3.698, + "step": 190350 + }, + { + "epoch": 2.1390494486667455, + "grad_norm": 1.2474559545516968, + "learning_rate": 1.1446899821370395e-05, + "loss": 3.6674, + "step": 190400 + }, + { + "epoch": 2.1396111738370887, + "grad_norm": 1.1839038133621216, + "learning_rate": 1.1444652908067544e-05, + "loss": 3.65, + "step": 190450 + }, + { + "epoch": 2.140172899007432, + "grad_norm": 1.0742826461791992, + "learning_rate": 1.1442405994764693e-05, + "loss": 3.7048, + "step": 190500 + }, + { + "epoch": 2.140734624177775, + "grad_norm": 1.9982452392578125, + "learning_rate": 1.1440159081461844e-05, + "loss": 3.6968, + "step": 190550 + }, + { + "epoch": 2.141296349348118, + "grad_norm": 1.16074800491333, + "learning_rate": 1.1437912168158993e-05, + "loss": 3.5948, + "step": 190600 + }, + { + "epoch": 2.141858074518461, + "grad_norm": 1.2510814666748047, + "learning_rate": 1.1435665254856143e-05, + "loss": 3.6732, + "step": 190650 + }, + { + "epoch": 2.1424197996888044, + "grad_norm": 1.1811882257461548, + "learning_rate": 1.1433418341553292e-05, + "loss": 3.7217, + "step": 190700 + }, + { + "epoch": 2.1429815248591475, + "grad_norm": 1.0726606845855713, + "learning_rate": 1.1431171428250443e-05, + "loss": 3.6865, + "step": 190750 + }, + { + "epoch": 2.1435432500294906, + "grad_norm": 1.2404667139053345, + "learning_rate": 1.1428924514947592e-05, + "loss": 3.639, + "step": 190800 + }, + { + "epoch": 2.144104975199834, + "grad_norm": 1.1091114282608032, + "learning_rate": 1.1426677601644742e-05, + "loss": 3.6803, + "step": 190850 + }, + { + "epoch": 2.144666700370177, + "grad_norm": 1.1242750883102417, + "learning_rate": 1.1424430688341893e-05, + "loss": 3.6574, + "step": 190900 + }, + { + "epoch": 2.14522842554052, + "grad_norm": 0.9307460784912109, + "learning_rate": 1.1422183775039042e-05, + "loss": 3.7273, + "step": 190950 + }, + { + "epoch": 2.145790150710863, + "grad_norm": 1.123029112815857, + "learning_rate": 1.1419936861736191e-05, + "loss": 3.6071, + "step": 191000 + }, + { + "epoch": 2.1463518758812064, + "grad_norm": 1.162349820137024, + "learning_rate": 1.141768994843334e-05, + "loss": 3.687, + "step": 191050 + }, + { + "epoch": 2.1469136010515495, + "grad_norm": 1.164348840713501, + "learning_rate": 1.1415443035130492e-05, + "loss": 3.6717, + "step": 191100 + }, + { + "epoch": 2.1474753262218926, + "grad_norm": 1.2084416151046753, + "learning_rate": 1.1413196121827641e-05, + "loss": 3.6662, + "step": 191150 + }, + { + "epoch": 2.1480370513922358, + "grad_norm": 1.1981921195983887, + "learning_rate": 1.141094920852479e-05, + "loss": 3.6786, + "step": 191200 + }, + { + "epoch": 2.148598776562579, + "grad_norm": 1.047624111175537, + "learning_rate": 1.1408702295221941e-05, + "loss": 3.7191, + "step": 191250 + }, + { + "epoch": 2.149160501732922, + "grad_norm": 1.154442548751831, + "learning_rate": 1.140645538191909e-05, + "loss": 3.6479, + "step": 191300 + }, + { + "epoch": 2.149722226903265, + "grad_norm": 1.0728384256362915, + "learning_rate": 1.140420846861624e-05, + "loss": 3.6669, + "step": 191350 + }, + { + "epoch": 2.1502839520736083, + "grad_norm": 1.2294445037841797, + "learning_rate": 1.140196155531339e-05, + "loss": 3.6674, + "step": 191400 + }, + { + "epoch": 2.1508456772439515, + "grad_norm": 1.1478815078735352, + "learning_rate": 1.139971464201054e-05, + "loss": 3.5864, + "step": 191450 + }, + { + "epoch": 2.1514074024142946, + "grad_norm": 1.1214373111724854, + "learning_rate": 1.139746772870769e-05, + "loss": 3.7497, + "step": 191500 + }, + { + "epoch": 2.1519691275846378, + "grad_norm": 1.183599829673767, + "learning_rate": 1.1395220815404839e-05, + "loss": 3.6244, + "step": 191550 + }, + { + "epoch": 2.152530852754981, + "grad_norm": 1.1228039264678955, + "learning_rate": 1.139297390210199e-05, + "loss": 3.617, + "step": 191600 + }, + { + "epoch": 2.1530925779253245, + "grad_norm": 1.067888617515564, + "learning_rate": 1.139072698879914e-05, + "loss": 3.6954, + "step": 191650 + }, + { + "epoch": 2.153654303095667, + "grad_norm": 1.085471510887146, + "learning_rate": 1.1388480075496289e-05, + "loss": 3.7293, + "step": 191700 + }, + { + "epoch": 2.1542160282660108, + "grad_norm": 1.1760141849517822, + "learning_rate": 1.1386233162193438e-05, + "loss": 3.6734, + "step": 191750 + }, + { + "epoch": 2.154777753436354, + "grad_norm": 1.1908788681030273, + "learning_rate": 1.1383986248890589e-05, + "loss": 3.6836, + "step": 191800 + }, + { + "epoch": 2.155339478606697, + "grad_norm": 1.1994457244873047, + "learning_rate": 1.1381739335587738e-05, + "loss": 3.5693, + "step": 191850 + }, + { + "epoch": 2.15590120377704, + "grad_norm": 1.2013858556747437, + "learning_rate": 1.1379492422284888e-05, + "loss": 3.6542, + "step": 191900 + }, + { + "epoch": 2.1564629289473833, + "grad_norm": 1.1068100929260254, + "learning_rate": 1.1377245508982037e-05, + "loss": 3.633, + "step": 191950 + }, + { + "epoch": 2.1570246541177265, + "grad_norm": 1.1132248640060425, + "learning_rate": 1.1374998595679188e-05, + "loss": 3.6764, + "step": 192000 + }, + { + "epoch": 2.1575863792880696, + "grad_norm": 1.1618642807006836, + "learning_rate": 1.1372751682376337e-05, + "loss": 3.6839, + "step": 192050 + }, + { + "epoch": 2.1581481044584128, + "grad_norm": 1.1055786609649658, + "learning_rate": 1.1370504769073487e-05, + "loss": 3.6788, + "step": 192100 + }, + { + "epoch": 2.158709829628756, + "grad_norm": 1.046073317527771, + "learning_rate": 1.1368257855770638e-05, + "loss": 3.6824, + "step": 192150 + }, + { + "epoch": 2.159271554799099, + "grad_norm": 1.0207507610321045, + "learning_rate": 1.1366010942467787e-05, + "loss": 3.6136, + "step": 192200 + }, + { + "epoch": 2.159833279969442, + "grad_norm": 1.2055473327636719, + "learning_rate": 1.1363764029164936e-05, + "loss": 3.6755, + "step": 192250 + }, + { + "epoch": 2.1603950051397853, + "grad_norm": 1.1619362831115723, + "learning_rate": 1.1361517115862085e-05, + "loss": 3.6353, + "step": 192300 + }, + { + "epoch": 2.1609567303101285, + "grad_norm": 1.0099173784255981, + "learning_rate": 1.1359270202559236e-05, + "loss": 3.6397, + "step": 192350 + }, + { + "epoch": 2.1615184554804716, + "grad_norm": 1.1548542976379395, + "learning_rate": 1.1357023289256386e-05, + "loss": 3.7073, + "step": 192400 + }, + { + "epoch": 2.1620801806508148, + "grad_norm": 1.1104629039764404, + "learning_rate": 1.1354776375953535e-05, + "loss": 3.7453, + "step": 192450 + }, + { + "epoch": 2.162641905821158, + "grad_norm": 3.932023525238037, + "learning_rate": 1.1352529462650686e-05, + "loss": 3.6289, + "step": 192500 + }, + { + "epoch": 2.163203630991501, + "grad_norm": 1.2070088386535645, + "learning_rate": 1.1350282549347835e-05, + "loss": 3.701, + "step": 192550 + }, + { + "epoch": 2.163765356161844, + "grad_norm": 1.2627888917922974, + "learning_rate": 1.1348035636044985e-05, + "loss": 3.6332, + "step": 192600 + }, + { + "epoch": 2.1643270813321873, + "grad_norm": 1.2985186576843262, + "learning_rate": 1.1345788722742134e-05, + "loss": 3.5978, + "step": 192650 + }, + { + "epoch": 2.1648888065025305, + "grad_norm": 1.1906691789627075, + "learning_rate": 1.1343541809439285e-05, + "loss": 3.6652, + "step": 192700 + }, + { + "epoch": 2.1654505316728736, + "grad_norm": 1.1376667022705078, + "learning_rate": 1.1341294896136434e-05, + "loss": 3.6271, + "step": 192750 + }, + { + "epoch": 2.1660122568432167, + "grad_norm": 1.0888322591781616, + "learning_rate": 1.1339047982833584e-05, + "loss": 3.5866, + "step": 192800 + }, + { + "epoch": 2.16657398201356, + "grad_norm": 1.0450643301010132, + "learning_rate": 1.1336801069530735e-05, + "loss": 3.6099, + "step": 192850 + }, + { + "epoch": 2.167135707183903, + "grad_norm": 1.1386446952819824, + "learning_rate": 1.1334554156227884e-05, + "loss": 3.5828, + "step": 192900 + }, + { + "epoch": 2.167697432354246, + "grad_norm": 1.1101601123809814, + "learning_rate": 1.1332307242925033e-05, + "loss": 3.6742, + "step": 192950 + }, + { + "epoch": 2.1682591575245898, + "grad_norm": 1.1992783546447754, + "learning_rate": 1.1330060329622183e-05, + "loss": 3.6665, + "step": 193000 + }, + { + "epoch": 2.1688208826949325, + "grad_norm": 1.1518826484680176, + "learning_rate": 1.1327813416319334e-05, + "loss": 3.6617, + "step": 193050 + }, + { + "epoch": 2.169382607865276, + "grad_norm": 1.4648700952529907, + "learning_rate": 1.1325566503016483e-05, + "loss": 3.6676, + "step": 193100 + }, + { + "epoch": 2.169944333035619, + "grad_norm": 1.2937272787094116, + "learning_rate": 1.1323319589713632e-05, + "loss": 3.7151, + "step": 193150 + }, + { + "epoch": 2.1705060582059623, + "grad_norm": 1.221721887588501, + "learning_rate": 1.1321072676410783e-05, + "loss": 3.6795, + "step": 193200 + }, + { + "epoch": 2.1710677833763055, + "grad_norm": 1.5011725425720215, + "learning_rate": 1.1318825763107933e-05, + "loss": 3.6012, + "step": 193250 + }, + { + "epoch": 2.1716295085466486, + "grad_norm": 1.0371367931365967, + "learning_rate": 1.1316578849805082e-05, + "loss": 3.635, + "step": 193300 + }, + { + "epoch": 2.1721912337169917, + "grad_norm": 1.062644124031067, + "learning_rate": 1.1314331936502231e-05, + "loss": 3.7003, + "step": 193350 + }, + { + "epoch": 2.172752958887335, + "grad_norm": 1.191779613494873, + "learning_rate": 1.1312085023199382e-05, + "loss": 3.6577, + "step": 193400 + }, + { + "epoch": 2.173314684057678, + "grad_norm": 1.106966257095337, + "learning_rate": 1.1309838109896532e-05, + "loss": 3.702, + "step": 193450 + }, + { + "epoch": 2.173876409228021, + "grad_norm": 1.474577784538269, + "learning_rate": 1.1307591196593681e-05, + "loss": 3.7128, + "step": 193500 + }, + { + "epoch": 2.1744381343983643, + "grad_norm": 0.9508491158485413, + "learning_rate": 1.1305344283290832e-05, + "loss": 3.6307, + "step": 193550 + }, + { + "epoch": 2.1749998595687074, + "grad_norm": 1.0904008150100708, + "learning_rate": 1.1303097369987981e-05, + "loss": 3.6612, + "step": 193600 + }, + { + "epoch": 2.1755615847390506, + "grad_norm": 1.1697651147842407, + "learning_rate": 1.130085045668513e-05, + "loss": 3.7044, + "step": 193650 + }, + { + "epoch": 2.1761233099093937, + "grad_norm": 1.2298531532287598, + "learning_rate": 1.129860354338228e-05, + "loss": 3.6504, + "step": 193700 + }, + { + "epoch": 2.176685035079737, + "grad_norm": 1.0923755168914795, + "learning_rate": 1.1296356630079431e-05, + "loss": 3.6549, + "step": 193750 + }, + { + "epoch": 2.17724676025008, + "grad_norm": 1.1315332651138306, + "learning_rate": 1.1294154655042635e-05, + "loss": 3.7113, + "step": 193800 + }, + { + "epoch": 2.177808485420423, + "grad_norm": 0.9924259781837463, + "learning_rate": 1.1291907741739784e-05, + "loss": 3.6409, + "step": 193850 + }, + { + "epoch": 2.1783702105907663, + "grad_norm": 1.2110371589660645, + "learning_rate": 1.1289660828436935e-05, + "loss": 3.6688, + "step": 193900 + }, + { + "epoch": 2.1789319357611094, + "grad_norm": 1.2357207536697388, + "learning_rate": 1.1287413915134085e-05, + "loss": 3.6865, + "step": 193950 + }, + { + "epoch": 2.1794936609314526, + "grad_norm": 1.288679838180542, + "learning_rate": 1.1285167001831234e-05, + "loss": 3.6362, + "step": 194000 + }, + { + "epoch": 2.1800553861017957, + "grad_norm": 1.194150686264038, + "learning_rate": 1.1282920088528383e-05, + "loss": 3.7321, + "step": 194050 + }, + { + "epoch": 2.180617111272139, + "grad_norm": 1.1033546924591064, + "learning_rate": 1.1280673175225534e-05, + "loss": 3.6577, + "step": 194100 + }, + { + "epoch": 2.181178836442482, + "grad_norm": 1.285281777381897, + "learning_rate": 1.1278426261922684e-05, + "loss": 3.525, + "step": 194150 + }, + { + "epoch": 2.181740561612825, + "grad_norm": 1.0182749032974243, + "learning_rate": 1.1276179348619833e-05, + "loss": 3.7052, + "step": 194200 + }, + { + "epoch": 2.1823022867831683, + "grad_norm": 1.1671384572982788, + "learning_rate": 1.1273932435316984e-05, + "loss": 3.6362, + "step": 194250 + }, + { + "epoch": 2.1828640119535114, + "grad_norm": 1.1830589771270752, + "learning_rate": 1.1271685522014133e-05, + "loss": 3.6723, + "step": 194300 + }, + { + "epoch": 2.183425737123855, + "grad_norm": 1.2304086685180664, + "learning_rate": 1.1269438608711283e-05, + "loss": 3.6114, + "step": 194350 + }, + { + "epoch": 2.1839874622941977, + "grad_norm": 3.535710573196411, + "learning_rate": 1.1267191695408432e-05, + "loss": 3.6511, + "step": 194400 + }, + { + "epoch": 2.1845491874645413, + "grad_norm": 1.1054655313491821, + "learning_rate": 1.1264944782105583e-05, + "loss": 3.5475, + "step": 194450 + }, + { + "epoch": 2.1851109126348844, + "grad_norm": 1.1622523069381714, + "learning_rate": 1.1262697868802732e-05, + "loss": 3.6228, + "step": 194500 + }, + { + "epoch": 2.1856726378052276, + "grad_norm": 1.2083780765533447, + "learning_rate": 1.1260450955499882e-05, + "loss": 3.6528, + "step": 194550 + }, + { + "epoch": 2.1862343629755707, + "grad_norm": 1.095506191253662, + "learning_rate": 1.1258204042197031e-05, + "loss": 3.7116, + "step": 194600 + }, + { + "epoch": 2.186796088145914, + "grad_norm": 1.1502091884613037, + "learning_rate": 1.1255957128894182e-05, + "loss": 3.6603, + "step": 194650 + }, + { + "epoch": 2.187357813316257, + "grad_norm": 1.1696006059646606, + "learning_rate": 1.1253710215591331e-05, + "loss": 3.6242, + "step": 194700 + }, + { + "epoch": 2.1879195384866, + "grad_norm": 1.3300443887710571, + "learning_rate": 1.125146330228848e-05, + "loss": 3.625, + "step": 194750 + }, + { + "epoch": 2.1884812636569433, + "grad_norm": 1.1268310546875, + "learning_rate": 1.1249216388985632e-05, + "loss": 3.6589, + "step": 194800 + }, + { + "epoch": 2.1890429888272864, + "grad_norm": 1.2160589694976807, + "learning_rate": 1.1246969475682781e-05, + "loss": 3.5945, + "step": 194850 + }, + { + "epoch": 2.1896047139976296, + "grad_norm": 1.0838061571121216, + "learning_rate": 1.124472256237993e-05, + "loss": 3.6457, + "step": 194900 + }, + { + "epoch": 2.1901664391679727, + "grad_norm": 1.2142306566238403, + "learning_rate": 1.124247564907708e-05, + "loss": 3.6174, + "step": 194950 + }, + { + "epoch": 2.190728164338316, + "grad_norm": 1.091429591178894, + "learning_rate": 1.124022873577423e-05, + "loss": 3.6036, + "step": 195000 + }, + { + "epoch": 2.191289889508659, + "grad_norm": 1.0193670988082886, + "learning_rate": 1.123798182247138e-05, + "loss": 3.6599, + "step": 195050 + }, + { + "epoch": 2.191851614679002, + "grad_norm": 1.2449196577072144, + "learning_rate": 1.123573490916853e-05, + "loss": 3.7411, + "step": 195100 + }, + { + "epoch": 2.1924133398493453, + "grad_norm": 1.2533265352249146, + "learning_rate": 1.123348799586568e-05, + "loss": 3.6701, + "step": 195150 + }, + { + "epoch": 2.1929750650196884, + "grad_norm": 1.0579513311386108, + "learning_rate": 1.1231286020828886e-05, + "loss": 3.5372, + "step": 195200 + }, + { + "epoch": 2.1935367901900316, + "grad_norm": 1.1262048482894897, + "learning_rate": 1.1229039107526035e-05, + "loss": 3.6519, + "step": 195250 + }, + { + "epoch": 2.1940985153603747, + "grad_norm": 1.3239110708236694, + "learning_rate": 1.1226792194223186e-05, + "loss": 3.674, + "step": 195300 + }, + { + "epoch": 2.194660240530718, + "grad_norm": 1.1064801216125488, + "learning_rate": 1.1224545280920336e-05, + "loss": 3.5913, + "step": 195350 + }, + { + "epoch": 2.195221965701061, + "grad_norm": 1.1677799224853516, + "learning_rate": 1.1222298367617485e-05, + "loss": 3.6571, + "step": 195400 + }, + { + "epoch": 2.195783690871404, + "grad_norm": 1.2749618291854858, + "learning_rate": 1.1220051454314636e-05, + "loss": 3.6984, + "step": 195450 + }, + { + "epoch": 2.1963454160417473, + "grad_norm": 1.1262245178222656, + "learning_rate": 1.1217804541011785e-05, + "loss": 3.6573, + "step": 195500 + }, + { + "epoch": 2.1969071412120904, + "grad_norm": 1.2008533477783203, + "learning_rate": 1.1215557627708935e-05, + "loss": 3.6885, + "step": 195550 + }, + { + "epoch": 2.1974688663824335, + "grad_norm": 1.3368501663208008, + "learning_rate": 1.1213310714406084e-05, + "loss": 3.56, + "step": 195600 + }, + { + "epoch": 2.1980305915527767, + "grad_norm": 1.2217975854873657, + "learning_rate": 1.1211063801103235e-05, + "loss": 3.6528, + "step": 195650 + }, + { + "epoch": 2.1985923167231203, + "grad_norm": 1.2406367063522339, + "learning_rate": 1.1208816887800384e-05, + "loss": 3.67, + "step": 195700 + }, + { + "epoch": 2.199154041893463, + "grad_norm": 1.1667741537094116, + "learning_rate": 1.1206569974497534e-05, + "loss": 3.6043, + "step": 195750 + }, + { + "epoch": 2.1997157670638066, + "grad_norm": 1.0640195608139038, + "learning_rate": 1.1204323061194685e-05, + "loss": 3.6652, + "step": 195800 + }, + { + "epoch": 2.2002774922341497, + "grad_norm": 1.1555191278457642, + "learning_rate": 1.1202076147891834e-05, + "loss": 3.6219, + "step": 195850 + }, + { + "epoch": 2.200839217404493, + "grad_norm": 1.0477038621902466, + "learning_rate": 1.1199829234588983e-05, + "loss": 3.5922, + "step": 195900 + }, + { + "epoch": 2.201400942574836, + "grad_norm": 1.1948083639144897, + "learning_rate": 1.1197582321286133e-05, + "loss": 3.6632, + "step": 195950 + }, + { + "epoch": 2.201962667745179, + "grad_norm": 1.0949130058288574, + "learning_rate": 1.1195335407983284e-05, + "loss": 3.5816, + "step": 196000 + }, + { + "epoch": 2.2025243929155223, + "grad_norm": 1.1196852922439575, + "learning_rate": 1.1193088494680433e-05, + "loss": 3.6256, + "step": 196050 + }, + { + "epoch": 2.2030861180858654, + "grad_norm": 1.0841516256332397, + "learning_rate": 1.1190841581377582e-05, + "loss": 3.6894, + "step": 196100 + }, + { + "epoch": 2.2036478432562085, + "grad_norm": 1.2244213819503784, + "learning_rate": 1.1188594668074733e-05, + "loss": 3.6416, + "step": 196150 + }, + { + "epoch": 2.2042095684265517, + "grad_norm": 1.1902669668197632, + "learning_rate": 1.1186347754771883e-05, + "loss": 3.6649, + "step": 196200 + }, + { + "epoch": 2.204771293596895, + "grad_norm": 1.4555490016937256, + "learning_rate": 1.1184100841469032e-05, + "loss": 3.7151, + "step": 196250 + }, + { + "epoch": 2.205333018767238, + "grad_norm": 1.236116647720337, + "learning_rate": 1.1181853928166181e-05, + "loss": 3.5958, + "step": 196300 + }, + { + "epoch": 2.205894743937581, + "grad_norm": 1.2802221775054932, + "learning_rate": 1.1179607014863332e-05, + "loss": 3.5966, + "step": 196350 + }, + { + "epoch": 2.2064564691079243, + "grad_norm": 1.3653600215911865, + "learning_rate": 1.1177360101560482e-05, + "loss": 3.6607, + "step": 196400 + }, + { + "epoch": 2.2070181942782674, + "grad_norm": 0.9888893365859985, + "learning_rate": 1.1175113188257631e-05, + "loss": 3.6654, + "step": 196450 + }, + { + "epoch": 2.2075799194486105, + "grad_norm": 1.0868043899536133, + "learning_rate": 1.1172866274954782e-05, + "loss": 3.6432, + "step": 196500 + }, + { + "epoch": 2.2081416446189537, + "grad_norm": 1.1037492752075195, + "learning_rate": 1.1170619361651931e-05, + "loss": 3.6218, + "step": 196550 + }, + { + "epoch": 2.208703369789297, + "grad_norm": 1.244435429573059, + "learning_rate": 1.116837244834908e-05, + "loss": 3.6909, + "step": 196600 + }, + { + "epoch": 2.20926509495964, + "grad_norm": 1.148008108139038, + "learning_rate": 1.116612553504623e-05, + "loss": 3.646, + "step": 196650 + }, + { + "epoch": 2.209826820129983, + "grad_norm": 1.1879488229751587, + "learning_rate": 1.1163878621743381e-05, + "loss": 3.7235, + "step": 196700 + }, + { + "epoch": 2.2103885453003262, + "grad_norm": 1.0928137302398682, + "learning_rate": 1.116163170844053e-05, + "loss": 3.6463, + "step": 196750 + }, + { + "epoch": 2.2109502704706694, + "grad_norm": 1.0006128549575806, + "learning_rate": 1.115938479513768e-05, + "loss": 3.7025, + "step": 196800 + }, + { + "epoch": 2.2115119956410125, + "grad_norm": 1.2530834674835205, + "learning_rate": 1.115713788183483e-05, + "loss": 3.6741, + "step": 196850 + }, + { + "epoch": 2.2120737208113557, + "grad_norm": 1.4288831949234009, + "learning_rate": 1.115489096853198e-05, + "loss": 3.6525, + "step": 196900 + }, + { + "epoch": 2.212635445981699, + "grad_norm": 1.1534000635147095, + "learning_rate": 1.115264405522913e-05, + "loss": 3.6866, + "step": 196950 + }, + { + "epoch": 2.213197171152042, + "grad_norm": 1.12404465675354, + "learning_rate": 1.1150397141926278e-05, + "loss": 3.7836, + "step": 197000 + }, + { + "epoch": 2.2137588963223855, + "grad_norm": 1.221832275390625, + "learning_rate": 1.114815022862343e-05, + "loss": 3.7044, + "step": 197050 + }, + { + "epoch": 2.2143206214927282, + "grad_norm": 1.2092190980911255, + "learning_rate": 1.1145903315320579e-05, + "loss": 3.6073, + "step": 197100 + }, + { + "epoch": 2.214882346663072, + "grad_norm": 1.0969897508621216, + "learning_rate": 1.1143656402017728e-05, + "loss": 3.6569, + "step": 197150 + }, + { + "epoch": 2.215444071833415, + "grad_norm": 1.1661725044250488, + "learning_rate": 1.1141409488714877e-05, + "loss": 3.6218, + "step": 197200 + }, + { + "epoch": 2.216005797003758, + "grad_norm": 3.4693822860717773, + "learning_rate": 1.1139162575412028e-05, + "loss": 3.6296, + "step": 197250 + }, + { + "epoch": 2.2165675221741012, + "grad_norm": 1.1358059644699097, + "learning_rate": 1.1136915662109178e-05, + "loss": 3.5186, + "step": 197300 + }, + { + "epoch": 2.2171292473444444, + "grad_norm": 1.1608294248580933, + "learning_rate": 1.1134668748806327e-05, + "loss": 3.6262, + "step": 197350 + }, + { + "epoch": 2.2176909725147875, + "grad_norm": 1.3317348957061768, + "learning_rate": 1.1132421835503478e-05, + "loss": 3.602, + "step": 197400 + }, + { + "epoch": 2.2182526976851307, + "grad_norm": 1.1245685815811157, + "learning_rate": 1.1130174922200627e-05, + "loss": 3.7152, + "step": 197450 + }, + { + "epoch": 2.218814422855474, + "grad_norm": 1.0280276536941528, + "learning_rate": 1.1127928008897777e-05, + "loss": 3.5854, + "step": 197500 + }, + { + "epoch": 2.219376148025817, + "grad_norm": 1.0780091285705566, + "learning_rate": 1.1125681095594926e-05, + "loss": 3.6664, + "step": 197550 + }, + { + "epoch": 2.21993787319616, + "grad_norm": 1.0677632093429565, + "learning_rate": 1.1123434182292077e-05, + "loss": 3.683, + "step": 197600 + }, + { + "epoch": 2.2204995983665032, + "grad_norm": 1.2442424297332764, + "learning_rate": 1.1121187268989226e-05, + "loss": 3.6738, + "step": 197650 + }, + { + "epoch": 2.2210613235368464, + "grad_norm": 1.1354444026947021, + "learning_rate": 1.1118940355686376e-05, + "loss": 3.6153, + "step": 197700 + }, + { + "epoch": 2.2216230487071895, + "grad_norm": 1.225928544998169, + "learning_rate": 1.1116693442383527e-05, + "loss": 3.6901, + "step": 197750 + }, + { + "epoch": 2.2221847738775327, + "grad_norm": 1.3080968856811523, + "learning_rate": 1.1114446529080676e-05, + "loss": 3.6879, + "step": 197800 + }, + { + "epoch": 2.222746499047876, + "grad_norm": 1.0205680131912231, + "learning_rate": 1.1112199615777825e-05, + "loss": 3.6512, + "step": 197850 + }, + { + "epoch": 2.223308224218219, + "grad_norm": 1.0382564067840576, + "learning_rate": 1.1109952702474975e-05, + "loss": 3.7031, + "step": 197900 + }, + { + "epoch": 2.223869949388562, + "grad_norm": 1.2341996431350708, + "learning_rate": 1.1107705789172126e-05, + "loss": 3.7014, + "step": 197950 + }, + { + "epoch": 2.2244316745589052, + "grad_norm": 1.1369259357452393, + "learning_rate": 1.1105458875869275e-05, + "loss": 3.7022, + "step": 198000 + }, + { + "epoch": 2.2249933997292484, + "grad_norm": 1.226236343383789, + "learning_rate": 1.1103211962566424e-05, + "loss": 3.5697, + "step": 198050 + }, + { + "epoch": 2.2255551248995915, + "grad_norm": 1.083716630935669, + "learning_rate": 1.1100965049263575e-05, + "loss": 3.5827, + "step": 198100 + }, + { + "epoch": 2.2261168500699346, + "grad_norm": 1.1048554182052612, + "learning_rate": 1.1098718135960725e-05, + "loss": 3.5583, + "step": 198150 + }, + { + "epoch": 2.226678575240278, + "grad_norm": 1.3492372035980225, + "learning_rate": 1.1096471222657874e-05, + "loss": 3.6775, + "step": 198200 + }, + { + "epoch": 2.227240300410621, + "grad_norm": 1.2858997583389282, + "learning_rate": 1.1094224309355023e-05, + "loss": 3.6434, + "step": 198250 + }, + { + "epoch": 2.227802025580964, + "grad_norm": 1.2529743909835815, + "learning_rate": 1.1091977396052174e-05, + "loss": 3.5534, + "step": 198300 + }, + { + "epoch": 2.228363750751307, + "grad_norm": 1.1517821550369263, + "learning_rate": 1.1089730482749324e-05, + "loss": 3.6645, + "step": 198350 + }, + { + "epoch": 2.228925475921651, + "grad_norm": 1.4201101064682007, + "learning_rate": 1.1087483569446473e-05, + "loss": 3.7434, + "step": 198400 + }, + { + "epoch": 2.2294872010919935, + "grad_norm": 1.2251349687576294, + "learning_rate": 1.1085236656143624e-05, + "loss": 3.6577, + "step": 198450 + }, + { + "epoch": 2.230048926262337, + "grad_norm": 1.2064093351364136, + "learning_rate": 1.1082989742840773e-05, + "loss": 3.57, + "step": 198500 + }, + { + "epoch": 2.2306106514326802, + "grad_norm": 1.2292957305908203, + "learning_rate": 1.1080742829537923e-05, + "loss": 3.6438, + "step": 198550 + }, + { + "epoch": 2.2311723766030234, + "grad_norm": 1.165260910987854, + "learning_rate": 1.1078495916235072e-05, + "loss": 3.6598, + "step": 198600 + }, + { + "epoch": 2.2317341017733665, + "grad_norm": 1.052223563194275, + "learning_rate": 1.1076249002932223e-05, + "loss": 3.7084, + "step": 198650 + }, + { + "epoch": 2.2322958269437096, + "grad_norm": 1.179465413093567, + "learning_rate": 1.1074002089629372e-05, + "loss": 3.7048, + "step": 198700 + }, + { + "epoch": 2.232857552114053, + "grad_norm": 1.2978014945983887, + "learning_rate": 1.1071755176326522e-05, + "loss": 3.6179, + "step": 198750 + }, + { + "epoch": 2.233419277284396, + "grad_norm": 1.4667948484420776, + "learning_rate": 1.106950826302367e-05, + "loss": 3.6239, + "step": 198800 + }, + { + "epoch": 2.233981002454739, + "grad_norm": 1.246903657913208, + "learning_rate": 1.1067261349720822e-05, + "loss": 3.6396, + "step": 198850 + }, + { + "epoch": 2.234542727625082, + "grad_norm": 0.9928773045539856, + "learning_rate": 1.1065014436417971e-05, + "loss": 3.6633, + "step": 198900 + }, + { + "epoch": 2.2351044527954254, + "grad_norm": 1.1761093139648438, + "learning_rate": 1.106276752311512e-05, + "loss": 3.6865, + "step": 198950 + }, + { + "epoch": 2.2356661779657685, + "grad_norm": 1.2706598043441772, + "learning_rate": 1.1060520609812271e-05, + "loss": 3.654, + "step": 199000 + }, + { + "epoch": 2.2362279031361116, + "grad_norm": 1.2302583456039429, + "learning_rate": 1.105827369650942e-05, + "loss": 3.6131, + "step": 199050 + }, + { + "epoch": 2.236789628306455, + "grad_norm": 1.1114083528518677, + "learning_rate": 1.105602678320657e-05, + "loss": 3.6453, + "step": 199100 + }, + { + "epoch": 2.237351353476798, + "grad_norm": 1.1350913047790527, + "learning_rate": 1.105377986990372e-05, + "loss": 3.676, + "step": 199150 + }, + { + "epoch": 2.237913078647141, + "grad_norm": 1.2548173666000366, + "learning_rate": 1.105153295660087e-05, + "loss": 3.6188, + "step": 199200 + }, + { + "epoch": 2.238474803817484, + "grad_norm": 1.1997712850570679, + "learning_rate": 1.104928604329802e-05, + "loss": 3.6246, + "step": 199250 + }, + { + "epoch": 2.2390365289878273, + "grad_norm": 1.1969828605651855, + "learning_rate": 1.1047039129995169e-05, + "loss": 3.704, + "step": 199300 + }, + { + "epoch": 2.2395982541581705, + "grad_norm": 1.2156494855880737, + "learning_rate": 1.104479221669232e-05, + "loss": 3.6953, + "step": 199350 + }, + { + "epoch": 2.2401599793285136, + "grad_norm": 1.2109017372131348, + "learning_rate": 1.104254530338947e-05, + "loss": 3.7038, + "step": 199400 + }, + { + "epoch": 2.2407217044988568, + "grad_norm": 1.0635199546813965, + "learning_rate": 1.1040298390086619e-05, + "loss": 3.6131, + "step": 199450 + }, + { + "epoch": 2.2412834296692, + "grad_norm": 1.0498710870742798, + "learning_rate": 1.1038051476783768e-05, + "loss": 3.6186, + "step": 199500 + }, + { + "epoch": 2.241845154839543, + "grad_norm": 1.3620375394821167, + "learning_rate": 1.1035804563480919e-05, + "loss": 3.6428, + "step": 199550 + }, + { + "epoch": 2.242406880009886, + "grad_norm": 1.0943694114685059, + "learning_rate": 1.1033557650178068e-05, + "loss": 3.6925, + "step": 199600 + }, + { + "epoch": 2.2429686051802293, + "grad_norm": 1.103467345237732, + "learning_rate": 1.1031310736875218e-05, + "loss": 3.6864, + "step": 199650 + }, + { + "epoch": 2.2435303303505725, + "grad_norm": 1.0348925590515137, + "learning_rate": 1.1029063823572369e-05, + "loss": 3.6311, + "step": 199700 + }, + { + "epoch": 2.244092055520916, + "grad_norm": 1.1905165910720825, + "learning_rate": 1.1026816910269518e-05, + "loss": 3.659, + "step": 199750 + }, + { + "epoch": 2.2446537806912588, + "grad_norm": 1.3361129760742188, + "learning_rate": 1.1024569996966667e-05, + "loss": 3.6541, + "step": 199800 + }, + { + "epoch": 2.2452155058616023, + "grad_norm": 1.1166316270828247, + "learning_rate": 1.1022323083663817e-05, + "loss": 3.5581, + "step": 199850 + }, + { + "epoch": 2.2457772310319455, + "grad_norm": 1.2856205701828003, + "learning_rate": 1.1020076170360968e-05, + "loss": 3.6256, + "step": 199900 + }, + { + "epoch": 2.2463389562022886, + "grad_norm": 1.3031456470489502, + "learning_rate": 1.1017829257058117e-05, + "loss": 3.6146, + "step": 199950 + }, + { + "epoch": 2.2469006813726318, + "grad_norm": 1.0035897493362427, + "learning_rate": 1.1015582343755266e-05, + "loss": 3.7184, + "step": 200000 + }, + { + "epoch": 2.247462406542975, + "grad_norm": 1.3020460605621338, + "learning_rate": 1.1013335430452417e-05, + "loss": 3.6684, + "step": 200050 + }, + { + "epoch": 2.248024131713318, + "grad_norm": 1.1402523517608643, + "learning_rate": 1.1011088517149567e-05, + "loss": 3.6102, + "step": 200100 + }, + { + "epoch": 2.248585856883661, + "grad_norm": 1.2541253566741943, + "learning_rate": 1.1008841603846716e-05, + "loss": 3.6028, + "step": 200150 + }, + { + "epoch": 2.2491475820540043, + "grad_norm": 1.1429928541183472, + "learning_rate": 1.1006594690543865e-05, + "loss": 3.664, + "step": 200200 + }, + { + "epoch": 2.2497093072243475, + "grad_norm": 1.4887467622756958, + "learning_rate": 1.1004392715507073e-05, + "loss": 3.6599, + "step": 200250 + }, + { + "epoch": 2.2502710323946906, + "grad_norm": 1.1874189376831055, + "learning_rate": 1.1002145802204222e-05, + "loss": 3.7105, + "step": 200300 + }, + { + "epoch": 2.2508327575650338, + "grad_norm": 1.4966840744018555, + "learning_rate": 1.0999898888901373e-05, + "loss": 3.6475, + "step": 200350 + }, + { + "epoch": 2.251394482735377, + "grad_norm": 1.068434476852417, + "learning_rate": 1.0997651975598523e-05, + "loss": 3.6619, + "step": 200400 + }, + { + "epoch": 2.25195620790572, + "grad_norm": 1.2671818733215332, + "learning_rate": 1.0995405062295672e-05, + "loss": 3.6184, + "step": 200450 + }, + { + "epoch": 2.252517933076063, + "grad_norm": 1.0523030757904053, + "learning_rate": 1.0993158148992821e-05, + "loss": 3.6023, + "step": 200500 + }, + { + "epoch": 2.2530796582464063, + "grad_norm": 1.0349535942077637, + "learning_rate": 1.0990911235689972e-05, + "loss": 3.6532, + "step": 200550 + }, + { + "epoch": 2.2536413834167495, + "grad_norm": 0.9204732179641724, + "learning_rate": 1.0988664322387121e-05, + "loss": 3.6567, + "step": 200600 + }, + { + "epoch": 2.2542031085870926, + "grad_norm": 1.0933384895324707, + "learning_rate": 1.098641740908427e-05, + "loss": 3.6977, + "step": 200650 + }, + { + "epoch": 2.2547648337574357, + "grad_norm": 1.305959939956665, + "learning_rate": 1.0984170495781422e-05, + "loss": 3.6408, + "step": 200700 + }, + { + "epoch": 2.255326558927779, + "grad_norm": 1.1364576816558838, + "learning_rate": 1.0981923582478571e-05, + "loss": 3.5922, + "step": 200750 + }, + { + "epoch": 2.255888284098122, + "grad_norm": 1.2036141157150269, + "learning_rate": 1.097967666917572e-05, + "loss": 3.6742, + "step": 200800 + }, + { + "epoch": 2.256450009268465, + "grad_norm": 1.1628385782241821, + "learning_rate": 1.097742975587287e-05, + "loss": 3.7339, + "step": 200850 + }, + { + "epoch": 2.2570117344388083, + "grad_norm": 1.0342094898223877, + "learning_rate": 1.097518284257002e-05, + "loss": 3.6503, + "step": 200900 + }, + { + "epoch": 2.2575734596091515, + "grad_norm": 1.196022629737854, + "learning_rate": 1.097293592926717e-05, + "loss": 3.6804, + "step": 200950 + }, + { + "epoch": 2.2581351847794946, + "grad_norm": 1.245009422302246, + "learning_rate": 1.097068901596432e-05, + "loss": 3.5853, + "step": 201000 + }, + { + "epoch": 2.2586969099498377, + "grad_norm": 1.203466534614563, + "learning_rate": 1.096844210266147e-05, + "loss": 3.6283, + "step": 201050 + }, + { + "epoch": 2.2592586351201813, + "grad_norm": 1.1776548624038696, + "learning_rate": 1.096619518935862e-05, + "loss": 3.637, + "step": 201100 + }, + { + "epoch": 2.259820360290524, + "grad_norm": 1.0840448141098022, + "learning_rate": 1.0963948276055769e-05, + "loss": 3.595, + "step": 201150 + }, + { + "epoch": 2.2603820854608676, + "grad_norm": 1.3171910047531128, + "learning_rate": 1.0961701362752918e-05, + "loss": 3.6929, + "step": 201200 + }, + { + "epoch": 2.2609438106312107, + "grad_norm": 0.9727421402931213, + "learning_rate": 1.095945444945007e-05, + "loss": 3.5353, + "step": 201250 + }, + { + "epoch": 2.261505535801554, + "grad_norm": 1.2239476442337036, + "learning_rate": 1.0957207536147219e-05, + "loss": 3.6409, + "step": 201300 + }, + { + "epoch": 2.262067260971897, + "grad_norm": 1.076357364654541, + "learning_rate": 1.0954960622844368e-05, + "loss": 3.6531, + "step": 201350 + }, + { + "epoch": 2.26262898614224, + "grad_norm": 1.1421380043029785, + "learning_rate": 1.0952713709541517e-05, + "loss": 3.6225, + "step": 201400 + }, + { + "epoch": 2.2631907113125833, + "grad_norm": 1.158110499382019, + "learning_rate": 1.0950466796238668e-05, + "loss": 3.6719, + "step": 201450 + }, + { + "epoch": 2.2637524364829265, + "grad_norm": 1.3112207651138306, + "learning_rate": 1.0948219882935818e-05, + "loss": 3.6902, + "step": 201500 + }, + { + "epoch": 2.2643141616532696, + "grad_norm": 1.178525686264038, + "learning_rate": 1.0945972969632967e-05, + "loss": 3.6311, + "step": 201550 + }, + { + "epoch": 2.2648758868236127, + "grad_norm": 1.0799179077148438, + "learning_rate": 1.0943726056330118e-05, + "loss": 3.6396, + "step": 201600 + }, + { + "epoch": 2.265437611993956, + "grad_norm": 1.1785821914672852, + "learning_rate": 1.0941479143027267e-05, + "loss": 3.6488, + "step": 201650 + }, + { + "epoch": 2.265999337164299, + "grad_norm": 1.0482996702194214, + "learning_rate": 1.0939232229724417e-05, + "loss": 3.6388, + "step": 201700 + }, + { + "epoch": 2.266561062334642, + "grad_norm": 1.2701849937438965, + "learning_rate": 1.0936985316421566e-05, + "loss": 3.6097, + "step": 201750 + }, + { + "epoch": 2.2671227875049853, + "grad_norm": 1.0427109003067017, + "learning_rate": 1.0934738403118717e-05, + "loss": 3.5898, + "step": 201800 + }, + { + "epoch": 2.2676845126753284, + "grad_norm": 1.1787889003753662, + "learning_rate": 1.0932491489815866e-05, + "loss": 3.5461, + "step": 201850 + }, + { + "epoch": 2.2682462378456716, + "grad_norm": 1.0887303352355957, + "learning_rate": 1.0930244576513016e-05, + "loss": 3.6988, + "step": 201900 + }, + { + "epoch": 2.2688079630160147, + "grad_norm": 1.3267873525619507, + "learning_rate": 1.0927997663210167e-05, + "loss": 3.6635, + "step": 201950 + }, + { + "epoch": 2.269369688186358, + "grad_norm": 1.1822171211242676, + "learning_rate": 1.0925750749907316e-05, + "loss": 3.596, + "step": 202000 + }, + { + "epoch": 2.269931413356701, + "grad_norm": 1.2599691152572632, + "learning_rate": 1.0923503836604465e-05, + "loss": 3.6804, + "step": 202050 + }, + { + "epoch": 2.270493138527044, + "grad_norm": 1.1934468746185303, + "learning_rate": 1.0921256923301615e-05, + "loss": 3.66, + "step": 202100 + }, + { + "epoch": 2.2710548636973873, + "grad_norm": 1.2453982830047607, + "learning_rate": 1.0919010009998766e-05, + "loss": 3.6684, + "step": 202150 + }, + { + "epoch": 2.2716165888677304, + "grad_norm": 0.9536491632461548, + "learning_rate": 1.0916763096695915e-05, + "loss": 3.6246, + "step": 202200 + }, + { + "epoch": 2.2721783140380736, + "grad_norm": 1.0329339504241943, + "learning_rate": 1.0914516183393064e-05, + "loss": 3.7197, + "step": 202250 + }, + { + "epoch": 2.2727400392084167, + "grad_norm": 1.1337761878967285, + "learning_rate": 1.0912269270090215e-05, + "loss": 3.6023, + "step": 202300 + }, + { + "epoch": 2.27330176437876, + "grad_norm": 0.9923123717308044, + "learning_rate": 1.0910022356787364e-05, + "loss": 3.6429, + "step": 202350 + }, + { + "epoch": 2.273863489549103, + "grad_norm": 1.7720701694488525, + "learning_rate": 1.0907775443484514e-05, + "loss": 3.6781, + "step": 202400 + }, + { + "epoch": 2.2744252147194466, + "grad_norm": 1.190482258796692, + "learning_rate": 1.0905573468447721e-05, + "loss": 3.6465, + "step": 202450 + }, + { + "epoch": 2.2749869398897893, + "grad_norm": 1.2313224077224731, + "learning_rate": 1.090332655514487e-05, + "loss": 3.6782, + "step": 202500 + }, + { + "epoch": 2.275548665060133, + "grad_norm": 1.0463004112243652, + "learning_rate": 1.090107964184202e-05, + "loss": 3.7569, + "step": 202550 + }, + { + "epoch": 2.276110390230476, + "grad_norm": 1.1098483800888062, + "learning_rate": 1.0898832728539171e-05, + "loss": 3.5805, + "step": 202600 + }, + { + "epoch": 2.276672115400819, + "grad_norm": 1.1954679489135742, + "learning_rate": 1.089658581523632e-05, + "loss": 3.6629, + "step": 202650 + }, + { + "epoch": 2.2772338405711623, + "grad_norm": 1.240631341934204, + "learning_rate": 1.089433890193347e-05, + "loss": 3.6044, + "step": 202700 + }, + { + "epoch": 2.2777955657415054, + "grad_norm": 1.313819169998169, + "learning_rate": 1.0892091988630619e-05, + "loss": 3.6146, + "step": 202750 + }, + { + "epoch": 2.2783572909118486, + "grad_norm": 1.261030912399292, + "learning_rate": 1.088984507532777e-05, + "loss": 3.6696, + "step": 202800 + }, + { + "epoch": 2.2789190160821917, + "grad_norm": 1.1849219799041748, + "learning_rate": 1.088759816202492e-05, + "loss": 3.7237, + "step": 202850 + }, + { + "epoch": 2.279480741252535, + "grad_norm": 1.3644628524780273, + "learning_rate": 1.0885351248722069e-05, + "loss": 3.6499, + "step": 202900 + }, + { + "epoch": 2.280042466422878, + "grad_norm": 1.3659747838974, + "learning_rate": 1.088310433541922e-05, + "loss": 3.66, + "step": 202950 + }, + { + "epoch": 2.280604191593221, + "grad_norm": 1.1541392803192139, + "learning_rate": 1.0880857422116369e-05, + "loss": 3.7414, + "step": 203000 + }, + { + "epoch": 2.2811659167635643, + "grad_norm": 1.039172649383545, + "learning_rate": 1.0878610508813518e-05, + "loss": 3.6835, + "step": 203050 + }, + { + "epoch": 2.2817276419339074, + "grad_norm": 1.0046977996826172, + "learning_rate": 1.0876363595510668e-05, + "loss": 3.6234, + "step": 203100 + }, + { + "epoch": 2.2822893671042506, + "grad_norm": 1.396032452583313, + "learning_rate": 1.0874116682207819e-05, + "loss": 3.6203, + "step": 203150 + }, + { + "epoch": 2.2828510922745937, + "grad_norm": 1.0471729040145874, + "learning_rate": 1.0871869768904968e-05, + "loss": 3.6176, + "step": 203200 + }, + { + "epoch": 2.283412817444937, + "grad_norm": 1.3599485158920288, + "learning_rate": 1.0869622855602117e-05, + "loss": 3.6163, + "step": 203250 + }, + { + "epoch": 2.28397454261528, + "grad_norm": 1.3429630994796753, + "learning_rate": 1.0867375942299268e-05, + "loss": 3.6847, + "step": 203300 + }, + { + "epoch": 2.284536267785623, + "grad_norm": 1.2321631908416748, + "learning_rate": 1.0865129028996418e-05, + "loss": 3.6505, + "step": 203350 + }, + { + "epoch": 2.2850979929559663, + "grad_norm": 1.2308857440948486, + "learning_rate": 1.0862882115693567e-05, + "loss": 3.6571, + "step": 203400 + }, + { + "epoch": 2.2856597181263094, + "grad_norm": 1.1756349802017212, + "learning_rate": 1.0860635202390716e-05, + "loss": 3.6543, + "step": 203450 + }, + { + "epoch": 2.2862214432966526, + "grad_norm": 1.045255422592163, + "learning_rate": 1.0858388289087867e-05, + "loss": 3.6134, + "step": 203500 + }, + { + "epoch": 2.2867831684669957, + "grad_norm": 1.206047534942627, + "learning_rate": 1.0856141375785017e-05, + "loss": 3.591, + "step": 203550 + }, + { + "epoch": 2.287344893637339, + "grad_norm": 1.1834707260131836, + "learning_rate": 1.0853894462482166e-05, + "loss": 3.6628, + "step": 203600 + }, + { + "epoch": 2.287906618807682, + "grad_norm": 0.9967352151870728, + "learning_rate": 1.0851647549179317e-05, + "loss": 3.6352, + "step": 203650 + }, + { + "epoch": 2.288468343978025, + "grad_norm": 1.293893814086914, + "learning_rate": 1.0849400635876466e-05, + "loss": 3.6823, + "step": 203700 + }, + { + "epoch": 2.2890300691483683, + "grad_norm": 1.2545384168624878, + "learning_rate": 1.0847153722573616e-05, + "loss": 3.6486, + "step": 203750 + }, + { + "epoch": 2.289591794318712, + "grad_norm": 1.224974513053894, + "learning_rate": 1.0844906809270765e-05, + "loss": 3.7078, + "step": 203800 + }, + { + "epoch": 2.2901535194890545, + "grad_norm": 1.20431649684906, + "learning_rate": 1.0842659895967916e-05, + "loss": 3.702, + "step": 203850 + }, + { + "epoch": 2.290715244659398, + "grad_norm": 1.206586241722107, + "learning_rate": 1.0840412982665065e-05, + "loss": 3.6062, + "step": 203900 + }, + { + "epoch": 2.2912769698297413, + "grad_norm": 1.0869084596633911, + "learning_rate": 1.0838166069362214e-05, + "loss": 3.6111, + "step": 203950 + }, + { + "epoch": 2.2918386950000844, + "grad_norm": 1.0805386304855347, + "learning_rate": 1.0835919156059364e-05, + "loss": 3.645, + "step": 204000 + }, + { + "epoch": 2.2924004201704276, + "grad_norm": 1.3582491874694824, + "learning_rate": 1.0833672242756515e-05, + "loss": 3.7398, + "step": 204050 + }, + { + "epoch": 2.2929621453407707, + "grad_norm": 1.1402126550674438, + "learning_rate": 1.0831425329453664e-05, + "loss": 3.5855, + "step": 204100 + }, + { + "epoch": 2.293523870511114, + "grad_norm": 1.2263522148132324, + "learning_rate": 1.0829178416150813e-05, + "loss": 3.6099, + "step": 204150 + }, + { + "epoch": 2.294085595681457, + "grad_norm": 1.1827775239944458, + "learning_rate": 1.0826931502847964e-05, + "loss": 3.6069, + "step": 204200 + }, + { + "epoch": 2.2946473208518, + "grad_norm": 1.1820076704025269, + "learning_rate": 1.0824684589545114e-05, + "loss": 3.5847, + "step": 204250 + }, + { + "epoch": 2.2952090460221433, + "grad_norm": 1.3091599941253662, + "learning_rate": 1.0822437676242263e-05, + "loss": 3.6601, + "step": 204300 + }, + { + "epoch": 2.2957707711924864, + "grad_norm": 1.1638920307159424, + "learning_rate": 1.0820190762939412e-05, + "loss": 3.6105, + "step": 204350 + }, + { + "epoch": 2.2963324963628295, + "grad_norm": 1.0978766679763794, + "learning_rate": 1.0817943849636563e-05, + "loss": 3.6773, + "step": 204400 + }, + { + "epoch": 2.2968942215331727, + "grad_norm": 1.1125136613845825, + "learning_rate": 1.0815696936333713e-05, + "loss": 3.6585, + "step": 204450 + }, + { + "epoch": 2.297455946703516, + "grad_norm": 1.2809821367263794, + "learning_rate": 1.0813450023030862e-05, + "loss": 3.6271, + "step": 204500 + }, + { + "epoch": 2.298017671873859, + "grad_norm": 1.2620893716812134, + "learning_rate": 1.0811203109728013e-05, + "loss": 3.648, + "step": 204550 + }, + { + "epoch": 2.298579397044202, + "grad_norm": 1.2011334896087646, + "learning_rate": 1.0808956196425162e-05, + "loss": 3.6975, + "step": 204600 + }, + { + "epoch": 2.2991411222145453, + "grad_norm": 1.1442394256591797, + "learning_rate": 1.0806709283122312e-05, + "loss": 3.5874, + "step": 204650 + }, + { + "epoch": 2.2997028473848884, + "grad_norm": 1.2460062503814697, + "learning_rate": 1.0804462369819461e-05, + "loss": 3.6958, + "step": 204700 + }, + { + "epoch": 2.3002645725552315, + "grad_norm": 1.1312328577041626, + "learning_rate": 1.0802215456516612e-05, + "loss": 3.6445, + "step": 204750 + }, + { + "epoch": 2.3008262977255747, + "grad_norm": 1.2142772674560547, + "learning_rate": 1.0799968543213761e-05, + "loss": 3.6515, + "step": 204800 + }, + { + "epoch": 2.301388022895918, + "grad_norm": 1.2299082279205322, + "learning_rate": 1.079772162991091e-05, + "loss": 3.6667, + "step": 204850 + }, + { + "epoch": 2.301949748066261, + "grad_norm": 1.1363869905471802, + "learning_rate": 1.0795474716608062e-05, + "loss": 3.692, + "step": 204900 + }, + { + "epoch": 2.302511473236604, + "grad_norm": 1.0149507522583008, + "learning_rate": 1.0793227803305211e-05, + "loss": 3.6699, + "step": 204950 + }, + { + "epoch": 2.3030731984069472, + "grad_norm": 1.3276242017745972, + "learning_rate": 1.079098089000236e-05, + "loss": 3.6628, + "step": 205000 + }, + { + "epoch": 2.3036349235772904, + "grad_norm": 1.2210326194763184, + "learning_rate": 1.078873397669951e-05, + "loss": 3.6833, + "step": 205050 + }, + { + "epoch": 2.3041966487476335, + "grad_norm": 1.1588329076766968, + "learning_rate": 1.078648706339666e-05, + "loss": 3.6682, + "step": 205100 + }, + { + "epoch": 2.304758373917977, + "grad_norm": 1.1705402135849, + "learning_rate": 1.078424015009381e-05, + "loss": 3.6154, + "step": 205150 + }, + { + "epoch": 2.30532009908832, + "grad_norm": 1.0182251930236816, + "learning_rate": 1.078199323679096e-05, + "loss": 3.6416, + "step": 205200 + }, + { + "epoch": 2.3058818242586634, + "grad_norm": 1.4481021165847778, + "learning_rate": 1.077974632348811e-05, + "loss": 3.5957, + "step": 205250 + }, + { + "epoch": 2.3064435494290065, + "grad_norm": 1.7506201267242432, + "learning_rate": 1.077749941018526e-05, + "loss": 3.6368, + "step": 205300 + }, + { + "epoch": 2.3070052745993497, + "grad_norm": 1.2117962837219238, + "learning_rate": 1.0775252496882409e-05, + "loss": 3.6373, + "step": 205350 + }, + { + "epoch": 2.307566999769693, + "grad_norm": 1.046059012413025, + "learning_rate": 1.0773005583579558e-05, + "loss": 3.6224, + "step": 205400 + }, + { + "epoch": 2.308128724940036, + "grad_norm": 1.2020015716552734, + "learning_rate": 1.077075867027671e-05, + "loss": 3.6837, + "step": 205450 + }, + { + "epoch": 2.308690450110379, + "grad_norm": 1.1153267621994019, + "learning_rate": 1.0768511756973859e-05, + "loss": 3.6947, + "step": 205500 + }, + { + "epoch": 2.3092521752807222, + "grad_norm": 1.2907932996749878, + "learning_rate": 1.0766264843671008e-05, + "loss": 3.7269, + "step": 205550 + }, + { + "epoch": 2.3098139004510654, + "grad_norm": 1.1694223880767822, + "learning_rate": 1.0764017930368157e-05, + "loss": 3.6491, + "step": 205600 + }, + { + "epoch": 2.3103756256214085, + "grad_norm": 1.2663607597351074, + "learning_rate": 1.0761771017065308e-05, + "loss": 3.6718, + "step": 205650 + }, + { + "epoch": 2.3109373507917517, + "grad_norm": 1.1763992309570312, + "learning_rate": 1.0759524103762457e-05, + "loss": 3.5671, + "step": 205700 + }, + { + "epoch": 2.311499075962095, + "grad_norm": 1.0478515625, + "learning_rate": 1.0757277190459607e-05, + "loss": 3.6266, + "step": 205750 + }, + { + "epoch": 2.312060801132438, + "grad_norm": 1.0747405290603638, + "learning_rate": 1.0755030277156758e-05, + "loss": 3.6551, + "step": 205800 + }, + { + "epoch": 2.312622526302781, + "grad_norm": 1.0794366598129272, + "learning_rate": 1.0752783363853907e-05, + "loss": 3.6115, + "step": 205850 + }, + { + "epoch": 2.3131842514731242, + "grad_norm": 1.2503647804260254, + "learning_rate": 1.0750536450551056e-05, + "loss": 3.6626, + "step": 205900 + }, + { + "epoch": 2.3137459766434674, + "grad_norm": 1.3416721820831299, + "learning_rate": 1.0748289537248206e-05, + "loss": 3.6613, + "step": 205950 + }, + { + "epoch": 2.3143077018138105, + "grad_norm": 1.128549337387085, + "learning_rate": 1.0746042623945357e-05, + "loss": 3.6929, + "step": 206000 + }, + { + "epoch": 2.3148694269841537, + "grad_norm": 1.3488574028015137, + "learning_rate": 1.0743795710642506e-05, + "loss": 3.6999, + "step": 206050 + }, + { + "epoch": 2.315431152154497, + "grad_norm": 1.2450639009475708, + "learning_rate": 1.0741548797339655e-05, + "loss": 3.6602, + "step": 206100 + }, + { + "epoch": 2.31599287732484, + "grad_norm": 1.137138843536377, + "learning_rate": 1.0739301884036806e-05, + "loss": 3.6707, + "step": 206150 + }, + { + "epoch": 2.316554602495183, + "grad_norm": 1.1490094661712646, + "learning_rate": 1.073714484726607e-05, + "loss": 3.6099, + "step": 206200 + }, + { + "epoch": 2.317116327665526, + "grad_norm": 1.1616986989974976, + "learning_rate": 1.073489793396322e-05, + "loss": 3.658, + "step": 206250 + }, + { + "epoch": 2.3176780528358694, + "grad_norm": 1.285482406616211, + "learning_rate": 1.073265102066037e-05, + "loss": 3.6447, + "step": 206300 + }, + { + "epoch": 2.3182397780062125, + "grad_norm": 1.3026942014694214, + "learning_rate": 1.0730404107357519e-05, + "loss": 3.7069, + "step": 206350 + }, + { + "epoch": 2.3188015031765556, + "grad_norm": 1.1542354822158813, + "learning_rate": 1.072815719405467e-05, + "loss": 3.724, + "step": 206400 + }, + { + "epoch": 2.319363228346899, + "grad_norm": 1.1903376579284668, + "learning_rate": 1.0725910280751819e-05, + "loss": 3.5683, + "step": 206450 + }, + { + "epoch": 2.3199249535172424, + "grad_norm": 1.0221582651138306, + "learning_rate": 1.0723663367448968e-05, + "loss": 3.6657, + "step": 206500 + }, + { + "epoch": 2.320486678687585, + "grad_norm": 1.1555572748184204, + "learning_rate": 1.072141645414612e-05, + "loss": 3.6727, + "step": 206550 + }, + { + "epoch": 2.3210484038579287, + "grad_norm": 1.0467361211776733, + "learning_rate": 1.0719169540843269e-05, + "loss": 3.6765, + "step": 206600 + }, + { + "epoch": 2.321610129028272, + "grad_norm": 1.1794931888580322, + "learning_rate": 1.0716922627540418e-05, + "loss": 3.7091, + "step": 206650 + }, + { + "epoch": 2.322171854198615, + "grad_norm": 1.255722999572754, + "learning_rate": 1.0714675714237567e-05, + "loss": 3.6278, + "step": 206700 + }, + { + "epoch": 2.322733579368958, + "grad_norm": 1.2511638402938843, + "learning_rate": 1.0712428800934718e-05, + "loss": 3.672, + "step": 206750 + }, + { + "epoch": 2.323295304539301, + "grad_norm": 1.034964680671692, + "learning_rate": 1.0710181887631868e-05, + "loss": 3.6508, + "step": 206800 + }, + { + "epoch": 2.3238570297096444, + "grad_norm": 1.3474657535552979, + "learning_rate": 1.0707934974329017e-05, + "loss": 3.6476, + "step": 206850 + }, + { + "epoch": 2.3244187548799875, + "grad_norm": 1.443680763244629, + "learning_rate": 1.0705688061026168e-05, + "loss": 3.6276, + "step": 206900 + }, + { + "epoch": 2.3249804800503306, + "grad_norm": 1.1959564685821533, + "learning_rate": 1.0703441147723317e-05, + "loss": 3.6052, + "step": 206950 + }, + { + "epoch": 2.325542205220674, + "grad_norm": 1.2000470161437988, + "learning_rate": 1.0701194234420466e-05, + "loss": 3.6673, + "step": 207000 + }, + { + "epoch": 2.326103930391017, + "grad_norm": 1.1703729629516602, + "learning_rate": 1.0698947321117616e-05, + "loss": 3.6643, + "step": 207050 + }, + { + "epoch": 2.32666565556136, + "grad_norm": 1.043171763420105, + "learning_rate": 1.0696700407814767e-05, + "loss": 3.65, + "step": 207100 + }, + { + "epoch": 2.327227380731703, + "grad_norm": 1.255172610282898, + "learning_rate": 1.0694453494511916e-05, + "loss": 3.6703, + "step": 207150 + }, + { + "epoch": 2.3277891059020464, + "grad_norm": 1.2360597848892212, + "learning_rate": 1.0692206581209065e-05, + "loss": 3.6407, + "step": 207200 + }, + { + "epoch": 2.3283508310723895, + "grad_norm": 1.3001174926757812, + "learning_rate": 1.0689959667906216e-05, + "loss": 3.6984, + "step": 207250 + }, + { + "epoch": 2.3289125562427326, + "grad_norm": 1.0854785442352295, + "learning_rate": 1.0687712754603366e-05, + "loss": 3.5829, + "step": 207300 + }, + { + "epoch": 2.3294742814130758, + "grad_norm": 1.53227698802948, + "learning_rate": 1.0685465841300515e-05, + "loss": 3.6344, + "step": 207350 + }, + { + "epoch": 2.330036006583419, + "grad_norm": 1.0187820196151733, + "learning_rate": 1.0683218927997664e-05, + "loss": 3.6426, + "step": 207400 + }, + { + "epoch": 2.330597731753762, + "grad_norm": 1.227591633796692, + "learning_rate": 1.0680972014694815e-05, + "loss": 3.6018, + "step": 207450 + }, + { + "epoch": 2.331159456924105, + "grad_norm": 1.2064884901046753, + "learning_rate": 1.0678725101391965e-05, + "loss": 3.6709, + "step": 207500 + }, + { + "epoch": 2.3317211820944483, + "grad_norm": 1.5153018236160278, + "learning_rate": 1.0676478188089114e-05, + "loss": 3.6663, + "step": 207550 + }, + { + "epoch": 2.3322829072647915, + "grad_norm": 0.9832055568695068, + "learning_rate": 1.0674231274786263e-05, + "loss": 3.6747, + "step": 207600 + }, + { + "epoch": 2.3328446324351346, + "grad_norm": 1.329431176185608, + "learning_rate": 1.0671984361483414e-05, + "loss": 3.6594, + "step": 207650 + }, + { + "epoch": 2.3334063576054778, + "grad_norm": 1.0778201818466187, + "learning_rate": 1.0669737448180564e-05, + "loss": 3.6171, + "step": 207700 + }, + { + "epoch": 2.3339680827758214, + "grad_norm": 1.1410647630691528, + "learning_rate": 1.0667490534877713e-05, + "loss": 3.64, + "step": 207750 + }, + { + "epoch": 2.334529807946164, + "grad_norm": 1.1544653177261353, + "learning_rate": 1.0665243621574864e-05, + "loss": 3.6708, + "step": 207800 + }, + { + "epoch": 2.3350915331165076, + "grad_norm": 1.1542446613311768, + "learning_rate": 1.0662996708272013e-05, + "loss": 3.6195, + "step": 207850 + }, + { + "epoch": 2.3356532582868503, + "grad_norm": 0.9414635896682739, + "learning_rate": 1.0660749794969163e-05, + "loss": 3.7031, + "step": 207900 + }, + { + "epoch": 2.336214983457194, + "grad_norm": 1.4521923065185547, + "learning_rate": 1.0658502881666312e-05, + "loss": 3.6119, + "step": 207950 + }, + { + "epoch": 2.336776708627537, + "grad_norm": 1.123447299003601, + "learning_rate": 1.0656255968363463e-05, + "loss": 3.6402, + "step": 208000 + }, + { + "epoch": 2.33733843379788, + "grad_norm": 1.0561689138412476, + "learning_rate": 1.0654009055060612e-05, + "loss": 3.6748, + "step": 208050 + }, + { + "epoch": 2.3379001589682233, + "grad_norm": 1.0808101892471313, + "learning_rate": 1.0651762141757762e-05, + "loss": 3.6409, + "step": 208100 + }, + { + "epoch": 2.3384618841385665, + "grad_norm": 1.194089651107788, + "learning_rate": 1.0649515228454913e-05, + "loss": 3.6534, + "step": 208150 + }, + { + "epoch": 2.3390236093089096, + "grad_norm": 2.9633257389068604, + "learning_rate": 1.0647268315152062e-05, + "loss": 3.6217, + "step": 208200 + }, + { + "epoch": 2.3395853344792528, + "grad_norm": 1.3349636793136597, + "learning_rate": 1.0645021401849211e-05, + "loss": 3.6271, + "step": 208250 + }, + { + "epoch": 2.340147059649596, + "grad_norm": 1.3026089668273926, + "learning_rate": 1.064277448854636e-05, + "loss": 3.6668, + "step": 208300 + }, + { + "epoch": 2.340708784819939, + "grad_norm": 4.803406715393066, + "learning_rate": 1.0640527575243512e-05, + "loss": 3.6234, + "step": 208350 + }, + { + "epoch": 2.341270509990282, + "grad_norm": 1.1711561679840088, + "learning_rate": 1.0638280661940661e-05, + "loss": 3.6529, + "step": 208400 + }, + { + "epoch": 2.3418322351606253, + "grad_norm": 1.007407784461975, + "learning_rate": 1.063603374863781e-05, + "loss": 3.646, + "step": 208450 + }, + { + "epoch": 2.3423939603309685, + "grad_norm": 1.0475504398345947, + "learning_rate": 1.0633786835334961e-05, + "loss": 3.767, + "step": 208500 + }, + { + "epoch": 2.3429556855013116, + "grad_norm": 1.1315566301345825, + "learning_rate": 1.063153992203211e-05, + "loss": 3.6716, + "step": 208550 + }, + { + "epoch": 2.3435174106716548, + "grad_norm": 1.1114345788955688, + "learning_rate": 1.062929300872926e-05, + "loss": 3.6099, + "step": 208600 + }, + { + "epoch": 2.344079135841998, + "grad_norm": 1.3237587213516235, + "learning_rate": 1.062704609542641e-05, + "loss": 3.697, + "step": 208650 + }, + { + "epoch": 2.344640861012341, + "grad_norm": 1.151031255722046, + "learning_rate": 1.062479918212356e-05, + "loss": 3.6629, + "step": 208700 + }, + { + "epoch": 2.345202586182684, + "grad_norm": 1.1777541637420654, + "learning_rate": 1.062255226882071e-05, + "loss": 3.676, + "step": 208750 + }, + { + "epoch": 2.3457643113530273, + "grad_norm": 1.126303791999817, + "learning_rate": 1.0620305355517859e-05, + "loss": 3.554, + "step": 208800 + }, + { + "epoch": 2.3463260365233705, + "grad_norm": 1.1741201877593994, + "learning_rate": 1.061805844221501e-05, + "loss": 3.6529, + "step": 208850 + }, + { + "epoch": 2.3468877616937136, + "grad_norm": 1.0780304670333862, + "learning_rate": 1.061581152891216e-05, + "loss": 3.6431, + "step": 208900 + }, + { + "epoch": 2.3474494868640567, + "grad_norm": 1.1062053442001343, + "learning_rate": 1.0613564615609308e-05, + "loss": 3.6463, + "step": 208950 + }, + { + "epoch": 2.3480112120344, + "grad_norm": 1.2727903127670288, + "learning_rate": 1.0611317702306458e-05, + "loss": 3.6653, + "step": 209000 + }, + { + "epoch": 2.348572937204743, + "grad_norm": 1.1275520324707031, + "learning_rate": 1.0609070789003609e-05, + "loss": 3.6461, + "step": 209050 + }, + { + "epoch": 2.3491346623750866, + "grad_norm": 1.3311638832092285, + "learning_rate": 1.0606823875700758e-05, + "loss": 3.702, + "step": 209100 + }, + { + "epoch": 2.3496963875454293, + "grad_norm": 1.1495685577392578, + "learning_rate": 1.0604576962397907e-05, + "loss": 3.6863, + "step": 209150 + }, + { + "epoch": 2.350258112715773, + "grad_norm": 1.2845324277877808, + "learning_rate": 1.0602330049095057e-05, + "loss": 3.6757, + "step": 209200 + }, + { + "epoch": 2.3508198378861156, + "grad_norm": 5.180753231048584, + "learning_rate": 1.0600083135792208e-05, + "loss": 3.6989, + "step": 209250 + }, + { + "epoch": 2.351381563056459, + "grad_norm": 1.39224374294281, + "learning_rate": 1.0597836222489357e-05, + "loss": 3.5864, + "step": 209300 + }, + { + "epoch": 2.3519432882268023, + "grad_norm": 1.1731806993484497, + "learning_rate": 1.0595589309186506e-05, + "loss": 3.6332, + "step": 209350 + }, + { + "epoch": 2.3525050133971455, + "grad_norm": 1.3183012008666992, + "learning_rate": 1.0593342395883657e-05, + "loss": 3.6614, + "step": 209400 + }, + { + "epoch": 2.3530667385674886, + "grad_norm": 1.3669489622116089, + "learning_rate": 1.0591095482580807e-05, + "loss": 3.6488, + "step": 209450 + }, + { + "epoch": 2.3536284637378317, + "grad_norm": 1.0421273708343506, + "learning_rate": 1.0588848569277956e-05, + "loss": 3.5317, + "step": 209500 + }, + { + "epoch": 2.354190188908175, + "grad_norm": 1.057143211364746, + "learning_rate": 1.0586601655975105e-05, + "loss": 3.6176, + "step": 209550 + }, + { + "epoch": 2.354751914078518, + "grad_norm": 1.0931198596954346, + "learning_rate": 1.0584354742672256e-05, + "loss": 3.6886, + "step": 209600 + }, + { + "epoch": 2.355313639248861, + "grad_norm": 1.4499849081039429, + "learning_rate": 1.0582107829369406e-05, + "loss": 3.7207, + "step": 209650 + }, + { + "epoch": 2.3558753644192043, + "grad_norm": 1.3522409200668335, + "learning_rate": 1.0579860916066555e-05, + "loss": 3.6915, + "step": 209700 + }, + { + "epoch": 2.3564370895895475, + "grad_norm": 1.271995186805725, + "learning_rate": 1.0577614002763706e-05, + "loss": 3.6232, + "step": 209750 + }, + { + "epoch": 2.3569988147598906, + "grad_norm": 1.169212818145752, + "learning_rate": 1.0575367089460855e-05, + "loss": 3.7393, + "step": 209800 + }, + { + "epoch": 2.3575605399302337, + "grad_norm": 1.1130293607711792, + "learning_rate": 1.0573120176158005e-05, + "loss": 3.6652, + "step": 209850 + }, + { + "epoch": 2.358122265100577, + "grad_norm": 1.2910044193267822, + "learning_rate": 1.0570873262855154e-05, + "loss": 3.7016, + "step": 209900 + }, + { + "epoch": 2.35868399027092, + "grad_norm": 1.1066601276397705, + "learning_rate": 1.0568626349552305e-05, + "loss": 3.6889, + "step": 209950 + }, + { + "epoch": 2.359245715441263, + "grad_norm": 1.2973527908325195, + "learning_rate": 1.0566379436249454e-05, + "loss": 3.6981, + "step": 210000 + }, + { + "epoch": 2.3598074406116063, + "grad_norm": 1.340416431427002, + "learning_rate": 1.0564132522946604e-05, + "loss": 3.5788, + "step": 210050 + }, + { + "epoch": 2.3603691657819494, + "grad_norm": 1.1649587154388428, + "learning_rate": 1.0561885609643755e-05, + "loss": 3.6245, + "step": 210100 + }, + { + "epoch": 2.3609308909522926, + "grad_norm": 1.2052050828933716, + "learning_rate": 1.0559638696340902e-05, + "loss": 3.7734, + "step": 210150 + }, + { + "epoch": 2.3614926161226357, + "grad_norm": 1.1914728879928589, + "learning_rate": 1.0557391783038052e-05, + "loss": 3.6131, + "step": 210200 + }, + { + "epoch": 2.362054341292979, + "grad_norm": 1.1473121643066406, + "learning_rate": 1.0555144869735201e-05, + "loss": 3.6201, + "step": 210250 + }, + { + "epoch": 2.362616066463322, + "grad_norm": 1.0467044115066528, + "learning_rate": 1.055289795643235e-05, + "loss": 3.6894, + "step": 210300 + }, + { + "epoch": 2.363177791633665, + "grad_norm": 1.103670358657837, + "learning_rate": 1.0550651043129501e-05, + "loss": 3.6476, + "step": 210350 + }, + { + "epoch": 2.3637395168040083, + "grad_norm": 1.2465933561325073, + "learning_rate": 1.054840412982665e-05, + "loss": 3.5981, + "step": 210400 + }, + { + "epoch": 2.364301241974352, + "grad_norm": 1.1501760482788086, + "learning_rate": 1.05461572165238e-05, + "loss": 3.6431, + "step": 210450 + }, + { + "epoch": 2.3648629671446946, + "grad_norm": 1.0407079458236694, + "learning_rate": 1.054391030322095e-05, + "loss": 3.6348, + "step": 210500 + }, + { + "epoch": 2.365424692315038, + "grad_norm": 1.2164947986602783, + "learning_rate": 1.05416633899181e-05, + "loss": 3.6887, + "step": 210550 + }, + { + "epoch": 2.365986417485381, + "grad_norm": 1.007479190826416, + "learning_rate": 1.053941647661525e-05, + "loss": 3.5831, + "step": 210600 + }, + { + "epoch": 2.3665481426557244, + "grad_norm": 1.0290968418121338, + "learning_rate": 1.0537169563312399e-05, + "loss": 3.5973, + "step": 210650 + }, + { + "epoch": 2.3671098678260676, + "grad_norm": 1.297382116317749, + "learning_rate": 1.053492265000955e-05, + "loss": 3.7142, + "step": 210700 + }, + { + "epoch": 2.3676715929964107, + "grad_norm": 1.215598464012146, + "learning_rate": 1.0532675736706699e-05, + "loss": 3.6544, + "step": 210750 + }, + { + "epoch": 2.368233318166754, + "grad_norm": 1.5987659692764282, + "learning_rate": 1.0530428823403848e-05, + "loss": 3.6862, + "step": 210800 + }, + { + "epoch": 2.368795043337097, + "grad_norm": 1.0816415548324585, + "learning_rate": 1.0528181910100998e-05, + "loss": 3.5947, + "step": 210850 + }, + { + "epoch": 2.36935676850744, + "grad_norm": 1.255016565322876, + "learning_rate": 1.0525934996798149e-05, + "loss": 3.5878, + "step": 210900 + }, + { + "epoch": 2.3699184936777833, + "grad_norm": 1.2578915357589722, + "learning_rate": 1.0523688083495298e-05, + "loss": 3.6189, + "step": 210950 + }, + { + "epoch": 2.3704802188481264, + "grad_norm": 1.1681090593338013, + "learning_rate": 1.0521441170192447e-05, + "loss": 3.6204, + "step": 211000 + }, + { + "epoch": 2.3710419440184696, + "grad_norm": 1.3190033435821533, + "learning_rate": 1.0519194256889598e-05, + "loss": 3.6094, + "step": 211050 + }, + { + "epoch": 2.3716036691888127, + "grad_norm": 1.0447425842285156, + "learning_rate": 1.0516947343586748e-05, + "loss": 3.6688, + "step": 211100 + }, + { + "epoch": 2.372165394359156, + "grad_norm": 1.2852200269699097, + "learning_rate": 1.0514700430283897e-05, + "loss": 3.639, + "step": 211150 + }, + { + "epoch": 2.372727119529499, + "grad_norm": 1.0362699031829834, + "learning_rate": 1.0512453516981046e-05, + "loss": 3.6384, + "step": 211200 + }, + { + "epoch": 2.373288844699842, + "grad_norm": 1.2051905393600464, + "learning_rate": 1.0510206603678197e-05, + "loss": 3.6728, + "step": 211250 + }, + { + "epoch": 2.3738505698701853, + "grad_norm": 1.088995337486267, + "learning_rate": 1.0507959690375347e-05, + "loss": 3.6041, + "step": 211300 + }, + { + "epoch": 2.3744122950405284, + "grad_norm": 1.1478296518325806, + "learning_rate": 1.0505712777072496e-05, + "loss": 3.6943, + "step": 211350 + }, + { + "epoch": 2.3749740202108716, + "grad_norm": 1.334021806716919, + "learning_rate": 1.0503465863769647e-05, + "loss": 3.6199, + "step": 211400 + }, + { + "epoch": 2.3755357453812147, + "grad_norm": 1.2082432508468628, + "learning_rate": 1.0501218950466796e-05, + "loss": 3.7031, + "step": 211450 + }, + { + "epoch": 2.376097470551558, + "grad_norm": 1.133233666419983, + "learning_rate": 1.0498972037163946e-05, + "loss": 3.7362, + "step": 211500 + }, + { + "epoch": 2.376659195721901, + "grad_norm": 1.2972453832626343, + "learning_rate": 1.0496725123861095e-05, + "loss": 3.6044, + "step": 211550 + }, + { + "epoch": 2.377220920892244, + "grad_norm": 1.0553796291351318, + "learning_rate": 1.0494478210558246e-05, + "loss": 3.6411, + "step": 211600 + }, + { + "epoch": 2.3777826460625873, + "grad_norm": 1.4199177026748657, + "learning_rate": 1.0492231297255395e-05, + "loss": 3.6847, + "step": 211650 + }, + { + "epoch": 2.3783443712329304, + "grad_norm": 1.332186222076416, + "learning_rate": 1.0489984383952545e-05, + "loss": 3.645, + "step": 211700 + }, + { + "epoch": 2.3789060964032736, + "grad_norm": 1.034148097038269, + "learning_rate": 1.0487737470649696e-05, + "loss": 3.6861, + "step": 211750 + }, + { + "epoch": 2.379467821573617, + "grad_norm": 1.1468724012374878, + "learning_rate": 1.0485490557346845e-05, + "loss": 3.6651, + "step": 211800 + }, + { + "epoch": 2.38002954674396, + "grad_norm": 1.186020851135254, + "learning_rate": 1.0483243644043994e-05, + "loss": 3.6186, + "step": 211850 + }, + { + "epoch": 2.3805912719143034, + "grad_norm": 1.140351414680481, + "learning_rate": 1.0480996730741144e-05, + "loss": 3.6236, + "step": 211900 + }, + { + "epoch": 2.381152997084646, + "grad_norm": 1.1734734773635864, + "learning_rate": 1.0478749817438295e-05, + "loss": 3.6766, + "step": 211950 + }, + { + "epoch": 2.3817147222549897, + "grad_norm": 1.1378458738327026, + "learning_rate": 1.0476502904135444e-05, + "loss": 3.6499, + "step": 212000 + }, + { + "epoch": 2.382276447425333, + "grad_norm": 1.2333191633224487, + "learning_rate": 1.0474255990832593e-05, + "loss": 3.7074, + "step": 212050 + }, + { + "epoch": 2.382838172595676, + "grad_norm": 1.0585025548934937, + "learning_rate": 1.0472009077529744e-05, + "loss": 3.6026, + "step": 212100 + }, + { + "epoch": 2.383399897766019, + "grad_norm": 1.2075213193893433, + "learning_rate": 1.0469762164226894e-05, + "loss": 3.7076, + "step": 212150 + }, + { + "epoch": 2.3839616229363623, + "grad_norm": 1.148812174797058, + "learning_rate": 1.0467515250924043e-05, + "loss": 3.6372, + "step": 212200 + }, + { + "epoch": 2.3845233481067054, + "grad_norm": 1.156119704246521, + "learning_rate": 1.0465268337621192e-05, + "loss": 3.6752, + "step": 212250 + }, + { + "epoch": 2.3850850732770486, + "grad_norm": 1.3167558908462524, + "learning_rate": 1.0463021424318343e-05, + "loss": 3.653, + "step": 212300 + }, + { + "epoch": 2.3856467984473917, + "grad_norm": 1.27252995967865, + "learning_rate": 1.0460774511015492e-05, + "loss": 3.6471, + "step": 212350 + }, + { + "epoch": 2.386208523617735, + "grad_norm": 1.3370920419692993, + "learning_rate": 1.0458527597712642e-05, + "loss": 3.6483, + "step": 212400 + }, + { + "epoch": 2.386770248788078, + "grad_norm": 1.3032448291778564, + "learning_rate": 1.0456280684409791e-05, + "loss": 3.6389, + "step": 212450 + }, + { + "epoch": 2.387331973958421, + "grad_norm": 1.0964500904083252, + "learning_rate": 1.0454033771106942e-05, + "loss": 3.6809, + "step": 212500 + }, + { + "epoch": 2.3878936991287643, + "grad_norm": 1.074172019958496, + "learning_rate": 1.0451786857804091e-05, + "loss": 3.7075, + "step": 212550 + }, + { + "epoch": 2.3884554242991074, + "grad_norm": 1.2807329893112183, + "learning_rate": 1.044953994450124e-05, + "loss": 3.6838, + "step": 212600 + }, + { + "epoch": 2.3890171494694505, + "grad_norm": 1.1744275093078613, + "learning_rate": 1.0447293031198392e-05, + "loss": 3.6299, + "step": 212650 + }, + { + "epoch": 2.3895788746397937, + "grad_norm": 1.3053351640701294, + "learning_rate": 1.0445046117895541e-05, + "loss": 3.6463, + "step": 212700 + }, + { + "epoch": 2.390140599810137, + "grad_norm": 1.2043471336364746, + "learning_rate": 1.044279920459269e-05, + "loss": 3.6402, + "step": 212750 + }, + { + "epoch": 2.39070232498048, + "grad_norm": 1.0350993871688843, + "learning_rate": 1.044055229128984e-05, + "loss": 3.6565, + "step": 212800 + }, + { + "epoch": 2.391264050150823, + "grad_norm": 1.1305289268493652, + "learning_rate": 1.0438350316253047e-05, + "loss": 3.7486, + "step": 212850 + }, + { + "epoch": 2.3918257753211662, + "grad_norm": 1.2447079420089722, + "learning_rate": 1.0436103402950197e-05, + "loss": 3.6249, + "step": 212900 + }, + { + "epoch": 2.3923875004915094, + "grad_norm": 1.3462215662002563, + "learning_rate": 1.0433856489647348e-05, + "loss": 3.6136, + "step": 212950 + }, + { + "epoch": 2.3929492256618525, + "grad_norm": 1.3543200492858887, + "learning_rate": 1.0431609576344497e-05, + "loss": 3.6838, + "step": 213000 + }, + { + "epoch": 2.3935109508321957, + "grad_norm": 1.217747449874878, + "learning_rate": 1.0429362663041646e-05, + "loss": 3.6542, + "step": 213050 + }, + { + "epoch": 2.394072676002539, + "grad_norm": 1.1525721549987793, + "learning_rate": 1.0427115749738797e-05, + "loss": 3.6055, + "step": 213100 + }, + { + "epoch": 2.3946344011728824, + "grad_norm": 1.1056920289993286, + "learning_rate": 1.0424868836435947e-05, + "loss": 3.6782, + "step": 213150 + }, + { + "epoch": 2.395196126343225, + "grad_norm": 0.9926613569259644, + "learning_rate": 1.0422621923133096e-05, + "loss": 3.6243, + "step": 213200 + }, + { + "epoch": 2.3957578515135687, + "grad_norm": 1.0209110975265503, + "learning_rate": 1.0420419948096304e-05, + "loss": 3.6288, + "step": 213250 + }, + { + "epoch": 2.396319576683912, + "grad_norm": 1.0990540981292725, + "learning_rate": 1.0418173034793453e-05, + "loss": 3.6718, + "step": 213300 + }, + { + "epoch": 2.396881301854255, + "grad_norm": 1.2863000631332397, + "learning_rate": 1.0415926121490602e-05, + "loss": 3.6767, + "step": 213350 + }, + { + "epoch": 2.397443027024598, + "grad_norm": 1.0847902297973633, + "learning_rate": 1.0413679208187753e-05, + "loss": 3.671, + "step": 213400 + }, + { + "epoch": 2.3980047521949412, + "grad_norm": 1.220410943031311, + "learning_rate": 1.0411432294884903e-05, + "loss": 3.5818, + "step": 213450 + }, + { + "epoch": 2.3985664773652844, + "grad_norm": 2.0756099224090576, + "learning_rate": 1.0409185381582052e-05, + "loss": 3.6215, + "step": 213500 + }, + { + "epoch": 2.3991282025356275, + "grad_norm": 1.3049434423446655, + "learning_rate": 1.0406938468279201e-05, + "loss": 3.6558, + "step": 213550 + }, + { + "epoch": 2.3996899277059707, + "grad_norm": 1.2601306438446045, + "learning_rate": 1.0404691554976352e-05, + "loss": 3.5866, + "step": 213600 + }, + { + "epoch": 2.400251652876314, + "grad_norm": 1.2760636806488037, + "learning_rate": 1.0402444641673501e-05, + "loss": 3.6674, + "step": 213650 + }, + { + "epoch": 2.400813378046657, + "grad_norm": 1.0744049549102783, + "learning_rate": 1.040019772837065e-05, + "loss": 3.6744, + "step": 213700 + }, + { + "epoch": 2.401375103217, + "grad_norm": 1.2868483066558838, + "learning_rate": 1.0397950815067802e-05, + "loss": 3.6368, + "step": 213750 + }, + { + "epoch": 2.4019368283873432, + "grad_norm": 1.3159387111663818, + "learning_rate": 1.0395703901764951e-05, + "loss": 3.6556, + "step": 213800 + }, + { + "epoch": 2.4024985535576864, + "grad_norm": 1.1377818584442139, + "learning_rate": 1.03934569884621e-05, + "loss": 3.6512, + "step": 213850 + }, + { + "epoch": 2.4030602787280295, + "grad_norm": 1.1791696548461914, + "learning_rate": 1.039121007515925e-05, + "loss": 3.6349, + "step": 213900 + }, + { + "epoch": 2.4036220038983727, + "grad_norm": 1.1945469379425049, + "learning_rate": 1.03889631618564e-05, + "loss": 3.6634, + "step": 213950 + }, + { + "epoch": 2.404183729068716, + "grad_norm": 1.2033038139343262, + "learning_rate": 1.038671624855355e-05, + "loss": 3.639, + "step": 214000 + }, + { + "epoch": 2.404745454239059, + "grad_norm": 1.6592442989349365, + "learning_rate": 1.03844693352507e-05, + "loss": 3.6357, + "step": 214050 + }, + { + "epoch": 2.405307179409402, + "grad_norm": 1.1835962533950806, + "learning_rate": 1.038222242194785e-05, + "loss": 3.6082, + "step": 214100 + }, + { + "epoch": 2.4058689045797452, + "grad_norm": 1.2160457372665405, + "learning_rate": 1.0379975508645e-05, + "loss": 3.6424, + "step": 214150 + }, + { + "epoch": 2.4064306297500884, + "grad_norm": 1.13961923122406, + "learning_rate": 1.0377773533608206e-05, + "loss": 3.6699, + "step": 214200 + }, + { + "epoch": 2.4069923549204315, + "grad_norm": 1.2178468704223633, + "learning_rate": 1.0375526620305357e-05, + "loss": 3.6302, + "step": 214250 + }, + { + "epoch": 2.4075540800907747, + "grad_norm": 1.180449366569519, + "learning_rate": 1.0373279707002506e-05, + "loss": 3.6606, + "step": 214300 + }, + { + "epoch": 2.408115805261118, + "grad_norm": 1.0457371473312378, + "learning_rate": 1.0371032793699655e-05, + "loss": 3.6764, + "step": 214350 + }, + { + "epoch": 2.408677530431461, + "grad_norm": 1.032880425453186, + "learning_rate": 1.0368785880396806e-05, + "loss": 3.6936, + "step": 214400 + }, + { + "epoch": 2.409239255601804, + "grad_norm": 1.0636664628982544, + "learning_rate": 1.0366538967093956e-05, + "loss": 3.605, + "step": 214450 + }, + { + "epoch": 2.4098009807721477, + "grad_norm": 1.0582681894302368, + "learning_rate": 1.0364292053791105e-05, + "loss": 3.7026, + "step": 214500 + }, + { + "epoch": 2.4103627059424904, + "grad_norm": 1.1255106925964355, + "learning_rate": 1.0362045140488254e-05, + "loss": 3.6631, + "step": 214550 + }, + { + "epoch": 2.410924431112834, + "grad_norm": 1.17404305934906, + "learning_rate": 1.0359798227185405e-05, + "loss": 3.6122, + "step": 214600 + }, + { + "epoch": 2.411486156283177, + "grad_norm": 1.4032615423202515, + "learning_rate": 1.0357551313882555e-05, + "loss": 3.6654, + "step": 214650 + }, + { + "epoch": 2.4120478814535202, + "grad_norm": 1.0625853538513184, + "learning_rate": 1.0355304400579704e-05, + "loss": 3.7238, + "step": 214700 + }, + { + "epoch": 2.4126096066238634, + "grad_norm": 1.2413190603256226, + "learning_rate": 1.0353057487276855e-05, + "loss": 3.7059, + "step": 214750 + }, + { + "epoch": 2.4131713317942065, + "grad_norm": 0.9958818554878235, + "learning_rate": 1.0350810573974004e-05, + "loss": 3.6565, + "step": 214800 + }, + { + "epoch": 2.4137330569645497, + "grad_norm": 1.2933051586151123, + "learning_rate": 1.0348563660671154e-05, + "loss": 3.649, + "step": 214850 + }, + { + "epoch": 2.414294782134893, + "grad_norm": 1.0213230848312378, + "learning_rate": 1.0346316747368303e-05, + "loss": 3.6311, + "step": 214900 + }, + { + "epoch": 2.414856507305236, + "grad_norm": 1.2674108743667603, + "learning_rate": 1.0344069834065454e-05, + "loss": 3.7243, + "step": 214950 + }, + { + "epoch": 2.415418232475579, + "grad_norm": 1.2302976846694946, + "learning_rate": 1.0341822920762603e-05, + "loss": 3.7059, + "step": 215000 + }, + { + "epoch": 2.415979957645922, + "grad_norm": 1.2191627025604248, + "learning_rate": 1.0339576007459753e-05, + "loss": 3.6319, + "step": 215050 + }, + { + "epoch": 2.4165416828162654, + "grad_norm": 1.215530514717102, + "learning_rate": 1.0337329094156902e-05, + "loss": 3.6553, + "step": 215100 + }, + { + "epoch": 2.4171034079866085, + "grad_norm": 1.2192349433898926, + "learning_rate": 1.0335082180854053e-05, + "loss": 3.6413, + "step": 215150 + }, + { + "epoch": 2.4176651331569516, + "grad_norm": 1.132501244544983, + "learning_rate": 1.0332835267551202e-05, + "loss": 3.6787, + "step": 215200 + }, + { + "epoch": 2.418226858327295, + "grad_norm": 1.065586805343628, + "learning_rate": 1.0330588354248351e-05, + "loss": 3.5863, + "step": 215250 + }, + { + "epoch": 2.418788583497638, + "grad_norm": 1.0597398281097412, + "learning_rate": 1.0328341440945502e-05, + "loss": 3.6315, + "step": 215300 + }, + { + "epoch": 2.419350308667981, + "grad_norm": 1.0700304508209229, + "learning_rate": 1.0326094527642652e-05, + "loss": 3.7679, + "step": 215350 + }, + { + "epoch": 2.419912033838324, + "grad_norm": 1.3021875619888306, + "learning_rate": 1.0323847614339801e-05, + "loss": 3.7039, + "step": 215400 + }, + { + "epoch": 2.4204737590086673, + "grad_norm": 1.3365675210952759, + "learning_rate": 1.032160070103695e-05, + "loss": 3.6874, + "step": 215450 + }, + { + "epoch": 2.4210354841790105, + "grad_norm": 1.261286735534668, + "learning_rate": 1.0319353787734101e-05, + "loss": 3.599, + "step": 215500 + }, + { + "epoch": 2.4215972093493536, + "grad_norm": 1.128183364868164, + "learning_rate": 1.031710687443125e-05, + "loss": 3.6616, + "step": 215550 + }, + { + "epoch": 2.4221589345196968, + "grad_norm": 1.0821489095687866, + "learning_rate": 1.03148599611284e-05, + "loss": 3.6034, + "step": 215600 + }, + { + "epoch": 2.42272065969004, + "grad_norm": 1.2359782457351685, + "learning_rate": 1.0312613047825551e-05, + "loss": 3.6662, + "step": 215650 + }, + { + "epoch": 2.423282384860383, + "grad_norm": 1.2059345245361328, + "learning_rate": 1.03103661345227e-05, + "loss": 3.5647, + "step": 215700 + }, + { + "epoch": 2.423844110030726, + "grad_norm": 1.7044291496276855, + "learning_rate": 1.030811922121985e-05, + "loss": 3.6759, + "step": 215750 + }, + { + "epoch": 2.4244058352010693, + "grad_norm": 1.320499062538147, + "learning_rate": 1.0305872307916999e-05, + "loss": 3.6642, + "step": 215800 + }, + { + "epoch": 2.424967560371413, + "grad_norm": 1.1735167503356934, + "learning_rate": 1.030362539461415e-05, + "loss": 3.6223, + "step": 215850 + }, + { + "epoch": 2.4255292855417556, + "grad_norm": 1.0095165967941284, + "learning_rate": 1.03013784813113e-05, + "loss": 3.6482, + "step": 215900 + }, + { + "epoch": 2.426091010712099, + "grad_norm": 1.0431257486343384, + "learning_rate": 1.0299131568008449e-05, + "loss": 3.5853, + "step": 215950 + }, + { + "epoch": 2.4266527358824423, + "grad_norm": 1.2413547039031982, + "learning_rate": 1.02968846547056e-05, + "loss": 3.6254, + "step": 216000 + }, + { + "epoch": 2.4272144610527855, + "grad_norm": 1.143583059310913, + "learning_rate": 1.0294637741402749e-05, + "loss": 3.6449, + "step": 216050 + }, + { + "epoch": 2.4277761862231286, + "grad_norm": 1.152782678604126, + "learning_rate": 1.0292390828099898e-05, + "loss": 3.6702, + "step": 216100 + }, + { + "epoch": 2.4283379113934718, + "grad_norm": 1.3686628341674805, + "learning_rate": 1.0290143914797048e-05, + "loss": 3.6322, + "step": 216150 + }, + { + "epoch": 2.428899636563815, + "grad_norm": 1.1874475479125977, + "learning_rate": 1.0287897001494199e-05, + "loss": 3.6287, + "step": 216200 + }, + { + "epoch": 2.429461361734158, + "grad_norm": 1.4857885837554932, + "learning_rate": 1.0285650088191348e-05, + "loss": 3.6677, + "step": 216250 + }, + { + "epoch": 2.430023086904501, + "grad_norm": 1.2705439329147339, + "learning_rate": 1.0283403174888497e-05, + "loss": 3.6905, + "step": 216300 + }, + { + "epoch": 2.4305848120748443, + "grad_norm": 1.2015371322631836, + "learning_rate": 1.0281156261585648e-05, + "loss": 3.6073, + "step": 216350 + }, + { + "epoch": 2.4311465372451875, + "grad_norm": 1.1872402429580688, + "learning_rate": 1.0278909348282798e-05, + "loss": 3.6097, + "step": 216400 + }, + { + "epoch": 2.4317082624155306, + "grad_norm": 1.4883981943130493, + "learning_rate": 1.0276662434979947e-05, + "loss": 3.6469, + "step": 216450 + }, + { + "epoch": 2.4322699875858738, + "grad_norm": 1.1034952402114868, + "learning_rate": 1.0274415521677096e-05, + "loss": 3.7157, + "step": 216500 + }, + { + "epoch": 2.432831712756217, + "grad_norm": 1.0829899311065674, + "learning_rate": 1.0272168608374247e-05, + "loss": 3.6375, + "step": 216550 + }, + { + "epoch": 2.43339343792656, + "grad_norm": 1.2546882629394531, + "learning_rate": 1.0269921695071397e-05, + "loss": 3.6124, + "step": 216600 + }, + { + "epoch": 2.433955163096903, + "grad_norm": 1.122017741203308, + "learning_rate": 1.0267674781768546e-05, + "loss": 3.6131, + "step": 216650 + }, + { + "epoch": 2.4345168882672463, + "grad_norm": 1.322969675064087, + "learning_rate": 1.0265427868465697e-05, + "loss": 3.6663, + "step": 216700 + }, + { + "epoch": 2.4350786134375895, + "grad_norm": 1.2598247528076172, + "learning_rate": 1.0263180955162846e-05, + "loss": 3.6465, + "step": 216750 + }, + { + "epoch": 2.4356403386079326, + "grad_norm": 1.169589877128601, + "learning_rate": 1.0260934041859996e-05, + "loss": 3.6928, + "step": 216800 + }, + { + "epoch": 2.4362020637782758, + "grad_norm": 1.1836838722229004, + "learning_rate": 1.0258687128557145e-05, + "loss": 3.7588, + "step": 216850 + }, + { + "epoch": 2.436763788948619, + "grad_norm": 1.3690922260284424, + "learning_rate": 1.0256440215254296e-05, + "loss": 3.7039, + "step": 216900 + }, + { + "epoch": 2.437325514118962, + "grad_norm": 1.2842140197753906, + "learning_rate": 1.0254193301951445e-05, + "loss": 3.5944, + "step": 216950 + }, + { + "epoch": 2.437887239289305, + "grad_norm": 1.17047119140625, + "learning_rate": 1.0251946388648594e-05, + "loss": 3.6459, + "step": 217000 + }, + { + "epoch": 2.4384489644596483, + "grad_norm": 1.1833974123001099, + "learning_rate": 1.0249699475345744e-05, + "loss": 3.6239, + "step": 217050 + }, + { + "epoch": 2.4390106896299915, + "grad_norm": 1.0102674961090088, + "learning_rate": 1.0247452562042895e-05, + "loss": 3.6602, + "step": 217100 + }, + { + "epoch": 2.4395724148003346, + "grad_norm": 1.1659404039382935, + "learning_rate": 1.0245205648740044e-05, + "loss": 3.6185, + "step": 217150 + }, + { + "epoch": 2.440134139970678, + "grad_norm": 1.080694317817688, + "learning_rate": 1.0242958735437193e-05, + "loss": 3.7509, + "step": 217200 + }, + { + "epoch": 2.440695865141021, + "grad_norm": 1.0265114307403564, + "learning_rate": 1.0240711822134344e-05, + "loss": 3.6257, + "step": 217250 + }, + { + "epoch": 2.4412575903113645, + "grad_norm": 1.1171821355819702, + "learning_rate": 1.0238464908831494e-05, + "loss": 3.6605, + "step": 217300 + }, + { + "epoch": 2.4418193154817076, + "grad_norm": 1.123995304107666, + "learning_rate": 1.0236217995528643e-05, + "loss": 3.6906, + "step": 217350 + }, + { + "epoch": 2.4423810406520507, + "grad_norm": 1.3015533685684204, + "learning_rate": 1.0233971082225792e-05, + "loss": 3.6372, + "step": 217400 + }, + { + "epoch": 2.442942765822394, + "grad_norm": 1.2588542699813843, + "learning_rate": 1.0231724168922943e-05, + "loss": 3.6604, + "step": 217450 + }, + { + "epoch": 2.443504490992737, + "grad_norm": 1.0011076927185059, + "learning_rate": 1.0229477255620093e-05, + "loss": 3.62, + "step": 217500 + }, + { + "epoch": 2.44406621616308, + "grad_norm": 1.1005232334136963, + "learning_rate": 1.0227230342317242e-05, + "loss": 3.6412, + "step": 217550 + }, + { + "epoch": 2.4446279413334233, + "grad_norm": 1.3701766729354858, + "learning_rate": 1.0224983429014393e-05, + "loss": 3.682, + "step": 217600 + }, + { + "epoch": 2.4451896665037665, + "grad_norm": 1.1460520029067993, + "learning_rate": 1.0222736515711542e-05, + "loss": 3.6485, + "step": 217650 + }, + { + "epoch": 2.4457513916741096, + "grad_norm": 1.1530455350875854, + "learning_rate": 1.0220489602408692e-05, + "loss": 3.6716, + "step": 217700 + }, + { + "epoch": 2.4463131168444527, + "grad_norm": 1.154485821723938, + "learning_rate": 1.0218242689105841e-05, + "loss": 3.6729, + "step": 217750 + }, + { + "epoch": 2.446874842014796, + "grad_norm": 1.321520209312439, + "learning_rate": 1.0215995775802992e-05, + "loss": 3.582, + "step": 217800 + }, + { + "epoch": 2.447436567185139, + "grad_norm": 1.4846552610397339, + "learning_rate": 1.0213748862500141e-05, + "loss": 3.7066, + "step": 217850 + }, + { + "epoch": 2.447998292355482, + "grad_norm": 2.0846121311187744, + "learning_rate": 1.021150194919729e-05, + "loss": 3.6042, + "step": 217900 + }, + { + "epoch": 2.4485600175258253, + "grad_norm": 1.1489448547363281, + "learning_rate": 1.0209255035894442e-05, + "loss": 3.6159, + "step": 217950 + }, + { + "epoch": 2.4491217426961684, + "grad_norm": 1.273450493812561, + "learning_rate": 1.0207008122591591e-05, + "loss": 3.5722, + "step": 218000 + }, + { + "epoch": 2.4496834678665116, + "grad_norm": 1.0643285512924194, + "learning_rate": 1.020476120928874e-05, + "loss": 3.6386, + "step": 218050 + }, + { + "epoch": 2.4502451930368547, + "grad_norm": 1.552664875984192, + "learning_rate": 1.020251429598589e-05, + "loss": 3.6878, + "step": 218100 + }, + { + "epoch": 2.450806918207198, + "grad_norm": 1.3038806915283203, + "learning_rate": 1.020026738268304e-05, + "loss": 3.6344, + "step": 218150 + }, + { + "epoch": 2.451368643377541, + "grad_norm": 1.5481359958648682, + "learning_rate": 1.019802046938019e-05, + "loss": 3.634, + "step": 218200 + }, + { + "epoch": 2.451930368547884, + "grad_norm": 0.9933804273605347, + "learning_rate": 1.019577355607734e-05, + "loss": 3.5348, + "step": 218250 + }, + { + "epoch": 2.4524920937182273, + "grad_norm": 0.9886470437049866, + "learning_rate": 1.019352664277449e-05, + "loss": 3.6177, + "step": 218300 + }, + { + "epoch": 2.4530538188885704, + "grad_norm": 1.268970251083374, + "learning_rate": 1.019127972947164e-05, + "loss": 3.7232, + "step": 218350 + }, + { + "epoch": 2.4536155440589136, + "grad_norm": 1.159010648727417, + "learning_rate": 1.0189032816168789e-05, + "loss": 3.7073, + "step": 218400 + }, + { + "epoch": 2.4541772692292567, + "grad_norm": 1.1107714176177979, + "learning_rate": 1.0186785902865938e-05, + "loss": 3.5887, + "step": 218450 + }, + { + "epoch": 2.4547389943996, + "grad_norm": 1.1322340965270996, + "learning_rate": 1.018453898956309e-05, + "loss": 3.6622, + "step": 218500 + }, + { + "epoch": 2.4553007195699434, + "grad_norm": 1.1565299034118652, + "learning_rate": 1.0182292076260239e-05, + "loss": 3.6327, + "step": 218550 + }, + { + "epoch": 2.455862444740286, + "grad_norm": 1.2363173961639404, + "learning_rate": 1.0180045162957388e-05, + "loss": 3.6067, + "step": 218600 + }, + { + "epoch": 2.4564241699106297, + "grad_norm": 1.071761965751648, + "learning_rate": 1.0177798249654537e-05, + "loss": 3.6748, + "step": 218650 + }, + { + "epoch": 2.456985895080973, + "grad_norm": 1.1876356601715088, + "learning_rate": 1.0175551336351688e-05, + "loss": 3.7111, + "step": 218700 + }, + { + "epoch": 2.457547620251316, + "grad_norm": 1.0165085792541504, + "learning_rate": 1.0173304423048838e-05, + "loss": 3.7227, + "step": 218750 + }, + { + "epoch": 2.458109345421659, + "grad_norm": 1.1822094917297363, + "learning_rate": 1.0171102448012045e-05, + "loss": 3.6007, + "step": 218800 + }, + { + "epoch": 2.4586710705920023, + "grad_norm": 1.026947021484375, + "learning_rate": 1.0168900472975251e-05, + "loss": 3.6632, + "step": 218850 + }, + { + "epoch": 2.4592327957623454, + "grad_norm": 1.3460332155227661, + "learning_rate": 1.0166653559672402e-05, + "loss": 3.6523, + "step": 218900 + }, + { + "epoch": 2.4597945209326886, + "grad_norm": 1.2409924268722534, + "learning_rate": 1.0164406646369551e-05, + "loss": 3.6089, + "step": 218950 + }, + { + "epoch": 2.4603562461030317, + "grad_norm": 1.1835811138153076, + "learning_rate": 1.01621597330667e-05, + "loss": 3.6637, + "step": 219000 + }, + { + "epoch": 2.460917971273375, + "grad_norm": 1.5220344066619873, + "learning_rate": 1.015991281976385e-05, + "loss": 3.6492, + "step": 219050 + }, + { + "epoch": 2.461479696443718, + "grad_norm": 1.036289930343628, + "learning_rate": 1.0157665906461001e-05, + "loss": 3.6771, + "step": 219100 + }, + { + "epoch": 2.462041421614061, + "grad_norm": 1.3220967054367065, + "learning_rate": 1.015541899315815e-05, + "loss": 3.6313, + "step": 219150 + }, + { + "epoch": 2.4626031467844043, + "grad_norm": 1.1639056205749512, + "learning_rate": 1.01531720798553e-05, + "loss": 3.6877, + "step": 219200 + }, + { + "epoch": 2.4631648719547474, + "grad_norm": 1.1655687093734741, + "learning_rate": 1.015092516655245e-05, + "loss": 3.6249, + "step": 219250 + }, + { + "epoch": 2.4637265971250906, + "grad_norm": 1.2269138097763062, + "learning_rate": 1.01486782532496e-05, + "loss": 3.6885, + "step": 219300 + }, + { + "epoch": 2.4642883222954337, + "grad_norm": 1.293755292892456, + "learning_rate": 1.014643133994675e-05, + "loss": 3.5937, + "step": 219350 + }, + { + "epoch": 2.464850047465777, + "grad_norm": 1.0094069242477417, + "learning_rate": 1.0144184426643899e-05, + "loss": 3.6618, + "step": 219400 + }, + { + "epoch": 2.46541177263612, + "grad_norm": 1.3151395320892334, + "learning_rate": 1.014193751334105e-05, + "loss": 3.6901, + "step": 219450 + }, + { + "epoch": 2.465973497806463, + "grad_norm": 1.2603576183319092, + "learning_rate": 1.0139690600038199e-05, + "loss": 3.637, + "step": 219500 + }, + { + "epoch": 2.4665352229768063, + "grad_norm": 1.1873706579208374, + "learning_rate": 1.0137443686735348e-05, + "loss": 3.6228, + "step": 219550 + }, + { + "epoch": 2.4670969481471494, + "grad_norm": 1.1385982036590576, + "learning_rate": 1.01351967734325e-05, + "loss": 3.6278, + "step": 219600 + }, + { + "epoch": 2.4676586733174926, + "grad_norm": 1.1630982160568237, + "learning_rate": 1.0132949860129649e-05, + "loss": 3.6334, + "step": 219650 + }, + { + "epoch": 2.4682203984878357, + "grad_norm": 1.1185778379440308, + "learning_rate": 1.0130702946826798e-05, + "loss": 3.6418, + "step": 219700 + }, + { + "epoch": 2.468782123658179, + "grad_norm": 1.111441969871521, + "learning_rate": 1.0128456033523947e-05, + "loss": 3.668, + "step": 219750 + }, + { + "epoch": 2.469343848828522, + "grad_norm": 1.442638635635376, + "learning_rate": 1.0126209120221098e-05, + "loss": 3.6877, + "step": 219800 + }, + { + "epoch": 2.469905573998865, + "grad_norm": 1.118112325668335, + "learning_rate": 1.0123962206918248e-05, + "loss": 3.5624, + "step": 219850 + }, + { + "epoch": 2.4704672991692087, + "grad_norm": 1.1270848512649536, + "learning_rate": 1.0121715293615397e-05, + "loss": 3.6369, + "step": 219900 + }, + { + "epoch": 2.4710290243395514, + "grad_norm": 1.3634648323059082, + "learning_rate": 1.0119468380312548e-05, + "loss": 3.6662, + "step": 219950 + }, + { + "epoch": 2.471590749509895, + "grad_norm": 1.1991126537322998, + "learning_rate": 1.0117221467009697e-05, + "loss": 3.6581, + "step": 220000 + }, + { + "epoch": 2.472152474680238, + "grad_norm": 1.2159489393234253, + "learning_rate": 1.0114974553706847e-05, + "loss": 3.5836, + "step": 220050 + }, + { + "epoch": 2.4727141998505813, + "grad_norm": 1.057368516921997, + "learning_rate": 1.0112727640403996e-05, + "loss": 3.6606, + "step": 220100 + }, + { + "epoch": 2.4732759250209244, + "grad_norm": 1.4949387311935425, + "learning_rate": 1.0110480727101147e-05, + "loss": 3.7073, + "step": 220150 + }, + { + "epoch": 2.4738376501912676, + "grad_norm": 1.039035677909851, + "learning_rate": 1.0108233813798296e-05, + "loss": 3.6282, + "step": 220200 + }, + { + "epoch": 2.4743993753616107, + "grad_norm": 1.3212014436721802, + "learning_rate": 1.0105986900495445e-05, + "loss": 3.6479, + "step": 220250 + }, + { + "epoch": 2.474961100531954, + "grad_norm": 1.0305522680282593, + "learning_rate": 1.0103739987192596e-05, + "loss": 3.6493, + "step": 220300 + }, + { + "epoch": 2.475522825702297, + "grad_norm": 1.081343412399292, + "learning_rate": 1.0101493073889746e-05, + "loss": 3.6103, + "step": 220350 + }, + { + "epoch": 2.47608455087264, + "grad_norm": 1.2654993534088135, + "learning_rate": 1.0099246160586895e-05, + "loss": 3.6338, + "step": 220400 + }, + { + "epoch": 2.4766462760429833, + "grad_norm": 2.206638813018799, + "learning_rate": 1.0096999247284044e-05, + "loss": 3.657, + "step": 220450 + }, + { + "epoch": 2.4772080012133264, + "grad_norm": 1.1787418127059937, + "learning_rate": 1.0094752333981195e-05, + "loss": 3.601, + "step": 220500 + }, + { + "epoch": 2.4777697263836695, + "grad_norm": 1.1602827310562134, + "learning_rate": 1.0092505420678345e-05, + "loss": 3.7573, + "step": 220550 + }, + { + "epoch": 2.4783314515540127, + "grad_norm": 1.0930272340774536, + "learning_rate": 1.0090258507375494e-05, + "loss": 3.5823, + "step": 220600 + }, + { + "epoch": 2.478893176724356, + "grad_norm": 1.2273497581481934, + "learning_rate": 1.0088011594072643e-05, + "loss": 3.593, + "step": 220650 + }, + { + "epoch": 2.479454901894699, + "grad_norm": 1.2213597297668457, + "learning_rate": 1.0085764680769794e-05, + "loss": 3.7434, + "step": 220700 + }, + { + "epoch": 2.480016627065042, + "grad_norm": 1.196616530418396, + "learning_rate": 1.0083517767466944e-05, + "loss": 3.5832, + "step": 220750 + }, + { + "epoch": 2.4805783522353853, + "grad_norm": 1.21841299533844, + "learning_rate": 1.0081270854164093e-05, + "loss": 3.7306, + "step": 220800 + }, + { + "epoch": 2.4811400774057284, + "grad_norm": 1.2195571660995483, + "learning_rate": 1.0079023940861244e-05, + "loss": 3.6752, + "step": 220850 + }, + { + "epoch": 2.4817018025760715, + "grad_norm": 1.2555859088897705, + "learning_rate": 1.0076777027558393e-05, + "loss": 3.6822, + "step": 220900 + }, + { + "epoch": 2.4822635277464147, + "grad_norm": 1.1204763650894165, + "learning_rate": 1.0074530114255543e-05, + "loss": 3.7223, + "step": 220950 + }, + { + "epoch": 2.482825252916758, + "grad_norm": 1.1556209325790405, + "learning_rate": 1.0072283200952692e-05, + "loss": 3.6588, + "step": 221000 + }, + { + "epoch": 2.483386978087101, + "grad_norm": 1.1292698383331299, + "learning_rate": 1.0070036287649843e-05, + "loss": 3.6401, + "step": 221050 + }, + { + "epoch": 2.483948703257444, + "grad_norm": 1.0915005207061768, + "learning_rate": 1.0067789374346992e-05, + "loss": 3.5738, + "step": 221100 + }, + { + "epoch": 2.4845104284277872, + "grad_norm": 1.2218918800354004, + "learning_rate": 1.0065542461044142e-05, + "loss": 3.6765, + "step": 221150 + }, + { + "epoch": 2.4850721535981304, + "grad_norm": 1.2162284851074219, + "learning_rate": 1.0063295547741293e-05, + "loss": 3.5944, + "step": 221200 + }, + { + "epoch": 2.485633878768474, + "grad_norm": 1.1479686498641968, + "learning_rate": 1.0061048634438442e-05, + "loss": 3.6261, + "step": 221250 + }, + { + "epoch": 2.4861956039388167, + "grad_norm": 1.1043280363082886, + "learning_rate": 1.0058801721135591e-05, + "loss": 3.6859, + "step": 221300 + }, + { + "epoch": 2.4867573291091603, + "grad_norm": 1.1242334842681885, + "learning_rate": 1.005655480783274e-05, + "loss": 3.6959, + "step": 221350 + }, + { + "epoch": 2.4873190542795034, + "grad_norm": 1.1252726316452026, + "learning_rate": 1.0054307894529892e-05, + "loss": 3.6519, + "step": 221400 + }, + { + "epoch": 2.4878807794498465, + "grad_norm": 1.095293641090393, + "learning_rate": 1.0052060981227041e-05, + "loss": 3.6378, + "step": 221450 + }, + { + "epoch": 2.4884425046201897, + "grad_norm": 1.1695057153701782, + "learning_rate": 1.004981406792419e-05, + "loss": 3.6797, + "step": 221500 + }, + { + "epoch": 2.489004229790533, + "grad_norm": 1.2546207904815674, + "learning_rate": 1.0047567154621341e-05, + "loss": 3.6618, + "step": 221550 + }, + { + "epoch": 2.489565954960876, + "grad_norm": 1.0123082399368286, + "learning_rate": 1.004532024131849e-05, + "loss": 3.6444, + "step": 221600 + }, + { + "epoch": 2.490127680131219, + "grad_norm": 1.2161242961883545, + "learning_rate": 1.004307332801564e-05, + "loss": 3.5852, + "step": 221650 + }, + { + "epoch": 2.4906894053015622, + "grad_norm": 0.9452434182167053, + "learning_rate": 1.004082641471279e-05, + "loss": 3.5919, + "step": 221700 + }, + { + "epoch": 2.4912511304719054, + "grad_norm": 1.1768102645874023, + "learning_rate": 1.003857950140994e-05, + "loss": 3.6244, + "step": 221750 + }, + { + "epoch": 2.4918128556422485, + "grad_norm": 1.2206100225448608, + "learning_rate": 1.003633258810709e-05, + "loss": 3.6372, + "step": 221800 + }, + { + "epoch": 2.4923745808125917, + "grad_norm": 1.3399410247802734, + "learning_rate": 1.0034085674804239e-05, + "loss": 3.6186, + "step": 221850 + }, + { + "epoch": 2.492936305982935, + "grad_norm": 1.1862549781799316, + "learning_rate": 1.003183876150139e-05, + "loss": 3.5653, + "step": 221900 + }, + { + "epoch": 2.493498031153278, + "grad_norm": 1.1225323677062988, + "learning_rate": 1.002959184819854e-05, + "loss": 3.6313, + "step": 221950 + }, + { + "epoch": 2.494059756323621, + "grad_norm": 1.3176363706588745, + "learning_rate": 1.0027344934895688e-05, + "loss": 3.6931, + "step": 222000 + }, + { + "epoch": 2.4946214814939642, + "grad_norm": 1.2463792562484741, + "learning_rate": 1.0025098021592838e-05, + "loss": 3.6632, + "step": 222050 + }, + { + "epoch": 2.4951832066643074, + "grad_norm": 1.0954043865203857, + "learning_rate": 1.0022851108289989e-05, + "loss": 3.6704, + "step": 222100 + }, + { + "epoch": 2.4957449318346505, + "grad_norm": 1.0819978713989258, + "learning_rate": 1.0020604194987138e-05, + "loss": 3.6596, + "step": 222150 + }, + { + "epoch": 2.4963066570049937, + "grad_norm": 1.0964871644973755, + "learning_rate": 1.0018357281684287e-05, + "loss": 3.6577, + "step": 222200 + }, + { + "epoch": 2.496868382175337, + "grad_norm": 1.0642293691635132, + "learning_rate": 1.0016110368381437e-05, + "loss": 3.6343, + "step": 222250 + }, + { + "epoch": 2.49743010734568, + "grad_norm": 1.3439793586730957, + "learning_rate": 1.0013863455078588e-05, + "loss": 3.6691, + "step": 222300 + }, + { + "epoch": 2.497991832516023, + "grad_norm": 1.1293972730636597, + "learning_rate": 1.0011616541775737e-05, + "loss": 3.6225, + "step": 222350 + }, + { + "epoch": 2.4985535576863662, + "grad_norm": 2.4220151901245117, + "learning_rate": 1.0009369628472886e-05, + "loss": 3.6662, + "step": 222400 + }, + { + "epoch": 2.4991152828567094, + "grad_norm": 1.14895498752594, + "learning_rate": 1.0007122715170037e-05, + "loss": 3.7285, + "step": 222450 + }, + { + "epoch": 2.4996770080270525, + "grad_norm": 1.3529636859893799, + "learning_rate": 1.0004875801867187e-05, + "loss": 3.6247, + "step": 222500 + }, + { + "epoch": 2.5002387331973956, + "grad_norm": 1.074212670326233, + "learning_rate": 1.0002628888564336e-05, + "loss": 3.6574, + "step": 222550 + }, + { + "epoch": 2.5008004583677392, + "grad_norm": 1.0517715215682983, + "learning_rate": 1.0000381975261485e-05, + "loss": 3.6177, + "step": 222600 + }, + { + "epoch": 2.501362183538082, + "grad_norm": 1.2327373027801514, + "learning_rate": 9.998135061958635e-06, + "loss": 3.6466, + "step": 222650 + }, + { + "epoch": 2.5019239087084255, + "grad_norm": 1.2518898248672485, + "learning_rate": 9.995888148655784e-06, + "loss": 3.5565, + "step": 222700 + }, + { + "epoch": 2.502485633878768, + "grad_norm": 1.2579315900802612, + "learning_rate": 9.993641235352935e-06, + "loss": 3.6041, + "step": 222750 + }, + { + "epoch": 2.503047359049112, + "grad_norm": 1.1023578643798828, + "learning_rate": 9.991394322050084e-06, + "loss": 3.5945, + "step": 222800 + }, + { + "epoch": 2.503609084219455, + "grad_norm": 1.2034040689468384, + "learning_rate": 9.989147408747234e-06, + "loss": 3.7039, + "step": 222850 + }, + { + "epoch": 2.504170809389798, + "grad_norm": 1.1180808544158936, + "learning_rate": 9.986900495444383e-06, + "loss": 3.6345, + "step": 222900 + }, + { + "epoch": 2.504732534560141, + "grad_norm": 1.0925536155700684, + "learning_rate": 9.984653582141534e-06, + "loss": 3.6766, + "step": 222950 + }, + { + "epoch": 2.5052942597304844, + "grad_norm": 1.1931333541870117, + "learning_rate": 9.982406668838683e-06, + "loss": 3.7071, + "step": 223000 + }, + { + "epoch": 2.5058559849008275, + "grad_norm": 1.371894359588623, + "learning_rate": 9.980159755535833e-06, + "loss": 3.6939, + "step": 223050 + }, + { + "epoch": 2.5064177100711706, + "grad_norm": 1.0637400150299072, + "learning_rate": 9.977912842232984e-06, + "loss": 3.6487, + "step": 223100 + }, + { + "epoch": 2.506979435241514, + "grad_norm": 1.3496999740600586, + "learning_rate": 9.975665928930133e-06, + "loss": 3.6769, + "step": 223150 + }, + { + "epoch": 2.507541160411857, + "grad_norm": 1.0121572017669678, + "learning_rate": 9.973419015627282e-06, + "loss": 3.7185, + "step": 223200 + }, + { + "epoch": 2.5081028855822, + "grad_norm": 1.0683531761169434, + "learning_rate": 9.971172102324432e-06, + "loss": 3.6865, + "step": 223250 + }, + { + "epoch": 2.508664610752543, + "grad_norm": 1.2914478778839111, + "learning_rate": 9.968925189021583e-06, + "loss": 3.6488, + "step": 223300 + }, + { + "epoch": 2.5092263359228864, + "grad_norm": 1.1898984909057617, + "learning_rate": 9.966678275718732e-06, + "loss": 3.6985, + "step": 223350 + }, + { + "epoch": 2.5097880610932295, + "grad_norm": 1.2759250402450562, + "learning_rate": 9.964431362415881e-06, + "loss": 3.5878, + "step": 223400 + }, + { + "epoch": 2.5103497862635726, + "grad_norm": 1.2521064281463623, + "learning_rate": 9.962184449113032e-06, + "loss": 3.5806, + "step": 223450 + }, + { + "epoch": 2.5109115114339158, + "grad_norm": 1.2106614112854004, + "learning_rate": 9.959937535810182e-06, + "loss": 3.7408, + "step": 223500 + }, + { + "epoch": 2.511473236604259, + "grad_norm": 1.4075654745101929, + "learning_rate": 9.957690622507331e-06, + "loss": 3.6301, + "step": 223550 + }, + { + "epoch": 2.512034961774602, + "grad_norm": 1.0811665058135986, + "learning_rate": 9.95544370920448e-06, + "loss": 3.6178, + "step": 223600 + }, + { + "epoch": 2.512596686944945, + "grad_norm": 1.3429867029190063, + "learning_rate": 9.953196795901631e-06, + "loss": 3.658, + "step": 223650 + }, + { + "epoch": 2.5131584121152883, + "grad_norm": 1.1472651958465576, + "learning_rate": 9.95094988259878e-06, + "loss": 3.6108, + "step": 223700 + }, + { + "epoch": 2.5137201372856315, + "grad_norm": 1.316075325012207, + "learning_rate": 9.94870296929593e-06, + "loss": 3.6201, + "step": 223750 + }, + { + "epoch": 2.5142818624559746, + "grad_norm": 1.3413081169128418, + "learning_rate": 9.94645605599308e-06, + "loss": 3.5898, + "step": 223800 + }, + { + "epoch": 2.514843587626318, + "grad_norm": 1.2931952476501465, + "learning_rate": 9.94420914269023e-06, + "loss": 3.6628, + "step": 223850 + }, + { + "epoch": 2.515405312796661, + "grad_norm": 1.4174492359161377, + "learning_rate": 9.94196222938738e-06, + "loss": 3.6464, + "step": 223900 + }, + { + "epoch": 2.5159670379670045, + "grad_norm": 1.0922467708587646, + "learning_rate": 9.939715316084529e-06, + "loss": 3.5746, + "step": 223950 + }, + { + "epoch": 2.516528763137347, + "grad_norm": 1.1403781175613403, + "learning_rate": 9.93746840278168e-06, + "loss": 3.6885, + "step": 224000 + }, + { + "epoch": 2.5170904883076908, + "grad_norm": 1.3727272748947144, + "learning_rate": 9.935221489478829e-06, + "loss": 3.6833, + "step": 224050 + }, + { + "epoch": 2.5176522134780335, + "grad_norm": 1.3506124019622803, + "learning_rate": 9.932974576175978e-06, + "loss": 3.6645, + "step": 224100 + }, + { + "epoch": 2.518213938648377, + "grad_norm": 1.1761014461517334, + "learning_rate": 9.93072766287313e-06, + "loss": 3.6475, + "step": 224150 + }, + { + "epoch": 2.51877566381872, + "grad_norm": 1.3753101825714111, + "learning_rate": 9.928480749570279e-06, + "loss": 3.594, + "step": 224200 + }, + { + "epoch": 2.5193373889890633, + "grad_norm": 1.7304960489273071, + "learning_rate": 9.926233836267428e-06, + "loss": 3.645, + "step": 224250 + }, + { + "epoch": 2.5198991141594065, + "grad_norm": 1.1977565288543701, + "learning_rate": 9.923986922964577e-06, + "loss": 3.6815, + "step": 224300 + }, + { + "epoch": 2.5204608393297496, + "grad_norm": 1.3336644172668457, + "learning_rate": 9.921740009661728e-06, + "loss": 3.6404, + "step": 224350 + }, + { + "epoch": 2.5210225645000928, + "grad_norm": 1.0081298351287842, + "learning_rate": 9.919493096358878e-06, + "loss": 3.5817, + "step": 224400 + }, + { + "epoch": 2.521584289670436, + "grad_norm": 1.0654035806655884, + "learning_rate": 9.917246183056027e-06, + "loss": 3.6469, + "step": 224450 + }, + { + "epoch": 2.522146014840779, + "grad_norm": 1.0444693565368652, + "learning_rate": 9.914999269753176e-06, + "loss": 3.5891, + "step": 224500 + }, + { + "epoch": 2.522707740011122, + "grad_norm": 1.1029542684555054, + "learning_rate": 9.912752356450327e-06, + "loss": 3.5915, + "step": 224550 + }, + { + "epoch": 2.5232694651814653, + "grad_norm": 1.1915572881698608, + "learning_rate": 9.910505443147477e-06, + "loss": 3.6211, + "step": 224600 + }, + { + "epoch": 2.5238311903518085, + "grad_norm": 2.208009958267212, + "learning_rate": 9.908258529844626e-06, + "loss": 3.6802, + "step": 224650 + }, + { + "epoch": 2.5243929155221516, + "grad_norm": 1.3922756910324097, + "learning_rate": 9.906011616541777e-06, + "loss": 3.595, + "step": 224700 + }, + { + "epoch": 2.5249546406924948, + "grad_norm": 1.2410928010940552, + "learning_rate": 9.903764703238926e-06, + "loss": 3.7118, + "step": 224750 + }, + { + "epoch": 2.525516365862838, + "grad_norm": 1.3258376121520996, + "learning_rate": 9.901517789936076e-06, + "loss": 3.6872, + "step": 224800 + }, + { + "epoch": 2.526078091033181, + "grad_norm": 1.3248200416564941, + "learning_rate": 9.899315814899283e-06, + "loss": 3.687, + "step": 224850 + }, + { + "epoch": 2.526639816203524, + "grad_norm": 1.1712646484375, + "learning_rate": 9.897068901596433e-06, + "loss": 3.6254, + "step": 224900 + }, + { + "epoch": 2.5272015413738673, + "grad_norm": 1.007655143737793, + "learning_rate": 9.894821988293582e-06, + "loss": 3.6162, + "step": 224950 + }, + { + "epoch": 2.5277632665442105, + "grad_norm": 1.2282979488372803, + "learning_rate": 9.892575074990733e-06, + "loss": 3.7098, + "step": 225000 + }, + { + "epoch": 2.5283249917145536, + "grad_norm": 1.0925374031066895, + "learning_rate": 9.890328161687882e-06, + "loss": 3.606, + "step": 225050 + }, + { + "epoch": 2.5288867168848967, + "grad_norm": 1.2417986392974854, + "learning_rate": 9.888081248385032e-06, + "loss": 3.5919, + "step": 225100 + }, + { + "epoch": 2.52944844205524, + "grad_norm": 1.144342303276062, + "learning_rate": 9.885834335082183e-06, + "loss": 3.6636, + "step": 225150 + }, + { + "epoch": 2.5300101672255835, + "grad_norm": 1.090048909187317, + "learning_rate": 9.883587421779332e-06, + "loss": 3.601, + "step": 225200 + }, + { + "epoch": 2.530571892395926, + "grad_norm": 1.1404286623001099, + "learning_rate": 9.881340508476481e-06, + "loss": 3.5664, + "step": 225250 + }, + { + "epoch": 2.5311336175662698, + "grad_norm": 1.2363505363464355, + "learning_rate": 9.87909359517363e-06, + "loss": 3.6045, + "step": 225300 + }, + { + "epoch": 2.5316953427366125, + "grad_norm": 1.2350670099258423, + "learning_rate": 9.876846681870781e-06, + "loss": 3.6679, + "step": 225350 + }, + { + "epoch": 2.532257067906956, + "grad_norm": 1.3042763471603394, + "learning_rate": 9.87459976856793e-06, + "loss": 3.6355, + "step": 225400 + }, + { + "epoch": 2.5328187930772987, + "grad_norm": 1.2034074068069458, + "learning_rate": 9.87235285526508e-06, + "loss": 3.6217, + "step": 225450 + }, + { + "epoch": 2.5333805182476423, + "grad_norm": 1.1341825723648071, + "learning_rate": 9.87010594196223e-06, + "loss": 3.5584, + "step": 225500 + }, + { + "epoch": 2.5339422434179855, + "grad_norm": 1.2538197040557861, + "learning_rate": 9.86785902865938e-06, + "loss": 3.6495, + "step": 225550 + }, + { + "epoch": 2.5345039685883286, + "grad_norm": 1.225101113319397, + "learning_rate": 9.86561211535653e-06, + "loss": 3.6399, + "step": 225600 + }, + { + "epoch": 2.5350656937586717, + "grad_norm": 1.322907567024231, + "learning_rate": 9.863365202053679e-06, + "loss": 3.6577, + "step": 225650 + }, + { + "epoch": 2.535627418929015, + "grad_norm": 1.14747154712677, + "learning_rate": 9.86111828875083e-06, + "loss": 3.6627, + "step": 225700 + }, + { + "epoch": 2.536189144099358, + "grad_norm": 1.3067673444747925, + "learning_rate": 9.85887137544798e-06, + "loss": 3.6451, + "step": 225750 + }, + { + "epoch": 2.536750869269701, + "grad_norm": 1.1978440284729004, + "learning_rate": 9.856624462145129e-06, + "loss": 3.6941, + "step": 225800 + }, + { + "epoch": 2.5373125944400443, + "grad_norm": 1.2274320125579834, + "learning_rate": 9.854377548842278e-06, + "loss": 3.5767, + "step": 225850 + }, + { + "epoch": 2.5378743196103875, + "grad_norm": 1.400112271308899, + "learning_rate": 9.852130635539429e-06, + "loss": 3.6707, + "step": 225900 + }, + { + "epoch": 2.5384360447807306, + "grad_norm": 1.0723315477371216, + "learning_rate": 9.849883722236578e-06, + "loss": 3.6521, + "step": 225950 + }, + { + "epoch": 2.5389977699510737, + "grad_norm": 1.1016207933425903, + "learning_rate": 9.847636808933728e-06, + "loss": 3.6211, + "step": 226000 + }, + { + "epoch": 2.539559495121417, + "grad_norm": 1.2424542903900146, + "learning_rate": 9.845389895630879e-06, + "loss": 3.6466, + "step": 226050 + }, + { + "epoch": 2.54012122029176, + "grad_norm": 1.113968014717102, + "learning_rate": 9.843142982328028e-06, + "loss": 3.6645, + "step": 226100 + }, + { + "epoch": 2.540682945462103, + "grad_norm": 1.185343623161316, + "learning_rate": 9.840896069025177e-06, + "loss": 3.6477, + "step": 226150 + }, + { + "epoch": 2.5412446706324463, + "grad_norm": 1.28215491771698, + "learning_rate": 9.838649155722327e-06, + "loss": 3.7339, + "step": 226200 + }, + { + "epoch": 2.5418063958027894, + "grad_norm": 1.1373237371444702, + "learning_rate": 9.836402242419478e-06, + "loss": 3.5953, + "step": 226250 + }, + { + "epoch": 2.5423681209731326, + "grad_norm": 1.2297579050064087, + "learning_rate": 9.834155329116627e-06, + "loss": 3.6917, + "step": 226300 + }, + { + "epoch": 2.5429298461434757, + "grad_norm": 1.1348350048065186, + "learning_rate": 9.831908415813776e-06, + "loss": 3.5935, + "step": 226350 + }, + { + "epoch": 2.543491571313819, + "grad_norm": 1.2301366329193115, + "learning_rate": 9.829661502510927e-06, + "loss": 3.5674, + "step": 226400 + }, + { + "epoch": 2.544053296484162, + "grad_norm": 1.1980699300765991, + "learning_rate": 9.827414589208077e-06, + "loss": 3.6758, + "step": 226450 + }, + { + "epoch": 2.544615021654505, + "grad_norm": 1.1047879457473755, + "learning_rate": 9.825167675905226e-06, + "loss": 3.7047, + "step": 226500 + }, + { + "epoch": 2.5451767468248487, + "grad_norm": 1.1126362085342407, + "learning_rate": 9.822920762602375e-06, + "loss": 3.5761, + "step": 226550 + }, + { + "epoch": 2.5457384719951914, + "grad_norm": 1.0520906448364258, + "learning_rate": 9.820673849299526e-06, + "loss": 3.5795, + "step": 226600 + }, + { + "epoch": 2.546300197165535, + "grad_norm": 1.475176215171814, + "learning_rate": 9.818426935996676e-06, + "loss": 3.6411, + "step": 226650 + }, + { + "epoch": 2.5468619223358777, + "grad_norm": 1.1460368633270264, + "learning_rate": 9.816180022693825e-06, + "loss": 3.6371, + "step": 226700 + }, + { + "epoch": 2.5474236475062213, + "grad_norm": 1.2125566005706787, + "learning_rate": 9.813933109390976e-06, + "loss": 3.6441, + "step": 226750 + }, + { + "epoch": 2.547985372676564, + "grad_norm": 1.3701915740966797, + "learning_rate": 9.811686196088125e-06, + "loss": 3.7155, + "step": 226800 + }, + { + "epoch": 2.5485470978469076, + "grad_norm": 1.1672906875610352, + "learning_rate": 9.809439282785275e-06, + "loss": 3.6298, + "step": 226850 + }, + { + "epoch": 2.5491088230172507, + "grad_norm": 1.2414804697036743, + "learning_rate": 9.807237307748482e-06, + "loss": 3.6688, + "step": 226900 + }, + { + "epoch": 2.549670548187594, + "grad_norm": 1.0830920934677124, + "learning_rate": 9.804990394445631e-06, + "loss": 3.5695, + "step": 226950 + }, + { + "epoch": 2.550232273357937, + "grad_norm": 1.1367555856704712, + "learning_rate": 9.80274348114278e-06, + "loss": 3.6027, + "step": 227000 + }, + { + "epoch": 2.55079399852828, + "grad_norm": 1.1189628839492798, + "learning_rate": 9.800541506105987e-06, + "loss": 3.6176, + "step": 227050 + }, + { + "epoch": 2.5513557236986233, + "grad_norm": 1.156973958015442, + "learning_rate": 9.798294592803138e-06, + "loss": 3.6535, + "step": 227100 + }, + { + "epoch": 2.5519174488689664, + "grad_norm": 1.1597468852996826, + "learning_rate": 9.796047679500287e-06, + "loss": 3.6358, + "step": 227150 + }, + { + "epoch": 2.5524791740393096, + "grad_norm": 1.1040409803390503, + "learning_rate": 9.793800766197436e-06, + "loss": 3.7609, + "step": 227200 + }, + { + "epoch": 2.5530408992096527, + "grad_norm": 1.3094677925109863, + "learning_rate": 9.791553852894586e-06, + "loss": 3.71, + "step": 227250 + }, + { + "epoch": 2.553602624379996, + "grad_norm": 1.2151713371276855, + "learning_rate": 9.789306939591737e-06, + "loss": 3.6331, + "step": 227300 + }, + { + "epoch": 2.554164349550339, + "grad_norm": 1.1243503093719482, + "learning_rate": 9.787060026288886e-06, + "loss": 3.597, + "step": 227350 + }, + { + "epoch": 2.554726074720682, + "grad_norm": 1.1748740673065186, + "learning_rate": 9.784813112986035e-06, + "loss": 3.6153, + "step": 227400 + }, + { + "epoch": 2.5552877998910253, + "grad_norm": 1.1404298543930054, + "learning_rate": 9.782566199683186e-06, + "loss": 3.6226, + "step": 227450 + }, + { + "epoch": 2.5558495250613684, + "grad_norm": 1.275992751121521, + "learning_rate": 9.780319286380336e-06, + "loss": 3.7076, + "step": 227500 + }, + { + "epoch": 2.5564112502317116, + "grad_norm": 1.2919789552688599, + "learning_rate": 9.778072373077485e-06, + "loss": 3.5933, + "step": 227550 + }, + { + "epoch": 2.5569729754020547, + "grad_norm": 1.1128894090652466, + "learning_rate": 9.775825459774634e-06, + "loss": 3.6055, + "step": 227600 + }, + { + "epoch": 2.557534700572398, + "grad_norm": 1.2390849590301514, + "learning_rate": 9.773578546471785e-06, + "loss": 3.6385, + "step": 227650 + }, + { + "epoch": 2.558096425742741, + "grad_norm": 1.1185600757598877, + "learning_rate": 9.771331633168935e-06, + "loss": 3.6285, + "step": 227700 + }, + { + "epoch": 2.558658150913084, + "grad_norm": 1.2565029859542847, + "learning_rate": 9.769084719866084e-06, + "loss": 3.6374, + "step": 227750 + }, + { + "epoch": 2.5592198760834273, + "grad_norm": 1.1560763120651245, + "learning_rate": 9.766837806563235e-06, + "loss": 3.686, + "step": 227800 + }, + { + "epoch": 2.5597816012537704, + "grad_norm": 1.5770220756530762, + "learning_rate": 9.764590893260384e-06, + "loss": 3.663, + "step": 227850 + }, + { + "epoch": 2.560343326424114, + "grad_norm": 1.2783591747283936, + "learning_rate": 9.762343979957534e-06, + "loss": 3.6719, + "step": 227900 + }, + { + "epoch": 2.5609050515944567, + "grad_norm": 1.4159845113754272, + "learning_rate": 9.760097066654683e-06, + "loss": 3.6422, + "step": 227950 + }, + { + "epoch": 2.5614667767648003, + "grad_norm": 1.1151092052459717, + "learning_rate": 9.757850153351834e-06, + "loss": 3.5841, + "step": 228000 + }, + { + "epoch": 2.562028501935143, + "grad_norm": 1.0429368019104004, + "learning_rate": 9.755603240048983e-06, + "loss": 3.5358, + "step": 228050 + }, + { + "epoch": 2.5625902271054866, + "grad_norm": 1.4416569471359253, + "learning_rate": 9.753356326746133e-06, + "loss": 3.6824, + "step": 228100 + }, + { + "epoch": 2.5631519522758293, + "grad_norm": 1.1972196102142334, + "learning_rate": 9.751109413443282e-06, + "loss": 3.6269, + "step": 228150 + }, + { + "epoch": 2.563713677446173, + "grad_norm": 1.1545695066452026, + "learning_rate": 9.748862500140433e-06, + "loss": 3.626, + "step": 228200 + }, + { + "epoch": 2.564275402616516, + "grad_norm": 1.23501455783844, + "learning_rate": 9.746615586837582e-06, + "loss": 3.6907, + "step": 228250 + }, + { + "epoch": 2.564837127786859, + "grad_norm": 1.1475234031677246, + "learning_rate": 9.744368673534731e-06, + "loss": 3.7481, + "step": 228300 + }, + { + "epoch": 2.5653988529572023, + "grad_norm": 1.705857276916504, + "learning_rate": 9.742121760231882e-06, + "loss": 3.605, + "step": 228350 + }, + { + "epoch": 2.5659605781275454, + "grad_norm": 1.0402103662490845, + "learning_rate": 9.739874846929032e-06, + "loss": 3.6349, + "step": 228400 + }, + { + "epoch": 2.5665223032978886, + "grad_norm": 1.1646581888198853, + "learning_rate": 9.737627933626181e-06, + "loss": 3.6811, + "step": 228450 + }, + { + "epoch": 2.5670840284682317, + "grad_norm": 1.327896237373352, + "learning_rate": 9.73538102032333e-06, + "loss": 3.7385, + "step": 228500 + }, + { + "epoch": 2.567645753638575, + "grad_norm": 1.330582618713379, + "learning_rate": 9.733134107020481e-06, + "loss": 3.7422, + "step": 228550 + }, + { + "epoch": 2.568207478808918, + "grad_norm": 1.0372596979141235, + "learning_rate": 9.73088719371763e-06, + "loss": 3.6206, + "step": 228600 + }, + { + "epoch": 2.568769203979261, + "grad_norm": 1.1130669116973877, + "learning_rate": 9.72864028041478e-06, + "loss": 3.7046, + "step": 228650 + }, + { + "epoch": 2.5693309291496043, + "grad_norm": 1.196959137916565, + "learning_rate": 9.726393367111931e-06, + "loss": 3.631, + "step": 228700 + }, + { + "epoch": 2.5698926543199474, + "grad_norm": 1.3812388181686401, + "learning_rate": 9.72414645380908e-06, + "loss": 3.6295, + "step": 228750 + }, + { + "epoch": 2.5704543794902905, + "grad_norm": 1.2576218843460083, + "learning_rate": 9.72189954050623e-06, + "loss": 3.6193, + "step": 228800 + }, + { + "epoch": 2.5710161046606337, + "grad_norm": 1.4126763343811035, + "learning_rate": 9.719652627203379e-06, + "loss": 3.5631, + "step": 228850 + }, + { + "epoch": 2.571577829830977, + "grad_norm": 1.189589262008667, + "learning_rate": 9.71740571390053e-06, + "loss": 3.5979, + "step": 228900 + }, + { + "epoch": 2.57213955500132, + "grad_norm": 1.1494877338409424, + "learning_rate": 9.71515880059768e-06, + "loss": 3.639, + "step": 228950 + }, + { + "epoch": 2.572701280171663, + "grad_norm": 1.1925745010375977, + "learning_rate": 9.712911887294829e-06, + "loss": 3.6837, + "step": 229000 + }, + { + "epoch": 2.5732630053420062, + "grad_norm": 1.2979105710983276, + "learning_rate": 9.71066497399198e-06, + "loss": 3.6697, + "step": 229050 + }, + { + "epoch": 2.5738247305123494, + "grad_norm": 1.1843870878219604, + "learning_rate": 9.708418060689129e-06, + "loss": 3.6816, + "step": 229100 + }, + { + "epoch": 2.5743864556826925, + "grad_norm": 0.9702261686325073, + "learning_rate": 9.706171147386278e-06, + "loss": 3.6774, + "step": 229150 + }, + { + "epoch": 2.5749481808530357, + "grad_norm": 1.1876819133758545, + "learning_rate": 9.703924234083428e-06, + "loss": 3.6931, + "step": 229200 + }, + { + "epoch": 2.5755099060233793, + "grad_norm": 1.1669557094573975, + "learning_rate": 9.701722259046635e-06, + "loss": 3.6682, + "step": 229250 + }, + { + "epoch": 2.576071631193722, + "grad_norm": 1.2973331212997437, + "learning_rate": 9.699475345743785e-06, + "loss": 3.6812, + "step": 229300 + }, + { + "epoch": 2.5766333563640655, + "grad_norm": 1.2905819416046143, + "learning_rate": 9.697228432440936e-06, + "loss": 3.7467, + "step": 229350 + }, + { + "epoch": 2.5771950815344082, + "grad_norm": 1.0218195915222168, + "learning_rate": 9.694981519138085e-06, + "loss": 3.6171, + "step": 229400 + }, + { + "epoch": 2.577756806704752, + "grad_norm": 1.2959119081497192, + "learning_rate": 9.692734605835234e-06, + "loss": 3.6238, + "step": 229450 + }, + { + "epoch": 2.5783185318750945, + "grad_norm": 1.1232720613479614, + "learning_rate": 9.690487692532384e-06, + "loss": 3.6235, + "step": 229500 + }, + { + "epoch": 2.578880257045438, + "grad_norm": 1.2635669708251953, + "learning_rate": 9.688240779229535e-06, + "loss": 3.6442, + "step": 229550 + }, + { + "epoch": 2.5794419822157812, + "grad_norm": 1.140675663948059, + "learning_rate": 9.685993865926684e-06, + "loss": 3.6595, + "step": 229600 + }, + { + "epoch": 2.5800037073861244, + "grad_norm": 1.2757861614227295, + "learning_rate": 9.683746952623833e-06, + "loss": 3.6604, + "step": 229650 + }, + { + "epoch": 2.5805654325564675, + "grad_norm": 1.026766300201416, + "learning_rate": 9.681500039320984e-06, + "loss": 3.676, + "step": 229700 + }, + { + "epoch": 2.5811271577268107, + "grad_norm": 1.2279552221298218, + "learning_rate": 9.679253126018134e-06, + "loss": 3.6094, + "step": 229750 + }, + { + "epoch": 2.581688882897154, + "grad_norm": 1.172577142715454, + "learning_rate": 9.677006212715283e-06, + "loss": 3.7079, + "step": 229800 + }, + { + "epoch": 2.582250608067497, + "grad_norm": 1.140169382095337, + "learning_rate": 9.674759299412432e-06, + "loss": 3.6405, + "step": 229850 + }, + { + "epoch": 2.58281233323784, + "grad_norm": 1.3780945539474487, + "learning_rate": 9.672512386109583e-06, + "loss": 3.6893, + "step": 229900 + }, + { + "epoch": 2.5833740584081832, + "grad_norm": 1.1179897785186768, + "learning_rate": 9.670265472806732e-06, + "loss": 3.605, + "step": 229950 + }, + { + "epoch": 2.5839357835785264, + "grad_norm": 1.0302451848983765, + "learning_rate": 9.668018559503882e-06, + "loss": 3.5907, + "step": 230000 + }, + { + "epoch": 2.5844975087488695, + "grad_norm": 1.1761894226074219, + "learning_rate": 9.665771646201033e-06, + "loss": 3.6571, + "step": 230050 + }, + { + "epoch": 2.5850592339192127, + "grad_norm": 1.1279114484786987, + "learning_rate": 9.663524732898182e-06, + "loss": 3.6152, + "step": 230100 + }, + { + "epoch": 2.585620959089556, + "grad_norm": 1.0731397867202759, + "learning_rate": 9.661277819595331e-06, + "loss": 3.6051, + "step": 230150 + }, + { + "epoch": 2.586182684259899, + "grad_norm": 1.0091617107391357, + "learning_rate": 9.65903090629248e-06, + "loss": 3.6454, + "step": 230200 + }, + { + "epoch": 2.586744409430242, + "grad_norm": 1.1999599933624268, + "learning_rate": 9.656783992989632e-06, + "loss": 3.5903, + "step": 230250 + }, + { + "epoch": 2.5873061346005852, + "grad_norm": 1.1237565279006958, + "learning_rate": 9.654537079686781e-06, + "loss": 3.6313, + "step": 230300 + }, + { + "epoch": 2.5878678597709284, + "grad_norm": 1.1436638832092285, + "learning_rate": 9.65229016638393e-06, + "loss": 3.684, + "step": 230350 + }, + { + "epoch": 2.5884295849412715, + "grad_norm": 1.419813632965088, + "learning_rate": 9.650043253081081e-06, + "loss": 3.6183, + "step": 230400 + }, + { + "epoch": 2.5889913101116147, + "grad_norm": 1.036826252937317, + "learning_rate": 9.64779633977823e-06, + "loss": 3.6608, + "step": 230450 + }, + { + "epoch": 2.589553035281958, + "grad_norm": 1.0370091199874878, + "learning_rate": 9.64554942647538e-06, + "loss": 3.6029, + "step": 230500 + }, + { + "epoch": 2.590114760452301, + "grad_norm": 1.214907169342041, + "learning_rate": 9.64330251317253e-06, + "loss": 3.6474, + "step": 230550 + }, + { + "epoch": 2.5906764856226445, + "grad_norm": 1.1329468488693237, + "learning_rate": 9.64105559986968e-06, + "loss": 3.6658, + "step": 230600 + }, + { + "epoch": 2.591238210792987, + "grad_norm": 1.250857949256897, + "learning_rate": 9.63880868656683e-06, + "loss": 3.7155, + "step": 230650 + }, + { + "epoch": 2.591799935963331, + "grad_norm": 1.4392398595809937, + "learning_rate": 9.636561773263979e-06, + "loss": 3.638, + "step": 230700 + }, + { + "epoch": 2.5923616611336735, + "grad_norm": 1.2033816576004028, + "learning_rate": 9.634314859961128e-06, + "loss": 3.595, + "step": 230750 + }, + { + "epoch": 2.592923386304017, + "grad_norm": 1.2748844623565674, + "learning_rate": 9.63206794665828e-06, + "loss": 3.6759, + "step": 230800 + }, + { + "epoch": 2.59348511147436, + "grad_norm": 1.1000982522964478, + "learning_rate": 9.629821033355429e-06, + "loss": 3.6866, + "step": 230850 + }, + { + "epoch": 2.5940468366447034, + "grad_norm": 1.2321597337722778, + "learning_rate": 9.627574120052578e-06, + "loss": 3.7767, + "step": 230900 + }, + { + "epoch": 2.5946085618150465, + "grad_norm": 1.126806378364563, + "learning_rate": 9.625327206749729e-06, + "loss": 3.7061, + "step": 230950 + }, + { + "epoch": 2.5951702869853897, + "grad_norm": 1.2558183670043945, + "learning_rate": 9.623080293446878e-06, + "loss": 3.6879, + "step": 231000 + }, + { + "epoch": 2.595732012155733, + "grad_norm": 1.1356263160705566, + "learning_rate": 9.620833380144028e-06, + "loss": 3.6204, + "step": 231050 + }, + { + "epoch": 2.596293737326076, + "grad_norm": 1.0545427799224854, + "learning_rate": 9.618586466841177e-06, + "loss": 3.6406, + "step": 231100 + }, + { + "epoch": 2.596855462496419, + "grad_norm": 1.0609054565429688, + "learning_rate": 9.616339553538328e-06, + "loss": 3.6456, + "step": 231150 + }, + { + "epoch": 2.597417187666762, + "grad_norm": 1.2802765369415283, + "learning_rate": 9.614092640235477e-06, + "loss": 3.6708, + "step": 231200 + }, + { + "epoch": 2.5979789128371054, + "grad_norm": 1.2823866605758667, + "learning_rate": 9.611845726932627e-06, + "loss": 3.6163, + "step": 231250 + }, + { + "epoch": 2.5985406380074485, + "grad_norm": 1.1320018768310547, + "learning_rate": 9.609598813629778e-06, + "loss": 3.6317, + "step": 231300 + }, + { + "epoch": 2.5991023631777916, + "grad_norm": 1.2215267419815063, + "learning_rate": 9.607351900326927e-06, + "loss": 3.7005, + "step": 231350 + }, + { + "epoch": 2.599664088348135, + "grad_norm": 1.1547940969467163, + "learning_rate": 9.605104987024076e-06, + "loss": 3.6928, + "step": 231400 + }, + { + "epoch": 2.600225813518478, + "grad_norm": 1.1811012029647827, + "learning_rate": 9.602858073721226e-06, + "loss": 3.7021, + "step": 231450 + }, + { + "epoch": 2.600787538688821, + "grad_norm": 1.4081757068634033, + "learning_rate": 9.600611160418377e-06, + "loss": 3.6676, + "step": 231500 + }, + { + "epoch": 2.601349263859164, + "grad_norm": 1.0522280931472778, + "learning_rate": 9.598364247115526e-06, + "loss": 3.6563, + "step": 231550 + }, + { + "epoch": 2.6019109890295073, + "grad_norm": 1.44118332862854, + "learning_rate": 9.596117333812675e-06, + "loss": 3.6493, + "step": 231600 + }, + { + "epoch": 2.6024727141998505, + "grad_norm": 1.143570899963379, + "learning_rate": 9.593870420509826e-06, + "loss": 3.6629, + "step": 231650 + }, + { + "epoch": 2.6030344393701936, + "grad_norm": 1.205647349357605, + "learning_rate": 9.591623507206975e-06, + "loss": 3.6862, + "step": 231700 + }, + { + "epoch": 2.6035961645405368, + "grad_norm": 1.1053411960601807, + "learning_rate": 9.589376593904125e-06, + "loss": 3.7092, + "step": 231750 + }, + { + "epoch": 2.60415788971088, + "grad_norm": 1.081027626991272, + "learning_rate": 9.587129680601274e-06, + "loss": 3.6065, + "step": 231800 + }, + { + "epoch": 2.604719614881223, + "grad_norm": 1.2975106239318848, + "learning_rate": 9.584882767298425e-06, + "loss": 3.6254, + "step": 231850 + }, + { + "epoch": 2.605281340051566, + "grad_norm": 1.1592826843261719, + "learning_rate": 9.582635853995574e-06, + "loss": 3.5702, + "step": 231900 + }, + { + "epoch": 2.60584306522191, + "grad_norm": 1.1324329376220703, + "learning_rate": 9.580388940692724e-06, + "loss": 3.6665, + "step": 231950 + }, + { + "epoch": 2.6064047903922525, + "grad_norm": 1.1697642803192139, + "learning_rate": 9.578142027389875e-06, + "loss": 3.6271, + "step": 232000 + }, + { + "epoch": 2.606966515562596, + "grad_norm": 1.2418887615203857, + "learning_rate": 9.575895114087024e-06, + "loss": 3.6288, + "step": 232050 + }, + { + "epoch": 2.6075282407329388, + "grad_norm": 1.1859896183013916, + "learning_rate": 9.573648200784173e-06, + "loss": 3.669, + "step": 232100 + }, + { + "epoch": 2.6080899659032823, + "grad_norm": 1.1547828912734985, + "learning_rate": 9.571401287481323e-06, + "loss": 3.7053, + "step": 232150 + }, + { + "epoch": 2.608651691073625, + "grad_norm": 1.3412789106369019, + "learning_rate": 9.569154374178474e-06, + "loss": 3.6156, + "step": 232200 + }, + { + "epoch": 2.6092134162439686, + "grad_norm": 1.2743942737579346, + "learning_rate": 9.566907460875623e-06, + "loss": 3.6579, + "step": 232250 + }, + { + "epoch": 2.6097751414143118, + "grad_norm": 1.14712393283844, + "learning_rate": 9.564660547572772e-06, + "loss": 3.6421, + "step": 232300 + }, + { + "epoch": 2.610336866584655, + "grad_norm": 1.2016804218292236, + "learning_rate": 9.562413634269922e-06, + "loss": 3.5883, + "step": 232350 + }, + { + "epoch": 2.610898591754998, + "grad_norm": 1.2405736446380615, + "learning_rate": 9.560166720967073e-06, + "loss": 3.6278, + "step": 232400 + }, + { + "epoch": 2.611460316925341, + "grad_norm": 1.0889028310775757, + "learning_rate": 9.557919807664222e-06, + "loss": 3.634, + "step": 232450 + }, + { + "epoch": 2.6120220420956843, + "grad_norm": 1.4650338888168335, + "learning_rate": 9.555672894361371e-06, + "loss": 3.6748, + "step": 232500 + }, + { + "epoch": 2.6125837672660275, + "grad_norm": 1.162473440170288, + "learning_rate": 9.553425981058522e-06, + "loss": 3.6959, + "step": 232550 + }, + { + "epoch": 2.6131454924363706, + "grad_norm": 1.3220645189285278, + "learning_rate": 9.551179067755672e-06, + "loss": 3.6216, + "step": 232600 + }, + { + "epoch": 2.6137072176067138, + "grad_norm": 1.1755160093307495, + "learning_rate": 9.548932154452821e-06, + "loss": 3.6655, + "step": 232650 + }, + { + "epoch": 2.614268942777057, + "grad_norm": 1.1696518659591675, + "learning_rate": 9.54668524114997e-06, + "loss": 3.6851, + "step": 232700 + }, + { + "epoch": 2.6148306679474, + "grad_norm": 1.214743733406067, + "learning_rate": 9.544438327847121e-06, + "loss": 3.6141, + "step": 232750 + }, + { + "epoch": 2.615392393117743, + "grad_norm": 1.0277268886566162, + "learning_rate": 9.54219141454427e-06, + "loss": 3.6633, + "step": 232800 + }, + { + "epoch": 2.6159541182880863, + "grad_norm": 1.1339294910430908, + "learning_rate": 9.53994450124142e-06, + "loss": 3.638, + "step": 232850 + }, + { + "epoch": 2.6165158434584295, + "grad_norm": 1.1576228141784668, + "learning_rate": 9.537697587938571e-06, + "loss": 3.5904, + "step": 232900 + }, + { + "epoch": 2.6170775686287726, + "grad_norm": 1.4079818725585938, + "learning_rate": 9.53545067463572e-06, + "loss": 3.6654, + "step": 232950 + }, + { + "epoch": 2.6176392937991158, + "grad_norm": 1.0673800706863403, + "learning_rate": 9.53320376133287e-06, + "loss": 3.6718, + "step": 233000 + }, + { + "epoch": 2.618201018969459, + "grad_norm": 1.1846091747283936, + "learning_rate": 9.530956848030019e-06, + "loss": 3.648, + "step": 233050 + }, + { + "epoch": 2.618762744139802, + "grad_norm": 1.1678266525268555, + "learning_rate": 9.52870993472717e-06, + "loss": 3.5739, + "step": 233100 + }, + { + "epoch": 2.619324469310145, + "grad_norm": 1.2650402784347534, + "learning_rate": 9.52646302142432e-06, + "loss": 3.6458, + "step": 233150 + }, + { + "epoch": 2.6198861944804883, + "grad_norm": 1.2277120351791382, + "learning_rate": 9.524216108121469e-06, + "loss": 3.6615, + "step": 233200 + }, + { + "epoch": 2.6204479196508315, + "grad_norm": 1.2441250085830688, + "learning_rate": 9.52196919481862e-06, + "loss": 3.6584, + "step": 233250 + }, + { + "epoch": 2.621009644821175, + "grad_norm": 1.3057355880737305, + "learning_rate": 9.519722281515769e-06, + "loss": 3.6591, + "step": 233300 + }, + { + "epoch": 2.6215713699915177, + "grad_norm": 1.1118351221084595, + "learning_rate": 9.517475368212918e-06, + "loss": 3.6061, + "step": 233350 + }, + { + "epoch": 2.6221330951618613, + "grad_norm": 1.1428354978561401, + "learning_rate": 9.515228454910068e-06, + "loss": 3.6541, + "step": 233400 + }, + { + "epoch": 2.622694820332204, + "grad_norm": 1.2182830572128296, + "learning_rate": 9.512981541607219e-06, + "loss": 3.6867, + "step": 233450 + }, + { + "epoch": 2.6232565455025476, + "grad_norm": 1.158341884613037, + "learning_rate": 9.510734628304368e-06, + "loss": 3.6546, + "step": 233500 + }, + { + "epoch": 2.6238182706728903, + "grad_norm": 1.13567054271698, + "learning_rate": 9.508487715001517e-06, + "loss": 3.6903, + "step": 233550 + }, + { + "epoch": 2.624379995843234, + "grad_norm": 1.0206648111343384, + "learning_rate": 9.506240801698668e-06, + "loss": 3.5835, + "step": 233600 + }, + { + "epoch": 2.624941721013577, + "grad_norm": 1.0175087451934814, + "learning_rate": 9.503993888395817e-06, + "loss": 3.7287, + "step": 233650 + }, + { + "epoch": 2.62550344618392, + "grad_norm": 0.9948505163192749, + "learning_rate": 9.501746975092967e-06, + "loss": 3.6272, + "step": 233700 + }, + { + "epoch": 2.6260651713542633, + "grad_norm": 1.4096300601959229, + "learning_rate": 9.499500061790116e-06, + "loss": 3.6122, + "step": 233750 + }, + { + "epoch": 2.6266268965246065, + "grad_norm": 1.0361156463623047, + "learning_rate": 9.497253148487267e-06, + "loss": 3.6412, + "step": 233800 + }, + { + "epoch": 2.6271886216949496, + "grad_norm": 1.1061748266220093, + "learning_rate": 9.495006235184416e-06, + "loss": 3.5806, + "step": 233850 + }, + { + "epoch": 2.6277503468652927, + "grad_norm": 1.3449162244796753, + "learning_rate": 9.492759321881566e-06, + "loss": 3.5348, + "step": 233900 + }, + { + "epoch": 2.628312072035636, + "grad_norm": 1.7299484014511108, + "learning_rate": 9.490512408578717e-06, + "loss": 3.6063, + "step": 233950 + }, + { + "epoch": 2.628873797205979, + "grad_norm": 1.1820323467254639, + "learning_rate": 9.488265495275866e-06, + "loss": 3.6409, + "step": 234000 + }, + { + "epoch": 2.629435522376322, + "grad_norm": 1.146406888961792, + "learning_rate": 9.486063520239072e-06, + "loss": 3.731, + "step": 234050 + }, + { + "epoch": 2.6299972475466653, + "grad_norm": 1.1521809101104736, + "learning_rate": 9.483816606936223e-06, + "loss": 3.6015, + "step": 234100 + }, + { + "epoch": 2.6305589727170084, + "grad_norm": 1.2219223976135254, + "learning_rate": 9.481569693633372e-06, + "loss": 3.6379, + "step": 234150 + }, + { + "epoch": 2.6311206978873516, + "grad_norm": 1.103837013244629, + "learning_rate": 9.479322780330522e-06, + "loss": 3.641, + "step": 234200 + }, + { + "epoch": 2.6316824230576947, + "grad_norm": 1.1956130266189575, + "learning_rate": 9.477075867027673e-06, + "loss": 3.5529, + "step": 234250 + }, + { + "epoch": 2.632244148228038, + "grad_norm": 1.3069959878921509, + "learning_rate": 9.474828953724822e-06, + "loss": 3.606, + "step": 234300 + }, + { + "epoch": 2.632805873398381, + "grad_norm": 1.2597901821136475, + "learning_rate": 9.472582040421971e-06, + "loss": 3.6377, + "step": 234350 + }, + { + "epoch": 2.633367598568724, + "grad_norm": 1.0955803394317627, + "learning_rate": 9.47033512711912e-06, + "loss": 3.7162, + "step": 234400 + }, + { + "epoch": 2.6339293237390673, + "grad_norm": 1.1958327293395996, + "learning_rate": 9.468088213816272e-06, + "loss": 3.5871, + "step": 234450 + }, + { + "epoch": 2.6344910489094104, + "grad_norm": 1.2950828075408936, + "learning_rate": 9.465841300513421e-06, + "loss": 3.6441, + "step": 234500 + }, + { + "epoch": 2.6350527740797536, + "grad_norm": 0.989854633808136, + "learning_rate": 9.46359438721057e-06, + "loss": 3.696, + "step": 234550 + }, + { + "epoch": 2.6356144992500967, + "grad_norm": 1.1626341342926025, + "learning_rate": 9.461347473907721e-06, + "loss": 3.653, + "step": 234600 + }, + { + "epoch": 2.6361762244204403, + "grad_norm": 1.0983715057373047, + "learning_rate": 9.45910056060487e-06, + "loss": 3.6619, + "step": 234650 + }, + { + "epoch": 2.636737949590783, + "grad_norm": 1.0216072797775269, + "learning_rate": 9.45685364730202e-06, + "loss": 3.6305, + "step": 234700 + }, + { + "epoch": 2.6372996747611266, + "grad_norm": 1.0913469791412354, + "learning_rate": 9.45460673399917e-06, + "loss": 3.6289, + "step": 234750 + }, + { + "epoch": 2.6378613999314693, + "grad_norm": 1.5100804567337036, + "learning_rate": 9.45235982069632e-06, + "loss": 3.6328, + "step": 234800 + }, + { + "epoch": 2.638423125101813, + "grad_norm": 1.1998149156570435, + "learning_rate": 9.45011290739347e-06, + "loss": 3.7158, + "step": 234850 + }, + { + "epoch": 2.6389848502721556, + "grad_norm": 1.3535040616989136, + "learning_rate": 9.447865994090619e-06, + "loss": 3.6558, + "step": 234900 + }, + { + "epoch": 2.639546575442499, + "grad_norm": 1.1371439695358276, + "learning_rate": 9.445619080787768e-06, + "loss": 3.6966, + "step": 234950 + }, + { + "epoch": 2.6401083006128423, + "grad_norm": 1.214625358581543, + "learning_rate": 9.44337216748492e-06, + "loss": 3.6085, + "step": 235000 + }, + { + "epoch": 2.6406700257831854, + "grad_norm": 1.1851134300231934, + "learning_rate": 9.441125254182069e-06, + "loss": 3.6476, + "step": 235050 + }, + { + "epoch": 2.6412317509535286, + "grad_norm": 1.106297254562378, + "learning_rate": 9.438878340879218e-06, + "loss": 3.6427, + "step": 235100 + }, + { + "epoch": 2.6417934761238717, + "grad_norm": 1.156652808189392, + "learning_rate": 9.436631427576369e-06, + "loss": 3.6887, + "step": 235150 + }, + { + "epoch": 2.642355201294215, + "grad_norm": 1.2252565622329712, + "learning_rate": 9.434384514273518e-06, + "loss": 3.6339, + "step": 235200 + }, + { + "epoch": 2.642916926464558, + "grad_norm": 0.9991495013237, + "learning_rate": 9.432137600970667e-06, + "loss": 3.591, + "step": 235250 + }, + { + "epoch": 2.643478651634901, + "grad_norm": 1.096312403678894, + "learning_rate": 9.429890687667817e-06, + "loss": 3.6028, + "step": 235300 + }, + { + "epoch": 2.6440403768052443, + "grad_norm": 1.1620893478393555, + "learning_rate": 9.427643774364968e-06, + "loss": 3.6738, + "step": 235350 + }, + { + "epoch": 2.6446021019755874, + "grad_norm": 1.4076168537139893, + "learning_rate": 9.425396861062117e-06, + "loss": 3.6402, + "step": 235400 + }, + { + "epoch": 2.6451638271459306, + "grad_norm": 1.1057208776474, + "learning_rate": 9.423149947759266e-06, + "loss": 3.6549, + "step": 235450 + }, + { + "epoch": 2.6457255523162737, + "grad_norm": 1.0851587057113647, + "learning_rate": 9.420903034456417e-06, + "loss": 3.6544, + "step": 235500 + }, + { + "epoch": 2.646287277486617, + "grad_norm": 1.1699048280715942, + "learning_rate": 9.418656121153567e-06, + "loss": 3.6056, + "step": 235550 + }, + { + "epoch": 2.64684900265696, + "grad_norm": 1.1354840993881226, + "learning_rate": 9.416409207850716e-06, + "loss": 3.6982, + "step": 235600 + }, + { + "epoch": 2.647410727827303, + "grad_norm": 1.257293701171875, + "learning_rate": 9.414162294547865e-06, + "loss": 3.6762, + "step": 235650 + }, + { + "epoch": 2.6479724529976463, + "grad_norm": 1.1046998500823975, + "learning_rate": 9.411960319511071e-06, + "loss": 3.6089, + "step": 235700 + }, + { + "epoch": 2.6485341781679894, + "grad_norm": 1.2434829473495483, + "learning_rate": 9.409713406208222e-06, + "loss": 3.6709, + "step": 235750 + }, + { + "epoch": 2.6490959033383326, + "grad_norm": 1.1442354917526245, + "learning_rate": 9.407466492905372e-06, + "loss": 3.5999, + "step": 235800 + }, + { + "epoch": 2.6496576285086757, + "grad_norm": 1.150447964668274, + "learning_rate": 9.405219579602521e-06, + "loss": 3.6673, + "step": 235850 + }, + { + "epoch": 2.650219353679019, + "grad_norm": 1.1089730262756348, + "learning_rate": 9.402972666299672e-06, + "loss": 3.5911, + "step": 235900 + }, + { + "epoch": 2.650781078849362, + "grad_norm": 1.2682679891586304, + "learning_rate": 9.400725752996821e-06, + "loss": 3.5348, + "step": 235950 + }, + { + "epoch": 2.6513428040197056, + "grad_norm": 1.3281710147857666, + "learning_rate": 9.39847883969397e-06, + "loss": 3.6362, + "step": 236000 + }, + { + "epoch": 2.6519045291900483, + "grad_norm": 1.3261010646820068, + "learning_rate": 9.39623192639112e-06, + "loss": 3.6157, + "step": 236050 + }, + { + "epoch": 2.652466254360392, + "grad_norm": 1.2928086519241333, + "learning_rate": 9.393985013088271e-06, + "loss": 3.6542, + "step": 236100 + }, + { + "epoch": 2.6530279795307345, + "grad_norm": 1.4209024906158447, + "learning_rate": 9.39173809978542e-06, + "loss": 3.7344, + "step": 236150 + }, + { + "epoch": 2.653589704701078, + "grad_norm": 1.31272554397583, + "learning_rate": 9.38949118648257e-06, + "loss": 3.7201, + "step": 236200 + }, + { + "epoch": 2.6541514298714213, + "grad_norm": 1.176992654800415, + "learning_rate": 9.38724427317972e-06, + "loss": 3.6173, + "step": 236250 + }, + { + "epoch": 2.6547131550417644, + "grad_norm": 1.160933256149292, + "learning_rate": 9.38499735987687e-06, + "loss": 3.642, + "step": 236300 + }, + { + "epoch": 2.6552748802121076, + "grad_norm": 1.5851428508758545, + "learning_rate": 9.38275044657402e-06, + "loss": 3.7491, + "step": 236350 + }, + { + "epoch": 2.6558366053824507, + "grad_norm": 1.163303256034851, + "learning_rate": 9.380503533271169e-06, + "loss": 3.63, + "step": 236400 + }, + { + "epoch": 2.656398330552794, + "grad_norm": 1.2374213933944702, + "learning_rate": 9.37825661996832e-06, + "loss": 3.6488, + "step": 236450 + }, + { + "epoch": 2.656960055723137, + "grad_norm": 1.2023411989212036, + "learning_rate": 9.376009706665469e-06, + "loss": 3.6569, + "step": 236500 + }, + { + "epoch": 2.65752178089348, + "grad_norm": 1.4848881959915161, + "learning_rate": 9.373762793362618e-06, + "loss": 3.6683, + "step": 236550 + }, + { + "epoch": 2.6580835060638233, + "grad_norm": 2.022050142288208, + "learning_rate": 9.371515880059767e-06, + "loss": 3.6196, + "step": 236600 + }, + { + "epoch": 2.6586452312341664, + "grad_norm": 1.1171576976776123, + "learning_rate": 9.369268966756918e-06, + "loss": 3.5738, + "step": 236650 + }, + { + "epoch": 2.6592069564045095, + "grad_norm": 1.1062504053115845, + "learning_rate": 9.367022053454068e-06, + "loss": 3.7431, + "step": 236700 + }, + { + "epoch": 2.6597686815748527, + "grad_norm": 1.2124685049057007, + "learning_rate": 9.364775140151217e-06, + "loss": 3.575, + "step": 236750 + }, + { + "epoch": 2.660330406745196, + "grad_norm": 1.1376625299453735, + "learning_rate": 9.362528226848368e-06, + "loss": 3.6897, + "step": 236800 + }, + { + "epoch": 2.660892131915539, + "grad_norm": 1.3555550575256348, + "learning_rate": 9.360281313545517e-06, + "loss": 3.6666, + "step": 236850 + }, + { + "epoch": 2.661453857085882, + "grad_norm": 3.010193109512329, + "learning_rate": 9.358034400242667e-06, + "loss": 3.6706, + "step": 236900 + }, + { + "epoch": 2.6620155822562253, + "grad_norm": 1.2489869594573975, + "learning_rate": 9.355787486939816e-06, + "loss": 3.6344, + "step": 236950 + }, + { + "epoch": 2.6625773074265684, + "grad_norm": 1.212320327758789, + "learning_rate": 9.353540573636967e-06, + "loss": 3.6851, + "step": 237000 + }, + { + "epoch": 2.6631390325969115, + "grad_norm": 1.0101131200790405, + "learning_rate": 9.351293660334116e-06, + "loss": 3.6492, + "step": 237050 + }, + { + "epoch": 2.6637007577672547, + "grad_norm": 1.0726019144058228, + "learning_rate": 9.349046747031266e-06, + "loss": 3.6297, + "step": 237100 + }, + { + "epoch": 2.664262482937598, + "grad_norm": 1.1873642206192017, + "learning_rate": 9.346799833728417e-06, + "loss": 3.6453, + "step": 237150 + }, + { + "epoch": 2.664824208107941, + "grad_norm": 1.2106003761291504, + "learning_rate": 9.344552920425566e-06, + "loss": 3.6412, + "step": 237200 + }, + { + "epoch": 2.665385933278284, + "grad_norm": 1.0715569257736206, + "learning_rate": 9.342306007122715e-06, + "loss": 3.6722, + "step": 237250 + }, + { + "epoch": 2.6659476584486272, + "grad_norm": 1.1939773559570312, + "learning_rate": 9.340059093819865e-06, + "loss": 3.6001, + "step": 237300 + }, + { + "epoch": 2.666509383618971, + "grad_norm": 1.160776138305664, + "learning_rate": 9.337812180517016e-06, + "loss": 3.6837, + "step": 237350 + }, + { + "epoch": 2.6670711087893135, + "grad_norm": 1.4269455671310425, + "learning_rate": 9.335565267214165e-06, + "loss": 3.7343, + "step": 237400 + }, + { + "epoch": 2.667632833959657, + "grad_norm": 1.1893212795257568, + "learning_rate": 9.333318353911314e-06, + "loss": 3.714, + "step": 237450 + }, + { + "epoch": 2.66819455913, + "grad_norm": 1.197709321975708, + "learning_rate": 9.331071440608465e-06, + "loss": 3.6083, + "step": 237500 + }, + { + "epoch": 2.6687562843003434, + "grad_norm": 1.1480973958969116, + "learning_rate": 9.328824527305615e-06, + "loss": 3.6273, + "step": 237550 + }, + { + "epoch": 2.6693180094706865, + "grad_norm": 1.0416712760925293, + "learning_rate": 9.326577614002764e-06, + "loss": 3.6305, + "step": 237600 + }, + { + "epoch": 2.6698797346410297, + "grad_norm": 1.234265923500061, + "learning_rate": 9.324330700699913e-06, + "loss": 3.6808, + "step": 237650 + }, + { + "epoch": 2.670441459811373, + "grad_norm": 1.2496914863586426, + "learning_rate": 9.322083787397064e-06, + "loss": 3.5479, + "step": 237700 + }, + { + "epoch": 2.671003184981716, + "grad_norm": 1.355756163597107, + "learning_rate": 9.319836874094214e-06, + "loss": 3.5549, + "step": 237750 + }, + { + "epoch": 2.671564910152059, + "grad_norm": 1.4441754817962646, + "learning_rate": 9.317589960791363e-06, + "loss": 3.6918, + "step": 237800 + }, + { + "epoch": 2.6721266353224022, + "grad_norm": 1.2543582916259766, + "learning_rate": 9.315343047488514e-06, + "loss": 3.6524, + "step": 237850 + }, + { + "epoch": 2.6726883604927454, + "grad_norm": 1.2836605310440063, + "learning_rate": 9.31314107245172e-06, + "loss": 3.6425, + "step": 237900 + }, + { + "epoch": 2.6732500856630885, + "grad_norm": 1.5076261758804321, + "learning_rate": 9.31089415914887e-06, + "loss": 3.6275, + "step": 237950 + }, + { + "epoch": 2.6738118108334317, + "grad_norm": 1.1220850944519043, + "learning_rate": 9.30864724584602e-06, + "loss": 3.6343, + "step": 238000 + }, + { + "epoch": 2.674373536003775, + "grad_norm": 1.0616127252578735, + "learning_rate": 9.30640033254317e-06, + "loss": 3.6229, + "step": 238050 + }, + { + "epoch": 2.674935261174118, + "grad_norm": 1.0698143243789673, + "learning_rate": 9.304153419240319e-06, + "loss": 3.5899, + "step": 238100 + }, + { + "epoch": 2.675496986344461, + "grad_norm": 1.200904369354248, + "learning_rate": 9.30190650593747e-06, + "loss": 3.6681, + "step": 238150 + }, + { + "epoch": 2.6760587115148042, + "grad_norm": 1.0132920742034912, + "learning_rate": 9.299659592634619e-06, + "loss": 3.6474, + "step": 238200 + }, + { + "epoch": 2.6766204366851474, + "grad_norm": 1.3051836490631104, + "learning_rate": 9.297412679331768e-06, + "loss": 3.609, + "step": 238250 + }, + { + "epoch": 2.6771821618554905, + "grad_norm": 1.1440683603286743, + "learning_rate": 9.295165766028918e-06, + "loss": 3.6477, + "step": 238300 + }, + { + "epoch": 2.6777438870258337, + "grad_norm": 0.9804529547691345, + "learning_rate": 9.292918852726069e-06, + "loss": 3.6513, + "step": 238350 + }, + { + "epoch": 2.678305612196177, + "grad_norm": 1.2794501781463623, + "learning_rate": 9.290671939423218e-06, + "loss": 3.6358, + "step": 238400 + }, + { + "epoch": 2.67886733736652, + "grad_norm": 1.1855015754699707, + "learning_rate": 9.288425026120367e-06, + "loss": 3.7059, + "step": 238450 + }, + { + "epoch": 2.679429062536863, + "grad_norm": 1.282378077507019, + "learning_rate": 9.286178112817518e-06, + "loss": 3.6186, + "step": 238500 + }, + { + "epoch": 2.6799907877072062, + "grad_norm": 1.1567116975784302, + "learning_rate": 9.283931199514668e-06, + "loss": 3.621, + "step": 238550 + }, + { + "epoch": 2.6805525128775494, + "grad_norm": 1.2169396877288818, + "learning_rate": 9.281684286211817e-06, + "loss": 3.581, + "step": 238600 + }, + { + "epoch": 2.6811142380478925, + "grad_norm": 1.1934642791748047, + "learning_rate": 9.279437372908966e-06, + "loss": 3.6494, + "step": 238650 + }, + { + "epoch": 2.681675963218236, + "grad_norm": 1.1107144355773926, + "learning_rate": 9.277190459606117e-06, + "loss": 3.5958, + "step": 238700 + }, + { + "epoch": 2.682237688388579, + "grad_norm": 1.163450002670288, + "learning_rate": 9.274943546303267e-06, + "loss": 3.6176, + "step": 238750 + }, + { + "epoch": 2.6827994135589224, + "grad_norm": 1.4309276342391968, + "learning_rate": 9.272696633000416e-06, + "loss": 3.6634, + "step": 238800 + }, + { + "epoch": 2.683361138729265, + "grad_norm": 1.1508939266204834, + "learning_rate": 9.270449719697567e-06, + "loss": 3.6055, + "step": 238850 + }, + { + "epoch": 2.6839228638996087, + "grad_norm": 1.1272213459014893, + "learning_rate": 9.268202806394716e-06, + "loss": 3.608, + "step": 238900 + }, + { + "epoch": 2.684484589069952, + "grad_norm": 1.311135172843933, + "learning_rate": 9.265955893091866e-06, + "loss": 3.6426, + "step": 238950 + }, + { + "epoch": 2.685046314240295, + "grad_norm": 1.0856311321258545, + "learning_rate": 9.263708979789015e-06, + "loss": 3.5447, + "step": 239000 + }, + { + "epoch": 2.685608039410638, + "grad_norm": 1.298883318901062, + "learning_rate": 9.261462066486166e-06, + "loss": 3.6805, + "step": 239050 + }, + { + "epoch": 2.6861697645809812, + "grad_norm": 1.2146790027618408, + "learning_rate": 9.259215153183315e-06, + "loss": 3.6835, + "step": 239100 + }, + { + "epoch": 2.6867314897513244, + "grad_norm": 1.240931510925293, + "learning_rate": 9.256968239880465e-06, + "loss": 3.5861, + "step": 239150 + }, + { + "epoch": 2.6872932149216675, + "grad_norm": 1.3513401746749878, + "learning_rate": 9.254721326577614e-06, + "loss": 3.6177, + "step": 239200 + }, + { + "epoch": 2.6878549400920106, + "grad_norm": 1.0338592529296875, + "learning_rate": 9.252474413274765e-06, + "loss": 3.5875, + "step": 239250 + }, + { + "epoch": 2.688416665262354, + "grad_norm": 1.1467785835266113, + "learning_rate": 9.250227499971914e-06, + "loss": 3.6338, + "step": 239300 + }, + { + "epoch": 2.688978390432697, + "grad_norm": 1.279463291168213, + "learning_rate": 9.247980586669064e-06, + "loss": 3.6839, + "step": 239350 + }, + { + "epoch": 2.68954011560304, + "grad_norm": 1.1546107530593872, + "learning_rate": 9.245733673366215e-06, + "loss": 3.6859, + "step": 239400 + }, + { + "epoch": 2.690101840773383, + "grad_norm": 1.0767256021499634, + "learning_rate": 9.243486760063364e-06, + "loss": 3.6337, + "step": 239450 + }, + { + "epoch": 2.6906635659437264, + "grad_norm": 1.2261996269226074, + "learning_rate": 9.241239846760513e-06, + "loss": 3.6972, + "step": 239500 + }, + { + "epoch": 2.6912252911140695, + "grad_norm": 1.0836470127105713, + "learning_rate": 9.238992933457663e-06, + "loss": 3.6677, + "step": 239550 + }, + { + "epoch": 2.6917870162844126, + "grad_norm": 1.1737556457519531, + "learning_rate": 9.236746020154814e-06, + "loss": 3.6386, + "step": 239600 + }, + { + "epoch": 2.692348741454756, + "grad_norm": 1.1945728063583374, + "learning_rate": 9.234499106851963e-06, + "loss": 3.6118, + "step": 239650 + }, + { + "epoch": 2.692910466625099, + "grad_norm": 1.3617130517959595, + "learning_rate": 9.232252193549112e-06, + "loss": 3.5948, + "step": 239700 + }, + { + "epoch": 2.693472191795442, + "grad_norm": 1.123518943786621, + "learning_rate": 9.230005280246263e-06, + "loss": 3.6197, + "step": 239750 + }, + { + "epoch": 2.694033916965785, + "grad_norm": 1.2159085273742676, + "learning_rate": 9.227758366943413e-06, + "loss": 3.5417, + "step": 239800 + }, + { + "epoch": 2.6945956421361283, + "grad_norm": 1.347956657409668, + "learning_rate": 9.225511453640562e-06, + "loss": 3.7454, + "step": 239850 + }, + { + "epoch": 2.6951573673064715, + "grad_norm": 1.34463632106781, + "learning_rate": 9.223264540337711e-06, + "loss": 3.6986, + "step": 239900 + }, + { + "epoch": 2.695719092476815, + "grad_norm": 1.1763185262680054, + "learning_rate": 9.221017627034862e-06, + "loss": 3.5904, + "step": 239950 + }, + { + "epoch": 2.6962808176471578, + "grad_norm": 1.2356795072555542, + "learning_rate": 9.218770713732011e-06, + "loss": 3.6617, + "step": 240000 + }, + { + "epoch": 2.6968425428175014, + "grad_norm": 2.081676721572876, + "learning_rate": 9.21652380042916e-06, + "loss": 3.6598, + "step": 240050 + }, + { + "epoch": 2.697404267987844, + "grad_norm": 1.215590238571167, + "learning_rate": 9.214276887126312e-06, + "loss": 3.6284, + "step": 240100 + }, + { + "epoch": 2.6979659931581876, + "grad_norm": 1.2818858623504639, + "learning_rate": 9.212029973823461e-06, + "loss": 3.7313, + "step": 240150 + }, + { + "epoch": 2.6985277183285303, + "grad_norm": 1.0892887115478516, + "learning_rate": 9.20978306052061e-06, + "loss": 3.6084, + "step": 240200 + }, + { + "epoch": 2.699089443498874, + "grad_norm": 1.2128300666809082, + "learning_rate": 9.20753614721776e-06, + "loss": 3.6177, + "step": 240250 + }, + { + "epoch": 2.699651168669217, + "grad_norm": 1.139426589012146, + "learning_rate": 9.20528923391491e-06, + "loss": 3.5944, + "step": 240300 + }, + { + "epoch": 2.70021289383956, + "grad_norm": 1.1543854475021362, + "learning_rate": 9.20304232061206e-06, + "loss": 3.6371, + "step": 240350 + }, + { + "epoch": 2.7007746190099033, + "grad_norm": 1.1960457563400269, + "learning_rate": 9.20079540730921e-06, + "loss": 3.6425, + "step": 240400 + }, + { + "epoch": 2.7013363441802465, + "grad_norm": 1.123557209968567, + "learning_rate": 9.19854849400636e-06, + "loss": 3.6176, + "step": 240450 + }, + { + "epoch": 2.7018980693505896, + "grad_norm": 1.2426378726959229, + "learning_rate": 9.19630158070351e-06, + "loss": 3.578, + "step": 240500 + }, + { + "epoch": 2.7024597945209328, + "grad_norm": 1.2549535036087036, + "learning_rate": 9.194054667400659e-06, + "loss": 3.6038, + "step": 240550 + }, + { + "epoch": 2.703021519691276, + "grad_norm": 1.2187061309814453, + "learning_rate": 9.191807754097808e-06, + "loss": 3.7156, + "step": 240600 + }, + { + "epoch": 2.703583244861619, + "grad_norm": 1.3443732261657715, + "learning_rate": 9.18956084079496e-06, + "loss": 3.72, + "step": 240650 + }, + { + "epoch": 2.704144970031962, + "grad_norm": 0.9994232058525085, + "learning_rate": 9.187313927492109e-06, + "loss": 3.6145, + "step": 240700 + }, + { + "epoch": 2.7047066952023053, + "grad_norm": 1.1979265213012695, + "learning_rate": 9.185067014189258e-06, + "loss": 3.6338, + "step": 240750 + }, + { + "epoch": 2.7052684203726485, + "grad_norm": 1.807159662246704, + "learning_rate": 9.182820100886409e-06, + "loss": 3.6061, + "step": 240800 + }, + { + "epoch": 2.7058301455429916, + "grad_norm": 1.0491582155227661, + "learning_rate": 9.180573187583558e-06, + "loss": 3.6531, + "step": 240850 + }, + { + "epoch": 2.7063918707133348, + "grad_norm": 1.0936228036880493, + "learning_rate": 9.178326274280708e-06, + "loss": 3.6325, + "step": 240900 + }, + { + "epoch": 2.706953595883678, + "grad_norm": 1.1391863822937012, + "learning_rate": 9.176079360977857e-06, + "loss": 3.5716, + "step": 240950 + }, + { + "epoch": 2.707515321054021, + "grad_norm": 1.4656747579574585, + "learning_rate": 9.173832447675008e-06, + "loss": 3.592, + "step": 241000 + }, + { + "epoch": 2.708077046224364, + "grad_norm": 1.1552832126617432, + "learning_rate": 9.171585534372157e-06, + "loss": 3.5734, + "step": 241050 + }, + { + "epoch": 2.7086387713947073, + "grad_norm": 1.0767759084701538, + "learning_rate": 9.169338621069307e-06, + "loss": 3.7697, + "step": 241100 + }, + { + "epoch": 2.7092004965650505, + "grad_norm": 1.158240795135498, + "learning_rate": 9.167091707766456e-06, + "loss": 3.6753, + "step": 241150 + }, + { + "epoch": 2.7097622217353936, + "grad_norm": 1.0932152271270752, + "learning_rate": 9.164844794463607e-06, + "loss": 3.7098, + "step": 241200 + }, + { + "epoch": 2.7103239469057367, + "grad_norm": 1.2083481550216675, + "learning_rate": 9.162597881160756e-06, + "loss": 3.6069, + "step": 241250 + }, + { + "epoch": 2.7108856720760803, + "grad_norm": 1.191329002380371, + "learning_rate": 9.160350967857906e-06, + "loss": 3.5987, + "step": 241300 + }, + { + "epoch": 2.711447397246423, + "grad_norm": 1.1355061531066895, + "learning_rate": 9.158104054555057e-06, + "loss": 3.5238, + "step": 241350 + }, + { + "epoch": 2.7120091224167666, + "grad_norm": 1.117626667022705, + "learning_rate": 9.155857141252206e-06, + "loss": 3.6962, + "step": 241400 + }, + { + "epoch": 2.7125708475871093, + "grad_norm": 1.2261159420013428, + "learning_rate": 9.153610227949355e-06, + "loss": 3.6318, + "step": 241450 + }, + { + "epoch": 2.713132572757453, + "grad_norm": 1.1712418794631958, + "learning_rate": 9.151363314646505e-06, + "loss": 3.6104, + "step": 241500 + }, + { + "epoch": 2.7136942979277956, + "grad_norm": 1.111906886100769, + "learning_rate": 9.149116401343656e-06, + "loss": 3.5495, + "step": 241550 + }, + { + "epoch": 2.714256023098139, + "grad_norm": 1.1478056907653809, + "learning_rate": 9.146869488040805e-06, + "loss": 3.6877, + "step": 241600 + }, + { + "epoch": 2.7148177482684823, + "grad_norm": 1.1401646137237549, + "learning_rate": 9.144622574737954e-06, + "loss": 3.6927, + "step": 241650 + }, + { + "epoch": 2.7153794734388255, + "grad_norm": 1.0862573385238647, + "learning_rate": 9.142375661435105e-06, + "loss": 3.6816, + "step": 241700 + }, + { + "epoch": 2.7159411986091686, + "grad_norm": 1.3958070278167725, + "learning_rate": 9.140128748132255e-06, + "loss": 3.6334, + "step": 241750 + }, + { + "epoch": 2.7165029237795117, + "grad_norm": 1.2459132671356201, + "learning_rate": 9.137881834829404e-06, + "loss": 3.6812, + "step": 241800 + }, + { + "epoch": 2.717064648949855, + "grad_norm": 1.3747844696044922, + "learning_rate": 9.135634921526553e-06, + "loss": 3.6143, + "step": 241850 + }, + { + "epoch": 2.717626374120198, + "grad_norm": 1.1211518049240112, + "learning_rate": 9.133388008223704e-06, + "loss": 3.5938, + "step": 241900 + }, + { + "epoch": 2.718188099290541, + "grad_norm": 1.055198073387146, + "learning_rate": 9.131141094920853e-06, + "loss": 3.6315, + "step": 241950 + }, + { + "epoch": 2.7187498244608843, + "grad_norm": 1.279075264930725, + "learning_rate": 9.128894181618003e-06, + "loss": 3.7427, + "step": 242000 + }, + { + "epoch": 2.7193115496312275, + "grad_norm": 1.2764612436294556, + "learning_rate": 9.126647268315154e-06, + "loss": 3.6734, + "step": 242050 + }, + { + "epoch": 2.7198732748015706, + "grad_norm": 1.2446975708007812, + "learning_rate": 9.124400355012303e-06, + "loss": 3.6952, + "step": 242100 + }, + { + "epoch": 2.7204349999719137, + "grad_norm": 1.241504430770874, + "learning_rate": 9.122153441709452e-06, + "loss": 3.6832, + "step": 242150 + }, + { + "epoch": 2.720996725142257, + "grad_norm": 1.3896424770355225, + "learning_rate": 9.119906528406602e-06, + "loss": 3.6839, + "step": 242200 + }, + { + "epoch": 2.7215584503126, + "grad_norm": 1.1340253353118896, + "learning_rate": 9.117659615103753e-06, + "loss": 3.6136, + "step": 242250 + }, + { + "epoch": 2.722120175482943, + "grad_norm": 1.163301944732666, + "learning_rate": 9.115412701800902e-06, + "loss": 3.5696, + "step": 242300 + }, + { + "epoch": 2.7226819006532863, + "grad_norm": 1.260135531425476, + "learning_rate": 9.113165788498051e-06, + "loss": 3.5426, + "step": 242350 + }, + { + "epoch": 2.7232436258236294, + "grad_norm": 1.1422429084777832, + "learning_rate": 9.110918875195202e-06, + "loss": 3.7006, + "step": 242400 + }, + { + "epoch": 2.7238053509939726, + "grad_norm": 1.1454027891159058, + "learning_rate": 9.108671961892352e-06, + "loss": 3.5909, + "step": 242450 + }, + { + "epoch": 2.7243670761643157, + "grad_norm": 1.1539671421051025, + "learning_rate": 9.106425048589501e-06, + "loss": 3.5875, + "step": 242500 + }, + { + "epoch": 2.724928801334659, + "grad_norm": 1.1983054876327515, + "learning_rate": 9.10417813528665e-06, + "loss": 3.6367, + "step": 242550 + }, + { + "epoch": 2.725490526505002, + "grad_norm": 2.16520094871521, + "learning_rate": 9.101931221983801e-06, + "loss": 3.5879, + "step": 242600 + }, + { + "epoch": 2.7260522516753456, + "grad_norm": 1.152925968170166, + "learning_rate": 9.09968430868095e-06, + "loss": 3.6638, + "step": 242650 + }, + { + "epoch": 2.7266139768456883, + "grad_norm": 1.5511201620101929, + "learning_rate": 9.0974373953781e-06, + "loss": 3.651, + "step": 242700 + }, + { + "epoch": 2.727175702016032, + "grad_norm": 1.3489755392074585, + "learning_rate": 9.09519048207525e-06, + "loss": 3.6506, + "step": 242750 + }, + { + "epoch": 2.7277374271863746, + "grad_norm": 1.2226667404174805, + "learning_rate": 9.0929435687724e-06, + "loss": 3.6772, + "step": 242800 + }, + { + "epoch": 2.728299152356718, + "grad_norm": 1.0761747360229492, + "learning_rate": 9.09069665546955e-06, + "loss": 3.6534, + "step": 242850 + }, + { + "epoch": 2.728860877527061, + "grad_norm": 1.2483551502227783, + "learning_rate": 9.088449742166699e-06, + "loss": 3.7027, + "step": 242900 + }, + { + "epoch": 2.7294226026974044, + "grad_norm": 1.4201072454452515, + "learning_rate": 9.08620282886385e-06, + "loss": 3.6614, + "step": 242950 + }, + { + "epoch": 2.7299843278677476, + "grad_norm": 1.1233983039855957, + "learning_rate": 9.083955915561e-06, + "loss": 3.6096, + "step": 243000 + }, + { + "epoch": 2.7305460530380907, + "grad_norm": 1.2304747104644775, + "learning_rate": 9.081709002258149e-06, + "loss": 3.6144, + "step": 243050 + }, + { + "epoch": 2.731107778208434, + "grad_norm": 1.2645460367202759, + "learning_rate": 9.079462088955298e-06, + "loss": 3.6092, + "step": 243100 + }, + { + "epoch": 2.731669503378777, + "grad_norm": 1.2786263227462769, + "learning_rate": 9.077215175652449e-06, + "loss": 3.5947, + "step": 243150 + }, + { + "epoch": 2.73223122854912, + "grad_norm": 1.286088228225708, + "learning_rate": 9.074968262349598e-06, + "loss": 3.6442, + "step": 243200 + }, + { + "epoch": 2.7327929537194633, + "grad_norm": 1.1732699871063232, + "learning_rate": 9.072721349046748e-06, + "loss": 3.6664, + "step": 243250 + }, + { + "epoch": 2.7333546788898064, + "grad_norm": 1.2338069677352905, + "learning_rate": 9.070474435743899e-06, + "loss": 3.6205, + "step": 243300 + }, + { + "epoch": 2.7339164040601496, + "grad_norm": 1.1352485418319702, + "learning_rate": 9.068227522441048e-06, + "loss": 3.6537, + "step": 243350 + }, + { + "epoch": 2.7344781292304927, + "grad_norm": 1.456569790840149, + "learning_rate": 9.065980609138197e-06, + "loss": 3.6952, + "step": 243400 + }, + { + "epoch": 2.735039854400836, + "grad_norm": 1.1619824171066284, + "learning_rate": 9.063733695835347e-06, + "loss": 3.6651, + "step": 243450 + }, + { + "epoch": 2.735601579571179, + "grad_norm": 1.2682414054870605, + "learning_rate": 9.061531720798554e-06, + "loss": 3.5823, + "step": 243500 + }, + { + "epoch": 2.736163304741522, + "grad_norm": 1.2526882886886597, + "learning_rate": 9.059284807495703e-06, + "loss": 3.6704, + "step": 243550 + }, + { + "epoch": 2.7367250299118653, + "grad_norm": 1.0567783117294312, + "learning_rate": 9.057037894192854e-06, + "loss": 3.6031, + "step": 243600 + }, + { + "epoch": 2.7372867550822084, + "grad_norm": 1.2400263547897339, + "learning_rate": 9.054790980890004e-06, + "loss": 3.6466, + "step": 243650 + }, + { + "epoch": 2.7378484802525516, + "grad_norm": 1.216665267944336, + "learning_rate": 9.052544067587153e-06, + "loss": 3.6579, + "step": 243700 + }, + { + "epoch": 2.7384102054228947, + "grad_norm": 1.3223114013671875, + "learning_rate": 9.050297154284302e-06, + "loss": 3.617, + "step": 243750 + }, + { + "epoch": 2.738971930593238, + "grad_norm": 1.1438169479370117, + "learning_rate": 9.048050240981453e-06, + "loss": 3.6198, + "step": 243800 + }, + { + "epoch": 2.739533655763581, + "grad_norm": 1.1510319709777832, + "learning_rate": 9.045803327678603e-06, + "loss": 3.6681, + "step": 243850 + }, + { + "epoch": 2.740095380933924, + "grad_norm": 1.2513720989227295, + "learning_rate": 9.043556414375752e-06, + "loss": 3.5245, + "step": 243900 + }, + { + "epoch": 2.7406571061042673, + "grad_norm": 1.0472750663757324, + "learning_rate": 9.041309501072903e-06, + "loss": 3.6285, + "step": 243950 + }, + { + "epoch": 2.741218831274611, + "grad_norm": 1.2637301683425903, + "learning_rate": 9.03906258777005e-06, + "loss": 3.6431, + "step": 244000 + }, + { + "epoch": 2.7417805564449536, + "grad_norm": 1.1572527885437012, + "learning_rate": 9.036815674467202e-06, + "loss": 3.6351, + "step": 244050 + }, + { + "epoch": 2.742342281615297, + "grad_norm": 1.313968539237976, + "learning_rate": 9.034568761164351e-06, + "loss": 3.6279, + "step": 244100 + }, + { + "epoch": 2.74290400678564, + "grad_norm": 1.2717763185501099, + "learning_rate": 9.0323218478615e-06, + "loss": 3.6438, + "step": 244150 + }, + { + "epoch": 2.7434657319559834, + "grad_norm": 1.3627235889434814, + "learning_rate": 9.03007493455865e-06, + "loss": 3.6907, + "step": 244200 + }, + { + "epoch": 2.744027457126326, + "grad_norm": 1.1469571590423584, + "learning_rate": 9.0278280212558e-06, + "loss": 3.6161, + "step": 244250 + }, + { + "epoch": 2.7445891822966697, + "grad_norm": 1.127825140953064, + "learning_rate": 9.025626046219007e-06, + "loss": 3.6382, + "step": 244300 + }, + { + "epoch": 2.745150907467013, + "grad_norm": 1.3997374773025513, + "learning_rate": 9.023379132916158e-06, + "loss": 3.6522, + "step": 244350 + }, + { + "epoch": 2.745712632637356, + "grad_norm": 1.309731125831604, + "learning_rate": 9.021132219613307e-06, + "loss": 3.6872, + "step": 244400 + }, + { + "epoch": 2.746274357807699, + "grad_norm": 1.3515589237213135, + "learning_rate": 9.018885306310456e-06, + "loss": 3.6513, + "step": 244450 + }, + { + "epoch": 2.7468360829780423, + "grad_norm": 1.3234989643096924, + "learning_rate": 9.016638393007606e-06, + "loss": 3.7034, + "step": 244500 + }, + { + "epoch": 2.7473978081483854, + "grad_norm": 1.0811554193496704, + "learning_rate": 9.014391479704757e-06, + "loss": 3.6828, + "step": 244550 + }, + { + "epoch": 2.7479595333187286, + "grad_norm": 1.234450340270996, + "learning_rate": 9.012144566401906e-06, + "loss": 3.7199, + "step": 244600 + }, + { + "epoch": 2.7485212584890717, + "grad_norm": 1.0604885816574097, + "learning_rate": 9.009897653099055e-06, + "loss": 3.6605, + "step": 244650 + }, + { + "epoch": 2.749082983659415, + "grad_norm": 1.215888500213623, + "learning_rate": 9.007650739796206e-06, + "loss": 3.5975, + "step": 244700 + }, + { + "epoch": 2.749644708829758, + "grad_norm": 1.1834688186645508, + "learning_rate": 9.005403826493356e-06, + "loss": 3.7012, + "step": 244750 + }, + { + "epoch": 2.750206434000101, + "grad_norm": 1.245441198348999, + "learning_rate": 9.003156913190505e-06, + "loss": 3.7003, + "step": 244800 + }, + { + "epoch": 2.7507681591704443, + "grad_norm": 1.0024648904800415, + "learning_rate": 9.000909999887654e-06, + "loss": 3.6394, + "step": 244850 + }, + { + "epoch": 2.7513298843407874, + "grad_norm": 1.1968562602996826, + "learning_rate": 8.998663086584805e-06, + "loss": 3.6342, + "step": 244900 + }, + { + "epoch": 2.7518916095111305, + "grad_norm": 1.2056657075881958, + "learning_rate": 8.996416173281954e-06, + "loss": 3.5545, + "step": 244950 + }, + { + "epoch": 2.7524533346814737, + "grad_norm": 1.3170746564865112, + "learning_rate": 8.994169259979104e-06, + "loss": 3.7121, + "step": 245000 + }, + { + "epoch": 2.753015059851817, + "grad_norm": 1.23115074634552, + "learning_rate": 8.991922346676255e-06, + "loss": 3.5795, + "step": 245050 + }, + { + "epoch": 2.75357678502216, + "grad_norm": 1.2126127481460571, + "learning_rate": 8.989675433373404e-06, + "loss": 3.5721, + "step": 245100 + }, + { + "epoch": 2.754138510192503, + "grad_norm": 1.1153799295425415, + "learning_rate": 8.987428520070553e-06, + "loss": 3.7114, + "step": 245150 + }, + { + "epoch": 2.7547002353628463, + "grad_norm": 1.2014546394348145, + "learning_rate": 8.985181606767703e-06, + "loss": 3.6424, + "step": 245200 + }, + { + "epoch": 2.7552619605331894, + "grad_norm": 1.1441682577133179, + "learning_rate": 8.982934693464854e-06, + "loss": 3.5702, + "step": 245250 + }, + { + "epoch": 2.7558236857035325, + "grad_norm": 1.2664340734481812, + "learning_rate": 8.980687780162003e-06, + "loss": 3.5781, + "step": 245300 + }, + { + "epoch": 2.756385410873876, + "grad_norm": 1.2148077487945557, + "learning_rate": 8.978440866859152e-06, + "loss": 3.6007, + "step": 245350 + }, + { + "epoch": 2.756947136044219, + "grad_norm": 1.3117756843566895, + "learning_rate": 8.976193953556302e-06, + "loss": 3.5995, + "step": 245400 + }, + { + "epoch": 2.7575088612145624, + "grad_norm": 1.3556278944015503, + "learning_rate": 8.973947040253453e-06, + "loss": 3.6414, + "step": 245450 + }, + { + "epoch": 2.758070586384905, + "grad_norm": 1.1827054023742676, + "learning_rate": 8.971700126950602e-06, + "loss": 3.6375, + "step": 245500 + }, + { + "epoch": 2.7586323115552487, + "grad_norm": 1.2030131816864014, + "learning_rate": 8.969453213647751e-06, + "loss": 3.671, + "step": 245550 + }, + { + "epoch": 2.7591940367255914, + "grad_norm": 1.1759666204452515, + "learning_rate": 8.967206300344902e-06, + "loss": 3.6656, + "step": 245600 + }, + { + "epoch": 2.759755761895935, + "grad_norm": 1.3242117166519165, + "learning_rate": 8.964959387042052e-06, + "loss": 3.6524, + "step": 245650 + }, + { + "epoch": 2.760317487066278, + "grad_norm": 1.2978301048278809, + "learning_rate": 8.962712473739201e-06, + "loss": 3.6528, + "step": 245700 + }, + { + "epoch": 2.7608792122366212, + "grad_norm": 1.304042100906372, + "learning_rate": 8.96046556043635e-06, + "loss": 3.6286, + "step": 245750 + }, + { + "epoch": 2.7614409374069644, + "grad_norm": 1.214600920677185, + "learning_rate": 8.958218647133501e-06, + "loss": 3.5943, + "step": 245800 + }, + { + "epoch": 2.7620026625773075, + "grad_norm": 1.1684232950210571, + "learning_rate": 8.95597173383065e-06, + "loss": 3.551, + "step": 245850 + }, + { + "epoch": 2.7625643877476507, + "grad_norm": 1.1153879165649414, + "learning_rate": 8.9537248205278e-06, + "loss": 3.6145, + "step": 245900 + }, + { + "epoch": 2.763126112917994, + "grad_norm": 1.3280410766601562, + "learning_rate": 8.951477907224951e-06, + "loss": 3.6442, + "step": 245950 + }, + { + "epoch": 2.763687838088337, + "grad_norm": 1.2092448472976685, + "learning_rate": 8.9492309939221e-06, + "loss": 3.7016, + "step": 246000 + }, + { + "epoch": 2.76424956325868, + "grad_norm": 1.1283891201019287, + "learning_rate": 8.94698408061925e-06, + "loss": 3.6287, + "step": 246050 + }, + { + "epoch": 2.7648112884290232, + "grad_norm": 1.2193617820739746, + "learning_rate": 8.944737167316399e-06, + "loss": 3.6846, + "step": 246100 + }, + { + "epoch": 2.7653730135993664, + "grad_norm": 1.167502760887146, + "learning_rate": 8.94249025401355e-06, + "loss": 3.6715, + "step": 246150 + }, + { + "epoch": 2.7659347387697095, + "grad_norm": 1.071590542793274, + "learning_rate": 8.9402433407107e-06, + "loss": 3.586, + "step": 246200 + }, + { + "epoch": 2.7664964639400527, + "grad_norm": 1.2695493698120117, + "learning_rate": 8.937996427407849e-06, + "loss": 3.6704, + "step": 246250 + }, + { + "epoch": 2.767058189110396, + "grad_norm": 1.181056261062622, + "learning_rate": 8.935749514105e-06, + "loss": 3.6904, + "step": 246300 + }, + { + "epoch": 2.767619914280739, + "grad_norm": 1.0909191370010376, + "learning_rate": 8.933502600802149e-06, + "loss": 3.5455, + "step": 246350 + }, + { + "epoch": 2.768181639451082, + "grad_norm": 1.1767804622650146, + "learning_rate": 8.931255687499298e-06, + "loss": 3.656, + "step": 246400 + }, + { + "epoch": 2.7687433646214252, + "grad_norm": 1.1259640455245972, + "learning_rate": 8.929008774196448e-06, + "loss": 3.5926, + "step": 246450 + }, + { + "epoch": 2.7693050897917684, + "grad_norm": 1.0753785371780396, + "learning_rate": 8.926761860893599e-06, + "loss": 3.6134, + "step": 246500 + }, + { + "epoch": 2.7698668149621115, + "grad_norm": 1.1701672077178955, + "learning_rate": 8.924514947590748e-06, + "loss": 3.5493, + "step": 246550 + }, + { + "epoch": 2.7704285401324547, + "grad_norm": 1.2856541872024536, + "learning_rate": 8.922268034287897e-06, + "loss": 3.7311, + "step": 246600 + }, + { + "epoch": 2.770990265302798, + "grad_norm": 1.198441505432129, + "learning_rate": 8.920021120985048e-06, + "loss": 3.6904, + "step": 246650 + }, + { + "epoch": 2.7715519904731414, + "grad_norm": 1.2065755128860474, + "learning_rate": 8.917774207682197e-06, + "loss": 3.5943, + "step": 246700 + }, + { + "epoch": 2.772113715643484, + "grad_norm": 1.2504088878631592, + "learning_rate": 8.915527294379347e-06, + "loss": 3.5905, + "step": 246750 + }, + { + "epoch": 2.7726754408138277, + "grad_norm": 1.1688779592514038, + "learning_rate": 8.913280381076496e-06, + "loss": 3.735, + "step": 246800 + }, + { + "epoch": 2.7732371659841704, + "grad_norm": 1.1989463567733765, + "learning_rate": 8.911033467773647e-06, + "loss": 3.5422, + "step": 246850 + }, + { + "epoch": 2.773798891154514, + "grad_norm": 1.1521341800689697, + "learning_rate": 8.908786554470796e-06, + "loss": 3.635, + "step": 246900 + }, + { + "epoch": 2.7743606163248566, + "grad_norm": 1.1931767463684082, + "learning_rate": 8.906539641167946e-06, + "loss": 3.6291, + "step": 246950 + }, + { + "epoch": 2.7749223414952002, + "grad_norm": 1.0621672868728638, + "learning_rate": 8.904292727865095e-06, + "loss": 3.6459, + "step": 247000 + }, + { + "epoch": 2.7754840666655434, + "grad_norm": 1.0527377128601074, + "learning_rate": 8.902045814562246e-06, + "loss": 3.6414, + "step": 247050 + }, + { + "epoch": 2.7760457918358865, + "grad_norm": 1.1050207614898682, + "learning_rate": 8.899798901259395e-06, + "loss": 3.626, + "step": 247100 + }, + { + "epoch": 2.7766075170062297, + "grad_norm": 1.1118500232696533, + "learning_rate": 8.897551987956545e-06, + "loss": 3.6147, + "step": 247150 + }, + { + "epoch": 2.777169242176573, + "grad_norm": 1.1933377981185913, + "learning_rate": 8.895305074653696e-06, + "loss": 3.6173, + "step": 247200 + }, + { + "epoch": 2.777730967346916, + "grad_norm": 1.0509549379348755, + "learning_rate": 8.893058161350845e-06, + "loss": 3.6746, + "step": 247250 + }, + { + "epoch": 2.778292692517259, + "grad_norm": 1.2486989498138428, + "learning_rate": 8.890811248047994e-06, + "loss": 3.6326, + "step": 247300 + }, + { + "epoch": 2.778854417687602, + "grad_norm": 1.1824915409088135, + "learning_rate": 8.888564334745144e-06, + "loss": 3.6884, + "step": 247350 + }, + { + "epoch": 2.7794161428579454, + "grad_norm": 1.0689419507980347, + "learning_rate": 8.886317421442295e-06, + "loss": 3.6419, + "step": 247400 + }, + { + "epoch": 2.7799778680282885, + "grad_norm": 1.365525484085083, + "learning_rate": 8.884070508139444e-06, + "loss": 3.655, + "step": 247450 + }, + { + "epoch": 2.7805395931986316, + "grad_norm": 1.170935034751892, + "learning_rate": 8.881823594836593e-06, + "loss": 3.6977, + "step": 247500 + }, + { + "epoch": 2.781101318368975, + "grad_norm": 1.2151670455932617, + "learning_rate": 8.879576681533744e-06, + "loss": 3.6232, + "step": 247550 + }, + { + "epoch": 2.781663043539318, + "grad_norm": 1.2371901273727417, + "learning_rate": 8.877329768230894e-06, + "loss": 3.6226, + "step": 247600 + }, + { + "epoch": 2.782224768709661, + "grad_norm": 1.160022497177124, + "learning_rate": 8.875082854928043e-06, + "loss": 3.5917, + "step": 247650 + }, + { + "epoch": 2.782786493880004, + "grad_norm": 1.082189679145813, + "learning_rate": 8.872835941625192e-06, + "loss": 3.5847, + "step": 247700 + }, + { + "epoch": 2.7833482190503473, + "grad_norm": 1.2220501899719238, + "learning_rate": 8.870589028322343e-06, + "loss": 3.6836, + "step": 247750 + }, + { + "epoch": 2.7839099442206905, + "grad_norm": 1.026036024093628, + "learning_rate": 8.868342115019493e-06, + "loss": 3.6766, + "step": 247800 + }, + { + "epoch": 2.7844716693910336, + "grad_norm": 1.0759615898132324, + "learning_rate": 8.866095201716642e-06, + "loss": 3.5203, + "step": 247850 + }, + { + "epoch": 2.7850333945613768, + "grad_norm": 1.0197969675064087, + "learning_rate": 8.863848288413793e-06, + "loss": 3.5947, + "step": 247900 + }, + { + "epoch": 2.78559511973172, + "grad_norm": 1.2278189659118652, + "learning_rate": 8.861601375110942e-06, + "loss": 3.6332, + "step": 247950 + }, + { + "epoch": 2.786156844902063, + "grad_norm": 1.2787965536117554, + "learning_rate": 8.859354461808092e-06, + "loss": 3.553, + "step": 248000 + }, + { + "epoch": 2.7867185700724066, + "grad_norm": 1.1631124019622803, + "learning_rate": 8.857107548505241e-06, + "loss": 3.5928, + "step": 248050 + }, + { + "epoch": 2.7872802952427493, + "grad_norm": 1.2057949304580688, + "learning_rate": 8.854860635202392e-06, + "loss": 3.5721, + "step": 248100 + }, + { + "epoch": 2.787842020413093, + "grad_norm": 1.0266140699386597, + "learning_rate": 8.852613721899541e-06, + "loss": 3.6113, + "step": 248150 + }, + { + "epoch": 2.7884037455834356, + "grad_norm": 1.293364405632019, + "learning_rate": 8.85036680859669e-06, + "loss": 3.698, + "step": 248200 + }, + { + "epoch": 2.788965470753779, + "grad_norm": 1.096808910369873, + "learning_rate": 8.848119895293842e-06, + "loss": 3.6375, + "step": 248250 + }, + { + "epoch": 2.789527195924122, + "grad_norm": 1.2148140668869019, + "learning_rate": 8.845872981990991e-06, + "loss": 3.6571, + "step": 248300 + }, + { + "epoch": 2.7900889210944655, + "grad_norm": 1.2298568487167358, + "learning_rate": 8.84362606868814e-06, + "loss": 3.5967, + "step": 248350 + }, + { + "epoch": 2.7906506462648086, + "grad_norm": 1.1999495029449463, + "learning_rate": 8.841424093651348e-06, + "loss": 3.6436, + "step": 248400 + }, + { + "epoch": 2.7912123714351518, + "grad_norm": 1.2359338998794556, + "learning_rate": 8.839177180348497e-06, + "loss": 3.6498, + "step": 248450 + }, + { + "epoch": 2.791774096605495, + "grad_norm": 1.1596742868423462, + "learning_rate": 8.836930267045646e-06, + "loss": 3.6416, + "step": 248500 + }, + { + "epoch": 2.792335821775838, + "grad_norm": 1.1785552501678467, + "learning_rate": 8.834728292008854e-06, + "loss": 3.714, + "step": 248550 + }, + { + "epoch": 2.792897546946181, + "grad_norm": 1.2940049171447754, + "learning_rate": 8.832481378706003e-06, + "loss": 3.6765, + "step": 248600 + }, + { + "epoch": 2.7934592721165243, + "grad_norm": 1.2178881168365479, + "learning_rate": 8.830234465403154e-06, + "loss": 3.6165, + "step": 248650 + }, + { + "epoch": 2.7940209972868675, + "grad_norm": 1.068784236907959, + "learning_rate": 8.827987552100304e-06, + "loss": 3.6239, + "step": 248700 + }, + { + "epoch": 2.7945827224572106, + "grad_norm": 1.4802403450012207, + "learning_rate": 8.825740638797453e-06, + "loss": 3.6999, + "step": 248750 + }, + { + "epoch": 2.7951444476275538, + "grad_norm": 1.1822059154510498, + "learning_rate": 8.823493725494602e-06, + "loss": 3.7425, + "step": 248800 + }, + { + "epoch": 2.795706172797897, + "grad_norm": 1.0425667762756348, + "learning_rate": 8.821246812191753e-06, + "loss": 3.7056, + "step": 248850 + }, + { + "epoch": 2.79626789796824, + "grad_norm": 1.1360570192337036, + "learning_rate": 8.818999898888903e-06, + "loss": 3.6334, + "step": 248900 + }, + { + "epoch": 2.796829623138583, + "grad_norm": 1.0428361892700195, + "learning_rate": 8.816752985586052e-06, + "loss": 3.6138, + "step": 248950 + }, + { + "epoch": 2.7973913483089263, + "grad_norm": 1.3827158212661743, + "learning_rate": 8.814506072283201e-06, + "loss": 3.669, + "step": 249000 + }, + { + "epoch": 2.7979530734792695, + "grad_norm": 1.1668068170547485, + "learning_rate": 8.812259158980352e-06, + "loss": 3.6294, + "step": 249050 + }, + { + "epoch": 2.7985147986496126, + "grad_norm": 1.1025125980377197, + "learning_rate": 8.810012245677502e-06, + "loss": 3.6196, + "step": 249100 + }, + { + "epoch": 2.7990765238199558, + "grad_norm": 1.403463363647461, + "learning_rate": 8.807765332374651e-06, + "loss": 3.6405, + "step": 249150 + }, + { + "epoch": 2.799638248990299, + "grad_norm": 1.200330138206482, + "learning_rate": 8.805518419071802e-06, + "loss": 3.6801, + "step": 249200 + }, + { + "epoch": 2.800199974160642, + "grad_norm": 1.2137800455093384, + "learning_rate": 8.803271505768951e-06, + "loss": 3.7161, + "step": 249250 + }, + { + "epoch": 2.800761699330985, + "grad_norm": 1.0471996068954468, + "learning_rate": 8.8010245924661e-06, + "loss": 3.6234, + "step": 249300 + }, + { + "epoch": 2.8013234245013283, + "grad_norm": 1.1346518993377686, + "learning_rate": 8.79877767916325e-06, + "loss": 3.7064, + "step": 249350 + }, + { + "epoch": 2.801885149671672, + "grad_norm": 1.2496943473815918, + "learning_rate": 8.796530765860401e-06, + "loss": 3.6554, + "step": 249400 + }, + { + "epoch": 2.8024468748420146, + "grad_norm": 1.213826298713684, + "learning_rate": 8.79428385255755e-06, + "loss": 3.655, + "step": 249450 + }, + { + "epoch": 2.803008600012358, + "grad_norm": 1.3619517087936401, + "learning_rate": 8.7920369392547e-06, + "loss": 3.5986, + "step": 249500 + }, + { + "epoch": 2.803570325182701, + "grad_norm": 1.109195590019226, + "learning_rate": 8.78979002595185e-06, + "loss": 3.6476, + "step": 249550 + }, + { + "epoch": 2.8041320503530445, + "grad_norm": 1.2254453897476196, + "learning_rate": 8.787543112649e-06, + "loss": 3.6908, + "step": 249600 + }, + { + "epoch": 2.804693775523387, + "grad_norm": 1.133043885231018, + "learning_rate": 8.78529619934615e-06, + "loss": 3.704, + "step": 249650 + }, + { + "epoch": 2.8052555006937308, + "grad_norm": 1.0666509866714478, + "learning_rate": 8.783049286043298e-06, + "loss": 3.6652, + "step": 249700 + }, + { + "epoch": 2.805817225864074, + "grad_norm": 1.1767330169677734, + "learning_rate": 8.78080237274045e-06, + "loss": 3.6484, + "step": 249750 + }, + { + "epoch": 2.806378951034417, + "grad_norm": 1.2071733474731445, + "learning_rate": 8.778555459437599e-06, + "loss": 3.5479, + "step": 249800 + }, + { + "epoch": 2.80694067620476, + "grad_norm": 1.0609846115112305, + "learning_rate": 8.776308546134748e-06, + "loss": 3.6491, + "step": 249850 + }, + { + "epoch": 2.8075024013751033, + "grad_norm": 1.076884388923645, + "learning_rate": 8.774061632831899e-06, + "loss": 3.6264, + "step": 249900 + }, + { + "epoch": 2.8080641265454465, + "grad_norm": 1.1630980968475342, + "learning_rate": 8.771814719529048e-06, + "loss": 3.5596, + "step": 249950 + }, + { + "epoch": 2.8086258517157896, + "grad_norm": 1.1163171529769897, + "learning_rate": 8.769567806226198e-06, + "loss": 3.6302, + "step": 250000 + }, + { + "epoch": 2.8091875768861327, + "grad_norm": 1.2870421409606934, + "learning_rate": 8.767320892923347e-06, + "loss": 3.6759, + "step": 250050 + }, + { + "epoch": 2.809749302056476, + "grad_norm": 1.2093838453292847, + "learning_rate": 8.765073979620498e-06, + "loss": 3.6711, + "step": 250100 + }, + { + "epoch": 2.810311027226819, + "grad_norm": 1.1980772018432617, + "learning_rate": 8.762827066317647e-06, + "loss": 3.6602, + "step": 250150 + }, + { + "epoch": 2.810872752397162, + "grad_norm": 1.2051241397857666, + "learning_rate": 8.760580153014797e-06, + "loss": 3.6949, + "step": 250200 + }, + { + "epoch": 2.8114344775675053, + "grad_norm": 1.3431698083877563, + "learning_rate": 8.758333239711948e-06, + "loss": 3.6473, + "step": 250250 + }, + { + "epoch": 2.8119962027378484, + "grad_norm": 1.0086201429367065, + "learning_rate": 8.756086326409097e-06, + "loss": 3.6524, + "step": 250300 + }, + { + "epoch": 2.8125579279081916, + "grad_norm": 1.1246020793914795, + "learning_rate": 8.753839413106246e-06, + "loss": 3.6584, + "step": 250350 + }, + { + "epoch": 2.8131196530785347, + "grad_norm": 1.124054193496704, + "learning_rate": 8.751592499803396e-06, + "loss": 3.6191, + "step": 250400 + }, + { + "epoch": 2.813681378248878, + "grad_norm": 1.1673346757888794, + "learning_rate": 8.749345586500547e-06, + "loss": 3.7155, + "step": 250450 + }, + { + "epoch": 2.814243103419221, + "grad_norm": 1.0143064260482788, + "learning_rate": 8.747098673197696e-06, + "loss": 3.6322, + "step": 250500 + }, + { + "epoch": 2.814804828589564, + "grad_norm": 1.2310535907745361, + "learning_rate": 8.744851759894845e-06, + "loss": 3.6532, + "step": 250550 + }, + { + "epoch": 2.8153665537599073, + "grad_norm": 1.1990960836410522, + "learning_rate": 8.742604846591995e-06, + "loss": 3.6442, + "step": 250600 + }, + { + "epoch": 2.8159282789302504, + "grad_norm": 1.0461411476135254, + "learning_rate": 8.740357933289146e-06, + "loss": 3.576, + "step": 250650 + }, + { + "epoch": 2.8164900041005936, + "grad_norm": 1.3315068483352661, + "learning_rate": 8.738155958252352e-06, + "loss": 3.7084, + "step": 250700 + }, + { + "epoch": 2.817051729270937, + "grad_norm": 1.2003432512283325, + "learning_rate": 8.735909044949501e-06, + "loss": 3.7478, + "step": 250750 + }, + { + "epoch": 2.81761345444128, + "grad_norm": 1.1496334075927734, + "learning_rate": 8.73366213164665e-06, + "loss": 3.6494, + "step": 250800 + }, + { + "epoch": 2.8181751796116234, + "grad_norm": 1.3857228755950928, + "learning_rate": 8.731415218343801e-06, + "loss": 3.7085, + "step": 250850 + }, + { + "epoch": 2.818736904781966, + "grad_norm": 1.5104680061340332, + "learning_rate": 8.72916830504095e-06, + "loss": 3.6291, + "step": 250900 + }, + { + "epoch": 2.8192986299523097, + "grad_norm": 1.228920817375183, + "learning_rate": 8.7269213917381e-06, + "loss": 3.7008, + "step": 250950 + }, + { + "epoch": 2.8198603551226524, + "grad_norm": 1.0269227027893066, + "learning_rate": 8.72467447843525e-06, + "loss": 3.6127, + "step": 251000 + }, + { + "epoch": 2.820422080292996, + "grad_norm": 1.380207896232605, + "learning_rate": 8.7224275651324e-06, + "loss": 3.6471, + "step": 251050 + }, + { + "epoch": 2.820983805463339, + "grad_norm": 1.117279052734375, + "learning_rate": 8.72018065182955e-06, + "loss": 3.6003, + "step": 251100 + }, + { + "epoch": 2.8215455306336823, + "grad_norm": 1.062842607498169, + "learning_rate": 8.717933738526699e-06, + "loss": 3.5711, + "step": 251150 + }, + { + "epoch": 2.8221072558040254, + "grad_norm": 0.9625957608222961, + "learning_rate": 8.71568682522385e-06, + "loss": 3.7088, + "step": 251200 + }, + { + "epoch": 2.8226689809743686, + "grad_norm": 1.1299883127212524, + "learning_rate": 8.713439911921e-06, + "loss": 3.6721, + "step": 251250 + }, + { + "epoch": 2.8232307061447117, + "grad_norm": 1.284261703491211, + "learning_rate": 8.711192998618148e-06, + "loss": 3.6396, + "step": 251300 + }, + { + "epoch": 2.823792431315055, + "grad_norm": 1.3281031847000122, + "learning_rate": 8.708946085315298e-06, + "loss": 3.6529, + "step": 251350 + }, + { + "epoch": 2.824354156485398, + "grad_norm": 1.1197929382324219, + "learning_rate": 8.706699172012449e-06, + "loss": 3.6708, + "step": 251400 + }, + { + "epoch": 2.824915881655741, + "grad_norm": 1.1786137819290161, + "learning_rate": 8.704452258709598e-06, + "loss": 3.616, + "step": 251450 + }, + { + "epoch": 2.8254776068260843, + "grad_norm": 1.1578930616378784, + "learning_rate": 8.702205345406747e-06, + "loss": 3.6754, + "step": 251500 + }, + { + "epoch": 2.8260393319964274, + "grad_norm": 1.159993052482605, + "learning_rate": 8.699958432103898e-06, + "loss": 3.6259, + "step": 251550 + }, + { + "epoch": 2.8266010571667706, + "grad_norm": 1.2689379453659058, + "learning_rate": 8.697711518801048e-06, + "loss": 3.6686, + "step": 251600 + }, + { + "epoch": 2.8271627823371137, + "grad_norm": 1.140148639678955, + "learning_rate": 8.695464605498197e-06, + "loss": 3.6117, + "step": 251650 + }, + { + "epoch": 2.827724507507457, + "grad_norm": 1.244386911392212, + "learning_rate": 8.693217692195346e-06, + "loss": 3.6526, + "step": 251700 + }, + { + "epoch": 2.8282862326778, + "grad_norm": 1.2680684328079224, + "learning_rate": 8.690970778892497e-06, + "loss": 3.6854, + "step": 251750 + }, + { + "epoch": 2.828847957848143, + "grad_norm": 1.140735149383545, + "learning_rate": 8.688723865589647e-06, + "loss": 3.6391, + "step": 251800 + }, + { + "epoch": 2.8294096830184863, + "grad_norm": 1.1599724292755127, + "learning_rate": 8.686476952286796e-06, + "loss": 3.6935, + "step": 251850 + }, + { + "epoch": 2.8299714081888294, + "grad_norm": 1.357729196548462, + "learning_rate": 8.684230038983947e-06, + "loss": 3.6596, + "step": 251900 + }, + { + "epoch": 2.8305331333591726, + "grad_norm": 1.1147730350494385, + "learning_rate": 8.681983125681096e-06, + "loss": 3.7163, + "step": 251950 + }, + { + "epoch": 2.8310948585295157, + "grad_norm": 1.0270379781723022, + "learning_rate": 8.679736212378246e-06, + "loss": 3.658, + "step": 252000 + }, + { + "epoch": 2.831656583699859, + "grad_norm": 1.3949124813079834, + "learning_rate": 8.677489299075395e-06, + "loss": 3.6602, + "step": 252050 + }, + { + "epoch": 2.8322183088702024, + "grad_norm": 1.193929672241211, + "learning_rate": 8.675242385772546e-06, + "loss": 3.6682, + "step": 252100 + }, + { + "epoch": 2.832780034040545, + "grad_norm": 1.161643624305725, + "learning_rate": 8.672995472469695e-06, + "loss": 3.7082, + "step": 252150 + }, + { + "epoch": 2.8333417592108887, + "grad_norm": 1.1729676723480225, + "learning_rate": 8.670748559166845e-06, + "loss": 3.6377, + "step": 252200 + }, + { + "epoch": 2.8339034843812314, + "grad_norm": 1.2947314977645874, + "learning_rate": 8.668501645863994e-06, + "loss": 3.6678, + "step": 252250 + }, + { + "epoch": 2.834465209551575, + "grad_norm": 1.1691313982009888, + "learning_rate": 8.666254732561145e-06, + "loss": 3.6623, + "step": 252300 + }, + { + "epoch": 2.8350269347219177, + "grad_norm": 1.1190428733825684, + "learning_rate": 8.664007819258294e-06, + "loss": 3.6095, + "step": 252350 + }, + { + "epoch": 2.8355886598922613, + "grad_norm": 1.2951792478561401, + "learning_rate": 8.661760905955444e-06, + "loss": 3.6125, + "step": 252400 + }, + { + "epoch": 2.8361503850626044, + "grad_norm": 1.2349745035171509, + "learning_rate": 8.659513992652595e-06, + "loss": 3.6809, + "step": 252450 + }, + { + "epoch": 2.8367121102329476, + "grad_norm": 1.2198148965835571, + "learning_rate": 8.657267079349744e-06, + "loss": 3.6674, + "step": 252500 + }, + { + "epoch": 2.8372738354032907, + "grad_norm": 1.2092562913894653, + "learning_rate": 8.655020166046893e-06, + "loss": 3.6697, + "step": 252550 + }, + { + "epoch": 2.837835560573634, + "grad_norm": 1.0556259155273438, + "learning_rate": 8.652773252744043e-06, + "loss": 3.7152, + "step": 252600 + }, + { + "epoch": 2.838397285743977, + "grad_norm": 1.175010323524475, + "learning_rate": 8.650526339441194e-06, + "loss": 3.5873, + "step": 252650 + }, + { + "epoch": 2.83895901091432, + "grad_norm": 1.234927773475647, + "learning_rate": 8.648279426138343e-06, + "loss": 3.5207, + "step": 252700 + }, + { + "epoch": 2.8395207360846633, + "grad_norm": 1.3387160301208496, + "learning_rate": 8.646032512835492e-06, + "loss": 3.6267, + "step": 252750 + }, + { + "epoch": 2.8400824612550064, + "grad_norm": 1.3436369895935059, + "learning_rate": 8.643785599532643e-06, + "loss": 3.6727, + "step": 252800 + }, + { + "epoch": 2.8406441864253495, + "grad_norm": 1.1480870246887207, + "learning_rate": 8.641538686229793e-06, + "loss": 3.5908, + "step": 252850 + }, + { + "epoch": 2.8412059115956927, + "grad_norm": 1.1611076593399048, + "learning_rate": 8.639291772926942e-06, + "loss": 3.5875, + "step": 252900 + }, + { + "epoch": 2.841767636766036, + "grad_norm": 1.1544500589370728, + "learning_rate": 8.637044859624091e-06, + "loss": 3.6649, + "step": 252950 + }, + { + "epoch": 2.842329361936379, + "grad_norm": 1.2998915910720825, + "learning_rate": 8.634797946321242e-06, + "loss": 3.6305, + "step": 253000 + }, + { + "epoch": 2.842891087106722, + "grad_norm": 1.068569302558899, + "learning_rate": 8.632551033018392e-06, + "loss": 3.5911, + "step": 253050 + }, + { + "epoch": 2.8434528122770653, + "grad_norm": 1.2081338167190552, + "learning_rate": 8.63030411971554e-06, + "loss": 3.7156, + "step": 253100 + }, + { + "epoch": 2.8440145374474084, + "grad_norm": 1.0115337371826172, + "learning_rate": 8.628057206412692e-06, + "loss": 3.5947, + "step": 253150 + }, + { + "epoch": 2.8445762626177515, + "grad_norm": 1.187606930732727, + "learning_rate": 8.625810293109841e-06, + "loss": 3.5828, + "step": 253200 + }, + { + "epoch": 2.8451379877880947, + "grad_norm": 1.1698554754257202, + "learning_rate": 8.62356337980699e-06, + "loss": 3.6434, + "step": 253250 + }, + { + "epoch": 2.845699712958438, + "grad_norm": 1.352535367012024, + "learning_rate": 8.62131646650414e-06, + "loss": 3.5652, + "step": 253300 + }, + { + "epoch": 2.846261438128781, + "grad_norm": 1.1410608291625977, + "learning_rate": 8.61906955320129e-06, + "loss": 3.6455, + "step": 253350 + }, + { + "epoch": 2.846823163299124, + "grad_norm": 1.1857903003692627, + "learning_rate": 8.61682263989844e-06, + "loss": 3.6229, + "step": 253400 + }, + { + "epoch": 2.8473848884694677, + "grad_norm": 1.272344708442688, + "learning_rate": 8.61457572659559e-06, + "loss": 3.6278, + "step": 253450 + }, + { + "epoch": 2.8479466136398104, + "grad_norm": 0.9948447346687317, + "learning_rate": 8.61232881329274e-06, + "loss": 3.6406, + "step": 253500 + }, + { + "epoch": 2.848508338810154, + "grad_norm": 1.1672770977020264, + "learning_rate": 8.61008189998989e-06, + "loss": 3.6908, + "step": 253550 + }, + { + "epoch": 2.8490700639804967, + "grad_norm": 1.1695061922073364, + "learning_rate": 8.607834986687039e-06, + "loss": 3.6661, + "step": 253600 + }, + { + "epoch": 2.8496317891508403, + "grad_norm": 1.0867644548416138, + "learning_rate": 8.605588073384188e-06, + "loss": 3.7171, + "step": 253650 + }, + { + "epoch": 2.850193514321183, + "grad_norm": 1.081315040588379, + "learning_rate": 8.60334116008134e-06, + "loss": 3.5856, + "step": 253700 + }, + { + "epoch": 2.8507552394915265, + "grad_norm": 1.0130645036697388, + "learning_rate": 8.601094246778489e-06, + "loss": 3.6423, + "step": 253750 + }, + { + "epoch": 2.8513169646618697, + "grad_norm": 1.2605386972427368, + "learning_rate": 8.598847333475638e-06, + "loss": 3.6737, + "step": 253800 + }, + { + "epoch": 2.851878689832213, + "grad_norm": 1.3743705749511719, + "learning_rate": 8.596600420172787e-06, + "loss": 3.6046, + "step": 253850 + }, + { + "epoch": 2.852440415002556, + "grad_norm": 1.1312140226364136, + "learning_rate": 8.594353506869938e-06, + "loss": 3.6586, + "step": 253900 + }, + { + "epoch": 2.853002140172899, + "grad_norm": 1.0801674127578735, + "learning_rate": 8.592106593567088e-06, + "loss": 3.7009, + "step": 253950 + }, + { + "epoch": 2.8535638653432422, + "grad_norm": 1.1705094575881958, + "learning_rate": 8.589859680264237e-06, + "loss": 3.6314, + "step": 254000 + }, + { + "epoch": 2.8541255905135854, + "grad_norm": 1.0673092603683472, + "learning_rate": 8.587612766961388e-06, + "loss": 3.6911, + "step": 254050 + }, + { + "epoch": 2.8546873156839285, + "grad_norm": 1.3116928339004517, + "learning_rate": 8.585365853658537e-06, + "loss": 3.7241, + "step": 254100 + }, + { + "epoch": 2.8552490408542717, + "grad_norm": 1.0950393676757812, + "learning_rate": 8.583118940355687e-06, + "loss": 3.6792, + "step": 254150 + }, + { + "epoch": 2.855810766024615, + "grad_norm": 1.0413347482681274, + "learning_rate": 8.580872027052836e-06, + "loss": 3.6527, + "step": 254200 + }, + { + "epoch": 2.856372491194958, + "grad_norm": 1.2501533031463623, + "learning_rate": 8.578625113749987e-06, + "loss": 3.5793, + "step": 254250 + }, + { + "epoch": 2.856934216365301, + "grad_norm": 1.2898277044296265, + "learning_rate": 8.576378200447136e-06, + "loss": 3.6846, + "step": 254300 + }, + { + "epoch": 2.8574959415356442, + "grad_norm": 1.5056421756744385, + "learning_rate": 8.574131287144286e-06, + "loss": 3.7161, + "step": 254350 + }, + { + "epoch": 2.8580576667059874, + "grad_norm": 1.0803170204162598, + "learning_rate": 8.571884373841437e-06, + "loss": 3.6196, + "step": 254400 + }, + { + "epoch": 2.8586193918763305, + "grad_norm": 1.1864174604415894, + "learning_rate": 8.569637460538586e-06, + "loss": 3.7141, + "step": 254450 + }, + { + "epoch": 2.8591811170466737, + "grad_norm": 1.1153154373168945, + "learning_rate": 8.567390547235735e-06, + "loss": 3.5161, + "step": 254500 + }, + { + "epoch": 2.859742842217017, + "grad_norm": 1.189439296722412, + "learning_rate": 8.565143633932885e-06, + "loss": 3.5413, + "step": 254550 + }, + { + "epoch": 2.86030456738736, + "grad_norm": 1.0375182628631592, + "learning_rate": 8.562896720630036e-06, + "loss": 3.6565, + "step": 254600 + }, + { + "epoch": 2.860866292557703, + "grad_norm": 1.4673399925231934, + "learning_rate": 8.560649807327185e-06, + "loss": 3.643, + "step": 254650 + }, + { + "epoch": 2.8614280177280462, + "grad_norm": 1.2392841577529907, + "learning_rate": 8.558402894024334e-06, + "loss": 3.6671, + "step": 254700 + }, + { + "epoch": 2.8619897428983894, + "grad_norm": 1.0539594888687134, + "learning_rate": 8.556155980721485e-06, + "loss": 3.6985, + "step": 254750 + }, + { + "epoch": 2.862551468068733, + "grad_norm": 1.079054832458496, + "learning_rate": 8.553909067418635e-06, + "loss": 3.5707, + "step": 254800 + }, + { + "epoch": 2.8631131932390756, + "grad_norm": 1.0152620077133179, + "learning_rate": 8.55170709238184e-06, + "loss": 3.5897, + "step": 254850 + }, + { + "epoch": 2.8636749184094192, + "grad_norm": 1.104443073272705, + "learning_rate": 8.549460179078991e-06, + "loss": 3.6845, + "step": 254900 + }, + { + "epoch": 2.864236643579762, + "grad_norm": 1.1426451206207275, + "learning_rate": 8.54721326577614e-06, + "loss": 3.6354, + "step": 254950 + }, + { + "epoch": 2.8647983687501055, + "grad_norm": 1.0804557800292969, + "learning_rate": 8.54496635247329e-06, + "loss": 3.6437, + "step": 255000 + }, + { + "epoch": 2.865360093920448, + "grad_norm": 1.399170994758606, + "learning_rate": 8.542719439170441e-06, + "loss": 3.6506, + "step": 255050 + }, + { + "epoch": 2.865921819090792, + "grad_norm": 1.1732996702194214, + "learning_rate": 8.54047252586759e-06, + "loss": 3.6017, + "step": 255100 + }, + { + "epoch": 2.866483544261135, + "grad_norm": 1.1958513259887695, + "learning_rate": 8.53822561256474e-06, + "loss": 3.7021, + "step": 255150 + }, + { + "epoch": 2.867045269431478, + "grad_norm": 1.0984779596328735, + "learning_rate": 8.535978699261889e-06, + "loss": 3.6368, + "step": 255200 + }, + { + "epoch": 2.8676069946018212, + "grad_norm": 1.165334701538086, + "learning_rate": 8.53373178595904e-06, + "loss": 3.68, + "step": 255250 + }, + { + "epoch": 2.8681687197721644, + "grad_norm": 1.0844663381576538, + "learning_rate": 8.53148487265619e-06, + "loss": 3.6386, + "step": 255300 + }, + { + "epoch": 2.8687304449425075, + "grad_norm": 1.0539356470108032, + "learning_rate": 8.529237959353339e-06, + "loss": 3.6779, + "step": 255350 + }, + { + "epoch": 2.8692921701128506, + "grad_norm": 1.1420127153396606, + "learning_rate": 8.52699104605049e-06, + "loss": 3.6267, + "step": 255400 + }, + { + "epoch": 2.869853895283194, + "grad_norm": 1.3625901937484741, + "learning_rate": 8.524744132747639e-06, + "loss": 3.6376, + "step": 255450 + }, + { + "epoch": 2.870415620453537, + "grad_norm": 1.2747001647949219, + "learning_rate": 8.522497219444788e-06, + "loss": 3.6448, + "step": 255500 + }, + { + "epoch": 2.87097734562388, + "grad_norm": 1.2826663255691528, + "learning_rate": 8.520250306141938e-06, + "loss": 3.623, + "step": 255550 + }, + { + "epoch": 2.871539070794223, + "grad_norm": 1.063071608543396, + "learning_rate": 8.518003392839089e-06, + "loss": 3.6853, + "step": 255600 + }, + { + "epoch": 2.8721007959645664, + "grad_norm": 1.176928997039795, + "learning_rate": 8.515756479536238e-06, + "loss": 3.6428, + "step": 255650 + }, + { + "epoch": 2.8726625211349095, + "grad_norm": 1.1733942031860352, + "learning_rate": 8.513509566233387e-06, + "loss": 3.6082, + "step": 255700 + }, + { + "epoch": 2.8732242463052526, + "grad_norm": 1.1952406167984009, + "learning_rate": 8.511262652930538e-06, + "loss": 3.6099, + "step": 255750 + }, + { + "epoch": 2.873785971475596, + "grad_norm": 1.1005576848983765, + "learning_rate": 8.509015739627688e-06, + "loss": 3.5694, + "step": 255800 + }, + { + "epoch": 2.874347696645939, + "grad_norm": 1.929207682609558, + "learning_rate": 8.506768826324837e-06, + "loss": 3.6943, + "step": 255850 + }, + { + "epoch": 2.874909421816282, + "grad_norm": 1.0577033758163452, + "learning_rate": 8.504521913021986e-06, + "loss": 3.6198, + "step": 255900 + }, + { + "epoch": 2.875471146986625, + "grad_norm": 0.9591113924980164, + "learning_rate": 8.502274999719137e-06, + "loss": 3.6189, + "step": 255950 + }, + { + "epoch": 2.8760328721569683, + "grad_norm": 1.1310296058654785, + "learning_rate": 8.500028086416287e-06, + "loss": 3.6907, + "step": 256000 + }, + { + "epoch": 2.8765945973273115, + "grad_norm": 1.4007959365844727, + "learning_rate": 8.497781173113436e-06, + "loss": 3.6813, + "step": 256050 + }, + { + "epoch": 2.8771563224976546, + "grad_norm": 1.0054186582565308, + "learning_rate": 8.495534259810587e-06, + "loss": 3.7063, + "step": 256100 + }, + { + "epoch": 2.877718047667998, + "grad_norm": 1.243122935295105, + "learning_rate": 8.493287346507736e-06, + "loss": 3.6054, + "step": 256150 + }, + { + "epoch": 2.878279772838341, + "grad_norm": 1.2411491870880127, + "learning_rate": 8.491040433204886e-06, + "loss": 3.7219, + "step": 256200 + }, + { + "epoch": 2.8788414980086845, + "grad_norm": 0.8729803562164307, + "learning_rate": 8.488793519902035e-06, + "loss": 3.6231, + "step": 256250 + }, + { + "epoch": 2.879403223179027, + "grad_norm": 1.0931246280670166, + "learning_rate": 8.486546606599186e-06, + "loss": 3.6765, + "step": 256300 + }, + { + "epoch": 2.879964948349371, + "grad_norm": 1.240601897239685, + "learning_rate": 8.484299693296335e-06, + "loss": 3.6464, + "step": 256350 + }, + { + "epoch": 2.8805266735197135, + "grad_norm": 1.2226136922836304, + "learning_rate": 8.482052779993485e-06, + "loss": 3.5675, + "step": 256400 + }, + { + "epoch": 2.881088398690057, + "grad_norm": 1.1992837190628052, + "learning_rate": 8.479805866690634e-06, + "loss": 3.6351, + "step": 256450 + }, + { + "epoch": 2.8816501238604, + "grad_norm": 1.183427095413208, + "learning_rate": 8.477558953387785e-06, + "loss": 3.6925, + "step": 256500 + }, + { + "epoch": 2.8822118490307433, + "grad_norm": 1.1084449291229248, + "learning_rate": 8.475312040084934e-06, + "loss": 3.5992, + "step": 256550 + }, + { + "epoch": 2.8827735742010865, + "grad_norm": 1.1920135021209717, + "learning_rate": 8.473065126782083e-06, + "loss": 3.5795, + "step": 256600 + }, + { + "epoch": 2.8833352993714296, + "grad_norm": 1.4281495809555054, + "learning_rate": 8.470818213479234e-06, + "loss": 3.5584, + "step": 256650 + }, + { + "epoch": 2.8838970245417728, + "grad_norm": 1.243395447731018, + "learning_rate": 8.468571300176384e-06, + "loss": 3.5865, + "step": 256700 + }, + { + "epoch": 2.884458749712116, + "grad_norm": 1.2723886966705322, + "learning_rate": 8.466324386873533e-06, + "loss": 3.6345, + "step": 256750 + }, + { + "epoch": 2.885020474882459, + "grad_norm": 1.2030029296875, + "learning_rate": 8.464077473570682e-06, + "loss": 3.7157, + "step": 256800 + }, + { + "epoch": 2.885582200052802, + "grad_norm": 1.2293763160705566, + "learning_rate": 8.461830560267833e-06, + "loss": 3.5977, + "step": 256850 + }, + { + "epoch": 2.8861439252231453, + "grad_norm": 1.2963725328445435, + "learning_rate": 8.459583646964983e-06, + "loss": 3.6438, + "step": 256900 + }, + { + "epoch": 2.8867056503934885, + "grad_norm": 1.3075108528137207, + "learning_rate": 8.457336733662132e-06, + "loss": 3.572, + "step": 256950 + }, + { + "epoch": 2.8872673755638316, + "grad_norm": 1.0276139974594116, + "learning_rate": 8.455089820359283e-06, + "loss": 3.6522, + "step": 257000 + }, + { + "epoch": 2.8878291007341748, + "grad_norm": 1.1830617189407349, + "learning_rate": 8.452842907056432e-06, + "loss": 3.6152, + "step": 257050 + }, + { + "epoch": 2.888390825904518, + "grad_norm": 1.220034122467041, + "learning_rate": 8.450595993753582e-06, + "loss": 3.6242, + "step": 257100 + }, + { + "epoch": 2.888952551074861, + "grad_norm": 1.1185407638549805, + "learning_rate": 8.448349080450731e-06, + "loss": 3.6761, + "step": 257150 + }, + { + "epoch": 2.889514276245204, + "grad_norm": 1.1763192415237427, + "learning_rate": 8.446102167147882e-06, + "loss": 3.7019, + "step": 257200 + }, + { + "epoch": 2.8900760014155473, + "grad_norm": 1.1944607496261597, + "learning_rate": 8.443900192111088e-06, + "loss": 3.6723, + "step": 257250 + }, + { + "epoch": 2.8906377265858905, + "grad_norm": 1.1896106004714966, + "learning_rate": 8.441653278808239e-06, + "loss": 3.6422, + "step": 257300 + }, + { + "epoch": 2.8911994517562336, + "grad_norm": 1.2760024070739746, + "learning_rate": 8.439406365505388e-06, + "loss": 3.648, + "step": 257350 + }, + { + "epoch": 2.8917611769265767, + "grad_norm": 1.2450604438781738, + "learning_rate": 8.437159452202538e-06, + "loss": 3.6805, + "step": 257400 + }, + { + "epoch": 2.89232290209692, + "grad_norm": 1.4671880006790161, + "learning_rate": 8.434912538899687e-06, + "loss": 3.5994, + "step": 257450 + }, + { + "epoch": 2.8928846272672635, + "grad_norm": 1.2130038738250732, + "learning_rate": 8.432665625596838e-06, + "loss": 3.675, + "step": 257500 + }, + { + "epoch": 2.893446352437606, + "grad_norm": 1.521599531173706, + "learning_rate": 8.430418712293987e-06, + "loss": 3.613, + "step": 257550 + }, + { + "epoch": 2.8940080776079498, + "grad_norm": 1.0279889106750488, + "learning_rate": 8.428171798991137e-06, + "loss": 3.6016, + "step": 257600 + }, + { + "epoch": 2.8945698027782925, + "grad_norm": 1.2140361070632935, + "learning_rate": 8.425924885688288e-06, + "loss": 3.5809, + "step": 257650 + }, + { + "epoch": 2.895131527948636, + "grad_norm": 1.276888370513916, + "learning_rate": 8.423677972385437e-06, + "loss": 3.6281, + "step": 257700 + }, + { + "epoch": 2.8956932531189787, + "grad_norm": 1.1586765050888062, + "learning_rate": 8.421431059082586e-06, + "loss": 3.6154, + "step": 257750 + }, + { + "epoch": 2.8962549782893223, + "grad_norm": 1.2609378099441528, + "learning_rate": 8.419184145779736e-06, + "loss": 3.5866, + "step": 257800 + }, + { + "epoch": 2.8968167034596655, + "grad_norm": 1.2265291213989258, + "learning_rate": 8.416937232476887e-06, + "loss": 3.6479, + "step": 257850 + }, + { + "epoch": 2.8973784286300086, + "grad_norm": 1.1802059412002563, + "learning_rate": 8.414690319174036e-06, + "loss": 3.656, + "step": 257900 + }, + { + "epoch": 2.8979401538003517, + "grad_norm": 1.4281153678894043, + "learning_rate": 8.412443405871185e-06, + "loss": 3.6283, + "step": 257950 + }, + { + "epoch": 2.898501878970695, + "grad_norm": 1.294285774230957, + "learning_rate": 8.410196492568336e-06, + "loss": 3.6803, + "step": 258000 + }, + { + "epoch": 2.899063604141038, + "grad_norm": 1.129927158355713, + "learning_rate": 8.407949579265485e-06, + "loss": 3.717, + "step": 258050 + }, + { + "epoch": 2.899625329311381, + "grad_norm": 1.1277838945388794, + "learning_rate": 8.405702665962635e-06, + "loss": 3.6004, + "step": 258100 + }, + { + "epoch": 2.9001870544817243, + "grad_norm": 1.3431695699691772, + "learning_rate": 8.403455752659784e-06, + "loss": 3.7418, + "step": 258150 + }, + { + "epoch": 2.9007487796520675, + "grad_norm": 1.0524933338165283, + "learning_rate": 8.401208839356935e-06, + "loss": 3.609, + "step": 258200 + }, + { + "epoch": 2.9013105048224106, + "grad_norm": 1.154251217842102, + "learning_rate": 8.398961926054083e-06, + "loss": 3.6867, + "step": 258250 + }, + { + "epoch": 2.9018722299927537, + "grad_norm": 1.2041724920272827, + "learning_rate": 8.396715012751234e-06, + "loss": 3.733, + "step": 258300 + }, + { + "epoch": 2.902433955163097, + "grad_norm": 1.2063065767288208, + "learning_rate": 8.394468099448383e-06, + "loss": 3.653, + "step": 258350 + }, + { + "epoch": 2.90299568033344, + "grad_norm": 1.105446219444275, + "learning_rate": 8.392221186145532e-06, + "loss": 3.6039, + "step": 258400 + }, + { + "epoch": 2.903557405503783, + "grad_norm": 1.1139814853668213, + "learning_rate": 8.389974272842682e-06, + "loss": 3.6277, + "step": 258450 + }, + { + "epoch": 2.9041191306741263, + "grad_norm": 1.2182289361953735, + "learning_rate": 8.387727359539833e-06, + "loss": 3.7213, + "step": 258500 + }, + { + "epoch": 2.9046808558444694, + "grad_norm": 1.2544748783111572, + "learning_rate": 8.385480446236982e-06, + "loss": 3.5964, + "step": 258550 + }, + { + "epoch": 2.9052425810148126, + "grad_norm": 1.1787946224212646, + "learning_rate": 8.383233532934131e-06, + "loss": 3.6215, + "step": 258600 + }, + { + "epoch": 2.9058043061851557, + "grad_norm": 1.1869672536849976, + "learning_rate": 8.380986619631282e-06, + "loss": 3.6412, + "step": 258650 + }, + { + "epoch": 2.906366031355499, + "grad_norm": 1.1628656387329102, + "learning_rate": 8.378739706328432e-06, + "loss": 3.6151, + "step": 258700 + }, + { + "epoch": 2.906927756525842, + "grad_norm": 1.1098726987838745, + "learning_rate": 8.376492793025581e-06, + "loss": 3.6668, + "step": 258750 + }, + { + "epoch": 2.907489481696185, + "grad_norm": 1.0621229410171509, + "learning_rate": 8.37424587972273e-06, + "loss": 3.7166, + "step": 258800 + }, + { + "epoch": 2.9080512068665287, + "grad_norm": 1.275601863861084, + "learning_rate": 8.371998966419881e-06, + "loss": 3.6822, + "step": 258850 + }, + { + "epoch": 2.9086129320368714, + "grad_norm": 1.1331881284713745, + "learning_rate": 8.36975205311703e-06, + "loss": 3.5805, + "step": 258900 + }, + { + "epoch": 2.909174657207215, + "grad_norm": 1.4205636978149414, + "learning_rate": 8.36750513981418e-06, + "loss": 3.6829, + "step": 258950 + }, + { + "epoch": 2.9097363823775577, + "grad_norm": 1.1129319667816162, + "learning_rate": 8.365258226511331e-06, + "loss": 3.723, + "step": 259000 + }, + { + "epoch": 2.9102981075479013, + "grad_norm": 1.158905267715454, + "learning_rate": 8.36301131320848e-06, + "loss": 3.6381, + "step": 259050 + }, + { + "epoch": 2.910859832718244, + "grad_norm": 0.996040403842926, + "learning_rate": 8.36076439990563e-06, + "loss": 3.6112, + "step": 259100 + }, + { + "epoch": 2.9114215578885876, + "grad_norm": 1.0164822340011597, + "learning_rate": 8.358517486602779e-06, + "loss": 3.6798, + "step": 259150 + }, + { + "epoch": 2.9119832830589307, + "grad_norm": 1.1888489723205566, + "learning_rate": 8.35627057329993e-06, + "loss": 3.6172, + "step": 259200 + }, + { + "epoch": 2.912545008229274, + "grad_norm": 1.2077574729919434, + "learning_rate": 8.35402365999708e-06, + "loss": 3.5739, + "step": 259250 + }, + { + "epoch": 2.913106733399617, + "grad_norm": 1.1407501697540283, + "learning_rate": 8.351776746694229e-06, + "loss": 3.6503, + "step": 259300 + }, + { + "epoch": 2.91366845856996, + "grad_norm": 1.205299735069275, + "learning_rate": 8.34952983339138e-06, + "loss": 3.5773, + "step": 259350 + }, + { + "epoch": 2.9142301837403033, + "grad_norm": 1.109458327293396, + "learning_rate": 8.347282920088529e-06, + "loss": 3.6248, + "step": 259400 + }, + { + "epoch": 2.9147919089106464, + "grad_norm": 1.1124844551086426, + "learning_rate": 8.345036006785678e-06, + "loss": 3.6433, + "step": 259450 + }, + { + "epoch": 2.9153536340809896, + "grad_norm": 1.4625402688980103, + "learning_rate": 8.342789093482828e-06, + "loss": 3.6646, + "step": 259500 + }, + { + "epoch": 2.9159153592513327, + "grad_norm": 1.2352852821350098, + "learning_rate": 8.340542180179979e-06, + "loss": 3.6157, + "step": 259550 + }, + { + "epoch": 2.916477084421676, + "grad_norm": 1.5141936540603638, + "learning_rate": 8.338295266877128e-06, + "loss": 3.641, + "step": 259600 + }, + { + "epoch": 2.917038809592019, + "grad_norm": 1.446658968925476, + "learning_rate": 8.336048353574277e-06, + "loss": 3.6134, + "step": 259650 + }, + { + "epoch": 2.917600534762362, + "grad_norm": 1.0663747787475586, + "learning_rate": 8.333801440271428e-06, + "loss": 3.6095, + "step": 259700 + }, + { + "epoch": 2.9181622599327053, + "grad_norm": 1.14284348487854, + "learning_rate": 8.331554526968578e-06, + "loss": 3.5978, + "step": 259750 + }, + { + "epoch": 2.9187239851030484, + "grad_norm": 1.254639744758606, + "learning_rate": 8.329307613665727e-06, + "loss": 3.669, + "step": 259800 + }, + { + "epoch": 2.9192857102733916, + "grad_norm": 1.1343966722488403, + "learning_rate": 8.327060700362876e-06, + "loss": 3.6259, + "step": 259850 + }, + { + "epoch": 2.9198474354437347, + "grad_norm": 0.9781042337417603, + "learning_rate": 8.324813787060027e-06, + "loss": 3.602, + "step": 259900 + }, + { + "epoch": 2.920409160614078, + "grad_norm": 1.1170762777328491, + "learning_rate": 8.322566873757176e-06, + "loss": 3.6324, + "step": 259950 + }, + { + "epoch": 2.920970885784421, + "grad_norm": 1.15255606174469, + "learning_rate": 8.320319960454326e-06, + "loss": 3.6607, + "step": 260000 + }, + { + "epoch": 2.921532610954764, + "grad_norm": 1.0240029096603394, + "learning_rate": 8.318073047151475e-06, + "loss": 3.6105, + "step": 260050 + }, + { + "epoch": 2.9220943361251073, + "grad_norm": 1.3248906135559082, + "learning_rate": 8.315826133848626e-06, + "loss": 3.6399, + "step": 260100 + }, + { + "epoch": 2.9226560612954504, + "grad_norm": 1.2919012308120728, + "learning_rate": 8.313579220545775e-06, + "loss": 3.6315, + "step": 260150 + }, + { + "epoch": 2.923217786465794, + "grad_norm": 1.1979879140853882, + "learning_rate": 8.311332307242925e-06, + "loss": 3.6296, + "step": 260200 + }, + { + "epoch": 2.9237795116361367, + "grad_norm": 1.1804121732711792, + "learning_rate": 8.309085393940076e-06, + "loss": 3.6084, + "step": 260250 + }, + { + "epoch": 2.9243412368064803, + "grad_norm": 1.197500467300415, + "learning_rate": 8.306838480637225e-06, + "loss": 3.6655, + "step": 260300 + }, + { + "epoch": 2.924902961976823, + "grad_norm": 1.2215604782104492, + "learning_rate": 8.304591567334374e-06, + "loss": 3.657, + "step": 260350 + }, + { + "epoch": 2.9254646871471666, + "grad_norm": 1.2738559246063232, + "learning_rate": 8.302344654031524e-06, + "loss": 3.657, + "step": 260400 + }, + { + "epoch": 2.9260264123175093, + "grad_norm": 1.082836389541626, + "learning_rate": 8.300097740728675e-06, + "loss": 3.6958, + "step": 260450 + }, + { + "epoch": 2.926588137487853, + "grad_norm": 1.264907717704773, + "learning_rate": 8.297850827425824e-06, + "loss": 3.6554, + "step": 260500 + }, + { + "epoch": 2.927149862658196, + "grad_norm": 1.1699469089508057, + "learning_rate": 8.295603914122973e-06, + "loss": 3.6287, + "step": 260550 + }, + { + "epoch": 2.927711587828539, + "grad_norm": 1.3663575649261475, + "learning_rate": 8.293401939086181e-06, + "loss": 3.6935, + "step": 260600 + }, + { + "epoch": 2.9282733129988823, + "grad_norm": 1.222821831703186, + "learning_rate": 8.29115502578333e-06, + "loss": 3.6179, + "step": 260650 + }, + { + "epoch": 2.9288350381692254, + "grad_norm": 1.0978554487228394, + "learning_rate": 8.28890811248048e-06, + "loss": 3.6874, + "step": 260700 + }, + { + "epoch": 2.9293967633395686, + "grad_norm": 1.1071609258651733, + "learning_rate": 8.28666119917763e-06, + "loss": 3.6344, + "step": 260750 + }, + { + "epoch": 2.9299584885099117, + "grad_norm": 1.1441104412078857, + "learning_rate": 8.28441428587478e-06, + "loss": 3.7294, + "step": 260800 + }, + { + "epoch": 2.930520213680255, + "grad_norm": 1.0611757040023804, + "learning_rate": 8.28216737257193e-06, + "loss": 3.5771, + "step": 260850 + }, + { + "epoch": 2.931081938850598, + "grad_norm": 1.0881798267364502, + "learning_rate": 8.27992045926908e-06, + "loss": 3.577, + "step": 260900 + }, + { + "epoch": 2.931643664020941, + "grad_norm": 1.148647427558899, + "learning_rate": 8.27767354596623e-06, + "loss": 3.6592, + "step": 260950 + }, + { + "epoch": 2.9322053891912843, + "grad_norm": 1.1440473794937134, + "learning_rate": 8.275426632663379e-06, + "loss": 3.6693, + "step": 261000 + }, + { + "epoch": 2.9327671143616274, + "grad_norm": 1.1127500534057617, + "learning_rate": 8.273179719360528e-06, + "loss": 3.6086, + "step": 261050 + }, + { + "epoch": 2.9333288395319705, + "grad_norm": 1.1406452655792236, + "learning_rate": 8.27093280605768e-06, + "loss": 3.6302, + "step": 261100 + }, + { + "epoch": 2.9338905647023137, + "grad_norm": 1.1945470571517944, + "learning_rate": 8.268730831020885e-06, + "loss": 3.7131, + "step": 261150 + }, + { + "epoch": 2.934452289872657, + "grad_norm": 1.1877835988998413, + "learning_rate": 8.266483917718036e-06, + "loss": 3.5216, + "step": 261200 + }, + { + "epoch": 2.935014015043, + "grad_norm": 1.5802574157714844, + "learning_rate": 8.264237004415185e-06, + "loss": 3.6957, + "step": 261250 + }, + { + "epoch": 2.935575740213343, + "grad_norm": 1.353175401687622, + "learning_rate": 8.261990091112335e-06, + "loss": 3.6262, + "step": 261300 + }, + { + "epoch": 2.9361374653836863, + "grad_norm": 1.2118300199508667, + "learning_rate": 8.259743177809486e-06, + "loss": 3.7121, + "step": 261350 + }, + { + "epoch": 2.9366991905540294, + "grad_norm": 1.281163215637207, + "learning_rate": 8.257496264506635e-06, + "loss": 3.6363, + "step": 261400 + }, + { + "epoch": 2.9372609157243725, + "grad_norm": 1.1473280191421509, + "learning_rate": 8.255249351203784e-06, + "loss": 3.6721, + "step": 261450 + }, + { + "epoch": 2.9378226408947157, + "grad_norm": 1.249420166015625, + "learning_rate": 8.253002437900934e-06, + "loss": 3.5846, + "step": 261500 + }, + { + "epoch": 2.9383843660650593, + "grad_norm": 1.2632077932357788, + "learning_rate": 8.250755524598085e-06, + "loss": 3.6419, + "step": 261550 + }, + { + "epoch": 2.938946091235402, + "grad_norm": 1.0302752256393433, + "learning_rate": 8.248508611295234e-06, + "loss": 3.5984, + "step": 261600 + }, + { + "epoch": 2.9395078164057455, + "grad_norm": 1.2598966360092163, + "learning_rate": 8.246261697992383e-06, + "loss": 3.6639, + "step": 261650 + }, + { + "epoch": 2.9400695415760882, + "grad_norm": 1.1481045484542847, + "learning_rate": 8.244014784689533e-06, + "loss": 3.6319, + "step": 261700 + }, + { + "epoch": 2.940631266746432, + "grad_norm": 1.176296591758728, + "learning_rate": 8.241767871386684e-06, + "loss": 3.6737, + "step": 261750 + }, + { + "epoch": 2.9411929919167745, + "grad_norm": 1.1154247522354126, + "learning_rate": 8.239520958083833e-06, + "loss": 3.5941, + "step": 261800 + }, + { + "epoch": 2.941754717087118, + "grad_norm": 1.2286163568496704, + "learning_rate": 8.237274044780982e-06, + "loss": 3.6045, + "step": 261850 + }, + { + "epoch": 2.9423164422574613, + "grad_norm": 1.1951974630355835, + "learning_rate": 8.235027131478133e-06, + "loss": 3.6135, + "step": 261900 + }, + { + "epoch": 2.9428781674278044, + "grad_norm": 1.1343259811401367, + "learning_rate": 8.232780218175283e-06, + "loss": 3.6668, + "step": 261950 + }, + { + "epoch": 2.9434398925981475, + "grad_norm": 1.4456795454025269, + "learning_rate": 8.230533304872432e-06, + "loss": 3.6074, + "step": 262000 + }, + { + "epoch": 2.9440016177684907, + "grad_norm": 1.2335236072540283, + "learning_rate": 8.228286391569581e-06, + "loss": 3.5383, + "step": 262050 + }, + { + "epoch": 2.944563342938834, + "grad_norm": 1.1757346391677856, + "learning_rate": 8.226039478266732e-06, + "loss": 3.6207, + "step": 262100 + }, + { + "epoch": 2.945125068109177, + "grad_norm": 1.1784416437149048, + "learning_rate": 8.223792564963882e-06, + "loss": 3.7592, + "step": 262150 + }, + { + "epoch": 2.94568679327952, + "grad_norm": 1.1022356748580933, + "learning_rate": 8.221545651661031e-06, + "loss": 3.5952, + "step": 262200 + }, + { + "epoch": 2.9462485184498632, + "grad_norm": 1.3761985301971436, + "learning_rate": 8.219298738358182e-06, + "loss": 3.6662, + "step": 262250 + }, + { + "epoch": 2.9468102436202064, + "grad_norm": 1.0535807609558105, + "learning_rate": 8.217051825055331e-06, + "loss": 3.608, + "step": 262300 + }, + { + "epoch": 2.9473719687905495, + "grad_norm": 1.0880423784255981, + "learning_rate": 8.21480491175248e-06, + "loss": 3.6238, + "step": 262350 + }, + { + "epoch": 2.9479336939608927, + "grad_norm": 0.931067705154419, + "learning_rate": 8.21255799844963e-06, + "loss": 3.6487, + "step": 262400 + }, + { + "epoch": 2.948495419131236, + "grad_norm": 1.377273678779602, + "learning_rate": 8.210311085146781e-06, + "loss": 3.6544, + "step": 262450 + }, + { + "epoch": 2.949057144301579, + "grad_norm": 1.1592832803726196, + "learning_rate": 8.20806417184393e-06, + "loss": 3.7145, + "step": 262500 + }, + { + "epoch": 2.949618869471922, + "grad_norm": 1.6375360488891602, + "learning_rate": 8.20581725854108e-06, + "loss": 3.6314, + "step": 262550 + }, + { + "epoch": 2.9501805946422652, + "grad_norm": 1.5016745328903198, + "learning_rate": 8.20357034523823e-06, + "loss": 3.6528, + "step": 262600 + }, + { + "epoch": 2.9507423198126084, + "grad_norm": 1.2260786294937134, + "learning_rate": 8.20132343193538e-06, + "loss": 3.6993, + "step": 262650 + }, + { + "epoch": 2.9513040449829515, + "grad_norm": 1.317764163017273, + "learning_rate": 8.19907651863253e-06, + "loss": 3.6427, + "step": 262700 + }, + { + "epoch": 2.9518657701532947, + "grad_norm": 1.0656663179397583, + "learning_rate": 8.196829605329679e-06, + "loss": 3.7057, + "step": 262750 + }, + { + "epoch": 2.952427495323638, + "grad_norm": 1.2142822742462158, + "learning_rate": 8.19458269202683e-06, + "loss": 3.7001, + "step": 262800 + }, + { + "epoch": 2.952989220493981, + "grad_norm": 1.0361478328704834, + "learning_rate": 8.192335778723979e-06, + "loss": 3.585, + "step": 262850 + }, + { + "epoch": 2.9535509456643245, + "grad_norm": 1.3071506023406982, + "learning_rate": 8.190088865421128e-06, + "loss": 3.6227, + "step": 262900 + }, + { + "epoch": 2.954112670834667, + "grad_norm": 1.0098520517349243, + "learning_rate": 8.18784195211828e-06, + "loss": 3.6593, + "step": 262950 + }, + { + "epoch": 2.954674396005011, + "grad_norm": 1.2244157791137695, + "learning_rate": 8.185595038815428e-06, + "loss": 3.661, + "step": 263000 + }, + { + "epoch": 2.9552361211753535, + "grad_norm": 1.1741385459899902, + "learning_rate": 8.183348125512578e-06, + "loss": 3.625, + "step": 263050 + }, + { + "epoch": 2.955797846345697, + "grad_norm": 1.1160554885864258, + "learning_rate": 8.181101212209727e-06, + "loss": 3.639, + "step": 263100 + }, + { + "epoch": 2.95635957151604, + "grad_norm": 1.3024271726608276, + "learning_rate": 8.178854298906878e-06, + "loss": 3.6041, + "step": 263150 + }, + { + "epoch": 2.9569212966863834, + "grad_norm": 1.1722395420074463, + "learning_rate": 8.176607385604027e-06, + "loss": 3.7005, + "step": 263200 + }, + { + "epoch": 2.9574830218567265, + "grad_norm": 1.129657506942749, + "learning_rate": 8.174360472301177e-06, + "loss": 3.6384, + "step": 263250 + }, + { + "epoch": 2.9580447470270697, + "grad_norm": 1.3723276853561401, + "learning_rate": 8.172113558998326e-06, + "loss": 3.6765, + "step": 263300 + }, + { + "epoch": 2.958606472197413, + "grad_norm": 1.1832383871078491, + "learning_rate": 8.169866645695477e-06, + "loss": 3.6051, + "step": 263350 + }, + { + "epoch": 2.959168197367756, + "grad_norm": 1.2455813884735107, + "learning_rate": 8.167619732392626e-06, + "loss": 3.6145, + "step": 263400 + }, + { + "epoch": 2.959729922538099, + "grad_norm": 1.0869892835617065, + "learning_rate": 8.165372819089776e-06, + "loss": 3.625, + "step": 263450 + }, + { + "epoch": 2.960291647708442, + "grad_norm": 0.987187922000885, + "learning_rate": 8.163125905786927e-06, + "loss": 3.6918, + "step": 263500 + }, + { + "epoch": 2.9608533728787854, + "grad_norm": 1.2373844385147095, + "learning_rate": 8.160878992484076e-06, + "loss": 3.6743, + "step": 263550 + }, + { + "epoch": 2.9614150980491285, + "grad_norm": 1.2048543691635132, + "learning_rate": 8.158632079181225e-06, + "loss": 3.6173, + "step": 263600 + }, + { + "epoch": 2.9619768232194716, + "grad_norm": 1.2274523973464966, + "learning_rate": 8.156385165878375e-06, + "loss": 3.6511, + "step": 263650 + }, + { + "epoch": 2.962538548389815, + "grad_norm": 1.2022584676742554, + "learning_rate": 8.154138252575526e-06, + "loss": 3.5976, + "step": 263700 + }, + { + "epoch": 2.963100273560158, + "grad_norm": 1.2064024209976196, + "learning_rate": 8.151891339272675e-06, + "loss": 3.5609, + "step": 263750 + }, + { + "epoch": 2.963661998730501, + "grad_norm": 1.0156909227371216, + "learning_rate": 8.149644425969824e-06, + "loss": 3.6531, + "step": 263800 + }, + { + "epoch": 2.964223723900844, + "grad_norm": 1.1341277360916138, + "learning_rate": 8.147397512666975e-06, + "loss": 3.6602, + "step": 263850 + }, + { + "epoch": 2.9647854490711874, + "grad_norm": 1.2537122964859009, + "learning_rate": 8.145195537630181e-06, + "loss": 3.6318, + "step": 263900 + }, + { + "epoch": 2.9653471742415305, + "grad_norm": 1.1408644914627075, + "learning_rate": 8.142948624327332e-06, + "loss": 3.5669, + "step": 263950 + }, + { + "epoch": 2.9659088994118736, + "grad_norm": 1.0433231592178345, + "learning_rate": 8.140701711024482e-06, + "loss": 3.6918, + "step": 264000 + }, + { + "epoch": 2.9664706245822168, + "grad_norm": 1.2975996732711792, + "learning_rate": 8.138454797721631e-06, + "loss": 3.6188, + "step": 264050 + }, + { + "epoch": 2.96703234975256, + "grad_norm": 1.2062848806381226, + "learning_rate": 8.13620788441878e-06, + "loss": 3.6281, + "step": 264100 + }, + { + "epoch": 2.967594074922903, + "grad_norm": 1.1855512857437134, + "learning_rate": 8.133960971115931e-06, + "loss": 3.6488, + "step": 264150 + }, + { + "epoch": 2.968155800093246, + "grad_norm": 1.1772892475128174, + "learning_rate": 8.13171405781308e-06, + "loss": 3.6896, + "step": 264200 + }, + { + "epoch": 2.96871752526359, + "grad_norm": 1.1771161556243896, + "learning_rate": 8.12946714451023e-06, + "loss": 3.5914, + "step": 264250 + }, + { + "epoch": 2.9692792504339325, + "grad_norm": 1.2375949621200562, + "learning_rate": 8.12722023120738e-06, + "loss": 3.5846, + "step": 264300 + }, + { + "epoch": 2.969840975604276, + "grad_norm": 1.2629023790359497, + "learning_rate": 8.12497331790453e-06, + "loss": 3.6376, + "step": 264350 + }, + { + "epoch": 2.9704027007746188, + "grad_norm": 1.022668480873108, + "learning_rate": 8.12272640460168e-06, + "loss": 3.6028, + "step": 264400 + }, + { + "epoch": 2.9709644259449624, + "grad_norm": 1.0815987586975098, + "learning_rate": 8.120479491298829e-06, + "loss": 3.6658, + "step": 264450 + }, + { + "epoch": 2.9715261511153055, + "grad_norm": 1.2520629167556763, + "learning_rate": 8.11823257799598e-06, + "loss": 3.6826, + "step": 264500 + }, + { + "epoch": 2.9720878762856486, + "grad_norm": 1.1516664028167725, + "learning_rate": 8.115985664693129e-06, + "loss": 3.647, + "step": 264550 + }, + { + "epoch": 2.9726496014559918, + "grad_norm": 1.0715162754058838, + "learning_rate": 8.113738751390278e-06, + "loss": 3.5649, + "step": 264600 + }, + { + "epoch": 2.973211326626335, + "grad_norm": 1.186252236366272, + "learning_rate": 8.111491838087428e-06, + "loss": 3.6375, + "step": 264650 + }, + { + "epoch": 2.973773051796678, + "grad_norm": 1.1471000909805298, + "learning_rate": 8.109244924784579e-06, + "loss": 3.634, + "step": 264700 + }, + { + "epoch": 2.974334776967021, + "grad_norm": 1.2380133867263794, + "learning_rate": 8.106998011481728e-06, + "loss": 3.6771, + "step": 264750 + }, + { + "epoch": 2.9748965021373643, + "grad_norm": 1.0482748746871948, + "learning_rate": 8.104751098178877e-06, + "loss": 3.628, + "step": 264800 + }, + { + "epoch": 2.9754582273077075, + "grad_norm": 1.1627472639083862, + "learning_rate": 8.102504184876028e-06, + "loss": 3.665, + "step": 264850 + }, + { + "epoch": 2.9760199524780506, + "grad_norm": 1.2711812257766724, + "learning_rate": 8.100257271573178e-06, + "loss": 3.6635, + "step": 264900 + }, + { + "epoch": 2.9765816776483938, + "grad_norm": 1.2783652544021606, + "learning_rate": 8.098010358270327e-06, + "loss": 3.6356, + "step": 264950 + }, + { + "epoch": 2.977143402818737, + "grad_norm": 1.17335045337677, + "learning_rate": 8.095763444967476e-06, + "loss": 3.596, + "step": 265000 + }, + { + "epoch": 2.97770512798908, + "grad_norm": 1.1174033880233765, + "learning_rate": 8.093516531664627e-06, + "loss": 3.6033, + "step": 265050 + }, + { + "epoch": 2.978266853159423, + "grad_norm": 1.2793207168579102, + "learning_rate": 8.091269618361777e-06, + "loss": 3.6049, + "step": 265100 + }, + { + "epoch": 2.9788285783297663, + "grad_norm": 1.1536370515823364, + "learning_rate": 8.089022705058926e-06, + "loss": 3.6841, + "step": 265150 + }, + { + "epoch": 2.9793903035001095, + "grad_norm": 1.2100703716278076, + "learning_rate": 8.086775791756077e-06, + "loss": 3.655, + "step": 265200 + }, + { + "epoch": 2.9799520286704526, + "grad_norm": 1.0454168319702148, + "learning_rate": 8.084528878453226e-06, + "loss": 3.598, + "step": 265250 + }, + { + "epoch": 2.9805137538407958, + "grad_norm": 1.0071556568145752, + "learning_rate": 8.082281965150376e-06, + "loss": 3.6403, + "step": 265300 + }, + { + "epoch": 2.981075479011139, + "grad_norm": 1.1551620960235596, + "learning_rate": 8.080035051847525e-06, + "loss": 3.643, + "step": 265350 + }, + { + "epoch": 2.981637204181482, + "grad_norm": 1.1440470218658447, + "learning_rate": 8.077788138544674e-06, + "loss": 3.6378, + "step": 265400 + }, + { + "epoch": 2.982198929351825, + "grad_norm": 1.1203893423080444, + "learning_rate": 8.075541225241824e-06, + "loss": 3.5999, + "step": 265450 + }, + { + "epoch": 2.9827606545221683, + "grad_norm": 1.146336317062378, + "learning_rate": 8.073294311938975e-06, + "loss": 3.6543, + "step": 265500 + }, + { + "epoch": 2.9833223796925115, + "grad_norm": 1.1753056049346924, + "learning_rate": 8.071047398636124e-06, + "loss": 3.6093, + "step": 265550 + }, + { + "epoch": 2.983884104862855, + "grad_norm": 1.254305124282837, + "learning_rate": 8.068800485333273e-06, + "loss": 3.6747, + "step": 265600 + }, + { + "epoch": 2.9844458300331977, + "grad_norm": 1.2076882123947144, + "learning_rate": 8.066553572030423e-06, + "loss": 3.661, + "step": 265650 + }, + { + "epoch": 2.9850075552035413, + "grad_norm": 1.1521852016448975, + "learning_rate": 8.064306658727574e-06, + "loss": 3.6043, + "step": 265700 + }, + { + "epoch": 2.985569280373884, + "grad_norm": 1.5098720788955688, + "learning_rate": 8.062059745424723e-06, + "loss": 3.6875, + "step": 265750 + }, + { + "epoch": 2.9861310055442276, + "grad_norm": 1.168294906616211, + "learning_rate": 8.059812832121872e-06, + "loss": 3.6478, + "step": 265800 + }, + { + "epoch": 2.9866927307145708, + "grad_norm": 1.1249010562896729, + "learning_rate": 8.057565918819023e-06, + "loss": 3.6628, + "step": 265850 + }, + { + "epoch": 2.987254455884914, + "grad_norm": 1.4948920011520386, + "learning_rate": 8.055319005516173e-06, + "loss": 3.5833, + "step": 265900 + }, + { + "epoch": 2.987816181055257, + "grad_norm": 1.4017730951309204, + "learning_rate": 8.053072092213322e-06, + "loss": 3.6423, + "step": 265950 + }, + { + "epoch": 2.9883779062256, + "grad_norm": 1.1269339323043823, + "learning_rate": 8.050825178910471e-06, + "loss": 3.6902, + "step": 266000 + }, + { + "epoch": 2.9889396313959433, + "grad_norm": 1.5000494718551636, + "learning_rate": 8.048578265607622e-06, + "loss": 3.5804, + "step": 266050 + }, + { + "epoch": 2.9895013565662865, + "grad_norm": 1.0375537872314453, + "learning_rate": 8.046331352304772e-06, + "loss": 3.6515, + "step": 266100 + }, + { + "epoch": 2.9900630817366296, + "grad_norm": 1.1606764793395996, + "learning_rate": 8.04408443900192e-06, + "loss": 3.6638, + "step": 266150 + }, + { + "epoch": 2.9906248069069727, + "grad_norm": 1.0183250904083252, + "learning_rate": 8.041837525699072e-06, + "loss": 3.6846, + "step": 266200 + }, + { + "epoch": 2.991186532077316, + "grad_norm": 1.2175195217132568, + "learning_rate": 8.039590612396221e-06, + "loss": 3.6546, + "step": 266250 + }, + { + "epoch": 2.991748257247659, + "grad_norm": 1.2877438068389893, + "learning_rate": 8.03734369909337e-06, + "loss": 3.6357, + "step": 266300 + }, + { + "epoch": 2.992309982418002, + "grad_norm": 1.2034518718719482, + "learning_rate": 8.03509678579052e-06, + "loss": 3.6697, + "step": 266350 + }, + { + "epoch": 2.9928717075883453, + "grad_norm": 1.2930511236190796, + "learning_rate": 8.03284987248767e-06, + "loss": 3.6098, + "step": 266400 + }, + { + "epoch": 2.9934334327586885, + "grad_norm": 1.302494764328003, + "learning_rate": 8.03060295918482e-06, + "loss": 3.6286, + "step": 266450 + }, + { + "epoch": 2.9939951579290316, + "grad_norm": 1.1157937049865723, + "learning_rate": 8.02835604588197e-06, + "loss": 3.5678, + "step": 266500 + }, + { + "epoch": 2.9945568830993747, + "grad_norm": 1.2050288915634155, + "learning_rate": 8.02610913257912e-06, + "loss": 3.5746, + "step": 266550 + }, + { + "epoch": 2.995118608269718, + "grad_norm": 1.3136342763900757, + "learning_rate": 8.02386221927627e-06, + "loss": 3.6396, + "step": 266600 + }, + { + "epoch": 2.995680333440061, + "grad_norm": 1.1844359636306763, + "learning_rate": 8.021615305973419e-06, + "loss": 3.66, + "step": 266650 + }, + { + "epoch": 2.996242058610404, + "grad_norm": 1.1723449230194092, + "learning_rate": 8.019368392670568e-06, + "loss": 3.5934, + "step": 266700 + }, + { + "epoch": 2.9968037837807473, + "grad_norm": 1.1537070274353027, + "learning_rate": 8.01712147936772e-06, + "loss": 3.6387, + "step": 266750 + }, + { + "epoch": 2.9973655089510904, + "grad_norm": 1.2313581705093384, + "learning_rate": 8.014874566064869e-06, + "loss": 3.6024, + "step": 266800 + }, + { + "epoch": 2.9979272341214336, + "grad_norm": 1.2923564910888672, + "learning_rate": 8.012627652762018e-06, + "loss": 3.6989, + "step": 266850 + }, + { + "epoch": 2.9984889592917767, + "grad_norm": 1.2889326810836792, + "learning_rate": 8.010380739459167e-06, + "loss": 3.7019, + "step": 266900 + }, + { + "epoch": 2.9990506844621203, + "grad_norm": 1.8831459283828735, + "learning_rate": 8.008133826156318e-06, + "loss": 3.679, + "step": 266950 + }, + { + "epoch": 2.999612409632463, + "grad_norm": 1.1464004516601562, + "learning_rate": 8.005886912853468e-06, + "loss": 3.6694, + "step": 267000 + }, + { + "epoch": 2.9999943827482967, + "eval_bleu": 6.0595, + "eval_gen_len": 21.5704, + "eval_loss": 3.4992141723632812, + "eval_runtime": 25714.3846, + "eval_samples_per_second": 13.846, + "eval_steps_per_second": 0.865, + "step": 267034 + } + ], + "logging_steps": 50, + "max_steps": 445055, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 100, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 7.847598620639232e+16, + "train_batch_size": 16, + "trial_name": null, + "trial_params": null +}