{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 2.999975471559273,
  "eval_steps": 500,
  "global_step": 7644,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.003924550516323677,
      "grad_norm": 17.851987080758406,
      "learning_rate": 8.695652173913044e-07,
      "loss": 1.6604,
      "step": 10
    },
    {
      "epoch": 0.007849101032647354,
      "grad_norm": 4.080985759495849,
      "learning_rate": 1.7391304347826088e-06,
      "loss": 1.6397,
      "step": 20
    },
    {
      "epoch": 0.011773651548971031,
      "grad_norm": 2.9426600727037857,
      "learning_rate": 2.6086956521739132e-06,
      "loss": 1.5109,
      "step": 30
    },
    {
      "epoch": 0.01569820206529471,
      "grad_norm": 2.083885999160059,
      "learning_rate": 3.4782608695652175e-06,
      "loss": 1.4323,
      "step": 40
    },
    {
      "epoch": 0.019622752581618386,
      "grad_norm": 5.06773100741999,
      "learning_rate": 4.347826086956522e-06,
      "loss": 1.4517,
      "step": 50
    },
    {
      "epoch": 0.023547303097942063,
      "grad_norm": 3.058219108148106,
      "learning_rate": 5.2173913043478265e-06,
      "loss": 1.3684,
      "step": 60
    },
    {
      "epoch": 0.02747185361426574,
      "grad_norm": 2.512374629159717,
      "learning_rate": 6.086956521739132e-06,
      "loss": 1.2584,
      "step": 70
    },
    {
      "epoch": 0.03139640413058942,
      "grad_norm": 2.083781423381198,
      "learning_rate": 6.956521739130435e-06,
      "loss": 1.2888,
      "step": 80
    },
    {
      "epoch": 0.035320954646913094,
      "grad_norm": 1.33754132363256,
      "learning_rate": 7.82608695652174e-06,
      "loss": 1.3061,
      "step": 90
    },
    {
      "epoch": 0.03924550516323677,
      "grad_norm": 5.056402397307596,
      "learning_rate": 8.695652173913044e-06,
      "loss": 1.3281,
      "step": 100
    },
    {
      "epoch": 0.04317005567956045,
      "grad_norm": 2.924060457678278,
      "learning_rate": 9.565217391304349e-06,
      "loss": 1.3235,
      "step": 110
    },
    {
      "epoch": 0.047094606195884126,
      "grad_norm": 2.2704861146708177,
      "learning_rate": 1.0434782608695653e-05,
      "loss": 1.2073,
      "step": 120
    },
    {
      "epoch": 0.0510191567122078,
      "grad_norm": 2.137074925911332,
      "learning_rate": 1.1304347826086957e-05,
      "loss": 1.2304,
      "step": 130
    },
    {
      "epoch": 0.05494370722853148,
      "grad_norm": 1.7728719097748167,
      "learning_rate": 1.2173913043478263e-05,
      "loss": 1.2688,
      "step": 140
    },
    {
      "epoch": 0.05886825774485516,
      "grad_norm": 4.131355974368673,
      "learning_rate": 1.3043478260869566e-05,
      "loss": 1.2972,
      "step": 150
    },
    {
      "epoch": 0.06279280826117883,
      "grad_norm": 2.799659453356947,
      "learning_rate": 1.391304347826087e-05,
      "loss": 1.2953,
      "step": 160
    },
    {
      "epoch": 0.06671735877750251,
      "grad_norm": 2.111913097871217,
      "learning_rate": 1.4782608695652174e-05,
      "loss": 1.2443,
      "step": 170
    },
    {
      "epoch": 0.07064190929382619,
      "grad_norm": 2.1912441872440005,
      "learning_rate": 1.565217391304348e-05,
      "loss": 1.236,
      "step": 180
    },
    {
      "epoch": 0.07456645981014987,
      "grad_norm": 2.234098380713297,
      "learning_rate": 1.6521739130434785e-05,
      "loss": 1.259,
      "step": 190
    },
    {
      "epoch": 0.07849101032647354,
      "grad_norm": 4.198102404688399,
      "learning_rate": 1.739130434782609e-05,
      "loss": 1.2817,
      "step": 200
    },
    {
      "epoch": 0.08241556084279722,
      "grad_norm": 3.019782787557433,
      "learning_rate": 1.8260869565217393e-05,
      "loss": 1.2664,
      "step": 210
    },
    {
      "epoch": 0.0863401113591209,
      "grad_norm": 2.0685750522996167,
      "learning_rate": 1.9130434782608697e-05,
      "loss": 1.2373,
      "step": 220
    },
    {
      "epoch": 0.09026466187544457,
      "grad_norm": 4.300018110272735,
      "learning_rate": 2e-05,
      "loss": 1.2275,
      "step": 230
    },
    {
      "epoch": 0.09418921239176825,
      "grad_norm": 1.7354497070996773,
      "learning_rate": 1.9999910223238215e-05,
      "loss": 1.2408,
      "step": 240
    },
    {
      "epoch": 0.09811376290809193,
      "grad_norm": 3.4204284789804036,
      "learning_rate": 1.999964089456483e-05,
      "loss": 1.2965,
      "step": 250
    },
    {
      "epoch": 0.1020383134244156,
      "grad_norm": 2.4941748811333215,
      "learning_rate": 1.9999192018815737e-05,
      "loss": 1.2779,
      "step": 260
    },
    {
      "epoch": 0.10596286394073928,
      "grad_norm": 1.9457868113217196,
      "learning_rate": 1.999856360405066e-05,
      "loss": 1.2323,
      "step": 270
    },
    {
      "epoch": 0.10988741445706296,
      "grad_norm": 1.6849461858955408,
      "learning_rate": 1.9997755661553007e-05,
      "loss": 1.2056,
      "step": 280
    },
    {
      "epoch": 0.11381196497338664,
      "grad_norm": 1.867640144655699,
      "learning_rate": 1.9996768205829667e-05,
      "loss": 1.2369,
      "step": 290
    },
    {
      "epoch": 0.11773651548971031,
      "grad_norm": 3.3865066094085785,
      "learning_rate": 1.9995601254610757e-05,
      "loss": 1.2683,
      "step": 300
    },
    {
      "epoch": 0.12166106600603399,
      "grad_norm": 2.397528528715023,
      "learning_rate": 1.99942548288493e-05,
      "loss": 1.2799,
      "step": 310
    },
    {
      "epoch": 0.12558561652235767,
      "grad_norm": 2.5820404632260487,
      "learning_rate": 1.9992728952720842e-05,
      "loss": 1.2019,
      "step": 320
    },
    {
      "epoch": 0.12951016703868135,
      "grad_norm": 1.7151124312301378,
      "learning_rate": 1.9991023653623028e-05,
      "loss": 1.2133,
      "step": 330
    },
    {
      "epoch": 0.13343471755500502,
      "grad_norm": 2.0107220916427164,
      "learning_rate": 1.9989138962175105e-05,
      "loss": 1.2405,
      "step": 340
    },
    {
      "epoch": 0.1373592680713287,
      "grad_norm": 2.8574288343315897,
      "learning_rate": 1.998707491221737e-05,
      "loss": 1.2588,
      "step": 350
    },
    {
      "epoch": 0.14128381858765238,
      "grad_norm": 2.4054121950102756,
      "learning_rate": 1.9984831540810567e-05,
      "loss": 1.2581,
      "step": 360
    },
    {
      "epoch": 0.14520836910397605,
      "grad_norm": 1.8485304425199756,
      "learning_rate": 1.9982408888235224e-05,
      "loss": 1.1938,
      "step": 370
    },
    {
      "epoch": 0.14913291962029973,
      "grad_norm": 1.5828252143457966,
      "learning_rate": 1.997980699799092e-05,
      "loss": 1.1862,
      "step": 380
    },
    {
      "epoch": 0.1530574701366234,
      "grad_norm": 1.2627808446236994,
      "learning_rate": 1.9977025916795503e-05,
      "loss": 1.2135,
      "step": 390
    },
    {
      "epoch": 0.15698202065294709,
      "grad_norm": 3.2564505072901664,
      "learning_rate": 1.997406569458428e-05,
      "loss": 1.2506,
      "step": 400
    },
    {
      "epoch": 0.16090657116927076,
      "grad_norm": 2.3862130817749825,
      "learning_rate": 1.997092638450907e-05,
      "loss": 1.2391,
      "step": 410
    },
    {
      "epoch": 0.16483112168559444,
      "grad_norm": 1.8315079964245025,
      "learning_rate": 1.9967608042937303e-05,
      "loss": 1.1829,
      "step": 420
    },
    {
      "epoch": 0.16875567220191812,
      "grad_norm": 1.8430052290908208,
      "learning_rate": 1.9964110729450966e-05,
      "loss": 1.205,
      "step": 430
    },
    {
      "epoch": 0.1726802227182418,
      "grad_norm": 1.5980747844446883,
      "learning_rate": 1.9960434506845555e-05,
      "loss": 1.2007,
      "step": 440
    },
    {
      "epoch": 0.17660477323456547,
      "grad_norm": 2.942407149873092,
      "learning_rate": 1.9956579441128942e-05,
      "loss": 1.2365,
      "step": 450
    },
    {
      "epoch": 0.18052932375088915,
      "grad_norm": 2.379751902649777,
      "learning_rate": 1.995254560152019e-05,
      "loss": 1.2276,
      "step": 460
    },
    {
      "epoch": 0.18445387426721283,
      "grad_norm": 1.7688369917233489,
      "learning_rate": 1.9948333060448314e-05,
      "loss": 1.1968,
      "step": 470
    },
    {
      "epoch": 0.1883784247835365,
      "grad_norm": 1.589083721261629,
      "learning_rate": 1.994394189355097e-05,
      "loss": 1.195,
      "step": 480
    },
    {
      "epoch": 0.19230297529986018,
      "grad_norm": 1.5496967150579632,
      "learning_rate": 1.9939372179673104e-05,
      "loss": 1.1999,
      "step": 490
    },
    {
      "epoch": 0.19622752581618386,
      "grad_norm": 2.718076034623976,
      "learning_rate": 1.9934624000865542e-05,
      "loss": 1.2349,
      "step": 500
    },
    {
      "epoch": 0.20015207633250753,
      "grad_norm": 2.3890997580290567,
      "learning_rate": 1.9929697442383514e-05,
      "loss": 1.2326,
      "step": 510
    },
    {
      "epoch": 0.2040766268488312,
      "grad_norm": 1.9579908821335397,
      "learning_rate": 1.9924592592685105e-05,
      "loss": 1.1975,
      "step": 520
    },
    {
      "epoch": 0.2080011773651549,
      "grad_norm": 1.5776501334782151,
      "learning_rate": 1.991930954342969e-05,
      "loss": 1.1816,
      "step": 530
    },
    {
      "epoch": 0.21192572788147856,
      "grad_norm": 1.1161863427171768,
      "learning_rate": 1.9913848389476283e-05,
      "loss": 1.1491,
      "step": 540
    },
    {
      "epoch": 0.21585027839780224,
      "grad_norm": 2.648197492260032,
      "learning_rate": 1.9908209228881826e-05,
      "loss": 1.2435,
      "step": 550
    },
    {
      "epoch": 0.21977482891412592,
      "grad_norm": 2.408343139002311,
      "learning_rate": 1.990239216289944e-05,
      "loss": 1.2316,
      "step": 560
    },
    {
      "epoch": 0.2236993794304496,
      "grad_norm": 1.8673008806345932,
      "learning_rate": 1.9896397295976585e-05,
      "loss": 1.1606,
      "step": 570
    },
    {
      "epoch": 0.22762392994677327,
      "grad_norm": 1.5191789990510323,
      "learning_rate": 1.9890224735753215e-05,
      "loss": 1.1647,
      "step": 580
    },
    {
      "epoch": 0.23154848046309695,
      "grad_norm": 1.2422189101767849,
      "learning_rate": 1.9883874593059825e-05,
      "loss": 1.1751,
      "step": 590
    },
    {
      "epoch": 0.23547303097942063,
      "grad_norm": 2.942648489159468,
      "learning_rate": 1.987734698191546e-05,
      "loss": 1.2334,
      "step": 600
    },
    {
      "epoch": 0.2393975814957443,
      "grad_norm": 2.4573102563716294,
      "learning_rate": 1.987064201952568e-05,
      "loss": 1.237,
      "step": 610
    },
    {
      "epoch": 0.24332213201206798,
      "grad_norm": 1.7727471642184118,
      "learning_rate": 1.9863759826280446e-05,
      "loss": 1.1665,
      "step": 620
    },
    {
      "epoch": 0.24724668252839166,
      "grad_norm": 1.5235132834345957,
      "learning_rate": 1.9856700525751967e-05,
      "loss": 1.1724,
      "step": 630
    },
    {
      "epoch": 0.25117123304471534,
      "grad_norm": 1.2376130762267767,
      "learning_rate": 1.9849464244692465e-05,
      "loss": 1.1712,
      "step": 640
    },
    {
      "epoch": 0.255095783561039,
      "grad_norm": 3.040275337730361,
      "learning_rate": 1.9842051113031922e-05,
      "loss": 1.2033,
      "step": 650
    },
    {
      "epoch": 0.2590203340773627,
      "grad_norm": 2.3310773148731445,
      "learning_rate": 1.9834461263875728e-05,
      "loss": 1.2291,
      "step": 660
    },
    {
      "epoch": 0.26294488459368637,
      "grad_norm": 1.7613401699558136,
      "learning_rate": 1.9826694833502295e-05,
      "loss": 1.2002,
      "step": 670
    },
    {
      "epoch": 0.26686943511001004,
      "grad_norm": 1.4212023760208006,
      "learning_rate": 1.9818751961360623e-05,
      "loss": 1.1806,
      "step": 680
    },
    {
      "epoch": 0.2707939856263337,
      "grad_norm": 1.5547355362596973,
      "learning_rate": 1.9810632790067773e-05,
      "loss": 1.203,
      "step": 690
    },
    {
      "epoch": 0.2747185361426574,
      "grad_norm": 2.864900900676303,
      "learning_rate": 1.9802337465406332e-05,
      "loss": 1.2255,
      "step": 700
    },
    {
      "epoch": 0.2786430866589811,
      "grad_norm": 2.273750000509486,
      "learning_rate": 1.9793866136321775e-05,
      "loss": 1.2176,
      "step": 710
    },
    {
      "epoch": 0.28256763717530475,
      "grad_norm": 1.8215540424542227,
      "learning_rate": 1.97852189549198e-05,
      "loss": 1.159,
      "step": 720
    },
    {
      "epoch": 0.28649218769162843,
      "grad_norm": 1.4139436194360697,
      "learning_rate": 1.9776396076463597e-05,
      "loss": 1.1552,
      "step": 730
    },
    {
      "epoch": 0.2904167382079521,
      "grad_norm": 1.2805284263458057,
      "learning_rate": 1.9767397659371058e-05,
      "loss": 1.1857,
      "step": 740
    },
    {
      "epoch": 0.2943412887242758,
      "grad_norm": 2.6772231556409025,
      "learning_rate": 1.975822386521193e-05,
      "loss": 1.2332,
      "step": 750
    },
    {
      "epoch": 0.29826583924059946,
      "grad_norm": 2.283135985364224,
      "learning_rate": 1.974887485870492e-05,
      "loss": 1.203,
      "step": 760
    },
    {
      "epoch": 0.30219038975692314,
      "grad_norm": 1.6219262272093848,
      "learning_rate": 1.973935080771474e-05,
      "loss": 1.1696,
      "step": 770
    },
    {
      "epoch": 0.3061149402732468,
      "grad_norm": 1.492581338030124,
      "learning_rate": 1.9729651883249075e-05,
      "loss": 1.1732,
      "step": 780
    },
    {
      "epoch": 0.3100394907895705,
      "grad_norm": 1.0942797906848452,
      "learning_rate": 1.9719778259455533e-05,
      "loss": 1.1665,
      "step": 790
    },
    {
      "epoch": 0.31396404130589417,
      "grad_norm": 2.7812656388019694,
      "learning_rate": 1.9709730113618507e-05,
      "loss": 1.224,
      "step": 800
    },
    {
      "epoch": 0.31788859182221785,
      "grad_norm": 2.2668037248648494,
      "learning_rate": 1.9699507626156e-05,
      "loss": 1.2128,
      "step": 810
    },
    {
      "epoch": 0.3218131423385415,
      "grad_norm": 1.653446544884032,
      "learning_rate": 1.9689110980616374e-05,
      "loss": 1.1697,
      "step": 820
    },
    {
      "epoch": 0.3257376928548652,
      "grad_norm": 1.6562721449500364,
      "learning_rate": 1.967854036367506e-05,
      "loss": 1.1786,
      "step": 830
    },
    {
      "epoch": 0.3296622433711889,
      "grad_norm": 1.3990148712849042,
      "learning_rate": 1.9667795965131215e-05,
      "loss": 1.1814,
      "step": 840
    },
    {
      "epoch": 0.33358679388751256,
      "grad_norm": 2.709511506933081,
      "learning_rate": 1.96568779779043e-05,
      "loss": 1.2135,
      "step": 850
    },
    {
      "epoch": 0.33751134440383623,
      "grad_norm": 2.1617540677777303,
      "learning_rate": 1.9645786598030617e-05,
      "loss": 1.1925,
      "step": 860
    },
    {
      "epoch": 0.3414358949201599,
      "grad_norm": 1.6647622938579412,
      "learning_rate": 1.9634522024659802e-05,
      "loss": 1.1654,
      "step": 870
    },
    {
      "epoch": 0.3453604454364836,
      "grad_norm": 1.445420557740109,
      "learning_rate": 1.9623084460051246e-05,
      "loss": 1.1318,
      "step": 880
    },
    {
      "epoch": 0.34928499595280726,
      "grad_norm": 1.2317074284184588,
      "learning_rate": 1.9611474109570446e-05,
      "loss": 1.1489,
      "step": 890
    },
    {
      "epoch": 0.35320954646913094,
      "grad_norm": 2.5433820603816497,
      "learning_rate": 1.9599691181685335e-05,
      "loss": 1.2242,
      "step": 900
    },
    {
      "epoch": 0.3571340969854546,
      "grad_norm": 2.147149465279269,
      "learning_rate": 1.9587735887962533e-05,
      "loss": 1.2087,
      "step": 910
    },
    {
      "epoch": 0.3610586475017783,
      "grad_norm": 1.6551085595896637,
      "learning_rate": 1.957560844306356e-05,
      "loss": 1.1401,
      "step": 920
    },
    {
      "epoch": 0.364983198018102,
      "grad_norm": 1.511777154462307,
      "learning_rate": 1.9563309064740955e-05,
      "loss": 1.1597,
      "step": 930
    },
    {
      "epoch": 0.36890774853442565,
      "grad_norm": 1.1094232746787396,
      "learning_rate": 1.955083797383439e-05,
      "loss": 1.1569,
      "step": 940
    },
    {
      "epoch": 0.3728322990507493,
      "grad_norm": 2.572815297222484,
      "learning_rate": 1.95381953942667e-05,
      "loss": 1.2189,
      "step": 950
    },
    {
      "epoch": 0.376756849567073,
      "grad_norm": 2.333811966623515,
      "learning_rate": 1.9525381553039852e-05,
      "loss": 1.2014,
      "step": 960
    },
    {
      "epoch": 0.3806814000833967,
      "grad_norm": 1.6717294454527916,
      "learning_rate": 1.951239668023088e-05,
      "loss": 1.1645,
      "step": 970
    },
    {
      "epoch": 0.38460595059972036,
      "grad_norm": 1.5026637953972823,
      "learning_rate": 1.9499241008987758e-05,
      "loss": 1.1632,
      "step": 980
    },
    {
      "epoch": 0.38853050111604404,
      "grad_norm": 1.2986707472970862,
      "learning_rate": 1.9485914775525193e-05,
      "loss": 1.1644,
      "step": 990
    },
    {
      "epoch": 0.3924550516323677,
      "grad_norm": 2.9685551602402573,
      "learning_rate": 1.9472418219120403e-05,
      "loss": 1.1866,
      "step": 1000
    },
    {
      "epoch": 0.3963796021486914,
      "grad_norm": 2.234634727478329,
      "learning_rate": 1.945875158210881e-05,
      "loss": 1.2102,
      "step": 1010
    },
    {
      "epoch": 0.40030415266501507,
      "grad_norm": 1.724781393052345,
      "learning_rate": 1.9444915109879704e-05,
      "loss": 1.1415,
      "step": 1020
    },
    {
      "epoch": 0.40422870318133874,
      "grad_norm": 1.4455353570141956,
      "learning_rate": 1.9430909050871815e-05,
      "loss": 1.1638,
      "step": 1030
    },
    {
      "epoch": 0.4081532536976624,
      "grad_norm": 1.1969939749246539,
      "learning_rate": 1.9416733656568868e-05,
      "loss": 1.1527,
      "step": 1040
    },
    {
      "epoch": 0.4120778042139861,
      "grad_norm": 2.5537155941782026,
      "learning_rate": 1.9402389181495063e-05,
      "loss": 1.2141,
      "step": 1050
    },
    {
      "epoch": 0.4160023547303098,
      "grad_norm": 2.08351625708691,
      "learning_rate": 1.9387875883210507e-05,
      "loss": 1.1907,
      "step": 1060
    },
    {
      "epoch": 0.41992690524663345,
      "grad_norm": 1.683716206941947,
      "learning_rate": 1.937319402230658e-05,
      "loss": 1.1538,
      "step": 1070
    },
    {
      "epoch": 0.42385145576295713,
      "grad_norm": 1.4654454547439344,
      "learning_rate": 1.935834386240127e-05,
      "loss": 1.1601,
      "step": 1080
    },
    {
      "epoch": 0.4277760062792808,
      "grad_norm": 1.1890264813491744,
      "learning_rate": 1.934332567013443e-05,
      "loss": 1.1569,
      "step": 1090
    },
    {
      "epoch": 0.4317005567956045,
      "grad_norm": 2.491064388570012,
      "learning_rate": 1.9328139715162994e-05,
      "loss": 1.2189,
      "step": 1100
    },
    {
      "epoch": 0.43562510731192816,
      "grad_norm": 2.1439066913442444,
      "learning_rate": 1.9312786270156135e-05,
      "loss": 1.1932,
      "step": 1110
    },
    {
      "epoch": 0.43954965782825184,
      "grad_norm": 1.669653473033775,
      "learning_rate": 1.9297265610790373e-05,
      "loss": 1.1387,
      "step": 1120
    },
    {
      "epoch": 0.4434742083445755,
      "grad_norm": 1.3843825017667384,
      "learning_rate": 1.9281578015744603e-05,
      "loss": 1.1376,
      "step": 1130
    },
    {
      "epoch": 0.4473987588608992,
      "grad_norm": 1.1753467269379858,
      "learning_rate": 1.9265723766695135e-05,
      "loss": 1.1481,
      "step": 1140
    },
    {
      "epoch": 0.45132330937722287,
      "grad_norm": 2.6835572127286818,
      "learning_rate": 1.9249703148310588e-05,
      "loss": 1.1767,
      "step": 1150
    },
    {
      "epoch": 0.45524785989354655,
      "grad_norm": 2.3542813633364106,
      "learning_rate": 1.9233516448246815e-05,
      "loss": 1.2115,
      "step": 1160
    },
    {
      "epoch": 0.4591724104098702,
      "grad_norm": 1.7788410116492113,
      "learning_rate": 1.9217163957141716e-05,
      "loss": 1.1595,
      "step": 1170
    },
    {
      "epoch": 0.4630969609261939,
      "grad_norm": 1.589344468558335,
      "learning_rate": 1.9200645968610036e-05,
      "loss": 1.1469,
      "step": 1180
    },
    {
      "epoch": 0.4670215114425176,
      "grad_norm": 1.3723348105531945,
      "learning_rate": 1.918396277923807e-05,
      "loss": 1.1257,
      "step": 1190
    },
    {
      "epoch": 0.47094606195884126,
      "grad_norm": 2.3205173632902674,
      "learning_rate": 1.9167114688578368e-05,
      "loss": 1.1712,
      "step": 1200
    },
    {
      "epoch": 0.47487061247516493,
      "grad_norm": 2.20379995717616,
      "learning_rate": 1.9150101999144338e-05,
      "loss": 1.198,
      "step": 1210
    },
    {
      "epoch": 0.4787951629914886,
      "grad_norm": 1.6571237734382616,
      "learning_rate": 1.9132925016404805e-05,
      "loss": 1.1346,
      "step": 1220
    },
    {
      "epoch": 0.4827197135078123,
      "grad_norm": 1.5174786415787016,
      "learning_rate": 1.911558404877855e-05,
      "loss": 1.1382,
      "step": 1230
    },
    {
      "epoch": 0.48664426402413596,
      "grad_norm": 1.0610840707954994,
      "learning_rate": 1.909807940762876e-05,
      "loss": 1.1223,
      "step": 1240
    },
    {
      "epoch": 0.49056881454045964,
      "grad_norm": 2.628300567508133,
      "learning_rate": 1.908041140725743e-05,
      "loss": 1.1758,
      "step": 1250
    },
    {
      "epoch": 0.4944933650567833,
      "grad_norm": 2.1593460670180655,
      "learning_rate": 1.9062580364899735e-05,
      "loss": 1.2182,
      "step": 1260
    },
    {
      "epoch": 0.498417915573107,
      "grad_norm": 1.6850637902364638,
      "learning_rate": 1.9044586600718323e-05,
      "loss": 1.1582,
      "step": 1270
    },
    {
      "epoch": 0.5023424660894307,
      "grad_norm": 1.6079103318853967,
      "learning_rate": 1.9026430437797568e-05,
      "loss": 1.1213,
      "step": 1280
    },
    {
      "epoch": 0.5062670166057543,
      "grad_norm": 1.2518743683340756,
      "learning_rate": 1.9008112202137777e-05,
      "loss": 1.1546,
      "step": 1290
    },
    {
      "epoch": 0.510191567122078,
      "grad_norm": 2.4116262224448057,
      "learning_rate": 1.898963222264932e-05,
      "loss": 1.1807,
      "step": 1300
    },
    {
      "epoch": 0.5141161176384017,
      "grad_norm": 2.301017308903272,
      "learning_rate": 1.8970990831146744e-05,
      "loss": 1.1837,
      "step": 1310
    },
    {
      "epoch": 0.5180406681547254,
      "grad_norm": 1.69892059072323,
      "learning_rate": 1.8952188362342804e-05,
      "loss": 1.1347,
      "step": 1320
    },
    {
      "epoch": 0.5219652186710491,
      "grad_norm": 1.399901320658771,
      "learning_rate": 1.8933225153842446e-05,
      "loss": 1.1464,
      "step": 1330
    },
    {
      "epoch": 0.5258897691873727,
      "grad_norm": 1.0557641784260816,
      "learning_rate": 1.8914101546136766e-05,
      "loss": 1.1349,
      "step": 1340
    },
    {
      "epoch": 0.5298143197036964,
      "grad_norm": 2.5362209062888295,
      "learning_rate": 1.889481788259688e-05,
      "loss": 1.1834,
      "step": 1350
    },
    {
      "epoch": 0.5337388702200201,
      "grad_norm": 2.1100281299468278,
      "learning_rate": 1.8875374509467757e-05,
      "loss": 1.2085,
      "step": 1360
    },
    {
      "epoch": 0.5376634207363438,
      "grad_norm": 1.656879350518556,
      "learning_rate": 1.8855771775862014e-05,
      "loss": 1.145,
      "step": 1370
    },
    {
      "epoch": 0.5415879712526674,
      "grad_norm": 1.3835758185671234,
      "learning_rate": 1.8836010033753637e-05,
      "loss": 1.1332,
      "step": 1380
    },
    {
      "epoch": 0.5455125217689911,
      "grad_norm": 1.3074526959475135,
      "learning_rate": 1.8816089637971674e-05,
      "loss": 1.1337,
      "step": 1390
    },
    {
      "epoch": 0.5494370722853148,
      "grad_norm": 2.471754592942074,
      "learning_rate": 1.879601094619385e-05,
      "loss": 1.177,
      "step": 1400
    },
    {
      "epoch": 0.5533616228016385,
      "grad_norm": 2.0297870153949953,
      "learning_rate": 1.877577431894015e-05,
      "loss": 1.2251,
      "step": 1410
    },
    {
      "epoch": 0.5572861733179622,
      "grad_norm": 1.6193945226941358,
      "learning_rate": 1.8755380119566343e-05,
      "loss": 1.0928,
      "step": 1420
    },
    {
      "epoch": 0.5612107238342858,
      "grad_norm": 1.4634472002355838,
      "learning_rate": 1.873482871425747e-05,
      "loss": 1.143,
      "step": 1430
    },
    {
      "epoch": 0.5651352743506095,
      "grad_norm": 1.2878805152891477,
      "learning_rate": 1.8714120472021252e-05,
      "loss": 1.1712,
      "step": 1440
    },
    {
      "epoch": 0.5690598248669332,
      "grad_norm": 2.6107789403264965,
      "learning_rate": 1.8693255764681476e-05,
      "loss": 1.1793,
      "step": 1450
    },
    {
      "epoch": 0.5729843753832569,
      "grad_norm": 2.101138870961313,
      "learning_rate": 1.867223496687131e-05,
      "loss": 1.1724,
      "step": 1460
    },
    {
      "epoch": 0.5769089258995805,
      "grad_norm": 1.559869838184,
      "learning_rate": 1.865105845602659e-05,
      "loss": 1.1569,
      "step": 1470
    },
    {
      "epoch": 0.5808334764159042,
      "grad_norm": 1.4484698696907943,
      "learning_rate": 1.8629726612379034e-05,
      "loss": 1.1461,
      "step": 1480
    },
    {
      "epoch": 0.5847580269322279,
      "grad_norm": 1.1551387246395677,
      "learning_rate": 1.86082398189494e-05,
      "loss": 1.1276,
      "step": 1490
    },
    {
      "epoch": 0.5886825774485516,
      "grad_norm": 2.3243966610365208,
      "learning_rate": 1.8586598461540647e-05,
      "loss": 1.1865,
      "step": 1500
    },
    {
      "epoch": 0.5926071279648752,
      "grad_norm": 2.045560797585921,
      "learning_rate": 1.8564802928730963e-05,
      "loss": 1.1981,
      "step": 1510
    },
    {
      "epoch": 0.5965316784811989,
      "grad_norm": 1.5774145920172018,
      "learning_rate": 1.8542853611866826e-05,
      "loss": 1.1475,
      "step": 1520
    },
    {
      "epoch": 0.6004562289975226,
      "grad_norm": 1.401143146614057,
      "learning_rate": 1.8520750905055948e-05,
      "loss": 1.1113,
      "step": 1530
    },
    {
      "epoch": 0.6043807795138463,
      "grad_norm": 1.0993576375496286,
      "learning_rate": 1.849849520516023e-05,
      "loss": 1.1196,
      "step": 1540
    },
    {
      "epoch": 0.60830533003017,
      "grad_norm": 2.6837789697900694,
      "learning_rate": 1.8476086911788588e-05,
      "loss": 1.1731,
      "step": 1550
    },
    {
      "epoch": 0.6122298805464936,
      "grad_norm": 2.2269178857118166,
      "learning_rate": 1.8453526427289836e-05,
      "loss": 1.1673,
      "step": 1560
    },
    {
      "epoch": 0.6161544310628173,
      "grad_norm": 1.6753938116918217,
      "learning_rate": 1.8430814156745424e-05,
      "loss": 1.1212,
      "step": 1570
    },
    {
      "epoch": 0.620078981579141,
      "grad_norm": 1.2926994229597162,
      "learning_rate": 1.8407950507962166e-05,
      "loss": 1.12,
      "step": 1580
    },
    {
      "epoch": 0.6240035320954647,
      "grad_norm": 1.1434774002781025,
      "learning_rate": 1.8384935891464938e-05,
      "loss": 1.1059,
      "step": 1590
    },
    {
      "epoch": 0.6279280826117883,
      "grad_norm": 2.5701272662383623,
      "learning_rate": 1.8361770720489287e-05,
      "loss": 1.1667,
      "step": 1600
    },
    {
      "epoch": 0.631852633128112,
      "grad_norm": 2.099590587250419,
      "learning_rate": 1.8338455410974017e-05,
      "loss": 1.1811,
      "step": 1610
    },
    {
      "epoch": 0.6357771836444357,
      "grad_norm": 1.5881595852499024,
      "learning_rate": 1.831499038155373e-05,
      "loss": 1.1198,
      "step": 1620
    },
    {
      "epoch": 0.6397017341607594,
      "grad_norm": 1.364474265009956,
      "learning_rate": 1.8291376053551293e-05,
      "loss": 1.1348,
      "step": 1630
    },
    {
      "epoch": 0.643626284677083,
      "grad_norm": 1.0132232392193459,
      "learning_rate": 1.8267612850970292e-05,
      "loss": 1.1341,
      "step": 1640
    },
    {
      "epoch": 0.6475508351934067,
      "grad_norm": 2.375671216469934,
      "learning_rate": 1.824370120048739e-05,
      "loss": 1.1971,
      "step": 1650
    },
    {
      "epoch": 0.6514753857097304,
      "grad_norm": 2.0072312065319142,
      "learning_rate": 1.8219641531444713e-05,
      "loss": 1.1696,
      "step": 1660
    },
    {
      "epoch": 0.6553999362260541,
      "grad_norm": 1.621521304969733,
      "learning_rate": 1.8195434275842088e-05,
      "loss": 1.1116,
      "step": 1670
    },
    {
      "epoch": 0.6593244867423778,
      "grad_norm": 1.289974630938439,
      "learning_rate": 1.817107986832932e-05,
      "loss": 1.1427,
      "step": 1680
    },
    {
      "epoch": 0.6632490372587014,
      "grad_norm": 1.2226882453760828,
      "learning_rate": 1.8146578746198374e-05,
      "loss": 1.1324,
      "step": 1690
    },
    {
      "epoch": 0.6671735877750251,
      "grad_norm": 2.6497361969234836,
      "learning_rate": 1.812193134937554e-05,
      "loss": 1.1518,
      "step": 1700
    },
    {
      "epoch": 0.6710981382913488,
      "grad_norm": 1.934779659069536,
      "learning_rate": 1.8097138120413503e-05,
      "loss": 1.1667,
      "step": 1710
    },
    {
      "epoch": 0.6750226888076725,
      "grad_norm": 1.682274234607041,
      "learning_rate": 1.8072199504483428e-05,
      "loss": 1.1094,
      "step": 1720
    },
    {
      "epoch": 0.6789472393239961,
      "grad_norm": 1.3189385241228773,
      "learning_rate": 1.8047115949366955e-05,
      "loss": 1.1485,
      "step": 1730
    },
    {
      "epoch": 0.6828717898403198,
      "grad_norm": 1.2380330513347648,
      "learning_rate": 1.8021887905448146e-05,
      "loss": 1.1228,
      "step": 1740
    },
    {
      "epoch": 0.6867963403566435,
      "grad_norm": 2.503615625647334,
      "learning_rate": 1.799651582570543e-05,
      "loss": 1.1545,
      "step": 1750
    },
    {
      "epoch": 0.6907208908729672,
      "grad_norm": 1.9970401432155471,
      "learning_rate": 1.7971000165703434e-05,
      "loss": 1.1698,
      "step": 1760
    },
    {
      "epoch": 0.6946454413892909,
      "grad_norm": 1.585682831800493,
      "learning_rate": 1.7945341383584818e-05,
      "loss": 1.12,
      "step": 1770
    },
    {
      "epoch": 0.6985699919056145,
      "grad_norm": 1.4103033699462193,
      "learning_rate": 1.7919539940062068e-05,
      "loss": 1.1375,
      "step": 1780
    },
    {
      "epoch": 0.7024945424219382,
      "grad_norm": 1.1741968139844532,
      "learning_rate": 1.7893596298409182e-05,
      "loss": 1.1045,
      "step": 1790
    },
    {
      "epoch": 0.7064190929382619,
      "grad_norm": 2.5435233808457265,
      "learning_rate": 1.7867510924453394e-05,
      "loss": 1.1561,
      "step": 1800
    },
    {
      "epoch": 0.7103436434545856,
      "grad_norm": 2.0058708638995744,
      "learning_rate": 1.784128428656678e-05,
      "loss": 1.1905,
      "step": 1810
    },
    {
      "epoch": 0.7142681939709092,
      "grad_norm": 1.5513764227014477,
      "learning_rate": 1.7814916855657872e-05,
      "loss": 1.116,
      "step": 1820
    },
    {
      "epoch": 0.7181927444872329,
      "grad_norm": 1.3841452634663314,
      "learning_rate": 1.7788409105163178e-05,
      "loss": 1.1359,
      "step": 1830
    },
    {
      "epoch": 0.7221172950035566,
      "grad_norm": 1.0616071904873385,
      "learning_rate": 1.7761761511038694e-05,
      "loss": 1.0973,
      "step": 1840
    },
    {
      "epoch": 0.7260418455198803,
      "grad_norm": 2.4472932789694726,
      "learning_rate": 1.773497455175137e-05,
      "loss": 1.1611,
      "step": 1850
    },
    {
      "epoch": 0.729966396036204,
      "grad_norm": 2.1478318961127325,
      "learning_rate": 1.7708048708270497e-05,
      "loss": 1.1637,
      "step": 1860
    },
    {
      "epoch": 0.7338909465525276,
      "grad_norm": 1.5986355699917554,
      "learning_rate": 1.7680984464059077e-05,
      "loss": 1.1179,
      "step": 1870
    },
    {
      "epoch": 0.7378154970688513,
      "grad_norm": 1.2845396362764854,
      "learning_rate": 1.7653782305065158e-05,
      "loss": 1.1407,
      "step": 1880
    },
    {
      "epoch": 0.741740047585175,
      "grad_norm": 1.1067586294132603,
      "learning_rate": 1.7626442719713083e-05,
      "loss": 1.1255,
      "step": 1890
    },
    {
      "epoch": 0.7456645981014987,
      "grad_norm": 2.2098441511746705,
      "learning_rate": 1.7598966198894746e-05,
      "loss": 1.1756,
      "step": 1900
    },
    {
      "epoch": 0.7495891486178223,
      "grad_norm": 1.9535681972911683,
      "learning_rate": 1.7571353235960754e-05,
      "loss": 1.1813,
      "step": 1910
    },
    {
      "epoch": 0.753513699134146,
      "grad_norm": 1.5198953597892402,
      "learning_rate": 1.7543604326711592e-05,
      "loss": 1.1157,
      "step": 1920
    },
    {
      "epoch": 0.7574382496504697,
      "grad_norm": 1.282264664888777,
      "learning_rate": 1.7515719969388697e-05,
      "loss": 1.1325,
      "step": 1930
    },
    {
      "epoch": 0.7613628001667934,
      "grad_norm": 1.1529117257538906,
      "learning_rate": 1.7487700664665536e-05,
      "loss": 1.1579,
      "step": 1940
    },
    {
      "epoch": 0.765287350683117,
      "grad_norm": 2.4479153069810877,
      "learning_rate": 1.7459546915638595e-05,
      "loss": 1.1548,
      "step": 1950
    },
    {
      "epoch": 0.7692119011994407,
      "grad_norm": 1.9930358021042167,
      "learning_rate": 1.743125922781836e-05,
      "loss": 1.1702,
      "step": 1960
    },
    {
      "epoch": 0.7731364517157644,
      "grad_norm": 1.5561233301314203,
      "learning_rate": 1.740283810912023e-05,
      "loss": 1.1098,
      "step": 1970
    },
    {
      "epoch": 0.7770610022320881,
      "grad_norm": 1.363877169879588,
      "learning_rate": 1.737428406985541e-05,
      "loss": 1.1276,
      "step": 1980
    },
    {
      "epoch": 0.7809855527484117,
      "grad_norm": 1.1118223017408846,
      "learning_rate": 1.7345597622721727e-05,
      "loss": 1.1143,
      "step": 1990
    },
    {
      "epoch": 0.7849101032647354,
      "grad_norm": 2.2103639445489707,
      "learning_rate": 1.7316779282794458e-05,
      "loss": 1.1436,
      "step": 2000
    },
    {
      "epoch": 0.7888346537810591,
      "grad_norm": 1.9866216610680039,
      "learning_rate": 1.728782956751705e-05,
      "loss": 1.1366,
      "step": 2010
    },
    {
      "epoch": 0.7927592042973828,
      "grad_norm": 1.557276150793299,
      "learning_rate": 1.725874899669183e-05,
      "loss": 1.1028,
      "step": 2020
    },
    {
      "epoch": 0.7966837548137065,
      "grad_norm": 1.3407353791664398,
      "learning_rate": 1.7229538092470708e-05,
      "loss": 1.121,
      "step": 2030
    },
    {
      "epoch": 0.8006083053300301,
      "grad_norm": 1.0452399263705143,
      "learning_rate": 1.7200197379345752e-05,
      "loss": 1.1052,
      "step": 2040
    },
    {
      "epoch": 0.8045328558463538,
      "grad_norm": 2.255129030276742,
      "learning_rate": 1.7170727384139808e-05,
      "loss": 1.1534,
      "step": 2050
    },
    {
      "epoch": 0.8084574063626775,
      "grad_norm": 2.023922420597953,
      "learning_rate": 1.7141128635997027e-05,
      "loss": 1.1536,
      "step": 2060
    },
    {
      "epoch": 0.8123819568790012,
      "grad_norm": 1.6459699923272906,
      "learning_rate": 1.711140166637336e-05,
      "loss": 1.1237,
      "step": 2070
    },
    {
      "epoch": 0.8163065073953248,
      "grad_norm": 1.3995835514158206,
      "learning_rate": 1.7081547009027014e-05,
      "loss": 1.1364,
      "step": 2080
    },
    {
      "epoch": 0.8202310579116485,
      "grad_norm": 1.1468866803278337,
      "learning_rate": 1.705156520000889e-05,
      "loss": 1.1055,
      "step": 2090
    },
    {
      "epoch": 0.8241556084279722,
      "grad_norm": 2.4534252010588626,
      "learning_rate": 1.702145677765293e-05,
      "loss": 1.1233,
      "step": 2100
    },
    {
      "epoch": 0.8280801589442959,
      "grad_norm": 2.069228109517972,
      "learning_rate": 1.6991222282566465e-05,
      "loss": 1.1368,
      "step": 2110
    },
    {
      "epoch": 0.8320047094606196,
      "grad_norm": 1.5353360725588796,
      "learning_rate": 1.696086225762051e-05,
      "loss": 1.0936,
      "step": 2120
    },
    {
      "epoch": 0.8359292599769432,
      "grad_norm": 1.31441567730565,
      "learning_rate": 1.6930377247940005e-05,
      "loss": 1.103,
      "step": 2130
    },
    {
      "epoch": 0.8398538104932669,
      "grad_norm": 0.9169234319686077,
      "learning_rate": 1.689976780089405e-05,
      "loss": 1.0933,
      "step": 2140
    },
    {
      "epoch": 0.8437783610095906,
      "grad_norm": 2.321894947275435,
      "learning_rate": 1.6869034466086046e-05,
      "loss": 1.1397,
      "step": 2150
    },
    {
      "epoch": 0.8477029115259143,
      "grad_norm": 2.0305512220158386,
      "learning_rate": 1.6838177795343847e-05,
      "loss": 1.1704,
      "step": 2160
    },
    {
      "epoch": 0.8516274620422379,
      "grad_norm": 1.5512604200141975,
      "learning_rate": 1.6807198342709858e-05,
      "loss": 1.1113,
      "step": 2170
    },
    {
      "epoch": 0.8555520125585616,
      "grad_norm": 1.3624069523785742,
      "learning_rate": 1.677609666443105e-05,
      "loss": 1.1355,
      "step": 2180
    },
    {
      "epoch": 0.8594765630748853,
      "grad_norm": 1.162420454215084,
      "learning_rate": 1.6744873318949032e-05,
      "loss": 1.1217,
      "step": 2190
    },
    {
      "epoch": 0.863401113591209,
      "grad_norm": 2.59296418698937,
      "learning_rate": 1.6713528866889966e-05,
      "loss": 1.1753,
      "step": 2200
    },
    {
      "epoch": 0.8673256641075326,
      "grad_norm": 2.005406504075907,
      "learning_rate": 1.6682063871054534e-05,
      "loss": 1.1596,
      "step": 2210
    },
    {
      "epoch": 0.8712502146238563,
      "grad_norm": 1.5322297264828286,
      "learning_rate": 1.6650478896407825e-05,
      "loss": 1.1093,
      "step": 2220
    },
    {
      "epoch": 0.87517476514018,
      "grad_norm": 1.3164373320619593,
      "learning_rate": 1.6618774510069187e-05,
      "loss": 1.109,
      "step": 2230
    },
    {
      "epoch": 0.8790993156565037,
      "grad_norm": 0.979869187444681,
      "learning_rate": 1.6586951281302046e-05,
      "loss": 1.1212,
      "step": 2240
    },
    {
      "epoch": 0.8830238661728274,
      "grad_norm": 2.4909588716430617,
      "learning_rate": 1.655500978150369e-05,
      "loss": 1.1286,
      "step": 2250
    },
    {
      "epoch": 0.886948416689151,
      "grad_norm": 2.022377055680929,
      "learning_rate": 1.6522950584195003e-05,
      "loss": 1.149,
      "step": 2260
    },
    {
      "epoch": 0.8908729672054747,
      "grad_norm": 1.522868856655693,
      "learning_rate": 1.649077426501017e-05,
      "loss": 1.0971,
      "step": 2270
    },
    {
      "epoch": 0.8947975177217984,
      "grad_norm": 1.3697259871875316,
      "learning_rate": 1.6458481401686334e-05,
      "loss": 1.0805,
      "step": 2280
    },
    {
      "epoch": 0.8987220682381221,
      "grad_norm": 1.018922203182855,
      "learning_rate": 1.6426072574053238e-05,
      "loss": 1.1299,
      "step": 2290
    },
    {
      "epoch": 0.9026466187544457,
      "grad_norm": 2.375176209683022,
      "learning_rate": 1.6393548364022803e-05,
      "loss": 1.1244,
      "step": 2300
    },
    {
      "epoch": 0.9065711692707694,
      "grad_norm": 1.9892422302580408,
      "learning_rate": 1.636090935557868e-05,
      "loss": 1.1825,
      "step": 2310
    },
    {
      "epoch": 0.9104957197870931,
      "grad_norm": 1.5724093274080302,
      "learning_rate": 1.632815613476576e-05,
      "loss": 1.1016,
      "step": 2320
    },
    {
      "epoch": 0.9144202703034168,
      "grad_norm": 1.2589150941461418,
      "learning_rate": 1.6295289289679674e-05,
      "loss": 1.1056,
      "step": 2330
    },
    {
      "epoch": 0.9183448208197404,
      "grad_norm": 0.946685256837567,
      "learning_rate": 1.62623094104562e-05,
      "loss": 1.0829,
      "step": 2340
    },
    {
      "epoch": 0.9222693713360641,
      "grad_norm": 2.352958803717975,
      "learning_rate": 1.6229217089260695e-05,
      "loss": 1.1514,
      "step": 2350
    },
    {
      "epoch": 0.9261939218523878,
      "grad_norm": 1.9500631427530646,
      "learning_rate": 1.6196012920277436e-05,
      "loss": 1.1563,
      "step": 2360
    },
    {
      "epoch": 0.9301184723687115,
      "grad_norm": 1.5860068083635046,
      "learning_rate": 1.616269749969899e-05,
      "loss": 1.0999,
      "step": 2370
    },
    {
      "epoch": 0.9340430228850352,
      "grad_norm": 1.2393963203713174,
      "learning_rate": 1.6129271425715458e-05,
      "loss": 1.1056,
      "step": 2380
    },
    {
      "epoch": 0.9379675734013588,
      "grad_norm": 1.0108366563366444,
      "learning_rate": 1.609573529850379e-05,
      "loss": 1.0886,
      "step": 2390
    },
    {
      "epoch": 0.9418921239176825,
      "grad_norm": 2.307306825085365,
      "learning_rate": 1.6062089720216956e-05,
      "loss": 1.125,
      "step": 2400
    },
    {
      "epoch": 0.9458166744340062,
      "grad_norm": 2.08922761238031,
      "learning_rate": 1.6028335294973182e-05,
      "loss": 1.1676,
      "step": 2410
    },
    {
      "epoch": 0.9497412249503299,
      "grad_norm": 1.5072970519469342,
      "learning_rate": 1.5994472628845054e-05,
      "loss": 1.0805,
      "step": 2420
    },
    {
      "epoch": 0.9536657754666535,
      "grad_norm": 1.208673658456983,
      "learning_rate": 1.5960502329848683e-05,
      "loss": 1.1023,
      "step": 2430
    },
    {
      "epoch": 0.9575903259829772,
      "grad_norm": 1.0701544169565054,
      "learning_rate": 1.5926425007932747e-05,
      "loss": 1.0802,
      "step": 2440
    },
    {
      "epoch": 0.9615148764993009,
      "grad_norm": 2.2259127710545745,
      "learning_rate": 1.5892241274967578e-05,
      "loss": 1.1306,
      "step": 2450
    },
    {
      "epoch": 0.9654394270156246,
      "grad_norm": 1.9613490972509378,
      "learning_rate": 1.5857951744734145e-05,
      "loss": 1.1527,
      "step": 2460
    },
    {
      "epoch": 0.9693639775319483,
      "grad_norm": 1.6339352423393527,
      "learning_rate": 1.5823557032913045e-05,
      "loss": 1.1173,
      "step": 2470
    },
    {
      "epoch": 0.9732885280482719,
      "grad_norm": 1.2845543184685153,
      "learning_rate": 1.5789057757073444e-05,
      "loss": 1.0858,
      "step": 2480
    },
    {
      "epoch": 0.9772130785645956,
      "grad_norm": 1.0634701364462926,
      "learning_rate": 1.5754454536662e-05,
      "loss": 1.0772,
      "step": 2490
    },
    {
      "epoch": 0.9811376290809193,
      "grad_norm": 2.5457603854360618,
      "learning_rate": 1.5719747992991723e-05,
      "loss": 1.1572,
      "step": 2500
    },
    {
      "epoch": 0.985062179597243,
      "grad_norm": 1.9734906496943037,
      "learning_rate": 1.568493874923084e-05,
      "loss": 1.1277,
      "step": 2510
    },
    {
      "epoch": 0.9889867301135666,
      "grad_norm": 1.5783237318057073,
      "learning_rate": 1.5650027430391584e-05,
      "loss": 1.0856,
      "step": 2520
    },
    {
      "epoch": 0.9929112806298903,
      "grad_norm": 1.4264738203605272,
      "learning_rate": 1.5615014663318993e-05,
      "loss": 1.1078,
      "step": 2530
    },
    {
      "epoch": 0.996835831146214,
      "grad_norm": 1.1620086010183999,
      "learning_rate": 1.5579901076679625e-05,
      "loss": 1.1097,
      "step": 2540
    },
    {
      "epoch": 0.999975471559273,
      "eval_loss": 0.9125259518623352,
      "eval_runtime": 1520.6591,
      "eval_samples_per_second": 16.44,
      "eval_steps_per_second": 4.11,
      "step": 2548
    },
    {
      "epoch": 1.0007849101032646,
      "grad_norm": 1.0252573802297382,
      "learning_rate": 1.5544687300950306e-05,
      "loss": 0.9338,
      "step": 2550
    },
    {
      "epoch": 1.0047094606195883,
      "grad_norm": 2.201134957778082,
      "learning_rate": 1.5509373968406792e-05,
      "loss": 0.9016,
      "step": 2560
    },
    {
      "epoch": 1.008634011135912,
      "grad_norm": 1.7915224952217608,
      "learning_rate": 1.5473961713112405e-05,
      "loss": 0.991,
      "step": 2570
    },
    {
      "epoch": 1.0125585616522357,
      "grad_norm": 1.7104141742083112,
      "learning_rate": 1.5438451170906672e-05,
      "loss": 0.9134,
      "step": 2580
    },
    {
      "epoch": 1.0164831121685594,
      "grad_norm": 1.488033671149185,
      "learning_rate": 1.5402842979393882e-05,
      "loss": 0.8688,
      "step": 2590
    },
    {
      "epoch": 1.020407662684883,
      "grad_norm": 1.1133098743882353,
      "learning_rate": 1.5367137777931673e-05,
      "loss": 0.8432,
      "step": 2600
    },
    {
      "epoch": 1.0243322132012067,
      "grad_norm": 2.0585351825127143,
      "learning_rate": 1.5331336207619507e-05,
      "loss": 0.8874,
      "step": 2610
    },
    {
      "epoch": 1.0282567637175304,
      "grad_norm": 1.867901103345647,
      "learning_rate": 1.5295438911287203e-05,
      "loss": 0.9336,
      "step": 2620
    },
    {
      "epoch": 1.032181314233854,
      "grad_norm": 1.6563402725505043,
      "learning_rate": 1.5259446533483357e-05,
      "loss": 0.8879,
      "step": 2630
    },
    {
      "epoch": 1.0361058647501777,
      "grad_norm": 1.4856230129943386,
      "learning_rate": 1.5223359720463796e-05,
      "loss": 0.859,
      "step": 2640
    },
    {
      "epoch": 1.0400304152665014,
      "grad_norm": 0.9711013069770768,
      "learning_rate": 1.5187179120179969e-05,
      "loss": 0.8288,
      "step": 2650
    },
    {
      "epoch": 1.043954965782825,
      "grad_norm": 1.9828694802282283,
      "learning_rate": 1.5150905382267299e-05,
      "loss": 0.8955,
      "step": 2660
    },
    {
      "epoch": 1.0478795162991488,
      "grad_norm": 2.0020449073070283,
      "learning_rate": 1.511453915803353e-05,
      "loss": 0.9694,
      "step": 2670
    },
    {
      "epoch": 1.0518040668154724,
      "grad_norm": 1.6466177405453537,
      "learning_rate": 1.5078081100447035e-05,
      "loss": 0.9115,
      "step": 2680
    },
    {
      "epoch": 1.0557286173317961,
      "grad_norm": 1.7037544484554887,
      "learning_rate": 1.5041531864125082e-05,
      "loss": 0.8493,
      "step": 2690
    },
    {
      "epoch": 1.0596531678481198,
      "grad_norm": 0.9649533896621292,
      "learning_rate": 1.5004892105322092e-05,
      "loss": 0.8204,
      "step": 2700
    },
    {
      "epoch": 1.0635777183644435,
      "grad_norm": 2.104909863577972,
      "learning_rate": 1.4968162481917836e-05,
      "loss": 0.9002,
      "step": 2710
    },
    {
      "epoch": 1.0675022688807672,
      "grad_norm": 2.0341252133554146,
      "learning_rate": 1.4931343653405652e-05,
      "loss": 0.9456,
      "step": 2720
    },
    {
      "epoch": 1.0714268193970908,
      "grad_norm": 1.5915526204280668,
      "learning_rate": 1.4894436280880578e-05,
      "loss": 0.8801,
      "step": 2730
    },
    {
      "epoch": 1.0753513699134145,
      "grad_norm": 1.5184799515062875,
      "learning_rate": 1.4857441027027486e-05,
      "loss": 0.8608,
      "step": 2740
    },
    {
      "epoch": 1.0792759204297382,
      "grad_norm": 1.0739540966168113,
      "learning_rate": 1.4820358556109202e-05,
      "loss": 0.8383,
      "step": 2750
    },
    {
      "epoch": 1.0832004709460619,
      "grad_norm": 2.3527635362598787,
      "learning_rate": 1.4783189533954555e-05,
      "loss": 0.8989,
      "step": 2760
    },
    {
      "epoch": 1.0871250214623855,
      "grad_norm": 1.8879516919488128,
      "learning_rate": 1.4745934627946432e-05,
      "loss": 0.9203,
      "step": 2770
    },
    {
      "epoch": 1.0910495719787092,
      "grad_norm": 1.6794025897082823,
      "learning_rate": 1.4708594507009806e-05,
      "loss": 0.8939,
      "step": 2780
    },
    {
      "epoch": 1.094974122495033,
      "grad_norm": 1.263872598631979,
      "learning_rate": 1.4671169841599695e-05,
      "loss": 0.8435,
      "step": 2790
    },
    {
      "epoch": 1.0988986730113566,
      "grad_norm": 0.9984868253029158,
      "learning_rate": 1.4633661303689157e-05,
      "loss": 0.8568,
      "step": 2800
    },
    {
      "epoch": 1.1028232235276803,
      "grad_norm": 2.060249688323629,
      "learning_rate": 1.4596069566757207e-05,
      "loss": 0.8882,
      "step": 2810
    },
    {
      "epoch": 1.106747774044004,
      "grad_norm": 1.920918977602427,
      "learning_rate": 1.4558395305776731e-05,
      "loss": 0.9299,
      "step": 2820
    },
    {
      "epoch": 1.1106723245603276,
      "grad_norm": 1.6211510195551506,
      "learning_rate": 1.4520639197202355e-05,
      "loss": 0.8898,
      "step": 2830
    },
    {
      "epoch": 1.1145968750766513,
      "grad_norm": 1.229775633396699,
      "learning_rate": 1.4482801918958312e-05,
      "loss": 0.8609,
      "step": 2840
    },
    {
      "epoch": 1.118521425592975,
      "grad_norm": 1.1229141852110762,
      "learning_rate": 1.4444884150426267e-05,
      "loss": 0.8456,
      "step": 2850
    },
    {
      "epoch": 1.1224459761092986,
      "grad_norm": 2.120911251605639,
      "learning_rate": 1.4406886572433113e-05,
      "loss": 0.8918,
      "step": 2860
    },
    {
      "epoch": 1.1263705266256223,
      "grad_norm": 2.0769262910789767,
      "learning_rate": 1.4368809867238754e-05,
      "loss": 0.9531,
      "step": 2870
    },
    {
      "epoch": 1.130295077141946,
      "grad_norm": 1.6225603088474205,
      "learning_rate": 1.4330654718523847e-05,
      "loss": 0.8979,
      "step": 2880
    },
    {
      "epoch": 1.1342196276582697,
      "grad_norm": 1.1891920263768887,
      "learning_rate": 1.4292421811377532e-05,
      "loss": 0.8697,
      "step": 2890
    },
    {
      "epoch": 1.1381441781745933,
      "grad_norm": 1.0159234075717296,
      "learning_rate": 1.4254111832285128e-05,
      "loss": 0.8353,
      "step": 2900
    },
    {
      "epoch": 1.142068728690917,
      "grad_norm": 2.0861820849148383,
      "learning_rate": 1.4215725469115806e-05,
      "loss": 0.8676,
      "step": 2910
    },
    {
      "epoch": 1.1459932792072407,
      "grad_norm": 1.9828582401916874,
      "learning_rate": 1.4177263411110249e-05,
      "loss": 0.9457,
      "step": 2920
    },
    {
      "epoch": 1.1499178297235644,
      "grad_norm": 1.6363978446581915,
      "learning_rate": 1.413872634886825e-05,
      "loss": 0.8682,
      "step": 2930
    },
    {
      "epoch": 1.153842380239888,
      "grad_norm": 1.414190445025758,
      "learning_rate": 1.4100114974336352e-05,
      "loss": 0.8663,
      "step": 2940
    },
    {
      "epoch": 1.1577669307562117,
      "grad_norm": 0.9877954021628765,
      "learning_rate": 1.4061429980795382e-05,
      "loss": 0.8439,
      "step": 2950
    },
    {
      "epoch": 1.1616914812725354,
      "grad_norm": 2.2358539386859726,
      "learning_rate": 1.4022672062848034e-05,
      "loss": 0.8819,
      "step": 2960
    },
    {
      "epoch": 1.165616031788859,
      "grad_norm": 1.8768050513117698,
      "learning_rate": 1.3983841916406383e-05,
      "loss": 0.9261,
      "step": 2970
    },
    {
      "epoch": 1.1695405823051828,
      "grad_norm": 1.6017361554290517,
      "learning_rate": 1.3944940238679384e-05,
      "loss": 0.898,
      "step": 2980
    },
    {
      "epoch": 1.1734651328215064,
      "grad_norm": 1.2526189890877126,
      "learning_rate": 1.390596772816037e-05,
      "loss": 0.8496,
      "step": 2990
    },
    {
      "epoch": 1.1773896833378301,
      "grad_norm": 1.0365383728001132,
      "learning_rate": 1.3866925084614501e-05,
      "loss": 0.8468,
      "step": 3000
    },
    {
      "epoch": 1.1813142338541538,
      "grad_norm": 2.224248057811622,
      "learning_rate": 1.3827813009066202e-05,
      "loss": 0.8759,
      "step": 3010
    },
    {
      "epoch": 1.1852387843704775,
      "grad_norm": 1.8752890091264613,
      "learning_rate": 1.3788632203786567e-05,
      "loss": 0.9297,
      "step": 3020
    },
    {
      "epoch": 1.1891633348868011,
      "grad_norm": 1.6600876268313813,
      "learning_rate": 1.374938337228076e-05,
      "loss": 0.87,
      "step": 3030
    },
    {
      "epoch": 1.1930878854031248,
      "grad_norm": 1.5276898634308227,
      "learning_rate": 1.3710067219275382e-05,
      "loss": 0.8693,
      "step": 3040
    },
    {
      "epoch": 1.1970124359194485,
      "grad_norm": 0.9561545850483195,
      "learning_rate": 1.3670684450705813e-05,
      "loss": 0.8369,
      "step": 3050
    },
    {
      "epoch": 1.2009369864357722,
      "grad_norm": 2.3827528867634307,
      "learning_rate": 1.3631235773703535e-05,
      "loss": 0.8932,
      "step": 3060
    },
    {
      "epoch": 1.2048615369520959,
      "grad_norm": 2.0991202817563828,
      "learning_rate": 1.3591721896583455e-05,
      "loss": 0.9404,
      "step": 3070
    },
    {
      "epoch": 1.2087860874684195,
      "grad_norm": 1.6251446131482838,
      "learning_rate": 1.3552143528831149e-05,
      "loss": 0.8804,
      "step": 3080
    },
    {
      "epoch": 1.2127106379847432,
      "grad_norm": 1.3451655629852488,
      "learning_rate": 1.3512501381090158e-05,
      "loss": 0.8529,
      "step": 3090
    },
    {
      "epoch": 1.2166351885010669,
      "grad_norm": 0.8385047694025927,
      "learning_rate": 1.3472796165149217e-05,
      "loss": 0.818,
      "step": 3100
    },
    {
      "epoch": 1.2205597390173906,
      "grad_norm": 2.267246006997812,
      "learning_rate": 1.3433028593929467e-05,
      "loss": 0.8779,
      "step": 3110
    },
    {
      "epoch": 1.2244842895337142,
      "grad_norm": 2.0703651009730035,
      "learning_rate": 1.3393199381471657e-05,
      "loss": 0.9371,
      "step": 3120
    },
    {
      "epoch": 1.228408840050038,
      "grad_norm": 1.6728269310666124,
      "learning_rate": 1.3353309242923336e-05,
      "loss": 0.862,
      "step": 3130
    },
    {
      "epoch": 1.2323333905663616,
      "grad_norm": 1.3943213550602511,
      "learning_rate": 1.3313358894525997e-05,
      "loss": 0.8734,
      "step": 3140
    },
    {
      "epoch": 1.2362579410826853,
      "grad_norm": 0.9417267831082166,
      "learning_rate": 1.327334905360222e-05,
      "loss": 0.819,
      "step": 3150
    },
    {
      "epoch": 1.240182491599009,
      "grad_norm": 2.03618285439807,
      "learning_rate": 1.3233280438542795e-05,
      "loss": 0.8671,
      "step": 3160
    },
    {
      "epoch": 1.2441070421153326,
      "grad_norm": 1.8648736136084698,
      "learning_rate": 1.319315376879383e-05,
      "loss": 0.945,
      "step": 3170
    },
    {
      "epoch": 1.2480315926316563,
      "grad_norm": 1.5679297693032792,
      "learning_rate": 1.3152969764843812e-05,
      "loss": 0.8778,
      "step": 3180
    },
    {
      "epoch": 1.25195614314798,
      "grad_norm": 1.375990035837938,
      "learning_rate": 1.3112729148210694e-05,
      "loss": 0.8501,
      "step": 3190
    },
    {
      "epoch": 1.2558806936643037,
      "grad_norm": 0.8945069929861201,
      "learning_rate": 1.3072432641428931e-05,
      "loss": 0.8555,
      "step": 3200
    },
    {
      "epoch": 1.2598052441806273,
      "grad_norm": 1.9973442354473332,
      "learning_rate": 1.3032080968036498e-05,
      "loss": 0.8922,
      "step": 3210
    },
    {
      "epoch": 1.263729794696951,
      "grad_norm": 2.1062119639954444,
      "learning_rate": 1.2991674852561904e-05,
      "loss": 0.929,
      "step": 3220
    },
    {
      "epoch": 1.2676543452132747,
      "grad_norm": 1.6116944823889532,
      "learning_rate": 1.2951215020511196e-05,
      "loss": 0.8672,
      "step": 3230
    },
    {
      "epoch": 1.2715788957295984,
      "grad_norm": 1.330101731052534,
      "learning_rate": 1.2910702198354915e-05,
      "loss": 0.853,
      "step": 3240
    },
    {
      "epoch": 1.275503446245922,
      "grad_norm": 1.0644575098102675,
      "learning_rate": 1.2870137113515053e-05,
      "loss": 0.8281,
      "step": 3250
    },
    {
      "epoch": 1.2794279967622457,
      "grad_norm": 2.4521439306633748,
      "learning_rate": 1.2829520494352004e-05,
      "loss": 0.8696,
      "step": 3260
    },
    {
      "epoch": 1.2833525472785694,
      "grad_norm": 2.0163802307404466,
      "learning_rate": 1.2788853070151477e-05,
      "loss": 0.9172,
      "step": 3270
    },
    {
      "epoch": 1.287277097794893,
      "grad_norm": 1.6281065777711845,
      "learning_rate": 1.2748135571111404e-05,
      "loss": 0.8644,
      "step": 3280
    },
    {
      "epoch": 1.2912016483112168,
      "grad_norm": 1.5009876264848987,
      "learning_rate": 1.2707368728328826e-05,
      "loss": 0.8654,
      "step": 3290
    },
    {
      "epoch": 1.2951261988275404,
      "grad_norm": 0.7853723853615091,
      "learning_rate": 1.2666553273786771e-05,
      "loss": 0.8031,
      "step": 3300
    },
    {
      "epoch": 1.299050749343864,
      "grad_norm": 2.189696126410242,
      "learning_rate": 1.2625689940341102e-05,
      "loss": 0.8795,
      "step": 3310
    },
    {
      "epoch": 1.3029752998601878,
      "grad_norm": 1.9334298716685878,
      "learning_rate": 1.2584779461707374e-05,
      "loss": 0.926,
      "step": 3320
    },
    {
      "epoch": 1.3068998503765115,
      "grad_norm": 1.5481872721801182,
      "learning_rate": 1.254382257244765e-05,
      "loss": 0.8565,
      "step": 3330
    },
    {
      "epoch": 1.3108244008928351,
      "grad_norm": 1.2607713095759323,
      "learning_rate": 1.2502820007957302e-05,
      "loss": 0.8505,
      "step": 3340
    },
    {
      "epoch": 1.3147489514091588,
      "grad_norm": 0.8930555325270236,
      "learning_rate": 1.2461772504451822e-05,
      "loss": 0.8323,
      "step": 3350
    },
    {
      "epoch": 1.3186735019254825,
      "grad_norm": 1.9338117374514918,
      "learning_rate": 1.2420680798953604e-05,
      "loss": 0.8754,
      "step": 3360
    },
    {
      "epoch": 1.3225980524418062,
      "grad_norm": 2.061421402861005,
      "learning_rate": 1.2379545629278693e-05,
      "loss": 0.9426,
      "step": 3370
    },
    {
      "epoch": 1.3265226029581298,
      "grad_norm": 1.588986245387613,
      "learning_rate": 1.233836773402356e-05,
      "loss": 0.867,
      "step": 3380
    },
    {
      "epoch": 1.3304471534744535,
      "grad_norm": 1.2886545052157778,
      "learning_rate": 1.229714785255182e-05,
      "loss": 0.868,
      "step": 3390
    },
    {
      "epoch": 1.3343717039907772,
      "grad_norm": 1.219703452074695,
      "learning_rate": 1.2255886724980974e-05,
      "loss": 0.8168,
      "step": 3400
    },
    {
      "epoch": 1.3382962545071009,
      "grad_norm": 2.236888625402866,
      "learning_rate": 1.2214585092169103e-05,
      "loss": 0.8794,
      "step": 3410
    },
    {
      "epoch": 1.3422208050234246,
      "grad_norm": 2.13039765109174,
      "learning_rate": 1.2173243695701575e-05,
      "loss": 0.9218,
      "step": 3420
    },
    {
      "epoch": 1.3461453555397482,
      "grad_norm": 1.652775626275893,
      "learning_rate": 1.213186327787773e-05,
      "loss": 0.8937,
      "step": 3430
    },
    {
      "epoch": 1.350069906056072,
      "grad_norm": 1.469984852272106,
      "learning_rate": 1.209044458169756e-05,
      "loss": 0.8582,
      "step": 3440
    },
    {
      "epoch": 1.3539944565723956,
      "grad_norm": 0.8843940147907386,
      "learning_rate": 1.2048988350848338e-05,
      "loss": 0.8537,
      "step": 3450
    },
    {
      "epoch": 1.3579190070887193,
      "grad_norm": 2.286808087504479,
      "learning_rate": 1.2007495329691301e-05,
      "loss": 0.8915,
      "step": 3460
    },
    {
      "epoch": 1.361843557605043,
      "grad_norm": 2.1637511011528585,
      "learning_rate": 1.1965966263248267e-05,
      "loss": 0.9545,
      "step": 3470
    },
    {
      "epoch": 1.3657681081213666,
      "grad_norm": 1.615967978820237,
      "learning_rate": 1.192440189718825e-05,
      "loss": 0.889,
      "step": 3480
    },
    {
      "epoch": 1.3696926586376903,
      "grad_norm": 1.3561598955089822,
      "learning_rate": 1.1882802977814092e-05,
      "loss": 0.8568,
      "step": 3490
    },
    {
      "epoch": 1.373617209154014,
      "grad_norm": 1.031049725329889,
      "learning_rate": 1.184117025204905e-05,
      "loss": 0.8065,
      "step": 3500
    },
    {
      "epoch": 1.3775417596703377,
      "grad_norm": 2.130942563304923,
      "learning_rate": 1.1799504467423382e-05,
      "loss": 0.8781,
      "step": 3510
    },
    {
      "epoch": 1.3814663101866613,
      "grad_norm": 1.914562387879776,
      "learning_rate": 1.1757806372060934e-05,
      "loss": 0.9244,
      "step": 3520
    },
    {
      "epoch": 1.385390860702985,
      "grad_norm": 1.6890782058344322,
      "learning_rate": 1.1716076714665701e-05,
      "loss": 0.8621,
      "step": 3530
    },
    {
      "epoch": 1.3893154112193087,
      "grad_norm": 1.2006002638188844,
      "learning_rate": 1.1674316244508381e-05,
      "loss": 0.8393,
      "step": 3540
    },
    {
      "epoch": 1.3932399617356324,
      "grad_norm": 1.168530668598062,
      "learning_rate": 1.1632525711412936e-05,
      "loss": 0.8402,
      "step": 3550
    },
    {
      "epoch": 1.397164512251956,
      "grad_norm": 2.214073914981144,
      "learning_rate": 1.1590705865743108e-05,
      "loss": 0.8794,
      "step": 3560
    },
    {
      "epoch": 1.4010890627682797,
      "grad_norm": 1.9175558977832043,
      "learning_rate": 1.1548857458388967e-05,
      "loss": 0.9226,
      "step": 3570
    },
    {
      "epoch": 1.4050136132846034,
      "grad_norm": 1.7516445773007234,
      "learning_rate": 1.1506981240753406e-05,
      "loss": 0.8688,
      "step": 3580
    },
    {
      "epoch": 1.408938163800927,
      "grad_norm": 1.236950976064499,
      "learning_rate": 1.1465077964738674e-05,
      "loss": 0.8604,
      "step": 3590
    },
    {
      "epoch": 1.4128627143172507,
      "grad_norm": 0.9946584575727137,
      "learning_rate": 1.1423148382732854e-05,
      "loss": 0.8198,
      "step": 3600
    },
    {
      "epoch": 1.4167872648335744,
      "grad_norm": 2.357981326812003,
      "learning_rate": 1.1381193247596365e-05,
      "loss": 0.8815,
      "step": 3610
    },
    {
      "epoch": 1.420711815349898,
      "grad_norm": 2.068839516965018,
      "learning_rate": 1.133921331264844e-05,
      "loss": 0.9218,
      "step": 3620
    },
    {
      "epoch": 1.4246363658662218,
      "grad_norm": 1.7728344994302516,
      "learning_rate": 1.1297209331653606e-05,
      "loss": 0.8679,
      "step": 3630
    },
    {
      "epoch": 1.4285609163825455,
      "grad_norm": 1.2048121431631056,
      "learning_rate": 1.1255182058808143e-05,
      "loss": 0.8544,
      "step": 3640
    },
    {
      "epoch": 1.4324854668988691,
      "grad_norm": 1.333275738578586,
      "learning_rate": 1.1213132248726541e-05,
      "loss": 0.8409,
      "step": 3650
    },
    {
      "epoch": 1.4364100174151928,
      "grad_norm": 2.206900104214621,
      "learning_rate": 1.1171060656427957e-05,
      "loss": 0.881,
      "step": 3660
    },
    {
      "epoch": 1.4403345679315165,
      "grad_norm": 2.0032350194617923,
      "learning_rate": 1.1128968037322654e-05,
      "loss": 0.9304,
      "step": 3670
    },
    {
      "epoch": 1.4442591184478402,
      "grad_norm": 1.899875816582961,
      "learning_rate": 1.1086855147198442e-05,
      "loss": 0.8756,
      "step": 3680
    },
    {
      "epoch": 1.4481836689641638,
      "grad_norm": 1.1750450632734881,
      "learning_rate": 1.1044722742207102e-05,
      "loss": 0.8516,
      "step": 3690
    },
    {
      "epoch": 1.4521082194804875,
      "grad_norm": 1.0654206477147208,
      "learning_rate": 1.1002571578850808e-05,
      "loss": 0.8287,
      "step": 3700
    },
    {
      "epoch": 1.4560327699968112,
      "grad_norm": 2.1156979827088698,
      "learning_rate": 1.0960402413968552e-05,
      "loss": 0.8747,
      "step": 3710
    },
    {
      "epoch": 1.4599573205131349,
      "grad_norm": 1.9613993823327838,
      "learning_rate": 1.0918216004722551e-05,
      "loss": 0.9248,
      "step": 3720
    },
    {
      "epoch": 1.4638818710294585,
      "grad_norm": 1.5896298448753268,
      "learning_rate": 1.0876013108584644e-05,
      "loss": 0.862,
      "step": 3730
    },
    {
      "epoch": 1.4678064215457822,
      "grad_norm": 1.2663265403929282,
      "learning_rate": 1.08337944833227e-05,
      "loss": 0.8671,
      "step": 3740
    },
    {
      "epoch": 1.471730972062106,
      "grad_norm": 0.8402022117082641,
      "learning_rate": 1.0791560886987016e-05,
      "loss": 0.8089,
      "step": 3750
    },
    {
      "epoch": 1.4756555225784296,
      "grad_norm": 2.2895246842794346,
      "learning_rate": 1.0749313077896697e-05,
      "loss": 0.8865,
      "step": 3760
    },
    {
      "epoch": 1.4795800730947533,
      "grad_norm": 2.005690344314872,
      "learning_rate": 1.0707051814626035e-05,
      "loss": 0.9195,
      "step": 3770
    },
    {
      "epoch": 1.483504623611077,
      "grad_norm": 1.6528392426188365,
      "learning_rate": 1.0664777855990909e-05,
      "loss": 0.8482,
      "step": 3780
    },
    {
      "epoch": 1.4874291741274006,
      "grad_norm": 1.3252339862500955,
      "learning_rate": 1.062249196103514e-05,
      "loss": 0.8633,
      "step": 3790
    },
    {
      "epoch": 1.4913537246437243,
      "grad_norm": 0.8384617708635065,
      "learning_rate": 1.0580194889016866e-05,
      "loss": 0.8424,
      "step": 3800
    },
    {
      "epoch": 1.495278275160048,
      "grad_norm": 2.2376729601071013,
      "learning_rate": 1.0537887399394926e-05,
      "loss": 0.8698,
      "step": 3810
    },
    {
      "epoch": 1.4992028256763716,
      "grad_norm": 2.0054017581982158,
      "learning_rate": 1.0495570251815204e-05,
      "loss": 0.9146,
      "step": 3820
    },
    {
      "epoch": 1.5031273761926953,
      "grad_norm": 1.620117177752491,
      "learning_rate": 1.0453244206096993e-05,
      "loss": 0.86,
      "step": 3830
    },
    {
      "epoch": 1.507051926709019,
      "grad_norm": 1.2414222987018593,
      "learning_rate": 1.0410910022219356e-05,
      "loss": 0.8462,
      "step": 3840
    },
    {
      "epoch": 1.5109764772253427,
      "grad_norm": 0.985434847255323,
      "learning_rate": 1.0368568460307482e-05,
      "loss": 0.8374,
      "step": 3850
    },
    {
      "epoch": 1.5149010277416664,
      "grad_norm": 2.414664203183688,
      "learning_rate": 1.0326220280619036e-05,
      "loss": 0.8643,
      "step": 3860
    },
    {
      "epoch": 1.51882557825799,
      "grad_norm": 1.959661011938626,
      "learning_rate": 1.0283866243530506e-05,
      "loss": 0.9216,
      "step": 3870
    },
    {
      "epoch": 1.5227501287743137,
      "grad_norm": 1.8216460410768873,
      "learning_rate": 1.0241507109523551e-05,
      "loss": 0.8557,
      "step": 3880
    },
    {
      "epoch": 1.5266746792906374,
      "grad_norm": 1.2379402019260293,
      "learning_rate": 1.019914363917135e-05,
      "loss": 0.8528,
      "step": 3890
    },
    {
      "epoch": 1.530599229806961,
      "grad_norm": 1.1791841499933475,
      "learning_rate": 1.0156776593124933e-05,
      "loss": 0.8409,
      "step": 3900
    },
    {
      "epoch": 1.5345237803232847,
      "grad_norm": 2.2744852259625294,
      "learning_rate": 1.0114406732099549e-05,
      "loss": 0.877,
      "step": 3910
    },
    {
      "epoch": 1.5384483308396084,
      "grad_norm": 1.9131218031571517,
      "learning_rate": 1.0072034816860979e-05,
      "loss": 0.9287,
      "step": 3920
    },
    {
      "epoch": 1.542372881355932,
      "grad_norm": 1.7840885206492576,
      "learning_rate": 1.0029661608211884e-05,
      "loss": 0.8511,
      "step": 3930
    },
    {
      "epoch": 1.5462974318722558,
      "grad_norm": 1.227268938632673,
      "learning_rate": 9.987287866978169e-06,
      "loss": 0.8535,
      "step": 3940
    },
    {
      "epoch": 1.5502219823885794,
      "grad_norm": 1.1621128799647606,
      "learning_rate": 9.944914353995277e-06,
      "loss": 0.8447,
      "step": 3950
    },
    {
      "epoch": 1.5541465329049031,
      "grad_norm": 2.2074040426072483,
      "learning_rate": 9.90254183009457e-06,
      "loss": 0.8529,
      "step": 3960
    },
    {
      "epoch": 1.5580710834212268,
      "grad_norm": 1.8956680374875223,
      "learning_rate": 9.860171056089646e-06,
      "loss": 0.9103,
      "step": 3970
    },
    {
      "epoch": 1.5619956339375505,
      "grad_norm": 1.8417783539473633,
      "learning_rate": 9.817802792762675e-06,
      "loss": 0.8619,
      "step": 3980
    },
    {
      "epoch": 1.5659201844538742,
      "grad_norm": 1.1239027585518444,
      "learning_rate": 9.775437800850764e-06,
      "loss": 0.8405,
      "step": 3990
    },
    {
      "epoch": 1.5698447349701978,
      "grad_norm": 1.093899462684469,
      "learning_rate": 9.73307684103226e-06,
      "loss": 0.8409,
      "step": 4000
    },
    {
      "epoch": 1.5737692854865215,
      "grad_norm": 2.2458835890160698,
      "learning_rate": 9.690720673913135e-06,
      "loss": 0.8331,
      "step": 4010
    },
    {
      "epoch": 1.5776938360028452,
      "grad_norm": 1.9788980494060489,
      "learning_rate": 9.648370060013279e-06,
      "loss": 0.9097,
      "step": 4020
    },
    {
      "epoch": 1.5816183865191689,
      "grad_norm": 1.627824694857222,
      "learning_rate": 9.606025759752895e-06,
      "loss": 0.8831,
      "step": 4030
    },
    {
      "epoch": 1.5855429370354925,
      "grad_norm": 1.2178062700728904,
      "learning_rate": 9.56368853343882e-06,
      "loss": 0.8462,
      "step": 4040
    },
    {
      "epoch": 1.5894674875518162,
      "grad_norm": 0.8688900774056482,
      "learning_rate": 9.52135914125086e-06,
      "loss": 0.8132,
      "step": 4050
    },
    {
      "epoch": 1.59339203806814,
      "grad_norm": 2.223343280091266,
      "learning_rate": 9.479038343228173e-06,
      "loss": 0.8987,
      "step": 4060
    },
    {
      "epoch": 1.5973165885844636,
      "grad_norm": 1.9082308916293935,
      "learning_rate": 9.436726899255596e-06,
      "loss": 0.9305,
      "step": 4070
    },
    {
      "epoch": 1.6012411391007872,
      "grad_norm": 1.6281376379032095,
      "learning_rate": 9.394425569050018e-06,
      "loss": 0.8806,
      "step": 4080
    },
    {
      "epoch": 1.605165689617111,
      "grad_norm": 1.3465856506420029,
      "learning_rate": 9.352135112146726e-06,
      "loss": 0.8553,
      "step": 4090
    },
    {
      "epoch": 1.6090902401334346,
      "grad_norm": 0.8680202506295652,
      "learning_rate": 9.309856287885775e-06,
      "loss": 0.8224,
      "step": 4100
    },
    {
      "epoch": 1.6130147906497583,
      "grad_norm": 2.0551986939006266,
      "learning_rate": 9.267589855398356e-06,
      "loss": 0.866,
      "step": 4110
    },
    {
      "epoch": 1.616939341166082,
      "grad_norm": 2.009568214502001,
      "learning_rate": 9.22533657359315e-06,
      "loss": 0.9291,
      "step": 4120
    },
    {
      "epoch": 1.6208638916824056,
      "grad_norm": 1.63900979800811,
      "learning_rate": 9.183097201142722e-06,
      "loss": 0.8596,
      "step": 4130
    },
    {
      "epoch": 1.6247884421987293,
      "grad_norm": 1.270695193750196,
      "learning_rate": 9.140872496469891e-06,
      "loss": 0.8496,
      "step": 4140
    },
    {
      "epoch": 1.628712992715053,
      "grad_norm": 0.800400571374905,
      "learning_rate": 9.098663217734102e-06,
      "loss": 0.8171,
      "step": 4150
    },
    {
      "epoch": 1.6326375432313767,
      "grad_norm": 2.239858328605921,
      "learning_rate": 9.056470122817836e-06,
      "loss": 0.8696,
      "step": 4160
    },
    {
      "epoch": 1.6365620937477003,
      "grad_norm": 2.0001741452838444,
      "learning_rate": 9.01429396931297e-06,
      "loss": 0.9174,
      "step": 4170
    },
    {
      "epoch": 1.640486644264024,
      "grad_norm": 1.6873988098563508,
      "learning_rate": 8.972135514507212e-06,
      "loss": 0.8725,
      "step": 4180
    },
    {
      "epoch": 1.6444111947803477,
      "grad_norm": 1.1871894670080978,
      "learning_rate": 8.92999551537046e-06,
      "loss": 0.819,
      "step": 4190
    },
    {
      "epoch": 1.6483357452966714,
      "grad_norm": 0.8255765266436247,
      "learning_rate": 8.88787472854126e-06,
      "loss": 0.8178,
      "step": 4200
    },
    {
      "epoch": 1.652260295812995,
      "grad_norm": 2.1797960381842927,
      "learning_rate": 8.845773910313168e-06,
      "loss": 0.8486,
      "step": 4210
    },
    {
      "epoch": 1.6561848463293187,
      "grad_norm": 1.9916857527591452,
      "learning_rate": 8.803693816621218e-06,
      "loss": 0.8947,
      "step": 4220
    },
    {
      "epoch": 1.6601093968456424,
      "grad_norm": 1.7685848462081732,
      "learning_rate": 8.761635203028319e-06,
      "loss": 0.8766,
      "step": 4230
    },
    {
      "epoch": 1.664033947361966,
      "grad_norm": 1.245584977919593,
      "learning_rate": 8.719598824711694e-06,
      "loss": 0.8337,
      "step": 4240
    },
    {
      "epoch": 1.6679584978782898,
      "grad_norm": 1.0048598136091462,
      "learning_rate": 8.677585436449332e-06,
      "loss": 0.8163,
      "step": 4250
    },
    {
      "epoch": 1.6718830483946134,
      "grad_norm": 2.398297940398292,
      "learning_rate": 8.635595792606419e-06,
      "loss": 0.8559,
      "step": 4260
    },
    {
      "epoch": 1.6758075989109371,
      "grad_norm": 1.9730845143153721,
      "learning_rate": 8.593630647121809e-06,
      "loss": 0.895,
      "step": 4270
    },
    {
      "epoch": 1.6797321494272608,
      "grad_norm": 1.6696729781305142,
      "learning_rate": 8.551690753494476e-06,
      "loss": 0.8508,
      "step": 4280
    },
    {
      "epoch": 1.6836566999435845,
      "grad_norm": 1.3530608242464415,
      "learning_rate": 8.509776864769982e-06,
      "loss": 0.8295,
      "step": 4290
    },
    {
      "epoch": 1.6875812504599081,
      "grad_norm": 0.8499210708063567,
      "learning_rate": 8.467889733526977e-06,
      "loss": 0.8245,
      "step": 4300
    },
    {
      "epoch": 1.6915058009762318,
      "grad_norm": 2.3773533821815067,
      "learning_rate": 8.426030111863654e-06,
      "loss": 0.8521,
      "step": 4310
    },
    {
      "epoch": 1.6954303514925555,
      "grad_norm": 2.077394013322105,
      "learning_rate": 8.384198751384272e-06,
      "loss": 0.9227,
      "step": 4320
    },
    {
      "epoch": 1.6993549020088792,
      "grad_norm": 1.6718741898436833,
      "learning_rate": 8.342396403185649e-06,
      "loss": 0.8448,
      "step": 4330
    },
    {
      "epoch": 1.7032794525252029,
      "grad_norm": 1.3584330125549005,
      "learning_rate": 8.300623817843673e-06,
      "loss": 0.8385,
      "step": 4340
    },
    {
      "epoch": 1.7072040030415265,
      "grad_norm": 0.9817563909977678,
      "learning_rate": 8.258881745399837e-06,
      "loss": 0.8062,
      "step": 4350
    },
    {
      "epoch": 1.7111285535578502,
      "grad_norm": 2.3439029636827127,
      "learning_rate": 8.217170935347756e-06,
      "loss": 0.8164,
      "step": 4360
    },
    {
      "epoch": 1.7150531040741739,
      "grad_norm": 2.016656786265483,
      "learning_rate": 8.17549213661973e-06,
      "loss": 0.8954,
      "step": 4370
    },
    {
      "epoch": 1.7189776545904976,
      "grad_norm": 1.7641936297392027,
      "learning_rate": 8.133846097573263e-06,
      "loss": 0.8658,
      "step": 4380
    },
    {
      "epoch": 1.7229022051068212,
      "grad_norm": 1.289046077866149,
      "learning_rate": 8.09223356597767e-06,
      "loss": 0.8291,
      "step": 4390
    },
    {
      "epoch": 1.726826755623145,
      "grad_norm": 0.8325407065873541,
      "learning_rate": 8.050655289000612e-06,
      "loss": 0.8168,
      "step": 4400
    },
    {
      "epoch": 1.7307513061394686,
      "grad_norm": 2.6948112568790124,
      "learning_rate": 8.009112013194707e-06,
      "loss": 0.8495,
      "step": 4410
    },
    {
      "epoch": 1.7346758566557923,
      "grad_norm": 1.9659980108114699,
      "learning_rate": 7.96760448448411e-06,
      "loss": 0.8947,
      "step": 4420
    },
    {
      "epoch": 1.738600407172116,
      "grad_norm": 1.8751083414206937,
      "learning_rate": 7.926133448151121e-06,
      "loss": 0.8493,
      "step": 4430
    },
    {
      "epoch": 1.7425249576884396,
      "grad_norm": 1.2387882054830557,
      "learning_rate": 7.884699648822816e-06,
      "loss": 0.8267,
      "step": 4440
    },
    {
      "epoch": 1.7464495082047633,
      "grad_norm": 1.1399825715598682,
      "learning_rate": 7.843303830457654e-06,
      "loss": 0.791,
      "step": 4450
    },
    {
      "epoch": 1.750374058721087,
      "grad_norm": 2.411702115321597,
      "learning_rate": 7.801946736332144e-06,
      "loss": 0.8578,
      "step": 4460
    },
    {
      "epoch": 1.7542986092374107,
      "grad_norm": 2.125672710751084,
      "learning_rate": 7.760629109027488e-06,
      "loss": 0.8945,
      "step": 4470
    },
    {
      "epoch": 1.7582231597537343,
      "grad_norm": 1.6589788634772225,
      "learning_rate": 7.719351690416234e-06,
      "loss": 0.8528,
      "step": 4480
    },
    {
      "epoch": 1.762147710270058,
      "grad_norm": 1.2339299024543553,
      "learning_rate": 7.678115221648983e-06,
      "loss": 0.8264,
      "step": 4490
    },
    {
      "epoch": 1.7660722607863817,
      "grad_norm": 0.8818646846150126,
      "learning_rate": 7.636920443141057e-06,
      "loss": 0.7858,
      "step": 4500
    },
    {
      "epoch": 1.7699968113027054,
      "grad_norm": 2.227068159454661,
      "learning_rate": 7.595768094559226e-06,
      "loss": 0.8546,
      "step": 4510
    },
    {
      "epoch": 1.773921361819029,
      "grad_norm": 2.0038773703981527,
      "learning_rate": 7.554658914808404e-06,
      "loss": 0.8974,
      "step": 4520
    },
    {
      "epoch": 1.7778459123353527,
      "grad_norm": 1.559249536864396,
      "learning_rate": 7.513593642018398e-06,
      "loss": 0.8488,
      "step": 4530
    },
    {
      "epoch": 1.7817704628516764,
      "grad_norm": 1.2969189118652353,
      "learning_rate": 7.472573013530657e-06,
      "loss": 0.8509,
      "step": 4540
    },
    {
      "epoch": 1.785695013368,
      "grad_norm": 0.8958343326615786,
      "learning_rate": 7.431597765885013e-06,
      "loss": 0.7997,
      "step": 4550
    },
    {
      "epoch": 1.7896195638843237,
      "grad_norm": 2.217098339251589,
      "learning_rate": 7.39066863480648e-06,
      "loss": 0.8348,
      "step": 4560
    },
    {
      "epoch": 1.7935441144006474,
      "grad_norm": 2.1265537658801117,
      "learning_rate": 7.349786355192023e-06,
      "loss": 0.8944,
      "step": 4570
    },
    {
      "epoch": 1.797468664916971,
      "grad_norm": 1.6797170624498905,
      "learning_rate": 7.308951661097379e-06,
      "loss": 0.8448,
      "step": 4580
    },
    {
      "epoch": 1.8013932154332948,
      "grad_norm": 1.1786118411152537,
      "learning_rate": 7.268165285723875e-06,
      "loss": 0.8474,
      "step": 4590
    },
    {
      "epoch": 1.8053177659496185,
      "grad_norm": 0.8503022381368873,
      "learning_rate": 7.227427961405245e-06,
      "loss": 0.7908,
      "step": 4600
    },
    {
      "epoch": 1.8092423164659421,
      "grad_norm": 2.3703275938509525,
      "learning_rate": 7.186740419594505e-06,
      "loss": 0.845,
      "step": 4610
    },
    {
      "epoch": 1.8131668669822658,
      "grad_norm": 2.0576172422877073,
      "learning_rate": 7.1461033908508004e-06,
      "loss": 0.9065,
      "step": 4620
    },
    {
      "epoch": 1.8170914174985895,
      "grad_norm": 1.6463925064837197,
      "learning_rate": 7.1055176048263085e-06,
      "loss": 0.842,
      "step": 4630
    },
    {
      "epoch": 1.8210159680149132,
      "grad_norm": 1.3242222800874879,
      "learning_rate": 7.0649837902531095e-06,
      "loss": 0.8499,
      "step": 4640
    },
    {
      "epoch": 1.8249405185312368,
      "grad_norm": 0.8538674555412706,
      "learning_rate": 7.0245026749301315e-06,
      "loss": 0.8046,
      "step": 4650
    },
    {
      "epoch": 1.8288650690475605,
      "grad_norm": 2.4490816400507516,
      "learning_rate": 6.984074985710068e-06,
      "loss": 0.8529,
      "step": 4660
    },
    {
      "epoch": 1.8327896195638842,
      "grad_norm": 2.0294960511407822,
      "learning_rate": 6.943701448486313e-06,
      "loss": 0.8992,
      "step": 4670
    },
    {
      "epoch": 1.8367141700802079,
      "grad_norm": 1.680488532778669,
      "learning_rate": 6.903382788179962e-06,
      "loss": 0.8566,
      "step": 4680
    },
    {
      "epoch": 1.8406387205965316,
      "grad_norm": 1.2281066012509496,
      "learning_rate": 6.8631197287267636e-06,
      "loss": 0.8376,
      "step": 4690
    },
    {
      "epoch": 1.8445632711128552,
      "grad_norm": 1.2188089446996344,
      "learning_rate": 6.82291299306414e-06,
      "loss": 0.8058,
      "step": 4700
    },
    {
      "epoch": 1.848487821629179,
      "grad_norm": 2.1748995033926497,
      "learning_rate": 6.782763303118194e-06,
      "loss": 0.8464,
      "step": 4710
    },
    {
      "epoch": 1.8524123721455026,
      "grad_norm": 2.0279495030924046,
      "learning_rate": 6.742671379790756e-06,
      "loss": 0.8782,
      "step": 4720
    },
    {
      "epoch": 1.8563369226618263,
      "grad_norm": 1.6914708882579899,
      "learning_rate": 6.702637942946441e-06,
      "loss": 0.8422,
      "step": 4730
    },
    {
      "epoch": 1.86026147317815,
      "grad_norm": 1.3001032946364874,
      "learning_rate": 6.662663711399705e-06,
      "loss": 0.8189,
      "step": 4740
    },
    {
      "epoch": 1.8641860236944736,
      "grad_norm": 0.9695568949982307,
      "learning_rate": 6.622749402901971e-06,
      "loss": 0.7972,
      "step": 4750
    },
    {
      "epoch": 1.8681105742107973,
      "grad_norm": 2.3580357832714696,
      "learning_rate": 6.5828957341287025e-06,
      "loss": 0.8602,
      "step": 4760
    },
    {
      "epoch": 1.872035124727121,
      "grad_norm": 2.053651822379713,
      "learning_rate": 6.5431034206665686e-06,
      "loss": 0.8946,
      "step": 4770
    },
    {
      "epoch": 1.8759596752434446,
      "grad_norm": 1.6781331457753144,
      "learning_rate": 6.503373177000582e-06,
      "loss": 0.8479,
      "step": 4780
    },
    {
      "epoch": 1.8798842257597683,
      "grad_norm": 1.3439461982630885,
      "learning_rate": 6.463705716501261e-06,
      "loss": 0.8108,
      "step": 4790
    },
    {
      "epoch": 1.883808776276092,
      "grad_norm": 0.8215956260267698,
      "learning_rate": 6.424101751411842e-06,
      "loss": 0.8124,
      "step": 4800
    },
    {
      "epoch": 1.8877333267924157,
      "grad_norm": 2.2909082972168275,
      "learning_rate": 6.3845619928354676e-06,
      "loss": 0.8253,
      "step": 4810
    },
    {
      "epoch": 1.8916578773087394,
      "grad_norm": 2.1043172809575057,
      "learning_rate": 6.345087150722441e-06,
      "loss": 0.8767,
      "step": 4820
    },
    {
      "epoch": 1.895582427825063,
      "grad_norm": 1.7239356576641465,
      "learning_rate": 6.305677933857455e-06,
      "loss": 0.8217,
      "step": 4830
    },
    {
      "epoch": 1.8995069783413867,
      "grad_norm": 1.4857325973939928,
      "learning_rate": 6.266335049846886e-06,
      "loss": 0.8415,
      "step": 4840
    },
    {
      "epoch": 1.9034315288577104,
      "grad_norm": 0.9334678918114299,
      "learning_rate": 6.227059205106085e-06,
      "loss": 0.7717,
      "step": 4850
    },
    {
      "epoch": 1.907356079374034,
      "grad_norm": 2.2849232645875297,
      "learning_rate": 6.187851104846676e-06,
      "loss": 0.846,
      "step": 4860
    },
    {
      "epoch": 1.9112806298903577,
      "grad_norm": 1.945016141615289,
      "learning_rate": 6.1487114530639205e-06,
      "loss": 0.8882,
      "step": 4870
    },
    {
      "epoch": 1.9152051804066814,
      "grad_norm": 1.7795684379972176,
      "learning_rate": 6.109640952524052e-06,
      "loss": 0.8329,
      "step": 4880
    },
    {
      "epoch": 1.919129730923005,
      "grad_norm": 1.2407252843374668,
      "learning_rate": 6.070640304751677e-06,
      "loss": 0.8251,
      "step": 4890
    },
    {
      "epoch": 1.9230542814393288,
      "grad_norm": 0.8194399568120128,
      "learning_rate": 6.031710210017171e-06,
      "loss": 0.7867,
      "step": 4900
    },
    {
      "epoch": 1.9269788319556524,
      "grad_norm": 2.3956208299915955,
      "learning_rate": 5.992851367324097e-06,
      "loss": 0.8433,
      "step": 4910
    },
    {
      "epoch": 1.9309033824719761,
      "grad_norm": 2.0827483565619915,
      "learning_rate": 5.954064474396675e-06,
      "loss": 0.885,
      "step": 4920
    },
    {
      "epoch": 1.9348279329882998,
      "grad_norm": 1.70357287888689,
      "learning_rate": 5.915350227667225e-06,
      "loss": 0.8385,
      "step": 4930
    },
    {
      "epoch": 1.9387524835046235,
      "grad_norm": 1.2786111455423548,
      "learning_rate": 5.876709322263696e-06,
      "loss": 0.8207,
      "step": 4940
    },
    {
      "epoch": 1.9426770340209472,
      "grad_norm": 0.7671237740151083,
      "learning_rate": 5.838142451997155e-06,
      "loss": 0.8048,
      "step": 4950
    },
    {
      "epoch": 1.9466015845372708,
      "grad_norm": 2.3292976163560546,
      "learning_rate": 5.799650309349348e-06,
      "loss": 0.8462,
      "step": 4960
    },
    {
      "epoch": 1.9505261350535945,
      "grad_norm": 2.077691017439424,
      "learning_rate": 5.761233585460265e-06,
      "loss": 0.9123,
      "step": 4970
    },
    {
      "epoch": 1.9544506855699182,
      "grad_norm": 1.6563795333206879,
      "learning_rate": 5.722892970115712e-06,
      "loss": 0.8154,
      "step": 4980
    },
    {
      "epoch": 1.9583752360862419,
      "grad_norm": 1.3593453278887475,
      "learning_rate": 5.684629151734949e-06,
      "loss": 0.8108,
      "step": 4990
    },
    {
      "epoch": 1.9622997866025655,
      "grad_norm": 0.8477008150249254,
      "learning_rate": 5.6464428173583174e-06,
      "loss": 0.7722,
      "step": 5000
    },
    {
      "epoch": 1.9662243371188892,
      "grad_norm": 2.3204148587243103,
      "learning_rate": 5.608334652634914e-06,
      "loss": 0.8387,
      "step": 5010
    },
    {
      "epoch": 1.970148887635213,
      "grad_norm": 2.08224549041347,
      "learning_rate": 5.570305341810252e-06,
      "loss": 0.8696,
      "step": 5020
    },
    {
      "epoch": 1.9740734381515366,
      "grad_norm": 1.6687767550667691,
      "learning_rate": 5.532355567714013e-06,
      "loss": 0.8536,
      "step": 5030
    },
    {
      "epoch": 1.9779979886678603,
      "grad_norm": 1.2616375376723235,
      "learning_rate": 5.494486011747761e-06,
      "loss": 0.8102,
      "step": 5040
    },
    {
      "epoch": 1.981922539184184,
      "grad_norm": 0.7574755635671572,
      "learning_rate": 5.4566973538727216e-06,
      "loss": 0.7677,
      "step": 5050
    },
    {
      "epoch": 1.9858470897005076,
      "grad_norm": 2.309973402310073,
      "learning_rate": 5.418990272597561e-06,
      "loss": 0.839,
      "step": 5060
    },
    {
      "epoch": 1.9897716402168313,
      "grad_norm": 2.196528214038538,
      "learning_rate": 5.381365444966205e-06,
      "loss": 0.8893,
      "step": 5070
    },
    {
      "epoch": 1.993696190733155,
      "grad_norm": 1.6515133028295628,
      "learning_rate": 5.3438235465456926e-06,
      "loss": 0.8053,
      "step": 5080
    },
    {
      "epoch": 1.9976207412494786,
      "grad_norm": 1.4943879150775243,
      "learning_rate": 5.306365251414043e-06,
      "loss": 0.798,
      "step": 5090
    },
    {
      "epoch": 1.999975471559273,
      "eval_loss": 0.6450071930885315,
      "eval_runtime": 1529.4328,
      "eval_samples_per_second": 16.346,
      "eval_steps_per_second": 4.086,
      "step": 5096
    },
    {
      "epoch": 2.0015698202065293,
      "grad_norm": 1.4784398750719419,
      "learning_rate": 5.268991232148137e-06,
      "loss": 0.8525,
      "step": 5100
    },
    {
      "epoch": 2.005494370722853,
      "grad_norm": 9.338133108180491,
      "learning_rate": 5.2317021598116635e-06,
      "loss": 0.5966,
      "step": 5110
    },
    {
      "epoch": 2.0094189212391766,
      "grad_norm": 1.8570667185487677,
      "learning_rate": 5.1944987039430535e-06,
      "loss": 0.6489,
      "step": 5120
    },
    {
      "epoch": 2.0133434717555003,
      "grad_norm": 2.1202392786052697,
      "learning_rate": 5.157381532543473e-06,
      "loss": 0.6784,
      "step": 5130
    },
    {
      "epoch": 2.017268022271824,
      "grad_norm": 1.700823660700944,
      "learning_rate": 5.120351312064802e-06,
      "loss": 0.6222,
      "step": 5140
    },
    {
      "epoch": 2.0211925727881477,
      "grad_norm": 1.3821129005652262,
      "learning_rate": 5.083408707397704e-06,
      "loss": 0.6147,
      "step": 5150
    },
    {
      "epoch": 2.0251171233044714,
      "grad_norm": 1.0951660096886453,
      "learning_rate": 5.046554381859663e-06,
      "loss": 0.5845,
      "step": 5160
    },
    {
      "epoch": 2.029041673820795,
      "grad_norm": 2.138790453625462,
      "learning_rate": 5.009788997183074e-06,
      "loss": 0.6237,
      "step": 5170
    },
    {
      "epoch": 2.0329662243371187,
      "grad_norm": 2.0579134568589597,
      "learning_rate": 4.973113213503379e-06,
      "loss": 0.705,
      "step": 5180
    },
    {
      "epoch": 2.0368907748534424,
      "grad_norm": 1.75667330374861,
      "learning_rate": 4.936527689347195e-06,
      "loss": 0.6389,
      "step": 5190
    },
    {
      "epoch": 2.040815325369766,
      "grad_norm": 1.308594348694505,
      "learning_rate": 4.9000330816205e-06,
      "loss": 0.6035,
      "step": 5200
    },
    {
      "epoch": 2.0447398758860897,
      "grad_norm": 0.9410941640978612,
      "learning_rate": 4.863630045596838e-06,
      "loss": 0.5541,
      "step": 5210
    },
    {
      "epoch": 2.0486644264024134,
      "grad_norm": 2.1522254372307357,
      "learning_rate": 4.8273192349055405e-06,
      "loss": 0.5952,
      "step": 5220
    },
    {
      "epoch": 2.052588976918737,
      "grad_norm": 2.036592335343225,
      "learning_rate": 4.791101301520016e-06,
      "loss": 0.6809,
      "step": 5230
    },
    {
      "epoch": 2.0565135274350608,
      "grad_norm": 1.686946791293678,
      "learning_rate": 4.754976895746007e-06,
      "loss": 0.6342,
      "step": 5240
    },
    {
      "epoch": 2.0604380779513845,
      "grad_norm": 1.3407168000815266,
      "learning_rate": 4.718946666209966e-06,
      "loss": 0.6237,
      "step": 5250
    },
    {
      "epoch": 2.064362628467708,
      "grad_norm": 0.8377432352577971,
      "learning_rate": 4.683011259847346e-06,
      "loss": 0.5427,
      "step": 5260
    },
    {
      "epoch": 2.068287178984032,
      "grad_norm": 2.1455140854622496,
      "learning_rate": 4.647171321891034e-06,
      "loss": 0.6384,
      "step": 5270
    },
    {
      "epoch": 2.0722117295003555,
      "grad_norm": 1.99774007410406,
      "learning_rate": 4.61142749585975e-06,
      "loss": 0.667,
      "step": 5280
    },
    {
      "epoch": 2.076136280016679,
      "grad_norm": 1.7474636373639787,
      "learning_rate": 4.575780423546476e-06,
      "loss": 0.6309,
      "step": 5290
    },
    {
      "epoch": 2.080060830533003,
      "grad_norm": 1.3359269678333927,
      "learning_rate": 4.540230745006962e-06,
      "loss": 0.5829,
      "step": 5300
    },
    {
      "epoch": 2.0839853810493265,
      "grad_norm": 0.7162710672069984,
      "learning_rate": 4.504779098548209e-06,
      "loss": 0.5332,
      "step": 5310
    },
    {
      "epoch": 2.08790993156565,
      "grad_norm": 2.0989757061484893,
      "learning_rate": 4.469426120717025e-06,
      "loss": 0.624,
      "step": 5320
    },
    {
      "epoch": 2.091834482081974,
      "grad_norm": 2.1426189322836975,
      "learning_rate": 4.434172446288579e-06,
      "loss": 0.6681,
      "step": 5330
    },
    {
      "epoch": 2.0957590325982975,
      "grad_norm": 1.776402952977803,
      "learning_rate": 4.399018708255018e-06,
      "loss": 0.6193,
      "step": 5340
    },
    {
      "epoch": 2.099683583114621,
      "grad_norm": 1.3008240753278815,
      "learning_rate": 4.363965537814102e-06,
      "loss": 0.6082,
      "step": 5350
    },
    {
      "epoch": 2.103608133630945,
      "grad_norm": 0.8996256226429636,
      "learning_rate": 4.329013564357848e-06,
      "loss": 0.5629,
      "step": 5360
    },
    {
      "epoch": 2.1075326841472686,
      "grad_norm": 2.0984085235237004,
      "learning_rate": 4.294163415461258e-06,
      "loss": 0.6169,
      "step": 5370
    },
    {
      "epoch": 2.1114572346635923,
      "grad_norm": 2.1055348326355667,
      "learning_rate": 4.259415716871037e-06,
      "loss": 0.6725,
      "step": 5380
    },
    {
      "epoch": 2.115381785179916,
      "grad_norm": 1.7564083013798226,
      "learning_rate": 4.224771092494355e-06,
      "loss": 0.6177,
      "step": 5390
    },
    {
      "epoch": 2.1193063356962396,
      "grad_norm": 1.3081730861592322,
      "learning_rate": 4.1902301643876555e-06,
      "loss": 0.5994,
      "step": 5400
    },
    {
      "epoch": 2.1232308862125633,
      "grad_norm": 0.7914811793114824,
      "learning_rate": 4.155793552745465e-06,
      "loss": 0.5642,
      "step": 5410
    },
    {
      "epoch": 2.127155436728887,
      "grad_norm": 2.4378415532004287,
      "learning_rate": 4.1214618758892865e-06,
      "loss": 0.6125,
      "step": 5420
    },
    {
      "epoch": 2.1310799872452106,
      "grad_norm": 2.0526253778498154,
      "learning_rate": 4.087235750256469e-06,
      "loss": 0.666,
      "step": 5430
    },
    {
      "epoch": 2.1350045377615343,
      "grad_norm": 1.8656396637088302,
      "learning_rate": 4.053115790389159e-06,
      "loss": 0.6394,
      "step": 5440
    },
    {
      "epoch": 2.138929088277858,
      "grad_norm": 1.209266875293654,
      "learning_rate": 4.019102608923262e-06,
      "loss": 0.6132,
      "step": 5450
    },
    {
      "epoch": 2.1428536387941817,
      "grad_norm": 1.1832551572175067,
      "learning_rate": 3.985196816577433e-06,
      "loss": 0.5475,
      "step": 5460
    },
    {
      "epoch": 2.1467781893105053,
      "grad_norm": 2.2143801889042276,
      "learning_rate": 3.951399022142127e-06,
      "loss": 0.608,
      "step": 5470
    },
    {
      "epoch": 2.150702739826829,
      "grad_norm": 2.128346152883008,
      "learning_rate": 3.917709832468641e-06,
      "loss": 0.6848,
      "step": 5480
    },
    {
      "epoch": 2.1546272903431527,
      "grad_norm": 1.6896984103464467,
      "learning_rate": 3.884129852458253e-06,
      "loss": 0.6284,
      "step": 5490
    },
    {
      "epoch": 2.1585518408594764,
      "grad_norm": 1.222694494025062,
      "learning_rate": 3.850659685051336e-06,
      "loss": 0.5898,
      "step": 5500
    },
    {
      "epoch": 2.1624763913758,
      "grad_norm": 0.8066312626727323,
      "learning_rate": 3.817299931216537e-06,
      "loss": 0.546,
      "step": 5510
    },
    {
      "epoch": 2.1664009418921237,
      "grad_norm": 2.05281043982966,
      "learning_rate": 3.784051189939996e-06,
      "loss": 0.6217,
      "step": 5520
    },
    {
      "epoch": 2.1703254924084474,
      "grad_norm": 2.0329973993617556,
      "learning_rate": 3.7509140582145707e-06,
      "loss": 0.6679,
      "step": 5530
    },
    {
      "epoch": 2.174250042924771,
      "grad_norm": 1.7930592907009166,
      "learning_rate": 3.7178891310291444e-06,
      "loss": 0.6302,
      "step": 5540
    },
    {
      "epoch": 2.1781745934410948,
      "grad_norm": 1.3304481186905044,
      "learning_rate": 3.6849770013579135e-06,
      "loss": 0.5972,
      "step": 5550
    },
    {
      "epoch": 2.1820991439574184,
      "grad_norm": 0.8862316368226267,
      "learning_rate": 3.652178260149768e-06,
      "loss": 0.5508,
      "step": 5560
    },
    {
      "epoch": 2.186023694473742,
      "grad_norm": 2.16438233599458,
      "learning_rate": 3.619493496317662e-06,
      "loss": 0.6113,
      "step": 5570
    },
    {
      "epoch": 2.189948244990066,
      "grad_norm": 2.097362156115183,
      "learning_rate": 3.5869232967280466e-06,
      "loss": 0.678,
      "step": 5580
    },
    {
      "epoch": 2.1938727955063895,
      "grad_norm": 1.7293983928389607,
      "learning_rate": 3.554468246190337e-06,
      "loss": 0.6255,
      "step": 5590
    },
    {
      "epoch": 2.197797346022713,
      "grad_norm": 1.2526910240559423,
      "learning_rate": 3.522128927446392e-06,
      "loss": 0.6191,
      "step": 5600
    },
    {
      "epoch": 2.201721896539037,
      "grad_norm": 0.779154938656075,
      "learning_rate": 3.489905921160083e-06,
      "loss": 0.5403,
      "step": 5610
    },
    {
      "epoch": 2.2056464470553605,
      "grad_norm": 2.0546436977432094,
      "learning_rate": 3.4577998059068354e-06,
      "loss": 0.6159,
      "step": 5620
    },
    {
      "epoch": 2.209570997571684,
      "grad_norm": 2.141525722727545,
      "learning_rate": 3.4258111581632634e-06,
      "loss": 0.6876,
      "step": 5630
    },
    {
      "epoch": 2.213495548088008,
      "grad_norm": 1.7486151559652525,
      "learning_rate": 3.3939405522968105e-06,
      "loss": 0.6232,
      "step": 5640
    },
    {
      "epoch": 2.2174200986043315,
      "grad_norm": 1.2463505933135222,
      "learning_rate": 3.362188560555434e-06,
      "loss": 0.603,
      "step": 5650
    },
    {
      "epoch": 2.221344649120655,
      "grad_norm": 0.7557156986288721,
      "learning_rate": 3.3305557530573363e-06,
      "loss": 0.5734,
      "step": 5660
    },
    {
      "epoch": 2.225269199636979,
      "grad_norm": 2.6186065252503994,
      "learning_rate": 3.2990426977807156e-06,
      "loss": 0.6169,
      "step": 5670
    },
    {
      "epoch": 2.2291937501533026,
      "grad_norm": 2.1713884725106314,
      "learning_rate": 3.2676499605535918e-06,
      "loss": 0.6557,
      "step": 5680
    },
    {
      "epoch": 2.2331183006696262,
      "grad_norm": 1.8496259928018195,
      "learning_rate": 3.2363781050436105e-06,
      "loss": 0.6224,
      "step": 5690
    },
    {
      "epoch": 2.23704285118595,
      "grad_norm": 1.3375119127996462,
      "learning_rate": 3.2052276927479677e-06,
      "loss": 0.6029,
      "step": 5700
    },
    {
      "epoch": 2.2409674017022736,
      "grad_norm": 0.847985655236947,
      "learning_rate": 3.1741992829832924e-06,
      "loss": 0.5552,
      "step": 5710
    },
    {
      "epoch": 2.2448919522185973,
      "grad_norm": 2.1771913040167212,
      "learning_rate": 3.143293432875607e-06,
      "loss": 0.6089,
      "step": 5720
    },
    {
      "epoch": 2.248816502734921,
      "grad_norm": 2.0426822342569837,
      "learning_rate": 3.112510697350348e-06,
      "loss": 0.6927,
      "step": 5730
    },
    {
      "epoch": 2.2527410532512446,
      "grad_norm": 1.8247017082924184,
      "learning_rate": 3.081851629122372e-06,
      "loss": 0.6389,
      "step": 5740
    },
    {
      "epoch": 2.2566656037675683,
      "grad_norm": 1.3403447493161933,
      "learning_rate": 3.051316778686055e-06,
      "loss": 0.5947,
      "step": 5750
    },
    {
      "epoch": 2.260590154283892,
      "grad_norm": 0.8089857773862223,
      "learning_rate": 3.0209066943053944e-06,
      "loss": 0.5622,
      "step": 5760
    },
    {
      "epoch": 2.2645147048002157,
      "grad_norm": 2.3577656985489477,
      "learning_rate": 2.990621922004172e-06,
      "loss": 0.5892,
      "step": 5770
    },
    {
      "epoch": 2.2684392553165393,
      "grad_norm": 2.1740666869738323,
      "learning_rate": 2.960463005556149e-06,
      "loss": 0.672,
      "step": 5780
    },
    {
      "epoch": 2.272363805832863,
      "grad_norm": 1.829981750443929,
      "learning_rate": 2.9304304864752886e-06,
      "loss": 0.6373,
      "step": 5790
    },
    {
      "epoch": 2.2762883563491867,
      "grad_norm": 1.3281957095626744,
      "learning_rate": 2.900524904006061e-06,
      "loss": 0.5975,
      "step": 5800
    },
    {
      "epoch": 2.2802129068655104,
      "grad_norm": 0.7087417039119808,
      "learning_rate": 2.87074679511373e-06,
      "loss": 0.5296,
      "step": 5810
    },
    {
      "epoch": 2.284137457381834,
      "grad_norm": 2.109191274545063,
      "learning_rate": 2.8410966944747377e-06,
      "loss": 0.5962,
      "step": 5820
    },
    {
      "epoch": 2.2880620078981577,
      "grad_norm": 2.241584819214679,
      "learning_rate": 2.8115751344670863e-06,
      "loss": 0.6636,
      "step": 5830
    },
    {
      "epoch": 2.2919865584144814,
      "grad_norm": 1.7605810701006008,
      "learning_rate": 2.782182645160789e-06,
      "loss": 0.6265,
      "step": 5840
    },
    {
      "epoch": 2.295911108930805,
      "grad_norm": 1.2836328256236162,
      "learning_rate": 2.7529197543083507e-06,
      "loss": 0.5931,
      "step": 5850
    },
    {
      "epoch": 2.2998356594471288,
      "grad_norm": 0.9519727219083821,
      "learning_rate": 2.7237869873352827e-06,
      "loss": 0.5509,
      "step": 5860
    },
    {
      "epoch": 2.3037602099634524,
      "grad_norm": 2.1895645275891704,
      "learning_rate": 2.6947848673306853e-06,
      "loss": 0.6199,
      "step": 5870
    },
    {
      "epoch": 2.307684760479776,
      "grad_norm": 2.0598817109009904,
      "learning_rate": 2.6659139150378377e-06,
      "loss": 0.6591,
      "step": 5880
    },
    {
      "epoch": 2.3116093109961,
      "grad_norm": 1.8143316778974414,
      "learning_rate": 2.6371746488448614e-06,
      "loss": 0.6347,
      "step": 5890
    },
    {
      "epoch": 2.3155338615124235,
      "grad_norm": 1.3541045070964877,
      "learning_rate": 2.6085675847754155e-06,
      "loss": 0.586,
      "step": 5900
    },
    {
      "epoch": 2.319458412028747,
      "grad_norm": 0.7923721169078987,
      "learning_rate": 2.5800932364794064e-06,
      "loss": 0.5212,
      "step": 5910
    },
    {
      "epoch": 2.323382962545071,
      "grad_norm": 2.794475468134139,
      "learning_rate": 2.5517521152237966e-06,
      "loss": 0.5974,
      "step": 5920
    },
    {
      "epoch": 2.3273075130613945,
      "grad_norm": 2.2143050890712255,
      "learning_rate": 2.5235447298834003e-06,
      "loss": 0.6684,
      "step": 5930
    },
    {
      "epoch": 2.331232063577718,
      "grad_norm": 1.7900709742899215,
      "learning_rate": 2.49547158693176e-06,
      "loss": 0.6278,
      "step": 5940
    },
    {
      "epoch": 2.335156614094042,
      "grad_norm": 1.271494366484403,
      "learning_rate": 2.4675331904320533e-06,
      "loss": 0.5929,
      "step": 5950
    },
    {
      "epoch": 2.3390811646103655,
      "grad_norm": 0.8556427431737861,
      "learning_rate": 2.43973004202803e-06,
      "loss": 0.5524,
      "step": 5960
    },
    {
      "epoch": 2.343005715126689,
      "grad_norm": 2.117180082843371,
      "learning_rate": 2.412062640935021e-06,
      "loss": 0.6013,
      "step": 5970
    },
    {
      "epoch": 2.346930265643013,
      "grad_norm": 2.1484003642018457,
      "learning_rate": 2.3845314839309563e-06,
      "loss": 0.6632,
      "step": 5980
    },
    {
      "epoch": 2.3508548161593366,
      "grad_norm": 1.8030873162593484,
      "learning_rate": 2.3571370653474656e-06,
      "loss": 0.6168,
      "step": 5990
    },
    {
      "epoch": 2.3547793666756602,
      "grad_norm": 1.2601737288351715,
      "learning_rate": 2.329879877060981e-06,
      "loss": 0.5886,
      "step": 6000
    },
    {
      "epoch": 2.358703917191984,
      "grad_norm": 0.9165582307879456,
      "learning_rate": 2.302760408483926e-06,
      "loss": 0.5428,
      "step": 6010
    },
    {
      "epoch": 2.3626284677083076,
      "grad_norm": 2.0984461367902605,
      "learning_rate": 2.275779146555915e-06,
      "loss": 0.6007,
      "step": 6020
    },
    {
      "epoch": 2.3665530182246313,
      "grad_norm": 2.1846524601461894,
      "learning_rate": 2.2489365757350132e-06,
      "loss": 0.664,
      "step": 6030
    },
    {
      "epoch": 2.370477568740955,
      "grad_norm": 1.7647491805175937,
      "learning_rate": 2.2222331779890393e-06,
      "loss": 0.6257,
      "step": 6040
    },
    {
      "epoch": 2.3744021192572786,
      "grad_norm": 1.3444624902761966,
      "learning_rate": 2.1956694327869043e-06,
      "loss": 0.6041,
      "step": 6050
    },
    {
      "epoch": 2.3783266697736023,
      "grad_norm": 0.8924579097538136,
      "learning_rate": 2.16924581709002e-06,
      "loss": 0.5369,
      "step": 6060
    },
    {
      "epoch": 2.382251220289926,
      "grad_norm": 2.1816842093278526,
      "learning_rate": 2.142962805343708e-06,
      "loss": 0.5806,
      "step": 6070
    },
    {
      "epoch": 2.3861757708062497,
      "grad_norm": 2.1315338386325138,
      "learning_rate": 2.1168208694687108e-06,
      "loss": 0.6934,
      "step": 6080
    },
    {
      "epoch": 2.3901003213225733,
      "grad_norm": 1.8401419512172745,
      "learning_rate": 2.0908204788526965e-06,
      "loss": 0.6473,
      "step": 6090
    },
    {
      "epoch": 2.394024871838897,
      "grad_norm": 1.256508614588367,
      "learning_rate": 2.064962100341842e-06,
      "loss": 0.6,
      "step": 6100
    },
    {
      "epoch": 2.3979494223552207,
      "grad_norm": 0.7725755529070791,
      "learning_rate": 2.039246198232446e-06,
      "loss": 0.5488,
      "step": 6110
    },
    {
      "epoch": 2.4018739728715444,
      "grad_norm": 2.1092659651397474,
      "learning_rate": 2.0136732342625874e-06,
      "loss": 0.5748,
      "step": 6120
    },
    {
      "epoch": 2.405798523387868,
      "grad_norm": 2.202730059861675,
      "learning_rate": 1.9882436676038477e-06,
      "loss": 0.6778,
      "step": 6130
    },
    {
      "epoch": 2.4097230739041917,
      "grad_norm": 1.7146488474319233,
      "learning_rate": 1.962957954853055e-06,
      "loss": 0.642,
      "step": 6140
    },
    {
      "epoch": 2.4136476244205154,
      "grad_norm": 1.2875382567695426,
      "learning_rate": 1.9378165500240943e-06,
      "loss": 0.5935,
      "step": 6150
    },
    {
      "epoch": 2.417572174936839,
      "grad_norm": 0.8159718677862676,
      "learning_rate": 1.912819904539749e-06,
      "loss": 0.556,
      "step": 6160
    },
    {
      "epoch": 2.4214967254531627,
      "grad_norm": 2.147420171464766,
      "learning_rate": 1.887968467223591e-06,
      "loss": 0.6084,
      "step": 6170
    },
    {
      "epoch": 2.4254212759694864,
      "grad_norm": 2.352418026966586,
      "learning_rate": 1.8632626842919398e-06,
      "loss": 0.6647,
      "step": 6180
    },
    {
      "epoch": 2.42934582648581,
      "grad_norm": 1.843956419568753,
      "learning_rate": 1.8387029993458273e-06,
      "loss": 0.6224,
      "step": 6190
    },
    {
      "epoch": 2.4332703770021338,
      "grad_norm": 1.215534258708622,
      "learning_rate": 1.8142898533630536e-06,
      "loss": 0.6116,
      "step": 6200
    },
    {
      "epoch": 2.4371949275184575,
      "grad_norm": 0.7790773570638417,
      "learning_rate": 1.7900236846902575e-06,
      "loss": 0.5395,
      "step": 6210
    },
    {
      "epoch": 2.441119478034781,
      "grad_norm": 2.274459819686287,
      "learning_rate": 1.765904929035046e-06,
      "loss": 0.6089,
      "step": 6220
    },
    {
      "epoch": 2.445044028551105,
      "grad_norm": 2.156030969614435,
      "learning_rate": 1.7419340194581803e-06,
      "loss": 0.6517,
      "step": 6230
    },
    {
      "epoch": 2.4489685790674285,
      "grad_norm": 1.8023260247461752,
      "learning_rate": 1.7181113863657805e-06,
      "loss": 0.6312,
      "step": 6240
    },
    {
      "epoch": 2.452893129583752,
      "grad_norm": 1.6128467278404588,
      "learning_rate": 1.6944374575016253e-06,
      "loss": 0.6097,
      "step": 6250
    },
    {
      "epoch": 2.456817680100076,
      "grad_norm": 0.7816281842546718,
      "learning_rate": 1.670912657939443e-06,
      "loss": 0.5411,
      "step": 6260
    },
    {
      "epoch": 2.4607422306163995,
      "grad_norm": 2.2598007072156028,
      "learning_rate": 1.6475374100753017e-06,
      "loss": 0.6139,
      "step": 6270
    },
    {
      "epoch": 2.464666781132723,
      "grad_norm": 2.1326684182909936,
      "learning_rate": 1.624312133620013e-06,
      "loss": 0.6849,
      "step": 6280
    },
    {
      "epoch": 2.468591331649047,
      "grad_norm": 1.793136525038962,
      "learning_rate": 1.6012372455915993e-06,
      "loss": 0.6165,
      "step": 6290
    },
    {
      "epoch": 2.4725158821653705,
      "grad_norm": 1.2436395590561673,
      "learning_rate": 1.5783131603078083e-06,
      "loss": 0.5958,
      "step": 6300
    },
    {
      "epoch": 2.4764404326816942,
      "grad_norm": 0.8178284156434679,
      "learning_rate": 1.555540289378663e-06,
      "loss": 0.542,
      "step": 6310
    },
    {
      "epoch": 2.480364983198018,
      "grad_norm": 2.219348641209819,
      "learning_rate": 1.532919041699089e-06,
      "loss": 0.6146,
      "step": 6320
    },
    {
      "epoch": 2.4842895337143416,
      "grad_norm": 2.1421139951368375,
      "learning_rate": 1.510449823441561e-06,
      "loss": 0.669,
      "step": 6330
    },
    {
      "epoch": 2.4882140842306653,
      "grad_norm": 1.7736069610642504,
      "learning_rate": 1.4881330380488014e-06,
      "loss": 0.6325,
      "step": 6340
    },
    {
      "epoch": 2.492138634746989,
      "grad_norm": 1.2940938084497826,
      "learning_rate": 1.4659690862265675e-06,
      "loss": 0.5918,
      "step": 6350
    },
    {
      "epoch": 2.4960631852633126,
      "grad_norm": 0.7888355624944882,
      "learning_rate": 1.4439583659364154e-06,
      "loss": 0.5432,
      "step": 6360
    },
    {
      "epoch": 2.4999877357796363,
      "grad_norm": 2.2684459134167003,
      "learning_rate": 1.4221012723885874e-06,
      "loss": 0.6068,
      "step": 6370
    },
    {
      "epoch": 2.50391228629596,
      "grad_norm": 2.303631890847662,
      "learning_rate": 1.400398198034897e-06,
      "loss": 0.6815,
      "step": 6380
    },
    {
      "epoch": 2.5078368368122836,
      "grad_norm": 1.725928594909339,
      "learning_rate": 1.3788495325616912e-06,
      "loss": 0.629,
      "step": 6390
    },
    {
      "epoch": 2.5117613873286073,
      "grad_norm": 1.2724712580718072,
      "learning_rate": 1.357455662882855e-06,
      "loss": 0.5858,
      "step": 6400
    },
    {
      "epoch": 2.515685937844931,
      "grad_norm": 0.8196671527768702,
      "learning_rate": 1.3362169731328534e-06,
      "loss": 0.543,
      "step": 6410
    },
    {
      "epoch": 2.5196104883612547,
      "grad_norm": 2.2121857648933614,
      "learning_rate": 1.3151338446598483e-06,
      "loss": 0.5918,
      "step": 6420
    },
    {
      "epoch": 2.5235350388775784,
      "grad_norm": 2.07329199306329,
      "learning_rate": 1.2942066560188349e-06,
      "loss": 0.65,
      "step": 6430
    },
    {
      "epoch": 2.527459589393902,
      "grad_norm": 1.7691554410042842,
      "learning_rate": 1.2734357829648624e-06,
      "loss": 0.6245,
      "step": 6440
    },
    {
      "epoch": 2.5313841399102257,
      "grad_norm": 1.2397356761301885,
      "learning_rate": 1.2528215984462766e-06,
      "loss": 0.5757,
      "step": 6450
    },
    {
      "epoch": 2.5353086904265494,
      "grad_norm": 0.8552163717265764,
      "learning_rate": 1.23236447259802e-06,
      "loss": 0.5636,
      "step": 6460
    },
    {
      "epoch": 2.539233240942873,
      "grad_norm": 2.019030277835733,
      "learning_rate": 1.2120647727349977e-06,
      "loss": 0.5962,
      "step": 6470
    },
    {
      "epoch": 2.5431577914591967,
      "grad_norm": 2.1298373917453928,
      "learning_rate": 1.1919228633454738e-06,
      "loss": 0.6936,
      "step": 6480
    },
    {
      "epoch": 2.5470823419755204,
      "grad_norm": 1.8011050211359714,
      "learning_rate": 1.1719391060845298e-06,
      "loss": 0.6272,
      "step": 6490
    },
    {
      "epoch": 2.551006892491844,
      "grad_norm": 1.221657027890025,
      "learning_rate": 1.152113859767565e-06,
      "loss": 0.6286,
      "step": 6500
    },
    {
      "epoch": 2.5549314430081678,
      "grad_norm": 0.8061467765642856,
      "learning_rate": 1.1324474803638653e-06,
      "loss": 0.5501,
      "step": 6510
    },
    {
      "epoch": 2.5588559935244914,
      "grad_norm": 2.051549539190188,
      "learning_rate": 1.1129403209902034e-06,
      "loss": 0.6067,
      "step": 6520
    },
    {
      "epoch": 2.562780544040815,
      "grad_norm": 2.3142399536514073,
      "learning_rate": 1.0935927319044959e-06,
      "loss": 0.6484,
      "step": 6530
    },
    {
      "epoch": 2.566705094557139,
      "grad_norm": 1.7767173325015122,
      "learning_rate": 1.0744050604995237e-06,
      "loss": 0.6047,
      "step": 6540
    },
    {
      "epoch": 2.5706296450734625,
      "grad_norm": 1.3099949970790892,
      "learning_rate": 1.0553776512966886e-06,
      "loss": 0.5826,
      "step": 6550
    },
    {
      "epoch": 2.574554195589786,
      "grad_norm": 0.8416372830973359,
      "learning_rate": 1.0365108459398277e-06,
      "loss": 0.5418,
      "step": 6560
    },
    {
      "epoch": 2.57847874610611,
      "grad_norm": 2.127337984518528,
      "learning_rate": 1.0178049831890768e-06,
      "loss": 0.6093,
      "step": 6570
    },
    {
      "epoch": 2.5824032966224335,
      "grad_norm": 2.252528047531405,
      "learning_rate": 9.992603989147941e-07,
      "loss": 0.6867,
      "step": 6580
    },
    {
      "epoch": 2.586327847138757,
      "grad_norm": 1.7132261228204282,
      "learning_rate": 9.808774260915243e-07,
      "loss": 0.6564,
      "step": 6590
    },
    {
      "epoch": 2.590252397655081,
      "grad_norm": 1.2934365792962297,
      "learning_rate": 9.626563947920231e-07,
      "loss": 0.5691,
      "step": 6600
    },
    {
      "epoch": 2.5941769481714045,
      "grad_norm": 0.8786324009918022,
      "learning_rate": 9.445976321813277e-07,
      "loss": 0.5383,
      "step": 6610
    },
    {
      "epoch": 2.598101498687728,
      "grad_norm": 2.2501075917295212,
      "learning_rate": 9.267014625108806e-07,
      "loss": 0.5817,
      "step": 6620
    },
    {
      "epoch": 2.602026049204052,
      "grad_norm": 2.2174253075426167,
      "learning_rate": 9.089682071127171e-07,
      "loss": 0.6744,
      "step": 6630
    },
    {
      "epoch": 2.6059505997203756,
      "grad_norm": 1.7616659355193223,
      "learning_rate": 8.91398184393687e-07,
      "loss": 0.6359,
      "step": 6640
    },
    {
      "epoch": 2.6098751502366992,
      "grad_norm": 1.3089314769046627,
      "learning_rate": 8.739917098297357e-07,
      "loss": 0.6045,
      "step": 6650
    },
    {
      "epoch": 2.613799700753023,
      "grad_norm": 0.7444796805297569,
      "learning_rate": 8.567490959602509e-07,
      "loss": 0.5295,
      "step": 6660
    },
    {
      "epoch": 2.6177242512693466,
      "grad_norm": 2.2630343377076287,
      "learning_rate": 8.396706523824372e-07,
      "loss": 0.6244,
      "step": 6670
    },
    {
      "epoch": 2.6216488017856703,
      "grad_norm": 2.0443361827637836,
      "learning_rate": 8.227566857457702e-07,
      "loss": 0.6894,
      "step": 6680
    },
    {
      "epoch": 2.625573352301994,
      "grad_norm": 1.7688362533725879,
      "learning_rate": 8.060074997464773e-07,
      "loss": 0.6192,
      "step": 6690
    },
    {
      "epoch": 2.6294979028183176,
      "grad_norm": 1.2229083921293311,
      "learning_rate": 7.894233951220953e-07,
      "loss": 0.5856,
      "step": 6700
    },
    {
      "epoch": 2.6334224533346413,
      "grad_norm": 0.76533431687556,
      "learning_rate": 7.730046696460691e-07,
      "loss": 0.53,
      "step": 6710
    },
    {
      "epoch": 2.637347003850965,
      "grad_norm": 2.264231653926727,
      "learning_rate": 7.567516181223966e-07,
      "loss": 0.5991,
      "step": 6720
    },
    {
      "epoch": 2.6412715543672887,
      "grad_norm": 2.1991521419444267,
      "learning_rate": 7.406645323803463e-07,
      "loss": 0.6315,
      "step": 6730
    },
    {
      "epoch": 2.6451961048836123,
      "grad_norm": 1.7239735369036862,
      "learning_rate": 7.247437012692104e-07,
      "loss": 0.6427,
      "step": 6740
    },
    {
      "epoch": 2.649120655399936,
      "grad_norm": 1.2790232035397695,
      "learning_rate": 7.089894106531214e-07,
      "loss": 0.594,
      "step": 6750
    },
    {
      "epoch": 2.6530452059162597,
      "grad_norm": 0.955171760712246,
      "learning_rate": 6.934019434059213e-07,
      "loss": 0.5533,
      "step": 6760
    },
    {
      "epoch": 2.6569697564325834,
      "grad_norm": 2.1955164062676644,
      "learning_rate": 6.779815794060718e-07,
      "loss": 0.5936,
      "step": 6770
    },
    {
      "epoch": 2.660894306948907,
      "grad_norm": 2.2404548484277664,
      "learning_rate": 6.627285955316476e-07,
      "loss": 0.6513,
      "step": 6780
    },
    {
      "epoch": 2.6648188574652307,
      "grad_norm": 1.8828883729369554,
      "learning_rate": 6.476432656553411e-07,
      "loss": 0.6286,
      "step": 6790
    },
    {
      "epoch": 2.6687434079815544,
      "grad_norm": 1.2443150219704873,
      "learning_rate": 6.327258606395736e-07,
      "loss": 0.5939,
      "step": 6800
    },
    {
      "epoch": 2.672667958497878,
      "grad_norm": 0.9014377738542545,
      "learning_rate": 6.179766483316041e-07,
      "loss": 0.5334,
      "step": 6810
    },
    {
      "epoch": 2.6765925090142018,
      "grad_norm": 2.183823820105367,
      "learning_rate": 6.03395893558737e-07,
      "loss": 0.5913,
      "step": 6820
    },
    {
      "epoch": 2.6805170595305254,
      "grad_norm": 2.1912679703089313,
      "learning_rate": 5.889838581235641e-07,
      "loss": 0.6719,
      "step": 6830
    },
    {
      "epoch": 2.684441610046849,
      "grad_norm": 1.7923155245248856,
      "learning_rate": 5.747408007992572e-07,
      "loss": 0.6208,
      "step": 6840
    },
    {
      "epoch": 2.688366160563173,
      "grad_norm": 1.2494561675837905,
      "learning_rate": 5.606669773249296e-07,
      "loss": 0.596,
      "step": 6850
    },
    {
      "epoch": 2.6922907110794965,
      "grad_norm": 0.907833148614821,
      "learning_rate": 5.467626404010407e-07,
      "loss": 0.5372,
      "step": 6860
    },
    {
      "epoch": 2.69621526159582,
      "grad_norm": 2.2934267800117856,
      "learning_rate": 5.330280396848619e-07,
      "loss": 0.609,
      "step": 6870
    },
    {
      "epoch": 2.700139812112144,
      "grad_norm": 2.114541613337409,
      "learning_rate": 5.194634217859851e-07,
      "loss": 0.6611,
      "step": 6880
    },
    {
      "epoch": 2.7040643626284675,
      "grad_norm": 1.754501104961409,
      "learning_rate": 5.060690302619053e-07,
      "loss": 0.6157,
      "step": 6890
    },
    {
      "epoch": 2.707988913144791,
      "grad_norm": 1.364912899052101,
      "learning_rate": 4.92845105613644e-07,
      "loss": 0.5929,
      "step": 6900
    },
    {
      "epoch": 2.711913463661115,
      "grad_norm": 0.7372286271827898,
      "learning_rate": 4.797918852814254e-07,
      "loss": 0.5314,
      "step": 6910
    },
    {
      "epoch": 2.7158380141774385,
      "grad_norm": 2.073475468780587,
      "learning_rate": 4.6690960364041973e-07,
      "loss": 0.5976,
      "step": 6920
    },
    {
      "epoch": 2.719762564693762,
      "grad_norm": 2.2358152929450963,
      "learning_rate": 4.5419849199653364e-07,
      "loss": 0.6623,
      "step": 6930
    },
    {
      "epoch": 2.723687115210086,
      "grad_norm": 1.815422184822613,
      "learning_rate": 4.416587785822568e-07,
      "loss": 0.6073,
      "step": 6940
    },
    {
      "epoch": 2.7276116657264096,
      "grad_norm": 1.350125241744432,
      "learning_rate": 4.2929068855256275e-07,
      "loss": 0.5984,
      "step": 6950
    },
    {
      "epoch": 2.7315362162427332,
      "grad_norm": 0.7223399634960578,
      "learning_rate": 4.170944439808622e-07,
      "loss": 0.5491,
      "step": 6960
    },
    {
      "epoch": 2.735460766759057,
      "grad_norm": 2.1836663497672455,
      "learning_rate": 4.0507026385502747e-07,
      "loss": 0.5941,
      "step": 6970
    },
    {
      "epoch": 2.7393853172753806,
      "grad_norm": 2.154640259731243,
      "learning_rate": 3.932183640734466e-07,
      "loss": 0.6781,
      "step": 6980
    },
    {
      "epoch": 2.7433098677917043,
      "grad_norm": 1.908154353492756,
      "learning_rate": 3.8153895744115767e-07,
      "loss": 0.6178,
      "step": 6990
    },
    {
      "epoch": 2.747234418308028,
      "grad_norm": 1.2339098485081115,
      "learning_rate": 3.700322536660228e-07,
      "loss": 0.5819,
      "step": 7000
    },
    {
      "epoch": 2.7511589688243516,
      "grad_norm": 0.9319742069385385,
      "learning_rate": 3.586984593549614e-07,
      "loss": 0.5296,
      "step": 7010
    },
    {
      "epoch": 2.7550835193406753,
      "grad_norm": 2.2839758727554424,
      "learning_rate": 3.475377780102451e-07,
      "loss": 0.5919,
      "step": 7020
    },
    {
      "epoch": 2.759008069856999,
      "grad_norm": 2.184552450870539,
      "learning_rate": 3.365504100258399e-07,
      "loss": 0.6341,
      "step": 7030
    },
    {
      "epoch": 2.7629326203733227,
      "grad_norm": 1.7194768639357731,
      "learning_rate": 3.2573655268380746e-07,
      "loss": 0.6252,
      "step": 7040
    },
    {
      "epoch": 2.7668571708896463,
      "grad_norm": 1.3255796782916087,
      "learning_rate": 3.1509640015076946e-07,
      "loss": 0.5879,
      "step": 7050
    },
    {
      "epoch": 2.77078172140597,
      "grad_norm": 0.8088254429550722,
      "learning_rate": 3.0463014347441255e-07,
      "loss": 0.5519,
      "step": 7060
    },
    {
      "epoch": 2.7747062719222937,
      "grad_norm": 2.202949542374798,
      "learning_rate": 2.9433797058006195e-07,
      "loss": 0.598,
      "step": 7070
    },
    {
      "epoch": 2.7786308224386174,
      "grad_norm": 2.192327750139598,
      "learning_rate": 2.842200662673111e-07,
      "loss": 0.6815,
      "step": 7080
    },
    {
      "epoch": 2.782555372954941,
      "grad_norm": 1.7866089200056756,
      "learning_rate": 2.7427661220669535e-07,
      "loss": 0.603,
      "step": 7090
    },
    {
      "epoch": 2.7864799234712647,
      "grad_norm": 1.2795477884768611,
      "learning_rate": 2.645077869364354e-07,
      "loss": 0.5773,
      "step": 7100
    },
    {
      "epoch": 2.7904044739875884,
      "grad_norm": 0.7183394998611272,
      "learning_rate": 2.5491376585923265e-07,
      "loss": 0.5313,
      "step": 7110
    },
    {
      "epoch": 2.794329024503912,
      "grad_norm": 2.308178999869327,
      "learning_rate": 2.4549472123911564e-07,
      "loss": 0.5919,
      "step": 7120
    },
    {
      "epoch": 2.7982535750202358,
      "grad_norm": 2.2719235337571813,
      "learning_rate": 2.362508221983484e-07,
      "loss": 0.6633,
      "step": 7130
    },
    {
      "epoch": 2.8021781255365594,
      "grad_norm": 1.8358853033909601,
      "learning_rate": 2.2718223471439815e-07,
      "loss": 0.6084,
      "step": 7140
    },
    {
      "epoch": 2.806102676052883,
      "grad_norm": 1.3561609455618762,
      "learning_rate": 2.182891216169447e-07,
      "loss": 0.5946,
      "step": 7150
    },
    {
      "epoch": 2.810027226569207,
      "grad_norm": 0.801043920678816,
      "learning_rate": 2.0957164258497031e-07,
      "loss": 0.5394,
      "step": 7160
    },
    {
      "epoch": 2.8139517770855305,
      "grad_norm": 2.3157120536637685,
      "learning_rate": 2.0102995414387983e-07,
      "loss": 0.5861,
      "step": 7170
    },
    {
      "epoch": 2.817876327601854,
      "grad_norm": 2.0452627880092886,
      "learning_rate": 1.9266420966270182e-07,
      "loss": 0.6566,
      "step": 7180
    },
    {
      "epoch": 2.821800878118178,
      "grad_norm": 1.883254183997997,
      "learning_rate": 1.8447455935132418e-07,
      "loss": 0.6011,
      "step": 7190
    },
    {
      "epoch": 2.8257254286345015,
      "grad_norm": 1.277403091223367,
      "learning_rate": 1.764611502578051e-07,
      "loss": 0.573,
      "step": 7200
    },
    {
      "epoch": 2.829649979150825,
      "grad_norm": 0.8701579907748704,
      "learning_rate": 1.6862412626572845e-07,
      "loss": 0.5748,
      "step": 7210
    },
    {
      "epoch": 2.833574529667149,
      "grad_norm": 2.2395045060776115,
      "learning_rate": 1.6096362809162047e-07,
      "loss": 0.5897,
      "step": 7220
    },
    {
      "epoch": 2.8374990801834725,
      "grad_norm": 2.1921791669212864,
      "learning_rate": 1.5347979328242613e-07,
      "loss": 0.6472,
      "step": 7230
    },
    {
      "epoch": 2.841423630699796,
      "grad_norm": 1.7683013517405388,
      "learning_rate": 1.461727562130344e-07,
      "loss": 0.6176,
      "step": 7240
    },
    {
      "epoch": 2.84534818121612,
      "grad_norm": 1.2599894592930116,
      "learning_rate": 1.3904264808387246e-07,
      "loss": 0.583,
      "step": 7250
    },
    {
      "epoch": 2.8492727317324436,
      "grad_norm": 0.8602212034932214,
      "learning_rate": 1.320895969185454e-07,
      "loss": 0.5383,
      "step": 7260
    },
    {
      "epoch": 2.8531972822487672,
      "grad_norm": 2.01097097204698,
      "learning_rate": 1.2531372756153458e-07,
      "loss": 0.5882,
      "step": 7270
    },
    {
      "epoch": 2.857121832765091,
      "grad_norm": 2.29115443827529,
      "learning_rate": 1.1871516167596186e-07,
      "loss": 0.659,
      "step": 7280
    },
    {
      "epoch": 2.8610463832814146,
      "grad_norm": 1.798405094313227,
      "learning_rate": 1.1229401774140447e-07,
      "loss": 0.6425,
      "step": 7290
    },
    {
      "epoch": 2.8649709337977383,
      "grad_norm": 1.2675180226420815,
      "learning_rate": 1.0605041105176128e-07,
      "loss": 0.5757,
      "step": 7300
    },
    {
      "epoch": 2.868895484314062,
      "grad_norm": 0.778167866423488,
      "learning_rate": 9.998445371319332e-08,
      "loss": 0.5289,
      "step": 7310
    },
    {
      "epoch": 2.8728200348303856,
      "grad_norm": 2.1530650160440845,
      "learning_rate": 9.409625464210093e-08,
      "loss": 0.582,
      "step": 7320
    },
    {
      "epoch": 2.8767445853467093,
      "grad_norm": 2.039982400238999,
      "learning_rate": 8.83859195631731e-08,
      "loss": 0.6649,
      "step": 7330
    },
    {
      "epoch": 2.880669135863033,
      "grad_norm": 1.8192255183888113,
      "learning_rate": 8.285355100748904e-08,
      "loss": 0.6083,
      "step": 7340
    },
    {
      "epoch": 2.8845936863793566,
      "grad_norm": 1.3511200742987195,
      "learning_rate": 7.749924831067401e-08,
      "loss": 0.5947,
      "step": 7350
    },
    {
      "epoch": 2.8885182368956803,
      "grad_norm": 0.8078794663885482,
      "learning_rate": 7.232310761112082e-08,
      "loss": 0.5189,
      "step": 7360
    },
    {
      "epoch": 2.892442787412004,
      "grad_norm": 2.086447914981744,
      "learning_rate": 6.732522184825896e-08,
      "loss": 0.5929,
      "step": 7370
    },
    {
      "epoch": 2.8963673379283277,
      "grad_norm": 2.2301352619297097,
      "learning_rate": 6.250568076088814e-08,
      "loss": 0.6598,
      "step": 7380
    },
    {
      "epoch": 2.9002918884446514,
      "grad_norm": 1.8389039532268878,
      "learning_rate": 5.7864570885567405e-08,
      "loss": 0.6154,
      "step": 7390
    },
    {
      "epoch": 2.904216438960975,
      "grad_norm": 1.2792199660600578,
      "learning_rate": 5.340197555505966e-08,
      "loss": 0.6012,
      "step": 7400
    },
    {
      "epoch": 2.9081409894772987,
      "grad_norm": 0.7325366163854787,
      "learning_rate": 4.911797489683734e-08,
      "loss": 0.5215,
      "step": 7410
    },
    {
      "epoch": 2.9120655399936224,
      "grad_norm": 1.8968115785034971,
      "learning_rate": 4.5012645831640225e-08,
      "loss": 0.5973,
      "step": 7420
    },
    {
      "epoch": 2.915990090509946,
      "grad_norm": 2.115373675723123,
      "learning_rate": 4.108606207209875e-08,
      "loss": 0.6792,
      "step": 7430
    },
    {
      "epoch": 2.9199146410262697,
      "grad_norm": 1.8029011243698139,
      "learning_rate": 3.7338294121407324e-08,
      "loss": 0.62,
      "step": 7440
    },
    {
      "epoch": 2.9238391915425934,
      "grad_norm": 1.288742701315245,
      "learning_rate": 3.376940927206196e-08,
      "loss": 0.5808,
      "step": 7450
    },
    {
      "epoch": 2.927763742058917,
      "grad_norm": 0.798534719587082,
      "learning_rate": 3.037947160464572e-08,
      "loss": 0.5186,
      "step": 7460
    },
    {
      "epoch": 2.9316882925752408,
      "grad_norm": 2.5141632319612954,
      "learning_rate": 2.716854198668517e-08,
      "loss": 0.577,
      "step": 7470
    },
    {
      "epoch": 2.9356128430915645,
      "grad_norm": 2.35827100419327,
      "learning_rate": 2.41366780715524e-08,
      "loss": 0.6523,
      "step": 7480
    },
    {
      "epoch": 2.939537393607888,
      "grad_norm": 1.8102877923296363,
      "learning_rate": 2.1283934297432472e-08,
      "loss": 0.6021,
      "step": 7490
    },
    {
      "epoch": 2.943461944124212,
      "grad_norm": 1.3570094551959164,
      "learning_rate": 1.861036188634424e-08,
      "loss": 0.5859,
      "step": 7500
    },
    {
      "epoch": 2.9473864946405355,
      "grad_norm": 0.8112341118577404,
      "learning_rate": 1.6116008843224395e-08,
      "loss": 0.5446,
      "step": 7510
    },
    {
      "epoch": 2.951311045156859,
      "grad_norm": 2.0316666239859726,
      "learning_rate": 1.3800919955058167e-08,
      "loss": 0.5938,
      "step": 7520
    },
    {
      "epoch": 2.955235595673183,
      "grad_norm": 2.24783418646791,
      "learning_rate": 1.1665136790084408e-08,
      "loss": 0.6663,
      "step": 7530
    },
    {
      "epoch": 2.9591601461895065,
      "grad_norm": 1.7761857680646513,
      "learning_rate": 9.708697697040636e-09,
      "loss": 0.6199,
      "step": 7540
    },
    {
      "epoch": 2.96308469670583,
      "grad_norm": 1.2633693500353296,
      "learning_rate": 7.931637804481362e-09,
      "loss": 0.5774,
      "step": 7550
    },
    {
      "epoch": 2.967009247222154,
      "grad_norm": 0.8065989220041421,
      "learning_rate": 6.333989020143039e-09,
      "loss": 0.5492,
      "step": 7560
    },
    {
      "epoch": 2.9709337977384775,
      "grad_norm": 2.3130844134472537,
      "learning_rate": 4.915780030372297e-09,
      "loss": 0.5792,
      "step": 7570
    },
    {
      "epoch": 2.974858348254801,
      "grad_norm": 2.1600268864719765,
      "learning_rate": 3.6770362996108033e-09,
      "loss": 0.6643,
      "step": 7580
    },
    {
      "epoch": 2.978782898771125,
      "grad_norm": 1.78858714847675,
      "learning_rate": 2.617780069940068e-09,
      "loss": 0.6227,
      "step": 7590
    },
    {
      "epoch": 2.9827074492874486,
      "grad_norm": 1.2711261056787222,
      "learning_rate": 1.738030360677323e-09,
      "loss": 0.5816,
      "step": 7600
    },
    {
      "epoch": 2.9866319998037723,
      "grad_norm": 0.8400903589420956,
      "learning_rate": 1.0378029680391254e-09,
      "loss": 0.5403,
      "step": 7610
    },
    {
      "epoch": 2.990556550320096,
      "grad_norm": 2.242115052909609,
      "learning_rate": 5.171104648549196e-10,
      "loss": 0.6179,
      "step": 7620
    },
    {
      "epoch": 2.9944811008364196,
      "grad_norm": 2.174676570322404,
      "learning_rate": 1.759622003427719e-10,
      "loss": 0.658,
      "step": 7630
    },
    {
      "epoch": 2.9984056513527433,
      "grad_norm": 1.9961793276061766,
      "learning_rate": 1.436429993950661e-11,
      "loss": 0.5866,
      "step": 7640
    },
    {
      "epoch": 2.999975471559273,
      "eval_loss": 0.5420118570327759,
      "eval_runtime": 1735.1416,
      "eval_samples_per_second": 14.408,
      "eval_steps_per_second": 3.602,
      "step": 7644
    },
    {
      "epoch": 2.999975471559273,
      "step": 7644,
      "total_flos": 1249712503734272.0,
      "train_loss": 0.20236909012093338,
      "train_runtime": 73063.4732,
      "train_samples_per_second": 13.392,
      "train_steps_per_second": 0.105
    }
  ],
  "logging_steps": 10,
  "max_steps": 7644,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 3,
  "save_steps": 100,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": true
      },
      "attributes": {}
    }
  },
  "total_flos": 1249712503734272.0,
  "train_batch_size": 1,
  "trial_name": null,
  "trial_params": null
}