{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 2.0,
  "eval_steps": 500,
  "global_step": 500,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.004,
      "grad_norm": 2.7120276745865266,
      "learning_rate": 1.2500000000000002e-07,
      "loss": 0.9478,
      "step": 1
    },
    {
      "epoch": 0.008,
      "grad_norm": 2.5535110233770237,
      "learning_rate": 2.5000000000000004e-07,
      "loss": 0.962,
      "step": 2
    },
    {
      "epoch": 0.012,
      "grad_norm": 2.6421564999361533,
      "learning_rate": 3.75e-07,
      "loss": 0.9483,
      "step": 3
    },
    {
      "epoch": 0.016,
      "grad_norm": 2.5842713407482596,
      "learning_rate": 5.000000000000001e-07,
      "loss": 0.9484,
      "step": 4
    },
    {
      "epoch": 0.02,
      "grad_norm": 2.5281267639976615,
      "learning_rate": 6.25e-07,
      "loss": 0.9512,
      "step": 5
    },
    {
      "epoch": 0.024,
      "grad_norm": 2.55003561639048,
      "learning_rate": 7.5e-07,
      "loss": 0.9185,
      "step": 6
    },
    {
      "epoch": 0.028,
      "grad_norm": 2.5014092783190947,
      "learning_rate": 8.75e-07,
      "loss": 0.9411,
      "step": 7
    },
    {
      "epoch": 0.032,
      "grad_norm": 2.5247715994428046,
      "learning_rate": 1.0000000000000002e-06,
      "loss": 0.9474,
      "step": 8
    },
    {
      "epoch": 0.036,
      "grad_norm": 2.564090844776098,
      "learning_rate": 1.125e-06,
      "loss": 0.9246,
      "step": 9
    },
    {
      "epoch": 0.04,
      "grad_norm": 2.5258707026382154,
      "learning_rate": 1.25e-06,
      "loss": 0.9078,
      "step": 10
    },
    {
      "epoch": 0.044,
      "grad_norm": 2.5471156175078464,
      "learning_rate": 1.3750000000000002e-06,
      "loss": 0.9134,
      "step": 11
    },
    {
      "epoch": 0.048,
      "grad_norm": 2.7125301372042467,
      "learning_rate": 1.5e-06,
      "loss": 0.8776,
      "step": 12
    },
    {
      "epoch": 0.052,
      "grad_norm": 2.150392048325219,
      "learning_rate": 1.6250000000000001e-06,
      "loss": 0.8549,
      "step": 13
    },
    {
      "epoch": 0.056,
      "grad_norm": 1.8184980673856375,
      "learning_rate": 1.75e-06,
      "loss": 0.8169,
      "step": 14
    },
    {
      "epoch": 0.06,
      "grad_norm": 1.7529184087805771,
      "learning_rate": 1.8750000000000003e-06,
      "loss": 0.7975,
      "step": 15
    },
    {
      "epoch": 0.064,
      "grad_norm": 1.7474840928555682,
      "learning_rate": 2.0000000000000003e-06,
      "loss": 0.8204,
      "step": 16
    },
    {
      "epoch": 0.068,
      "grad_norm": 1.522251241150409,
      "learning_rate": 2.125e-06,
      "loss": 0.8301,
      "step": 17
    },
    {
      "epoch": 0.072,
      "grad_norm": 0.9782588646220091,
      "learning_rate": 2.25e-06,
      "loss": 0.7839,
      "step": 18
    },
    {
      "epoch": 0.076,
      "grad_norm": 0.975162832649703,
      "learning_rate": 2.375e-06,
      "loss": 0.7876,
      "step": 19
    },
    {
      "epoch": 0.08,
      "grad_norm": 0.8952616849532462,
      "learning_rate": 2.5e-06,
      "loss": 0.7589,
      "step": 20
    },
    {
      "epoch": 0.084,
      "grad_norm": 0.8469187411478404,
      "learning_rate": 2.6250000000000003e-06,
      "loss": 0.7567,
      "step": 21
    },
    {
      "epoch": 0.088,
      "grad_norm": 0.7683035597272257,
      "learning_rate": 2.7500000000000004e-06,
      "loss": 0.7598,
      "step": 22
    },
    {
      "epoch": 0.092,
      "grad_norm": 0.6779719620181055,
      "learning_rate": 2.875e-06,
      "loss": 0.7431,
      "step": 23
    },
    {
      "epoch": 0.096,
      "grad_norm": 0.5998651036445499,
      "learning_rate": 3e-06,
      "loss": 0.7348,
      "step": 24
    },
    {
      "epoch": 0.1,
      "grad_norm": 0.6798779251188986,
      "learning_rate": 3.125e-06,
      "loss": 0.7271,
      "step": 25
    },
    {
      "epoch": 0.104,
      "grad_norm": 0.7459028436934305,
      "learning_rate": 3.2500000000000002e-06,
      "loss": 0.7011,
      "step": 26
    },
    {
      "epoch": 0.108,
      "grad_norm": 0.7661666147630528,
      "learning_rate": 3.3750000000000003e-06,
      "loss": 0.7209,
      "step": 27
    },
    {
      "epoch": 0.112,
      "grad_norm": 0.7109730987249342,
      "learning_rate": 3.5e-06,
      "loss": 0.7118,
      "step": 28
    },
    {
      "epoch": 0.116,
      "grad_norm": 0.6778217985942258,
      "learning_rate": 3.625e-06,
      "loss": 0.7295,
      "step": 29
    },
    {
      "epoch": 0.12,
      "grad_norm": 0.6073429084137083,
      "learning_rate": 3.7500000000000005e-06,
      "loss": 0.7114,
      "step": 30
    },
    {
      "epoch": 0.124,
      "grad_norm": 0.5325413632521706,
      "learning_rate": 3.875e-06,
      "loss": 0.7117,
      "step": 31
    },
    {
      "epoch": 0.128,
      "grad_norm": 0.5170445125401208,
      "learning_rate": 4.000000000000001e-06,
      "loss": 0.7132,
      "step": 32
    },
    {
      "epoch": 0.132,
      "grad_norm": 0.47459002896590347,
      "learning_rate": 4.125e-06,
      "loss": 0.6913,
      "step": 33
    },
    {
      "epoch": 0.136,
      "grad_norm": 0.46611890152884594,
      "learning_rate": 4.25e-06,
      "loss": 0.6893,
      "step": 34
    },
    {
      "epoch": 0.14,
      "grad_norm": 0.5282002224942279,
      "learning_rate": 4.3750000000000005e-06,
      "loss": 0.7059,
      "step": 35
    },
    {
      "epoch": 0.144,
      "grad_norm": 0.516098581069478,
      "learning_rate": 4.5e-06,
      "loss": 0.7068,
      "step": 36
    },
    {
      "epoch": 0.148,
      "grad_norm": 0.4946613953020982,
      "learning_rate": 4.625000000000001e-06,
      "loss": 0.6854,
      "step": 37
    },
    {
      "epoch": 0.152,
      "grad_norm": 0.45539519616935026,
      "learning_rate": 4.75e-06,
      "loss": 0.689,
      "step": 38
    },
    {
      "epoch": 0.156,
      "grad_norm": 0.43098619176364883,
      "learning_rate": 4.875e-06,
      "loss": 0.6904,
      "step": 39
    },
    {
      "epoch": 0.16,
      "grad_norm": 0.42693567901762375,
      "learning_rate": 5e-06,
      "loss": 0.6713,
      "step": 40
    },
    {
      "epoch": 0.164,
      "grad_norm": 0.42916236211888303,
      "learning_rate": 4.999941696797974e-06,
      "loss": 0.6929,
      "step": 41
    },
    {
      "epoch": 0.168,
      "grad_norm": 0.46320958381108696,
      "learning_rate": 4.9997667899113055e-06,
      "loss": 0.6601,
      "step": 42
    },
    {
      "epoch": 0.172,
      "grad_norm": 0.4650598350234117,
      "learning_rate": 4.9994752874981e-06,
      "loss": 0.6627,
      "step": 43
    },
    {
      "epoch": 0.176,
      "grad_norm": 0.4899967026584299,
      "learning_rate": 4.999067203154777e-06,
      "loss": 0.6583,
      "step": 44
    },
    {
      "epoch": 0.18,
      "grad_norm": 0.4562804538758281,
      "learning_rate": 4.998542555915435e-06,
      "loss": 0.6712,
      "step": 45
    },
    {
      "epoch": 0.184,
      "grad_norm": 0.36547047335003885,
      "learning_rate": 4.997901370250966e-06,
      "loss": 0.6495,
      "step": 46
    },
    {
      "epoch": 0.188,
      "grad_norm": 0.37624066260466155,
      "learning_rate": 4.997143676067913e-06,
      "loss": 0.6703,
      "step": 47
    },
    {
      "epoch": 0.192,
      "grad_norm": 0.36653954523845933,
      "learning_rate": 4.99626950870707e-06,
      "loss": 0.6781,
      "step": 48
    },
    {
      "epoch": 0.196,
      "grad_norm": 0.3477765639450634,
      "learning_rate": 4.995278908941845e-06,
      "loss": 0.6745,
      "step": 49
    },
    {
      "epoch": 0.2,
      "grad_norm": 0.3342296714847872,
      "learning_rate": 4.994171922976349e-06,
      "loss": 0.6771,
      "step": 50
    },
    {
      "epoch": 0.204,
      "grad_norm": 0.37958792690250714,
      "learning_rate": 4.9929486024432405e-06,
      "loss": 0.6546,
      "step": 51
    },
    {
      "epoch": 0.208,
      "grad_norm": 0.40870653964939113,
      "learning_rate": 4.991609004401324e-06,
      "loss": 0.6528,
      "step": 52
    },
    {
      "epoch": 0.212,
      "grad_norm": 0.364192633459814,
      "learning_rate": 4.990153191332885e-06,
      "loss": 0.6458,
      "step": 53
    },
    {
      "epoch": 0.216,
      "grad_norm": 0.30668416089546924,
      "learning_rate": 4.988581231140772e-06,
      "loss": 0.672,
      "step": 54
    },
    {
      "epoch": 0.22,
      "grad_norm": 0.33255921507240616,
      "learning_rate": 4.986893197145238e-06,
      "loss": 0.6575,
      "step": 55
    },
    {
      "epoch": 0.224,
      "grad_norm": 0.312428536241603,
      "learning_rate": 4.985089168080509e-06,
      "loss": 0.6629,
      "step": 56
    },
    {
      "epoch": 0.228,
      "grad_norm": 0.2961748326875195,
      "learning_rate": 4.983169228091125e-06,
      "loss": 0.6514,
      "step": 57
    },
    {
      "epoch": 0.232,
      "grad_norm": 0.30513153025857964,
      "learning_rate": 4.981133466728004e-06,
      "loss": 0.6547,
      "step": 58
    },
    {
      "epoch": 0.236,
      "grad_norm": 0.3274000044829569,
      "learning_rate": 4.978981978944271e-06,
      "loss": 0.6514,
      "step": 59
    },
    {
      "epoch": 0.24,
      "grad_norm": 0.3083977406610579,
      "learning_rate": 4.976714865090827e-06,
      "loss": 0.6422,
      "step": 60
    },
    {
      "epoch": 0.244,
      "grad_norm": 0.3052292638265554,
      "learning_rate": 4.97433223091167e-06,
      "loss": 0.6355,
      "step": 61
    },
    {
      "epoch": 0.248,
      "grad_norm": 0.2959593176715387,
      "learning_rate": 4.971834187538963e-06,
      "loss": 0.6551,
      "step": 62
    },
    {
      "epoch": 0.252,
      "grad_norm": 0.29042322618949995,
      "learning_rate": 4.9692208514878445e-06,
      "loss": 0.6581,
      "step": 63
    },
    {
      "epoch": 0.256,
      "grad_norm": 0.3149031676109748,
      "learning_rate": 4.966492344651006e-06,
      "loss": 0.6594,
      "step": 64
    },
    {
      "epoch": 0.26,
      "grad_norm": 0.2874450280533471,
      "learning_rate": 4.963648794292992e-06,
      "loss": 0.6488,
      "step": 65
    },
    {
      "epoch": 0.264,
      "grad_norm": 0.27325192501172824,
      "learning_rate": 4.960690333044279e-06,
      "loss": 0.6351,
      "step": 66
    },
    {
      "epoch": 0.268,
      "grad_norm": 0.2817944445002008,
      "learning_rate": 4.957617098895076e-06,
      "loss": 0.6265,
      "step": 67
    },
    {
      "epoch": 0.272,
      "grad_norm": 0.281980621240097,
      "learning_rate": 4.954429235188897e-06,
      "loss": 0.6432,
      "step": 68
    },
    {
      "epoch": 0.276,
      "grad_norm": 0.2962526841715817,
      "learning_rate": 4.951126890615871e-06,
      "loss": 0.6354,
      "step": 69
    },
    {
      "epoch": 0.28,
      "grad_norm": 0.28005347365677186,
      "learning_rate": 4.947710219205808e-06,
      "loss": 0.6548,
      "step": 70
    },
    {
      "epoch": 0.284,
      "grad_norm": 0.26960325063378,
      "learning_rate": 4.944179380321015e-06,
      "loss": 0.6125,
      "step": 71
    },
    {
      "epoch": 0.288,
      "grad_norm": 0.2848666105192882,
      "learning_rate": 4.940534538648862e-06,
      "loss": 0.6078,
      "step": 72
    },
    {
      "epoch": 0.292,
      "grad_norm": 0.276353304899713,
      "learning_rate": 4.936775864194101e-06,
      "loss": 0.6524,
      "step": 73
    },
    {
      "epoch": 0.296,
      "grad_norm": 0.27790715839174,
      "learning_rate": 4.932903532270939e-06,
      "loss": 0.6275,
      "step": 74
    },
    {
      "epoch": 0.3,
      "grad_norm": 0.2636499704034946,
      "learning_rate": 4.928917723494854e-06,
      "loss": 0.6422,
      "step": 75
    },
    {
      "epoch": 0.304,
      "grad_norm": 0.28133008746598603,
      "learning_rate": 4.924818623774178e-06,
      "loss": 0.6355,
      "step": 76
    },
    {
      "epoch": 0.308,
      "grad_norm": 0.2779410264901997,
      "learning_rate": 4.920606424301424e-06,
      "loss": 0.6408,
      "step": 77
    },
    {
      "epoch": 0.312,
      "grad_norm": 0.2897057665750394,
      "learning_rate": 4.916281321544362e-06,
      "loss": 0.6402,
      "step": 78
    },
    {
      "epoch": 0.316,
      "grad_norm": 0.2761996704216673,
      "learning_rate": 4.911843517236867e-06,
      "loss": 0.653,
      "step": 79
    },
    {
      "epoch": 0.32,
      "grad_norm": 0.29413202875425243,
      "learning_rate": 4.907293218369499e-06,
      "loss": 0.6298,
      "step": 80
    },
    {
      "epoch": 0.324,
      "grad_norm": 0.2861680279307969,
      "learning_rate": 4.9026306371798526e-06,
      "loss": 0.6553,
      "step": 81
    },
    {
      "epoch": 0.328,
      "grad_norm": 0.2848851913581759,
      "learning_rate": 4.897855991142658e-06,
      "loss": 0.6076,
      "step": 82
    },
    {
      "epoch": 0.332,
      "grad_norm": 0.2687848093353044,
      "learning_rate": 4.892969502959639e-06,
      "loss": 0.6311,
      "step": 83
    },
    {
      "epoch": 0.336,
      "grad_norm": 0.27061397240144314,
      "learning_rate": 4.8879714005491205e-06,
      "loss": 0.6148,
      "step": 84
    },
    {
      "epoch": 0.34,
      "grad_norm": 0.2663532374960906,
      "learning_rate": 4.882861917035403e-06,
      "loss": 0.6412,
      "step": 85
    },
    {
      "epoch": 0.344,
      "grad_norm": 0.27752950332016424,
      "learning_rate": 4.8776412907378845e-06,
      "loss": 0.642,
      "step": 86
    },
    {
      "epoch": 0.348,
      "grad_norm": 0.2772792789989426,
      "learning_rate": 4.87230976515995e-06,
      "loss": 0.6184,
      "step": 87
    },
    {
      "epoch": 0.352,
      "grad_norm": 0.29699617925202587,
      "learning_rate": 4.8668675889776095e-06,
      "loss": 0.6275,
      "step": 88
    },
    {
      "epoch": 0.356,
      "grad_norm": 0.2736517750983177,
      "learning_rate": 4.861315016027902e-06,
      "loss": 0.6347,
      "step": 89
    },
    {
      "epoch": 0.36,
      "grad_norm": 0.2871346395772671,
      "learning_rate": 4.855652305297052e-06,
      "loss": 0.6132,
      "step": 90
    },
    {
      "epoch": 0.364,
      "grad_norm": 0.27131438304958966,
      "learning_rate": 4.849879720908394e-06,
      "loss": 0.6026,
      "step": 91
    },
    {
      "epoch": 0.368,
      "grad_norm": 0.2701400539855458,
      "learning_rate": 4.843997532110051e-06,
      "loss": 0.6385,
      "step": 92
    },
    {
      "epoch": 0.372,
      "grad_norm": 0.2870430498497826,
      "learning_rate": 4.8380060132623776e-06,
      "loss": 0.624,
      "step": 93
    },
    {
      "epoch": 0.376,
      "grad_norm": 0.2889305346509075,
      "learning_rate": 4.83190544382516e-06,
      "loss": 0.6375,
      "step": 94
    },
    {
      "epoch": 0.38,
      "grad_norm": 0.2832554152572974,
      "learning_rate": 4.825696108344583e-06,
      "loss": 0.6348,
      "step": 95
    },
    {
      "epoch": 0.384,
      "grad_norm": 0.282963454209734,
      "learning_rate": 4.819378296439962e-06,
      "loss": 0.6425,
      "step": 96
    },
    {
      "epoch": 0.388,
      "grad_norm": 0.28136329883499284,
      "learning_rate": 4.812952302790226e-06,
      "loss": 0.6238,
      "step": 97
    },
    {
      "epoch": 0.392,
      "grad_norm": 0.2751896445755537,
      "learning_rate": 4.80641842712018e-06,
      "loss": 0.6453,
      "step": 98
    },
    {
      "epoch": 0.396,
      "grad_norm": 0.2802941094985609,
      "learning_rate": 4.799776974186523e-06,
      "loss": 0.6362,
      "step": 99
    },
    {
      "epoch": 0.4,
      "grad_norm": 0.2816828378634338,
      "learning_rate": 4.793028253763633e-06,
      "loss": 0.6394,
      "step": 100
    },
    {
      "epoch": 0.404,
      "grad_norm": 0.28942806582562414,
      "learning_rate": 4.786172580629118e-06,
      "loss": 0.6106,
      "step": 101
    },
    {
      "epoch": 0.408,
      "grad_norm": 0.28396225609673553,
      "learning_rate": 4.7792102745491345e-06,
      "loss": 0.6302,
      "step": 102
    },
    {
      "epoch": 0.412,
      "grad_norm": 0.269131748983874,
      "learning_rate": 4.772141660263472e-06,
      "loss": 0.6247,
      "step": 103
    },
    {
      "epoch": 0.416,
      "grad_norm": 0.28932510717249166,
      "learning_rate": 4.764967067470409e-06,
      "loss": 0.6201,
      "step": 104
    },
    {
      "epoch": 0.42,
      "grad_norm": 0.2734040302258933,
      "learning_rate": 4.757686830811332e-06,
      "loss": 0.6092,
      "step": 105
    },
    {
      "epoch": 0.424,
      "grad_norm": 0.28628481564556507,
      "learning_rate": 4.750301289855128e-06,
      "loss": 0.6284,
      "step": 106
    },
    {
      "epoch": 0.428,
      "grad_norm": 0.30932729909286755,
      "learning_rate": 4.742810789082345e-06,
      "loss": 0.6332,
      "step": 107
    },
    {
      "epoch": 0.432,
      "grad_norm": 0.2707515621211506,
      "learning_rate": 4.735215677869129e-06,
      "loss": 0.6107,
      "step": 108
    },
    {
      "epoch": 0.436,
      "grad_norm": 0.29859925040242785,
      "learning_rate": 4.72751631047092e-06,
      "loss": 0.6477,
      "step": 109
    },
    {
      "epoch": 0.44,
      "grad_norm": 0.2974079849667466,
      "learning_rate": 4.7197130460059385e-06,
      "loss": 0.632,
      "step": 110
    },
    {
      "epoch": 0.444,
      "grad_norm": 0.28109309454092835,
      "learning_rate": 4.711806248438428e-06,
      "loss": 0.6308,
      "step": 111
    },
    {
      "epoch": 0.448,
      "grad_norm": 0.287627270908265,
      "learning_rate": 4.7037962865616795e-06,
      "loss": 0.6322,
      "step": 112
    },
    {
      "epoch": 0.452,
      "grad_norm": 0.2859506113795605,
      "learning_rate": 4.695683533980835e-06,
      "loss": 0.6196,
      "step": 113
    },
    {
      "epoch": 0.456,
      "grad_norm": 0.27768505471724575,
      "learning_rate": 4.687468369095457e-06,
      "loss": 0.6107,
      "step": 114
    },
    {
      "epoch": 0.46,
      "grad_norm": 0.2985912133376052,
      "learning_rate": 4.679151175081879e-06,
      "loss": 0.6316,
      "step": 115
    },
    {
      "epoch": 0.464,
      "grad_norm": 0.27790254502289174,
      "learning_rate": 4.6707323398753346e-06,
      "loss": 0.6194,
      "step": 116
    },
    {
      "epoch": 0.468,
      "grad_norm": 0.2770496916475714,
      "learning_rate": 4.662212256151865e-06,
      "loss": 0.5938,
      "step": 117
    },
    {
      "epoch": 0.472,
      "grad_norm": 0.2750067208531331,
      "learning_rate": 4.6535913213100005e-06,
      "loss": 0.6125,
      "step": 118
    },
    {
      "epoch": 0.476,
      "grad_norm": 0.2640533408637943,
      "learning_rate": 4.644869937452224e-06,
      "loss": 0.6245,
      "step": 119
    },
    {
      "epoch": 0.48,
      "grad_norm": 0.27591929390872805,
      "learning_rate": 4.636048511366222e-06,
      "loss": 0.6186,
      "step": 120
    },
    {
      "epoch": 0.484,
      "grad_norm": 0.29643778935625803,
      "learning_rate": 4.627127454505902e-06,
      "loss": 0.6086,
      "step": 121
    },
    {
      "epoch": 0.488,
      "grad_norm": 0.28076149524670235,
      "learning_rate": 4.618107182972209e-06,
      "loss": 0.6158,
      "step": 122
    },
    {
      "epoch": 0.492,
      "grad_norm": 0.28214439623978305,
      "learning_rate": 4.6089881174937146e-06,
      "loss": 0.6299,
      "step": 123
    },
    {
      "epoch": 0.496,
      "grad_norm": 0.2885430172419291,
      "learning_rate": 4.599770683406992e-06,
      "loss": 0.6367,
      "step": 124
    },
    {
      "epoch": 0.5,
      "grad_norm": 0.28978271139767015,
      "learning_rate": 4.590455310636778e-06,
      "loss": 0.6248,
      "step": 125
    },
    {
      "epoch": 0.504,
      "grad_norm": 0.2763777911865909,
      "learning_rate": 4.58104243367592e-06,
      "loss": 0.6235,
      "step": 126
    },
    {
      "epoch": 0.508,
      "grad_norm": 0.2788581607213461,
      "learning_rate": 4.571532491565115e-06,
      "loss": 0.6201,
      "step": 127
    },
    {
      "epoch": 0.512,
      "grad_norm": 0.28522419436506885,
      "learning_rate": 4.561925927872421e-06,
      "loss": 0.6287,
      "step": 128
    },
    {
      "epoch": 0.516,
      "grad_norm": 0.2759422238312871,
      "learning_rate": 4.55222319067258e-06,
      "loss": 0.6105,
      "step": 129
    },
    {
      "epoch": 0.52,
      "grad_norm": 0.271521828303117,
      "learning_rate": 4.542424732526105e-06,
      "loss": 0.6004,
      "step": 130
    },
    {
      "epoch": 0.524,
      "grad_norm": 0.2832768486502443,
      "learning_rate": 4.532531010458188e-06,
      "loss": 0.6438,
      "step": 131
    },
    {
      "epoch": 0.528,
      "grad_norm": 0.28545986352466657,
      "learning_rate": 4.522542485937369e-06,
      "loss": 0.6147,
      "step": 132
    },
    {
      "epoch": 0.532,
      "grad_norm": 0.2843650568512383,
      "learning_rate": 4.512459624854017e-06,
      "loss": 0.6347,
      "step": 133
    },
    {
      "epoch": 0.536,
      "grad_norm": 0.2758779923686556,
      "learning_rate": 4.5022828974986044e-06,
      "loss": 0.6111,
      "step": 134
    },
    {
      "epoch": 0.54,
      "grad_norm": 0.28471240078326554,
      "learning_rate": 4.4920127785397615e-06,
      "loss": 0.6161,
      "step": 135
    },
    {
      "epoch": 0.544,
      "grad_norm": 0.27215538114487603,
      "learning_rate": 4.481649747002146e-06,
      "loss": 0.6019,
      "step": 136
    },
    {
      "epoch": 0.548,
      "grad_norm": 0.27161590017753495,
      "learning_rate": 4.471194286244094e-06,
      "loss": 0.6229,
      "step": 137
    },
    {
      "epoch": 0.552,
      "grad_norm": 0.2786884282741861,
      "learning_rate": 4.460646883935079e-06,
      "loss": 0.6217,
      "step": 138
    },
    {
      "epoch": 0.556,
      "grad_norm": 0.29095908793086706,
      "learning_rate": 4.4500080320329615e-06,
      "loss": 0.6212,
      "step": 139
    },
    {
      "epoch": 0.56,
      "grad_norm": 0.2797512942233689,
      "learning_rate": 4.43927822676105e-06,
      "loss": 0.6183,
      "step": 140
    },
    {
      "epoch": 0.564,
      "grad_norm": 0.2701904530059608,
      "learning_rate": 4.428457968584945e-06,
      "loss": 0.6067,
      "step": 141
    },
    {
      "epoch": 0.568,
      "grad_norm": 0.2924071263588622,
      "learning_rate": 4.417547762189207e-06,
      "loss": 0.6167,
      "step": 142
    },
    {
      "epoch": 0.572,
      "grad_norm": 0.2684300131690406,
      "learning_rate": 4.40654811645381e-06,
      "loss": 0.6185,
      "step": 143
    },
    {
      "epoch": 0.576,
      "grad_norm": 0.2774759359262972,
      "learning_rate": 4.395459544430407e-06,
      "loss": 0.602,
      "step": 144
    },
    {
      "epoch": 0.58,
      "grad_norm": 0.2808643430953345,
      "learning_rate": 4.384282563318403e-06,
      "loss": 0.598,
      "step": 145
    },
    {
      "epoch": 0.584,
      "grad_norm": 0.27015365579319356,
      "learning_rate": 4.373017694440828e-06,
      "loss": 0.5857,
      "step": 146
    },
    {
      "epoch": 0.588,
      "grad_norm": 0.2856861787094523,
      "learning_rate": 4.361665463220023e-06,
      "loss": 0.6206,
      "step": 147
    },
    {
      "epoch": 0.592,
      "grad_norm": 0.28199517014381215,
      "learning_rate": 4.35022639915313e-06,
      "loss": 0.6094,
      "step": 148
    },
    {
      "epoch": 0.596,
      "grad_norm": 0.27010314199532126,
      "learning_rate": 4.338701035787403e-06,
      "loss": 0.5947,
      "step": 149
    },
    {
      "epoch": 0.6,
      "grad_norm": 0.2601892508343049,
      "learning_rate": 4.32708991069531e-06,
      "loss": 0.5871,
      "step": 150
    },
    {
      "epoch": 0.604,
      "grad_norm": 0.27825064672033506,
      "learning_rate": 4.315393565449472e-06,
      "loss": 0.6093,
      "step": 151
    },
    {
      "epoch": 0.608,
      "grad_norm": 0.27958608277896724,
      "learning_rate": 4.30361254559739e-06,
      "loss": 0.5951,
      "step": 152
    },
    {
      "epoch": 0.612,
      "grad_norm": 0.2758779818206466,
      "learning_rate": 4.291747400636009e-06,
      "loss": 0.6062,
      "step": 153
    },
    {
      "epoch": 0.616,
      "grad_norm": 0.29572303953208817,
      "learning_rate": 4.279798683986084e-06,
      "loss": 0.605,
      "step": 154
    },
    {
      "epoch": 0.62,
      "grad_norm": 0.28194612739384267,
      "learning_rate": 4.267766952966369e-06,
      "loss": 0.6078,
      "step": 155
    },
    {
      "epoch": 0.624,
      "grad_norm": 0.28615376291544004,
      "learning_rate": 4.255652768767619e-06,
      "loss": 0.6319,
      "step": 156
    },
    {
      "epoch": 0.628,
      "grad_norm": 0.26959219285273633,
      "learning_rate": 4.243456696426415e-06,
      "loss": 0.5968,
      "step": 157
    },
    {
      "epoch": 0.632,
      "grad_norm": 0.27878753771339543,
      "learning_rate": 4.2311793047988145e-06,
      "loss": 0.6214,
      "step": 158
    },
    {
      "epoch": 0.636,
      "grad_norm": 0.2779844282953486,
      "learning_rate": 4.218821166533813e-06,
      "loss": 0.5964,
      "step": 159
    },
    {
      "epoch": 0.64,
      "grad_norm": 0.2767287929857217,
      "learning_rate": 4.206382858046636e-06,
      "loss": 0.6187,
      "step": 160
    },
    {
      "epoch": 0.644,
      "grad_norm": 0.2652936251998452,
      "learning_rate": 4.193864959491853e-06,
      "loss": 0.5897,
      "step": 161
    },
    {
      "epoch": 0.648,
      "grad_norm": 0.26227694980471933,
      "learning_rate": 4.181268054736319e-06,
      "loss": 0.6107,
      "step": 162
    },
    {
      "epoch": 0.652,
      "grad_norm": 0.2690441499487734,
      "learning_rate": 4.16859273133194e-06,
      "loss": 0.6012,
      "step": 163
    },
    {
      "epoch": 0.656,
      "grad_norm": 0.26934906424793176,
      "learning_rate": 4.15583958048827e-06,
      "loss": 0.6086,
      "step": 164
    },
    {
      "epoch": 0.66,
      "grad_norm": 0.2727839454931186,
      "learning_rate": 4.143009197044932e-06,
      "loss": 0.6156,
      "step": 165
    },
    {
      "epoch": 0.664,
      "grad_norm": 0.2767715664709689,
      "learning_rate": 4.130102179443877e-06,
      "loss": 0.607,
      "step": 166
    },
    {
      "epoch": 0.668,
      "grad_norm": 0.27462389864805775,
      "learning_rate": 4.117119129701468e-06,
      "loss": 0.598,
      "step": 167
    },
    {
      "epoch": 0.672,
      "grad_norm": 0.28124958800487015,
      "learning_rate": 4.104060653380403e-06,
      "loss": 0.6174,
      "step": 168
    },
    {
      "epoch": 0.676,
      "grad_norm": 0.26867080247614167,
      "learning_rate": 4.090927359561469e-06,
      "loss": 0.6222,
      "step": 169
    },
    {
      "epoch": 0.68,
      "grad_norm": 0.27329020109654967,
      "learning_rate": 4.077719860815132e-06,
      "loss": 0.6174,
      "step": 170
    },
    {
      "epoch": 0.684,
      "grad_norm": 0.2598239429892548,
      "learning_rate": 4.064438773172966e-06,
      "loss": 0.5949,
      "step": 171
    },
    {
      "epoch": 0.688,
      "grad_norm": 0.26610910407219807,
      "learning_rate": 4.051084716098921e-06,
      "loss": 0.5876,
      "step": 172
    },
    {
      "epoch": 0.692,
      "grad_norm": 0.28832248653224085,
      "learning_rate": 4.037658312460424e-06,
      "loss": 0.6038,
      "step": 173
    },
    {
      "epoch": 0.696,
      "grad_norm": 0.27995126745782395,
      "learning_rate": 4.024160188499337e-06,
      "loss": 0.6024,
      "step": 174
    },
    {
      "epoch": 0.7,
      "grad_norm": 0.2708024464743442,
      "learning_rate": 4.010590973802737e-06,
      "loss": 0.6166,
      "step": 175
    },
    {
      "epoch": 0.704,
      "grad_norm": 0.27379773644645394,
      "learning_rate": 3.996951301273556e-06,
      "loss": 0.6172,
      "step": 176
    },
    {
      "epoch": 0.708,
      "grad_norm": 0.2704403624062539,
      "learning_rate": 3.983241807101064e-06,
      "loss": 0.5848,
      "step": 177
    },
    {
      "epoch": 0.712,
      "grad_norm": 0.26367325554187204,
      "learning_rate": 3.969463130731183e-06,
      "loss": 0.6084,
      "step": 178
    },
    {
      "epoch": 0.716,
      "grad_norm": 0.2714449492216179,
      "learning_rate": 3.955615914836678e-06,
      "loss": 0.6067,
      "step": 179
    },
    {
      "epoch": 0.72,
      "grad_norm": 0.27396192782433526,
      "learning_rate": 3.941700805287169e-06,
      "loss": 0.6049,
      "step": 180
    },
    {
      "epoch": 0.724,
      "grad_norm": 0.2712108680127688,
      "learning_rate": 3.927718451119009e-06,
      "loss": 0.5981,
      "step": 181
    },
    {
      "epoch": 0.728,
      "grad_norm": 0.27016877733602884,
      "learning_rate": 3.913669504505015e-06,
      "loss": 0.6148,
      "step": 182
    },
    {
      "epoch": 0.732,
      "grad_norm": 0.2986908827790219,
      "learning_rate": 3.8995546207240455e-06,
      "loss": 0.6293,
      "step": 183
    },
    {
      "epoch": 0.736,
      "grad_norm": 0.27281610268420575,
      "learning_rate": 3.8853744581304376e-06,
      "loss": 0.5937,
      "step": 184
    },
    {
      "epoch": 0.74,
      "grad_norm": 0.28387181164952147,
      "learning_rate": 3.871129678123297e-06,
      "loss": 0.6098,
      "step": 185
    },
    {
      "epoch": 0.744,
      "grad_norm": 0.2740312224605285,
      "learning_rate": 3.856820945115655e-06,
      "loss": 0.6078,
      "step": 186
    },
    {
      "epoch": 0.748,
      "grad_norm": 0.2662514099930545,
      "learning_rate": 3.84244892650347e-06,
      "loss": 0.6254,
      "step": 187
    },
    {
      "epoch": 0.752,
      "grad_norm": 0.26802564374459203,
      "learning_rate": 3.828014292634508e-06,
      "loss": 0.6121,
      "step": 188
    },
    {
      "epoch": 0.756,
      "grad_norm": 0.28248517647364846,
      "learning_rate": 3.813517716777069e-06,
      "loss": 0.6202,
      "step": 189
    },
    {
      "epoch": 0.76,
      "grad_norm": 0.272622897496479,
      "learning_rate": 3.798959875088584e-06,
      "loss": 0.5901,
      "step": 190
    },
    {
      "epoch": 0.764,
      "grad_norm": 0.27197490333330376,
      "learning_rate": 3.7843414465840823e-06,
      "loss": 0.5856,
      "step": 191
    },
    {
      "epoch": 0.768,
      "grad_norm": 0.26663311683845875,
      "learning_rate": 3.769663113104516e-06,
      "loss": 0.5907,
      "step": 192
    },
    {
      "epoch": 0.772,
      "grad_norm": 0.2714585682015405,
      "learning_rate": 3.7549255592849575e-06,
      "loss": 0.6072,
      "step": 193
    },
    {
      "epoch": 0.776,
      "grad_norm": 0.2766267849608307,
      "learning_rate": 3.7401294725226707e-06,
      "loss": 0.6158,
      "step": 194
    },
    {
      "epoch": 0.78,
      "grad_norm": 0.26291754258948374,
      "learning_rate": 3.7252755429450437e-06,
      "loss": 0.5921,
      "step": 195
    },
    {
      "epoch": 0.784,
      "grad_norm": 0.26530027759256725,
      "learning_rate": 3.7103644633774015e-06,
      "loss": 0.5841,
      "step": 196
    },
    {
      "epoch": 0.788,
      "grad_norm": 0.26634011298693916,
      "learning_rate": 3.695396929310693e-06,
      "loss": 0.6147,
      "step": 197
    },
    {
      "epoch": 0.792,
      "grad_norm": 0.26354297421036926,
      "learning_rate": 3.680373638869047e-06,
      "loss": 0.6061,
      "step": 198
    },
    {
      "epoch": 0.796,
      "grad_norm": 0.2738313781435172,
      "learning_rate": 3.665295292777214e-06,
      "loss": 0.5903,
      "step": 199
    },
    {
      "epoch": 0.8,
      "grad_norm": 0.27041603557150606,
      "learning_rate": 3.650162594327881e-06,
      "loss": 0.6216,
      "step": 200
    },
    {
      "epoch": 0.804,
      "grad_norm": 0.29233209893761647,
      "learning_rate": 3.634976249348867e-06,
      "loss": 0.6221,
      "step": 201
    },
    {
      "epoch": 0.808,
      "grad_norm": 0.28236083977418097,
      "learning_rate": 3.6197369661702052e-06,
      "loss": 0.6048,
      "step": 202
    },
    {
      "epoch": 0.812,
      "grad_norm": 0.2610570975246162,
      "learning_rate": 3.604445455591099e-06,
      "loss": 0.586,
      "step": 203
    },
    {
      "epoch": 0.816,
      "grad_norm": 0.27792495585124566,
      "learning_rate": 3.589102430846773e-06,
      "loss": 0.6052,
      "step": 204
    },
    {
      "epoch": 0.82,
      "grad_norm": 0.27390708264043134,
      "learning_rate": 3.5737086075752054e-06,
      "loss": 0.5968,
      "step": 205
    },
    {
      "epoch": 0.824,
      "grad_norm": 0.26341409551542055,
      "learning_rate": 3.5582647037837446e-06,
      "loss": 0.6128,
      "step": 206
    },
    {
      "epoch": 0.828,
      "grad_norm": 0.2659397773506794,
      "learning_rate": 3.5427714398156267e-06,
      "loss": 0.6171,
      "step": 207
    },
    {
      "epoch": 0.832,
      "grad_norm": 0.278302186061793,
      "learning_rate": 3.527229538316371e-06,
      "loss": 0.6001,
      "step": 208
    },
    {
      "epoch": 0.836,
      "grad_norm": 0.27935693059901906,
      "learning_rate": 3.5116397242000748e-06,
      "loss": 0.5915,
      "step": 209
    },
    {
      "epoch": 0.84,
      "grad_norm": 0.2681762304000699,
      "learning_rate": 3.4960027246156043e-06,
      "loss": 0.5982,
      "step": 210
    },
    {
      "epoch": 0.844,
      "grad_norm": 0.26833511905783713,
      "learning_rate": 3.480319268912676e-06,
      "loss": 0.5823,
      "step": 211
    },
    {
      "epoch": 0.848,
      "grad_norm": 0.27282574698411466,
      "learning_rate": 3.4645900886078388e-06,
      "loss": 0.6098,
      "step": 212
    },
    {
      "epoch": 0.852,
      "grad_norm": 0.2841325964241835,
      "learning_rate": 3.448815917350355e-06,
      "loss": 0.6054,
      "step": 213
    },
    {
      "epoch": 0.856,
      "grad_norm": 0.27811416712297765,
      "learning_rate": 3.432997490887979e-06,
      "loss": 0.6071,
      "step": 214
    },
    {
      "epoch": 0.86,
      "grad_norm": 0.26897184119138856,
      "learning_rate": 3.417135547032642e-06,
      "loss": 0.612,
      "step": 215
    },
    {
      "epoch": 0.864,
      "grad_norm": 0.2752083088143504,
      "learning_rate": 3.4012308256260366e-06,
      "loss": 0.6189,
      "step": 216
    },
    {
      "epoch": 0.868,
      "grad_norm": 0.27146348269262077,
      "learning_rate": 3.385284068505113e-06,
      "loss": 0.5914,
      "step": 217
    },
    {
      "epoch": 0.872,
      "grad_norm": 0.2616553689610195,
      "learning_rate": 3.369296019467473e-06,
      "loss": 0.5935,
      "step": 218
    },
    {
      "epoch": 0.876,
      "grad_norm": 0.27439799989957114,
      "learning_rate": 3.3532674242366764e-06,
      "loss": 0.5815,
      "step": 219
    },
    {
      "epoch": 0.88,
      "grad_norm": 0.27322752130009204,
      "learning_rate": 3.3371990304274654e-06,
      "loss": 0.593,
      "step": 220
    },
    {
      "epoch": 0.884,
      "grad_norm": 0.2776915569429837,
      "learning_rate": 3.3210915875108895e-06,
      "loss": 0.636,
      "step": 221
    },
    {
      "epoch": 0.888,
      "grad_norm": 0.27021599999486623,
      "learning_rate": 3.304945846779346e-06,
      "loss": 0.618,
      "step": 222
    },
    {
      "epoch": 0.892,
      "grad_norm": 0.27748742860539916,
      "learning_rate": 3.2887625613115427e-06,
      "loss": 0.5937,
      "step": 223
    },
    {
      "epoch": 0.896,
      "grad_norm": 0.28302784990496294,
      "learning_rate": 3.272542485937369e-06,
      "loss": 0.6093,
      "step": 224
    },
    {
      "epoch": 0.9,
      "grad_norm": 0.2759446162478654,
      "learning_rate": 3.25628637720269e-06,
      "loss": 0.6261,
      "step": 225
    },
    {
      "epoch": 0.904,
      "grad_norm": 0.2794223898275526,
      "learning_rate": 3.239994993334059e-06,
      "loss": 0.6098,
      "step": 226
    },
    {
      "epoch": 0.908,
      "grad_norm": 0.2863521541541822,
      "learning_rate": 3.2236690942033523e-06,
      "loss": 0.6122,
      "step": 227
    },
    {
      "epoch": 0.912,
      "grad_norm": 0.2810413939687701,
      "learning_rate": 3.207309441292325e-06,
      "loss": 0.6193,
      "step": 228
    },
    {
      "epoch": 0.916,
      "grad_norm": 0.2686724574589154,
      "learning_rate": 3.1909167976570977e-06,
      "loss": 0.5847,
      "step": 229
    },
    {
      "epoch": 0.92,
      "grad_norm": 0.27668513115261734,
      "learning_rate": 3.174491927892561e-06,
      "loss": 0.6083,
      "step": 230
    },
    {
      "epoch": 0.924,
      "grad_norm": 0.2838278304951842,
      "learning_rate": 3.158035598096715e-06,
      "loss": 0.597,
      "step": 231
    },
    {
      "epoch": 0.928,
      "grad_norm": 0.2702238327135827,
      "learning_rate": 3.1415485758349344e-06,
      "loss": 0.5884,
      "step": 232
    },
    {
      "epoch": 0.932,
      "grad_norm": 0.2677368696260293,
      "learning_rate": 3.1250316301041727e-06,
      "loss": 0.5835,
      "step": 233
    },
    {
      "epoch": 0.936,
      "grad_norm": 0.2848905107931979,
      "learning_rate": 3.1084855312970897e-06,
      "loss": 0.6255,
      "step": 234
    },
    {
      "epoch": 0.94,
      "grad_norm": 0.2788843417615313,
      "learning_rate": 3.091911051166117e-06,
      "loss": 0.6215,
      "step": 235
    },
    {
      "epoch": 0.944,
      "grad_norm": 0.27680715854768223,
      "learning_rate": 3.0753089627874668e-06,
      "loss": 0.6022,
      "step": 236
    },
    {
      "epoch": 0.948,
      "grad_norm": 0.26897413893634964,
      "learning_rate": 3.0586800405250677e-06,
      "loss": 0.6194,
      "step": 237
    },
    {
      "epoch": 0.952,
      "grad_norm": 0.2657249388294363,
      "learning_rate": 3.0420250599944525e-06,
      "loss": 0.5884,
      "step": 238
    },
    {
      "epoch": 0.956,
      "grad_norm": 0.26576882569871957,
      "learning_rate": 3.0253447980265754e-06,
      "loss": 0.5949,
      "step": 239
    },
    {
      "epoch": 0.96,
      "grad_norm": 0.2754408264706426,
      "learning_rate": 3.0086400326315853e-06,
      "loss": 0.5747,
      "step": 240
    },
    {
      "epoch": 0.964,
      "grad_norm": 0.26333215813151695,
      "learning_rate": 2.9919115429625295e-06,
      "loss": 0.6035,
      "step": 241
    },
    {
      "epoch": 0.968,
      "grad_norm": 0.2687016604382336,
      "learning_rate": 2.9751601092790185e-06,
      "loss": 0.6057,
      "step": 242
    },
    {
      "epoch": 0.972,
      "grad_norm": 0.28837866031127346,
      "learning_rate": 2.958386512910831e-06,
      "loss": 0.5737,
      "step": 243
    },
    {
      "epoch": 0.976,
      "grad_norm": 0.27805411098388116,
      "learning_rate": 2.941591536221469e-06,
      "loss": 0.6022,
      "step": 244
    },
    {
      "epoch": 0.98,
      "grad_norm": 0.27696163335286905,
      "learning_rate": 2.924775962571667e-06,
      "loss": 0.6081,
      "step": 245
    },
    {
      "epoch": 0.984,
      "grad_norm": 0.28312029387331156,
      "learning_rate": 2.907940576282856e-06,
      "loss": 0.6178,
      "step": 246
    },
    {
      "epoch": 0.988,
      "grad_norm": 0.2516197979856304,
      "learning_rate": 2.8910861626005774e-06,
      "loss": 0.5812,
      "step": 247
    },
    {
      "epoch": 0.992,
      "grad_norm": 0.25765553700425475,
      "learning_rate": 2.8742135076578608e-06,
      "loss": 0.5868,
      "step": 248
    },
    {
      "epoch": 0.996,
      "grad_norm": 0.27540024093918797,
      "learning_rate": 2.857323398438554e-06,
      "loss": 0.5976,
      "step": 249
    },
    {
      "epoch": 1.0,
      "grad_norm": 0.2637219036362386,
      "learning_rate": 2.840416622740617e-06,
      "loss": 0.6106,
      "step": 250
    },
    {
      "epoch": 1.004,
      "grad_norm": 0.3270308432987493,
      "learning_rate": 2.8234939691393765e-06,
      "loss": 0.5662,
      "step": 251
    },
    {
      "epoch": 1.008,
      "grad_norm": 0.3032460726332443,
      "learning_rate": 2.8065562269507464e-06,
      "loss": 0.589,
      "step": 252
    },
    {
      "epoch": 1.012,
      "grad_norm": 0.2761928570311231,
      "learning_rate": 2.789604186194411e-06,
      "loss": 0.5654,
      "step": 253
    },
    {
      "epoch": 1.016,
      "grad_norm": 0.27190181641802363,
      "learning_rate": 2.7726386375569748e-06,
      "loss": 0.5908,
      "step": 254
    },
    {
      "epoch": 1.02,
      "grad_norm": 0.27382272887323905,
      "learning_rate": 2.7556603723550855e-06,
      "loss": 0.553,
      "step": 255
    },
    {
      "epoch": 1.024,
      "grad_norm": 0.2699869769385883,
      "learning_rate": 2.7386701824985257e-06,
      "loss": 0.5517,
      "step": 256
    },
    {
      "epoch": 1.028,
      "grad_norm": 0.2740431770397761,
      "learning_rate": 2.721668860453271e-06,
      "loss": 0.5611,
      "step": 257
    },
    {
      "epoch": 1.032,
      "grad_norm": 0.27006742872479766,
      "learning_rate": 2.7046571992045334e-06,
      "loss": 0.5609,
      "step": 258
    },
    {
      "epoch": 1.036,
      "grad_norm": 0.2853628863130536,
      "learning_rate": 2.6876359922197703e-06,
      "loss": 0.5839,
      "step": 259
    },
    {
      "epoch": 1.04,
      "grad_norm": 0.2674934886902205,
      "learning_rate": 2.670606033411678e-06,
      "loss": 0.5692,
      "step": 260
    },
    {
      "epoch": 1.044,
      "grad_norm": 0.2717961476035451,
      "learning_rate": 2.653568117101159e-06,
      "loss": 0.5586,
      "step": 261
    },
    {
      "epoch": 1.048,
      "grad_norm": 0.2886664885195019,
      "learning_rate": 2.636523037980275e-06,
      "loss": 0.5592,
      "step": 262
    },
    {
      "epoch": 1.052,
      "grad_norm": 0.2913975175671213,
      "learning_rate": 2.6194715910751806e-06,
      "loss": 0.5588,
      "step": 263
    },
    {
      "epoch": 1.056,
      "grad_norm": 0.26635335287134043,
      "learning_rate": 2.602414571709036e-06,
      "loss": 0.5518,
      "step": 264
    },
    {
      "epoch": 1.06,
      "grad_norm": 0.27589381073650543,
      "learning_rate": 2.58535277546492e-06,
      "loss": 0.5588,
      "step": 265
    },
    {
      "epoch": 1.064,
      "grad_norm": 0.26891839582782207,
      "learning_rate": 2.5682869981487154e-06,
      "loss": 0.5671,
      "step": 266
    },
    {
      "epoch": 1.068,
      "grad_norm": 0.2699654255248958,
      "learning_rate": 2.5512180357519913e-06,
      "loss": 0.5663,
      "step": 267
    },
    {
      "epoch": 1.072,
      "grad_norm": 0.27268944362101294,
      "learning_rate": 2.5341466844148775e-06,
      "loss": 0.5689,
      "step": 268
    },
    {
      "epoch": 1.076,
      "grad_norm": 0.2895176432362157,
      "learning_rate": 2.5170737403889334e-06,
      "loss": 0.555,
      "step": 269
    },
    {
      "epoch": 1.08,
      "grad_norm": 0.275834303493974,
      "learning_rate": 2.5e-06,
      "loss": 0.5652,
      "step": 270
    },
    {
      "epoch": 1.084,
      "grad_norm": 0.27109148773857455,
      "learning_rate": 2.4829262596110674e-06,
      "loss": 0.5531,
      "step": 271
    },
    {
      "epoch": 1.088,
      "grad_norm": 0.2885944970109634,
      "learning_rate": 2.465853315585123e-06,
      "loss": 0.5645,
      "step": 272
    },
    {
      "epoch": 1.092,
      "grad_norm": 0.3325900759228949,
      "learning_rate": 2.44878196424801e-06,
      "loss": 0.567,
      "step": 273
    },
    {
      "epoch": 1.096,
      "grad_norm": 0.2656725633978504,
      "learning_rate": 2.431713001851286e-06,
      "loss": 0.5645,
      "step": 274
    },
    {
      "epoch": 1.1,
      "grad_norm": 0.28515147451371026,
      "learning_rate": 2.4146472245350804e-06,
      "loss": 0.5627,
      "step": 275
    },
    {
      "epoch": 1.104,
      "grad_norm": 0.2772295325997671,
      "learning_rate": 2.3975854282909645e-06,
      "loss": 0.5606,
      "step": 276
    },
    {
      "epoch": 1.108,
      "grad_norm": 0.2732258740956759,
      "learning_rate": 2.3805284089248203e-06,
      "loss": 0.558,
      "step": 277
    },
    {
      "epoch": 1.112,
      "grad_norm": 0.2657630056993197,
      "learning_rate": 2.3634769620197253e-06,
      "loss": 0.5639,
      "step": 278
    },
    {
      "epoch": 1.116,
      "grad_norm": 0.2741360647437433,
      "learning_rate": 2.3464318828988416e-06,
      "loss": 0.539,
      "step": 279
    },
    {
      "epoch": 1.12,
      "grad_norm": 0.27924091679374063,
      "learning_rate": 2.3293939665883233e-06,
      "loss": 0.563,
      "step": 280
    },
    {
      "epoch": 1.124,
      "grad_norm": 0.2663368697744116,
      "learning_rate": 2.3123640077802305e-06,
      "loss": 0.5639,
      "step": 281
    },
    {
      "epoch": 1.1280000000000001,
      "grad_norm": 0.2635159398782791,
      "learning_rate": 2.2953428007954682e-06,
      "loss": 0.5225,
      "step": 282
    },
    {
      "epoch": 1.1320000000000001,
      "grad_norm": 0.26790514294280293,
      "learning_rate": 2.2783311395467304e-06,
      "loss": 0.5556,
      "step": 283
    },
    {
      "epoch": 1.1360000000000001,
      "grad_norm": 0.27007506595154834,
      "learning_rate": 2.261329817501475e-06,
      "loss": 0.545,
      "step": 284
    },
    {
      "epoch": 1.1400000000000001,
      "grad_norm": 0.27274462523361476,
      "learning_rate": 2.2443396276449145e-06,
      "loss": 0.5724,
      "step": 285
    },
    {
      "epoch": 1.144,
      "grad_norm": 0.25666672253669337,
      "learning_rate": 2.2273613624430256e-06,
      "loss": 0.541,
      "step": 286
    },
    {
      "epoch": 1.148,
      "grad_norm": 0.30831325626025924,
      "learning_rate": 2.2103958138055897e-06,
      "loss": 0.5544,
      "step": 287
    },
    {
      "epoch": 1.152,
      "grad_norm": 0.26679849744140743,
      "learning_rate": 2.1934437730492544e-06,
      "loss": 0.5544,
      "step": 288
    },
    {
      "epoch": 1.156,
      "grad_norm": 0.26795918133510693,
      "learning_rate": 2.1765060308606243e-06,
      "loss": 0.5726,
      "step": 289
    },
    {
      "epoch": 1.16,
      "grad_norm": 0.26692330671607206,
      "learning_rate": 2.159583377259384e-06,
      "loss": 0.5771,
      "step": 290
    },
    {
      "epoch": 1.164,
      "grad_norm": 0.26800012106006976,
      "learning_rate": 2.142676601561447e-06,
      "loss": 0.5376,
      "step": 291
    },
    {
      "epoch": 1.168,
      "grad_norm": 0.25520584565446575,
      "learning_rate": 2.1257864923421405e-06,
      "loss": 0.5439,
      "step": 292
    },
    {
      "epoch": 1.172,
      "grad_norm": 0.26220655980779767,
      "learning_rate": 2.1089138373994226e-06,
      "loss": 0.552,
      "step": 293
    },
    {
      "epoch": 1.176,
      "grad_norm": 0.2599229946929615,
      "learning_rate": 2.092059423717145e-06,
      "loss": 0.5591,
      "step": 294
    },
    {
      "epoch": 1.18,
      "grad_norm": 0.26226465656771997,
      "learning_rate": 2.0752240374283334e-06,
      "loss": 0.5424,
      "step": 295
    },
    {
      "epoch": 1.184,
      "grad_norm": 0.27097583397605235,
      "learning_rate": 2.0584084637785316e-06,
      "loss": 0.5613,
      "step": 296
    },
    {
      "epoch": 1.188,
      "grad_norm": 0.25487494718289816,
      "learning_rate": 2.0416134870891697e-06,
      "loss": 0.5531,
      "step": 297
    },
    {
      "epoch": 1.192,
      "grad_norm": 0.258627059695691,
      "learning_rate": 2.0248398907209827e-06,
      "loss": 0.5603,
      "step": 298
    },
    {
      "epoch": 1.196,
      "grad_norm": 0.27660938201914487,
      "learning_rate": 2.008088457037472e-06,
      "loss": 0.5648,
      "step": 299
    },
    {
      "epoch": 1.2,
      "grad_norm": 0.26596618229804614,
      "learning_rate": 1.991359967368416e-06,
      "loss": 0.574,
      "step": 300
    },
    {
      "epoch": 1.204,
      "grad_norm": 0.2618039991994125,
      "learning_rate": 1.9746552019734246e-06,
      "loss": 0.5492,
      "step": 301
    },
    {
      "epoch": 1.208,
      "grad_norm": 0.272654754048079,
      "learning_rate": 1.957974940005548e-06,
      "loss": 0.5767,
      "step": 302
    },
    {
      "epoch": 1.212,
      "grad_norm": 0.26671171090168844,
      "learning_rate": 1.9413199594749327e-06,
      "loss": 0.5338,
      "step": 303
    },
    {
      "epoch": 1.216,
      "grad_norm": 0.270832034752946,
      "learning_rate": 1.9246910372125345e-06,
      "loss": 0.5647,
      "step": 304
    },
    {
      "epoch": 1.22,
      "grad_norm": 0.2643013014532281,
      "learning_rate": 1.9080889488338833e-06,
      "loss": 0.5597,
      "step": 305
    },
    {
      "epoch": 1.224,
      "grad_norm": 0.27092629093151294,
      "learning_rate": 1.8915144687029107e-06,
      "loss": 0.5485,
      "step": 306
    },
    {
      "epoch": 1.228,
      "grad_norm": 0.26613235310331923,
      "learning_rate": 1.874968369895828e-06,
      "loss": 0.552,
      "step": 307
    },
    {
      "epoch": 1.232,
      "grad_norm": 0.2582406507610385,
      "learning_rate": 1.8584514241650667e-06,
      "loss": 0.5477,
      "step": 308
    },
    {
      "epoch": 1.236,
      "grad_norm": 0.2687239346512157,
      "learning_rate": 1.8419644019032868e-06,
      "loss": 0.5716,
      "step": 309
    },
    {
      "epoch": 1.24,
      "grad_norm": 0.2634130471010307,
      "learning_rate": 1.8255080721074391e-06,
      "loss": 0.5511,
      "step": 310
    },
    {
      "epoch": 1.244,
      "grad_norm": 0.27164814594203673,
      "learning_rate": 1.8090832023429022e-06,
      "loss": 0.5605,
      "step": 311
    },
    {
      "epoch": 1.248,
      "grad_norm": 0.26506994687071844,
      "learning_rate": 1.792690558707675e-06,
      "loss": 0.5522,
      "step": 312
    },
    {
      "epoch": 1.252,
      "grad_norm": 0.2710404134246902,
      "learning_rate": 1.7763309057966487e-06,
      "loss": 0.5754,
      "step": 313
    },
    {
      "epoch": 1.256,
      "grad_norm": 0.2644648635617096,
      "learning_rate": 1.7600050066659418e-06,
      "loss": 0.5542,
      "step": 314
    },
    {
      "epoch": 1.26,
      "grad_norm": 0.2563441049550618,
      "learning_rate": 1.7437136227973108e-06,
      "loss": 0.5428,
      "step": 315
    },
    {
      "epoch": 1.264,
      "grad_norm": 0.2781486176634604,
      "learning_rate": 1.7274575140626318e-06,
      "loss": 0.5469,
      "step": 316
    },
    {
      "epoch": 1.268,
      "grad_norm": 0.26498925525964123,
      "learning_rate": 1.7112374386884583e-06,
      "loss": 0.5671,
      "step": 317
    },
    {
      "epoch": 1.272,
      "grad_norm": 0.27752994625791155,
      "learning_rate": 1.695054153220655e-06,
      "loss": 0.5437,
      "step": 318
    },
    {
      "epoch": 1.276,
      "grad_norm": 0.2602835745087367,
      "learning_rate": 1.678908412489111e-06,
      "loss": 0.5602,
      "step": 319
    },
    {
      "epoch": 1.28,
      "grad_norm": 0.2665909274717665,
      "learning_rate": 1.6628009695725348e-06,
      "loss": 0.5762,
      "step": 320
    },
    {
      "epoch": 1.284,
      "grad_norm": 0.2531045714341802,
      "learning_rate": 1.6467325757633242e-06,
      "loss": 0.5674,
      "step": 321
    },
    {
      "epoch": 1.288,
      "grad_norm": 0.2621120701043234,
      "learning_rate": 1.630703980532528e-06,
      "loss": 0.5598,
      "step": 322
    },
    {
      "epoch": 1.292,
      "grad_norm": 0.275203318844759,
      "learning_rate": 1.6147159314948873e-06,
      "loss": 0.564,
      "step": 323
    },
    {
      "epoch": 1.296,
      "grad_norm": 0.2603786935372431,
      "learning_rate": 1.5987691743739636e-06,
      "loss": 0.5535,
      "step": 324
    },
    {
      "epoch": 1.3,
      "grad_norm": 0.26776057829072847,
      "learning_rate": 1.5828644529673592e-06,
      "loss": 0.5627,
      "step": 325
    },
    {
      "epoch": 1.304,
      "grad_norm": 0.2623048715119549,
      "learning_rate": 1.5670025091120219e-06,
      "loss": 0.5685,
      "step": 326
    },
    {
      "epoch": 1.308,
      "grad_norm": 0.26608080729860306,
      "learning_rate": 1.5511840826496462e-06,
      "loss": 0.5695,
      "step": 327
    },
    {
      "epoch": 1.312,
      "grad_norm": 0.2673157087455161,
      "learning_rate": 1.5354099113921614e-06,
      "loss": 0.5354,
      "step": 328
    },
    {
      "epoch": 1.316,
      "grad_norm": 0.2754778027664838,
      "learning_rate": 1.519680731087325e-06,
      "loss": 0.5705,
      "step": 329
    },
    {
      "epoch": 1.32,
      "grad_norm": 0.2697793809470272,
      "learning_rate": 1.5039972753843966e-06,
      "loss": 0.5748,
      "step": 330
    },
    {
      "epoch": 1.324,
      "grad_norm": 0.2737068013002947,
      "learning_rate": 1.488360275799926e-06,
      "loss": 0.5554,
      "step": 331
    },
    {
      "epoch": 1.328,
      "grad_norm": 0.2618826312103745,
      "learning_rate": 1.4727704616836297e-06,
      "loss": 0.5516,
      "step": 332
    },
    {
      "epoch": 1.332,
      "grad_norm": 0.2610637671077678,
      "learning_rate": 1.457228560184374e-06,
      "loss": 0.5491,
      "step": 333
    },
    {
      "epoch": 1.336,
      "grad_norm": 0.2615954274949455,
      "learning_rate": 1.441735296216256e-06,
      "loss": 0.5556,
      "step": 334
    },
    {
      "epoch": 1.34,
      "grad_norm": 0.2650842665446321,
      "learning_rate": 1.4262913924247956e-06,
      "loss": 0.5475,
      "step": 335
    },
    {
      "epoch": 1.3439999999999999,
      "grad_norm": 0.25585711289460494,
      "learning_rate": 1.4108975691532273e-06,
      "loss": 0.5549,
      "step": 336
    },
    {
      "epoch": 1.3479999999999999,
      "grad_norm": 0.25250951856873494,
      "learning_rate": 1.3955545444089017e-06,
      "loss": 0.5601,
      "step": 337
    },
    {
      "epoch": 1.3519999999999999,
      "grad_norm": 0.2569876765702743,
      "learning_rate": 1.3802630338297956e-06,
      "loss": 0.548,
      "step": 338
    },
    {
      "epoch": 1.3559999999999999,
      "grad_norm": 0.28203813821342777,
      "learning_rate": 1.3650237506511333e-06,
      "loss": 0.566,
      "step": 339
    },
    {
      "epoch": 1.3599999999999999,
      "grad_norm": 0.27081394807954995,
      "learning_rate": 1.3498374056721198e-06,
      "loss": 0.5594,
      "step": 340
    },
    {
      "epoch": 1.3639999999999999,
      "grad_norm": 0.26811473148644754,
      "learning_rate": 1.334704707222787e-06,
      "loss": 0.5558,
      "step": 341
    },
    {
      "epoch": 1.3679999999999999,
      "grad_norm": 0.2635422950339071,
      "learning_rate": 1.3196263611309539e-06,
      "loss": 0.5707,
      "step": 342
    },
    {
      "epoch": 1.3719999999999999,
      "grad_norm": 0.25640933129102816,
      "learning_rate": 1.3046030706893079e-06,
      "loss": 0.5551,
      "step": 343
    },
    {
      "epoch": 1.376,
      "grad_norm": 0.26144553856379565,
      "learning_rate": 1.2896355366226e-06,
      "loss": 0.5689,
      "step": 344
    },
    {
      "epoch": 1.38,
      "grad_norm": 0.24921356601339556,
      "learning_rate": 1.2747244570549578e-06,
      "loss": 0.5223,
      "step": 345
    },
    {
      "epoch": 1.384,
      "grad_norm": 0.2739025094336896,
      "learning_rate": 1.2598705274773299e-06,
      "loss": 0.5553,
      "step": 346
    },
    {
      "epoch": 1.388,
      "grad_norm": 0.26711249323543973,
      "learning_rate": 1.2450744407150427e-06,
      "loss": 0.5411,
      "step": 347
    },
    {
      "epoch": 1.392,
      "grad_norm": 0.2695324913604517,
      "learning_rate": 1.2303368868954848e-06,
      "loss": 0.567,
      "step": 348
    },
    {
      "epoch": 1.396,
      "grad_norm": 0.25781431882124334,
      "learning_rate": 1.215658553415918e-06,
      "loss": 0.5674,
      "step": 349
    },
    {
      "epoch": 1.4,
      "grad_norm": 0.2594657444454727,
      "learning_rate": 1.2010401249114166e-06,
      "loss": 0.5366,
      "step": 350
    },
    {
      "epoch": 1.404,
      "grad_norm": 0.26265811514812376,
      "learning_rate": 1.1864822832229319e-06,
      "loss": 0.5465,
      "step": 351
    },
    {
      "epoch": 1.408,
      "grad_norm": 0.26623313611028854,
      "learning_rate": 1.1719857073654923e-06,
      "loss": 0.564,
      "step": 352
    },
    {
      "epoch": 1.412,
      "grad_norm": 0.2702849646886806,
      "learning_rate": 1.1575510734965305e-06,
      "loss": 0.5521,
      "step": 353
    },
    {
      "epoch": 1.416,
      "grad_norm": 0.2689048371097966,
      "learning_rate": 1.1431790548843464e-06,
      "loss": 0.5527,
      "step": 354
    },
    {
      "epoch": 1.42,
      "grad_norm": 0.25819241256540093,
      "learning_rate": 1.1288703218767027e-06,
      "loss": 0.54,
      "step": 355
    },
    {
      "epoch": 1.424,
      "grad_norm": 0.25437896341243765,
      "learning_rate": 1.1146255418695635e-06,
      "loss": 0.528,
      "step": 356
    },
    {
      "epoch": 1.428,
      "grad_norm": 0.25406879267543436,
      "learning_rate": 1.1004453792759547e-06,
      "loss": 0.5471,
      "step": 357
    },
    {
      "epoch": 1.432,
      "grad_norm": 0.2562241134177116,
      "learning_rate": 1.0863304954949856e-06,
      "loss": 0.55,
      "step": 358
    },
    {
      "epoch": 1.436,
      "grad_norm": 0.26542441620839063,
      "learning_rate": 1.072281548880992e-06,
      "loss": 0.5678,
      "step": 359
    },
    {
      "epoch": 1.44,
      "grad_norm": 0.269318418298225,
      "learning_rate": 1.0582991947128324e-06,
      "loss": 0.5615,
      "step": 360
    },
    {
      "epoch": 1.444,
      "grad_norm": 0.2691764014021926,
      "learning_rate": 1.0443840851633227e-06,
      "loss": 0.565,
      "step": 361
    },
    {
      "epoch": 1.448,
      "grad_norm": 0.2594862635555543,
      "learning_rate": 1.0305368692688175e-06,
      "loss": 0.5666,
      "step": 362
    },
    {
      "epoch": 1.452,
      "grad_norm": 0.25315547667760335,
      "learning_rate": 1.0167581928989373e-06,
      "loss": 0.5555,
      "step": 363
    },
    {
      "epoch": 1.456,
      "grad_norm": 0.2613895044115319,
      "learning_rate": 1.0030486987264436e-06,
      "loss": 0.5776,
      "step": 364
    },
    {
      "epoch": 1.46,
      "grad_norm": 0.2631431546714004,
      "learning_rate": 9.89409026197264e-07,
      "loss": 0.5365,
      "step": 365
    },
    {
      "epoch": 1.464,
      "grad_norm": 0.2647751182203821,
      "learning_rate": 9.758398115006637e-07,
      "loss": 0.5573,
      "step": 366
    },
    {
      "epoch": 1.468,
      "grad_norm": 0.26208143115660865,
      "learning_rate": 9.623416875395763e-07,
      "loss": 0.5605,
      "step": 367
    },
    {
      "epoch": 1.472,
      "grad_norm": 0.2704474380702507,
      "learning_rate": 9.489152839010799e-07,
      "loss": 0.5697,
      "step": 368
    },
    {
      "epoch": 1.476,
      "grad_norm": 0.275761932132816,
      "learning_rate": 9.355612268270339e-07,
      "loss": 0.5586,
      "step": 369
    },
    {
      "epoch": 1.48,
      "grad_norm": 0.27242480864427504,
      "learning_rate": 9.222801391848688e-07,
      "loss": 0.5573,
      "step": 370
    },
    {
      "epoch": 1.484,
      "grad_norm": 0.254674372509468,
      "learning_rate": 9.090726404385319e-07,
      "loss": 0.5377,
      "step": 371
    },
    {
      "epoch": 1.488,
      "grad_norm": 0.25888213496864965,
      "learning_rate": 8.959393466195973e-07,
      "loss": 0.5278,
      "step": 372
    },
    {
      "epoch": 1.492,
      "grad_norm": 0.2527033545759895,
      "learning_rate": 8.828808702985325e-07,
      "loss": 0.5475,
      "step": 373
    },
    {
      "epoch": 1.496,
      "grad_norm": 0.2669973822070058,
      "learning_rate": 8.69897820556124e-07,
      "loss": 0.5652,
      "step": 374
    },
    {
      "epoch": 1.5,
      "grad_norm": 0.26903334117277744,
      "learning_rate": 8.569908029550686e-07,
      "loss": 0.5555,
      "step": 375
    },
    {
      "epoch": 1.504,
      "grad_norm": 0.2634312253069988,
      "learning_rate": 8.441604195117315e-07,
      "loss": 0.5581,
      "step": 376
    },
    {
      "epoch": 1.508,
      "grad_norm": 0.2683641729946301,
      "learning_rate": 8.31407268668061e-07,
      "loss": 0.5541,
      "step": 377
    },
    {
      "epoch": 1.512,
      "grad_norm": 0.26616933690609357,
      "learning_rate": 8.187319452636821e-07,
      "loss": 0.5564,
      "step": 378
    },
    {
      "epoch": 1.516,
      "grad_norm": 0.257110593824904,
      "learning_rate": 8.061350405081484e-07,
      "loss": 0.5571,
      "step": 379
    },
    {
      "epoch": 1.52,
      "grad_norm": 0.26192694577938985,
      "learning_rate": 7.936171419533653e-07,
      "loss": 0.5643,
      "step": 380
    },
    {
      "epoch": 1.524,
      "grad_norm": 0.2706323728611535,
      "learning_rate": 7.811788334661871e-07,
      "loss": 0.5493,
      "step": 381
    },
    {
      "epoch": 1.528,
      "grad_norm": 0.2567197095741578,
      "learning_rate": 7.688206952011862e-07,
      "loss": 0.559,
      "step": 382
    },
    {
      "epoch": 1.532,
      "grad_norm": 0.26232934311415484,
      "learning_rate": 7.565433035735859e-07,
      "loss": 0.557,
      "step": 383
    },
    {
      "epoch": 1.536,
      "grad_norm": 0.26498083181616333,
      "learning_rate": 7.443472312323824e-07,
      "loss": 0.5519,
      "step": 384
    },
    {
      "epoch": 1.54,
      "grad_norm": 0.2696289770109232,
      "learning_rate": 7.322330470336314e-07,
      "loss": 0.5674,
      "step": 385
    },
    {
      "epoch": 1.544,
      "grad_norm": 0.275304780317595,
      "learning_rate": 7.202013160139159e-07,
      "loss": 0.5367,
      "step": 386
    },
    {
      "epoch": 1.548,
      "grad_norm": 0.26693717335236866,
      "learning_rate": 7.082525993639916e-07,
      "loss": 0.5587,
      "step": 387
    },
    {
      "epoch": 1.552,
      "grad_norm": 0.26002682618175454,
      "learning_rate": 6.963874544026109e-07,
      "loss": 0.5496,
      "step": 388
    },
    {
      "epoch": 1.556,
      "grad_norm": 0.2622047147420331,
      "learning_rate": 6.846064345505296e-07,
      "loss": 0.5725,
      "step": 389
    },
    {
      "epoch": 1.56,
      "grad_norm": 0.26684263893910143,
      "learning_rate": 6.729100893046897e-07,
      "loss": 0.5294,
      "step": 390
    },
    {
      "epoch": 1.564,
      "grad_norm": 0.26176221742469546,
      "learning_rate": 6.612989642125977e-07,
      "loss": 0.5657,
      "step": 391
    },
    {
      "epoch": 1.568,
      "grad_norm": 0.26631983610262083,
      "learning_rate": 6.497736008468703e-07,
      "loss": 0.5685,
      "step": 392
    },
    {
      "epoch": 1.572,
      "grad_norm": 0.26179902426984153,
      "learning_rate": 6.383345367799784e-07,
      "loss": 0.554,
      "step": 393
    },
    {
      "epoch": 1.576,
      "grad_norm": 0.27346453641516727,
      "learning_rate": 6.269823055591726e-07,
      "loss": 0.5623,
      "step": 394
    },
    {
      "epoch": 1.58,
      "grad_norm": 0.270925236344549,
      "learning_rate": 6.157174366815979e-07,
      "loss": 0.5588,
      "step": 395
    },
    {
      "epoch": 1.584,
      "grad_norm": 0.25741948324451286,
      "learning_rate": 6.045404555695935e-07,
      "loss": 0.5403,
      "step": 396
    },
    {
      "epoch": 1.588,
      "grad_norm": 0.2613501970796598,
      "learning_rate": 5.934518835461908e-07,
      "loss": 0.5493,
      "step": 397
    },
    {
      "epoch": 1.592,
      "grad_norm": 0.2611536278189242,
      "learning_rate": 5.824522378107936e-07,
      "loss": 0.5738,
      "step": 398
    },
    {
      "epoch": 1.596,
      "grad_norm": 0.2675151064196103,
      "learning_rate": 5.715420314150547e-07,
      "loss": 0.5538,
      "step": 399
    },
    {
      "epoch": 1.6,
      "grad_norm": 0.2617887614907697,
      "learning_rate": 5.607217732389503e-07,
      "loss": 0.5476,
      "step": 400
    },
    {
      "epoch": 1.604,
      "grad_norm": 0.25503541212802705,
      "learning_rate": 5.499919679670385e-07,
      "loss": 0.5462,
      "step": 401
    },
    {
      "epoch": 1.608,
      "grad_norm": 0.2572004653637581,
      "learning_rate": 5.393531160649221e-07,
      "loss": 0.5555,
      "step": 402
    },
    {
      "epoch": 1.612,
      "grad_norm": 0.2565584643823191,
      "learning_rate": 5.288057137559066e-07,
      "loss": 0.5519,
      "step": 403
    },
    {
      "epoch": 1.616,
      "grad_norm": 0.2542793603223879,
      "learning_rate": 5.183502529978548e-07,
      "loss": 0.553,
      "step": 404
    },
    {
      "epoch": 1.62,
      "grad_norm": 0.26475636006034603,
      "learning_rate": 5.079872214602388e-07,
      "loss": 0.5508,
      "step": 405
    },
    {
      "epoch": 1.624,
      "grad_norm": 0.2579561028420663,
      "learning_rate": 4.977171025013961e-07,
      "loss": 0.5445,
      "step": 406
    },
    {
      "epoch": 1.6280000000000001,
      "grad_norm": 0.2609669688970827,
      "learning_rate": 4.875403751459825e-07,
      "loss": 0.5568,
      "step": 407
    },
    {
      "epoch": 1.6320000000000001,
      "grad_norm": 0.25816208452692024,
      "learning_rate": 4.774575140626317e-07,
      "loss": 0.5452,
      "step": 408
    },
    {
      "epoch": 1.6360000000000001,
      "grad_norm": 0.2583681473741294,
      "learning_rate": 4.6746898954181217e-07,
      "loss": 0.5587,
      "step": 409
    },
    {
      "epoch": 1.6400000000000001,
      "grad_norm": 0.2531012017534651,
      "learning_rate": 4.5757526747389506e-07,
      "loss": 0.5641,
      "step": 410
    },
    {
      "epoch": 1.6440000000000001,
      "grad_norm": 0.260155681123512,
      "learning_rate": 4.4777680932742124e-07,
      "loss": 0.5481,
      "step": 411
    },
    {
      "epoch": 1.6480000000000001,
      "grad_norm": 0.2683235093149167,
      "learning_rate": 4.380740721275786e-07,
      "loss": 0.5572,
      "step": 412
    },
    {
      "epoch": 1.6520000000000001,
      "grad_norm": 0.2564907831166984,
      "learning_rate": 4.2846750843488523e-07,
      "loss": 0.5416,
      "step": 413
    },
    {
      "epoch": 1.6560000000000001,
      "grad_norm": 0.25696609634112416,
      "learning_rate": 4.189575663240794e-07,
      "loss": 0.5549,
      "step": 414
    },
    {
      "epoch": 1.6600000000000001,
      "grad_norm": 0.2578454982930163,
      "learning_rate": 4.095446893632235e-07,
      "loss": 0.5399,
      "step": 415
    },
    {
      "epoch": 1.6640000000000001,
      "grad_norm": 0.25813226813140566,
      "learning_rate": 4.002293165930088e-07,
      "loss": 0.5585,
      "step": 416
    },
    {
      "epoch": 1.6680000000000001,
      "grad_norm": 0.25934385596480936,
      "learning_rate": 3.9101188250628557e-07,
      "loss": 0.5526,
      "step": 417
    },
    {
      "epoch": 1.6720000000000002,
      "grad_norm": 0.2617210666443342,
      "learning_rate": 3.818928170277911e-07,
      "loss": 0.5572,
      "step": 418
    },
    {
      "epoch": 1.6760000000000002,
      "grad_norm": 0.2571918169332575,
      "learning_rate": 3.7287254549409846e-07,
      "loss": 0.5581,
      "step": 419
    },
    {
      "epoch": 1.6800000000000002,
      "grad_norm": 0.26192520360464283,
      "learning_rate": 3.639514886337786e-07,
      "loss": 0.5532,
      "step": 420
    },
    {
      "epoch": 1.6840000000000002,
      "grad_norm": 0.260902445787394,
      "learning_rate": 3.551300625477763e-07,
      "loss": 0.5572,
      "step": 421
    },
    {
      "epoch": 1.688,
      "grad_norm": 0.2694942892729709,
      "learning_rate": 3.4640867869000036e-07,
      "loss": 0.5567,
      "step": 422
    },
    {
      "epoch": 1.692,
      "grad_norm": 0.266015788224457,
      "learning_rate": 3.3778774384813555e-07,
      "loss": 0.5601,
      "step": 423
    },
    {
      "epoch": 1.696,
      "grad_norm": 0.25988296706658587,
      "learning_rate": 3.292676601246661e-07,
      "loss": 0.5514,
      "step": 424
    },
    {
      "epoch": 1.7,
      "grad_norm": 0.27277792689483177,
      "learning_rate": 3.208488249181216e-07,
      "loss": 0.5454,
      "step": 425
    },
    {
      "epoch": 1.704,
      "grad_norm": 0.2765149685595683,
      "learning_rate": 3.125316309045434e-07,
      "loss": 0.5782,
      "step": 426
    },
    {
      "epoch": 1.708,
      "grad_norm": 0.2672590762478569,
      "learning_rate": 3.0431646601916556e-07,
      "loss": 0.5481,
      "step": 427
    },
    {
      "epoch": 1.712,
      "grad_norm": 0.2469424719148706,
      "learning_rate": 2.962037134383211e-07,
      "loss": 0.5248,
      "step": 428
    },
    {
      "epoch": 1.716,
      "grad_norm": 0.25442988775386594,
      "learning_rate": 2.881937515615732e-07,
      "loss": 0.5401,
      "step": 429
    },
    {
      "epoch": 1.72,
      "grad_norm": 0.25371799199659134,
      "learning_rate": 2.80286953994062e-07,
      "loss": 0.5361,
      "step": 430
    },
    {
      "epoch": 1.724,
      "grad_norm": 0.2589009719913934,
      "learning_rate": 2.7248368952908055e-07,
      "loss": 0.5675,
      "step": 431
    },
    {
      "epoch": 1.728,
      "grad_norm": 0.2512096194587606,
      "learning_rate": 2.647843221308721e-07,
      "loss": 0.5614,
      "step": 432
    },
    {
      "epoch": 1.732,
      "grad_norm": 0.25846242566159466,
      "learning_rate": 2.571892109176552e-07,
      "loss": 0.5419,
      "step": 433
    },
    {
      "epoch": 1.736,
      "grad_norm": 0.2546107738468481,
      "learning_rate": 2.496987101448728e-07,
      "loss": 0.5434,
      "step": 434
    },
    {
      "epoch": 1.74,
      "grad_norm": 0.25610395036355743,
      "learning_rate": 2.423131691886682e-07,
      "loss": 0.5527,
      "step": 435
    },
    {
      "epoch": 1.744,
      "grad_norm": 0.2538387917642117,
      "learning_rate": 2.3503293252959136e-07,
      "loss": 0.5561,
      "step": 436
    },
    {
      "epoch": 1.748,
      "grad_norm": 0.2584543393061303,
      "learning_rate": 2.278583397365286e-07,
      "loss": 0.5499,
      "step": 437
    },
    {
      "epoch": 1.752,
      "grad_norm": 0.2529672816521981,
      "learning_rate": 2.2078972545086647e-07,
      "loss": 0.5567,
      "step": 438
    },
    {
      "epoch": 1.756,
      "grad_norm": 0.25666575194151114,
      "learning_rate": 2.138274193708828e-07,
      "loss": 0.5765,
      "step": 439
    },
    {
      "epoch": 1.76,
      "grad_norm": 0.2579853829854033,
      "learning_rate": 2.0697174623636795e-07,
      "loss": 0.5782,
      "step": 440
    },
    {
      "epoch": 1.764,
      "grad_norm": 0.24941376249415378,
      "learning_rate": 2.002230258134777e-07,
      "loss": 0.5485,
      "step": 441
    },
    {
      "epoch": 1.768,
      "grad_norm": 0.25532064995305964,
      "learning_rate": 1.9358157287982099e-07,
      "loss": 0.5581,
      "step": 442
    },
    {
      "epoch": 1.772,
      "grad_norm": 0.258850758081898,
      "learning_rate": 1.8704769720977484e-07,
      "loss": 0.5447,
      "step": 443
    },
    {
      "epoch": 1.776,
      "grad_norm": 0.2510020854350655,
      "learning_rate": 1.8062170356003854e-07,
      "loss": 0.5494,
      "step": 444
    },
    {
      "epoch": 1.78,
      "grad_norm": 0.2611883344797634,
      "learning_rate": 1.743038916554171e-07,
      "loss": 0.5502,
      "step": 445
    },
    {
      "epoch": 1.784,
      "grad_norm": 0.26036595606020646,
      "learning_rate": 1.680945561748412e-07,
      "loss": 0.5454,
      "step": 446
    },
    {
      "epoch": 1.788,
      "grad_norm": 0.25246138722084965,
      "learning_rate": 1.6199398673762328e-07,
      "loss": 0.5531,
      "step": 447
    },
    {
      "epoch": 1.792,
      "grad_norm": 0.25741307398655594,
      "learning_rate": 1.5600246788994938e-07,
      "loss": 0.5394,
      "step": 448
    },
    {
      "epoch": 1.796,
      "grad_norm": 0.2588772149571607,
      "learning_rate": 1.5012027909160675e-07,
      "loss": 0.5552,
      "step": 449
    },
    {
      "epoch": 1.8,
      "grad_norm": 0.2617888726984671,
      "learning_rate": 1.44347694702949e-07,
      "loss": 0.5538,
      "step": 450
    },
    {
      "epoch": 1.804,
      "grad_norm": 0.2750715668908645,
      "learning_rate": 1.3868498397209884e-07,
      "loss": 0.5541,
      "step": 451
    },
    {
      "epoch": 1.808,
      "grad_norm": 0.26245053284364556,
      "learning_rate": 1.3313241102239056e-07,
      "loss": 0.5421,
      "step": 452
    },
    {
      "epoch": 1.812,
      "grad_norm": 0.26161552294885054,
      "learning_rate": 1.2769023484005049e-07,
      "loss": 0.5561,
      "step": 453
    },
    {
      "epoch": 1.8159999999999998,
      "grad_norm": 0.2623120144249456,
      "learning_rate": 1.223587092621162e-07,
      "loss": 0.5625,
      "step": 454
    },
    {
      "epoch": 1.8199999999999998,
      "grad_norm": 0.2591087788623649,
      "learning_rate": 1.1713808296459794e-07,
      "loss": 0.554,
      "step": 455
    },
    {
      "epoch": 1.8239999999999998,
      "grad_norm": 0.268216634704575,
      "learning_rate": 1.120285994508799e-07,
      "loss": 0.55,
      "step": 456
    },
    {
      "epoch": 1.8279999999999998,
      "grad_norm": 0.24896430869123282,
      "learning_rate": 1.0703049704036128e-07,
      "loss": 0.5433,
      "step": 457
    },
    {
      "epoch": 1.8319999999999999,
      "grad_norm": 0.26570965421595016,
      "learning_rate": 1.0214400885734194e-07,
      "loss": 0.5428,
      "step": 458
    },
    {
      "epoch": 1.8359999999999999,
      "grad_norm": 0.25167063380283183,
      "learning_rate": 9.73693628201483e-08,
      "loss": 0.5607,
      "step": 459
    },
    {
      "epoch": 1.8399999999999999,
      "grad_norm": 0.25822111998602776,
      "learning_rate": 9.270678163050218e-08,
      "loss": 0.5518,
      "step": 460
    },
    {
      "epoch": 1.8439999999999999,
      "grad_norm": 0.2633665190460336,
      "learning_rate": 8.815648276313343e-08,
      "loss": 0.5485,
      "step": 461
    },
    {
      "epoch": 1.8479999999999999,
      "grad_norm": 0.26717689797937383,
      "learning_rate": 8.371867845563819e-08,
      "loss": 0.5253,
      "step": 462
    },
    {
      "epoch": 1.8519999999999999,
      "grad_norm": 0.2639494215386647,
      "learning_rate": 7.939357569857697e-08,
      "loss": 0.5423,
      "step": 463
    },
    {
      "epoch": 1.8559999999999999,
      "grad_norm": 0.25843197317699923,
      "learning_rate": 7.518137622582189e-08,
      "loss": 0.5668,
      "step": 464
    },
    {
      "epoch": 1.8599999999999999,
      "grad_norm": 0.2534754343376283,
      "learning_rate": 7.108227650514637e-08,
      "loss": 0.5419,
      "step": 465
    },
    {
      "epoch": 1.8639999999999999,
      "grad_norm": 0.2506776201660858,
      "learning_rate": 6.70964677290617e-08,
      "loss": 0.5496,
      "step": 466
    },
    {
      "epoch": 1.8679999999999999,
      "grad_norm": 0.28500846055614565,
      "learning_rate": 6.322413580589881e-08,
      "loss": 0.5371,
      "step": 467
    },
    {
      "epoch": 1.8719999999999999,
      "grad_norm": 0.2558787240227369,
      "learning_rate": 5.946546135113862e-08,
      "loss": 0.5521,
      "step": 468
    },
    {
      "epoch": 1.876,
      "grad_norm": 0.24995974502835572,
      "learning_rate": 5.5820619678985947e-08,
      "loss": 0.5569,
      "step": 469
    },
    {
      "epoch": 1.88,
      "grad_norm": 0.2558205709053614,
      "learning_rate": 5.2289780794192726e-08,
      "loss": 0.5472,
      "step": 470
    },
    {
      "epoch": 1.884,
      "grad_norm": 0.2614512604532457,
      "learning_rate": 4.8873109384129927e-08,
      "loss": 0.5392,
      "step": 471
    },
    {
      "epoch": 1.888,
      "grad_norm": 0.2611020235601397,
      "learning_rate": 4.557076481110367e-08,
      "loss": 0.5632,
      "step": 472
    },
    {
      "epoch": 1.892,
      "grad_norm": 0.2555589904873057,
      "learning_rate": 4.238290110492427e-08,
      "loss": 0.5445,
      "step": 473
    },
    {
      "epoch": 1.896,
      "grad_norm": 0.2590925947063543,
      "learning_rate": 3.930966695572136e-08,
      "loss": 0.5511,
      "step": 474
    },
    {
      "epoch": 1.9,
      "grad_norm": 0.25584381954615376,
      "learning_rate": 3.635120570700784e-08,
      "loss": 0.5554,
      "step": 475
    },
    {
      "epoch": 1.904,
      "grad_norm": 0.2598744096425038,
      "learning_rate": 3.3507655348995194e-08,
      "loss": 0.5449,
      "step": 476
    },
    {
      "epoch": 1.908,
      "grad_norm": 0.2638610026594681,
      "learning_rate": 3.077914851215585e-08,
      "loss": 0.5591,
      "step": 477
    },
    {
      "epoch": 1.912,
      "grad_norm": 0.25151758051160417,
      "learning_rate": 2.8165812461038166e-08,
      "loss": 0.5499,
      "step": 478
    },
    {
      "epoch": 1.916,
      "grad_norm": 0.254508630104069,
      "learning_rate": 2.566776908832974e-08,
      "loss": 0.5374,
      "step": 479
    },
    {
      "epoch": 1.92,
      "grad_norm": 0.2575393452132042,
      "learning_rate": 2.3285134909173113e-08,
      "loss": 0.5354,
      "step": 480
    },
    {
      "epoch": 1.924,
      "grad_norm": 0.25136797567604285,
      "learning_rate": 2.1018021055729256e-08,
      "loss": 0.5525,
      "step": 481
    },
    {
      "epoch": 1.928,
      "grad_norm": 0.26449210060783246,
      "learning_rate": 1.886653327199617e-08,
      "loss": 0.5486,
      "step": 482
    },
    {
      "epoch": 1.932,
      "grad_norm": 0.24948013880023046,
      "learning_rate": 1.6830771908875054e-08,
      "loss": 0.5547,
      "step": 483
    },
    {
      "epoch": 1.936,
      "grad_norm": 0.2630268839649341,
      "learning_rate": 1.4910831919490997e-08,
      "loss": 0.5601,
      "step": 484
    },
    {
      "epoch": 1.94,
      "grad_norm": 0.2605899821024887,
      "learning_rate": 1.3106802854762901e-08,
      "loss": 0.565,
      "step": 485
    },
    {
      "epoch": 1.944,
      "grad_norm": 0.24702510074753287,
      "learning_rate": 1.1418768859227935e-08,
      "loss": 0.5519,
      "step": 486
    },
    {
      "epoch": 1.948,
      "grad_norm": 0.2513877665742416,
      "learning_rate": 9.846808667115515e-09,
      "loss": 0.5247,
      "step": 487
    },
    {
      "epoch": 1.952,
      "grad_norm": 0.2564857438108711,
      "learning_rate": 8.390995598676067e-09,
      "loss": 0.5545,
      "step": 488
    },
    {
      "epoch": 1.956,
      "grad_norm": 0.2547711549147179,
      "learning_rate": 7.051397556760153e-09,
      "loss": 0.5592,
      "step": 489
    },
    {
      "epoch": 1.96,
      "grad_norm": 0.2585526225777786,
      "learning_rate": 5.828077023651846e-09,
      "loss": 0.5279,
      "step": 490
    },
    {
      "epoch": 1.964,
      "grad_norm": 0.25380028383316505,
      "learning_rate": 4.721091058154936e-09,
      "loss": 0.5604,
      "step": 491
    },
    {
      "epoch": 1.968,
      "grad_norm": 0.2610939223735134,
      "learning_rate": 3.730491292930072e-09,
      "loss": 0.5623,
      "step": 492
    },
    {
      "epoch": 1.972,
      "grad_norm": 0.2586724319177905,
      "learning_rate": 2.8563239320880676e-09,
      "loss": 0.5502,
      "step": 493
    },
    {
      "epoch": 1.976,
      "grad_norm": 0.2607818153868037,
      "learning_rate": 2.0986297490338536e-09,
      "loss": 0.553,
      "step": 494
    },
    {
      "epoch": 1.98,
      "grad_norm": 0.25584510179263337,
      "learning_rate": 1.4574440845649406e-09,
      "loss": 0.5633,
      "step": 495
    },
    {
      "epoch": 1.984,
      "grad_norm": 0.2633673860402372,
      "learning_rate": 9.32796845223294e-10,
      "loss": 0.5659,
      "step": 496
    },
    {
      "epoch": 1.988,
      "grad_norm": 0.2656283591233252,
      "learning_rate": 5.247125019003374e-10,
      "loss": 0.5562,
      "step": 497
    },
    {
      "epoch": 1.992,
      "grad_norm": 0.2572974511141831,
      "learning_rate": 2.3321008869481296e-10,
      "loss": 0.5575,
      "step": 498
    },
    {
      "epoch": 1.996,
      "grad_norm": 0.2611191928635677,
      "learning_rate": 5.830320202654394e-11,
      "loss": 0.5552,
      "step": 499
    },
    {
      "epoch": 2.0,
      "grad_norm": 0.2598503292836428,
      "learning_rate": 0.0,
      "loss": 0.565,
      "step": 500
    }
  ],
  "logging_steps": 1,
  "max_steps": 500,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 2,
  "save_steps": 125,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": true
      },
      "attributes": {}
    }
  },
  "total_flos": 6.140491770384876e+17,
  "train_batch_size": 1,
  "trial_name": null,
  "trial_params": null
}