XCoder-Complexity-Scorer / trainer_state.json
banksy235's picture
Upload folder using huggingface_hub
1fe30f5 verified
raw
history blame
149 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9994672349493873,
"eval_steps": 500,
"global_step": 938,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0,
"grad_norm": 0.0,
"learning_rate": 0.0,
"loss": 4.4688,
"step": 1
},
{
"epoch": 0.0,
"grad_norm": 0.0,
"learning_rate": 0.0,
"loss": 5.7051,
"step": 2
},
{
"epoch": 0.0,
"grad_norm": 0.0,
"learning_rate": 0.0,
"loss": 4.6101,
"step": 3
},
{
"epoch": 0.0,
"grad_norm": 0.0,
"learning_rate": 0.0,
"loss": 5.1161,
"step": 4
},
{
"epoch": 0.01,
"grad_norm": 0.0,
"learning_rate": 0.0,
"loss": 4.7314,
"step": 5
},
{
"epoch": 0.01,
"grad_norm": 0.0,
"learning_rate": 0.0,
"loss": 5.4174,
"step": 6
},
{
"epoch": 0.01,
"grad_norm": 0.0,
"learning_rate": 0.0,
"loss": 5.3351,
"step": 7
},
{
"epoch": 0.01,
"grad_norm": 267.06263341170535,
"learning_rate": 2.1276595744680852e-07,
"loss": 4.3761,
"step": 8
},
{
"epoch": 0.01,
"grad_norm": 256.6399066161345,
"learning_rate": 4.2553191489361704e-07,
"loss": 4.911,
"step": 9
},
{
"epoch": 0.01,
"grad_norm": 259.8979037908876,
"learning_rate": 6.382978723404255e-07,
"loss": 4.5067,
"step": 10
},
{
"epoch": 0.01,
"grad_norm": 206.34210154966675,
"learning_rate": 8.510638297872341e-07,
"loss": 4.1617,
"step": 11
},
{
"epoch": 0.01,
"grad_norm": 206.34210154966675,
"learning_rate": 8.510638297872341e-07,
"loss": 4.7024,
"step": 12
},
{
"epoch": 0.01,
"grad_norm": 104.25967551689651,
"learning_rate": 1.0638297872340427e-06,
"loss": 2.8783,
"step": 13
},
{
"epoch": 0.01,
"grad_norm": 104.25967551689651,
"learning_rate": 1.0638297872340427e-06,
"loss": 3.6247,
"step": 14
},
{
"epoch": 0.02,
"grad_norm": 301.56418557125875,
"learning_rate": 1.276595744680851e-06,
"loss": 3.8274,
"step": 15
},
{
"epoch": 0.02,
"grad_norm": 158.0102677208041,
"learning_rate": 1.4893617021276596e-06,
"loss": 2.5054,
"step": 16
},
{
"epoch": 0.02,
"grad_norm": 144.94261605863426,
"learning_rate": 1.7021276595744682e-06,
"loss": 2.6479,
"step": 17
},
{
"epoch": 0.02,
"grad_norm": 77.83778037631348,
"learning_rate": 1.9148936170212767e-06,
"loss": 1.7461,
"step": 18
},
{
"epoch": 0.02,
"grad_norm": 110.09097679168819,
"learning_rate": 2.1276595744680853e-06,
"loss": 2.005,
"step": 19
},
{
"epoch": 0.02,
"grad_norm": 64.67189678957659,
"learning_rate": 2.340425531914894e-06,
"loss": 1.5011,
"step": 20
},
{
"epoch": 0.02,
"grad_norm": 85.70086135587184,
"learning_rate": 2.553191489361702e-06,
"loss": 1.4231,
"step": 21
},
{
"epoch": 0.02,
"grad_norm": 51.55499709016579,
"learning_rate": 2.765957446808511e-06,
"loss": 1.3443,
"step": 22
},
{
"epoch": 0.02,
"grad_norm": 39.458031493922036,
"learning_rate": 2.978723404255319e-06,
"loss": 1.1782,
"step": 23
},
{
"epoch": 0.03,
"grad_norm": 30.718044498953375,
"learning_rate": 3.191489361702128e-06,
"loss": 1.0572,
"step": 24
},
{
"epoch": 0.03,
"grad_norm": 54.03307642588333,
"learning_rate": 3.4042553191489363e-06,
"loss": 1.0557,
"step": 25
},
{
"epoch": 0.03,
"grad_norm": 45.15061216435745,
"learning_rate": 3.6170212765957453e-06,
"loss": 0.9468,
"step": 26
},
{
"epoch": 0.03,
"grad_norm": 85.20725035566016,
"learning_rate": 3.8297872340425535e-06,
"loss": 1.1611,
"step": 27
},
{
"epoch": 0.03,
"grad_norm": 61.03528945297965,
"learning_rate": 4.042553191489362e-06,
"loss": 0.9764,
"step": 28
},
{
"epoch": 0.03,
"grad_norm": 32.57984313596529,
"learning_rate": 4.255319148936171e-06,
"loss": 0.9542,
"step": 29
},
{
"epoch": 0.03,
"grad_norm": 46.71878018713139,
"learning_rate": 4.468085106382979e-06,
"loss": 1.0575,
"step": 30
},
{
"epoch": 0.03,
"grad_norm": 28.228382153829948,
"learning_rate": 4.680851063829788e-06,
"loss": 0.8927,
"step": 31
},
{
"epoch": 0.03,
"grad_norm": 32.305084533281175,
"learning_rate": 4.893617021276596e-06,
"loss": 0.9158,
"step": 32
},
{
"epoch": 0.04,
"grad_norm": 38.80822759339413,
"learning_rate": 5.106382978723404e-06,
"loss": 1.0127,
"step": 33
},
{
"epoch": 0.04,
"grad_norm": 16.765291089343936,
"learning_rate": 5.319148936170213e-06,
"loss": 0.8579,
"step": 34
},
{
"epoch": 0.04,
"grad_norm": 46.21887734855352,
"learning_rate": 5.531914893617022e-06,
"loss": 1.0595,
"step": 35
},
{
"epoch": 0.04,
"grad_norm": 33.51542422627954,
"learning_rate": 5.744680851063831e-06,
"loss": 0.7873,
"step": 36
},
{
"epoch": 0.04,
"grad_norm": 20.62520698421828,
"learning_rate": 5.957446808510638e-06,
"loss": 0.7463,
"step": 37
},
{
"epoch": 0.04,
"grad_norm": 32.86540760673144,
"learning_rate": 6.170212765957447e-06,
"loss": 0.9393,
"step": 38
},
{
"epoch": 0.04,
"grad_norm": 16.78339589062748,
"learning_rate": 6.382978723404256e-06,
"loss": 0.7036,
"step": 39
},
{
"epoch": 0.04,
"grad_norm": 20.638888081681362,
"learning_rate": 6.595744680851064e-06,
"loss": 0.8178,
"step": 40
},
{
"epoch": 0.04,
"grad_norm": 24.957942318493615,
"learning_rate": 6.808510638297873e-06,
"loss": 0.8192,
"step": 41
},
{
"epoch": 0.04,
"grad_norm": 18.753926303171866,
"learning_rate": 7.021276595744682e-06,
"loss": 0.8053,
"step": 42
},
{
"epoch": 0.05,
"grad_norm": 20.380738488767815,
"learning_rate": 7.234042553191491e-06,
"loss": 0.7775,
"step": 43
},
{
"epoch": 0.05,
"grad_norm": 15.358291004654351,
"learning_rate": 7.446808510638298e-06,
"loss": 0.768,
"step": 44
},
{
"epoch": 0.05,
"grad_norm": 28.01853531392082,
"learning_rate": 7.659574468085107e-06,
"loss": 0.6495,
"step": 45
},
{
"epoch": 0.05,
"grad_norm": 90.12246442870764,
"learning_rate": 7.872340425531916e-06,
"loss": 1.1254,
"step": 46
},
{
"epoch": 0.05,
"grad_norm": 24.048650390369733,
"learning_rate": 8.085106382978723e-06,
"loss": 0.8769,
"step": 47
},
{
"epoch": 0.05,
"grad_norm": 79.51846666706089,
"learning_rate": 8.297872340425532e-06,
"loss": 1.2554,
"step": 48
},
{
"epoch": 0.05,
"grad_norm": 20.341370812622667,
"learning_rate": 8.510638297872341e-06,
"loss": 0.7153,
"step": 49
},
{
"epoch": 0.05,
"grad_norm": 48.50233018085147,
"learning_rate": 8.72340425531915e-06,
"loss": 1.0358,
"step": 50
},
{
"epoch": 0.05,
"grad_norm": 27.71805876780665,
"learning_rate": 8.936170212765958e-06,
"loss": 0.7539,
"step": 51
},
{
"epoch": 0.06,
"grad_norm": 44.56567439721268,
"learning_rate": 9.148936170212767e-06,
"loss": 1.0363,
"step": 52
},
{
"epoch": 0.06,
"grad_norm": 43.96614589042312,
"learning_rate": 9.361702127659576e-06,
"loss": 1.0823,
"step": 53
},
{
"epoch": 0.06,
"grad_norm": 36.336438544851525,
"learning_rate": 9.574468085106385e-06,
"loss": 0.9189,
"step": 54
},
{
"epoch": 0.06,
"grad_norm": 20.360154033813146,
"learning_rate": 9.787234042553192e-06,
"loss": 0.7138,
"step": 55
},
{
"epoch": 0.06,
"grad_norm": 16.415639113466074,
"learning_rate": 1e-05,
"loss": 0.8165,
"step": 56
},
{
"epoch": 0.06,
"grad_norm": 17.154150115372023,
"learning_rate": 1.0212765957446808e-05,
"loss": 0.8003,
"step": 57
},
{
"epoch": 0.06,
"grad_norm": 23.485978008365016,
"learning_rate": 1.0425531914893619e-05,
"loss": 0.8017,
"step": 58
},
{
"epoch": 0.06,
"grad_norm": 18.921319860871982,
"learning_rate": 1.0638297872340426e-05,
"loss": 0.6803,
"step": 59
},
{
"epoch": 0.06,
"grad_norm": 28.458932361446607,
"learning_rate": 1.0851063829787233e-05,
"loss": 0.9205,
"step": 60
},
{
"epoch": 0.06,
"grad_norm": 16.224593364238817,
"learning_rate": 1.1063829787234044e-05,
"loss": 0.7286,
"step": 61
},
{
"epoch": 0.07,
"grad_norm": 30.78498419558807,
"learning_rate": 1.1276595744680851e-05,
"loss": 0.9564,
"step": 62
},
{
"epoch": 0.07,
"grad_norm": 20.29389328745947,
"learning_rate": 1.1489361702127662e-05,
"loss": 0.8442,
"step": 63
},
{
"epoch": 0.07,
"grad_norm": 18.23986171950765,
"learning_rate": 1.170212765957447e-05,
"loss": 0.7971,
"step": 64
},
{
"epoch": 0.07,
"grad_norm": 13.607742168916245,
"learning_rate": 1.1914893617021277e-05,
"loss": 0.7479,
"step": 65
},
{
"epoch": 0.07,
"grad_norm": 15.589517783052612,
"learning_rate": 1.2127659574468087e-05,
"loss": 0.6971,
"step": 66
},
{
"epoch": 0.07,
"grad_norm": 11.730568689593209,
"learning_rate": 1.2340425531914895e-05,
"loss": 0.6136,
"step": 67
},
{
"epoch": 0.07,
"grad_norm": 17.919191925814953,
"learning_rate": 1.2553191489361702e-05,
"loss": 0.6786,
"step": 68
},
{
"epoch": 0.07,
"grad_norm": 46.93984753189794,
"learning_rate": 1.2765957446808513e-05,
"loss": 1.1316,
"step": 69
},
{
"epoch": 0.07,
"grad_norm": 32.097199934080784,
"learning_rate": 1.297872340425532e-05,
"loss": 0.8888,
"step": 70
},
{
"epoch": 0.08,
"grad_norm": 23.37535729938417,
"learning_rate": 1.3191489361702127e-05,
"loss": 0.7891,
"step": 71
},
{
"epoch": 0.08,
"grad_norm": 16.127458600501658,
"learning_rate": 1.3404255319148938e-05,
"loss": 0.5976,
"step": 72
},
{
"epoch": 0.08,
"grad_norm": 13.056670968499086,
"learning_rate": 1.3617021276595745e-05,
"loss": 0.6602,
"step": 73
},
{
"epoch": 0.08,
"grad_norm": 18.80577219024569,
"learning_rate": 1.3829787234042556e-05,
"loss": 0.7343,
"step": 74
},
{
"epoch": 0.08,
"grad_norm": 18.0606093200878,
"learning_rate": 1.4042553191489363e-05,
"loss": 0.7229,
"step": 75
},
{
"epoch": 0.08,
"grad_norm": 13.62137898397275,
"learning_rate": 1.425531914893617e-05,
"loss": 0.7897,
"step": 76
},
{
"epoch": 0.08,
"grad_norm": 10.783139067640777,
"learning_rate": 1.4468085106382981e-05,
"loss": 0.5573,
"step": 77
},
{
"epoch": 0.08,
"grad_norm": 28.888686230080594,
"learning_rate": 1.4680851063829789e-05,
"loss": 0.7468,
"step": 78
},
{
"epoch": 0.08,
"grad_norm": 24.696181606034965,
"learning_rate": 1.4893617021276596e-05,
"loss": 0.6625,
"step": 79
},
{
"epoch": 0.09,
"grad_norm": 32.68489697451064,
"learning_rate": 1.5106382978723407e-05,
"loss": 1.0193,
"step": 80
},
{
"epoch": 0.09,
"grad_norm": 20.801300816024295,
"learning_rate": 1.5319148936170214e-05,
"loss": 0.6352,
"step": 81
},
{
"epoch": 0.09,
"grad_norm": 10.55513433584252,
"learning_rate": 1.5531914893617023e-05,
"loss": 0.7007,
"step": 82
},
{
"epoch": 0.09,
"grad_norm": 15.716140126443527,
"learning_rate": 1.5744680851063832e-05,
"loss": 0.7187,
"step": 83
},
{
"epoch": 0.09,
"grad_norm": 21.85726232694579,
"learning_rate": 1.595744680851064e-05,
"loss": 0.7349,
"step": 84
},
{
"epoch": 0.09,
"grad_norm": 20.964587712026038,
"learning_rate": 1.6170212765957446e-05,
"loss": 0.7143,
"step": 85
},
{
"epoch": 0.09,
"grad_norm": 8.250030396732411,
"learning_rate": 1.6382978723404255e-05,
"loss": 0.6892,
"step": 86
},
{
"epoch": 0.09,
"grad_norm": 21.20006063299826,
"learning_rate": 1.6595744680851064e-05,
"loss": 0.8109,
"step": 87
},
{
"epoch": 0.09,
"grad_norm": 19.09608912821583,
"learning_rate": 1.6808510638297873e-05,
"loss": 0.8082,
"step": 88
},
{
"epoch": 0.09,
"grad_norm": 20.52502129407863,
"learning_rate": 1.7021276595744682e-05,
"loss": 0.6783,
"step": 89
},
{
"epoch": 0.1,
"grad_norm": 39.20021747253875,
"learning_rate": 1.723404255319149e-05,
"loss": 0.9711,
"step": 90
},
{
"epoch": 0.1,
"grad_norm": 31.095085040142717,
"learning_rate": 1.74468085106383e-05,
"loss": 0.8283,
"step": 91
},
{
"epoch": 0.1,
"grad_norm": 37.26207899740689,
"learning_rate": 1.765957446808511e-05,
"loss": 1.0039,
"step": 92
},
{
"epoch": 0.1,
"grad_norm": 16.273798667976465,
"learning_rate": 1.7872340425531915e-05,
"loss": 0.6814,
"step": 93
},
{
"epoch": 0.1,
"grad_norm": 32.374520273410155,
"learning_rate": 1.8085106382978724e-05,
"loss": 0.9674,
"step": 94
},
{
"epoch": 0.1,
"grad_norm": 34.748891951713134,
"learning_rate": 1.8297872340425533e-05,
"loss": 1.1996,
"step": 95
},
{
"epoch": 0.1,
"grad_norm": 26.013510708052237,
"learning_rate": 1.8510638297872342e-05,
"loss": 0.9372,
"step": 96
},
{
"epoch": 0.1,
"grad_norm": 19.99677186530767,
"learning_rate": 1.872340425531915e-05,
"loss": 0.8718,
"step": 97
},
{
"epoch": 0.1,
"grad_norm": 16.62199424485097,
"learning_rate": 1.893617021276596e-05,
"loss": 0.9384,
"step": 98
},
{
"epoch": 0.11,
"grad_norm": 20.111532243609883,
"learning_rate": 1.914893617021277e-05,
"loss": 0.9813,
"step": 99
},
{
"epoch": 0.11,
"grad_norm": 15.191733826484391,
"learning_rate": 1.9361702127659575e-05,
"loss": 0.8069,
"step": 100
},
{
"epoch": 0.11,
"grad_norm": 23.926673285223416,
"learning_rate": 1.9574468085106384e-05,
"loss": 0.8308,
"step": 101
},
{
"epoch": 0.11,
"grad_norm": 23.218285916806728,
"learning_rate": 1.9787234042553193e-05,
"loss": 0.9754,
"step": 102
},
{
"epoch": 0.11,
"grad_norm": 17.76104315857583,
"learning_rate": 2e-05,
"loss": 0.8006,
"step": 103
},
{
"epoch": 0.11,
"grad_norm": 20.836820707421783,
"learning_rate": 1.9999930723752516e-05,
"loss": 0.8543,
"step": 104
},
{
"epoch": 0.11,
"grad_norm": 17.0438780130028,
"learning_rate": 1.9999722895969904e-05,
"loss": 0.8443,
"step": 105
},
{
"epoch": 0.11,
"grad_norm": 22.36760800612908,
"learning_rate": 1.9999376519531668e-05,
"loss": 0.8841,
"step": 106
},
{
"epoch": 0.11,
"grad_norm": 11.874378681555248,
"learning_rate": 1.999889159923694e-05,
"loss": 0.6501,
"step": 107
},
{
"epoch": 0.12,
"grad_norm": 17.245286092714633,
"learning_rate": 1.9998268141804412e-05,
"loss": 0.8119,
"step": 108
},
{
"epoch": 0.12,
"grad_norm": 16.693402688517093,
"learning_rate": 1.9997506155872246e-05,
"loss": 0.7683,
"step": 109
},
{
"epoch": 0.12,
"grad_norm": 18.100428165378997,
"learning_rate": 1.999660565199794e-05,
"loss": 0.8259,
"step": 110
},
{
"epoch": 0.12,
"grad_norm": 10.547692312079072,
"learning_rate": 1.9995566642658208e-05,
"loss": 0.6651,
"step": 111
},
{
"epoch": 0.12,
"grad_norm": 10.19017430503209,
"learning_rate": 1.9994389142248775e-05,
"loss": 0.7619,
"step": 112
},
{
"epoch": 0.12,
"grad_norm": 13.012636488876366,
"learning_rate": 1.999307316708421e-05,
"loss": 0.6931,
"step": 113
},
{
"epoch": 0.12,
"grad_norm": 13.672739554862506,
"learning_rate": 1.9991618735397674e-05,
"loss": 0.7014,
"step": 114
},
{
"epoch": 0.12,
"grad_norm": 11.268690129927412,
"learning_rate": 1.9990025867340683e-05,
"loss": 0.7041,
"step": 115
},
{
"epoch": 0.12,
"grad_norm": 13.835706135922516,
"learning_rate": 1.998829458498282e-05,
"loss": 0.7568,
"step": 116
},
{
"epoch": 0.12,
"grad_norm": 12.580781013528279,
"learning_rate": 1.998642491231143e-05,
"loss": 0.7155,
"step": 117
},
{
"epoch": 0.13,
"grad_norm": 11.88807897059113,
"learning_rate": 1.9984416875231303e-05,
"loss": 0.7374,
"step": 118
},
{
"epoch": 0.13,
"grad_norm": 15.58619342416967,
"learning_rate": 1.9982270501564286e-05,
"loss": 0.8355,
"step": 119
},
{
"epoch": 0.13,
"grad_norm": 14.670349037930592,
"learning_rate": 1.9979985821048925e-05,
"loss": 0.623,
"step": 120
},
{
"epoch": 0.13,
"grad_norm": 17.596509727191997,
"learning_rate": 1.997756286534004e-05,
"loss": 0.7664,
"step": 121
},
{
"epoch": 0.13,
"grad_norm": 16.388736277383142,
"learning_rate": 1.9975001668008282e-05,
"loss": 0.6749,
"step": 122
},
{
"epoch": 0.13,
"grad_norm": 11.99481967344947,
"learning_rate": 1.9972302264539686e-05,
"loss": 0.6724,
"step": 123
},
{
"epoch": 0.13,
"grad_norm": 20.53435074276929,
"learning_rate": 1.9969464692335153e-05,
"loss": 0.7242,
"step": 124
},
{
"epoch": 0.13,
"grad_norm": 11.475488488305645,
"learning_rate": 1.996648899070996e-05,
"loss": 0.7357,
"step": 125
},
{
"epoch": 0.13,
"grad_norm": 25.553564042295182,
"learning_rate": 1.996337520089319e-05,
"loss": 0.7927,
"step": 126
},
{
"epoch": 0.14,
"grad_norm": 8.109394353226076,
"learning_rate": 1.9960123366027187e-05,
"loss": 0.7367,
"step": 127
},
{
"epoch": 0.14,
"grad_norm": 8.254315086779002,
"learning_rate": 1.995673353116692e-05,
"loss": 0.5336,
"step": 128
},
{
"epoch": 0.14,
"grad_norm": 28.75262073669196,
"learning_rate": 1.995320574327941e-05,
"loss": 0.9592,
"step": 129
},
{
"epoch": 0.14,
"grad_norm": 22.09118124557931,
"learning_rate": 1.9949540051243032e-05,
"loss": 0.6935,
"step": 130
},
{
"epoch": 0.14,
"grad_norm": 19.95406582271437,
"learning_rate": 1.9945736505846866e-05,
"loss": 0.8254,
"step": 131
},
{
"epoch": 0.14,
"grad_norm": 13.525596071741782,
"learning_rate": 1.9941795159789983e-05,
"loss": 0.6531,
"step": 132
},
{
"epoch": 0.14,
"grad_norm": 9.891408326838208,
"learning_rate": 1.9937716067680712e-05,
"loss": 0.7162,
"step": 133
},
{
"epoch": 0.14,
"grad_norm": 23.11321488500328,
"learning_rate": 1.9933499286035896e-05,
"loss": 0.7962,
"step": 134
},
{
"epoch": 0.14,
"grad_norm": 14.482317722843773,
"learning_rate": 1.9929144873280092e-05,
"loss": 0.7631,
"step": 135
},
{
"epoch": 0.14,
"grad_norm": 19.49882212964946,
"learning_rate": 1.992465288974478e-05,
"loss": 0.7617,
"step": 136
},
{
"epoch": 0.15,
"grad_norm": 14.59122368824527,
"learning_rate": 1.992002339766751e-05,
"loss": 0.6722,
"step": 137
},
{
"epoch": 0.15,
"grad_norm": 14.253058587354259,
"learning_rate": 1.9915256461191054e-05,
"loss": 0.6869,
"step": 138
},
{
"epoch": 0.15,
"grad_norm": 22.646528058916722,
"learning_rate": 1.99103521463625e-05,
"loss": 0.858,
"step": 139
},
{
"epoch": 0.15,
"grad_norm": 19.599804799046947,
"learning_rate": 1.9905310521132353e-05,
"loss": 0.8475,
"step": 140
},
{
"epoch": 0.15,
"grad_norm": 7.459796818444622,
"learning_rate": 1.9900131655353597e-05,
"loss": 0.6509,
"step": 141
},
{
"epoch": 0.15,
"grad_norm": 14.795548081947178,
"learning_rate": 1.9894815620780705e-05,
"loss": 0.781,
"step": 142
},
{
"epoch": 0.15,
"grad_norm": 19.287512528853632,
"learning_rate": 1.9889362491068658e-05,
"loss": 0.7946,
"step": 143
},
{
"epoch": 0.15,
"grad_norm": 11.780467440280454,
"learning_rate": 1.9883772341771937e-05,
"loss": 0.6937,
"step": 144
},
{
"epoch": 0.15,
"grad_norm": 17.590934140385546,
"learning_rate": 1.9878045250343445e-05,
"loss": 0.8808,
"step": 145
},
{
"epoch": 0.16,
"grad_norm": 16.151381586194,
"learning_rate": 1.9872181296133473e-05,
"loss": 0.8076,
"step": 146
},
{
"epoch": 0.16,
"grad_norm": 10.326349558310788,
"learning_rate": 1.986618056038856e-05,
"loss": 0.8052,
"step": 147
},
{
"epoch": 0.16,
"grad_norm": 9.391393835716272,
"learning_rate": 1.9860043126250407e-05,
"loss": 0.7323,
"step": 148
},
{
"epoch": 0.16,
"grad_norm": 13.654055394779414,
"learning_rate": 1.9853769078754685e-05,
"loss": 0.7861,
"step": 149
},
{
"epoch": 0.16,
"grad_norm": 13.20685189778904,
"learning_rate": 1.9847358504829896e-05,
"loss": 0.7136,
"step": 150
},
{
"epoch": 0.16,
"grad_norm": 8.948428090341855,
"learning_rate": 1.9840811493296134e-05,
"loss": 0.6935,
"step": 151
},
{
"epoch": 0.16,
"grad_norm": 9.441488025541567,
"learning_rate": 1.9834128134863885e-05,
"loss": 0.7222,
"step": 152
},
{
"epoch": 0.16,
"grad_norm": 12.322015322175998,
"learning_rate": 1.982730852213274e-05,
"loss": 0.7606,
"step": 153
},
{
"epoch": 0.16,
"grad_norm": 6.325468800857368,
"learning_rate": 1.982035274959014e-05,
"loss": 0.5735,
"step": 154
},
{
"epoch": 0.17,
"grad_norm": 6.406666981105762,
"learning_rate": 1.9813260913610048e-05,
"loss": 0.5348,
"step": 155
},
{
"epoch": 0.17,
"grad_norm": 17.612437547183127,
"learning_rate": 1.9806033112451616e-05,
"loss": 0.6477,
"step": 156
},
{
"epoch": 0.17,
"grad_norm": 18.78091352132794,
"learning_rate": 1.9798669446257844e-05,
"loss": 0.7878,
"step": 157
},
{
"epoch": 0.17,
"grad_norm": 15.546105512429289,
"learning_rate": 1.979117001705415e-05,
"loss": 0.852,
"step": 158
},
{
"epoch": 0.17,
"grad_norm": 16.923786294030265,
"learning_rate": 1.9783534928747006e-05,
"loss": 0.7754,
"step": 159
},
{
"epoch": 0.17,
"grad_norm": 13.076565312797653,
"learning_rate": 1.9775764287122464e-05,
"loss": 0.7425,
"step": 160
},
{
"epoch": 0.17,
"grad_norm": 14.51221649438122,
"learning_rate": 1.9767858199844697e-05,
"loss": 0.693,
"step": 161
},
{
"epoch": 0.17,
"grad_norm": 16.76943181344085,
"learning_rate": 1.9759816776454527e-05,
"loss": 0.8376,
"step": 162
},
{
"epoch": 0.17,
"grad_norm": 10.230125019828748,
"learning_rate": 1.9751640128367872e-05,
"loss": 0.7238,
"step": 163
},
{
"epoch": 0.17,
"grad_norm": 24.10748192789658,
"learning_rate": 1.9743328368874237e-05,
"loss": 0.8927,
"step": 164
},
{
"epoch": 0.18,
"grad_norm": 24.836153220553506,
"learning_rate": 1.973488161313512e-05,
"loss": 0.8796,
"step": 165
},
{
"epoch": 0.18,
"grad_norm": 15.039263981577998,
"learning_rate": 1.972629997818243e-05,
"loss": 0.7177,
"step": 166
},
{
"epoch": 0.18,
"grad_norm": 7.166311769976143,
"learning_rate": 1.9717583582916862e-05,
"loss": 0.7274,
"step": 167
},
{
"epoch": 0.18,
"grad_norm": 13.36802727443076,
"learning_rate": 1.970873254810625e-05,
"loss": 0.7425,
"step": 168
},
{
"epoch": 0.18,
"grad_norm": 12.26776202136434,
"learning_rate": 1.969974699638388e-05,
"loss": 0.7671,
"step": 169
},
{
"epoch": 0.18,
"grad_norm": 14.871789262668347,
"learning_rate": 1.969062705224682e-05,
"loss": 0.8251,
"step": 170
},
{
"epoch": 0.18,
"grad_norm": 13.01131507667653,
"learning_rate": 1.968137284205417e-05,
"loss": 0.8131,
"step": 171
},
{
"epoch": 0.18,
"grad_norm": 8.692440461337334,
"learning_rate": 1.967198449402532e-05,
"loss": 0.7195,
"step": 172
},
{
"epoch": 0.18,
"grad_norm": 15.307857384021617,
"learning_rate": 1.966246213823818e-05,
"loss": 0.7753,
"step": 173
},
{
"epoch": 0.19,
"grad_norm": 22.444147986646815,
"learning_rate": 1.9652805906627356e-05,
"loss": 0.9884,
"step": 174
},
{
"epoch": 0.19,
"grad_norm": 14.639613594215758,
"learning_rate": 1.9643015932982355e-05,
"loss": 0.7357,
"step": 175
},
{
"epoch": 0.19,
"grad_norm": 10.34217870581945,
"learning_rate": 1.9633092352945698e-05,
"loss": 0.7147,
"step": 176
},
{
"epoch": 0.19,
"grad_norm": 12.490262046586464,
"learning_rate": 1.9623035304011062e-05,
"loss": 0.7384,
"step": 177
},
{
"epoch": 0.19,
"grad_norm": 15.433175693682523,
"learning_rate": 1.9612844925521375e-05,
"loss": 0.7652,
"step": 178
},
{
"epoch": 0.19,
"grad_norm": 6.045080843254174,
"learning_rate": 1.960252135866687e-05,
"loss": 0.8282,
"step": 179
},
{
"epoch": 0.19,
"grad_norm": 8.861252768013037,
"learning_rate": 1.9592064746483143e-05,
"loss": 0.7317,
"step": 180
},
{
"epoch": 0.19,
"grad_norm": 7.8221459133117435,
"learning_rate": 1.9581475233849165e-05,
"loss": 0.5407,
"step": 181
},
{
"epoch": 0.19,
"grad_norm": 12.879262754926978,
"learning_rate": 1.957075296748527e-05,
"loss": 0.6644,
"step": 182
},
{
"epoch": 0.19,
"grad_norm": 5.635984955428702,
"learning_rate": 1.9559898095951137e-05,
"loss": 0.6606,
"step": 183
},
{
"epoch": 0.2,
"grad_norm": 7.056126169759745,
"learning_rate": 1.9548910769643724e-05,
"loss": 0.5217,
"step": 184
},
{
"epoch": 0.2,
"grad_norm": 5.817340227592049,
"learning_rate": 1.953779114079517e-05,
"loss": 0.694,
"step": 185
},
{
"epoch": 0.2,
"grad_norm": 11.240599667838348,
"learning_rate": 1.9526539363470715e-05,
"loss": 0.6613,
"step": 186
},
{
"epoch": 0.2,
"grad_norm": 6.377305729246939,
"learning_rate": 1.9515155593566536e-05,
"loss": 0.6127,
"step": 187
},
{
"epoch": 0.2,
"grad_norm": 12.70631731046457,
"learning_rate": 1.9503639988807605e-05,
"loss": 0.648,
"step": 188
},
{
"epoch": 0.2,
"grad_norm": 7.740556879132694,
"learning_rate": 1.9491992708745502e-05,
"loss": 0.6363,
"step": 189
},
{
"epoch": 0.2,
"grad_norm": 9.838244865368074,
"learning_rate": 1.94802139147562e-05,
"loss": 0.5207,
"step": 190
},
{
"epoch": 0.2,
"grad_norm": 12.292539349351499,
"learning_rate": 1.946830377003782e-05,
"loss": 0.7129,
"step": 191
},
{
"epoch": 0.2,
"grad_norm": 9.434980765550153,
"learning_rate": 1.94562624396084e-05,
"loss": 0.623,
"step": 192
},
{
"epoch": 0.21,
"grad_norm": 17.774682137642376,
"learning_rate": 1.9444090090303567e-05,
"loss": 0.8399,
"step": 193
},
{
"epoch": 0.21,
"grad_norm": 9.97338739760112,
"learning_rate": 1.9431786890774265e-05,
"loss": 0.5893,
"step": 194
},
{
"epoch": 0.21,
"grad_norm": 11.708748373782319,
"learning_rate": 1.941935301148439e-05,
"loss": 0.7569,
"step": 195
},
{
"epoch": 0.21,
"grad_norm": 10.640144063322005,
"learning_rate": 1.9406788624708438e-05,
"loss": 0.694,
"step": 196
},
{
"epoch": 0.21,
"grad_norm": 14.454447375580173,
"learning_rate": 1.939409390452913e-05,
"loss": 0.7807,
"step": 197
},
{
"epoch": 0.21,
"grad_norm": 10.35715254572469,
"learning_rate": 1.9381269026834975e-05,
"loss": 0.6799,
"step": 198
},
{
"epoch": 0.21,
"grad_norm": 6.919008432462478,
"learning_rate": 1.9368314169317858e-05,
"loss": 0.6883,
"step": 199
},
{
"epoch": 0.21,
"grad_norm": 9.137624019639443,
"learning_rate": 1.935522951147056e-05,
"loss": 0.5919,
"step": 200
},
{
"epoch": 0.21,
"grad_norm": 15.35731927734825,
"learning_rate": 1.9342015234584277e-05,
"loss": 0.5982,
"step": 201
},
{
"epoch": 0.22,
"grad_norm": 28.1752923954513,
"learning_rate": 1.932867152174612e-05,
"loss": 0.6482,
"step": 202
},
{
"epoch": 0.22,
"grad_norm": 28.1752923954513,
"learning_rate": 1.932867152174612e-05,
"loss": 1.748,
"step": 203
},
{
"epoch": 0.22,
"grad_norm": 1587.2702484305357,
"learning_rate": 1.9315198557836555e-05,
"loss": 1.6079,
"step": 204
},
{
"epoch": 0.22,
"grad_norm": 1587.2702484305357,
"learning_rate": 1.9315198557836555e-05,
"loss": 11.678,
"step": 205
},
{
"epoch": 0.22,
"grad_norm": 2434.8463463776825,
"learning_rate": 1.9301596529526856e-05,
"loss": 11.5182,
"step": 206
},
{
"epoch": 0.22,
"grad_norm": 2901.204727033937,
"learning_rate": 1.928786562527652e-05,
"loss": 9.7123,
"step": 207
},
{
"epoch": 0.22,
"grad_norm": 490.65682411342476,
"learning_rate": 1.9274006035330654e-05,
"loss": 3.1282,
"step": 208
},
{
"epoch": 0.22,
"grad_norm": 480.76138720661936,
"learning_rate": 1.9260017951717334e-05,
"loss": 2.4949,
"step": 209
},
{
"epoch": 0.22,
"grad_norm": 143.8799125203908,
"learning_rate": 1.9245901568244945e-05,
"loss": 2.1616,
"step": 210
},
{
"epoch": 0.22,
"grad_norm": 226.80553179811704,
"learning_rate": 1.9231657080499507e-05,
"loss": 1.9539,
"step": 211
},
{
"epoch": 0.23,
"grad_norm": 210.9739319214792,
"learning_rate": 1.921728468584195e-05,
"loss": 1.7932,
"step": 212
},
{
"epoch": 0.23,
"grad_norm": 102.69163082309667,
"learning_rate": 1.9202784583405386e-05,
"loss": 1.5288,
"step": 213
},
{
"epoch": 0.23,
"grad_norm": 80.79475053684271,
"learning_rate": 1.9188156974092355e-05,
"loss": 1.4608,
"step": 214
},
{
"epoch": 0.23,
"grad_norm": 131.37203972527402,
"learning_rate": 1.9173402060572028e-05,
"loss": 1.3191,
"step": 215
},
{
"epoch": 0.23,
"grad_norm": 641.6313774584162,
"learning_rate": 1.915852004727742e-05,
"loss": 1.8877,
"step": 216
},
{
"epoch": 0.23,
"grad_norm": 535.4790750541629,
"learning_rate": 1.9143511140402532e-05,
"loss": 1.4452,
"step": 217
},
{
"epoch": 0.23,
"grad_norm": 548.3152023208836,
"learning_rate": 1.9128375547899518e-05,
"loss": 1.9607,
"step": 218
},
{
"epoch": 0.23,
"grad_norm": 568.5307973204539,
"learning_rate": 1.9113113479475784e-05,
"loss": 1.8329,
"step": 219
},
{
"epoch": 0.23,
"grad_norm": 247.07029020503853,
"learning_rate": 1.9097725146591103e-05,
"loss": 1.2675,
"step": 220
},
{
"epoch": 0.24,
"grad_norm": 483.1919344273096,
"learning_rate": 1.908221076245466e-05,
"loss": 3.1382,
"step": 221
},
{
"epoch": 0.24,
"grad_norm": 435.67666262712555,
"learning_rate": 1.9066570542022122e-05,
"loss": 2.9579,
"step": 222
},
{
"epoch": 0.24,
"grad_norm": 106.36477945538998,
"learning_rate": 1.905080470199264e-05,
"loss": 1.9584,
"step": 223
},
{
"epoch": 0.24,
"grad_norm": 150.21543919990003,
"learning_rate": 1.9034913460805868e-05,
"loss": 1.5164,
"step": 224
},
{
"epoch": 0.24,
"grad_norm": 265.0211172938759,
"learning_rate": 1.901889703863891e-05,
"loss": 1.4234,
"step": 225
},
{
"epoch": 0.24,
"grad_norm": 135.22687197035,
"learning_rate": 1.90027556574033e-05,
"loss": 1.3242,
"step": 226
},
{
"epoch": 0.24,
"grad_norm": 69.09048458856181,
"learning_rate": 1.8986489540741895e-05,
"loss": 1.1573,
"step": 227
},
{
"epoch": 0.24,
"grad_norm": 317.48370008905204,
"learning_rate": 1.89700989140258e-05,
"loss": 1.6469,
"step": 228
},
{
"epoch": 0.24,
"grad_norm": 178.1220954711751,
"learning_rate": 1.8953584004351243e-05,
"loss": 1.3997,
"step": 229
},
{
"epoch": 0.25,
"grad_norm": 58.594800768267845,
"learning_rate": 1.8936945040536413e-05,
"loss": 1.4203,
"step": 230
},
{
"epoch": 0.25,
"grad_norm": 100.8540607851877,
"learning_rate": 1.892018225311831e-05,
"loss": 1.3959,
"step": 231
},
{
"epoch": 0.25,
"grad_norm": 40.196456568889715,
"learning_rate": 1.8903295874349528e-05,
"loss": 1.3036,
"step": 232
},
{
"epoch": 0.25,
"grad_norm": 49.160977991667785,
"learning_rate": 1.8886286138195063e-05,
"loss": 1.2698,
"step": 233
},
{
"epoch": 0.25,
"grad_norm": 38.84069117366192,
"learning_rate": 1.8869153280329054e-05,
"loss": 1.1846,
"step": 234
},
{
"epoch": 0.25,
"grad_norm": 43.81901383324617,
"learning_rate": 1.885189753813152e-05,
"loss": 1.2088,
"step": 235
},
{
"epoch": 0.25,
"grad_norm": 19.950274945570303,
"learning_rate": 1.8834519150685074e-05,
"loss": 1.0781,
"step": 236
},
{
"epoch": 0.25,
"grad_norm": 50.927777642645886,
"learning_rate": 1.8817018358771612e-05,
"loss": 1.207,
"step": 237
},
{
"epoch": 0.25,
"grad_norm": 45.137241355795545,
"learning_rate": 1.879939540486897e-05,
"loss": 1.1345,
"step": 238
},
{
"epoch": 0.25,
"grad_norm": 49.611501553831054,
"learning_rate": 1.8781650533147572e-05,
"loss": 1.1675,
"step": 239
},
{
"epoch": 0.26,
"grad_norm": 60.82753635007021,
"learning_rate": 1.876378398946704e-05,
"loss": 1.1098,
"step": 240
},
{
"epoch": 0.26,
"grad_norm": 36.20738946945509,
"learning_rate": 1.87457960213728e-05,
"loss": 1.0201,
"step": 241
},
{
"epoch": 0.26,
"grad_norm": 46.88562809493131,
"learning_rate": 1.8727686878092638e-05,
"loss": 1.0511,
"step": 242
},
{
"epoch": 0.26,
"grad_norm": 38.81307588678313,
"learning_rate": 1.8709456810533248e-05,
"loss": 0.9694,
"step": 243
},
{
"epoch": 0.26,
"grad_norm": 29.29930215775943,
"learning_rate": 1.8691106071276763e-05,
"loss": 0.9828,
"step": 244
},
{
"epoch": 0.26,
"grad_norm": 54.42807589658664,
"learning_rate": 1.867263491457726e-05,
"loss": 1.1109,
"step": 245
},
{
"epoch": 0.26,
"grad_norm": 29.031387361113243,
"learning_rate": 1.8654043596357215e-05,
"loss": 1.0194,
"step": 246
},
{
"epoch": 0.26,
"grad_norm": 35.21872276907598,
"learning_rate": 1.8635332374203993e-05,
"loss": 1.0552,
"step": 247
},
{
"epoch": 0.26,
"grad_norm": 22.43165286258535,
"learning_rate": 1.8616501507366233e-05,
"loss": 0.9364,
"step": 248
},
{
"epoch": 0.27,
"grad_norm": 29.83402427558291,
"learning_rate": 1.85975512567503e-05,
"loss": 1.0535,
"step": 249
},
{
"epoch": 0.27,
"grad_norm": 36.54199147143507,
"learning_rate": 1.8578481884916645e-05,
"loss": 0.9772,
"step": 250
},
{
"epoch": 0.27,
"grad_norm": 16.06774723588221,
"learning_rate": 1.8559293656076167e-05,
"loss": 0.9044,
"step": 251
},
{
"epoch": 0.27,
"grad_norm": 14.242434397235082,
"learning_rate": 1.8539986836086568e-05,
"loss": 0.8656,
"step": 252
},
{
"epoch": 0.27,
"grad_norm": 34.06087523793793,
"learning_rate": 1.8520561692448655e-05,
"loss": 0.9776,
"step": 253
},
{
"epoch": 0.27,
"grad_norm": 23.65421594444165,
"learning_rate": 1.850101849430264e-05,
"loss": 0.9739,
"step": 254
},
{
"epoch": 0.27,
"grad_norm": 53.1622333489428,
"learning_rate": 1.848135751242441e-05,
"loss": 0.9659,
"step": 255
},
{
"epoch": 0.27,
"grad_norm": 27.743193122343396,
"learning_rate": 1.8461579019221775e-05,
"loss": 0.9828,
"step": 256
},
{
"epoch": 0.27,
"grad_norm": 36.32985964499645,
"learning_rate": 1.8441683288730686e-05,
"loss": 1.0724,
"step": 257
},
{
"epoch": 0.27,
"grad_norm": 57.208457798230874,
"learning_rate": 1.8421670596611463e-05,
"loss": 1.1399,
"step": 258
},
{
"epoch": 0.28,
"grad_norm": 36.265365003042525,
"learning_rate": 1.840154122014494e-05,
"loss": 1.061,
"step": 259
},
{
"epoch": 0.28,
"grad_norm": 50.18323851957246,
"learning_rate": 1.8381295438228656e-05,
"loss": 0.9041,
"step": 260
},
{
"epoch": 0.28,
"grad_norm": 23.117360915951355,
"learning_rate": 1.836093353137297e-05,
"loss": 0.974,
"step": 261
},
{
"epoch": 0.28,
"grad_norm": 50.588104484420946,
"learning_rate": 1.8340455781697175e-05,
"loss": 0.9803,
"step": 262
},
{
"epoch": 0.28,
"grad_norm": 43.25313730795733,
"learning_rate": 1.831986247292561e-05,
"loss": 0.8178,
"step": 263
},
{
"epoch": 0.28,
"grad_norm": 25.22535296579869,
"learning_rate": 1.8299153890383705e-05,
"loss": 0.9212,
"step": 264
},
{
"epoch": 0.28,
"grad_norm": 24.74441945494526,
"learning_rate": 1.8278330320994035e-05,
"loss": 0.9241,
"step": 265
},
{
"epoch": 0.28,
"grad_norm": 24.98014366767532,
"learning_rate": 1.8257392053272345e-05,
"loss": 0.9996,
"step": 266
},
{
"epoch": 0.28,
"grad_norm": 30.455334182511248,
"learning_rate": 1.823633937732357e-05,
"loss": 0.8491,
"step": 267
},
{
"epoch": 0.29,
"grad_norm": 35.02803006376697,
"learning_rate": 1.8215172584837776e-05,
"loss": 0.9924,
"step": 268
},
{
"epoch": 0.29,
"grad_norm": 21.742297963510474,
"learning_rate": 1.8193891969086164e-05,
"loss": 0.9068,
"step": 269
},
{
"epoch": 0.29,
"grad_norm": 16.27634444074666,
"learning_rate": 1.817249782491697e-05,
"loss": 0.9532,
"step": 270
},
{
"epoch": 0.29,
"grad_norm": 27.206570441621835,
"learning_rate": 1.8150990448751393e-05,
"loss": 0.9585,
"step": 271
},
{
"epoch": 0.29,
"grad_norm": 9.683718717075713,
"learning_rate": 1.8129370138579507e-05,
"loss": 0.8514,
"step": 272
},
{
"epoch": 0.29,
"grad_norm": 46.207488174236765,
"learning_rate": 1.8107637193956102e-05,
"loss": 0.9744,
"step": 273
},
{
"epoch": 0.29,
"grad_norm": 24.195105770115998,
"learning_rate": 1.8085791915996538e-05,
"loss": 0.994,
"step": 274
},
{
"epoch": 0.29,
"grad_norm": 33.17832980287206,
"learning_rate": 1.8063834607372603e-05,
"loss": 1.1351,
"step": 275
},
{
"epoch": 0.29,
"grad_norm": 18.66629676451458,
"learning_rate": 1.804176557230828e-05,
"loss": 0.7503,
"step": 276
},
{
"epoch": 0.3,
"grad_norm": 35.58783412196323,
"learning_rate": 1.8019585116575554e-05,
"loss": 0.8252,
"step": 277
},
{
"epoch": 0.3,
"grad_norm": 98.83505787049795,
"learning_rate": 1.799729354749018e-05,
"loss": 0.772,
"step": 278
},
{
"epoch": 0.3,
"grad_norm": 21.601957059226525,
"learning_rate": 1.7974891173907406e-05,
"loss": 0.7751,
"step": 279
},
{
"epoch": 0.3,
"grad_norm": 25.682884674147235,
"learning_rate": 1.7952378306217705e-05,
"loss": 0.8812,
"step": 280
},
{
"epoch": 0.3,
"grad_norm": 14.543550272731224,
"learning_rate": 1.792975525634248e-05,
"loss": 0.8405,
"step": 281
},
{
"epoch": 0.3,
"grad_norm": 49.50597218889562,
"learning_rate": 1.790702233772973e-05,
"loss": 0.7983,
"step": 282
},
{
"epoch": 0.3,
"grad_norm": 28.92796112130337,
"learning_rate": 1.7884179865349713e-05,
"loss": 0.8288,
"step": 283
},
{
"epoch": 0.3,
"grad_norm": 31.306112233207404,
"learning_rate": 1.786122815569058e-05,
"loss": 1.0268,
"step": 284
},
{
"epoch": 0.3,
"grad_norm": 23.72305170543349,
"learning_rate": 1.7838167526754002e-05,
"loss": 0.9805,
"step": 285
},
{
"epoch": 0.3,
"grad_norm": 30.976885389109945,
"learning_rate": 1.7814998298050744e-05,
"loss": 0.9433,
"step": 286
},
{
"epoch": 0.31,
"grad_norm": 9.017186183397232,
"learning_rate": 1.7791720790596242e-05,
"loss": 0.9042,
"step": 287
},
{
"epoch": 0.31,
"grad_norm": 24.453663623228078,
"learning_rate": 1.7768335326906182e-05,
"loss": 0.916,
"step": 288
},
{
"epoch": 0.31,
"grad_norm": 16.661532775610848,
"learning_rate": 1.774484223099199e-05,
"loss": 0.8476,
"step": 289
},
{
"epoch": 0.31,
"grad_norm": 75.49348483174468,
"learning_rate": 1.7721241828356377e-05,
"loss": 0.9214,
"step": 290
},
{
"epoch": 0.31,
"grad_norm": 28.038342380284316,
"learning_rate": 1.7697534445988804e-05,
"loss": 0.5674,
"step": 291
},
{
"epoch": 0.31,
"grad_norm": 26.918490245230988,
"learning_rate": 1.7673720412360973e-05,
"loss": 0.7465,
"step": 292
},
{
"epoch": 0.31,
"grad_norm": 64.73808049621005,
"learning_rate": 1.7649800057422256e-05,
"loss": 0.7832,
"step": 293
},
{
"epoch": 0.31,
"grad_norm": 21.87842844322322,
"learning_rate": 1.7625773712595147e-05,
"loss": 0.5816,
"step": 294
},
{
"epoch": 0.31,
"grad_norm": 15.934437709078225,
"learning_rate": 1.760164171077064e-05,
"loss": 0.7923,
"step": 295
},
{
"epoch": 0.32,
"grad_norm": 50.69489297444858,
"learning_rate": 1.7577404386303646e-05,
"loss": 0.6258,
"step": 296
},
{
"epoch": 0.32,
"grad_norm": 21.105815660081543,
"learning_rate": 1.755306207500834e-05,
"loss": 0.7464,
"step": 297
},
{
"epoch": 0.32,
"grad_norm": 21.183565369421174,
"learning_rate": 1.7528615114153524e-05,
"loss": 0.6985,
"step": 298
},
{
"epoch": 0.32,
"grad_norm": 15.244682840971725,
"learning_rate": 1.750406384245793e-05,
"loss": 0.767,
"step": 299
},
{
"epoch": 0.32,
"grad_norm": 10.358767383759252,
"learning_rate": 1.747940860008556e-05,
"loss": 0.6928,
"step": 300
},
{
"epoch": 0.32,
"grad_norm": 39.96772160224854,
"learning_rate": 1.7454649728640944e-05,
"loss": 0.6887,
"step": 301
},
{
"epoch": 0.32,
"grad_norm": 16.221102491098176,
"learning_rate": 1.742978757116443e-05,
"loss": 0.6359,
"step": 302
},
{
"epoch": 0.32,
"grad_norm": 36.830429589303506,
"learning_rate": 1.7404822472127406e-05,
"loss": 0.6717,
"step": 303
},
{
"epoch": 0.32,
"grad_norm": 17.661289846973503,
"learning_rate": 1.7379754777427554e-05,
"loss": 0.6609,
"step": 304
},
{
"epoch": 0.32,
"grad_norm": 8.91929875914757,
"learning_rate": 1.7354584834384036e-05,
"loss": 0.6851,
"step": 305
},
{
"epoch": 0.33,
"grad_norm": 36.39317884774958,
"learning_rate": 1.7329312991732687e-05,
"loss": 0.8089,
"step": 306
},
{
"epoch": 0.33,
"grad_norm": 20.66408093673374,
"learning_rate": 1.73039395996212e-05,
"loss": 0.6783,
"step": 307
},
{
"epoch": 0.33,
"grad_norm": 13.03603460389346,
"learning_rate": 1.727846500960425e-05,
"loss": 0.7026,
"step": 308
},
{
"epoch": 0.33,
"grad_norm": 17.971533560556523,
"learning_rate": 1.725288957463864e-05,
"loss": 0.7989,
"step": 309
},
{
"epoch": 0.33,
"grad_norm": 9.851305189485732,
"learning_rate": 1.7227213649078395e-05,
"loss": 0.6135,
"step": 310
},
{
"epoch": 0.33,
"grad_norm": 13.9093394062407,
"learning_rate": 1.720143758866988e-05,
"loss": 0.7099,
"step": 311
},
{
"epoch": 0.33,
"grad_norm": 15.918012101752101,
"learning_rate": 1.7175561750546833e-05,
"loss": 0.6836,
"step": 312
},
{
"epoch": 0.33,
"grad_norm": 9.424349721892472,
"learning_rate": 1.7149586493225453e-05,
"loss": 0.6301,
"step": 313
},
{
"epoch": 0.33,
"grad_norm": 15.040725484530588,
"learning_rate": 1.7123512176599413e-05,
"loss": 0.5931,
"step": 314
},
{
"epoch": 0.34,
"grad_norm": 10.404277489273346,
"learning_rate": 1.709733916193487e-05,
"loss": 0.5216,
"step": 315
},
{
"epoch": 0.34,
"grad_norm": 10.351015057753036,
"learning_rate": 1.7071067811865477e-05,
"loss": 0.6885,
"step": 316
},
{
"epoch": 0.34,
"grad_norm": 12.500602158481879,
"learning_rate": 1.704469849038734e-05,
"loss": 0.66,
"step": 317
},
{
"epoch": 0.34,
"grad_norm": 9.587643268920083,
"learning_rate": 1.7018231562853987e-05,
"loss": 0.6073,
"step": 318
},
{
"epoch": 0.34,
"grad_norm": 13.14738192750954,
"learning_rate": 1.6991667395971306e-05,
"loss": 0.6194,
"step": 319
},
{
"epoch": 0.34,
"grad_norm": 11.464802706452456,
"learning_rate": 1.6965006357792454e-05,
"loss": 0.4891,
"step": 320
},
{
"epoch": 0.34,
"grad_norm": 6.153912186441177,
"learning_rate": 1.6938248817712767e-05,
"loss": 0.5314,
"step": 321
},
{
"epoch": 0.34,
"grad_norm": 13.329406945514984,
"learning_rate": 1.691139514646464e-05,
"loss": 0.7992,
"step": 322
},
{
"epoch": 0.34,
"grad_norm": 13.044412872566445,
"learning_rate": 1.6884445716112388e-05,
"loss": 0.7126,
"step": 323
},
{
"epoch": 0.35,
"grad_norm": 8.05329073835129,
"learning_rate": 1.685740090004709e-05,
"loss": 0.5594,
"step": 324
},
{
"epoch": 0.35,
"grad_norm": 12.513875641703635,
"learning_rate": 1.6830261072981423e-05,
"loss": 0.6085,
"step": 325
},
{
"epoch": 0.35,
"grad_norm": 12.059036404622644,
"learning_rate": 1.680302661094446e-05,
"loss": 0.5629,
"step": 326
},
{
"epoch": 0.35,
"grad_norm": 10.324831974232412,
"learning_rate": 1.677569789127647e-05,
"loss": 0.6268,
"step": 327
},
{
"epoch": 0.35,
"grad_norm": 9.702271050156932,
"learning_rate": 1.6748275292623678e-05,
"loss": 0.621,
"step": 328
},
{
"epoch": 0.35,
"grad_norm": 11.784449310834715,
"learning_rate": 1.6720759194933037e-05,
"loss": 0.6856,
"step": 329
},
{
"epoch": 0.35,
"grad_norm": 8.598925774309697,
"learning_rate": 1.669314997944694e-05,
"loss": 0.6918,
"step": 330
},
{
"epoch": 0.35,
"grad_norm": 11.857440160524549,
"learning_rate": 1.666544802869796e-05,
"loss": 0.6879,
"step": 331
},
{
"epoch": 0.35,
"grad_norm": 8.032666301516855,
"learning_rate": 1.6637653726503533e-05,
"loss": 0.5792,
"step": 332
},
{
"epoch": 0.35,
"grad_norm": 7.0937478718250295,
"learning_rate": 1.660976745796065e-05,
"loss": 0.6607,
"step": 333
},
{
"epoch": 0.36,
"grad_norm": 9.465209749684965,
"learning_rate": 1.658178960944052e-05,
"loss": 0.6535,
"step": 334
},
{
"epoch": 0.36,
"grad_norm": 15.131095507571468,
"learning_rate": 1.655372056858322e-05,
"loss": 0.6438,
"step": 335
},
{
"epoch": 0.36,
"grad_norm": 17.304531600026,
"learning_rate": 1.6525560724292303e-05,
"loss": 0.7653,
"step": 336
},
{
"epoch": 0.36,
"grad_norm": 8.695700770875131,
"learning_rate": 1.6497310466729448e-05,
"loss": 0.6045,
"step": 337
},
{
"epoch": 0.36,
"grad_norm": 11.95231240235872,
"learning_rate": 1.646897018730902e-05,
"loss": 0.6791,
"step": 338
},
{
"epoch": 0.36,
"grad_norm": 9.486589536350005,
"learning_rate": 1.6440540278692656e-05,
"loss": 0.5519,
"step": 339
},
{
"epoch": 0.36,
"grad_norm": 10.961929493661737,
"learning_rate": 1.6412021134783835e-05,
"loss": 0.7537,
"step": 340
},
{
"epoch": 0.36,
"grad_norm": 9.852616880372151,
"learning_rate": 1.6383413150722413e-05,
"loss": 0.5775,
"step": 341
},
{
"epoch": 0.36,
"grad_norm": 4.770203374130839,
"learning_rate": 1.6354716722879152e-05,
"loss": 0.5684,
"step": 342
},
{
"epoch": 0.37,
"grad_norm": 6.203239240182946,
"learning_rate": 1.6325932248850206e-05,
"loss": 0.5391,
"step": 343
},
{
"epoch": 0.37,
"grad_norm": 14.714310109670022,
"learning_rate": 1.6297060127451656e-05,
"loss": 0.6426,
"step": 344
},
{
"epoch": 0.37,
"grad_norm": 16.664110650870665,
"learning_rate": 1.626810075871394e-05,
"loss": 0.7493,
"step": 345
},
{
"epoch": 0.37,
"grad_norm": 10.82617102662038,
"learning_rate": 1.6239054543876345e-05,
"loss": 0.6214,
"step": 346
},
{
"epoch": 0.37,
"grad_norm": 13.189719768118387,
"learning_rate": 1.6209921885381418e-05,
"loss": 0.647,
"step": 347
},
{
"epoch": 0.37,
"grad_norm": 6.638748272318463,
"learning_rate": 1.6180703186869414e-05,
"loss": 0.6062,
"step": 348
},
{
"epoch": 0.37,
"grad_norm": 9.318868865822735,
"learning_rate": 1.615139885317269e-05,
"loss": 0.5877,
"step": 349
},
{
"epoch": 0.37,
"grad_norm": 13.07963389547572,
"learning_rate": 1.6122009290310097e-05,
"loss": 0.697,
"step": 350
},
{
"epoch": 0.37,
"grad_norm": 8.538286649824611,
"learning_rate": 1.6092534905481367e-05,
"loss": 0.605,
"step": 351
},
{
"epoch": 0.38,
"grad_norm": 14.845192579391924,
"learning_rate": 1.6062976107061454e-05,
"loss": 0.5283,
"step": 352
},
{
"epoch": 0.38,
"grad_norm": 11.398188334173918,
"learning_rate": 1.6033333304594886e-05,
"loss": 0.7799,
"step": 353
},
{
"epoch": 0.38,
"grad_norm": 11.098935636927996,
"learning_rate": 1.6003606908790082e-05,
"loss": 0.6517,
"step": 354
},
{
"epoch": 0.38,
"grad_norm": 9.295708766284237,
"learning_rate": 1.5973797331513674e-05,
"loss": 0.8253,
"step": 355
},
{
"epoch": 0.38,
"grad_norm": 15.251818246107037,
"learning_rate": 1.5943904985784797e-05,
"loss": 0.729,
"step": 356
},
{
"epoch": 0.38,
"grad_norm": 5.329126700593149,
"learning_rate": 1.5913930285769356e-05,
"loss": 0.5526,
"step": 357
},
{
"epoch": 0.38,
"grad_norm": 12.120816580430638,
"learning_rate": 1.5883873646774296e-05,
"loss": 0.699,
"step": 358
},
{
"epoch": 0.38,
"grad_norm": 10.153946398501363,
"learning_rate": 1.5853735485241858e-05,
"loss": 0.6955,
"step": 359
},
{
"epoch": 0.38,
"grad_norm": 6.094377398479003,
"learning_rate": 1.582351621874378e-05,
"loss": 0.6237,
"step": 360
},
{
"epoch": 0.38,
"grad_norm": 12.749739058129057,
"learning_rate": 1.579321626597554e-05,
"loss": 0.7169,
"step": 361
},
{
"epoch": 0.39,
"grad_norm": 8.376136137589857,
"learning_rate": 1.5762836046750546e-05,
"loss": 0.6543,
"step": 362
},
{
"epoch": 0.39,
"grad_norm": 6.832733254275183,
"learning_rate": 1.573237598199432e-05,
"loss": 0.6495,
"step": 363
},
{
"epoch": 0.39,
"grad_norm": 10.567138540731836,
"learning_rate": 1.570183649373865e-05,
"loss": 0.6424,
"step": 364
},
{
"epoch": 0.39,
"grad_norm": 4.242489418395466,
"learning_rate": 1.5671218005115767e-05,
"loss": 0.5227,
"step": 365
},
{
"epoch": 0.39,
"grad_norm": 10.186721290405504,
"learning_rate": 1.5640520940352476e-05,
"loss": 0.6831,
"step": 366
},
{
"epoch": 0.39,
"grad_norm": 5.2392014967068405,
"learning_rate": 1.5609745724764264e-05,
"loss": 0.5989,
"step": 367
},
{
"epoch": 0.39,
"grad_norm": 7.53125873380665,
"learning_rate": 1.5578892784749414e-05,
"loss": 0.5461,
"step": 368
},
{
"epoch": 0.39,
"grad_norm": 11.308114294485758,
"learning_rate": 1.5547962547783126e-05,
"loss": 0.5895,
"step": 369
},
{
"epoch": 0.39,
"grad_norm": 9.200894044889376,
"learning_rate": 1.5516955442411533e-05,
"loss": 0.7269,
"step": 370
},
{
"epoch": 0.4,
"grad_norm": 7.3282431817133,
"learning_rate": 1.5485871898245824e-05,
"loss": 0.5479,
"step": 371
},
{
"epoch": 0.4,
"grad_norm": 14.868640801063636,
"learning_rate": 1.5454712345956254e-05,
"loss": 0.6124,
"step": 372
},
{
"epoch": 0.4,
"grad_norm": 18.938959881407193,
"learning_rate": 1.54234772172662e-05,
"loss": 0.6996,
"step": 373
},
{
"epoch": 0.4,
"grad_norm": 9.41122569113249,
"learning_rate": 1.539216694494616e-05,
"loss": 0.7736,
"step": 374
},
{
"epoch": 0.4,
"grad_norm": 13.076645506243146,
"learning_rate": 1.536078196280777e-05,
"loss": 0.6184,
"step": 375
},
{
"epoch": 0.4,
"grad_norm": 14.603964846269042,
"learning_rate": 1.532932270569778e-05,
"loss": 0.7387,
"step": 376
},
{
"epoch": 0.4,
"grad_norm": 7.409964871939044,
"learning_rate": 1.5297789609492062e-05,
"loss": 0.6705,
"step": 377
},
{
"epoch": 0.4,
"grad_norm": 10.532417778256553,
"learning_rate": 1.526618311108952e-05,
"loss": 0.6359,
"step": 378
},
{
"epoch": 0.4,
"grad_norm": 7.918617892191258,
"learning_rate": 1.5234503648406075e-05,
"loss": 0.5597,
"step": 379
},
{
"epoch": 0.4,
"grad_norm": 8.984191116234184,
"learning_rate": 1.5202751660368594e-05,
"loss": 0.5886,
"step": 380
},
{
"epoch": 0.41,
"grad_norm": 8.875342349884049,
"learning_rate": 1.5170927586908787e-05,
"loss": 0.579,
"step": 381
},
{
"epoch": 0.41,
"grad_norm": 12.368851666890444,
"learning_rate": 1.513903186895713e-05,
"loss": 0.684,
"step": 382
},
{
"epoch": 0.41,
"grad_norm": 8.674893219974978,
"learning_rate": 1.5107064948436758e-05,
"loss": 0.5412,
"step": 383
},
{
"epoch": 0.41,
"grad_norm": 11.644842239089794,
"learning_rate": 1.5075027268257328e-05,
"loss": 0.6165,
"step": 384
},
{
"epoch": 0.41,
"grad_norm": 24.313226821351616,
"learning_rate": 1.5042919272308895e-05,
"loss": 1.0199,
"step": 385
},
{
"epoch": 0.41,
"grad_norm": 12.280306305433697,
"learning_rate": 1.5010741405455751e-05,
"loss": 0.5976,
"step": 386
},
{
"epoch": 0.41,
"grad_norm": 5.830472782170163,
"learning_rate": 1.4978494113530268e-05,
"loss": 0.5179,
"step": 387
},
{
"epoch": 0.41,
"grad_norm": 14.70235378916328,
"learning_rate": 1.4946177843326726e-05,
"loss": 0.7092,
"step": 388
},
{
"epoch": 0.41,
"grad_norm": 17.543216455121254,
"learning_rate": 1.4913793042595109e-05,
"loss": 0.8279,
"step": 389
},
{
"epoch": 0.42,
"grad_norm": 12.859056834147074,
"learning_rate": 1.4881340160034909e-05,
"loss": 0.7344,
"step": 390
},
{
"epoch": 0.42,
"grad_norm": 9.382814693311708,
"learning_rate": 1.4848819645288915e-05,
"loss": 0.5823,
"step": 391
},
{
"epoch": 0.42,
"grad_norm": 10.602701451692775,
"learning_rate": 1.4816231948936967e-05,
"loss": 0.706,
"step": 392
},
{
"epoch": 0.42,
"grad_norm": 9.699931667366167,
"learning_rate": 1.4783577522489733e-05,
"loss": 0.6573,
"step": 393
},
{
"epoch": 0.42,
"grad_norm": 8.280832106689648,
"learning_rate": 1.475085681838244e-05,
"loss": 0.6462,
"step": 394
},
{
"epoch": 0.42,
"grad_norm": 10.316666897706078,
"learning_rate": 1.4718070289968602e-05,
"loss": 0.6581,
"step": 395
},
{
"epoch": 0.42,
"grad_norm": 7.552089443149507,
"learning_rate": 1.4685218391513752e-05,
"loss": 0.6516,
"step": 396
},
{
"epoch": 0.42,
"grad_norm": 10.041760330765845,
"learning_rate": 1.4652301578189141e-05,
"loss": 0.7184,
"step": 397
},
{
"epoch": 0.42,
"grad_norm": 11.3264716915839,
"learning_rate": 1.4619320306065432e-05,
"loss": 0.6023,
"step": 398
},
{
"epoch": 0.43,
"grad_norm": 9.31395109143123,
"learning_rate": 1.4586275032106373e-05,
"loss": 0.6248,
"step": 399
},
{
"epoch": 0.43,
"grad_norm": 16.084796485739954,
"learning_rate": 1.4553166214162486e-05,
"loss": 0.5707,
"step": 400
},
{
"epoch": 0.43,
"grad_norm": 5.559676215156368,
"learning_rate": 1.4519994310964697e-05,
"loss": 0.545,
"step": 401
},
{
"epoch": 0.43,
"grad_norm": 7.5228362636067425,
"learning_rate": 1.4486759782118012e-05,
"loss": 0.634,
"step": 402
},
{
"epoch": 0.43,
"grad_norm": 10.426847253583567,
"learning_rate": 1.4453463088095108e-05,
"loss": 0.613,
"step": 403
},
{
"epoch": 0.43,
"grad_norm": 6.094049952534055,
"learning_rate": 1.4420104690229992e-05,
"loss": 0.5793,
"step": 404
},
{
"epoch": 0.43,
"grad_norm": 6.245104416676842,
"learning_rate": 1.4386685050711593e-05,
"loss": 0.5631,
"step": 405
},
{
"epoch": 0.43,
"grad_norm": 7.453663730544318,
"learning_rate": 1.4353204632577354e-05,
"loss": 0.5591,
"step": 406
},
{
"epoch": 0.43,
"grad_norm": 10.466741619757949,
"learning_rate": 1.4319663899706818e-05,
"loss": 0.7189,
"step": 407
},
{
"epoch": 0.43,
"grad_norm": 6.220499980519979,
"learning_rate": 1.4286063316815209e-05,
"loss": 0.4383,
"step": 408
},
{
"epoch": 0.44,
"grad_norm": 10.892786776066355,
"learning_rate": 1.4252403349446986e-05,
"loss": 0.5664,
"step": 409
},
{
"epoch": 0.44,
"grad_norm": 6.174114043802528,
"learning_rate": 1.4218684463969396e-05,
"loss": 0.5716,
"step": 410
},
{
"epoch": 0.44,
"grad_norm": 4.9332645630935605,
"learning_rate": 1.4184907127566006e-05,
"loss": 0.553,
"step": 411
},
{
"epoch": 0.44,
"grad_norm": 7.331476371683697,
"learning_rate": 1.4151071808230246e-05,
"loss": 0.5706,
"step": 412
},
{
"epoch": 0.44,
"grad_norm": 9.54440691864147,
"learning_rate": 1.4117178974758903e-05,
"loss": 0.6349,
"step": 413
},
{
"epoch": 0.44,
"grad_norm": 9.51093808442514,
"learning_rate": 1.4083229096745644e-05,
"loss": 0.6076,
"step": 414
},
{
"epoch": 0.44,
"grad_norm": 7.416880704206411,
"learning_rate": 1.404922264457449e-05,
"loss": 0.5238,
"step": 415
},
{
"epoch": 0.44,
"grad_norm": 9.30112890680624,
"learning_rate": 1.4015160089413331e-05,
"loss": 0.6277,
"step": 416
},
{
"epoch": 0.44,
"grad_norm": 9.964620600235138,
"learning_rate": 1.3981041903207364e-05,
"loss": 0.6501,
"step": 417
},
{
"epoch": 0.45,
"grad_norm": 8.049380804954936,
"learning_rate": 1.3946868558672569e-05,
"loss": 0.5853,
"step": 418
},
{
"epoch": 0.45,
"grad_norm": 9.941935672040696,
"learning_rate": 1.3912640529289163e-05,
"loss": 0.6632,
"step": 419
},
{
"epoch": 0.45,
"grad_norm": 4.961145697377557,
"learning_rate": 1.3878358289295032e-05,
"loss": 0.5542,
"step": 420
},
{
"epoch": 0.45,
"grad_norm": 4.788363887274814,
"learning_rate": 1.3844022313679167e-05,
"loss": 0.5203,
"step": 421
},
{
"epoch": 0.45,
"grad_norm": 12.11801121511047,
"learning_rate": 1.380963307817507e-05,
"loss": 0.6661,
"step": 422
},
{
"epoch": 0.45,
"grad_norm": 11.34675588679876,
"learning_rate": 1.3775191059254185e-05,
"loss": 0.6116,
"step": 423
},
{
"epoch": 0.45,
"grad_norm": 6.678425818352266,
"learning_rate": 1.3740696734119279e-05,
"loss": 0.6401,
"step": 424
},
{
"epoch": 0.45,
"grad_norm": 12.160178505852342,
"learning_rate": 1.3706150580697826e-05,
"loss": 0.6138,
"step": 425
},
{
"epoch": 0.45,
"grad_norm": 6.094569373598361,
"learning_rate": 1.3671553077635404e-05,
"loss": 0.5602,
"step": 426
},
{
"epoch": 0.45,
"grad_norm": 5.908165508182039,
"learning_rate": 1.3636904704289053e-05,
"loss": 0.5606,
"step": 427
},
{
"epoch": 0.46,
"grad_norm": 12.500164793835593,
"learning_rate": 1.3602205940720628e-05,
"loss": 0.5601,
"step": 428
},
{
"epoch": 0.46,
"grad_norm": 5.981711152222135,
"learning_rate": 1.3567457267690152e-05,
"loss": 0.5011,
"step": 429
},
{
"epoch": 0.46,
"grad_norm": 6.390064988076557,
"learning_rate": 1.3532659166649165e-05,
"loss": 0.515,
"step": 430
},
{
"epoch": 0.46,
"grad_norm": 11.203440074699836,
"learning_rate": 1.3497812119734037e-05,
"loss": 0.7016,
"step": 431
},
{
"epoch": 0.46,
"grad_norm": 13.733785334526278,
"learning_rate": 1.3462916609759298e-05,
"loss": 0.6623,
"step": 432
},
{
"epoch": 0.46,
"grad_norm": 15.73461768409232,
"learning_rate": 1.342797312021094e-05,
"loss": 0.7519,
"step": 433
},
{
"epoch": 0.46,
"grad_norm": 8.804555804165531,
"learning_rate": 1.3392982135239736e-05,
"loss": 0.531,
"step": 434
},
{
"epoch": 0.46,
"grad_norm": 10.63905280022859,
"learning_rate": 1.3357944139654508e-05,
"loss": 0.5914,
"step": 435
},
{
"epoch": 0.46,
"grad_norm": 14.738317600036572,
"learning_rate": 1.3322859618915431e-05,
"loss": 0.6387,
"step": 436
},
{
"epoch": 0.47,
"grad_norm": 9.315538586392796,
"learning_rate": 1.3287729059127288e-05,
"loss": 0.5793,
"step": 437
},
{
"epoch": 0.47,
"grad_norm": 7.330889488527192,
"learning_rate": 1.3252552947032755e-05,
"loss": 0.5955,
"step": 438
},
{
"epoch": 0.47,
"grad_norm": 5.492945122022015,
"learning_rate": 1.3217331770005639e-05,
"loss": 0.4936,
"step": 439
},
{
"epoch": 0.47,
"grad_norm": 6.447759334210931,
"learning_rate": 1.3182066016044135e-05,
"loss": 0.6306,
"step": 440
},
{
"epoch": 0.47,
"grad_norm": 10.58800825139178,
"learning_rate": 1.3146756173764061e-05,
"loss": 0.682,
"step": 441
},
{
"epoch": 0.47,
"grad_norm": 11.358838931871405,
"learning_rate": 1.3111402732392098e-05,
"loss": 0.7864,
"step": 442
},
{
"epoch": 0.47,
"grad_norm": 8.567212453461783,
"learning_rate": 1.3076006181758989e-05,
"loss": 0.5594,
"step": 443
},
{
"epoch": 0.47,
"grad_norm": 8.546126016623386,
"learning_rate": 1.3040567012292779e-05,
"loss": 0.6419,
"step": 444
},
{
"epoch": 0.47,
"grad_norm": 11.957735628493818,
"learning_rate": 1.3005085715012003e-05,
"loss": 0.632,
"step": 445
},
{
"epoch": 0.48,
"grad_norm": 10.122315964153882,
"learning_rate": 1.2969562781518885e-05,
"loss": 0.6758,
"step": 446
},
{
"epoch": 0.48,
"grad_norm": 10.274234862260917,
"learning_rate": 1.2933998703992531e-05,
"loss": 0.4632,
"step": 447
},
{
"epoch": 0.48,
"grad_norm": 9.56291976312044,
"learning_rate": 1.2898393975182113e-05,
"loss": 0.7572,
"step": 448
},
{
"epoch": 0.48,
"grad_norm": 11.309501354950024,
"learning_rate": 1.2862749088400026e-05,
"loss": 0.5429,
"step": 449
},
{
"epoch": 0.48,
"grad_norm": 5.4111570184934195,
"learning_rate": 1.2827064537515075e-05,
"loss": 0.5896,
"step": 450
},
{
"epoch": 0.48,
"grad_norm": 4.403745251090165,
"learning_rate": 1.279134081694561e-05,
"loss": 0.6544,
"step": 451
},
{
"epoch": 0.48,
"grad_norm": 5.2380739150735565,
"learning_rate": 1.2755578421652696e-05,
"loss": 0.4995,
"step": 452
},
{
"epoch": 0.48,
"grad_norm": 6.1969762293333135,
"learning_rate": 1.2719777847133241e-05,
"loss": 0.6669,
"step": 453
},
{
"epoch": 0.48,
"grad_norm": 6.009545288953141,
"learning_rate": 1.2683939589413139e-05,
"loss": 0.4661,
"step": 454
},
{
"epoch": 0.48,
"grad_norm": 11.862108381435066,
"learning_rate": 1.2648064145040392e-05,
"loss": 0.5745,
"step": 455
},
{
"epoch": 0.49,
"grad_norm": 8.371435047081937,
"learning_rate": 1.2612152011078233e-05,
"loss": 0.581,
"step": 456
},
{
"epoch": 0.49,
"grad_norm": 6.570729807144181,
"learning_rate": 1.2576203685098233e-05,
"loss": 0.5283,
"step": 457
},
{
"epoch": 0.49,
"grad_norm": 15.891490498465696,
"learning_rate": 1.2540219665173423e-05,
"loss": 0.6488,
"step": 458
},
{
"epoch": 0.49,
"grad_norm": 11.348604428897819,
"learning_rate": 1.2504200449871378e-05,
"loss": 0.6554,
"step": 459
},
{
"epoch": 0.49,
"grad_norm": 10.1223452644342,
"learning_rate": 1.2468146538247308e-05,
"loss": 0.6271,
"step": 460
},
{
"epoch": 0.49,
"grad_norm": 14.257981744049157,
"learning_rate": 1.2432058429837153e-05,
"loss": 0.694,
"step": 461
},
{
"epoch": 0.49,
"grad_norm": 14.906690269642054,
"learning_rate": 1.2395936624650664e-05,
"loss": 0.6699,
"step": 462
},
{
"epoch": 0.49,
"grad_norm": 7.166307639896277,
"learning_rate": 1.2359781623164465e-05,
"loss": 0.5572,
"step": 463
},
{
"epoch": 0.49,
"grad_norm": 13.74157282336694,
"learning_rate": 1.2323593926315113e-05,
"loss": 0.7576,
"step": 464
},
{
"epoch": 0.5,
"grad_norm": 9.280109663389997,
"learning_rate": 1.2287374035492184e-05,
"loss": 0.5046,
"step": 465
},
{
"epoch": 0.5,
"grad_norm": 10.090941451576859,
"learning_rate": 1.2251122452531301e-05,
"loss": 0.5604,
"step": 466
},
{
"epoch": 0.5,
"grad_norm": 5.903088605165492,
"learning_rate": 1.2214839679707193e-05,
"loss": 0.5624,
"step": 467
},
{
"epoch": 0.5,
"grad_norm": 10.167630601617299,
"learning_rate": 1.2178526219726723e-05,
"loss": 0.5295,
"step": 468
},
{
"epoch": 0.5,
"grad_norm": 8.331133592273517,
"learning_rate": 1.2142182575721946e-05,
"loss": 0.6144,
"step": 469
},
{
"epoch": 0.5,
"grad_norm": 9.705907066152204,
"learning_rate": 1.2105809251243113e-05,
"loss": 0.6389,
"step": 470
},
{
"epoch": 0.5,
"grad_norm": 11.02758523623926,
"learning_rate": 1.2069406750251713e-05,
"loss": 0.666,
"step": 471
},
{
"epoch": 0.5,
"grad_norm": 7.447181404677012,
"learning_rate": 1.2032975577113474e-05,
"loss": 0.7216,
"step": 472
},
{
"epoch": 0.5,
"grad_norm": 7.658897343141703,
"learning_rate": 1.1996516236591398e-05,
"loss": 0.6087,
"step": 473
},
{
"epoch": 0.51,
"grad_norm": 4.187819572120313,
"learning_rate": 1.1960029233838737e-05,
"loss": 0.5356,
"step": 474
},
{
"epoch": 0.51,
"grad_norm": 5.7176090622427385,
"learning_rate": 1.1923515074392022e-05,
"loss": 0.619,
"step": 475
},
{
"epoch": 0.51,
"grad_norm": 6.823057157296125,
"learning_rate": 1.1886974264164037e-05,
"loss": 0.6037,
"step": 476
},
{
"epoch": 0.51,
"grad_norm": 11.059981117618824,
"learning_rate": 1.1850407309436831e-05,
"loss": 0.604,
"step": 477
},
{
"epoch": 0.51,
"grad_norm": 9.504143330556396,
"learning_rate": 1.181381471685468e-05,
"loss": 0.647,
"step": 478
},
{
"epoch": 0.51,
"grad_norm": 6.557755314159755,
"learning_rate": 1.1777196993417087e-05,
"loss": 0.4595,
"step": 479
},
{
"epoch": 0.51,
"grad_norm": 6.9247206639927095,
"learning_rate": 1.1740554646471742e-05,
"loss": 0.5228,
"step": 480
},
{
"epoch": 0.51,
"grad_norm": 10.232093839271188,
"learning_rate": 1.1703888183707513e-05,
"loss": 0.6062,
"step": 481
},
{
"epoch": 0.51,
"grad_norm": 6.510487958239591,
"learning_rate": 1.166719811314738e-05,
"loss": 0.5219,
"step": 482
},
{
"epoch": 0.51,
"grad_norm": 8.968572470896715,
"learning_rate": 1.1630484943141428e-05,
"loss": 0.5209,
"step": 483
},
{
"epoch": 0.52,
"grad_norm": 5.0494617841107265,
"learning_rate": 1.1593749182359788e-05,
"loss": 0.5948,
"step": 484
},
{
"epoch": 0.52,
"grad_norm": 11.340103876058729,
"learning_rate": 1.1556991339785595e-05,
"loss": 0.6203,
"step": 485
},
{
"epoch": 0.52,
"grad_norm": 10.146196176770191,
"learning_rate": 1.1520211924707919e-05,
"loss": 0.7785,
"step": 486
},
{
"epoch": 0.52,
"grad_norm": 4.43271626514912,
"learning_rate": 1.1483411446714744e-05,
"loss": 0.4862,
"step": 487
},
{
"epoch": 0.52,
"grad_norm": 7.824102062875491,
"learning_rate": 1.1446590415685863e-05,
"loss": 0.6113,
"step": 488
},
{
"epoch": 0.52,
"grad_norm": 6.336647655938098,
"learning_rate": 1.1409749341785859e-05,
"loss": 0.5343,
"step": 489
},
{
"epoch": 0.52,
"grad_norm": 4.855528076878574,
"learning_rate": 1.1372888735456988e-05,
"loss": 0.5014,
"step": 490
},
{
"epoch": 0.52,
"grad_norm": 8.244191689910387,
"learning_rate": 1.1336009107412162e-05,
"loss": 0.64,
"step": 491
},
{
"epoch": 0.52,
"grad_norm": 9.72186897530395,
"learning_rate": 1.1299110968627822e-05,
"loss": 0.5919,
"step": 492
},
{
"epoch": 0.53,
"grad_norm": 6.153546625651434,
"learning_rate": 1.1262194830336888e-05,
"loss": 0.4917,
"step": 493
},
{
"epoch": 0.53,
"grad_norm": 8.261625036086883,
"learning_rate": 1.1225261204021662e-05,
"loss": 0.5592,
"step": 494
},
{
"epoch": 0.53,
"grad_norm": 7.32183514687012,
"learning_rate": 1.118831060140676e-05,
"loss": 0.4822,
"step": 495
},
{
"epoch": 0.53,
"grad_norm": 6.309347895522231,
"learning_rate": 1.1151343534451995e-05,
"loss": 0.5727,
"step": 496
},
{
"epoch": 0.53,
"grad_norm": 7.8065524764950265,
"learning_rate": 1.1114360515345301e-05,
"loss": 0.6533,
"step": 497
},
{
"epoch": 0.53,
"grad_norm": 9.553172955626218,
"learning_rate": 1.107736205649564e-05,
"loss": 0.4895,
"step": 498
},
{
"epoch": 0.53,
"grad_norm": 8.353357382799153,
"learning_rate": 1.1040348670525889e-05,
"loss": 0.5254,
"step": 499
},
{
"epoch": 0.53,
"grad_norm": 11.74491323882519,
"learning_rate": 1.1003320870265741e-05,
"loss": 0.6004,
"step": 500
},
{
"epoch": 0.53,
"grad_norm": 3.4228605283872797,
"learning_rate": 1.096627916874461e-05,
"loss": 0.4572,
"step": 501
},
{
"epoch": 0.53,
"grad_norm": 11.366613402895425,
"learning_rate": 1.0929224079184514e-05,
"loss": 0.648,
"step": 502
},
{
"epoch": 0.54,
"grad_norm": 5.915082226142763,
"learning_rate": 1.0892156114992963e-05,
"loss": 0.5606,
"step": 503
},
{
"epoch": 0.54,
"grad_norm": 5.481987693582871,
"learning_rate": 1.0855075789755845e-05,
"loss": 0.5845,
"step": 504
},
{
"epoch": 0.54,
"grad_norm": 6.226052256005356,
"learning_rate": 1.0817983617230326e-05,
"loss": 0.5025,
"step": 505
},
{
"epoch": 0.54,
"grad_norm": 8.531553272430466,
"learning_rate": 1.0780880111337704e-05,
"loss": 0.5657,
"step": 506
},
{
"epoch": 0.54,
"grad_norm": 6.998952295190077,
"learning_rate": 1.0743765786156313e-05,
"loss": 0.7092,
"step": 507
},
{
"epoch": 0.54,
"grad_norm": 6.513406245121362,
"learning_rate": 1.0706641155914384e-05,
"loss": 0.548,
"step": 508
},
{
"epoch": 0.54,
"grad_norm": 8.48581353131954,
"learning_rate": 1.066950673498294e-05,
"loss": 0.5551,
"step": 509
},
{
"epoch": 0.54,
"grad_norm": 5.34639942444931,
"learning_rate": 1.0632363037868638e-05,
"loss": 0.4334,
"step": 510
},
{
"epoch": 0.54,
"grad_norm": 11.269718113412772,
"learning_rate": 1.0595210579206676e-05,
"loss": 0.5803,
"step": 511
},
{
"epoch": 0.55,
"grad_norm": 11.41953037054744,
"learning_rate": 1.0558049873753637e-05,
"loss": 0.5186,
"step": 512
},
{
"epoch": 0.55,
"grad_norm": 11.07638179235055,
"learning_rate": 1.0520881436380366e-05,
"loss": 0.6146,
"step": 513
},
{
"epoch": 0.55,
"grad_norm": 11.357643871564994,
"learning_rate": 1.0483705782064835e-05,
"loss": 0.4825,
"step": 514
},
{
"epoch": 0.55,
"grad_norm": 8.12832951837023,
"learning_rate": 1.0446523425885008e-05,
"loss": 0.5972,
"step": 515
},
{
"epoch": 0.55,
"grad_norm": 4.573339318696573,
"learning_rate": 1.040933488301171e-05,
"loss": 0.4511,
"step": 516
},
{
"epoch": 0.55,
"grad_norm": 7.488115795627891,
"learning_rate": 1.0372140668701483e-05,
"loss": 0.5612,
"step": 517
},
{
"epoch": 0.55,
"grad_norm": 4.416399464656011,
"learning_rate": 1.0334941298289441e-05,
"loss": 0.5678,
"step": 518
},
{
"epoch": 0.55,
"grad_norm": 8.472247115461284,
"learning_rate": 1.0297737287182144e-05,
"loss": 0.5397,
"step": 519
},
{
"epoch": 0.55,
"grad_norm": 7.278212668194716,
"learning_rate": 1.026052915085045e-05,
"loss": 0.5597,
"step": 520
},
{
"epoch": 0.56,
"grad_norm": 5.04716478079048,
"learning_rate": 1.022331740482237e-05,
"loss": 0.5014,
"step": 521
},
{
"epoch": 0.56,
"grad_norm": 11.570053771414813,
"learning_rate": 1.0186102564675928e-05,
"loss": 0.661,
"step": 522
},
{
"epoch": 0.56,
"grad_norm": 8.829505233768511,
"learning_rate": 1.014888514603202e-05,
"loss": 0.5232,
"step": 523
},
{
"epoch": 0.56,
"grad_norm": 9.542821950950582,
"learning_rate": 1.0111665664547267e-05,
"loss": 0.5823,
"step": 524
},
{
"epoch": 0.56,
"grad_norm": 4.63327242048301,
"learning_rate": 1.0074444635906875e-05,
"loss": 0.5625,
"step": 525
},
{
"epoch": 0.56,
"grad_norm": 6.365741719759197,
"learning_rate": 1.0037222575817476e-05,
"loss": 0.6587,
"step": 526
},
{
"epoch": 0.56,
"grad_norm": 8.488178372143695,
"learning_rate": 1e-05,
"loss": 0.5501,
"step": 527
},
{
"epoch": 0.56,
"grad_norm": 9.897138663353587,
"learning_rate": 9.962777424182527e-06,
"loss": 0.6661,
"step": 528
},
{
"epoch": 0.56,
"grad_norm": 10.569907059227898,
"learning_rate": 9.92555536409313e-06,
"loss": 0.5958,
"step": 529
},
{
"epoch": 0.56,
"grad_norm": 3.736421534561604,
"learning_rate": 9.888334335452734e-06,
"loss": 0.5303,
"step": 530
},
{
"epoch": 0.57,
"grad_norm": 8.836116584372025,
"learning_rate": 9.85111485396798e-06,
"loss": 0.6466,
"step": 531
},
{
"epoch": 0.57,
"grad_norm": 7.181117148918435,
"learning_rate": 9.813897435324075e-06,
"loss": 0.54,
"step": 532
},
{
"epoch": 0.57,
"grad_norm": 4.785898028898109,
"learning_rate": 9.776682595177633e-06,
"loss": 0.4985,
"step": 533
},
{
"epoch": 0.57,
"grad_norm": 11.140585777357792,
"learning_rate": 9.739470849149554e-06,
"loss": 0.6496,
"step": 534
},
{
"epoch": 0.57,
"grad_norm": 3.531792374263498,
"learning_rate": 9.702262712817857e-06,
"loss": 0.5364,
"step": 535
},
{
"epoch": 0.57,
"grad_norm": 6.601039058908546,
"learning_rate": 9.66505870171056e-06,
"loss": 0.5116,
"step": 536
},
{
"epoch": 0.57,
"grad_norm": 8.927040349981295,
"learning_rate": 9.627859331298522e-06,
"loss": 0.5219,
"step": 537
},
{
"epoch": 0.57,
"grad_norm": 6.1651706502501264,
"learning_rate": 9.590665116988293e-06,
"loss": 0.5608,
"step": 538
},
{
"epoch": 0.57,
"grad_norm": 9.745244839875141,
"learning_rate": 9.553476574114993e-06,
"loss": 0.5524,
"step": 539
},
{
"epoch": 0.58,
"grad_norm": 7.008580981680822,
"learning_rate": 9.51629421793517e-06,
"loss": 0.5662,
"step": 540
},
{
"epoch": 0.58,
"grad_norm": 7.687272495275157,
"learning_rate": 9.479118563619638e-06,
"loss": 0.5806,
"step": 541
},
{
"epoch": 0.58,
"grad_norm": 5.593165759891498,
"learning_rate": 9.441950126246365e-06,
"loss": 0.5447,
"step": 542
},
{
"epoch": 0.58,
"grad_norm": 4.518154487773691,
"learning_rate": 9.404789420793327e-06,
"loss": 0.5662,
"step": 543
},
{
"epoch": 0.58,
"grad_norm": 5.979882247197896,
"learning_rate": 9.367636962131364e-06,
"loss": 0.5499,
"step": 544
},
{
"epoch": 0.58,
"grad_norm": 10.248432776308547,
"learning_rate": 9.330493265017062e-06,
"loss": 0.6341,
"step": 545
},
{
"epoch": 0.58,
"grad_norm": 7.4913799071398275,
"learning_rate": 9.29335884408562e-06,
"loss": 0.4995,
"step": 546
},
{
"epoch": 0.58,
"grad_norm": 12.97254908941589,
"learning_rate": 9.25623421384369e-06,
"loss": 0.5575,
"step": 547
},
{
"epoch": 0.58,
"grad_norm": 7.7639027583981095,
"learning_rate": 9.2191198886623e-06,
"loss": 0.5819,
"step": 548
},
{
"epoch": 0.58,
"grad_norm": 7.6543085882238895,
"learning_rate": 9.182016382769678e-06,
"loss": 0.6647,
"step": 549
},
{
"epoch": 0.59,
"grad_norm": 5.7656791466080355,
"learning_rate": 9.144924210244155e-06,
"loss": 0.6249,
"step": 550
},
{
"epoch": 0.59,
"grad_norm": 5.2481708291921505,
"learning_rate": 9.107843885007042e-06,
"loss": 0.5478,
"step": 551
},
{
"epoch": 0.59,
"grad_norm": 10.21885595901045,
"learning_rate": 9.070775920815488e-06,
"loss": 0.5803,
"step": 552
},
{
"epoch": 0.59,
"grad_norm": 7.389626516947618,
"learning_rate": 9.033720831255391e-06,
"loss": 0.5334,
"step": 553
},
{
"epoch": 0.59,
"grad_norm": 11.910444324111724,
"learning_rate": 8.996679129734266e-06,
"loss": 0.5788,
"step": 554
},
{
"epoch": 0.59,
"grad_norm": 10.70692641976889,
"learning_rate": 8.959651329474115e-06,
"loss": 0.5155,
"step": 555
},
{
"epoch": 0.59,
"grad_norm": 5.6372805638926,
"learning_rate": 8.922637943504362e-06,
"loss": 0.5881,
"step": 556
},
{
"epoch": 0.59,
"grad_norm": 9.8644472116661,
"learning_rate": 8.8856394846547e-06,
"loss": 0.5072,
"step": 557
},
{
"epoch": 0.59,
"grad_norm": 11.356856786195184,
"learning_rate": 8.848656465548007e-06,
"loss": 0.549,
"step": 558
},
{
"epoch": 0.6,
"grad_norm": 11.998300599364015,
"learning_rate": 8.811689398593245e-06,
"loss": 0.5355,
"step": 559
},
{
"epoch": 0.6,
"grad_norm": 5.815584273089378,
"learning_rate": 8.774738795978341e-06,
"loss": 0.4673,
"step": 560
},
{
"epoch": 0.6,
"grad_norm": 9.210932501903986,
"learning_rate": 8.737805169663113e-06,
"loss": 0.5137,
"step": 561
},
{
"epoch": 0.6,
"grad_norm": 10.604624719913076,
"learning_rate": 8.700889031372183e-06,
"loss": 0.6092,
"step": 562
},
{
"epoch": 0.6,
"grad_norm": 6.733937738438429,
"learning_rate": 8.663990892587839e-06,
"loss": 0.5617,
"step": 563
},
{
"epoch": 0.6,
"grad_norm": 6.687819879267525,
"learning_rate": 8.627111264543012e-06,
"loss": 0.4319,
"step": 564
},
{
"epoch": 0.6,
"grad_norm": 11.209429725910061,
"learning_rate": 8.590250658214148e-06,
"loss": 0.5174,
"step": 565
},
{
"epoch": 0.6,
"grad_norm": 5.337704061513362,
"learning_rate": 8.553409584314138e-06,
"loss": 0.4295,
"step": 566
},
{
"epoch": 0.6,
"grad_norm": 9.126368763120784,
"learning_rate": 8.516588553285258e-06,
"loss": 0.4436,
"step": 567
},
{
"epoch": 0.61,
"grad_norm": 10.856129898989279,
"learning_rate": 8.479788075292083e-06,
"loss": 0.6126,
"step": 568
},
{
"epoch": 0.61,
"grad_norm": 5.163627248000245,
"learning_rate": 8.443008660214409e-06,
"loss": 0.5352,
"step": 569
},
{
"epoch": 0.61,
"grad_norm": 6.372067670711978,
"learning_rate": 8.406250817640212e-06,
"loss": 0.5076,
"step": 570
},
{
"epoch": 0.61,
"grad_norm": 6.426831912031028,
"learning_rate": 8.369515056858575e-06,
"loss": 0.6041,
"step": 571
},
{
"epoch": 0.61,
"grad_norm": 6.985038214862641,
"learning_rate": 8.332801886852624e-06,
"loss": 0.6586,
"step": 572
},
{
"epoch": 0.61,
"grad_norm": 6.593624344866879,
"learning_rate": 8.296111816292494e-06,
"loss": 0.6751,
"step": 573
},
{
"epoch": 0.61,
"grad_norm": 7.554153671309798,
"learning_rate": 8.259445353528261e-06,
"loss": 0.5276,
"step": 574
},
{
"epoch": 0.61,
"grad_norm": 9.224509289941837,
"learning_rate": 8.222803006582915e-06,
"loss": 0.5993,
"step": 575
},
{
"epoch": 0.61,
"grad_norm": 16.50239364012574,
"learning_rate": 8.186185283145325e-06,
"loss": 0.6259,
"step": 576
},
{
"epoch": 0.61,
"grad_norm": 15.120105685997352,
"learning_rate": 8.149592690563172e-06,
"loss": 0.6467,
"step": 577
},
{
"epoch": 0.62,
"grad_norm": 7.220612802422961,
"learning_rate": 8.113025735835965e-06,
"loss": 0.5715,
"step": 578
},
{
"epoch": 0.62,
"grad_norm": 8.562863959435413,
"learning_rate": 8.076484925607983e-06,
"loss": 0.5466,
"step": 579
},
{
"epoch": 0.62,
"grad_norm": 12.490548007030851,
"learning_rate": 8.039970766161264e-06,
"loss": 0.5514,
"step": 580
},
{
"epoch": 0.62,
"grad_norm": 6.5845388402784435,
"learning_rate": 8.003483763408604e-06,
"loss": 0.5687,
"step": 581
},
{
"epoch": 0.62,
"grad_norm": 8.50284284365726,
"learning_rate": 7.967024422886528e-06,
"loss": 0.4796,
"step": 582
},
{
"epoch": 0.62,
"grad_norm": 7.038271102438535,
"learning_rate": 7.930593249748289e-06,
"loss": 0.5228,
"step": 583
},
{
"epoch": 0.62,
"grad_norm": 11.889834227941325,
"learning_rate": 7.894190748756892e-06,
"loss": 0.4577,
"step": 584
},
{
"epoch": 0.62,
"grad_norm": 6.711588374541616,
"learning_rate": 7.857817424278056e-06,
"loss": 0.498,
"step": 585
},
{
"epoch": 0.62,
"grad_norm": 7.966852766279544,
"learning_rate": 7.821473780273279e-06,
"loss": 0.5579,
"step": 586
},
{
"epoch": 0.63,
"grad_norm": 8.417280128263915,
"learning_rate": 7.785160320292812e-06,
"loss": 0.4942,
"step": 587
},
{
"epoch": 0.63,
"grad_norm": 4.894439384156987,
"learning_rate": 7.7488775474687e-06,
"loss": 0.5082,
"step": 588
},
{
"epoch": 0.63,
"grad_norm": 7.960251253611578,
"learning_rate": 7.712625964507818e-06,
"loss": 0.4997,
"step": 589
},
{
"epoch": 0.63,
"grad_norm": 8.640093158656311,
"learning_rate": 7.676406073684892e-06,
"loss": 0.5362,
"step": 590
},
{
"epoch": 0.63,
"grad_norm": 11.82405754186769,
"learning_rate": 7.64021837683554e-06,
"loss": 0.5292,
"step": 591
},
{
"epoch": 0.63,
"grad_norm": 5.258983750312472,
"learning_rate": 7.604063375349338e-06,
"loss": 0.547,
"step": 592
},
{
"epoch": 0.63,
"grad_norm": 10.013781925100409,
"learning_rate": 7.567941570162849e-06,
"loss": 0.538,
"step": 593
},
{
"epoch": 0.63,
"grad_norm": 7.8051606786973755,
"learning_rate": 7.531853461752696e-06,
"loss": 0.5122,
"step": 594
},
{
"epoch": 0.63,
"grad_norm": 7.2586272669558145,
"learning_rate": 7.495799550128625e-06,
"loss": 0.4468,
"step": 595
},
{
"epoch": 0.64,
"grad_norm": 5.774720107719125,
"learning_rate": 7.459780334826578e-06,
"loss": 0.5275,
"step": 596
},
{
"epoch": 0.64,
"grad_norm": 7.511213422739932,
"learning_rate": 7.423796314901769e-06,
"loss": 0.5768,
"step": 597
},
{
"epoch": 0.64,
"grad_norm": 13.06441655084729,
"learning_rate": 7.387847988921772e-06,
"loss": 0.6067,
"step": 598
},
{
"epoch": 0.64,
"grad_norm": 9.818511674108413,
"learning_rate": 7.351935854959608e-06,
"loss": 0.5669,
"step": 599
},
{
"epoch": 0.64,
"grad_norm": 12.085632705436097,
"learning_rate": 7.31606041058686e-06,
"loss": 0.6871,
"step": 600
},
{
"epoch": 0.64,
"grad_norm": 6.460535141416251,
"learning_rate": 7.2802221528667604e-06,
"loss": 0.4964,
"step": 601
},
{
"epoch": 0.64,
"grad_norm": 3.7229407170275715,
"learning_rate": 7.244421578347307e-06,
"loss": 0.4216,
"step": 602
},
{
"epoch": 0.64,
"grad_norm": 8.837577427746787,
"learning_rate": 7.208659183054393e-06,
"loss": 0.5951,
"step": 603
},
{
"epoch": 0.64,
"grad_norm": 4.340949956633562,
"learning_rate": 7.172935462484931e-06,
"loss": 0.3946,
"step": 604
},
{
"epoch": 0.64,
"grad_norm": 5.8235051717682405,
"learning_rate": 7.137250911599978e-06,
"loss": 0.4795,
"step": 605
},
{
"epoch": 0.65,
"grad_norm": 5.498783950116057,
"learning_rate": 7.101606024817889e-06,
"loss": 0.5192,
"step": 606
},
{
"epoch": 0.65,
"grad_norm": 3.1814150008692534,
"learning_rate": 7.066001296007469e-06,
"loss": 0.3841,
"step": 607
},
{
"epoch": 0.65,
"grad_norm": 5.5757391334979225,
"learning_rate": 7.030437218481116e-06,
"loss": 0.4353,
"step": 608
},
{
"epoch": 0.65,
"grad_norm": 13.254243576587063,
"learning_rate": 6.9949142849880015e-06,
"loss": 0.5673,
"step": 609
},
{
"epoch": 0.65,
"grad_norm": 4.1536903593774515,
"learning_rate": 6.959432987707223e-06,
"loss": 0.4192,
"step": 610
},
{
"epoch": 0.65,
"grad_norm": 3.4786455475133287,
"learning_rate": 6.9239938182410126e-06,
"loss": 0.4249,
"step": 611
},
{
"epoch": 0.65,
"grad_norm": 7.5677310862475915,
"learning_rate": 6.888597267607906e-06,
"loss": 0.4949,
"step": 612
},
{
"epoch": 0.65,
"grad_norm": 8.368860704109643,
"learning_rate": 6.8532438262359404e-06,
"loss": 0.4154,
"step": 613
},
{
"epoch": 0.65,
"grad_norm": 5.058418452049737,
"learning_rate": 6.817933983955867e-06,
"loss": 0.4951,
"step": 614
},
{
"epoch": 0.66,
"grad_norm": 6.42488252810342,
"learning_rate": 6.7826682299943635e-06,
"loss": 0.5948,
"step": 615
},
{
"epoch": 0.66,
"grad_norm": 7.090667184322379,
"learning_rate": 6.747447052967246e-06,
"loss": 0.564,
"step": 616
},
{
"epoch": 0.66,
"grad_norm": 9.915670648118702,
"learning_rate": 6.712270940872713e-06,
"loss": 0.4821,
"step": 617
},
{
"epoch": 0.66,
"grad_norm": 10.251562644163112,
"learning_rate": 6.677140381084573e-06,
"loss": 0.5258,
"step": 618
},
{
"epoch": 0.66,
"grad_norm": 5.692081829738655,
"learning_rate": 6.642055860345494e-06,
"loss": 0.4832,
"step": 619
},
{
"epoch": 0.66,
"grad_norm": 4.7513407845962545,
"learning_rate": 6.607017864760265e-06,
"loss": 0.4669,
"step": 620
},
{
"epoch": 0.66,
"grad_norm": 7.302874981115006,
"learning_rate": 6.572026879789064e-06,
"loss": 0.4492,
"step": 621
},
{
"epoch": 0.66,
"grad_norm": 13.730001115703725,
"learning_rate": 6.537083390240706e-06,
"loss": 0.5786,
"step": 622
},
{
"epoch": 0.66,
"grad_norm": 5.826147970750949,
"learning_rate": 6.502187880265969e-06,
"loss": 0.3817,
"step": 623
},
{
"epoch": 0.66,
"grad_norm": 8.854083264240913,
"learning_rate": 6.467340833350837e-06,
"loss": 0.4938,
"step": 624
},
{
"epoch": 0.67,
"grad_norm": 5.711323962217592,
"learning_rate": 6.43254273230985e-06,
"loss": 0.4934,
"step": 625
},
{
"epoch": 0.67,
"grad_norm": 6.916231992838763,
"learning_rate": 6.3977940592793766e-06,
"loss": 0.5225,
"step": 626
},
{
"epoch": 0.67,
"grad_norm": 9.605613693440635,
"learning_rate": 6.36309529571095e-06,
"loss": 0.527,
"step": 627
},
{
"epoch": 0.67,
"grad_norm": 7.226506391255898,
"learning_rate": 6.328446922364595e-06,
"loss": 0.4771,
"step": 628
},
{
"epoch": 0.67,
"grad_norm": 5.211969056714009,
"learning_rate": 6.293849419302179e-06,
"loss": 0.5099,
"step": 629
},
{
"epoch": 0.67,
"grad_norm": 11.660847177274526,
"learning_rate": 6.259303265880725e-06,
"loss": 0.5479,
"step": 630
},
{
"epoch": 0.67,
"grad_norm": 8.043560392139083,
"learning_rate": 6.224808940745814e-06,
"loss": 0.5182,
"step": 631
},
{
"epoch": 0.67,
"grad_norm": 8.052748988432281,
"learning_rate": 6.190366921824933e-06,
"loss": 0.6176,
"step": 632
},
{
"epoch": 0.67,
"grad_norm": 7.108683500222242,
"learning_rate": 6.155977686320837e-06,
"loss": 0.5878,
"step": 633
},
{
"epoch": 0.68,
"grad_norm": 11.823467597819771,
"learning_rate": 6.121641710704972e-06,
"loss": 0.4971,
"step": 634
},
{
"epoch": 0.68,
"grad_norm": 10.40651530100889,
"learning_rate": 6.087359470710841e-06,
"loss": 0.4427,
"step": 635
},
{
"epoch": 0.68,
"grad_norm": 11.976145453386437,
"learning_rate": 6.053131441327431e-06,
"loss": 0.5098,
"step": 636
},
{
"epoch": 0.68,
"grad_norm": 7.971563751964657,
"learning_rate": 6.018958096792642e-06,
"loss": 0.5695,
"step": 637
},
{
"epoch": 0.68,
"grad_norm": 6.129677385908451,
"learning_rate": 5.98483991058667e-06,
"loss": 0.4808,
"step": 638
},
{
"epoch": 0.68,
"grad_norm": 4.479658413986526,
"learning_rate": 5.950777355425511e-06,
"loss": 0.4775,
"step": 639
},
{
"epoch": 0.68,
"grad_norm": 4.309077539414612,
"learning_rate": 5.916770903254363e-06,
"loss": 0.4718,
"step": 640
},
{
"epoch": 0.68,
"grad_norm": 5.726658232065661,
"learning_rate": 5.8828210252411e-06,
"loss": 0.5147,
"step": 641
},
{
"epoch": 0.68,
"grad_norm": 6.432897045805533,
"learning_rate": 5.848928191769753e-06,
"loss": 0.5047,
"step": 642
},
{
"epoch": 0.69,
"grad_norm": 6.06082150623531,
"learning_rate": 5.815092872433994e-06,
"loss": 0.5296,
"step": 643
},
{
"epoch": 0.69,
"grad_norm": 7.760932921939526,
"learning_rate": 5.781315536030607e-06,
"loss": 0.5339,
"step": 644
},
{
"epoch": 0.69,
"grad_norm": 6.0611679377908825,
"learning_rate": 5.74759665055302e-06,
"loss": 0.4212,
"step": 645
},
{
"epoch": 0.69,
"grad_norm": 6.4875374667111565,
"learning_rate": 5.713936683184795e-06,
"loss": 0.5394,
"step": 646
},
{
"epoch": 0.69,
"grad_norm": 5.850168689232539,
"learning_rate": 5.680336100293182e-06,
"loss": 0.4512,
"step": 647
},
{
"epoch": 0.69,
"grad_norm": 9.444171703118819,
"learning_rate": 5.6467953674226505e-06,
"loss": 0.3942,
"step": 648
},
{
"epoch": 0.69,
"grad_norm": 8.582206095093797,
"learning_rate": 5.613314949288409e-06,
"loss": 0.4582,
"step": 649
},
{
"epoch": 0.69,
"grad_norm": 11.757390924671101,
"learning_rate": 5.579895309770009e-06,
"loss": 0.5435,
"step": 650
},
{
"epoch": 0.69,
"grad_norm": 5.02260927644731,
"learning_rate": 5.546536911904896e-06,
"loss": 0.4719,
"step": 651
},
{
"epoch": 0.69,
"grad_norm": 9.22536236771837,
"learning_rate": 5.513240217881992e-06,
"loss": 0.4668,
"step": 652
},
{
"epoch": 0.7,
"grad_norm": 10.157555843180285,
"learning_rate": 5.4800056890353025e-06,
"loss": 0.5409,
"step": 653
},
{
"epoch": 0.7,
"grad_norm": 12.323174404868091,
"learning_rate": 5.44683378583752e-06,
"loss": 0.5631,
"step": 654
},
{
"epoch": 0.7,
"grad_norm": 8.185618668543789,
"learning_rate": 5.4137249678936265e-06,
"loss": 0.4284,
"step": 655
},
{
"epoch": 0.7,
"grad_norm": 8.516855530899237,
"learning_rate": 5.380679693934569e-06,
"loss": 0.4221,
"step": 656
},
{
"epoch": 0.7,
"grad_norm": 11.008555199751408,
"learning_rate": 5.347698421810861e-06,
"loss": 0.5924,
"step": 657
},
{
"epoch": 0.7,
"grad_norm": 15.331344015670625,
"learning_rate": 5.31478160848625e-06,
"loss": 0.642,
"step": 658
},
{
"epoch": 0.7,
"grad_norm": 10.747880055046071,
"learning_rate": 5.2819297100314e-06,
"loss": 0.5417,
"step": 659
},
{
"epoch": 0.7,
"grad_norm": 15.482764407878273,
"learning_rate": 5.249143181617562e-06,
"loss": 0.63,
"step": 660
},
{
"epoch": 0.7,
"grad_norm": 4.174929188215974,
"learning_rate": 5.216422477510267e-06,
"loss": 0.4989,
"step": 661
},
{
"epoch": 0.71,
"grad_norm": 3.810136037355045,
"learning_rate": 5.183768051063036e-06,
"loss": 0.4893,
"step": 662
},
{
"epoch": 0.71,
"grad_norm": 6.8676100081227744,
"learning_rate": 5.151180354711087e-06,
"loss": 0.5768,
"step": 663
},
{
"epoch": 0.71,
"grad_norm": 10.044174931284044,
"learning_rate": 5.11865983996509e-06,
"loss": 0.4937,
"step": 664
},
{
"epoch": 0.71,
"grad_norm": 5.716263993029701,
"learning_rate": 5.0862069574048956e-06,
"loss": 0.4918,
"step": 665
},
{
"epoch": 0.71,
"grad_norm": 5.327730475446738,
"learning_rate": 5.053822156673276e-06,
"loss": 0.5112,
"step": 666
},
{
"epoch": 0.71,
"grad_norm": 9.70768865377853,
"learning_rate": 5.021505886469733e-06,
"loss": 0.5682,
"step": 667
},
{
"epoch": 0.71,
"grad_norm": 7.449826876587748,
"learning_rate": 4.989258594544252e-06,
"loss": 0.5914,
"step": 668
},
{
"epoch": 0.71,
"grad_norm": 6.246979693998428,
"learning_rate": 4.957080727691107e-06,
"loss": 0.5022,
"step": 669
},
{
"epoch": 0.71,
"grad_norm": 6.234169934241622,
"learning_rate": 4.9249727317426765e-06,
"loss": 0.5086,
"step": 670
},
{
"epoch": 0.71,
"grad_norm": 8.795959544866905,
"learning_rate": 4.892935051563243e-06,
"loss": 0.5631,
"step": 671
},
{
"epoch": 0.72,
"grad_norm": 6.585824557633543,
"learning_rate": 4.86096813104287e-06,
"loss": 0.5387,
"step": 672
},
{
"epoch": 0.72,
"grad_norm": 9.412528445371356,
"learning_rate": 4.829072413091219e-06,
"loss": 0.5903,
"step": 673
},
{
"epoch": 0.72,
"grad_norm": 4.934030658656449,
"learning_rate": 4.797248339631411e-06,
"loss": 0.5905,
"step": 674
},
{
"epoch": 0.72,
"grad_norm": 7.225794451726558,
"learning_rate": 4.765496351593927e-06,
"loss": 0.595,
"step": 675
},
{
"epoch": 0.72,
"grad_norm": 10.387771035955092,
"learning_rate": 4.733816888910483e-06,
"loss": 0.4901,
"step": 676
},
{
"epoch": 0.72,
"grad_norm": 5.585297554337063,
"learning_rate": 4.7022103905079405e-06,
"loss": 0.3997,
"step": 677
},
{
"epoch": 0.72,
"grad_norm": 5.2187286296388375,
"learning_rate": 4.67067729430222e-06,
"loss": 0.4856,
"step": 678
},
{
"epoch": 0.72,
"grad_norm": 3.4821810592954296,
"learning_rate": 4.639218037192235e-06,
"loss": 0.4361,
"step": 679
},
{
"epoch": 0.72,
"grad_norm": 3.959673908934417,
"learning_rate": 4.607833055053843e-06,
"loss": 0.3775,
"step": 680
},
{
"epoch": 0.73,
"grad_norm": 5.129251200576683,
"learning_rate": 4.576522782733802e-06,
"loss": 0.4947,
"step": 681
},
{
"epoch": 0.73,
"grad_norm": 4.8501294404892,
"learning_rate": 4.545287654043751e-06,
"loss": 0.3936,
"step": 682
},
{
"epoch": 0.73,
"grad_norm": 5.318966235176163,
"learning_rate": 4.514128101754183e-06,
"loss": 0.4649,
"step": 683
},
{
"epoch": 0.73,
"grad_norm": 7.407733161742192,
"learning_rate": 4.483044557588473e-06,
"loss": 0.5102,
"step": 684
},
{
"epoch": 0.73,
"grad_norm": 4.965311089759532,
"learning_rate": 4.45203745221688e-06,
"loss": 0.457,
"step": 685
},
{
"epoch": 0.73,
"grad_norm": 7.5517122869625135,
"learning_rate": 4.4211072152505866e-06,
"loss": 0.436,
"step": 686
},
{
"epoch": 0.73,
"grad_norm": 5.082245336395077,
"learning_rate": 4.3902542752357415e-06,
"loss": 0.4936,
"step": 687
},
{
"epoch": 0.73,
"grad_norm": 9.843843848674208,
"learning_rate": 4.359479059647527e-06,
"loss": 0.4336,
"step": 688
},
{
"epoch": 0.73,
"grad_norm": 5.231213013499434,
"learning_rate": 4.3287819948842334e-06,
"loss": 0.4247,
"step": 689
},
{
"epoch": 0.74,
"grad_norm": 6.305006534689214,
"learning_rate": 4.2981635062613556e-06,
"loss": 0.4826,
"step": 690
},
{
"epoch": 0.74,
"grad_norm": 5.533371229509814,
"learning_rate": 4.267624018005686e-06,
"loss": 0.4353,
"step": 691
},
{
"epoch": 0.74,
"grad_norm": 7.108534233713205,
"learning_rate": 4.237163953249453e-06,
"loss": 0.5991,
"step": 692
},
{
"epoch": 0.74,
"grad_norm": 4.76045650818351,
"learning_rate": 4.206783734024463e-06,
"loss": 0.4514,
"step": 693
},
{
"epoch": 0.74,
"grad_norm": 8.886091309499616,
"learning_rate": 4.176483781256224e-06,
"loss": 0.4023,
"step": 694
},
{
"epoch": 0.74,
"grad_norm": 7.524320382507302,
"learning_rate": 4.1462645147581456e-06,
"loss": 0.4514,
"step": 695
},
{
"epoch": 0.74,
"grad_norm": 7.482440196178911,
"learning_rate": 4.116126353225703e-06,
"loss": 0.5668,
"step": 696
},
{
"epoch": 0.74,
"grad_norm": 9.764506225206542,
"learning_rate": 4.086069714230646e-06,
"loss": 0.539,
"step": 697
},
{
"epoch": 0.74,
"grad_norm": 8.336719846814763,
"learning_rate": 4.056095014215208e-06,
"loss": 0.6426,
"step": 698
},
{
"epoch": 0.74,
"grad_norm": 4.148296296213888,
"learning_rate": 4.0262026684863295e-06,
"loss": 0.3376,
"step": 699
},
{
"epoch": 0.75,
"grad_norm": 9.203116559209336,
"learning_rate": 3.99639309120992e-06,
"loss": 0.5199,
"step": 700
},
{
"epoch": 0.75,
"grad_norm": 4.576118300820756,
"learning_rate": 3.96666669540512e-06,
"loss": 0.3731,
"step": 701
},
{
"epoch": 0.75,
"grad_norm": 9.971893856018951,
"learning_rate": 3.937023892938549e-06,
"loss": 0.591,
"step": 702
},
{
"epoch": 0.75,
"grad_norm": 5.809308796580302,
"learning_rate": 3.907465094518636e-06,
"loss": 0.5063,
"step": 703
},
{
"epoch": 0.75,
"grad_norm": 5.798443674221982,
"learning_rate": 3.877990709689905e-06,
"loss": 0.462,
"step": 704
},
{
"epoch": 0.75,
"grad_norm": 9.114741488810273,
"learning_rate": 3.8486011468273145e-06,
"loss": 0.5972,
"step": 705
},
{
"epoch": 0.75,
"grad_norm": 10.35491156365141,
"learning_rate": 3.819296813130588e-06,
"loss": 0.46,
"step": 706
},
{
"epoch": 0.75,
"grad_norm": 8.13256163586952,
"learning_rate": 3.790078114618586e-06,
"loss": 0.46,
"step": 707
},
{
"epoch": 0.75,
"grad_norm": 8.953602041543364,
"learning_rate": 3.760945456123656e-06,
"loss": 0.5296,
"step": 708
},
{
"epoch": 0.76,
"grad_norm": 6.830059498599811,
"learning_rate": 3.731899241286061e-06,
"loss": 0.4798,
"step": 709
},
{
"epoch": 0.76,
"grad_norm": 8.333772044542284,
"learning_rate": 3.702939872548348e-06,
"loss": 0.5452,
"step": 710
},
{
"epoch": 0.76,
"grad_norm": 7.7830518181331785,
"learning_rate": 3.6740677511497958e-06,
"loss": 0.52,
"step": 711
},
{
"epoch": 0.76,
"grad_norm": 9.48244358766429,
"learning_rate": 3.6452832771208524e-06,
"loss": 0.4604,
"step": 712
},
{
"epoch": 0.76,
"grad_norm": 5.325755301496669,
"learning_rate": 3.616586849277587e-06,
"loss": 0.4801,
"step": 713
},
{
"epoch": 0.76,
"grad_norm": 7.883051056212922,
"learning_rate": 3.587978865216165e-06,
"loss": 0.4396,
"step": 714
},
{
"epoch": 0.76,
"grad_norm": 7.231451053733015,
"learning_rate": 3.559459721307349e-06,
"loss": 0.5395,
"step": 715
},
{
"epoch": 0.76,
"grad_norm": 7.736510298846521,
"learning_rate": 3.5310298126909814e-06,
"loss": 0.4596,
"step": 716
},
{
"epoch": 0.76,
"grad_norm": 5.986545004035281,
"learning_rate": 3.5026895332705504e-06,
"loss": 0.4593,
"step": 717
},
{
"epoch": 0.77,
"grad_norm": 4.438626475367621,
"learning_rate": 3.474439275707698e-06,
"loss": 0.397,
"step": 718
},
{
"epoch": 0.77,
"grad_norm": 9.623477478063533,
"learning_rate": 3.4462794314167846e-06,
"loss": 0.509,
"step": 719
},
{
"epoch": 0.77,
"grad_norm": 9.394298451398283,
"learning_rate": 3.418210390559481e-06,
"loss": 0.4663,
"step": 720
},
{
"epoch": 0.77,
"grad_norm": 9.315469015310178,
"learning_rate": 3.390232542039352e-06,
"loss": 0.5065,
"step": 721
},
{
"epoch": 0.77,
"grad_norm": 6.092259066777256,
"learning_rate": 3.3623462734964697e-06,
"loss": 0.4477,
"step": 722
},
{
"epoch": 0.77,
"grad_norm": 4.319138234974373,
"learning_rate": 3.3345519713020445e-06,
"loss": 0.3925,
"step": 723
},
{
"epoch": 0.77,
"grad_norm": 6.358450139297925,
"learning_rate": 3.3068500205530596e-06,
"loss": 0.4679,
"step": 724
},
{
"epoch": 0.77,
"grad_norm": 7.701373209390687,
"learning_rate": 3.2792408050669634e-06,
"loss": 0.409,
"step": 725
},
{
"epoch": 0.77,
"grad_norm": 9.55351492544124,
"learning_rate": 3.2517247073763237e-06,
"loss": 0.4771,
"step": 726
},
{
"epoch": 0.77,
"grad_norm": 4.7780971794647025,
"learning_rate": 3.2243021087235336e-06,
"loss": 0.4689,
"step": 727
},
{
"epoch": 0.78,
"grad_norm": 7.114890807561552,
"learning_rate": 3.196973389055542e-06,
"loss": 0.4797,
"step": 728
},
{
"epoch": 0.78,
"grad_norm": 7.072041698936169,
"learning_rate": 3.16973892701858e-06,
"loss": 0.5768,
"step": 729
},
{
"epoch": 0.78,
"grad_norm": 5.961275677943624,
"learning_rate": 3.142599099952912e-06,
"loss": 0.3881,
"step": 730
},
{
"epoch": 0.78,
"grad_norm": 4.06251591074102,
"learning_rate": 3.115554283887614e-06,
"loss": 0.375,
"step": 731
},
{
"epoch": 0.78,
"grad_norm": 7.326058092329529,
"learning_rate": 3.088604853535361e-06,
"loss": 0.5388,
"step": 732
},
{
"epoch": 0.78,
"grad_norm": 5.274081659887408,
"learning_rate": 3.0617511822872337e-06,
"loss": 0.4742,
"step": 733
},
{
"epoch": 0.78,
"grad_norm": 4.793423500693722,
"learning_rate": 3.03499364220755e-06,
"loss": 0.4056,
"step": 734
},
{
"epoch": 0.78,
"grad_norm": 5.269835865291424,
"learning_rate": 3.0083326040286977e-06,
"loss": 0.4332,
"step": 735
},
{
"epoch": 0.78,
"grad_norm": 6.413197977206429,
"learning_rate": 2.9817684371460155e-06,
"loss": 0.5452,
"step": 736
},
{
"epoch": 0.79,
"grad_norm": 8.581131926197953,
"learning_rate": 2.9553015096126638e-06,
"loss": 0.4393,
"step": 737
},
{
"epoch": 0.79,
"grad_norm": 4.5160115188573045,
"learning_rate": 2.9289321881345257e-06,
"loss": 0.4465,
"step": 738
},
{
"epoch": 0.79,
"grad_norm": 5.2592392037223235,
"learning_rate": 2.902660838065131e-06,
"loss": 0.5082,
"step": 739
},
{
"epoch": 0.79,
"grad_norm": 7.05269405170704,
"learning_rate": 2.8764878234005876e-06,
"loss": 0.372,
"step": 740
},
{
"epoch": 0.79,
"grad_norm": 5.274091473047936,
"learning_rate": 2.8504135067745463e-06,
"loss": 0.4121,
"step": 741
},
{
"epoch": 0.79,
"grad_norm": 7.3781029800169655,
"learning_rate": 2.8244382494531675e-06,
"loss": 0.4067,
"step": 742
},
{
"epoch": 0.79,
"grad_norm": 8.32474239381458,
"learning_rate": 2.798562411330126e-06,
"loss": 0.3679,
"step": 743
},
{
"epoch": 0.79,
"grad_norm": 12.450031335391019,
"learning_rate": 2.772786350921608e-06,
"loss": 0.5521,
"step": 744
},
{
"epoch": 0.79,
"grad_norm": 7.249666510251534,
"learning_rate": 2.7471104253613645e-06,
"loss": 0.3938,
"step": 745
},
{
"epoch": 0.79,
"grad_norm": 7.2873098383670065,
"learning_rate": 2.721534990395752e-06,
"loss": 0.4884,
"step": 746
},
{
"epoch": 0.8,
"grad_norm": 5.438711289922824,
"learning_rate": 2.6960604003788014e-06,
"loss": 0.364,
"step": 747
},
{
"epoch": 0.8,
"grad_norm": 9.79198662580905,
"learning_rate": 2.670687008267313e-06,
"loss": 0.3911,
"step": 748
},
{
"epoch": 0.8,
"grad_norm": 11.420021502949318,
"learning_rate": 2.6454151656159666e-06,
"loss": 0.4454,
"step": 749
},
{
"epoch": 0.8,
"grad_norm": 8.204149460627178,
"learning_rate": 2.6202452225724463e-06,
"loss": 0.4276,
"step": 750
},
{
"epoch": 0.8,
"grad_norm": 10.999756173661007,
"learning_rate": 2.5951775278725956e-06,
"loss": 0.5623,
"step": 751
},
{
"epoch": 0.8,
"grad_norm": 8.568105116698392,
"learning_rate": 2.5702124288355723e-06,
"loss": 0.3292,
"step": 752
},
{
"epoch": 0.8,
"grad_norm": 5.56439861064704,
"learning_rate": 2.545350271359055e-06,
"loss": 0.3879,
"step": 753
},
{
"epoch": 0.8,
"grad_norm": 8.106162701050645,
"learning_rate": 2.520591399914444e-06,
"loss": 0.364,
"step": 754
},
{
"epoch": 0.8,
"grad_norm": 10.29859188596712,
"learning_rate": 2.495936157542074e-06,
"loss": 0.4791,
"step": 755
},
{
"epoch": 0.81,
"grad_norm": 7.657220905887133,
"learning_rate": 2.471384885846482e-06,
"loss": 0.4902,
"step": 756
},
{
"epoch": 0.81,
"grad_norm": 6.191995160106613,
"learning_rate": 2.4469379249916614e-06,
"loss": 0.4585,
"step": 757
},
{
"epoch": 0.81,
"grad_norm": 7.623947016373425,
"learning_rate": 2.422595613696356e-06,
"loss": 0.42,
"step": 758
},
{
"epoch": 0.81,
"grad_norm": 5.456113345068306,
"learning_rate": 2.3983582892293642e-06,
"loss": 0.3625,
"step": 759
},
{
"epoch": 0.81,
"grad_norm": 9.722994068985187,
"learning_rate": 2.3742262874048583e-06,
"loss": 0.5455,
"step": 760
},
{
"epoch": 0.81,
"grad_norm": 10.210377059221097,
"learning_rate": 2.3501999425777433e-06,
"loss": 0.4177,
"step": 761
},
{
"epoch": 0.81,
"grad_norm": 9.986453075138304,
"learning_rate": 2.326279587639031e-06,
"loss": 0.4519,
"step": 762
},
{
"epoch": 0.81,
"grad_norm": 7.444568050178271,
"learning_rate": 2.3024655540111984e-06,
"loss": 0.4707,
"step": 763
},
{
"epoch": 0.81,
"grad_norm": 6.882963377525352,
"learning_rate": 2.2787581716436257e-06,
"loss": 0.4142,
"step": 764
},
{
"epoch": 0.82,
"grad_norm": 5.621837753619931,
"learning_rate": 2.255157769008011e-06,
"loss": 0.3837,
"step": 765
},
{
"epoch": 0.82,
"grad_norm": 8.979013360736435,
"learning_rate": 2.2316646730938197e-06,
"loss": 0.4511,
"step": 766
},
{
"epoch": 0.82,
"grad_norm": 9.387346843179332,
"learning_rate": 2.2082792094037585e-06,
"loss": 0.3837,
"step": 767
},
{
"epoch": 0.82,
"grad_norm": 11.152697788037019,
"learning_rate": 2.185001701949263e-06,
"loss": 0.3815,
"step": 768
},
{
"epoch": 0.82,
"grad_norm": 7.316279559405673,
"learning_rate": 2.1618324732459993e-06,
"loss": 0.5763,
"step": 769
},
{
"epoch": 0.82,
"grad_norm": 4.60864460318581,
"learning_rate": 2.138771844309423e-06,
"loss": 0.3784,
"step": 770
},
{
"epoch": 0.82,
"grad_norm": 8.893733826091104,
"learning_rate": 2.1158201346502927e-06,
"loss": 0.3993,
"step": 771
},
{
"epoch": 0.82,
"grad_norm": 8.857140540604258,
"learning_rate": 2.092977662270275e-06,
"loss": 0.3334,
"step": 772
},
{
"epoch": 0.82,
"grad_norm": 9.000582629375513,
"learning_rate": 2.0702447436575223e-06,
"loss": 0.5377,
"step": 773
},
{
"epoch": 0.82,
"grad_norm": 11.407276883996886,
"learning_rate": 2.0476216937822967e-06,
"loss": 0.4657,
"step": 774
},
{
"epoch": 0.83,
"grad_norm": 9.343576283040107,
"learning_rate": 2.0251088260925967e-06,
"loss": 0.5038,
"step": 775
},
{
"epoch": 0.83,
"grad_norm": 5.742436296561597,
"learning_rate": 2.002706452509824e-06,
"loss": 0.462,
"step": 776
},
{
"epoch": 0.83,
"grad_norm": 9.651440289066235,
"learning_rate": 1.9804148834244465e-06,
"loss": 0.4588,
"step": 777
},
{
"epoch": 0.83,
"grad_norm": 5.795036174327087,
"learning_rate": 1.9582344276917233e-06,
"loss": 0.5565,
"step": 778
},
{
"epoch": 0.83,
"grad_norm": 7.475417877764932,
"learning_rate": 1.9361653926274016e-06,
"loss": 0.3644,
"step": 779
},
{
"epoch": 0.83,
"grad_norm": 7.2305893277419235,
"learning_rate": 1.9142080840034647e-06,
"loss": 0.451,
"step": 780
},
{
"epoch": 0.83,
"grad_norm": 7.9485644539877365,
"learning_rate": 1.8923628060439037e-06,
"loss": 0.5415,
"step": 781
},
{
"epoch": 0.83,
"grad_norm": 7.339019736495735,
"learning_rate": 1.8706298614204942e-06,
"loss": 0.4258,
"step": 782
},
{
"epoch": 0.83,
"grad_norm": 6.229226290130147,
"learning_rate": 1.8490095512486072e-06,
"loss": 0.372,
"step": 783
},
{
"epoch": 0.84,
"grad_norm": 8.318686997660855,
"learning_rate": 1.8275021750830368e-06,
"loss": 0.4337,
"step": 784
},
{
"epoch": 0.84,
"grad_norm": 11.383779868869713,
"learning_rate": 1.8061080309138379e-06,
"loss": 0.5517,
"step": 785
},
{
"epoch": 0.84,
"grad_norm": 4.319009051803344,
"learning_rate": 1.7848274151622236e-06,
"loss": 0.4032,
"step": 786
},
{
"epoch": 0.84,
"grad_norm": 6.434125961202376,
"learning_rate": 1.7636606226764353e-06,
"loss": 0.4165,
"step": 787
},
{
"epoch": 0.84,
"grad_norm": 5.041886706286037,
"learning_rate": 1.742607946727658e-06,
"loss": 0.4621,
"step": 788
},
{
"epoch": 0.84,
"grad_norm": 6.183213056857235,
"learning_rate": 1.7216696790059718e-06,
"loss": 0.423,
"step": 789
},
{
"epoch": 0.84,
"grad_norm": 6.373680744920821,
"learning_rate": 1.7008461096162999e-06,
"loss": 0.5464,
"step": 790
},
{
"epoch": 0.84,
"grad_norm": 11.493069464165945,
"learning_rate": 1.6801375270743925e-06,
"loss": 0.4518,
"step": 791
},
{
"epoch": 0.84,
"grad_norm": 9.805183898852514,
"learning_rate": 1.6595442183028266e-06,
"loss": 0.4836,
"step": 792
},
{
"epoch": 0.84,
"grad_norm": 8.05887231146552,
"learning_rate": 1.6390664686270342e-06,
"loss": 0.3711,
"step": 793
},
{
"epoch": 0.85,
"grad_norm": 6.975427069143952,
"learning_rate": 1.6187045617713459e-06,
"loss": 0.4149,
"step": 794
},
{
"epoch": 0.85,
"grad_norm": 7.272454723310934,
"learning_rate": 1.5984587798550633e-06,
"loss": 0.4357,
"step": 795
},
{
"epoch": 0.85,
"grad_norm": 6.941951052506783,
"learning_rate": 1.578329403388541e-06,
"loss": 0.3835,
"step": 796
},
{
"epoch": 0.85,
"grad_norm": 5.800438948527679,
"learning_rate": 1.5583167112693153e-06,
"loss": 0.4408,
"step": 797
},
{
"epoch": 0.85,
"grad_norm": 5.3480222782611175,
"learning_rate": 1.5384209807782291e-06,
"loss": 0.4553,
"step": 798
},
{
"epoch": 0.85,
"grad_norm": 5.909917855292465,
"learning_rate": 1.518642487575591e-06,
"loss": 0.497,
"step": 799
},
{
"epoch": 0.85,
"grad_norm": 7.683087880393552,
"learning_rate": 1.49898150569736e-06,
"loss": 0.4535,
"step": 800
},
{
"epoch": 0.85,
"grad_norm": 7.446327856244201,
"learning_rate": 1.4794383075513453e-06,
"loss": 0.3873,
"step": 801
},
{
"epoch": 0.85,
"grad_norm": 6.360170134685136,
"learning_rate": 1.460013163913433e-06,
"loss": 0.5037,
"step": 802
},
{
"epoch": 0.86,
"grad_norm": 5.053653216562189,
"learning_rate": 1.4407063439238333e-06,
"loss": 0.4954,
"step": 803
},
{
"epoch": 0.86,
"grad_norm": 4.2483525961998945,
"learning_rate": 1.421518115083359e-06,
"loss": 0.3751,
"step": 804
},
{
"epoch": 0.86,
"grad_norm": 6.345969600489226,
"learning_rate": 1.4024487432497013e-06,
"loss": 0.4143,
"step": 805
},
{
"epoch": 0.86,
"grad_norm": 5.646370329937249,
"learning_rate": 1.3834984926337657e-06,
"loss": 0.3376,
"step": 806
},
{
"epoch": 0.86,
"grad_norm": 10.69680196625431,
"learning_rate": 1.36466762579601e-06,
"loss": 0.5439,
"step": 807
},
{
"epoch": 0.86,
"grad_norm": 9.34187811581496,
"learning_rate": 1.3459564036427841e-06,
"loss": 0.4761,
"step": 808
},
{
"epoch": 0.86,
"grad_norm": 8.322291418258063,
"learning_rate": 1.3273650854227438e-06,
"loss": 0.4699,
"step": 809
},
{
"epoch": 0.86,
"grad_norm": 6.089668194777211,
"learning_rate": 1.3088939287232383e-06,
"loss": 0.4135,
"step": 810
},
{
"epoch": 0.86,
"grad_norm": 6.349508519106518,
"learning_rate": 1.2905431894667552e-06,
"loss": 0.4429,
"step": 811
},
{
"epoch": 0.87,
"grad_norm": 6.0067928128468955,
"learning_rate": 1.272313121907366e-06,
"loss": 0.5506,
"step": 812
},
{
"epoch": 0.87,
"grad_norm": 5.057193603204914,
"learning_rate": 1.2542039786272008e-06,
"loss": 0.4029,
"step": 813
},
{
"epoch": 0.87,
"grad_norm": 6.9264232438262,
"learning_rate": 1.236216010532959e-06,
"loss": 0.399,
"step": 814
},
{
"epoch": 0.87,
"grad_norm": 8.111117092935682,
"learning_rate": 1.218349466852432e-06,
"loss": 0.3969,
"step": 815
},
{
"epoch": 0.87,
"grad_norm": 12.617119783922481,
"learning_rate": 1.200604595131033e-06,
"loss": 0.3632,
"step": 816
},
{
"epoch": 0.87,
"grad_norm": 4.997880390654607,
"learning_rate": 1.1829816412283912e-06,
"loss": 0.3727,
"step": 817
},
{
"epoch": 0.87,
"grad_norm": 9.925641311100412,
"learning_rate": 1.1654808493149283e-06,
"loss": 0.4779,
"step": 818
},
{
"epoch": 0.87,
"grad_norm": 10.259671670493303,
"learning_rate": 1.1481024618684821e-06,
"loss": 0.4792,
"step": 819
},
{
"epoch": 0.87,
"grad_norm": 4.666467514007945,
"learning_rate": 1.1308467196709505e-06,
"loss": 0.3665,
"step": 820
},
{
"epoch": 0.87,
"grad_norm": 5.8390766285836895,
"learning_rate": 1.1137138618049403e-06,
"loss": 0.4426,
"step": 821
},
{
"epoch": 0.88,
"grad_norm": 4.079596026020399,
"learning_rate": 1.096704125650474e-06,
"loss": 0.3858,
"step": 822
},
{
"epoch": 0.88,
"grad_norm": 6.355034616703483,
"learning_rate": 1.079817746881696e-06,
"loss": 0.4322,
"step": 823
},
{
"epoch": 0.88,
"grad_norm": 4.964202739872203,
"learning_rate": 1.0630549594635896e-06,
"loss": 0.4246,
"step": 824
},
{
"epoch": 0.88,
"grad_norm": 9.981500108734732,
"learning_rate": 1.0464159956487596e-06,
"loss": 0.4048,
"step": 825
},
{
"epoch": 0.88,
"grad_norm": 5.314444267142196,
"learning_rate": 1.0299010859742009e-06,
"loss": 0.4226,
"step": 826
},
{
"epoch": 0.88,
"grad_norm": 6.711068748283136,
"learning_rate": 1.013510459258108e-06,
"loss": 0.4626,
"step": 827
},
{
"epoch": 0.88,
"grad_norm": 6.87746096070976,
"learning_rate": 9.972443425967026e-07,
"loss": 0.4762,
"step": 828
},
{
"epoch": 0.88,
"grad_norm": 5.584995566992966,
"learning_rate": 9.811029613610913e-07,
"loss": 0.4638,
"step": 829
},
{
"epoch": 0.88,
"grad_norm": 5.476555647757321,
"learning_rate": 9.650865391941345e-07,
"loss": 0.4174,
"step": 830
},
{
"epoch": 0.89,
"grad_norm": 5.334567697843729,
"learning_rate": 9.491952980073604e-07,
"loss": 0.3264,
"step": 831
},
{
"epoch": 0.89,
"grad_norm": 7.864733082511921,
"learning_rate": 9.334294579778813e-07,
"loss": 0.3613,
"step": 832
},
{
"epoch": 0.89,
"grad_norm": 8.345283693272092,
"learning_rate": 9.177892375453413e-07,
"loss": 0.5469,
"step": 833
},
{
"epoch": 0.89,
"grad_norm": 8.885940166363653,
"learning_rate": 9.022748534088988e-07,
"loss": 0.4455,
"step": 834
},
{
"epoch": 0.89,
"grad_norm": 7.55591370496932,
"learning_rate": 8.86886520524216e-07,
"loss": 0.4356,
"step": 835
},
{
"epoch": 0.89,
"grad_norm": 7.5174900881299465,
"learning_rate": 8.716244521004846e-07,
"loss": 0.4074,
"step": 836
},
{
"epoch": 0.89,
"grad_norm": 7.226452727128449,
"learning_rate": 8.564888595974718e-07,
"loss": 0.473,
"step": 837
},
{
"epoch": 0.89,
"grad_norm": 7.586424458860782,
"learning_rate": 8.414799527225826e-07,
"loss": 0.4757,
"step": 838
},
{
"epoch": 0.89,
"grad_norm": 5.51882334743589,
"learning_rate": 8.265979394279732e-07,
"loss": 0.4272,
"step": 839
},
{
"epoch": 0.9,
"grad_norm": 6.945131528688248,
"learning_rate": 8.118430259076493e-07,
"loss": 0.3003,
"step": 840
},
{
"epoch": 0.9,
"grad_norm": 7.957372619224033,
"learning_rate": 7.972154165946155e-07,
"loss": 0.4485,
"step": 841
},
{
"epoch": 0.9,
"grad_norm": 7.203632243202649,
"learning_rate": 7.827153141580501e-07,
"loss": 0.384,
"step": 842
},
{
"epoch": 0.9,
"grad_norm": 5.973567270845765,
"learning_rate": 7.683429195004932e-07,
"loss": 0.415,
"step": 843
},
{
"epoch": 0.9,
"grad_norm": 7.742846323788447,
"learning_rate": 7.540984317550548e-07,
"loss": 0.4261,
"step": 844
},
{
"epoch": 0.9,
"grad_norm": 4.971511012508995,
"learning_rate": 7.399820482826692e-07,
"loss": 0.3073,
"step": 845
},
{
"epoch": 0.9,
"grad_norm": 7.417662373162329,
"learning_rate": 7.25993964669347e-07,
"loss": 0.4293,
"step": 846
},
{
"epoch": 0.9,
"grad_norm": 8.474162845843916,
"learning_rate": 7.12134374723481e-07,
"loss": 0.5288,
"step": 847
},
{
"epoch": 0.9,
"grad_norm": 6.74629233498178,
"learning_rate": 6.984034704731479e-07,
"loss": 0.379,
"step": 848
},
{
"epoch": 0.9,
"grad_norm": 6.51287661591934,
"learning_rate": 6.848014421634497e-07,
"loss": 0.4254,
"step": 849
},
{
"epoch": 0.91,
"grad_norm": 8.203119397161476,
"learning_rate": 6.71328478253881e-07,
"loss": 0.4818,
"step": 850
},
{
"epoch": 0.91,
"grad_norm": 7.091765460672564,
"learning_rate": 6.579847654157234e-07,
"loss": 0.3937,
"step": 851
},
{
"epoch": 0.91,
"grad_norm": 5.848337659332669,
"learning_rate": 6.447704885294437e-07,
"loss": 0.3223,
"step": 852
},
{
"epoch": 0.91,
"grad_norm": 11.861790191542282,
"learning_rate": 6.316858306821449e-07,
"loss": 0.3955,
"step": 853
},
{
"epoch": 0.91,
"grad_norm": 5.89823677782504,
"learning_rate": 6.18730973165027e-07,
"loss": 0.52,
"step": 854
},
{
"epoch": 0.91,
"grad_norm": 7.297372429370729,
"learning_rate": 6.05906095470874e-07,
"loss": 0.5625,
"step": 855
},
{
"epoch": 0.91,
"grad_norm": 8.018266432717773,
"learning_rate": 5.932113752915659e-07,
"loss": 0.5943,
"step": 856
},
{
"epoch": 0.91,
"grad_norm": 5.441022930966068,
"learning_rate": 5.806469885156163e-07,
"loss": 0.3584,
"step": 857
},
{
"epoch": 0.91,
"grad_norm": 7.923651656498159,
"learning_rate": 5.682131092257381e-07,
"loss": 0.4374,
"step": 858
},
{
"epoch": 0.92,
"grad_norm": 4.401381730215806,
"learning_rate": 5.55909909696436e-07,
"loss": 0.4548,
"step": 859
},
{
"epoch": 0.92,
"grad_norm": 9.323436877274784,
"learning_rate": 5.437375603916039e-07,
"loss": 0.3949,
"step": 860
},
{
"epoch": 0.92,
"grad_norm": 7.6852951141865455,
"learning_rate": 5.316962299621808e-07,
"loss": 0.4357,
"step": 861
},
{
"epoch": 0.92,
"grad_norm": 5.468965570424503,
"learning_rate": 5.19786085243803e-07,
"loss": 0.4446,
"step": 862
},
{
"epoch": 0.92,
"grad_norm": 6.3812841292130855,
"learning_rate": 5.080072912544987e-07,
"loss": 0.4428,
"step": 863
},
{
"epoch": 0.92,
"grad_norm": 10.852851737536733,
"learning_rate": 4.96360011192396e-07,
"loss": 0.4991,
"step": 864
},
{
"epoch": 0.92,
"grad_norm": 5.7584503904522775,
"learning_rate": 4.848444064334679e-07,
"loss": 0.5114,
"step": 865
},
{
"epoch": 0.92,
"grad_norm": 6.376038086977472,
"learning_rate": 4.734606365292871e-07,
"loss": 0.3782,
"step": 866
},
{
"epoch": 0.92,
"grad_norm": 6.224991564333229,
"learning_rate": 4.6220885920483014e-07,
"loss": 0.3527,
"step": 867
},
{
"epoch": 0.92,
"grad_norm": 5.203231071738429,
"learning_rate": 4.510892303562797e-07,
"loss": 0.3518,
"step": 868
},
{
"epoch": 0.93,
"grad_norm": 5.884197350261294,
"learning_rate": 4.401019040488652e-07,
"loss": 0.4493,
"step": 869
},
{
"epoch": 0.93,
"grad_norm": 8.438339641746209,
"learning_rate": 4.2924703251473355e-07,
"loss": 0.5192,
"step": 870
},
{
"epoch": 0.93,
"grad_norm": 9.929683884140545,
"learning_rate": 4.1852476615083957e-07,
"loss": 0.4769,
"step": 871
},
{
"epoch": 0.93,
"grad_norm": 7.444778653156325,
"learning_rate": 4.0793525351685905e-07,
"loss": 0.3915,
"step": 872
},
{
"epoch": 0.93,
"grad_norm": 5.55575801178513,
"learning_rate": 3.974786413331311e-07,
"loss": 0.3861,
"step": 873
},
{
"epoch": 0.93,
"grad_norm": 5.154553375207641,
"learning_rate": 3.8715507447862455e-07,
"loss": 0.3713,
"step": 874
},
{
"epoch": 0.93,
"grad_norm": 9.250721341718682,
"learning_rate": 3.7696469598893727e-07,
"loss": 0.4575,
"step": 875
},
{
"epoch": 0.93,
"grad_norm": 5.654967363493775,
"learning_rate": 3.669076470543054e-07,
"loss": 0.4686,
"step": 876
},
{
"epoch": 0.93,
"grad_norm": 5.710919914401177,
"learning_rate": 3.569840670176483e-07,
"loss": 0.4339,
"step": 877
},
{
"epoch": 0.94,
"grad_norm": 6.7079973379018085,
"learning_rate": 3.4719409337264473e-07,
"loss": 0.4238,
"step": 878
},
{
"epoch": 0.94,
"grad_norm": 10.924750798758199,
"learning_rate": 3.3753786176182303e-07,
"loss": 0.4977,
"step": 879
},
{
"epoch": 0.94,
"grad_norm": 4.9608766209441315,
"learning_rate": 3.2801550597468033e-07,
"loss": 0.2732,
"step": 880
},
{
"epoch": 0.94,
"grad_norm": 7.448404330000091,
"learning_rate": 3.186271579458333e-07,
"loss": 0.4483,
"step": 881
},
{
"epoch": 0.94,
"grad_norm": 5.022846446335248,
"learning_rate": 3.093729477531826e-07,
"loss": 0.4363,
"step": 882
},
{
"epoch": 0.94,
"grad_norm": 7.140771043711367,
"learning_rate": 3.002530036161222e-07,
"loss": 0.4566,
"step": 883
},
{
"epoch": 0.94,
"grad_norm": 5.57307898414647,
"learning_rate": 2.912674518937553e-07,
"loss": 0.3565,
"step": 884
},
{
"epoch": 0.94,
"grad_norm": 7.503093403901029,
"learning_rate": 2.824164170831389e-07,
"loss": 0.3495,
"step": 885
},
{
"epoch": 0.94,
"grad_norm": 8.266577699990709,
"learning_rate": 2.737000218175712e-07,
"loss": 0.458,
"step": 886
},
{
"epoch": 0.95,
"grad_norm": 8.007643327554275,
"learning_rate": 2.651183868648821e-07,
"loss": 0.3738,
"step": 887
},
{
"epoch": 0.95,
"grad_norm": 6.8532106769801215,
"learning_rate": 2.566716311257655e-07,
"loss": 0.4602,
"step": 888
},
{
"epoch": 0.95,
"grad_norm": 8.226940714871725,
"learning_rate": 2.483598716321289e-07,
"loss": 0.3939,
"step": 889
},
{
"epoch": 0.95,
"grad_norm": 7.35412581775661,
"learning_rate": 2.4018322354547483e-07,
"loss": 0.4912,
"step": 890
},
{
"epoch": 0.95,
"grad_norm": 5.189789718764145,
"learning_rate": 2.321418001553022e-07,
"loss": 0.3516,
"step": 891
},
{
"epoch": 0.95,
"grad_norm": 5.643929108838239,
"learning_rate": 2.2423571287753853e-07,
"loss": 0.3811,
"step": 892
},
{
"epoch": 0.95,
"grad_norm": 5.764646780403799,
"learning_rate": 2.1646507125299587e-07,
"loss": 0.4442,
"step": 893
},
{
"epoch": 0.95,
"grad_norm": 5.42774429745915,
"learning_rate": 2.0882998294585177e-07,
"loss": 0.3746,
"step": 894
},
{
"epoch": 0.95,
"grad_norm": 6.171794557047173,
"learning_rate": 2.013305537421606e-07,
"loss": 0.447,
"step": 895
},
{
"epoch": 0.95,
"grad_norm": 5.130410450658874,
"learning_rate": 1.9396688754838355e-07,
"loss": 0.3469,
"step": 896
},
{
"epoch": 0.96,
"grad_norm": 7.39720002220979,
"learning_rate": 1.867390863899543e-07,
"loss": 0.3783,
"step": 897
},
{
"epoch": 0.96,
"grad_norm": 7.90865739350042,
"learning_rate": 1.796472504098612e-07,
"loss": 0.4422,
"step": 898
},
{
"epoch": 0.96,
"grad_norm": 5.815959678060295,
"learning_rate": 1.726914778672606e-07,
"loss": 0.378,
"step": 899
},
{
"epoch": 0.96,
"grad_norm": 5.081919878256831,
"learning_rate": 1.6587186513611685e-07,
"loss": 0.3205,
"step": 900
},
{
"epoch": 0.96,
"grad_norm": 6.567585188289477,
"learning_rate": 1.5918850670386677e-07,
"loss": 0.44,
"step": 901
},
{
"epoch": 0.96,
"grad_norm": 7.533656800692689,
"learning_rate": 1.5264149517010718e-07,
"loss": 0.4362,
"step": 902
},
{
"epoch": 0.96,
"grad_norm": 5.4628885270524865,
"learning_rate": 1.4623092124531613e-07,
"loss": 0.4497,
"step": 903
},
{
"epoch": 0.96,
"grad_norm": 5.858826602545033,
"learning_rate": 1.399568737495971e-07,
"loss": 0.43,
"step": 904
},
{
"epoch": 0.96,
"grad_norm": 8.572231557747521,
"learning_rate": 1.3381943961144118e-07,
"loss": 0.3925,
"step": 905
},
{
"epoch": 0.97,
"grad_norm": 5.873954819161939,
"learning_rate": 1.2781870386653016e-07,
"loss": 0.3683,
"step": 906
},
{
"epoch": 0.97,
"grad_norm": 4.581416916538669,
"learning_rate": 1.2195474965655652e-07,
"loss": 0.2704,
"step": 907
},
{
"epoch": 0.97,
"grad_norm": 6.965146521509025,
"learning_rate": 1.1622765822806636e-07,
"loss": 0.3669,
"step": 908
},
{
"epoch": 0.97,
"grad_norm": 5.699569364212127,
"learning_rate": 1.1063750893134273e-07,
"loss": 0.4125,
"step": 909
},
{
"epoch": 0.97,
"grad_norm": 7.0422218435015855,
"learning_rate": 1.0518437921929747e-07,
"loss": 0.3156,
"step": 910
},
{
"epoch": 0.97,
"grad_norm": 9.329828931384606,
"learning_rate": 9.986834464640328e-08,
"loss": 0.4834,
"step": 911
},
{
"epoch": 0.97,
"grad_norm": 5.211053129225831,
"learning_rate": 9.468947886764557e-08,
"loss": 0.4188,
"step": 912
},
{
"epoch": 0.97,
"grad_norm": 5.70280316894895,
"learning_rate": 8.964785363750228e-08,
"loss": 0.3474,
"step": 913
},
{
"epoch": 0.97,
"grad_norm": 5.824204755244337,
"learning_rate": 8.474353880894904e-08,
"loss": 0.2739,
"step": 914
},
{
"epoch": 0.97,
"grad_norm": 7.610004827598348,
"learning_rate": 7.997660233249105e-08,
"loss": 0.4407,
"step": 915
},
{
"epoch": 0.98,
"grad_norm": 6.558806283300632,
"learning_rate": 7.534711025522168e-08,
"loss": 0.3602,
"step": 916
},
{
"epoch": 0.98,
"grad_norm": 6.692057122360635,
"learning_rate": 7.08551267199098e-08,
"loss": 0.4439,
"step": 917
},
{
"epoch": 0.98,
"grad_norm": 6.4300501195085555,
"learning_rate": 6.650071396410718e-08,
"loss": 0.428,
"step": 918
},
{
"epoch": 0.98,
"grad_norm": 6.994677903035368,
"learning_rate": 6.22839323192892e-08,
"loss": 0.517,
"step": 919
},
{
"epoch": 0.98,
"grad_norm": 7.2674316715963325,
"learning_rate": 5.820484021001993e-08,
"loss": 0.3041,
"step": 920
},
{
"epoch": 0.98,
"grad_norm": 8.22191067382583,
"learning_rate": 5.426349415313503e-08,
"loss": 0.4481,
"step": 921
},
{
"epoch": 0.98,
"grad_norm": 5.065088639546076,
"learning_rate": 5.0459948756967914e-08,
"loss": 0.3333,
"step": 922
},
{
"epoch": 0.98,
"grad_norm": 8.073376899592944,
"learning_rate": 4.679425672059035e-08,
"loss": 0.4493,
"step": 923
},
{
"epoch": 0.98,
"grad_norm": 6.784518754930513,
"learning_rate": 4.326646883307972e-08,
"loss": 0.4454,
"step": 924
},
{
"epoch": 0.99,
"grad_norm": 8.57495143340707,
"learning_rate": 3.987663397281627e-08,
"loss": 0.4892,
"step": 925
},
{
"epoch": 0.99,
"grad_norm": 5.72551957891776,
"learning_rate": 3.662479910681027e-08,
"loss": 0.4259,
"step": 926
},
{
"epoch": 0.99,
"grad_norm": 4.571968724650214,
"learning_rate": 3.3511009290042585e-08,
"loss": 0.364,
"step": 927
},
{
"epoch": 0.99,
"grad_norm": 7.563468099974791,
"learning_rate": 3.0535307664848466e-08,
"loss": 0.4774,
"step": 928
},
{
"epoch": 0.99,
"grad_norm": 6.8428676776213555,
"learning_rate": 2.7697735460316954e-08,
"loss": 0.4322,
"step": 929
},
{
"epoch": 0.99,
"grad_norm": 3.9770591980462786,
"learning_rate": 2.499833199171908e-08,
"loss": 0.3225,
"step": 930
},
{
"epoch": 0.99,
"grad_norm": 5.666940466332362,
"learning_rate": 2.2437134659962777e-08,
"loss": 0.4234,
"step": 931
},
{
"epoch": 0.99,
"grad_norm": 5.153046102603702,
"learning_rate": 2.0014178951076602e-08,
"loss": 0.471,
"step": 932
},
{
"epoch": 0.99,
"grad_norm": 5.709388029240762,
"learning_rate": 1.7729498435716808e-08,
"loss": 0.4391,
"step": 933
},
{
"epoch": 1.0,
"grad_norm": 5.983125443983794,
"learning_rate": 1.5583124768701052e-08,
"loss": 0.483,
"step": 934
},
{
"epoch": 1.0,
"grad_norm": 6.445238846878224,
"learning_rate": 1.3575087688570965e-08,
"loss": 0.3558,
"step": 935
},
{
"epoch": 1.0,
"grad_norm": 6.816452865113282,
"learning_rate": 1.1705415017183585e-08,
"loss": 0.413,
"step": 936
},
{
"epoch": 1.0,
"grad_norm": 5.457722190574727,
"learning_rate": 9.974132659319457e-09,
"loss": 0.3282,
"step": 937
},
{
"epoch": 1.0,
"grad_norm": 6.969886154201284,
"learning_rate": 8.381264602327354e-09,
"loss": 0.438,
"step": 938
},
{
"epoch": 1.0,
"step": 938,
"total_flos": 373627158528000.0,
"train_loss": 0.7472518730773601,
"train_runtime": 19046.2174,
"train_samples_per_second": 6.306,
"train_steps_per_second": 0.049
}
],
"logging_steps": 1.0,
"max_steps": 938,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 200,
"total_flos": 373627158528000.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}