tktung's picture
Upload folder using huggingface_hub
4326884 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.9920127795527156,
"eval_steps": 500,
"global_step": 624,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.006389776357827476,
"grad_norm": 2.055291493195234,
"learning_rate": 3.125e-06,
"loss": 1.695,
"step": 1
},
{
"epoch": 0.012779552715654952,
"grad_norm": 2.0685233500522586,
"learning_rate": 6.25e-06,
"loss": 1.6748,
"step": 2
},
{
"epoch": 0.025559105431309903,
"grad_norm": 2.325735299422439,
"learning_rate": 1.25e-05,
"loss": 1.6964,
"step": 4
},
{
"epoch": 0.038338658146964855,
"grad_norm": 0.4729866673863026,
"learning_rate": 1.8750000000000002e-05,
"loss": 1.4325,
"step": 6
},
{
"epoch": 0.051118210862619806,
"grad_norm": 0.482620239981458,
"learning_rate": 2.5e-05,
"loss": 1.3874,
"step": 8
},
{
"epoch": 0.06389776357827476,
"grad_norm": 1.6728433474079003,
"learning_rate": 3.125e-05,
"loss": 1.4689,
"step": 10
},
{
"epoch": 0.07667731629392971,
"grad_norm": 0.3405987431283081,
"learning_rate": 3.7500000000000003e-05,
"loss": 1.3127,
"step": 12
},
{
"epoch": 0.08945686900958466,
"grad_norm": 0.2323496464888272,
"learning_rate": 4.375e-05,
"loss": 1.2639,
"step": 14
},
{
"epoch": 0.10223642172523961,
"grad_norm": 0.18809974511784008,
"learning_rate": 5e-05,
"loss": 1.2401,
"step": 16
},
{
"epoch": 0.11501597444089456,
"grad_norm": 0.18997340619225084,
"learning_rate": 5.6250000000000005e-05,
"loss": 1.2084,
"step": 18
},
{
"epoch": 0.12779552715654952,
"grad_norm": 0.15504216343509883,
"learning_rate": 6.25e-05,
"loss": 1.1855,
"step": 20
},
{
"epoch": 0.14057507987220447,
"grad_norm": 0.12848416587626313,
"learning_rate": 6.875e-05,
"loss": 1.146,
"step": 22
},
{
"epoch": 0.15335463258785942,
"grad_norm": 0.09889252813730416,
"learning_rate": 7.500000000000001e-05,
"loss": 1.1357,
"step": 24
},
{
"epoch": 0.16613418530351437,
"grad_norm": 0.09024188902019939,
"learning_rate": 8.125000000000001e-05,
"loss": 1.1096,
"step": 26
},
{
"epoch": 0.17891373801916932,
"grad_norm": 0.08133676595279006,
"learning_rate": 8.75e-05,
"loss": 1.0913,
"step": 28
},
{
"epoch": 0.19169329073482427,
"grad_norm": 0.0978463769637292,
"learning_rate": 9.375e-05,
"loss": 1.0679,
"step": 30
},
{
"epoch": 0.20447284345047922,
"grad_norm": 0.07943889170723487,
"learning_rate": 0.0001,
"loss": 1.075,
"step": 32
},
{
"epoch": 0.21725239616613418,
"grad_norm": 0.08240884428512509,
"learning_rate": 9.99971838728789e-05,
"loss": 1.075,
"step": 34
},
{
"epoch": 0.23003194888178913,
"grad_norm": 0.08253986997481327,
"learning_rate": 9.998873580873848e-05,
"loss": 1.0652,
"step": 36
},
{
"epoch": 0.24281150159744408,
"grad_norm": 0.07954648039103362,
"learning_rate": 9.997465675921163e-05,
"loss": 1.0519,
"step": 38
},
{
"epoch": 0.25559105431309903,
"grad_norm": 0.0776223200815433,
"learning_rate": 9.995494831023409e-05,
"loss": 1.0094,
"step": 40
},
{
"epoch": 0.268370607028754,
"grad_norm": 0.08000844411167178,
"learning_rate": 9.992961268186573e-05,
"loss": 1.0074,
"step": 42
},
{
"epoch": 0.28115015974440893,
"grad_norm": 0.0689657212250583,
"learning_rate": 9.989865272804063e-05,
"loss": 1.0087,
"step": 44
},
{
"epoch": 0.2939297124600639,
"grad_norm": 0.0722150479128947,
"learning_rate": 9.986207193624536e-05,
"loss": 1.0067,
"step": 46
},
{
"epoch": 0.30670926517571884,
"grad_norm": 0.06646168454668608,
"learning_rate": 9.981987442712633e-05,
"loss": 0.9837,
"step": 48
},
{
"epoch": 0.3194888178913738,
"grad_norm": 0.06815852582234988,
"learning_rate": 9.977206495402554e-05,
"loss": 1.0024,
"step": 50
},
{
"epoch": 0.33226837060702874,
"grad_norm": 0.07469571057420442,
"learning_rate": 9.971864890244513e-05,
"loss": 0.9606,
"step": 52
},
{
"epoch": 0.3450479233226837,
"grad_norm": 0.07160841663430713,
"learning_rate": 9.965963228944078e-05,
"loss": 0.9681,
"step": 54
},
{
"epoch": 0.35782747603833864,
"grad_norm": 0.06954866095292117,
"learning_rate": 9.959502176294383e-05,
"loss": 0.951,
"step": 56
},
{
"epoch": 0.3706070287539936,
"grad_norm": 0.06598684065212063,
"learning_rate": 9.95248246010126e-05,
"loss": 0.9501,
"step": 58
},
{
"epoch": 0.38338658146964855,
"grad_norm": 0.12103302407814338,
"learning_rate": 9.944904871101228e-05,
"loss": 0.9713,
"step": 60
},
{
"epoch": 0.3961661341853035,
"grad_norm": 0.07330981053456032,
"learning_rate": 9.936770262872443e-05,
"loss": 0.9283,
"step": 62
},
{
"epoch": 0.40894568690095845,
"grad_norm": 0.06537535724415816,
"learning_rate": 9.928079551738543e-05,
"loss": 0.9118,
"step": 64
},
{
"epoch": 0.4217252396166134,
"grad_norm": 0.07457609795137939,
"learning_rate": 9.918833716665419e-05,
"loss": 0.9279,
"step": 66
},
{
"epoch": 0.43450479233226835,
"grad_norm": 0.07491122165043795,
"learning_rate": 9.909033799150946e-05,
"loss": 0.935,
"step": 68
},
{
"epoch": 0.4472843450479233,
"grad_norm": 0.06781283989008571,
"learning_rate": 9.898680903107666e-05,
"loss": 0.9361,
"step": 70
},
{
"epoch": 0.46006389776357826,
"grad_norm": 0.07160916695151898,
"learning_rate": 9.887776194738432e-05,
"loss": 0.9159,
"step": 72
},
{
"epoch": 0.4728434504792332,
"grad_norm": 0.0681941013678725,
"learning_rate": 9.876320902405042e-05,
"loss": 0.8779,
"step": 74
},
{
"epoch": 0.48562300319488816,
"grad_norm": 0.07482319269062407,
"learning_rate": 9.864316316489873e-05,
"loss": 0.8825,
"step": 76
},
{
"epoch": 0.4984025559105431,
"grad_norm": 0.08697975313543096,
"learning_rate": 9.851763789250525e-05,
"loss": 0.922,
"step": 78
},
{
"epoch": 0.5111821086261981,
"grad_norm": 0.09978612068745818,
"learning_rate": 9.838664734667495e-05,
"loss": 0.8894,
"step": 80
},
{
"epoch": 0.5239616613418531,
"grad_norm": 0.09384667638421956,
"learning_rate": 9.825020628284896e-05,
"loss": 0.8593,
"step": 82
},
{
"epoch": 0.536741214057508,
"grad_norm": 0.06932081799385038,
"learning_rate": 9.810833007044247e-05,
"loss": 0.8662,
"step": 84
},
{
"epoch": 0.549520766773163,
"grad_norm": 0.10358699944795004,
"learning_rate": 9.796103469111351e-05,
"loss": 0.8723,
"step": 86
},
{
"epoch": 0.5623003194888179,
"grad_norm": 0.07169243369499742,
"learning_rate": 9.780833673696254e-05,
"loss": 0.8482,
"step": 88
},
{
"epoch": 0.5750798722044729,
"grad_norm": 0.1050406308556227,
"learning_rate": 9.76502534086636e-05,
"loss": 0.8496,
"step": 90
},
{
"epoch": 0.5878594249201278,
"grad_norm": 0.07201905690967678,
"learning_rate": 9.74868025135266e-05,
"loss": 0.8291,
"step": 92
},
{
"epoch": 0.6006389776357828,
"grad_norm": 1.2625349021090781,
"learning_rate": 9.731800246349148e-05,
"loss": 0.8503,
"step": 94
},
{
"epoch": 0.6134185303514377,
"grad_norm": 0.17981258022070712,
"learning_rate": 9.714387227305422e-05,
"loss": 0.8231,
"step": 96
},
{
"epoch": 0.6261980830670927,
"grad_norm": 0.07561478832740967,
"learning_rate": 9.696443155712486e-05,
"loss": 0.8119,
"step": 98
},
{
"epoch": 0.6389776357827476,
"grad_norm": 0.08195686915168865,
"learning_rate": 9.67797005288181e-05,
"loss": 0.7926,
"step": 100
},
{
"epoch": 0.6517571884984026,
"grad_norm": 0.0890476280007116,
"learning_rate": 9.65896999971763e-05,
"loss": 0.8039,
"step": 102
},
{
"epoch": 0.6645367412140575,
"grad_norm": 0.07738578891457887,
"learning_rate": 9.639445136482548e-05,
"loss": 0.7721,
"step": 104
},
{
"epoch": 0.6773162939297125,
"grad_norm": 0.0743037172920425,
"learning_rate": 9.619397662556435e-05,
"loss": 0.794,
"step": 106
},
{
"epoch": 0.6900958466453674,
"grad_norm": 0.08803835897602165,
"learning_rate": 9.598829836188694e-05,
"loss": 0.7721,
"step": 108
},
{
"epoch": 0.7028753993610224,
"grad_norm": 0.07702819696223887,
"learning_rate": 9.577743974243874e-05,
"loss": 0.7765,
"step": 110
},
{
"epoch": 0.7156549520766773,
"grad_norm": 0.07473535070111323,
"learning_rate": 9.55614245194068e-05,
"loss": 0.7598,
"step": 112
},
{
"epoch": 0.7284345047923323,
"grad_norm": 0.08433756541496004,
"learning_rate": 9.534027702584425e-05,
"loss": 0.7727,
"step": 114
},
{
"epoch": 0.7412140575079872,
"grad_norm": 0.07483257658817612,
"learning_rate": 9.511402217292926e-05,
"loss": 0.7465,
"step": 116
},
{
"epoch": 0.7539936102236422,
"grad_norm": 0.0880318685591304,
"learning_rate": 9.488268544715896e-05,
"loss": 0.7321,
"step": 118
},
{
"epoch": 0.7667731629392971,
"grad_norm": 0.07719604899450865,
"learning_rate": 9.464629290747842e-05,
"loss": 0.7624,
"step": 120
},
{
"epoch": 0.7795527156549521,
"grad_norm": 0.0733176421376437,
"learning_rate": 9.440487118234535e-05,
"loss": 0.6975,
"step": 122
},
{
"epoch": 0.792332268370607,
"grad_norm": 0.07051701385784455,
"learning_rate": 9.415844746673047e-05,
"loss": 0.7127,
"step": 124
},
{
"epoch": 0.805111821086262,
"grad_norm": 0.0729787562181977,
"learning_rate": 9.390704951905411e-05,
"loss": 0.7503,
"step": 126
},
{
"epoch": 0.8178913738019169,
"grad_norm": 0.07128874732953779,
"learning_rate": 9.365070565805941e-05,
"loss": 0.6941,
"step": 128
},
{
"epoch": 0.8306709265175719,
"grad_norm": 0.07804844381711577,
"learning_rate": 9.338944475962237e-05,
"loss": 0.7197,
"step": 130
},
{
"epoch": 0.8434504792332268,
"grad_norm": 0.08207580744924538,
"learning_rate": 9.312329625349902e-05,
"loss": 0.7134,
"step": 132
},
{
"epoch": 0.8562300319488818,
"grad_norm": 0.10268159904999394,
"learning_rate": 9.285229012001047e-05,
"loss": 0.705,
"step": 134
},
{
"epoch": 0.8690095846645367,
"grad_norm": 0.07097527094154266,
"learning_rate": 9.257645688666556e-05,
"loss": 0.7036,
"step": 136
},
{
"epoch": 0.8817891373801917,
"grad_norm": 0.07284443178958877,
"learning_rate": 9.22958276247223e-05,
"loss": 0.7313,
"step": 138
},
{
"epoch": 0.8945686900958466,
"grad_norm": 0.07294697279525543,
"learning_rate": 9.201043394568773e-05,
"loss": 0.6847,
"step": 140
},
{
"epoch": 0.9073482428115016,
"grad_norm": 0.0725032039002937,
"learning_rate": 9.172030799775699e-05,
"loss": 0.6877,
"step": 142
},
{
"epoch": 0.9201277955271565,
"grad_norm": 0.06708836437156662,
"learning_rate": 9.142548246219212e-05,
"loss": 0.6837,
"step": 144
},
{
"epoch": 0.9329073482428115,
"grad_norm": 0.07361178534656698,
"learning_rate": 9.112599054964057e-05,
"loss": 0.6522,
"step": 146
},
{
"epoch": 0.9456869009584664,
"grad_norm": 0.06961060997060975,
"learning_rate": 9.082186599639428e-05,
"loss": 0.6732,
"step": 148
},
{
"epoch": 0.9584664536741214,
"grad_norm": 0.06369267112915664,
"learning_rate": 9.051314306058933e-05,
"loss": 0.6615,
"step": 150
},
{
"epoch": 0.9712460063897763,
"grad_norm": 0.06667729772792583,
"learning_rate": 9.019985651834703e-05,
"loss": 0.6742,
"step": 152
},
{
"epoch": 0.9840255591054313,
"grad_norm": 0.07052786453330319,
"learning_rate": 8.988204165985649e-05,
"loss": 0.6365,
"step": 154
},
{
"epoch": 0.9968051118210862,
"grad_norm": 0.06352217971127558,
"learning_rate": 8.955973428539944e-05,
"loss": 0.6531,
"step": 156
},
{
"epoch": 1.011182108626198,
"grad_norm": 0.0907023898699884,
"learning_rate": 8.923297070131737e-05,
"loss": 0.6986,
"step": 158
},
{
"epoch": 1.023961661341853,
"grad_norm": 0.06588723514264389,
"learning_rate": 8.890178771592199e-05,
"loss": 0.4221,
"step": 160
},
{
"epoch": 1.036741214057508,
"grad_norm": 0.07457104912562523,
"learning_rate": 8.856622263534875e-05,
"loss": 0.4375,
"step": 162
},
{
"epoch": 1.049520766773163,
"grad_norm": 0.08716030746078077,
"learning_rate": 8.822631325935463e-05,
"loss": 0.4633,
"step": 164
},
{
"epoch": 1.0623003194888179,
"grad_norm": 0.07564657660605784,
"learning_rate": 8.788209787706015e-05,
"loss": 0.4149,
"step": 166
},
{
"epoch": 1.0750798722044728,
"grad_norm": 0.2601478494309565,
"learning_rate": 8.753361526263621e-05,
"loss": 0.4644,
"step": 168
},
{
"epoch": 1.0878594249201279,
"grad_norm": 0.07236244361689621,
"learning_rate": 8.718090467093654e-05,
"loss": 0.445,
"step": 170
},
{
"epoch": 1.1006389776357828,
"grad_norm": 0.07360308087849284,
"learning_rate": 8.682400583307562e-05,
"loss": 0.4189,
"step": 172
},
{
"epoch": 1.1134185303514377,
"grad_norm": 0.06934965586236702,
"learning_rate": 8.646295895195333e-05,
"loss": 0.4168,
"step": 174
},
{
"epoch": 1.1261980830670926,
"grad_norm": 0.06652725595095291,
"learning_rate": 8.609780469772623e-05,
"loss": 0.4332,
"step": 176
},
{
"epoch": 1.1389776357827477,
"grad_norm": 0.06493423808775205,
"learning_rate": 8.572858420322627e-05,
"loss": 0.4126,
"step": 178
},
{
"epoch": 1.1517571884984026,
"grad_norm": 0.07224306242862681,
"learning_rate": 8.535533905932738e-05,
"loss": 0.4639,
"step": 180
},
{
"epoch": 1.1645367412140575,
"grad_norm": 0.06325420247080109,
"learning_rate": 8.497811131026046e-05,
"loss": 0.4097,
"step": 182
},
{
"epoch": 1.1773162939297124,
"grad_norm": 0.05960690196531746,
"learning_rate": 8.459694344887732e-05,
"loss": 0.4258,
"step": 184
},
{
"epoch": 1.1900958466453675,
"grad_norm": 0.06526403248406679,
"learning_rate": 8.421187841186402e-05,
"loss": 0.4453,
"step": 186
},
{
"epoch": 1.2028753993610224,
"grad_norm": 0.06754636177295095,
"learning_rate": 8.382295957490436e-05,
"loss": 0.4277,
"step": 188
},
{
"epoch": 1.2156549520766773,
"grad_norm": 0.11883404710840821,
"learning_rate": 8.343023074779368e-05,
"loss": 0.4386,
"step": 190
},
{
"epoch": 1.2284345047923322,
"grad_norm": 0.07793571463351197,
"learning_rate": 8.303373616950408e-05,
"loss": 0.4072,
"step": 192
},
{
"epoch": 1.2412140575079873,
"grad_norm": 0.06518657342856102,
"learning_rate": 8.263352050320094e-05,
"loss": 0.4396,
"step": 194
},
{
"epoch": 1.2539936102236422,
"grad_norm": 0.05974282037032855,
"learning_rate": 8.222962883121196e-05,
"loss": 0.4016,
"step": 196
},
{
"epoch": 1.266773162939297,
"grad_norm": 0.0693639502217822,
"learning_rate": 8.182210664994878e-05,
"loss": 0.3808,
"step": 198
},
{
"epoch": 1.279552715654952,
"grad_norm": 0.06127831754623801,
"learning_rate": 8.141099986478212e-05,
"loss": 0.3961,
"step": 200
},
{
"epoch": 1.292332268370607,
"grad_norm": 0.06755312065722066,
"learning_rate": 8.099635478487064e-05,
"loss": 0.3894,
"step": 202
},
{
"epoch": 1.305111821086262,
"grad_norm": 0.0584212869146413,
"learning_rate": 8.057821811794458e-05,
"loss": 0.414,
"step": 204
},
{
"epoch": 1.317891373801917,
"grad_norm": 0.05983512956529008,
"learning_rate": 8.015663696504422e-05,
"loss": 0.3634,
"step": 206
},
{
"epoch": 1.330670926517572,
"grad_norm": 0.05778218969166584,
"learning_rate": 7.973165881521434e-05,
"loss": 0.4233,
"step": 208
},
{
"epoch": 1.343450479233227,
"grad_norm": 0.058310021079803646,
"learning_rate": 7.930333154015466e-05,
"loss": 0.4061,
"step": 210
},
{
"epoch": 1.3562300319488818,
"grad_norm": 0.0642143238679532,
"learning_rate": 7.88717033888274e-05,
"loss": 0.4083,
"step": 212
},
{
"epoch": 1.3690095846645367,
"grad_norm": 0.05656381877721736,
"learning_rate": 7.843682298202235e-05,
"loss": 0.4033,
"step": 214
},
{
"epoch": 1.3817891373801916,
"grad_norm": 0.05518190162844295,
"learning_rate": 7.799873930687978e-05,
"loss": 0.3953,
"step": 216
},
{
"epoch": 1.3945686900958467,
"grad_norm": 0.05903661851778338,
"learning_rate": 7.755750171137246e-05,
"loss": 0.4096,
"step": 218
},
{
"epoch": 1.4073482428115016,
"grad_norm": 0.05833074145436464,
"learning_rate": 7.711315989874677e-05,
"loss": 0.4151,
"step": 220
},
{
"epoch": 1.4201277955271565,
"grad_norm": 0.05919878363690307,
"learning_rate": 7.666576392192389e-05,
"loss": 0.39,
"step": 222
},
{
"epoch": 1.4329073482428116,
"grad_norm": 0.05913664327254173,
"learning_rate": 7.621536417786159e-05,
"loss": 0.4005,
"step": 224
},
{
"epoch": 1.4456869009584665,
"grad_norm": 0.0640842931075253,
"learning_rate": 7.576201140187727e-05,
"loss": 0.4165,
"step": 226
},
{
"epoch": 1.4584664536741214,
"grad_norm": 0.062131879810909965,
"learning_rate": 7.530575666193283e-05,
"loss": 0.3891,
"step": 228
},
{
"epoch": 1.4712460063897763,
"grad_norm": 0.06992276137309804,
"learning_rate": 7.484665135288213e-05,
"loss": 0.3971,
"step": 230
},
{
"epoch": 1.4840255591054312,
"grad_norm": 0.06078790664861669,
"learning_rate": 7.438474719068173e-05,
"loss": 0.3961,
"step": 232
},
{
"epoch": 1.4968051118210863,
"grad_norm": 0.06922648734675908,
"learning_rate": 7.392009620656513e-05,
"loss": 0.4331,
"step": 234
},
{
"epoch": 1.5095846645367412,
"grad_norm": 0.05766139832102871,
"learning_rate": 7.345275074118185e-05,
"loss": 0.4182,
"step": 236
},
{
"epoch": 1.5223642172523961,
"grad_norm": 0.06292873231888371,
"learning_rate": 7.298276343870151e-05,
"loss": 0.4061,
"step": 238
},
{
"epoch": 1.5351437699680512,
"grad_norm": 0.06000860844713537,
"learning_rate": 7.251018724088367e-05,
"loss": 0.4023,
"step": 240
},
{
"epoch": 1.547923322683706,
"grad_norm": 0.0585777107714916,
"learning_rate": 7.203507538111423e-05,
"loss": 0.3855,
"step": 242
},
{
"epoch": 1.560702875399361,
"grad_norm": 0.0571671995255021,
"learning_rate": 7.155748137840892e-05,
"loss": 0.3951,
"step": 244
},
{
"epoch": 1.573482428115016,
"grad_norm": 0.053447175708899994,
"learning_rate": 7.107745903138472e-05,
"loss": 0.3745,
"step": 246
},
{
"epoch": 1.5862619808306708,
"grad_norm": 0.055736902711725635,
"learning_rate": 7.059506241219965e-05,
"loss": 0.3911,
"step": 248
},
{
"epoch": 1.599041533546326,
"grad_norm": 0.05715355824554817,
"learning_rate": 7.011034586046176e-05,
"loss": 0.4043,
"step": 250
},
{
"epoch": 1.6118210862619808,
"grad_norm": 0.06030447320081754,
"learning_rate": 6.962336397710819e-05,
"loss": 0.3899,
"step": 252
},
{
"epoch": 1.6246006389776357,
"grad_norm": 0.061239135474291606,
"learning_rate": 6.91341716182545e-05,
"loss": 0.4246,
"step": 254
},
{
"epoch": 1.6373801916932909,
"grad_norm": 0.05695235071864785,
"learning_rate": 6.864282388901544e-05,
"loss": 0.3953,
"step": 256
},
{
"epoch": 1.6501597444089455,
"grad_norm": 0.05308868251491366,
"learning_rate": 6.814937613729766e-05,
"loss": 0.4103,
"step": 258
},
{
"epoch": 1.6629392971246006,
"grad_norm": 0.054046791633493914,
"learning_rate": 6.765388394756504e-05,
"loss": 0.4059,
"step": 260
},
{
"epoch": 1.6757188498402555,
"grad_norm": 0.05148697040730548,
"learning_rate": 6.715640313457733e-05,
"loss": 0.3767,
"step": 262
},
{
"epoch": 1.6884984025559104,
"grad_norm": 0.05318569591896447,
"learning_rate": 6.665698973710288e-05,
"loss": 0.3708,
"step": 264
},
{
"epoch": 1.7012779552715656,
"grad_norm": 0.05196719070381999,
"learning_rate": 6.615570001160626e-05,
"loss": 0.4042,
"step": 266
},
{
"epoch": 1.7140575079872205,
"grad_norm": 0.05632881769869459,
"learning_rate": 6.565259042591113e-05,
"loss": 0.3987,
"step": 268
},
{
"epoch": 1.7268370607028753,
"grad_norm": 0.05470059818193366,
"learning_rate": 6.514771765283942e-05,
"loss": 0.3973,
"step": 270
},
{
"epoch": 1.7396166134185305,
"grad_norm": 0.056351811449582394,
"learning_rate": 6.464113856382752e-05,
"loss": 0.3864,
"step": 272
},
{
"epoch": 1.7523961661341851,
"grad_norm": 0.05831258279981057,
"learning_rate": 6.413291022251989e-05,
"loss": 0.4041,
"step": 274
},
{
"epoch": 1.7651757188498403,
"grad_norm": 0.053467450310740065,
"learning_rate": 6.362308987834115e-05,
"loss": 0.3814,
"step": 276
},
{
"epoch": 1.7779552715654952,
"grad_norm": 0.051287152623381335,
"learning_rate": 6.311173496004723e-05,
"loss": 0.395,
"step": 278
},
{
"epoch": 1.79073482428115,
"grad_norm": 0.05429714498773308,
"learning_rate": 6.259890306925627e-05,
"loss": 0.3821,
"step": 280
},
{
"epoch": 1.8035143769968052,
"grad_norm": 0.057523653580626326,
"learning_rate": 6.208465197396013e-05,
"loss": 0.3984,
"step": 282
},
{
"epoch": 1.81629392971246,
"grad_norm": 0.05724842136937287,
"learning_rate": 6.156903960201709e-05,
"loss": 0.4181,
"step": 284
},
{
"epoch": 1.829073482428115,
"grad_norm": 0.052227309043480996,
"learning_rate": 6.105212403462651e-05,
"loss": 0.4049,
"step": 286
},
{
"epoch": 1.84185303514377,
"grad_norm": 0.04967908325326877,
"learning_rate": 6.0533963499786314e-05,
"loss": 0.4117,
"step": 288
},
{
"epoch": 1.854632587859425,
"grad_norm": 0.05539898234566285,
"learning_rate": 6.001461636573397e-05,
"loss": 0.4006,
"step": 290
},
{
"epoch": 1.8674121405750799,
"grad_norm": 0.05795414669880149,
"learning_rate": 5.949414113437142e-05,
"loss": 0.386,
"step": 292
},
{
"epoch": 1.880191693290735,
"grad_norm": 0.050446841270231885,
"learning_rate": 5.897259643467527e-05,
"loss": 0.3842,
"step": 294
},
{
"epoch": 1.8929712460063897,
"grad_norm": 0.052453051506198604,
"learning_rate": 5.8450041016092464e-05,
"loss": 0.3525,
"step": 296
},
{
"epoch": 1.9057507987220448,
"grad_norm": 0.052803823491155276,
"learning_rate": 5.792653374192245e-05,
"loss": 0.3963,
"step": 298
},
{
"epoch": 1.9185303514376997,
"grad_norm": 0.05180901601155745,
"learning_rate": 5.7402133582686576e-05,
"loss": 0.3798,
"step": 300
},
{
"epoch": 1.9313099041533546,
"grad_norm": 0.05166645429890597,
"learning_rate": 5.6876899609485256e-05,
"loss": 0.3838,
"step": 302
},
{
"epoch": 1.9440894568690097,
"grad_norm": 0.05306354741968808,
"learning_rate": 5.6350890987343944e-05,
"loss": 0.4165,
"step": 304
},
{
"epoch": 1.9568690095846646,
"grad_norm": 0.0860975722690725,
"learning_rate": 5.582416696854853e-05,
"loss": 0.3737,
"step": 306
},
{
"epoch": 1.9696485623003195,
"grad_norm": 0.05323286133666828,
"learning_rate": 5.5296786885970805e-05,
"loss": 0.3889,
"step": 308
},
{
"epoch": 1.9824281150159746,
"grad_norm": 0.05299665331057226,
"learning_rate": 5.476881014638491e-05,
"loss": 0.3896,
"step": 310
},
{
"epoch": 1.9952076677316293,
"grad_norm": 0.05157945275339266,
"learning_rate": 5.4240296223775465e-05,
"loss": 0.3637,
"step": 312
},
{
"epoch": 2.009584664536741,
"grad_norm": 0.09139947660133817,
"learning_rate": 5.3711304652638126e-05,
"loss": 0.3775,
"step": 314
},
{
"epoch": 2.022364217252396,
"grad_norm": 0.10130414532724454,
"learning_rate": 5.318189502127332e-05,
"loss": 0.2112,
"step": 316
},
{
"epoch": 2.0351437699680512,
"grad_norm": 0.0633333619180165,
"learning_rate": 5.265212696507387e-05,
"loss": 0.2004,
"step": 318
},
{
"epoch": 2.047923322683706,
"grad_norm": 0.0668276114954086,
"learning_rate": 5.212206015980742e-05,
"loss": 0.2019,
"step": 320
},
{
"epoch": 2.060702875399361,
"grad_norm": 0.05942503367303514,
"learning_rate": 5.159175431489424e-05,
"loss": 0.1978,
"step": 322
},
{
"epoch": 2.073482428115016,
"grad_norm": 0.07284145764738766,
"learning_rate": 5.1061269166681183e-05,
"loss": 0.1935,
"step": 324
},
{
"epoch": 2.086261980830671,
"grad_norm": 0.052260140697323494,
"learning_rate": 5.053066447171282e-05,
"loss": 0.1854,
"step": 326
},
{
"epoch": 2.099041533546326,
"grad_norm": 0.05754923159453662,
"learning_rate": 5e-05,
"loss": 0.1965,
"step": 328
},
{
"epoch": 2.1118210862619806,
"grad_norm": 0.05500397186780569,
"learning_rate": 4.94693355282872e-05,
"loss": 0.1827,
"step": 330
},
{
"epoch": 2.1246006389776357,
"grad_norm": 0.061606661346763424,
"learning_rate": 4.893873083331882e-05,
"loss": 0.2008,
"step": 332
},
{
"epoch": 2.137380191693291,
"grad_norm": 0.05678242709297541,
"learning_rate": 4.840824568510579e-05,
"loss": 0.1853,
"step": 334
},
{
"epoch": 2.1501597444089455,
"grad_norm": 0.054080318070508115,
"learning_rate": 4.78779398401926e-05,
"loss": 0.1952,
"step": 336
},
{
"epoch": 2.1629392971246006,
"grad_norm": 0.057204881343756786,
"learning_rate": 4.734787303492615e-05,
"loss": 0.1778,
"step": 338
},
{
"epoch": 2.1757188498402558,
"grad_norm": 0.6941487667655994,
"learning_rate": 4.6818104978726685e-05,
"loss": 0.219,
"step": 340
},
{
"epoch": 2.1884984025559104,
"grad_norm": 0.06999590590614403,
"learning_rate": 4.628869534736187e-05,
"loss": 0.181,
"step": 342
},
{
"epoch": 2.2012779552715656,
"grad_norm": 0.07558854262088241,
"learning_rate": 4.575970377622456e-05,
"loss": 0.2349,
"step": 344
},
{
"epoch": 2.2140575079872207,
"grad_norm": 0.07120027160683609,
"learning_rate": 4.52311898536151e-05,
"loss": 0.1993,
"step": 346
},
{
"epoch": 2.2268370607028753,
"grad_norm": 0.05697032990090494,
"learning_rate": 4.47032131140292e-05,
"loss": 0.1739,
"step": 348
},
{
"epoch": 2.2396166134185305,
"grad_norm": 0.06092977319118132,
"learning_rate": 4.4175833031451473e-05,
"loss": 0.188,
"step": 350
},
{
"epoch": 2.252396166134185,
"grad_norm": 0.05900721095602371,
"learning_rate": 4.364910901265606e-05,
"loss": 0.1778,
"step": 352
},
{
"epoch": 2.2651757188498403,
"grad_norm": 0.08992850669862418,
"learning_rate": 4.3123100390514756e-05,
"loss": 0.1838,
"step": 354
},
{
"epoch": 2.2779552715654954,
"grad_norm": 0.059213794143429914,
"learning_rate": 4.2597866417313436e-05,
"loss": 0.1902,
"step": 356
},
{
"epoch": 2.29073482428115,
"grad_norm": 0.051525349318871976,
"learning_rate": 4.207346625807756e-05,
"loss": 0.1784,
"step": 358
},
{
"epoch": 2.303514376996805,
"grad_norm": 0.055922862481655594,
"learning_rate": 4.1549958983907555e-05,
"loss": 0.1827,
"step": 360
},
{
"epoch": 2.31629392971246,
"grad_norm": 0.054189632126131766,
"learning_rate": 4.102740356532473e-05,
"loss": 0.186,
"step": 362
},
{
"epoch": 2.329073482428115,
"grad_norm": 0.06298745746452741,
"learning_rate": 4.050585886562858e-05,
"loss": 0.1854,
"step": 364
},
{
"epoch": 2.34185303514377,
"grad_norm": 0.06476475169367538,
"learning_rate": 3.998538363426605e-05,
"loss": 0.1794,
"step": 366
},
{
"epoch": 2.3546325878594248,
"grad_norm": 0.05187178001518817,
"learning_rate": 3.94660365002137e-05,
"loss": 0.1817,
"step": 368
},
{
"epoch": 2.36741214057508,
"grad_norm": 0.05110076217610542,
"learning_rate": 3.894787596537352e-05,
"loss": 0.1757,
"step": 370
},
{
"epoch": 2.380191693290735,
"grad_norm": 0.061027606854849537,
"learning_rate": 3.843096039798293e-05,
"loss": 0.1888,
"step": 372
},
{
"epoch": 2.3929712460063897,
"grad_norm": 0.05689282057128392,
"learning_rate": 3.791534802603988e-05,
"loss": 0.1972,
"step": 374
},
{
"epoch": 2.405750798722045,
"grad_norm": 0.05144327012401281,
"learning_rate": 3.740109693074375e-05,
"loss": 0.1975,
"step": 376
},
{
"epoch": 2.4185303514377,
"grad_norm": 0.07243681779987425,
"learning_rate": 3.68882650399528e-05,
"loss": 0.1865,
"step": 378
},
{
"epoch": 2.4313099041533546,
"grad_norm": 0.11601839655528177,
"learning_rate": 3.637691012165886e-05,
"loss": 0.1977,
"step": 380
},
{
"epoch": 2.4440894568690097,
"grad_norm": 0.05323975029748036,
"learning_rate": 3.586708977748012e-05,
"loss": 0.1873,
"step": 382
},
{
"epoch": 2.4568690095846644,
"grad_norm": 0.0499469664737551,
"learning_rate": 3.5358861436172485e-05,
"loss": 0.1832,
"step": 384
},
{
"epoch": 2.4696485623003195,
"grad_norm": 0.05043024991533826,
"learning_rate": 3.485228234716058e-05,
"loss": 0.1821,
"step": 386
},
{
"epoch": 2.4824281150159746,
"grad_norm": 0.054685112352780986,
"learning_rate": 3.434740957408889e-05,
"loss": 0.1816,
"step": 388
},
{
"epoch": 2.4952076677316293,
"grad_norm": 0.057237969167094144,
"learning_rate": 3.3844299988393755e-05,
"loss": 0.1909,
"step": 390
},
{
"epoch": 2.5079872204472844,
"grad_norm": 0.05134273506646416,
"learning_rate": 3.334301026289712e-05,
"loss": 0.1782,
"step": 392
},
{
"epoch": 2.520766773162939,
"grad_norm": 0.049993934417102925,
"learning_rate": 3.284359686542269e-05,
"loss": 0.1928,
"step": 394
},
{
"epoch": 2.533546325878594,
"grad_norm": 0.06457823051474779,
"learning_rate": 3.234611605243496e-05,
"loss": 0.196,
"step": 396
},
{
"epoch": 2.5463258785942493,
"grad_norm": 0.051805062617152425,
"learning_rate": 3.1850623862702344e-05,
"loss": 0.1881,
"step": 398
},
{
"epoch": 2.559105431309904,
"grad_norm": 0.049188541484928724,
"learning_rate": 3.135717611098458e-05,
"loss": 0.1806,
"step": 400
},
{
"epoch": 2.571884984025559,
"grad_norm": 0.05687592017078177,
"learning_rate": 3.086582838174551e-05,
"loss": 0.1784,
"step": 402
},
{
"epoch": 2.584664536741214,
"grad_norm": 0.05098573657706369,
"learning_rate": 3.0376636022891812e-05,
"loss": 0.1932,
"step": 404
},
{
"epoch": 2.597444089456869,
"grad_norm": 0.052376381772842893,
"learning_rate": 2.9889654139538246e-05,
"loss": 0.1889,
"step": 406
},
{
"epoch": 2.610223642172524,
"grad_norm": 0.05031660077056393,
"learning_rate": 2.9404937587800375e-05,
"loss": 0.1769,
"step": 408
},
{
"epoch": 2.623003194888179,
"grad_norm": 0.04930354808056054,
"learning_rate": 2.8922540968615286e-05,
"loss": 0.1685,
"step": 410
},
{
"epoch": 2.635782747603834,
"grad_norm": 0.06709139465230578,
"learning_rate": 2.8442518621591086e-05,
"loss": 0.1785,
"step": 412
},
{
"epoch": 2.648562300319489,
"grad_norm": 0.0503489735828908,
"learning_rate": 2.7964924618885778e-05,
"loss": 0.1689,
"step": 414
},
{
"epoch": 2.661341853035144,
"grad_norm": 0.05047783892143097,
"learning_rate": 2.748981275911633e-05,
"loss": 0.1808,
"step": 416
},
{
"epoch": 2.6741214057507987,
"grad_norm": 0.04955419672921838,
"learning_rate": 2.701723656129851e-05,
"loss": 0.1727,
"step": 418
},
{
"epoch": 2.686900958466454,
"grad_norm": 0.04769759775271665,
"learning_rate": 2.6547249258818164e-05,
"loss": 0.1708,
"step": 420
},
{
"epoch": 2.6996805111821085,
"grad_norm": 0.050324696099913684,
"learning_rate": 2.607990379343489e-05,
"loss": 0.1817,
"step": 422
},
{
"epoch": 2.7124600638977636,
"grad_norm": 0.05249496210993974,
"learning_rate": 2.5615252809318284e-05,
"loss": 0.1836,
"step": 424
},
{
"epoch": 2.7252396166134183,
"grad_norm": 0.0472378188955872,
"learning_rate": 2.5153348647117857e-05,
"loss": 0.1736,
"step": 426
},
{
"epoch": 2.7380191693290734,
"grad_norm": 0.049243154928981264,
"learning_rate": 2.469424333806718e-05,
"loss": 0.1675,
"step": 428
},
{
"epoch": 2.7507987220447285,
"grad_norm": 0.05096273109137321,
"learning_rate": 2.4237988598122752e-05,
"loss": 0.1658,
"step": 430
},
{
"epoch": 2.763578274760383,
"grad_norm": 0.0514806212844811,
"learning_rate": 2.3784635822138424e-05,
"loss": 0.1922,
"step": 432
},
{
"epoch": 2.7763578274760383,
"grad_norm": 0.05006269553229606,
"learning_rate": 2.333423607807613e-05,
"loss": 0.1887,
"step": 434
},
{
"epoch": 2.7891373801916934,
"grad_norm": 0.04935551516167026,
"learning_rate": 2.288684010125325e-05,
"loss": 0.1763,
"step": 436
},
{
"epoch": 2.801916932907348,
"grad_norm": 0.05353903496894845,
"learning_rate": 2.2442498288627556e-05,
"loss": 0.1944,
"step": 438
},
{
"epoch": 2.8146964856230032,
"grad_norm": 0.04697149845887787,
"learning_rate": 2.2001260693120233e-05,
"loss": 0.1672,
"step": 440
},
{
"epoch": 2.8274760383386583,
"grad_norm": 0.054384654770629585,
"learning_rate": 2.156317701797766e-05,
"loss": 0.1807,
"step": 442
},
{
"epoch": 2.840255591054313,
"grad_norm": 0.04684823569442938,
"learning_rate": 2.1128296611172593e-05,
"loss": 0.171,
"step": 444
},
{
"epoch": 2.853035143769968,
"grad_norm": 0.0498371244165766,
"learning_rate": 2.0696668459845355e-05,
"loss": 0.1827,
"step": 446
},
{
"epoch": 2.8658146964856233,
"grad_norm": 0.04969475724913098,
"learning_rate": 2.026834118478567e-05,
"loss": 0.1749,
"step": 448
},
{
"epoch": 2.878594249201278,
"grad_norm": 0.051902756416916496,
"learning_rate": 1.98433630349558e-05,
"loss": 0.1891,
"step": 450
},
{
"epoch": 2.891373801916933,
"grad_norm": 0.05102564026340021,
"learning_rate": 1.9421781882055444e-05,
"loss": 0.1849,
"step": 452
},
{
"epoch": 2.9041533546325877,
"grad_norm": 0.05200929870376942,
"learning_rate": 1.9003645215129355e-05,
"loss": 0.1891,
"step": 454
},
{
"epoch": 2.916932907348243,
"grad_norm": 0.05083154953396676,
"learning_rate": 1.858900013521788e-05,
"loss": 0.179,
"step": 456
},
{
"epoch": 2.9297124600638975,
"grad_norm": 0.049127219472404525,
"learning_rate": 1.817789335005121e-05,
"loss": 0.17,
"step": 458
},
{
"epoch": 2.9424920127795526,
"grad_norm": 0.049677004679461886,
"learning_rate": 1.777037116878804e-05,
"loss": 0.1831,
"step": 460
},
{
"epoch": 2.9552715654952078,
"grad_norm": 0.054496479788018075,
"learning_rate": 1.7366479496799077e-05,
"loss": 0.1843,
"step": 462
},
{
"epoch": 2.9680511182108624,
"grad_norm": 0.04820092295738451,
"learning_rate": 1.6966263830495936e-05,
"loss": 0.1685,
"step": 464
},
{
"epoch": 2.9808306709265175,
"grad_norm": 0.04915420841884947,
"learning_rate": 1.656976925220633e-05,
"loss": 0.1875,
"step": 466
},
{
"epoch": 2.9936102236421727,
"grad_norm": 0.07661474807504913,
"learning_rate": 1.6177040425095662e-05,
"loss": 0.1891,
"step": 468
},
{
"epoch": 3.0079872204472844,
"grad_norm": 0.07655695803476695,
"learning_rate": 1.5788121588135975e-05,
"loss": 0.1837,
"step": 470
},
{
"epoch": 3.0207667731629395,
"grad_norm": 0.060916330302725,
"learning_rate": 1.5403056551122697e-05,
"loss": 0.0872,
"step": 472
},
{
"epoch": 3.033546325878594,
"grad_norm": 0.052542395235648506,
"learning_rate": 1.5021888689739549e-05,
"loss": 0.0778,
"step": 474
},
{
"epoch": 3.0463258785942493,
"grad_norm": 0.20368087770560855,
"learning_rate": 1.4644660940672627e-05,
"loss": 0.102,
"step": 476
},
{
"epoch": 3.059105431309904,
"grad_norm": 0.10396707161226072,
"learning_rate": 1.427141579677374e-05,
"loss": 0.083,
"step": 478
},
{
"epoch": 3.071884984025559,
"grad_norm": 0.04599720220665865,
"learning_rate": 1.3902195302273779e-05,
"loss": 0.0757,
"step": 480
},
{
"epoch": 3.084664536741214,
"grad_norm": 0.056109340867354925,
"learning_rate": 1.3537041048046695e-05,
"loss": 0.081,
"step": 482
},
{
"epoch": 3.097444089456869,
"grad_norm": 0.048015102375770044,
"learning_rate": 1.3175994166924394e-05,
"loss": 0.0802,
"step": 484
},
{
"epoch": 3.110223642172524,
"grad_norm": 0.04645228076024571,
"learning_rate": 1.2819095329063469e-05,
"loss": 0.0787,
"step": 486
},
{
"epoch": 3.123003194888179,
"grad_norm": 0.04637085498651796,
"learning_rate": 1.246638473736378e-05,
"loss": 0.0839,
"step": 488
},
{
"epoch": 3.135782747603834,
"grad_norm": 0.05039074009256794,
"learning_rate": 1.2117902122939861e-05,
"loss": 0.0812,
"step": 490
},
{
"epoch": 3.148562300319489,
"grad_norm": 0.05079569512274489,
"learning_rate": 1.1773686740645384e-05,
"loss": 0.0797,
"step": 492
},
{
"epoch": 3.1613418530351436,
"grad_norm": 0.04286375870307716,
"learning_rate": 1.1433777364651271e-05,
"loss": 0.0737,
"step": 494
},
{
"epoch": 3.1741214057507987,
"grad_norm": 0.03982951021947898,
"learning_rate": 1.1098212284078036e-05,
"loss": 0.0722,
"step": 496
},
{
"epoch": 3.186900958466454,
"grad_norm": 0.0446624849328897,
"learning_rate": 1.076702929868264e-05,
"loss": 0.079,
"step": 498
},
{
"epoch": 3.1996805111821085,
"grad_norm": 0.04376807908723891,
"learning_rate": 1.0440265714600572e-05,
"loss": 0.0837,
"step": 500
},
{
"epoch": 3.2124600638977636,
"grad_norm": 0.04087367539850916,
"learning_rate": 1.0117958340143507e-05,
"loss": 0.076,
"step": 502
},
{
"epoch": 3.2252396166134187,
"grad_norm": 0.04066584417219993,
"learning_rate": 9.800143481652979e-06,
"loss": 0.0701,
"step": 504
},
{
"epoch": 3.2380191693290734,
"grad_norm": 0.08215263649470263,
"learning_rate": 9.48685693941067e-06,
"loss": 0.0776,
"step": 506
},
{
"epoch": 3.2507987220447285,
"grad_norm": 0.0437601284673361,
"learning_rate": 9.17813400360572e-06,
"loss": 0.0764,
"step": 508
},
{
"epoch": 3.263578274760383,
"grad_norm": 0.04382435518426366,
"learning_rate": 8.874009450359427e-06,
"loss": 0.0826,
"step": 510
},
{
"epoch": 3.2763578274760383,
"grad_norm": 0.04095610913441161,
"learning_rate": 8.574517537807897e-06,
"loss": 0.0753,
"step": 512
},
{
"epoch": 3.2891373801916934,
"grad_norm": 0.040525949300480126,
"learning_rate": 8.279692002243027e-06,
"loss": 0.0764,
"step": 514
},
{
"epoch": 3.301916932907348,
"grad_norm": 0.043675586209021296,
"learning_rate": 7.989566054312287e-06,
"loss": 0.0817,
"step": 516
},
{
"epoch": 3.3146964856230032,
"grad_norm": 0.04319420448553361,
"learning_rate": 7.704172375277691e-06,
"loss": 0.0759,
"step": 518
},
{
"epoch": 3.3274760383386583,
"grad_norm": 0.044446852802239034,
"learning_rate": 7.423543113334436e-06,
"loss": 0.0813,
"step": 520
},
{
"epoch": 3.340255591054313,
"grad_norm": 0.09121973616663154,
"learning_rate": 7.14770987998954e-06,
"loss": 0.0838,
"step": 522
},
{
"epoch": 3.353035143769968,
"grad_norm": 0.05879997473879583,
"learning_rate": 6.876703746500984e-06,
"loss": 0.0738,
"step": 524
},
{
"epoch": 3.365814696485623,
"grad_norm": 0.04667273388126841,
"learning_rate": 6.610555240377652e-06,
"loss": 0.0787,
"step": 526
},
{
"epoch": 3.378594249201278,
"grad_norm": 0.042105033545020404,
"learning_rate": 6.349294341940593e-06,
"loss": 0.0801,
"step": 528
},
{
"epoch": 3.391373801916933,
"grad_norm": 0.0407975465413022,
"learning_rate": 6.092950480945897e-06,
"loss": 0.0735,
"step": 530
},
{
"epoch": 3.4041533546325877,
"grad_norm": 0.04234912863253251,
"learning_rate": 5.841552533269534e-06,
"loss": 0.0772,
"step": 532
},
{
"epoch": 3.416932907348243,
"grad_norm": 0.04032120711392374,
"learning_rate": 5.595128817654638e-06,
"loss": 0.0749,
"step": 534
},
{
"epoch": 3.4297124600638975,
"grad_norm": 0.041050930036482094,
"learning_rate": 5.353707092521582e-06,
"loss": 0.0769,
"step": 536
},
{
"epoch": 3.4424920127795526,
"grad_norm": 0.043382176933190755,
"learning_rate": 5.117314552841052e-06,
"loss": 0.0767,
"step": 538
},
{
"epoch": 3.4552715654952078,
"grad_norm": 0.039240502138117625,
"learning_rate": 4.885977827070748e-06,
"loss": 0.0721,
"step": 540
},
{
"epoch": 3.4680511182108624,
"grad_norm": 0.040812347040587296,
"learning_rate": 4.659722974155767e-06,
"loss": 0.1114,
"step": 542
},
{
"epoch": 3.4808306709265175,
"grad_norm": 0.0423787622918925,
"learning_rate": 4.43857548059321e-06,
"loss": 0.0778,
"step": 544
},
{
"epoch": 3.4936102236421727,
"grad_norm": 0.042228923687598445,
"learning_rate": 4.2225602575612755e-06,
"loss": 0.0814,
"step": 546
},
{
"epoch": 3.5063897763578273,
"grad_norm": 0.0407267289339222,
"learning_rate": 4.011701638113063e-06,
"loss": 0.0782,
"step": 548
},
{
"epoch": 3.5191693290734825,
"grad_norm": 0.0389855165359938,
"learning_rate": 3.8060233744356633e-06,
"loss": 0.0789,
"step": 550
},
{
"epoch": 3.5319488817891376,
"grad_norm": 0.040904703617676376,
"learning_rate": 3.605548635174533e-06,
"loss": 0.078,
"step": 552
},
{
"epoch": 3.5447284345047922,
"grad_norm": 0.04093280012624571,
"learning_rate": 3.410300002823691e-06,
"loss": 0.0777,
"step": 554
},
{
"epoch": 3.5575079872204474,
"grad_norm": 0.042904856841507744,
"learning_rate": 3.220299471181898e-06,
"loss": 0.0757,
"step": 556
},
{
"epoch": 3.5702875399361025,
"grad_norm": 0.0436449067886704,
"learning_rate": 3.035568442875136e-06,
"loss": 0.0798,
"step": 558
},
{
"epoch": 3.583067092651757,
"grad_norm": 0.035664776931118955,
"learning_rate": 2.85612772694579e-06,
"loss": 0.0632,
"step": 560
},
{
"epoch": 3.5958466453674123,
"grad_norm": 0.03847526723825484,
"learning_rate": 2.6819975365085237e-06,
"loss": 0.0744,
"step": 562
},
{
"epoch": 3.608626198083067,
"grad_norm": 0.039939236612970476,
"learning_rate": 2.5131974864734066e-06,
"loss": 0.0794,
"step": 564
},
{
"epoch": 3.621405750798722,
"grad_norm": 0.040388305870748,
"learning_rate": 2.349746591336405e-06,
"loss": 0.0718,
"step": 566
},
{
"epoch": 3.6341853035143767,
"grad_norm": 0.04232813426430434,
"learning_rate": 2.191663263037458e-06,
"loss": 0.0769,
"step": 568
},
{
"epoch": 3.646964856230032,
"grad_norm": 0.04213845492527589,
"learning_rate": 2.0389653088865036e-06,
"loss": 0.0728,
"step": 570
},
{
"epoch": 3.659744408945687,
"grad_norm": 0.04098999517730541,
"learning_rate": 1.8916699295575324e-06,
"loss": 0.0724,
"step": 572
},
{
"epoch": 3.6725239616613417,
"grad_norm": 0.037533240934183365,
"learning_rate": 1.7497937171510547e-06,
"loss": 0.0709,
"step": 574
},
{
"epoch": 3.6853035143769968,
"grad_norm": 0.039040304607963414,
"learning_rate": 1.6133526533250565e-06,
"loss": 0.0756,
"step": 576
},
{
"epoch": 3.698083067092652,
"grad_norm": 0.04065729024121047,
"learning_rate": 1.4823621074947503e-06,
"loss": 0.0774,
"step": 578
},
{
"epoch": 3.7108626198083066,
"grad_norm": 0.04252602887603373,
"learning_rate": 1.3568368351012717e-06,
"loss": 0.0824,
"step": 580
},
{
"epoch": 3.7236421725239617,
"grad_norm": 0.04343672134882273,
"learning_rate": 1.236790975949592e-06,
"loss": 0.074,
"step": 582
},
{
"epoch": 3.736421725239617,
"grad_norm": 0.0403766223584342,
"learning_rate": 1.1222380526156928e-06,
"loss": 0.0755,
"step": 584
},
{
"epoch": 3.7492012779552715,
"grad_norm": 0.04234625541762105,
"learning_rate": 1.0131909689233442e-06,
"loss": 0.0814,
"step": 586
},
{
"epoch": 3.7619808306709266,
"grad_norm": 0.03861507912847567,
"learning_rate": 9.096620084905472e-07,
"loss": 0.0664,
"step": 588
},
{
"epoch": 3.7747603833865817,
"grad_norm": 0.041892733843973705,
"learning_rate": 8.11662833345822e-07,
"loss": 0.0832,
"step": 590
},
{
"epoch": 3.7875399361022364,
"grad_norm": 0.05413889236863839,
"learning_rate": 7.192044826145771e-07,
"loss": 0.0921,
"step": 592
},
{
"epoch": 3.8003194888178915,
"grad_norm": 0.04010280150213322,
"learning_rate": 6.322973712755697e-07,
"loss": 0.0752,
"step": 594
},
{
"epoch": 3.813099041533546,
"grad_norm": 0.04321423418162425,
"learning_rate": 5.509512889877333e-07,
"loss": 0.0781,
"step": 596
},
{
"epoch": 3.8258785942492013,
"grad_norm": 0.04049679761598481,
"learning_rate": 4.7517539898741524e-07,
"loss": 0.0694,
"step": 598
},
{
"epoch": 3.838658146964856,
"grad_norm": 0.04258434666712487,
"learning_rate": 4.049782370561583e-07,
"loss": 0.0756,
"step": 600
},
{
"epoch": 3.851437699680511,
"grad_norm": 0.03927978960342531,
"learning_rate": 3.4036771055923066e-07,
"loss": 0.075,
"step": 602
},
{
"epoch": 3.864217252396166,
"grad_norm": 0.04093422273122725,
"learning_rate": 2.813510975548772e-07,
"loss": 0.0793,
"step": 604
},
{
"epoch": 3.876996805111821,
"grad_norm": 0.0433141394014271,
"learning_rate": 2.2793504597447002e-07,
"loss": 0.0796,
"step": 606
},
{
"epoch": 3.889776357827476,
"grad_norm": 0.04198937065288365,
"learning_rate": 1.8012557287367392e-07,
"loss": 0.0753,
"step": 608
},
{
"epoch": 3.902555910543131,
"grad_norm": 0.043002763720086865,
"learning_rate": 1.379280637546443e-07,
"loss": 0.0917,
"step": 610
},
{
"epoch": 3.915335463258786,
"grad_norm": 0.042376326172823574,
"learning_rate": 1.0134727195937333e-07,
"loss": 0.0747,
"step": 612
},
{
"epoch": 3.928115015974441,
"grad_norm": 0.040600294489722695,
"learning_rate": 7.038731813426291e-08,
"loss": 0.0714,
"step": 614
},
{
"epoch": 3.940894568690096,
"grad_norm": 0.0395711578920217,
"learning_rate": 4.5051689765929214e-08,
"loss": 0.0801,
"step": 616
},
{
"epoch": 3.9536741214057507,
"grad_norm": 0.03943811103107873,
"learning_rate": 2.534324078837802e-08,
"loss": 0.074,
"step": 618
},
{
"epoch": 3.966453674121406,
"grad_norm": 0.041905340316324986,
"learning_rate": 1.1264191261528557e-08,
"loss": 0.0771,
"step": 620
},
{
"epoch": 3.979233226837061,
"grad_norm": 0.06686218682304546,
"learning_rate": 2.8161271211024633e-09,
"loss": 0.0811,
"step": 622
},
{
"epoch": 3.9920127795527156,
"grad_norm": 0.03876030687059135,
"learning_rate": 0.0,
"loss": 0.076,
"step": 624
}
],
"logging_steps": 2,
"max_steps": 624,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 2.5262683933881926e+19,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}