murugeshmarvel's picture
Upload folder using huggingface_hub
4eda3ce verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 4.995079950799508,
"eval_steps": 102,
"global_step": 1015,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.004920049200492005,
"grad_norm": 2.296875,
"learning_rate": 1.0000000000000002e-06,
"loss": 0.2408,
"step": 1
},
{
"epoch": 0.00984009840098401,
"grad_norm": 2.8125,
"learning_rate": 2.0000000000000003e-06,
"loss": 0.2626,
"step": 2
},
{
"epoch": 0.014760147601476014,
"grad_norm": 2.484375,
"learning_rate": 3e-06,
"loss": 0.2438,
"step": 3
},
{
"epoch": 0.01968019680196802,
"grad_norm": 2.375,
"learning_rate": 4.000000000000001e-06,
"loss": 0.2335,
"step": 4
},
{
"epoch": 0.024600246002460024,
"grad_norm": 2.453125,
"learning_rate": 5e-06,
"loss": 0.229,
"step": 5
},
{
"epoch": 0.02952029520295203,
"grad_norm": 1.78125,
"learning_rate": 6e-06,
"loss": 0.2125,
"step": 6
},
{
"epoch": 0.03444034440344403,
"grad_norm": 1.4765625,
"learning_rate": 7e-06,
"loss": 0.1421,
"step": 7
},
{
"epoch": 0.03936039360393604,
"grad_norm": 1.53125,
"learning_rate": 8.000000000000001e-06,
"loss": 0.1727,
"step": 8
},
{
"epoch": 0.04428044280442804,
"grad_norm": 1.4296875,
"learning_rate": 9e-06,
"loss": 0.1599,
"step": 9
},
{
"epoch": 0.04920049200492005,
"grad_norm": 1.7265625,
"learning_rate": 1e-05,
"loss": 0.2048,
"step": 10
},
{
"epoch": 0.05412054120541206,
"grad_norm": 2.125,
"learning_rate": 9.99997557091071e-06,
"loss": 0.1458,
"step": 11
},
{
"epoch": 0.05904059040590406,
"grad_norm": 1.6015625,
"learning_rate": 9.99990228388155e-06,
"loss": 0.1435,
"step": 12
},
{
"epoch": 0.06396063960639606,
"grad_norm": 1.2578125,
"learning_rate": 9.999780139628657e-06,
"loss": 0.1791,
"step": 13
},
{
"epoch": 0.06888068880688807,
"grad_norm": 1.375,
"learning_rate": 9.999609139345578e-06,
"loss": 0.1412,
"step": 14
},
{
"epoch": 0.07380073800738007,
"grad_norm": 1.0859375,
"learning_rate": 9.999389284703265e-06,
"loss": 0.1416,
"step": 15
},
{
"epoch": 0.07872078720787208,
"grad_norm": 1.21875,
"learning_rate": 9.99912057785006e-06,
"loss": 0.1255,
"step": 16
},
{
"epoch": 0.08364083640836409,
"grad_norm": 0.9140625,
"learning_rate": 9.998803021411668e-06,
"loss": 0.1201,
"step": 17
},
{
"epoch": 0.08856088560885608,
"grad_norm": 1.6875,
"learning_rate": 9.99843661849113e-06,
"loss": 0.2039,
"step": 18
},
{
"epoch": 0.09348093480934809,
"grad_norm": 0.96875,
"learning_rate": 9.998021372668809e-06,
"loss": 0.1352,
"step": 19
},
{
"epoch": 0.0984009840098401,
"grad_norm": 0.9296875,
"learning_rate": 9.99755728800233e-06,
"loss": 0.1448,
"step": 20
},
{
"epoch": 0.1033210332103321,
"grad_norm": 1.109375,
"learning_rate": 9.997044369026563e-06,
"loss": 0.1537,
"step": 21
},
{
"epoch": 0.10824108241082411,
"grad_norm": 1.0234375,
"learning_rate": 9.996482620753565e-06,
"loss": 0.1506,
"step": 22
},
{
"epoch": 0.11316113161131611,
"grad_norm": 1.109375,
"learning_rate": 9.995872048672535e-06,
"loss": 0.1474,
"step": 23
},
{
"epoch": 0.11808118081180811,
"grad_norm": 0.8671875,
"learning_rate": 9.99521265874976e-06,
"loss": 0.1215,
"step": 24
},
{
"epoch": 0.12300123001230012,
"grad_norm": 1.0859375,
"learning_rate": 9.994504457428557e-06,
"loss": 0.126,
"step": 25
},
{
"epoch": 0.12792127921279212,
"grad_norm": 1.015625,
"learning_rate": 9.993747451629217e-06,
"loss": 0.1417,
"step": 26
},
{
"epoch": 0.13284132841328414,
"grad_norm": 1.15625,
"learning_rate": 9.992941648748917e-06,
"loss": 0.1163,
"step": 27
},
{
"epoch": 0.13776137761377613,
"grad_norm": 1.046875,
"learning_rate": 9.992087056661676e-06,
"loss": 0.1332,
"step": 28
},
{
"epoch": 0.14268142681426815,
"grad_norm": 0.8671875,
"learning_rate": 9.991183683718253e-06,
"loss": 0.1269,
"step": 29
},
{
"epoch": 0.14760147601476015,
"grad_norm": 1.0,
"learning_rate": 9.99023153874608e-06,
"loss": 0.1071,
"step": 30
},
{
"epoch": 0.15252152521525214,
"grad_norm": 1.078125,
"learning_rate": 9.989230631049171e-06,
"loss": 0.1415,
"step": 31
},
{
"epoch": 0.15744157441574416,
"grad_norm": 1.1015625,
"learning_rate": 9.988180970408032e-06,
"loss": 0.1439,
"step": 32
},
{
"epoch": 0.16236162361623616,
"grad_norm": 1.046875,
"learning_rate": 9.987082567079563e-06,
"loss": 0.1333,
"step": 33
},
{
"epoch": 0.16728167281672818,
"grad_norm": 0.9296875,
"learning_rate": 9.985935431796962e-06,
"loss": 0.134,
"step": 34
},
{
"epoch": 0.17220172201722017,
"grad_norm": 1.1015625,
"learning_rate": 9.984739575769619e-06,
"loss": 0.1315,
"step": 35
},
{
"epoch": 0.17712177121771217,
"grad_norm": 1.0625,
"learning_rate": 9.983495010683e-06,
"loss": 0.1258,
"step": 36
},
{
"epoch": 0.1820418204182042,
"grad_norm": 0.87890625,
"learning_rate": 9.982201748698543e-06,
"loss": 0.1141,
"step": 37
},
{
"epoch": 0.18696186961869618,
"grad_norm": 1.0703125,
"learning_rate": 9.980859802453533e-06,
"loss": 0.1437,
"step": 38
},
{
"epoch": 0.1918819188191882,
"grad_norm": 0.95703125,
"learning_rate": 9.97946918506098e-06,
"loss": 0.1326,
"step": 39
},
{
"epoch": 0.1968019680196802,
"grad_norm": 0.87109375,
"learning_rate": 9.978029910109491e-06,
"loss": 0.0979,
"step": 40
},
{
"epoch": 0.2017220172201722,
"grad_norm": 0.8203125,
"learning_rate": 9.976541991663136e-06,
"loss": 0.1086,
"step": 41
},
{
"epoch": 0.2066420664206642,
"grad_norm": 1.2265625,
"learning_rate": 9.975005444261312e-06,
"loss": 0.1504,
"step": 42
},
{
"epoch": 0.2115621156211562,
"grad_norm": 0.88671875,
"learning_rate": 9.973420282918603e-06,
"loss": 0.1282,
"step": 43
},
{
"epoch": 0.21648216482164823,
"grad_norm": 0.75,
"learning_rate": 9.971786523124623e-06,
"loss": 0.0807,
"step": 44
},
{
"epoch": 0.22140221402214022,
"grad_norm": 0.77734375,
"learning_rate": 9.970104180843881e-06,
"loss": 0.1084,
"step": 45
},
{
"epoch": 0.22632226322263221,
"grad_norm": 1.046875,
"learning_rate": 9.968373272515612e-06,
"loss": 0.1385,
"step": 46
},
{
"epoch": 0.23124231242312424,
"grad_norm": 1.0078125,
"learning_rate": 9.966593815053623e-06,
"loss": 0.1538,
"step": 47
},
{
"epoch": 0.23616236162361623,
"grad_norm": 0.82421875,
"learning_rate": 9.964765825846123e-06,
"loss": 0.1035,
"step": 48
},
{
"epoch": 0.24108241082410825,
"grad_norm": 0.96484375,
"learning_rate": 9.962889322755556e-06,
"loss": 0.1055,
"step": 49
},
{
"epoch": 0.24600246002460024,
"grad_norm": 0.7734375,
"learning_rate": 9.960964324118428e-06,
"loss": 0.108,
"step": 50
},
{
"epoch": 0.25092250922509224,
"grad_norm": 0.94921875,
"learning_rate": 9.958990848745123e-06,
"loss": 0.1407,
"step": 51
},
{
"epoch": 0.25584255842558423,
"grad_norm": 0.890625,
"learning_rate": 9.956968915919725e-06,
"loss": 0.1281,
"step": 52
},
{
"epoch": 0.2607626076260763,
"grad_norm": 1.2109375,
"learning_rate": 9.954898545399822e-06,
"loss": 0.1094,
"step": 53
},
{
"epoch": 0.2656826568265683,
"grad_norm": 0.8984375,
"learning_rate": 9.952779757416326e-06,
"loss": 0.1341,
"step": 54
},
{
"epoch": 0.27060270602706027,
"grad_norm": 1.1484375,
"learning_rate": 9.950612572673255e-06,
"loss": 0.1497,
"step": 55
},
{
"epoch": 0.27552275522755226,
"grad_norm": 1.0859375,
"learning_rate": 9.948397012347553e-06,
"loss": 0.146,
"step": 56
},
{
"epoch": 0.28044280442804426,
"grad_norm": 0.94140625,
"learning_rate": 9.946133098088868e-06,
"loss": 0.1294,
"step": 57
},
{
"epoch": 0.2853628536285363,
"grad_norm": 1.046875,
"learning_rate": 9.943820852019345e-06,
"loss": 0.1228,
"step": 58
},
{
"epoch": 0.2902829028290283,
"grad_norm": 0.890625,
"learning_rate": 9.941460296733409e-06,
"loss": 0.0948,
"step": 59
},
{
"epoch": 0.2952029520295203,
"grad_norm": 1.09375,
"learning_rate": 9.939051455297548e-06,
"loss": 0.1337,
"step": 60
},
{
"epoch": 0.3001230012300123,
"grad_norm": 0.90234375,
"learning_rate": 9.936594351250082e-06,
"loss": 0.1267,
"step": 61
},
{
"epoch": 0.3050430504305043,
"grad_norm": 0.6640625,
"learning_rate": 9.934089008600937e-06,
"loss": 0.0852,
"step": 62
},
{
"epoch": 0.30996309963099633,
"grad_norm": 0.89453125,
"learning_rate": 9.931535451831408e-06,
"loss": 0.0853,
"step": 63
},
{
"epoch": 0.3148831488314883,
"grad_norm": 1.2578125,
"learning_rate": 9.928933705893924e-06,
"loss": 0.104,
"step": 64
},
{
"epoch": 0.3198031980319803,
"grad_norm": 0.96484375,
"learning_rate": 9.926283796211796e-06,
"loss": 0.0999,
"step": 65
},
{
"epoch": 0.3247232472324723,
"grad_norm": 0.83203125,
"learning_rate": 9.923585748678976e-06,
"loss": 0.1169,
"step": 66
},
{
"epoch": 0.3296432964329643,
"grad_norm": 0.984375,
"learning_rate": 9.920839589659803e-06,
"loss": 0.1413,
"step": 67
},
{
"epoch": 0.33456334563345635,
"grad_norm": 0.90625,
"learning_rate": 9.918045345988743e-06,
"loss": 0.0995,
"step": 68
},
{
"epoch": 0.33948339483394835,
"grad_norm": 0.95703125,
"learning_rate": 9.915203044970125e-06,
"loss": 0.1242,
"step": 69
},
{
"epoch": 0.34440344403444034,
"grad_norm": 1.078125,
"learning_rate": 9.91231271437788e-06,
"loss": 0.0816,
"step": 70
},
{
"epoch": 0.34932349323493234,
"grad_norm": 0.96875,
"learning_rate": 9.909374382455268e-06,
"loss": 0.1365,
"step": 71
},
{
"epoch": 0.35424354243542433,
"grad_norm": 0.91796875,
"learning_rate": 9.906388077914598e-06,
"loss": 0.1388,
"step": 72
},
{
"epoch": 0.3591635916359164,
"grad_norm": 1.0546875,
"learning_rate": 9.903353829936943e-06,
"loss": 0.1291,
"step": 73
},
{
"epoch": 0.3640836408364084,
"grad_norm": 0.8828125,
"learning_rate": 9.900271668171879e-06,
"loss": 0.1014,
"step": 74
},
{
"epoch": 0.36900369003690037,
"grad_norm": 1.0625,
"learning_rate": 9.89714162273716e-06,
"loss": 0.1653,
"step": 75
},
{
"epoch": 0.37392373923739236,
"grad_norm": 0.76953125,
"learning_rate": 9.893963724218456e-06,
"loss": 0.1097,
"step": 76
},
{
"epoch": 0.37884378843788435,
"grad_norm": 1.09375,
"learning_rate": 9.890738003669029e-06,
"loss": 0.1509,
"step": 77
},
{
"epoch": 0.3837638376383764,
"grad_norm": 1.15625,
"learning_rate": 9.887464492609447e-06,
"loss": 0.119,
"step": 78
},
{
"epoch": 0.3886838868388684,
"grad_norm": 0.90234375,
"learning_rate": 9.884143223027267e-06,
"loss": 0.1354,
"step": 79
},
{
"epoch": 0.3936039360393604,
"grad_norm": 0.84375,
"learning_rate": 9.880774227376727e-06,
"loss": 0.0938,
"step": 80
},
{
"epoch": 0.3985239852398524,
"grad_norm": 0.890625,
"learning_rate": 9.877357538578426e-06,
"loss": 0.1072,
"step": 81
},
{
"epoch": 0.4034440344403444,
"grad_norm": 0.8125,
"learning_rate": 9.873893190018995e-06,
"loss": 0.099,
"step": 82
},
{
"epoch": 0.40836408364083643,
"grad_norm": 1.0703125,
"learning_rate": 9.870381215550796e-06,
"loss": 0.1036,
"step": 83
},
{
"epoch": 0.4132841328413284,
"grad_norm": 0.7734375,
"learning_rate": 9.866821649491558e-06,
"loss": 0.0779,
"step": 84
},
{
"epoch": 0.4182041820418204,
"grad_norm": 1.140625,
"learning_rate": 9.863214526624065e-06,
"loss": 0.1597,
"step": 85
},
{
"epoch": 0.4231242312423124,
"grad_norm": 0.96875,
"learning_rate": 9.85955988219581e-06,
"loss": 0.1283,
"step": 86
},
{
"epoch": 0.4280442804428044,
"grad_norm": 1.1171875,
"learning_rate": 9.855857751918644e-06,
"loss": 0.1327,
"step": 87
},
{
"epoch": 0.43296432964329645,
"grad_norm": 0.80859375,
"learning_rate": 9.852108171968438e-06,
"loss": 0.1,
"step": 88
},
{
"epoch": 0.43788437884378845,
"grad_norm": 1.0625,
"learning_rate": 9.848311178984718e-06,
"loss": 0.1653,
"step": 89
},
{
"epoch": 0.44280442804428044,
"grad_norm": 0.69921875,
"learning_rate": 9.844466810070319e-06,
"loss": 0.0841,
"step": 90
},
{
"epoch": 0.44772447724477243,
"grad_norm": 0.8828125,
"learning_rate": 9.840575102791013e-06,
"loss": 0.1193,
"step": 91
},
{
"epoch": 0.45264452644526443,
"grad_norm": 0.9921875,
"learning_rate": 9.836636095175148e-06,
"loss": 0.1187,
"step": 92
},
{
"epoch": 0.4575645756457565,
"grad_norm": 0.95703125,
"learning_rate": 9.832649825713266e-06,
"loss": 0.1198,
"step": 93
},
{
"epoch": 0.46248462484624847,
"grad_norm": 0.96875,
"learning_rate": 9.828616333357743e-06,
"loss": 0.1148,
"step": 94
},
{
"epoch": 0.46740467404674046,
"grad_norm": 0.79296875,
"learning_rate": 9.8245356575224e-06,
"loss": 0.1074,
"step": 95
},
{
"epoch": 0.47232472324723246,
"grad_norm": 0.7109375,
"learning_rate": 9.820407838082109e-06,
"loss": 0.1193,
"step": 96
},
{
"epoch": 0.47724477244772445,
"grad_norm": 0.66796875,
"learning_rate": 9.816232915372423e-06,
"loss": 0.0722,
"step": 97
},
{
"epoch": 0.4821648216482165,
"grad_norm": 0.9765625,
"learning_rate": 9.812010930189163e-06,
"loss": 0.0929,
"step": 98
},
{
"epoch": 0.4870848708487085,
"grad_norm": 0.89453125,
"learning_rate": 9.80774192378803e-06,
"loss": 0.0973,
"step": 99
},
{
"epoch": 0.4920049200492005,
"grad_norm": 1.1328125,
"learning_rate": 9.803425937884202e-06,
"loss": 0.1335,
"step": 100
},
{
"epoch": 0.4969249692496925,
"grad_norm": 0.77734375,
"learning_rate": 9.799063014651918e-06,
"loss": 0.1169,
"step": 101
},
{
"epoch": 0.5018450184501845,
"grad_norm": 0.93359375,
"learning_rate": 9.794653196724077e-06,
"loss": 0.1159,
"step": 102
},
{
"epoch": 0.5018450184501845,
"eval_loss": 0.12004748731851578,
"eval_runtime": 203.4298,
"eval_samples_per_second": 8.042,
"eval_steps_per_second": 1.342,
"step": 102
},
{
"epoch": 0.5067650676506765,
"grad_norm": 0.9375,
"learning_rate": 9.790196527191811e-06,
"loss": 0.1142,
"step": 103
},
{
"epoch": 0.5116851168511685,
"grad_norm": 0.8203125,
"learning_rate": 9.785693049604073e-06,
"loss": 0.1113,
"step": 104
},
{
"epoch": 0.5166051660516605,
"grad_norm": 0.80859375,
"learning_rate": 9.781142807967205e-06,
"loss": 0.1035,
"step": 105
},
{
"epoch": 0.5215252152521526,
"grad_norm": 0.91796875,
"learning_rate": 9.776545846744509e-06,
"loss": 0.1238,
"step": 106
},
{
"epoch": 0.5264452644526445,
"grad_norm": 1.4765625,
"learning_rate": 9.771902210855819e-06,
"loss": 0.0921,
"step": 107
},
{
"epoch": 0.5313653136531366,
"grad_norm": 1.1015625,
"learning_rate": 9.767211945677051e-06,
"loss": 0.1459,
"step": 108
},
{
"epoch": 0.5362853628536285,
"grad_norm": 1.0,
"learning_rate": 9.762475097039767e-06,
"loss": 0.101,
"step": 109
},
{
"epoch": 0.5412054120541205,
"grad_norm": 1.078125,
"learning_rate": 9.757691711230728e-06,
"loss": 0.1259,
"step": 110
},
{
"epoch": 0.5461254612546126,
"grad_norm": 0.8828125,
"learning_rate": 9.752861834991436e-06,
"loss": 0.1146,
"step": 111
},
{
"epoch": 0.5510455104551045,
"grad_norm": 1.0078125,
"learning_rate": 9.747985515517683e-06,
"loss": 0.1609,
"step": 112
},
{
"epoch": 0.5559655596555966,
"grad_norm": 0.94921875,
"learning_rate": 9.743062800459089e-06,
"loss": 0.1242,
"step": 113
},
{
"epoch": 0.5608856088560885,
"grad_norm": 0.9765625,
"learning_rate": 9.738093737918627e-06,
"loss": 0.1095,
"step": 114
},
{
"epoch": 0.5658056580565806,
"grad_norm": 1.0,
"learning_rate": 9.733078376452172e-06,
"loss": 0.1083,
"step": 115
},
{
"epoch": 0.5707257072570726,
"grad_norm": 1.0390625,
"learning_rate": 9.728016765068004e-06,
"loss": 0.1428,
"step": 116
},
{
"epoch": 0.5756457564575646,
"grad_norm": 0.859375,
"learning_rate": 9.722908953226349e-06,
"loss": 0.1073,
"step": 117
},
{
"epoch": 0.5805658056580566,
"grad_norm": 0.8125,
"learning_rate": 9.717754990838882e-06,
"loss": 0.1049,
"step": 118
},
{
"epoch": 0.5854858548585485,
"grad_norm": 1.140625,
"learning_rate": 9.712554928268245e-06,
"loss": 0.1479,
"step": 119
},
{
"epoch": 0.5904059040590406,
"grad_norm": 0.90234375,
"learning_rate": 9.707308816327557e-06,
"loss": 0.1195,
"step": 120
},
{
"epoch": 0.5953259532595326,
"grad_norm": 0.82421875,
"learning_rate": 9.702016706279914e-06,
"loss": 0.1048,
"step": 121
},
{
"epoch": 0.6002460024600246,
"grad_norm": 0.984375,
"learning_rate": 9.696678649837883e-06,
"loss": 0.1073,
"step": 122
},
{
"epoch": 0.6051660516605166,
"grad_norm": 1.0859375,
"learning_rate": 9.691294699163013e-06,
"loss": 0.1308,
"step": 123
},
{
"epoch": 0.6100861008610086,
"grad_norm": 0.86328125,
"learning_rate": 9.685864906865303e-06,
"loss": 0.1248,
"step": 124
},
{
"epoch": 0.6150061500615006,
"grad_norm": 0.8515625,
"learning_rate": 9.680389326002708e-06,
"loss": 0.1312,
"step": 125
},
{
"epoch": 0.6199261992619927,
"grad_norm": 0.71875,
"learning_rate": 9.67486801008061e-06,
"loss": 0.0868,
"step": 126
},
{
"epoch": 0.6248462484624846,
"grad_norm": 0.76953125,
"learning_rate": 9.669301013051297e-06,
"loss": 0.0859,
"step": 127
},
{
"epoch": 0.6297662976629766,
"grad_norm": 0.734375,
"learning_rate": 9.663688389313435e-06,
"loss": 0.0878,
"step": 128
},
{
"epoch": 0.6346863468634686,
"grad_norm": 1.0234375,
"learning_rate": 9.658030193711538e-06,
"loss": 0.1229,
"step": 129
},
{
"epoch": 0.6396063960639606,
"grad_norm": 0.97265625,
"learning_rate": 9.652326481535434e-06,
"loss": 0.1439,
"step": 130
},
{
"epoch": 0.6445264452644527,
"grad_norm": 0.703125,
"learning_rate": 9.646577308519719e-06,
"loss": 0.1063,
"step": 131
},
{
"epoch": 0.6494464944649446,
"grad_norm": 0.9609375,
"learning_rate": 9.640782730843219e-06,
"loss": 0.1023,
"step": 132
},
{
"epoch": 0.6543665436654367,
"grad_norm": 0.984375,
"learning_rate": 9.634942805128433e-06,
"loss": 0.1002,
"step": 133
},
{
"epoch": 0.6592865928659286,
"grad_norm": 0.76953125,
"learning_rate": 9.629057588440993e-06,
"loss": 0.0978,
"step": 134
},
{
"epoch": 0.6642066420664207,
"grad_norm": 0.8671875,
"learning_rate": 9.623127138289087e-06,
"loss": 0.1042,
"step": 135
},
{
"epoch": 0.6691266912669127,
"grad_norm": 0.8515625,
"learning_rate": 9.617151512622918e-06,
"loss": 0.0873,
"step": 136
},
{
"epoch": 0.6740467404674046,
"grad_norm": 0.91796875,
"learning_rate": 9.611130769834121e-06,
"loss": 0.1118,
"step": 137
},
{
"epoch": 0.6789667896678967,
"grad_norm": 0.9375,
"learning_rate": 9.6050649687552e-06,
"loss": 0.13,
"step": 138
},
{
"epoch": 0.6838868388683886,
"grad_norm": 0.828125,
"learning_rate": 9.598954168658956e-06,
"loss": 0.1096,
"step": 139
},
{
"epoch": 0.6888068880688807,
"grad_norm": 0.86328125,
"learning_rate": 9.592798429257899e-06,
"loss": 0.1219,
"step": 140
},
{
"epoch": 0.6937269372693727,
"grad_norm": 0.95703125,
"learning_rate": 9.586597810703674e-06,
"loss": 0.1002,
"step": 141
},
{
"epoch": 0.6986469864698647,
"grad_norm": 0.765625,
"learning_rate": 9.580352373586468e-06,
"loss": 0.0765,
"step": 142
},
{
"epoch": 0.7035670356703567,
"grad_norm": 0.90234375,
"learning_rate": 9.574062178934414e-06,
"loss": 0.1227,
"step": 143
},
{
"epoch": 0.7084870848708487,
"grad_norm": 0.82421875,
"learning_rate": 9.567727288213005e-06,
"loss": 0.1052,
"step": 144
},
{
"epoch": 0.7134071340713407,
"grad_norm": 0.87109375,
"learning_rate": 9.561347763324484e-06,
"loss": 0.0909,
"step": 145
},
{
"epoch": 0.7183271832718328,
"grad_norm": 1.0546875,
"learning_rate": 9.554923666607245e-06,
"loss": 0.1219,
"step": 146
},
{
"epoch": 0.7232472324723247,
"grad_norm": 0.98046875,
"learning_rate": 9.548455060835223e-06,
"loss": 0.1077,
"step": 147
},
{
"epoch": 0.7281672816728167,
"grad_norm": 0.94921875,
"learning_rate": 9.541942009217273e-06,
"loss": 0.132,
"step": 148
},
{
"epoch": 0.7330873308733087,
"grad_norm": 1.03125,
"learning_rate": 9.535384575396565e-06,
"loss": 0.1462,
"step": 149
},
{
"epoch": 0.7380073800738007,
"grad_norm": 0.79296875,
"learning_rate": 9.528782823449954e-06,
"loss": 0.0951,
"step": 150
},
{
"epoch": 0.7429274292742928,
"grad_norm": 0.99609375,
"learning_rate": 9.522136817887353e-06,
"loss": 0.1451,
"step": 151
},
{
"epoch": 0.7478474784747847,
"grad_norm": 0.81640625,
"learning_rate": 9.51544662365111e-06,
"loss": 0.072,
"step": 152
},
{
"epoch": 0.7527675276752768,
"grad_norm": 0.80078125,
"learning_rate": 9.508712306115366e-06,
"loss": 0.1202,
"step": 153
},
{
"epoch": 0.7576875768757687,
"grad_norm": 0.890625,
"learning_rate": 9.501933931085416e-06,
"loss": 0.1068,
"step": 154
},
{
"epoch": 0.7626076260762608,
"grad_norm": 0.8984375,
"learning_rate": 9.495111564797073e-06,
"loss": 0.1091,
"step": 155
},
{
"epoch": 0.7675276752767528,
"grad_norm": 0.84375,
"learning_rate": 9.488245273916016e-06,
"loss": 0.0982,
"step": 156
},
{
"epoch": 0.7724477244772447,
"grad_norm": 0.828125,
"learning_rate": 9.481335125537138e-06,
"loss": 0.1064,
"step": 157
},
{
"epoch": 0.7773677736777368,
"grad_norm": 0.8828125,
"learning_rate": 9.47438118718389e-06,
"loss": 0.1027,
"step": 158
},
{
"epoch": 0.7822878228782287,
"grad_norm": 0.765625,
"learning_rate": 9.467383526807626e-06,
"loss": 0.0917,
"step": 159
},
{
"epoch": 0.7872078720787208,
"grad_norm": 0.875,
"learning_rate": 9.460342212786933e-06,
"loss": 0.0975,
"step": 160
},
{
"epoch": 0.7921279212792128,
"grad_norm": 1.03125,
"learning_rate": 9.453257313926969e-06,
"loss": 0.0837,
"step": 161
},
{
"epoch": 0.7970479704797048,
"grad_norm": 0.68359375,
"learning_rate": 9.44612889945878e-06,
"loss": 0.093,
"step": 162
},
{
"epoch": 0.8019680196801968,
"grad_norm": 0.7578125,
"learning_rate": 9.43895703903864e-06,
"loss": 0.1094,
"step": 163
},
{
"epoch": 0.8068880688806888,
"grad_norm": 1.015625,
"learning_rate": 9.431741802747355e-06,
"loss": 0.1081,
"step": 164
},
{
"epoch": 0.8118081180811808,
"grad_norm": 0.96875,
"learning_rate": 9.424483261089584e-06,
"loss": 0.1207,
"step": 165
},
{
"epoch": 0.8167281672816729,
"grad_norm": 0.67578125,
"learning_rate": 9.417181484993154e-06,
"loss": 0.0717,
"step": 166
},
{
"epoch": 0.8216482164821648,
"grad_norm": 0.99609375,
"learning_rate": 9.40983654580836e-06,
"loss": 0.1013,
"step": 167
},
{
"epoch": 0.8265682656826568,
"grad_norm": 0.87890625,
"learning_rate": 9.402448515307274e-06,
"loss": 0.1122,
"step": 168
},
{
"epoch": 0.8314883148831488,
"grad_norm": 0.99609375,
"learning_rate": 9.395017465683036e-06,
"loss": 0.1234,
"step": 169
},
{
"epoch": 0.8364083640836408,
"grad_norm": 0.85546875,
"learning_rate": 9.387543469549156e-06,
"loss": 0.1259,
"step": 170
},
{
"epoch": 0.8413284132841329,
"grad_norm": 0.78515625,
"learning_rate": 9.380026599938804e-06,
"loss": 0.1007,
"step": 171
},
{
"epoch": 0.8462484624846248,
"grad_norm": 0.95703125,
"learning_rate": 9.372466930304091e-06,
"loss": 0.1148,
"step": 172
},
{
"epoch": 0.8511685116851169,
"grad_norm": 0.765625,
"learning_rate": 9.364864534515353e-06,
"loss": 0.11,
"step": 173
},
{
"epoch": 0.8560885608856088,
"grad_norm": 1.0078125,
"learning_rate": 9.357219486860434e-06,
"loss": 0.1255,
"step": 174
},
{
"epoch": 0.8610086100861009,
"grad_norm": 0.8671875,
"learning_rate": 9.349531862043952e-06,
"loss": 0.0973,
"step": 175
},
{
"epoch": 0.8659286592865929,
"grad_norm": 0.8671875,
"learning_rate": 9.34180173518658e-06,
"loss": 0.0872,
"step": 176
},
{
"epoch": 0.8708487084870848,
"grad_norm": 0.796875,
"learning_rate": 9.3340291818243e-06,
"loss": 0.089,
"step": 177
},
{
"epoch": 0.8757687576875769,
"grad_norm": 1.0078125,
"learning_rate": 9.32621427790767e-06,
"loss": 0.1148,
"step": 178
},
{
"epoch": 0.8806888068880688,
"grad_norm": 0.78515625,
"learning_rate": 9.318357099801087e-06,
"loss": 0.0896,
"step": 179
},
{
"epoch": 0.8856088560885609,
"grad_norm": 0.9296875,
"learning_rate": 9.310457724282034e-06,
"loss": 0.1104,
"step": 180
},
{
"epoch": 0.8905289052890529,
"grad_norm": 0.859375,
"learning_rate": 9.302516228540328e-06,
"loss": 0.1115,
"step": 181
},
{
"epoch": 0.8954489544895449,
"grad_norm": 0.921875,
"learning_rate": 9.294532690177373e-06,
"loss": 0.12,
"step": 182
},
{
"epoch": 0.9003690036900369,
"grad_norm": 1.0625,
"learning_rate": 9.286507187205399e-06,
"loss": 0.137,
"step": 183
},
{
"epoch": 0.9052890528905289,
"grad_norm": 0.80859375,
"learning_rate": 9.278439798046697e-06,
"loss": 0.1273,
"step": 184
},
{
"epoch": 0.9102091020910209,
"grad_norm": 0.96484375,
"learning_rate": 9.270330601532855e-06,
"loss": 0.1339,
"step": 185
},
{
"epoch": 0.915129151291513,
"grad_norm": 0.90234375,
"learning_rate": 9.262179676903986e-06,
"loss": 0.1011,
"step": 186
},
{
"epoch": 0.9200492004920049,
"grad_norm": 0.90234375,
"learning_rate": 9.25398710380796e-06,
"loss": 0.1096,
"step": 187
},
{
"epoch": 0.9249692496924969,
"grad_norm": 0.8125,
"learning_rate": 9.245752962299612e-06,
"loss": 0.1316,
"step": 188
},
{
"epoch": 0.9298892988929889,
"grad_norm": 0.93359375,
"learning_rate": 9.237477332839975e-06,
"loss": 0.0908,
"step": 189
},
{
"epoch": 0.9348093480934809,
"grad_norm": 0.97265625,
"learning_rate": 9.229160296295488e-06,
"loss": 0.138,
"step": 190
},
{
"epoch": 0.939729397293973,
"grad_norm": 0.81640625,
"learning_rate": 9.220801933937199e-06,
"loss": 0.1259,
"step": 191
},
{
"epoch": 0.9446494464944649,
"grad_norm": 0.92578125,
"learning_rate": 9.212402327439982e-06,
"loss": 0.117,
"step": 192
},
{
"epoch": 0.949569495694957,
"grad_norm": 1.0859375,
"learning_rate": 9.20396155888173e-06,
"loss": 0.1405,
"step": 193
},
{
"epoch": 0.9544895448954489,
"grad_norm": 1.015625,
"learning_rate": 9.19547971074256e-06,
"loss": 0.1154,
"step": 194
},
{
"epoch": 0.959409594095941,
"grad_norm": 0.8828125,
"learning_rate": 9.186956865904004e-06,
"loss": 0.1203,
"step": 195
},
{
"epoch": 0.964329643296433,
"grad_norm": 0.71875,
"learning_rate": 9.178393107648193e-06,
"loss": 0.1009,
"step": 196
},
{
"epoch": 0.9692496924969249,
"grad_norm": 0.859375,
"learning_rate": 9.169788519657056e-06,
"loss": 0.1023,
"step": 197
},
{
"epoch": 0.974169741697417,
"grad_norm": 1.015625,
"learning_rate": 9.161143186011492e-06,
"loss": 0.1402,
"step": 198
},
{
"epoch": 0.9790897908979089,
"grad_norm": 0.828125,
"learning_rate": 9.15245719119055e-06,
"loss": 0.1065,
"step": 199
},
{
"epoch": 0.984009840098401,
"grad_norm": 0.8515625,
"learning_rate": 9.143730620070609e-06,
"loss": 0.097,
"step": 200
},
{
"epoch": 0.988929889298893,
"grad_norm": 0.87109375,
"learning_rate": 9.134963557924543e-06,
"loss": 0.0979,
"step": 201
},
{
"epoch": 0.993849938499385,
"grad_norm": 0.87109375,
"learning_rate": 9.12615609042089e-06,
"loss": 0.1012,
"step": 202
},
{
"epoch": 0.998769987699877,
"grad_norm": 0.7578125,
"learning_rate": 9.11730830362301e-06,
"loss": 0.0764,
"step": 203
},
{
"epoch": 1.0030750307503076,
"grad_norm": 1.9296875,
"learning_rate": 9.10842028398826e-06,
"loss": 0.1883,
"step": 204
},
{
"epoch": 1.0030750307503076,
"eval_loss": 0.11158129572868347,
"eval_runtime": 204.933,
"eval_samples_per_second": 7.983,
"eval_steps_per_second": 1.332,
"step": 204
},
{
"epoch": 1.0079950799507995,
"grad_norm": 1.09375,
"learning_rate": 9.099492118367123e-06,
"loss": 0.1227,
"step": 205
},
{
"epoch": 1.0129151291512914,
"grad_norm": 0.80078125,
"learning_rate": 9.090523894002386e-06,
"loss": 0.0759,
"step": 206
},
{
"epoch": 1.0178351783517836,
"grad_norm": 1.234375,
"learning_rate": 9.081515698528267e-06,
"loss": 0.1026,
"step": 207
},
{
"epoch": 1.0227552275522755,
"grad_norm": 1.296875,
"learning_rate": 9.072467619969574e-06,
"loss": 0.1144,
"step": 208
},
{
"epoch": 1.0276752767527675,
"grad_norm": 1.2578125,
"learning_rate": 9.06337974674083e-06,
"loss": 0.1129,
"step": 209
},
{
"epoch": 1.0325953259532596,
"grad_norm": 1.5078125,
"learning_rate": 9.054252167645426e-06,
"loss": 0.081,
"step": 210
},
{
"epoch": 1.0375153751537516,
"grad_norm": 1.1640625,
"learning_rate": 9.045084971874738e-06,
"loss": 0.1211,
"step": 211
},
{
"epoch": 1.0424354243542435,
"grad_norm": 1.0546875,
"learning_rate": 9.035878249007264e-06,
"loss": 0.0857,
"step": 212
},
{
"epoch": 1.0473554735547355,
"grad_norm": 1.0703125,
"learning_rate": 9.026632089007745e-06,
"loss": 0.1407,
"step": 213
},
{
"epoch": 1.0522755227552276,
"grad_norm": 1.1484375,
"learning_rate": 9.017346582226289e-06,
"loss": 0.0835,
"step": 214
},
{
"epoch": 1.0571955719557196,
"grad_norm": 1.21875,
"learning_rate": 9.008021819397488e-06,
"loss": 0.1263,
"step": 215
},
{
"epoch": 1.0621156211562115,
"grad_norm": 1.2109375,
"learning_rate": 8.998657891639523e-06,
"loss": 0.0747,
"step": 216
},
{
"epoch": 1.0670356703567037,
"grad_norm": 0.89453125,
"learning_rate": 8.989254890453289e-06,
"loss": 0.09,
"step": 217
},
{
"epoch": 1.0719557195571956,
"grad_norm": 1.0,
"learning_rate": 8.979812907721485e-06,
"loss": 0.0924,
"step": 218
},
{
"epoch": 1.0768757687576875,
"grad_norm": 1.1171875,
"learning_rate": 8.97033203570773e-06,
"loss": 0.1017,
"step": 219
},
{
"epoch": 1.0817958179581795,
"grad_norm": 0.9921875,
"learning_rate": 8.960812367055646e-06,
"loss": 0.1033,
"step": 220
},
{
"epoch": 1.0867158671586716,
"grad_norm": 1.3515625,
"learning_rate": 8.951253994787975e-06,
"loss": 0.1425,
"step": 221
},
{
"epoch": 1.0916359163591636,
"grad_norm": 1.140625,
"learning_rate": 8.941657012305644e-06,
"loss": 0.1076,
"step": 222
},
{
"epoch": 1.0965559655596555,
"grad_norm": 0.9296875,
"learning_rate": 8.93202151338687e-06,
"loss": 0.081,
"step": 223
},
{
"epoch": 1.1014760147601477,
"grad_norm": 0.90234375,
"learning_rate": 8.922347592186237e-06,
"loss": 0.096,
"step": 224
},
{
"epoch": 1.1063960639606396,
"grad_norm": 0.79296875,
"learning_rate": 8.912635343233784e-06,
"loss": 0.0681,
"step": 225
},
{
"epoch": 1.1113161131611315,
"grad_norm": 1.28125,
"learning_rate": 8.902884861434066e-06,
"loss": 0.117,
"step": 226
},
{
"epoch": 1.1162361623616237,
"grad_norm": 1.2734375,
"learning_rate": 8.89309624206524e-06,
"loss": 0.1133,
"step": 227
},
{
"epoch": 1.1211562115621156,
"grad_norm": 0.96484375,
"learning_rate": 8.883269580778127e-06,
"loss": 0.0933,
"step": 228
},
{
"epoch": 1.1260762607626076,
"grad_norm": 1.2265625,
"learning_rate": 8.873404973595284e-06,
"loss": 0.1243,
"step": 229
},
{
"epoch": 1.1309963099630997,
"grad_norm": 1.6796875,
"learning_rate": 8.863502516910058e-06,
"loss": 0.0959,
"step": 230
},
{
"epoch": 1.1359163591635917,
"grad_norm": 1.0390625,
"learning_rate": 8.853562307485649e-06,
"loss": 0.1095,
"step": 231
},
{
"epoch": 1.1408364083640836,
"grad_norm": 1.3984375,
"learning_rate": 8.84358444245416e-06,
"loss": 0.1128,
"step": 232
},
{
"epoch": 1.1457564575645756,
"grad_norm": 1.046875,
"learning_rate": 8.833569019315654e-06,
"loss": 0.098,
"step": 233
},
{
"epoch": 1.1506765067650677,
"grad_norm": 1.6328125,
"learning_rate": 8.8235161359372e-06,
"loss": 0.1001,
"step": 234
},
{
"epoch": 1.1555965559655597,
"grad_norm": 1.1796875,
"learning_rate": 8.81342589055191e-06,
"loss": 0.1213,
"step": 235
},
{
"epoch": 1.1605166051660516,
"grad_norm": 1.078125,
"learning_rate": 8.803298381757987e-06,
"loss": 0.0891,
"step": 236
},
{
"epoch": 1.1654366543665438,
"grad_norm": 0.9375,
"learning_rate": 8.793133708517759e-06,
"loss": 0.0878,
"step": 237
},
{
"epoch": 1.1703567035670357,
"grad_norm": 1.0625,
"learning_rate": 8.782931970156708e-06,
"loss": 0.0967,
"step": 238
},
{
"epoch": 1.1752767527675276,
"grad_norm": 1.171875,
"learning_rate": 8.772693266362504e-06,
"loss": 0.0976,
"step": 239
},
{
"epoch": 1.1801968019680196,
"grad_norm": 0.875,
"learning_rate": 8.762417697184034e-06,
"loss": 0.0737,
"step": 240
},
{
"epoch": 1.1851168511685117,
"grad_norm": 1.109375,
"learning_rate": 8.752105363030414e-06,
"loss": 0.1093,
"step": 241
},
{
"epoch": 1.1900369003690037,
"grad_norm": 1.0703125,
"learning_rate": 8.741756364670018e-06,
"loss": 0.0853,
"step": 242
},
{
"epoch": 1.1949569495694956,
"grad_norm": 1.1640625,
"learning_rate": 8.731370803229488e-06,
"loss": 0.0941,
"step": 243
},
{
"epoch": 1.1998769987699878,
"grad_norm": 1.125,
"learning_rate": 8.720948780192747e-06,
"loss": 0.0935,
"step": 244
},
{
"epoch": 1.2047970479704797,
"grad_norm": 1.109375,
"learning_rate": 8.710490397400007e-06,
"loss": 0.1011,
"step": 245
},
{
"epoch": 1.2097170971709716,
"grad_norm": 1.171875,
"learning_rate": 8.699995757046773e-06,
"loss": 0.1141,
"step": 246
},
{
"epoch": 1.2146371463714638,
"grad_norm": 1.421875,
"learning_rate": 8.689464961682853e-06,
"loss": 0.1281,
"step": 247
},
{
"epoch": 1.2195571955719557,
"grad_norm": 1.1953125,
"learning_rate": 8.678898114211338e-06,
"loss": 0.1005,
"step": 248
},
{
"epoch": 1.2244772447724477,
"grad_norm": 1.1171875,
"learning_rate": 8.668295317887615e-06,
"loss": 0.1019,
"step": 249
},
{
"epoch": 1.2293972939729398,
"grad_norm": 1.0703125,
"learning_rate": 8.657656676318346e-06,
"loss": 0.1101,
"step": 250
},
{
"epoch": 1.2343173431734318,
"grad_norm": 1.2265625,
"learning_rate": 8.646982293460461e-06,
"loss": 0.1366,
"step": 251
},
{
"epoch": 1.2392373923739237,
"grad_norm": 1.21875,
"learning_rate": 8.63627227362014e-06,
"loss": 0.0937,
"step": 252
},
{
"epoch": 1.2441574415744157,
"grad_norm": 1.2265625,
"learning_rate": 8.625526721451799e-06,
"loss": 0.1273,
"step": 253
},
{
"epoch": 1.2490774907749078,
"grad_norm": 1.125,
"learning_rate": 8.614745741957054e-06,
"loss": 0.097,
"step": 254
},
{
"epoch": 1.2539975399753998,
"grad_norm": 1.4453125,
"learning_rate": 8.603929440483714e-06,
"loss": 0.1064,
"step": 255
},
{
"epoch": 1.2589175891758917,
"grad_norm": 1.1953125,
"learning_rate": 8.593077922724733e-06,
"loss": 0.118,
"step": 256
},
{
"epoch": 1.2638376383763839,
"grad_norm": 1.0625,
"learning_rate": 8.582191294717192e-06,
"loss": 0.0993,
"step": 257
},
{
"epoch": 1.2687576875768758,
"grad_norm": 1.34375,
"learning_rate": 8.571269662841253e-06,
"loss": 0.0659,
"step": 258
},
{
"epoch": 1.2736777367773677,
"grad_norm": 1.3828125,
"learning_rate": 8.560313133819124e-06,
"loss": 0.12,
"step": 259
},
{
"epoch": 1.2785977859778597,
"grad_norm": 1.390625,
"learning_rate": 8.549321814714018e-06,
"loss": 0.0868,
"step": 260
},
{
"epoch": 1.2835178351783518,
"grad_norm": 1.2734375,
"learning_rate": 8.538295812929096e-06,
"loss": 0.0902,
"step": 261
},
{
"epoch": 1.2884378843788438,
"grad_norm": 1.09375,
"learning_rate": 8.527235236206438e-06,
"loss": 0.0966,
"step": 262
},
{
"epoch": 1.293357933579336,
"grad_norm": 0.984375,
"learning_rate": 8.516140192625962e-06,
"loss": 0.0993,
"step": 263
},
{
"epoch": 1.2982779827798279,
"grad_norm": 0.97265625,
"learning_rate": 8.5050107906044e-06,
"loss": 0.0865,
"step": 264
},
{
"epoch": 1.3031980319803198,
"grad_norm": 1.1640625,
"learning_rate": 8.49384713889421e-06,
"loss": 0.089,
"step": 265
},
{
"epoch": 1.3081180811808117,
"grad_norm": 1.046875,
"learning_rate": 8.482649346582529e-06,
"loss": 0.0956,
"step": 266
},
{
"epoch": 1.3130381303813037,
"grad_norm": 1.1953125,
"learning_rate": 8.471417523090109e-06,
"loss": 0.0985,
"step": 267
},
{
"epoch": 1.3179581795817958,
"grad_norm": 0.875,
"learning_rate": 8.460151778170231e-06,
"loss": 0.0789,
"step": 268
},
{
"epoch": 1.3228782287822878,
"grad_norm": 1.3515625,
"learning_rate": 8.448852221907657e-06,
"loss": 0.0995,
"step": 269
},
{
"epoch": 1.32779827798278,
"grad_norm": 1.1328125,
"learning_rate": 8.43751896471753e-06,
"loss": 0.1006,
"step": 270
},
{
"epoch": 1.3327183271832719,
"grad_norm": 1.1328125,
"learning_rate": 8.426152117344314e-06,
"loss": 0.1119,
"step": 271
},
{
"epoch": 1.3376383763837638,
"grad_norm": 1.3359375,
"learning_rate": 8.414751790860697e-06,
"loss": 0.1329,
"step": 272
},
{
"epoch": 1.3425584255842558,
"grad_norm": 1.0234375,
"learning_rate": 8.40331809666652e-06,
"loss": 0.0863,
"step": 273
},
{
"epoch": 1.347478474784748,
"grad_norm": 1.265625,
"learning_rate": 8.391851146487675e-06,
"loss": 0.1255,
"step": 274
},
{
"epoch": 1.3523985239852399,
"grad_norm": 1.453125,
"learning_rate": 8.380351052375023e-06,
"loss": 0.1031,
"step": 275
},
{
"epoch": 1.3573185731857318,
"grad_norm": 1.03125,
"learning_rate": 8.368817926703296e-06,
"loss": 0.0798,
"step": 276
},
{
"epoch": 1.362238622386224,
"grad_norm": 1.28125,
"learning_rate": 8.357251882169993e-06,
"loss": 0.0923,
"step": 277
},
{
"epoch": 1.367158671586716,
"grad_norm": 1.0859375,
"learning_rate": 8.345653031794292e-06,
"loss": 0.0831,
"step": 278
},
{
"epoch": 1.3720787207872078,
"grad_norm": 1.09375,
"learning_rate": 8.33402148891593e-06,
"loss": 0.071,
"step": 279
},
{
"epoch": 1.3769987699876998,
"grad_norm": 1.1015625,
"learning_rate": 8.32235736719411e-06,
"loss": 0.0697,
"step": 280
},
{
"epoch": 1.381918819188192,
"grad_norm": 0.97265625,
"learning_rate": 8.310660780606376e-06,
"loss": 0.0704,
"step": 281
},
{
"epoch": 1.3868388683886839,
"grad_norm": 1.34375,
"learning_rate": 8.298931843447517e-06,
"loss": 0.1128,
"step": 282
},
{
"epoch": 1.391758917589176,
"grad_norm": 1.2421875,
"learning_rate": 8.28717067032843e-06,
"loss": 0.1112,
"step": 283
},
{
"epoch": 1.396678966789668,
"grad_norm": 0.90625,
"learning_rate": 8.275377376175015e-06,
"loss": 0.0798,
"step": 284
},
{
"epoch": 1.40159901599016,
"grad_norm": 1.1484375,
"learning_rate": 8.263552076227048e-06,
"loss": 0.1072,
"step": 285
},
{
"epoch": 1.4065190651906518,
"grad_norm": 1.0390625,
"learning_rate": 8.251694886037052e-06,
"loss": 0.0953,
"step": 286
},
{
"epoch": 1.4114391143911438,
"grad_norm": 1.21875,
"learning_rate": 8.239805921469168e-06,
"loss": 0.1145,
"step": 287
},
{
"epoch": 1.416359163591636,
"grad_norm": 1.4453125,
"learning_rate": 8.227885298698029e-06,
"loss": 0.1047,
"step": 288
},
{
"epoch": 1.4212792127921279,
"grad_norm": 1.046875,
"learning_rate": 8.21593313420762e-06,
"loss": 0.0905,
"step": 289
},
{
"epoch": 1.42619926199262,
"grad_norm": 0.9609375,
"learning_rate": 8.203949544790131e-06,
"loss": 0.1013,
"step": 290
},
{
"epoch": 1.431119311193112,
"grad_norm": 1.2109375,
"learning_rate": 8.19193464754484e-06,
"loss": 0.085,
"step": 291
},
{
"epoch": 1.436039360393604,
"grad_norm": 1.2421875,
"learning_rate": 8.179888559876943e-06,
"loss": 0.1029,
"step": 292
},
{
"epoch": 1.4409594095940959,
"grad_norm": 1.1796875,
"learning_rate": 8.16781139949642e-06,
"loss": 0.1262,
"step": 293
},
{
"epoch": 1.445879458794588,
"grad_norm": 1.421875,
"learning_rate": 8.155703284416884e-06,
"loss": 0.125,
"step": 294
},
{
"epoch": 1.45079950799508,
"grad_norm": 0.8984375,
"learning_rate": 8.143564332954426e-06,
"loss": 0.0796,
"step": 295
},
{
"epoch": 1.455719557195572,
"grad_norm": 1.4453125,
"learning_rate": 8.131394663726452e-06,
"loss": 0.1024,
"step": 296
},
{
"epoch": 1.460639606396064,
"grad_norm": 1.109375,
"learning_rate": 8.119194395650545e-06,
"loss": 0.0994,
"step": 297
},
{
"epoch": 1.465559655596556,
"grad_norm": 1.15625,
"learning_rate": 8.106963647943273e-06,
"loss": 0.1104,
"step": 298
},
{
"epoch": 1.470479704797048,
"grad_norm": 1.359375,
"learning_rate": 8.09470254011905e-06,
"loss": 0.1446,
"step": 299
},
{
"epoch": 1.4753997539975399,
"grad_norm": 1.0078125,
"learning_rate": 8.082411191988956e-06,
"loss": 0.1061,
"step": 300
},
{
"epoch": 1.480319803198032,
"grad_norm": 0.953125,
"learning_rate": 8.070089723659567e-06,
"loss": 0.074,
"step": 301
},
{
"epoch": 1.485239852398524,
"grad_norm": 1.296875,
"learning_rate": 8.057738255531781e-06,
"loss": 0.0852,
"step": 302
},
{
"epoch": 1.4901599015990161,
"grad_norm": 1.015625,
"learning_rate": 8.045356908299647e-06,
"loss": 0.091,
"step": 303
},
{
"epoch": 1.495079950799508,
"grad_norm": 1.0,
"learning_rate": 8.032945802949179e-06,
"loss": 0.0849,
"step": 304
},
{
"epoch": 1.5,
"grad_norm": 0.98828125,
"learning_rate": 8.02050506075718e-06,
"loss": 0.1125,
"step": 305
},
{
"epoch": 1.504920049200492,
"grad_norm": 0.8984375,
"learning_rate": 8.008034803290045e-06,
"loss": 0.086,
"step": 306
},
{
"epoch": 1.504920049200492,
"eval_loss": 0.10929346084594727,
"eval_runtime": 203.7138,
"eval_samples_per_second": 8.031,
"eval_steps_per_second": 1.34,
"step": 306
},
{
"epoch": 1.5098400984009839,
"grad_norm": 1.0390625,
"learning_rate": 7.995535152402592e-06,
"loss": 0.1118,
"step": 307
},
{
"epoch": 1.514760147601476,
"grad_norm": 1.203125,
"learning_rate": 7.983006230236855e-06,
"loss": 0.0959,
"step": 308
},
{
"epoch": 1.519680196801968,
"grad_norm": 1.234375,
"learning_rate": 7.970448159220897e-06,
"loss": 0.0916,
"step": 309
},
{
"epoch": 1.5246002460024601,
"grad_norm": 1.21875,
"learning_rate": 7.957861062067614e-06,
"loss": 0.1117,
"step": 310
},
{
"epoch": 1.529520295202952,
"grad_norm": 0.87109375,
"learning_rate": 7.945245061773531e-06,
"loss": 0.073,
"step": 311
},
{
"epoch": 1.534440344403444,
"grad_norm": 0.96875,
"learning_rate": 7.93260028161761e-06,
"loss": 0.072,
"step": 312
},
{
"epoch": 1.539360393603936,
"grad_norm": 1.046875,
"learning_rate": 7.919926845160037e-06,
"loss": 0.1082,
"step": 313
},
{
"epoch": 1.5442804428044279,
"grad_norm": 0.9375,
"learning_rate": 7.907224876241015e-06,
"loss": 0.0828,
"step": 314
},
{
"epoch": 1.54920049200492,
"grad_norm": 0.9296875,
"learning_rate": 7.894494498979558e-06,
"loss": 0.0924,
"step": 315
},
{
"epoch": 1.5541205412054122,
"grad_norm": 1.203125,
"learning_rate": 7.881735837772274e-06,
"loss": 0.1099,
"step": 316
},
{
"epoch": 1.5590405904059041,
"grad_norm": 0.984375,
"learning_rate": 7.868949017292153e-06,
"loss": 0.0909,
"step": 317
},
{
"epoch": 1.563960639606396,
"grad_norm": 1.09375,
"learning_rate": 7.856134162487346e-06,
"loss": 0.0986,
"step": 318
},
{
"epoch": 1.568880688806888,
"grad_norm": 0.94140625,
"learning_rate": 7.843291398579946e-06,
"loss": 0.1045,
"step": 319
},
{
"epoch": 1.57380073800738,
"grad_norm": 1.0234375,
"learning_rate": 7.830420851064767e-06,
"loss": 0.0971,
"step": 320
},
{
"epoch": 1.5787207872078721,
"grad_norm": 1.03125,
"learning_rate": 7.817522645708106e-06,
"loss": 0.0974,
"step": 321
},
{
"epoch": 1.583640836408364,
"grad_norm": 1.015625,
"learning_rate": 7.804596908546529e-06,
"loss": 0.1178,
"step": 322
},
{
"epoch": 1.5885608856088562,
"grad_norm": 1.1015625,
"learning_rate": 7.791643765885632e-06,
"loss": 0.093,
"step": 323
},
{
"epoch": 1.5934809348093482,
"grad_norm": 1.015625,
"learning_rate": 7.778663344298804e-06,
"loss": 0.0798,
"step": 324
},
{
"epoch": 1.59840098400984,
"grad_norm": 0.8828125,
"learning_rate": 7.765655770625997e-06,
"loss": 0.0991,
"step": 325
},
{
"epoch": 1.603321033210332,
"grad_norm": 0.93359375,
"learning_rate": 7.752621171972486e-06,
"loss": 0.0673,
"step": 326
},
{
"epoch": 1.608241082410824,
"grad_norm": 0.94140625,
"learning_rate": 7.739559675707615e-06,
"loss": 0.094,
"step": 327
},
{
"epoch": 1.6131611316113161,
"grad_norm": 1.15625,
"learning_rate": 7.726471409463572e-06,
"loss": 0.1286,
"step": 328
},
{
"epoch": 1.618081180811808,
"grad_norm": 1.3125,
"learning_rate": 7.713356501134124e-06,
"loss": 0.0938,
"step": 329
},
{
"epoch": 1.6230012300123002,
"grad_norm": 1.203125,
"learning_rate": 7.70021507887338e-06,
"loss": 0.1141,
"step": 330
},
{
"epoch": 1.6279212792127922,
"grad_norm": 0.71875,
"learning_rate": 7.687047271094528e-06,
"loss": 0.0708,
"step": 331
},
{
"epoch": 1.632841328413284,
"grad_norm": 0.96875,
"learning_rate": 7.673853206468593e-06,
"loss": 0.0816,
"step": 332
},
{
"epoch": 1.637761377613776,
"grad_norm": 1.3515625,
"learning_rate": 7.660633013923164e-06,
"loss": 0.1224,
"step": 333
},
{
"epoch": 1.642681426814268,
"grad_norm": 1.265625,
"learning_rate": 7.64738682264115e-06,
"loss": 0.1012,
"step": 334
},
{
"epoch": 1.6476014760147601,
"grad_norm": 0.8984375,
"learning_rate": 7.634114762059504e-06,
"loss": 0.103,
"step": 335
},
{
"epoch": 1.6525215252152523,
"grad_norm": 0.9375,
"learning_rate": 7.62081696186797e-06,
"loss": 0.0773,
"step": 336
},
{
"epoch": 1.6574415744157442,
"grad_norm": 0.98828125,
"learning_rate": 7.607493552007805e-06,
"loss": 0.1176,
"step": 337
},
{
"epoch": 1.6623616236162362,
"grad_norm": 0.9765625,
"learning_rate": 7.5941446626705175e-06,
"loss": 0.0996,
"step": 338
},
{
"epoch": 1.6672816728167281,
"grad_norm": 0.96484375,
"learning_rate": 7.580770424296591e-06,
"loss": 0.1001,
"step": 339
},
{
"epoch": 1.67220172201722,
"grad_norm": 0.8359375,
"learning_rate": 7.56737096757421e-06,
"loss": 0.1052,
"step": 340
},
{
"epoch": 1.6771217712177122,
"grad_norm": 1.28125,
"learning_rate": 7.553946423437988e-06,
"loss": 0.1104,
"step": 341
},
{
"epoch": 1.6820418204182042,
"grad_norm": 1.078125,
"learning_rate": 7.540496923067675e-06,
"loss": 0.1092,
"step": 342
},
{
"epoch": 1.6869618696186963,
"grad_norm": 1.1640625,
"learning_rate": 7.527022597886895e-06,
"loss": 0.0994,
"step": 343
},
{
"epoch": 1.6918819188191883,
"grad_norm": 1.484375,
"learning_rate": 7.513523579561839e-06,
"loss": 0.0787,
"step": 344
},
{
"epoch": 1.6968019680196802,
"grad_norm": 1.125,
"learning_rate": 7.500000000000001e-06,
"loss": 0.0986,
"step": 345
},
{
"epoch": 1.7017220172201721,
"grad_norm": 1.0546875,
"learning_rate": 7.486451991348872e-06,
"loss": 0.0806,
"step": 346
},
{
"epoch": 1.706642066420664,
"grad_norm": 1.2890625,
"learning_rate": 7.472879685994658e-06,
"loss": 0.1059,
"step": 347
},
{
"epoch": 1.7115621156211562,
"grad_norm": 1.0234375,
"learning_rate": 7.459283216560982e-06,
"loss": 0.0947,
"step": 348
},
{
"epoch": 1.7164821648216482,
"grad_norm": 1.015625,
"learning_rate": 7.445662715907591e-06,
"loss": 0.0897,
"step": 349
},
{
"epoch": 1.7214022140221403,
"grad_norm": 1.046875,
"learning_rate": 7.432018317129056e-06,
"loss": 0.1083,
"step": 350
},
{
"epoch": 1.7263222632226323,
"grad_norm": 0.8984375,
"learning_rate": 7.418350153553471e-06,
"loss": 0.07,
"step": 351
},
{
"epoch": 1.7312423124231242,
"grad_norm": 1.0703125,
"learning_rate": 7.40465835874115e-06,
"loss": 0.0942,
"step": 352
},
{
"epoch": 1.7361623616236161,
"grad_norm": 0.9765625,
"learning_rate": 7.390943066483327e-06,
"loss": 0.0881,
"step": 353
},
{
"epoch": 1.741082410824108,
"grad_norm": 1.046875,
"learning_rate": 7.377204410800839e-06,
"loss": 0.1029,
"step": 354
},
{
"epoch": 1.7460024600246002,
"grad_norm": 0.96484375,
"learning_rate": 7.363442525942827e-06,
"loss": 0.093,
"step": 355
},
{
"epoch": 1.7509225092250924,
"grad_norm": 1.1328125,
"learning_rate": 7.349657546385414e-06,
"loss": 0.0857,
"step": 356
},
{
"epoch": 1.7558425584255843,
"grad_norm": 0.9375,
"learning_rate": 7.335849606830402e-06,
"loss": 0.1016,
"step": 357
},
{
"epoch": 1.7607626076260763,
"grad_norm": 1.015625,
"learning_rate": 7.322018842203942e-06,
"loss": 0.1107,
"step": 358
},
{
"epoch": 1.7656826568265682,
"grad_norm": 1.15625,
"learning_rate": 7.308165387655231e-06,
"loss": 0.1411,
"step": 359
},
{
"epoch": 1.7706027060270602,
"grad_norm": 0.89453125,
"learning_rate": 7.294289378555179e-06,
"loss": 0.1037,
"step": 360
},
{
"epoch": 1.7755227552275523,
"grad_norm": 0.87109375,
"learning_rate": 7.2803909504950935e-06,
"loss": 0.0838,
"step": 361
},
{
"epoch": 1.7804428044280443,
"grad_norm": 0.99609375,
"learning_rate": 7.266470239285347e-06,
"loss": 0.0947,
"step": 362
},
{
"epoch": 1.7853628536285364,
"grad_norm": 1.0625,
"learning_rate": 7.252527380954062e-06,
"loss": 0.0904,
"step": 363
},
{
"epoch": 1.7902829028290284,
"grad_norm": 1.1640625,
"learning_rate": 7.238562511745768e-06,
"loss": 0.0877,
"step": 364
},
{
"epoch": 1.7952029520295203,
"grad_norm": 1.015625,
"learning_rate": 7.2245757681200835e-06,
"loss": 0.0871,
"step": 365
},
{
"epoch": 1.8001230012300122,
"grad_norm": 1.09375,
"learning_rate": 7.210567286750368e-06,
"loss": 0.0898,
"step": 366
},
{
"epoch": 1.8050430504305042,
"grad_norm": 1.140625,
"learning_rate": 7.196537204522401e-06,
"loss": 0.0826,
"step": 367
},
{
"epoch": 1.8099630996309963,
"grad_norm": 1.109375,
"learning_rate": 7.182485658533036e-06,
"loss": 0.0991,
"step": 368
},
{
"epoch": 1.8148831488314883,
"grad_norm": 1.0078125,
"learning_rate": 7.168412786088857e-06,
"loss": 0.0775,
"step": 369
},
{
"epoch": 1.8198031980319804,
"grad_norm": 0.8984375,
"learning_rate": 7.1543187247048525e-06,
"loss": 0.0915,
"step": 370
},
{
"epoch": 1.8247232472324724,
"grad_norm": 1.109375,
"learning_rate": 7.140203612103052e-06,
"loss": 0.0983,
"step": 371
},
{
"epoch": 1.8296432964329643,
"grad_norm": 0.89453125,
"learning_rate": 7.1260675862111986e-06,
"loss": 0.086,
"step": 372
},
{
"epoch": 1.8345633456334562,
"grad_norm": 1.34375,
"learning_rate": 7.111910785161381e-06,
"loss": 0.1281,
"step": 373
},
{
"epoch": 1.8394833948339482,
"grad_norm": 1.3203125,
"learning_rate": 7.0977333472887076e-06,
"loss": 0.0902,
"step": 374
},
{
"epoch": 1.8444034440344403,
"grad_norm": 1.234375,
"learning_rate": 7.083535411129934e-06,
"loss": 0.0996,
"step": 375
},
{
"epoch": 1.8493234932349325,
"grad_norm": 1.09375,
"learning_rate": 7.069317115422121e-06,
"loss": 0.0887,
"step": 376
},
{
"epoch": 1.8542435424354244,
"grad_norm": 1.15625,
"learning_rate": 7.055078599101275e-06,
"loss": 0.1046,
"step": 377
},
{
"epoch": 1.8591635916359164,
"grad_norm": 1.09375,
"learning_rate": 7.040820001300992e-06,
"loss": 0.1243,
"step": 378
},
{
"epoch": 1.8640836408364083,
"grad_norm": 1.1328125,
"learning_rate": 7.026541461351092e-06,
"loss": 0.0664,
"step": 379
},
{
"epoch": 1.8690036900369003,
"grad_norm": 1.0859375,
"learning_rate": 7.01224311877627e-06,
"loss": 0.1314,
"step": 380
},
{
"epoch": 1.8739237392373924,
"grad_norm": 0.82421875,
"learning_rate": 6.997925113294717e-06,
"loss": 0.0811,
"step": 381
},
{
"epoch": 1.8788437884378844,
"grad_norm": 0.9921875,
"learning_rate": 6.983587584816769e-06,
"loss": 0.0776,
"step": 382
},
{
"epoch": 1.8837638376383765,
"grad_norm": 1.359375,
"learning_rate": 6.969230673443531e-06,
"loss": 0.116,
"step": 383
},
{
"epoch": 1.8886838868388685,
"grad_norm": 0.9765625,
"learning_rate": 6.9548545194655115e-06,
"loss": 0.0983,
"step": 384
},
{
"epoch": 1.8936039360393604,
"grad_norm": 1.28125,
"learning_rate": 6.9404592633612486e-06,
"loss": 0.1323,
"step": 385
},
{
"epoch": 1.8985239852398523,
"grad_norm": 1.328125,
"learning_rate": 6.926045045795944e-06,
"loss": 0.1106,
"step": 386
},
{
"epoch": 1.9034440344403443,
"grad_norm": 1.3046875,
"learning_rate": 6.911612007620077e-06,
"loss": 0.0884,
"step": 387
},
{
"epoch": 1.9083640836408364,
"grad_norm": 0.953125,
"learning_rate": 6.897160289868042e-06,
"loss": 0.101,
"step": 388
},
{
"epoch": 1.9132841328413284,
"grad_norm": 0.81640625,
"learning_rate": 6.8826900337567595e-06,
"loss": 0.0907,
"step": 389
},
{
"epoch": 1.9182041820418205,
"grad_norm": 1.0078125,
"learning_rate": 6.8682013806842985e-06,
"loss": 0.083,
"step": 390
},
{
"epoch": 1.9231242312423125,
"grad_norm": 0.9765625,
"learning_rate": 6.853694472228504e-06,
"loss": 0.0821,
"step": 391
},
{
"epoch": 1.9280442804428044,
"grad_norm": 1.046875,
"learning_rate": 6.839169450145595e-06,
"loss": 0.0605,
"step": 392
},
{
"epoch": 1.9329643296432963,
"grad_norm": 1.0234375,
"learning_rate": 6.824626456368802e-06,
"loss": 0.1001,
"step": 393
},
{
"epoch": 1.9378843788437883,
"grad_norm": 1.203125,
"learning_rate": 6.810065633006956e-06,
"loss": 0.0936,
"step": 394
},
{
"epoch": 1.9428044280442804,
"grad_norm": 1.046875,
"learning_rate": 6.795487122343124e-06,
"loss": 0.0866,
"step": 395
},
{
"epoch": 1.9477244772447726,
"grad_norm": 1.3125,
"learning_rate": 6.7808910668331985e-06,
"loss": 0.1078,
"step": 396
},
{
"epoch": 1.9526445264452645,
"grad_norm": 1.109375,
"learning_rate": 6.766277609104518e-06,
"loss": 0.0854,
"step": 397
},
{
"epoch": 1.9575645756457565,
"grad_norm": 0.9453125,
"learning_rate": 6.751646891954466e-06,
"loss": 0.0848,
"step": 398
},
{
"epoch": 1.9624846248462484,
"grad_norm": 1.0234375,
"learning_rate": 6.736999058349085e-06,
"loss": 0.0861,
"step": 399
},
{
"epoch": 1.9674046740467404,
"grad_norm": 1.4296875,
"learning_rate": 6.722334251421665e-06,
"loss": 0.108,
"step": 400
},
{
"epoch": 1.9723247232472325,
"grad_norm": 0.76171875,
"learning_rate": 6.707652614471359e-06,
"loss": 0.0691,
"step": 401
},
{
"epoch": 1.9772447724477245,
"grad_norm": 1.4765625,
"learning_rate": 6.692954290961774e-06,
"loss": 0.1234,
"step": 402
},
{
"epoch": 1.9821648216482166,
"grad_norm": 0.9140625,
"learning_rate": 6.678239424519575e-06,
"loss": 0.0785,
"step": 403
},
{
"epoch": 1.9870848708487086,
"grad_norm": 1.328125,
"learning_rate": 6.6635081589330745e-06,
"loss": 0.096,
"step": 404
},
{
"epoch": 1.9920049200492005,
"grad_norm": 1.25,
"learning_rate": 6.648760638150833e-06,
"loss": 0.1109,
"step": 405
},
{
"epoch": 1.9969249692496924,
"grad_norm": 1.125,
"learning_rate": 6.6339970062802526e-06,
"loss": 0.0892,
"step": 406
},
{
"epoch": 2.0030750307503076,
"grad_norm": 0.83984375,
"learning_rate": 6.619217407586167e-06,
"loss": 0.065,
"step": 407
},
{
"epoch": 2.0079950799507995,
"grad_norm": 1.2578125,
"learning_rate": 6.604421986489428e-06,
"loss": 0.1044,
"step": 408
},
{
"epoch": 2.0079950799507995,
"eval_loss": 0.10599377751350403,
"eval_runtime": 203.102,
"eval_samples_per_second": 8.055,
"eval_steps_per_second": 1.344,
"step": 408
},
{
"epoch": 2.0129151291512914,
"grad_norm": 0.8828125,
"learning_rate": 6.589610887565503e-06,
"loss": 0.0784,
"step": 409
},
{
"epoch": 2.0178351783517834,
"grad_norm": 0.9609375,
"learning_rate": 6.574784255543052e-06,
"loss": 0.0788,
"step": 410
},
{
"epoch": 2.0227552275522753,
"grad_norm": 0.81640625,
"learning_rate": 6.559942235302527e-06,
"loss": 0.0876,
"step": 411
},
{
"epoch": 2.0276752767527677,
"grad_norm": 1.125,
"learning_rate": 6.545084971874738e-06,
"loss": 0.0702,
"step": 412
},
{
"epoch": 2.0325953259532596,
"grad_norm": 1.703125,
"learning_rate": 6.530212610439455e-06,
"loss": 0.1014,
"step": 413
},
{
"epoch": 2.0375153751537516,
"grad_norm": 0.87890625,
"learning_rate": 6.5153252963239736e-06,
"loss": 0.0846,
"step": 414
},
{
"epoch": 2.0424354243542435,
"grad_norm": 1.34375,
"learning_rate": 6.500423175001705e-06,
"loss": 0.1422,
"step": 415
},
{
"epoch": 2.0473554735547355,
"grad_norm": 1.1796875,
"learning_rate": 6.485506392090749e-06,
"loss": 0.1003,
"step": 416
},
{
"epoch": 2.0522755227552274,
"grad_norm": 1.1328125,
"learning_rate": 6.470575093352477e-06,
"loss": 0.1012,
"step": 417
},
{
"epoch": 2.0571955719557193,
"grad_norm": 1.1484375,
"learning_rate": 6.4556294246900994e-06,
"loss": 0.0778,
"step": 418
},
{
"epoch": 2.0621156211562117,
"grad_norm": 1.109375,
"learning_rate": 6.440669532147246e-06,
"loss": 0.0975,
"step": 419
},
{
"epoch": 2.0670356703567037,
"grad_norm": 0.91015625,
"learning_rate": 6.4256955619065375e-06,
"loss": 0.0772,
"step": 420
},
{
"epoch": 2.0719557195571956,
"grad_norm": 1.078125,
"learning_rate": 6.410707660288154e-06,
"loss": 0.1044,
"step": 421
},
{
"epoch": 2.0768757687576875,
"grad_norm": 1.1796875,
"learning_rate": 6.395705973748414e-06,
"loss": 0.0868,
"step": 422
},
{
"epoch": 2.0817958179581795,
"grad_norm": 1.234375,
"learning_rate": 6.3806906488783305e-06,
"loss": 0.1051,
"step": 423
},
{
"epoch": 2.0867158671586714,
"grad_norm": 1.1171875,
"learning_rate": 6.36566183240219e-06,
"loss": 0.0572,
"step": 424
},
{
"epoch": 2.091635916359164,
"grad_norm": 1.078125,
"learning_rate": 6.350619671176111e-06,
"loss": 0.1055,
"step": 425
},
{
"epoch": 2.0965559655596557,
"grad_norm": 0.9140625,
"learning_rate": 6.335564312186615e-06,
"loss": 0.0956,
"step": 426
},
{
"epoch": 2.1014760147601477,
"grad_norm": 1.75,
"learning_rate": 6.3204959025491844e-06,
"loss": 0.0727,
"step": 427
},
{
"epoch": 2.1063960639606396,
"grad_norm": 1.1015625,
"learning_rate": 6.30541458950683e-06,
"loss": 0.1014,
"step": 428
},
{
"epoch": 2.1113161131611315,
"grad_norm": 1.1953125,
"learning_rate": 6.2903205204286474e-06,
"loss": 0.0968,
"step": 429
},
{
"epoch": 2.1162361623616235,
"grad_norm": 1.21875,
"learning_rate": 6.275213842808383e-06,
"loss": 0.0856,
"step": 430
},
{
"epoch": 2.1211562115621154,
"grad_norm": 1.359375,
"learning_rate": 6.260094704262986e-06,
"loss": 0.0773,
"step": 431
},
{
"epoch": 2.126076260762608,
"grad_norm": 1.125,
"learning_rate": 6.244963252531171e-06,
"loss": 0.0802,
"step": 432
},
{
"epoch": 2.1309963099630997,
"grad_norm": 1.21875,
"learning_rate": 6.229819635471972e-06,
"loss": 0.0809,
"step": 433
},
{
"epoch": 2.1359163591635917,
"grad_norm": 1.1171875,
"learning_rate": 6.2146640010633e-06,
"loss": 0.0833,
"step": 434
},
{
"epoch": 2.1408364083640836,
"grad_norm": 0.984375,
"learning_rate": 6.19949649740049e-06,
"loss": 0.0784,
"step": 435
},
{
"epoch": 2.1457564575645756,
"grad_norm": 0.9453125,
"learning_rate": 6.184317272694866e-06,
"loss": 0.0885,
"step": 436
},
{
"epoch": 2.1506765067650675,
"grad_norm": 1.140625,
"learning_rate": 6.16912647527228e-06,
"loss": 0.0763,
"step": 437
},
{
"epoch": 2.15559655596556,
"grad_norm": 1.1484375,
"learning_rate": 6.1539242535716704e-06,
"loss": 0.0805,
"step": 438
},
{
"epoch": 2.160516605166052,
"grad_norm": 1.2421875,
"learning_rate": 6.138710756143613e-06,
"loss": 0.1051,
"step": 439
},
{
"epoch": 2.1654366543665438,
"grad_norm": 0.96875,
"learning_rate": 6.123486131648859e-06,
"loss": 0.0741,
"step": 440
},
{
"epoch": 2.1703567035670357,
"grad_norm": 1.171875,
"learning_rate": 6.108250528856895e-06,
"loss": 0.0725,
"step": 441
},
{
"epoch": 2.1752767527675276,
"grad_norm": 1.0859375,
"learning_rate": 6.0930040966444815e-06,
"loss": 0.0957,
"step": 442
},
{
"epoch": 2.1801968019680196,
"grad_norm": 1.421875,
"learning_rate": 6.077746983994198e-06,
"loss": 0.1151,
"step": 443
},
{
"epoch": 2.1851168511685115,
"grad_norm": 1.1328125,
"learning_rate": 6.062479339992993e-06,
"loss": 0.0939,
"step": 444
},
{
"epoch": 2.190036900369004,
"grad_norm": 1.0234375,
"learning_rate": 6.047201313830724e-06,
"loss": 0.0944,
"step": 445
},
{
"epoch": 2.194956949569496,
"grad_norm": 1.0859375,
"learning_rate": 6.031913054798692e-06,
"loss": 0.0825,
"step": 446
},
{
"epoch": 2.1998769987699878,
"grad_norm": 0.94921875,
"learning_rate": 6.016614712288198e-06,
"loss": 0.0732,
"step": 447
},
{
"epoch": 2.2047970479704797,
"grad_norm": 1.2421875,
"learning_rate": 6.0013064357890715e-06,
"loss": 0.0749,
"step": 448
},
{
"epoch": 2.2097170971709716,
"grad_norm": 1.3828125,
"learning_rate": 5.985988374888216e-06,
"loss": 0.0935,
"step": 449
},
{
"epoch": 2.2146371463714636,
"grad_norm": 1.0625,
"learning_rate": 5.970660679268139e-06,
"loss": 0.0762,
"step": 450
},
{
"epoch": 2.2195571955719555,
"grad_norm": 1.3671875,
"learning_rate": 5.955323498705501e-06,
"loss": 0.0929,
"step": 451
},
{
"epoch": 2.224477244772448,
"grad_norm": 1.3046875,
"learning_rate": 5.9399769830696404e-06,
"loss": 0.1067,
"step": 452
},
{
"epoch": 2.22939729397294,
"grad_norm": 0.9609375,
"learning_rate": 5.924621282321123e-06,
"loss": 0.0736,
"step": 453
},
{
"epoch": 2.234317343173432,
"grad_norm": 1.0078125,
"learning_rate": 5.909256546510257e-06,
"loss": 0.0699,
"step": 454
},
{
"epoch": 2.2392373923739237,
"grad_norm": 1.0,
"learning_rate": 5.893882925775648e-06,
"loss": 0.0792,
"step": 455
},
{
"epoch": 2.2441574415744157,
"grad_norm": 1.15625,
"learning_rate": 5.878500570342714e-06,
"loss": 0.0874,
"step": 456
},
{
"epoch": 2.2490774907749076,
"grad_norm": 1.296875,
"learning_rate": 5.86310963052223e-06,
"loss": 0.1197,
"step": 457
},
{
"epoch": 2.2539975399753995,
"grad_norm": 1.21875,
"learning_rate": 5.847710256708854e-06,
"loss": 0.1147,
"step": 458
},
{
"epoch": 2.258917589175892,
"grad_norm": 0.9453125,
"learning_rate": 5.832302599379657e-06,
"loss": 0.0647,
"step": 459
},
{
"epoch": 2.263837638376384,
"grad_norm": 0.86328125,
"learning_rate": 5.816886809092651e-06,
"loss": 0.0794,
"step": 460
},
{
"epoch": 2.268757687576876,
"grad_norm": 1.1171875,
"learning_rate": 5.8014630364853275e-06,
"loss": 0.0787,
"step": 461
},
{
"epoch": 2.2736777367773677,
"grad_norm": 1.375,
"learning_rate": 5.7860314322731705e-06,
"loss": 0.1395,
"step": 462
},
{
"epoch": 2.2785977859778597,
"grad_norm": 1.6640625,
"learning_rate": 5.770592147248197e-06,
"loss": 0.083,
"step": 463
},
{
"epoch": 2.2835178351783516,
"grad_norm": 1.390625,
"learning_rate": 5.755145332277472e-06,
"loss": 0.1035,
"step": 464
},
{
"epoch": 2.288437884378844,
"grad_norm": 1.2265625,
"learning_rate": 5.73969113830165e-06,
"loss": 0.1229,
"step": 465
},
{
"epoch": 2.293357933579336,
"grad_norm": 0.96875,
"learning_rate": 5.72422971633348e-06,
"loss": 0.0847,
"step": 466
},
{
"epoch": 2.298277982779828,
"grad_norm": 1.2734375,
"learning_rate": 5.70876121745635e-06,
"loss": 0.0944,
"step": 467
},
{
"epoch": 2.30319803198032,
"grad_norm": 0.8828125,
"learning_rate": 5.69328579282279e-06,
"loss": 0.0729,
"step": 468
},
{
"epoch": 2.3081180811808117,
"grad_norm": 1.2890625,
"learning_rate": 5.677803593653018e-06,
"loss": 0.076,
"step": 469
},
{
"epoch": 2.3130381303813037,
"grad_norm": 1.3515625,
"learning_rate": 5.66231477123344e-06,
"loss": 0.0779,
"step": 470
},
{
"epoch": 2.3179581795817956,
"grad_norm": 1.09375,
"learning_rate": 5.646819476915189e-06,
"loss": 0.0596,
"step": 471
},
{
"epoch": 2.322878228782288,
"grad_norm": 0.98046875,
"learning_rate": 5.631317862112636e-06,
"loss": 0.1077,
"step": 472
},
{
"epoch": 2.32779827798278,
"grad_norm": 1.3125,
"learning_rate": 5.615810078301912e-06,
"loss": 0.1024,
"step": 473
},
{
"epoch": 2.332718327183272,
"grad_norm": 1.3984375,
"learning_rate": 5.600296277019434e-06,
"loss": 0.1357,
"step": 474
},
{
"epoch": 2.337638376383764,
"grad_norm": 1.09375,
"learning_rate": 5.584776609860414e-06,
"loss": 0.0891,
"step": 475
},
{
"epoch": 2.3425584255842558,
"grad_norm": 1.1328125,
"learning_rate": 5.569251228477386e-06,
"loss": 0.0744,
"step": 476
},
{
"epoch": 2.3474784747847477,
"grad_norm": 0.984375,
"learning_rate": 5.553720284578723e-06,
"loss": 0.0669,
"step": 477
},
{
"epoch": 2.35239852398524,
"grad_norm": 1.015625,
"learning_rate": 5.538183929927152e-06,
"loss": 0.0819,
"step": 478
},
{
"epoch": 2.357318573185732,
"grad_norm": 1.140625,
"learning_rate": 5.522642316338268e-06,
"loss": 0.0794,
"step": 479
},
{
"epoch": 2.362238622386224,
"grad_norm": 1.03125,
"learning_rate": 5.507095595679059e-06,
"loss": 0.0887,
"step": 480
},
{
"epoch": 2.367158671586716,
"grad_norm": 1.4296875,
"learning_rate": 5.491543919866417e-06,
"loss": 0.0863,
"step": 481
},
{
"epoch": 2.372078720787208,
"grad_norm": 0.9140625,
"learning_rate": 5.47598744086565e-06,
"loss": 0.0844,
"step": 482
},
{
"epoch": 2.3769987699876998,
"grad_norm": 1.0625,
"learning_rate": 5.460426310689006e-06,
"loss": 0.0712,
"step": 483
},
{
"epoch": 2.3819188191881917,
"grad_norm": 1.359375,
"learning_rate": 5.4448606813941805e-06,
"loss": 0.0768,
"step": 484
},
{
"epoch": 2.3868388683886836,
"grad_norm": 1.1015625,
"learning_rate": 5.42929070508283e-06,
"loss": 0.079,
"step": 485
},
{
"epoch": 2.391758917589176,
"grad_norm": 0.96875,
"learning_rate": 5.413716533899096e-06,
"loss": 0.0853,
"step": 486
},
{
"epoch": 2.396678966789668,
"grad_norm": 1.1171875,
"learning_rate": 5.3981383200281004e-06,
"loss": 0.1153,
"step": 487
},
{
"epoch": 2.40159901599016,
"grad_norm": 1.1171875,
"learning_rate": 5.382556215694478e-06,
"loss": 0.0533,
"step": 488
},
{
"epoch": 2.406519065190652,
"grad_norm": 1.234375,
"learning_rate": 5.366970373160873e-06,
"loss": 0.0905,
"step": 489
},
{
"epoch": 2.411439114391144,
"grad_norm": 1.03125,
"learning_rate": 5.351380944726465e-06,
"loss": 0.0787,
"step": 490
},
{
"epoch": 2.416359163591636,
"grad_norm": 1.21875,
"learning_rate": 5.335788082725467e-06,
"loss": 0.0781,
"step": 491
},
{
"epoch": 2.421279212792128,
"grad_norm": 0.90234375,
"learning_rate": 5.3201919395256475e-06,
"loss": 0.0689,
"step": 492
},
{
"epoch": 2.42619926199262,
"grad_norm": 1.2265625,
"learning_rate": 5.304592667526835e-06,
"loss": 0.0849,
"step": 493
},
{
"epoch": 2.431119311193112,
"grad_norm": 1.109375,
"learning_rate": 5.288990419159433e-06,
"loss": 0.1141,
"step": 494
},
{
"epoch": 2.436039360393604,
"grad_norm": 0.85546875,
"learning_rate": 5.2733853468829295e-06,
"loss": 0.0842,
"step": 495
},
{
"epoch": 2.440959409594096,
"grad_norm": 1.1796875,
"learning_rate": 5.257777603184408e-06,
"loss": 0.0599,
"step": 496
},
{
"epoch": 2.445879458794588,
"grad_norm": 0.859375,
"learning_rate": 5.24216734057705e-06,
"loss": 0.0906,
"step": 497
},
{
"epoch": 2.4507995079950797,
"grad_norm": 1.84375,
"learning_rate": 5.226554711598659e-06,
"loss": 0.0934,
"step": 498
},
{
"epoch": 2.455719557195572,
"grad_norm": 0.75390625,
"learning_rate": 5.210939868810156e-06,
"loss": 0.0783,
"step": 499
},
{
"epoch": 2.460639606396064,
"grad_norm": 1.078125,
"learning_rate": 5.195322964794098e-06,
"loss": 0.0944,
"step": 500
},
{
"epoch": 2.465559655596556,
"grad_norm": 1.0234375,
"learning_rate": 5.1797041521531795e-06,
"loss": 0.0831,
"step": 501
},
{
"epoch": 2.470479704797048,
"grad_norm": 1.3828125,
"learning_rate": 5.16408358350875e-06,
"loss": 0.1166,
"step": 502
},
{
"epoch": 2.47539975399754,
"grad_norm": 1.0625,
"learning_rate": 5.1484614114993156e-06,
"loss": 0.089,
"step": 503
},
{
"epoch": 2.480319803198032,
"grad_norm": 1.546875,
"learning_rate": 5.132837788779049e-06,
"loss": 0.1219,
"step": 504
},
{
"epoch": 2.485239852398524,
"grad_norm": 0.95703125,
"learning_rate": 5.117212868016303e-06,
"loss": 0.0802,
"step": 505
},
{
"epoch": 2.490159901599016,
"grad_norm": 0.90625,
"learning_rate": 5.101586801892109e-06,
"loss": 0.0862,
"step": 506
},
{
"epoch": 2.495079950799508,
"grad_norm": 1.46875,
"learning_rate": 5.085959743098693e-06,
"loss": 0.0834,
"step": 507
},
{
"epoch": 2.5,
"grad_norm": 1.125,
"learning_rate": 5.07033184433798e-06,
"loss": 0.0764,
"step": 508
},
{
"epoch": 2.504920049200492,
"grad_norm": 1.21875,
"learning_rate": 5.054703258320107e-06,
"loss": 0.0789,
"step": 509
},
{
"epoch": 2.509840098400984,
"grad_norm": 1.0390625,
"learning_rate": 5.03907413776192e-06,
"loss": 0.0909,
"step": 510
},
{
"epoch": 2.509840098400984,
"eval_loss": 0.10507026314735413,
"eval_runtime": 204.2755,
"eval_samples_per_second": 8.009,
"eval_steps_per_second": 1.336,
"step": 510
},
{
"epoch": 2.514760147601476,
"grad_norm": 0.88671875,
"learning_rate": 5.0234446353854934e-06,
"loss": 0.0751,
"step": 511
},
{
"epoch": 2.5196801968019678,
"grad_norm": 1.1796875,
"learning_rate": 5.00781490391663e-06,
"loss": 0.1011,
"step": 512
},
{
"epoch": 2.52460024600246,
"grad_norm": 1.2890625,
"learning_rate": 4.992185096083372e-06,
"loss": 0.1007,
"step": 513
},
{
"epoch": 2.529520295202952,
"grad_norm": 1.0625,
"learning_rate": 4.976555364614509e-06,
"loss": 0.0879,
"step": 514
},
{
"epoch": 2.534440344403444,
"grad_norm": 1.3203125,
"learning_rate": 4.96092586223808e-06,
"loss": 0.0806,
"step": 515
},
{
"epoch": 2.539360393603936,
"grad_norm": 1.1875,
"learning_rate": 4.9452967416798945e-06,
"loss": 0.0916,
"step": 516
},
{
"epoch": 2.544280442804428,
"grad_norm": 1.4296875,
"learning_rate": 4.929668155662021e-06,
"loss": 0.0825,
"step": 517
},
{
"epoch": 2.5492004920049203,
"grad_norm": 1.1953125,
"learning_rate": 4.914040256901309e-06,
"loss": 0.0761,
"step": 518
},
{
"epoch": 2.554120541205412,
"grad_norm": 1.1015625,
"learning_rate": 4.898413198107892e-06,
"loss": 0.0657,
"step": 519
},
{
"epoch": 2.559040590405904,
"grad_norm": 1.625,
"learning_rate": 4.882787131983698e-06,
"loss": 0.1164,
"step": 520
},
{
"epoch": 2.563960639606396,
"grad_norm": 1.0078125,
"learning_rate": 4.867162211220952e-06,
"loss": 0.0794,
"step": 521
},
{
"epoch": 2.568880688806888,
"grad_norm": 1.0859375,
"learning_rate": 4.851538588500687e-06,
"loss": 0.0943,
"step": 522
},
{
"epoch": 2.57380073800738,
"grad_norm": 1.1953125,
"learning_rate": 4.835916416491251e-06,
"loss": 0.0911,
"step": 523
},
{
"epoch": 2.578720787207872,
"grad_norm": 0.9140625,
"learning_rate": 4.820295847846822e-06,
"loss": 0.0579,
"step": 524
},
{
"epoch": 2.583640836408364,
"grad_norm": 1.125,
"learning_rate": 4.804677035205903e-06,
"loss": 0.0955,
"step": 525
},
{
"epoch": 2.588560885608856,
"grad_norm": 1.484375,
"learning_rate": 4.789060131189845e-06,
"loss": 0.0879,
"step": 526
},
{
"epoch": 2.593480934809348,
"grad_norm": 1.140625,
"learning_rate": 4.773445288401343e-06,
"loss": 0.089,
"step": 527
},
{
"epoch": 2.59840098400984,
"grad_norm": 0.9375,
"learning_rate": 4.75783265942295e-06,
"loss": 0.0871,
"step": 528
},
{
"epoch": 2.603321033210332,
"grad_norm": 1.1640625,
"learning_rate": 4.742222396815593e-06,
"loss": 0.1021,
"step": 529
},
{
"epoch": 2.608241082410824,
"grad_norm": 1.0859375,
"learning_rate": 4.726614653117071e-06,
"loss": 0.0909,
"step": 530
},
{
"epoch": 2.6131611316113164,
"grad_norm": 1.234375,
"learning_rate": 4.711009580840569e-06,
"loss": 0.1082,
"step": 531
},
{
"epoch": 2.6180811808118083,
"grad_norm": 1.203125,
"learning_rate": 4.695407332473166e-06,
"loss": 0.083,
"step": 532
},
{
"epoch": 2.6230012300123002,
"grad_norm": 1.0859375,
"learning_rate": 4.679808060474354e-06,
"loss": 0.0745,
"step": 533
},
{
"epoch": 2.627921279212792,
"grad_norm": 1.0703125,
"learning_rate": 4.6642119172745345e-06,
"loss": 0.0776,
"step": 534
},
{
"epoch": 2.632841328413284,
"grad_norm": 0.9453125,
"learning_rate": 4.6486190552735375e-06,
"loss": 0.0676,
"step": 535
},
{
"epoch": 2.637761377613776,
"grad_norm": 1.0703125,
"learning_rate": 4.633029626839128e-06,
"loss": 0.0794,
"step": 536
},
{
"epoch": 2.642681426814268,
"grad_norm": 1.0859375,
"learning_rate": 4.617443784305524e-06,
"loss": 0.1136,
"step": 537
},
{
"epoch": 2.64760147601476,
"grad_norm": 1.1875,
"learning_rate": 4.601861679971901e-06,
"loss": 0.0747,
"step": 538
},
{
"epoch": 2.6525215252152523,
"grad_norm": 1.625,
"learning_rate": 4.5862834661009074e-06,
"loss": 0.0982,
"step": 539
},
{
"epoch": 2.6574415744157442,
"grad_norm": 1.046875,
"learning_rate": 4.57070929491717e-06,
"loss": 0.0788,
"step": 540
},
{
"epoch": 2.662361623616236,
"grad_norm": 1.4765625,
"learning_rate": 4.555139318605821e-06,
"loss": 0.0857,
"step": 541
},
{
"epoch": 2.667281672816728,
"grad_norm": 1.1796875,
"learning_rate": 4.539573689310995e-06,
"loss": 0.09,
"step": 542
},
{
"epoch": 2.67220172201722,
"grad_norm": 1.4609375,
"learning_rate": 4.524012559134352e-06,
"loss": 0.0934,
"step": 543
},
{
"epoch": 2.6771217712177124,
"grad_norm": 1.0234375,
"learning_rate": 4.508456080133584e-06,
"loss": 0.0742,
"step": 544
},
{
"epoch": 2.6820418204182044,
"grad_norm": 1.046875,
"learning_rate": 4.492904404320942e-06,
"loss": 0.0677,
"step": 545
},
{
"epoch": 2.6869618696186963,
"grad_norm": 1.078125,
"learning_rate": 4.477357683661734e-06,
"loss": 0.072,
"step": 546
},
{
"epoch": 2.6918819188191883,
"grad_norm": 1.2578125,
"learning_rate": 4.461816070072851e-06,
"loss": 0.0948,
"step": 547
},
{
"epoch": 2.69680196801968,
"grad_norm": 1.0625,
"learning_rate": 4.446279715421277e-06,
"loss": 0.0846,
"step": 548
},
{
"epoch": 2.701722017220172,
"grad_norm": 1.3828125,
"learning_rate": 4.430748771522615e-06,
"loss": 0.0952,
"step": 549
},
{
"epoch": 2.706642066420664,
"grad_norm": 1.09375,
"learning_rate": 4.415223390139588e-06,
"loss": 0.0797,
"step": 550
},
{
"epoch": 2.711562115621156,
"grad_norm": 1.0078125,
"learning_rate": 4.399703722980569e-06,
"loss": 0.0822,
"step": 551
},
{
"epoch": 2.716482164821648,
"grad_norm": 1.328125,
"learning_rate": 4.3841899216980895e-06,
"loss": 0.1041,
"step": 552
},
{
"epoch": 2.7214022140221403,
"grad_norm": 1.21875,
"learning_rate": 4.368682137887365e-06,
"loss": 0.0884,
"step": 553
},
{
"epoch": 2.7263222632226323,
"grad_norm": 1.0546875,
"learning_rate": 4.3531805230848116e-06,
"loss": 0.0799,
"step": 554
},
{
"epoch": 2.731242312423124,
"grad_norm": 1.375,
"learning_rate": 4.337685228766561e-06,
"loss": 0.1055,
"step": 555
},
{
"epoch": 2.736162361623616,
"grad_norm": 1.15625,
"learning_rate": 4.322196406346984e-06,
"loss": 0.1206,
"step": 556
},
{
"epoch": 2.741082410824108,
"grad_norm": 1.2265625,
"learning_rate": 4.30671420717721e-06,
"loss": 0.085,
"step": 557
},
{
"epoch": 2.7460024600246005,
"grad_norm": 1.4765625,
"learning_rate": 4.291238782543652e-06,
"loss": 0.1146,
"step": 558
},
{
"epoch": 2.7509225092250924,
"grad_norm": 1.125,
"learning_rate": 4.275770283666521e-06,
"loss": 0.0742,
"step": 559
},
{
"epoch": 2.7558425584255843,
"grad_norm": 1.2109375,
"learning_rate": 4.260308861698351e-06,
"loss": 0.091,
"step": 560
},
{
"epoch": 2.7607626076260763,
"grad_norm": 1.171875,
"learning_rate": 4.244854667722527e-06,
"loss": 0.0741,
"step": 561
},
{
"epoch": 2.765682656826568,
"grad_norm": 1.1328125,
"learning_rate": 4.229407852751806e-06,
"loss": 0.0738,
"step": 562
},
{
"epoch": 2.77060270602706,
"grad_norm": 1.3359375,
"learning_rate": 4.21396856772683e-06,
"loss": 0.0999,
"step": 563
},
{
"epoch": 2.775522755227552,
"grad_norm": 1.0703125,
"learning_rate": 4.198536963514674e-06,
"loss": 0.0619,
"step": 564
},
{
"epoch": 2.780442804428044,
"grad_norm": 1.0390625,
"learning_rate": 4.183113190907349e-06,
"loss": 0.0841,
"step": 565
},
{
"epoch": 2.7853628536285364,
"grad_norm": 0.96484375,
"learning_rate": 4.1676974006203456e-06,
"loss": 0.0738,
"step": 566
},
{
"epoch": 2.7902829028290284,
"grad_norm": 1.21875,
"learning_rate": 4.152289743291148e-06,
"loss": 0.1018,
"step": 567
},
{
"epoch": 2.7952029520295203,
"grad_norm": 1.0234375,
"learning_rate": 4.136890369477773e-06,
"loss": 0.0664,
"step": 568
},
{
"epoch": 2.8001230012300122,
"grad_norm": 1.09375,
"learning_rate": 4.121499429657287e-06,
"loss": 0.0769,
"step": 569
},
{
"epoch": 2.805043050430504,
"grad_norm": 1.0234375,
"learning_rate": 4.106117074224354e-06,
"loss": 0.0779,
"step": 570
},
{
"epoch": 2.8099630996309966,
"grad_norm": 1.203125,
"learning_rate": 4.090743453489744e-06,
"loss": 0.0958,
"step": 571
},
{
"epoch": 2.8148831488314885,
"grad_norm": 1.4609375,
"learning_rate": 4.07537871767888e-06,
"loss": 0.1121,
"step": 572
},
{
"epoch": 2.8198031980319804,
"grad_norm": 1.4375,
"learning_rate": 4.060023016930359e-06,
"loss": 0.1435,
"step": 573
},
{
"epoch": 2.8247232472324724,
"grad_norm": 1.2265625,
"learning_rate": 4.044676501294501e-06,
"loss": 0.1366,
"step": 574
},
{
"epoch": 2.8296432964329643,
"grad_norm": 1.171875,
"learning_rate": 4.029339320731862e-06,
"loss": 0.0755,
"step": 575
},
{
"epoch": 2.8345633456334562,
"grad_norm": 1.0625,
"learning_rate": 4.0140116251117865e-06,
"loss": 0.0837,
"step": 576
},
{
"epoch": 2.839483394833948,
"grad_norm": 1.25,
"learning_rate": 3.998693564210929e-06,
"loss": 0.1097,
"step": 577
},
{
"epoch": 2.84440344403444,
"grad_norm": 1.0078125,
"learning_rate": 3.983385287711803e-06,
"loss": 0.0652,
"step": 578
},
{
"epoch": 2.8493234932349325,
"grad_norm": 0.94921875,
"learning_rate": 3.96808694520131e-06,
"loss": 0.0785,
"step": 579
},
{
"epoch": 2.8542435424354244,
"grad_norm": 1.3125,
"learning_rate": 3.952798686169279e-06,
"loss": 0.0947,
"step": 580
},
{
"epoch": 2.8591635916359164,
"grad_norm": 1.1796875,
"learning_rate": 3.937520660007008e-06,
"loss": 0.0803,
"step": 581
},
{
"epoch": 2.8640836408364083,
"grad_norm": 0.8203125,
"learning_rate": 3.9222530160058025e-06,
"loss": 0.0664,
"step": 582
},
{
"epoch": 2.8690036900369003,
"grad_norm": 1.0703125,
"learning_rate": 3.90699590335552e-06,
"loss": 0.0737,
"step": 583
},
{
"epoch": 2.8739237392373926,
"grad_norm": 1.6875,
"learning_rate": 3.891749471143106e-06,
"loss": 0.1152,
"step": 584
},
{
"epoch": 2.8788437884378846,
"grad_norm": 1.0,
"learning_rate": 3.876513868351142e-06,
"loss": 0.0932,
"step": 585
},
{
"epoch": 2.8837638376383765,
"grad_norm": 1.6015625,
"learning_rate": 3.861289243856388e-06,
"loss": 0.1052,
"step": 586
},
{
"epoch": 2.8886838868388685,
"grad_norm": 1.5234375,
"learning_rate": 3.84607574642833e-06,
"loss": 0.1152,
"step": 587
},
{
"epoch": 2.8936039360393604,
"grad_norm": 1.1796875,
"learning_rate": 3.830873524727722e-06,
"loss": 0.083,
"step": 588
},
{
"epoch": 2.8985239852398523,
"grad_norm": 1.2109375,
"learning_rate": 3.815682727305136e-06,
"loss": 0.0824,
"step": 589
},
{
"epoch": 2.9034440344403443,
"grad_norm": 1.171875,
"learning_rate": 3.800503502599511e-06,
"loss": 0.0933,
"step": 590
},
{
"epoch": 2.908364083640836,
"grad_norm": 0.91015625,
"learning_rate": 3.7853359989367023e-06,
"loss": 0.0831,
"step": 591
},
{
"epoch": 2.913284132841328,
"grad_norm": 1.1484375,
"learning_rate": 3.7701803645280296e-06,
"loss": 0.0954,
"step": 592
},
{
"epoch": 2.9182041820418205,
"grad_norm": 1.046875,
"learning_rate": 3.7550367474688315e-06,
"loss": 0.0829,
"step": 593
},
{
"epoch": 2.9231242312423125,
"grad_norm": 1.703125,
"learning_rate": 3.739905295737015e-06,
"loss": 0.0905,
"step": 594
},
{
"epoch": 2.9280442804428044,
"grad_norm": 1.3125,
"learning_rate": 3.7247861571916183e-06,
"loss": 0.0838,
"step": 595
},
{
"epoch": 2.9329643296432963,
"grad_norm": 1.2109375,
"learning_rate": 3.7096794795713542e-06,
"loss": 0.1098,
"step": 596
},
{
"epoch": 2.9378843788437883,
"grad_norm": 1.140625,
"learning_rate": 3.6945854104931726e-06,
"loss": 0.0776,
"step": 597
},
{
"epoch": 2.9428044280442807,
"grad_norm": 1.203125,
"learning_rate": 3.6795040974508164e-06,
"loss": 0.0694,
"step": 598
},
{
"epoch": 2.9477244772447726,
"grad_norm": 1.2890625,
"learning_rate": 3.6644356878133862e-06,
"loss": 0.0996,
"step": 599
},
{
"epoch": 2.9526445264452645,
"grad_norm": 1.1953125,
"learning_rate": 3.6493803288238894e-06,
"loss": 0.0801,
"step": 600
},
{
"epoch": 2.9575645756457565,
"grad_norm": 1.265625,
"learning_rate": 3.634338167597812e-06,
"loss": 0.0904,
"step": 601
},
{
"epoch": 2.9624846248462484,
"grad_norm": 1.3125,
"learning_rate": 3.61930935112167e-06,
"loss": 0.0676,
"step": 602
},
{
"epoch": 2.9674046740467404,
"grad_norm": 1.265625,
"learning_rate": 3.6042940262515867e-06,
"loss": 0.0985,
"step": 603
},
{
"epoch": 2.9723247232472323,
"grad_norm": 1.109375,
"learning_rate": 3.5892923397118473e-06,
"loss": 0.0816,
"step": 604
},
{
"epoch": 2.9772447724477242,
"grad_norm": 1.1328125,
"learning_rate": 3.5743044380934655e-06,
"loss": 0.076,
"step": 605
},
{
"epoch": 2.9821648216482166,
"grad_norm": 0.97265625,
"learning_rate": 3.5593304678527547e-06,
"loss": 0.0932,
"step": 606
},
{
"epoch": 2.9870848708487086,
"grad_norm": 1.140625,
"learning_rate": 3.544370575309902e-06,
"loss": 0.0982,
"step": 607
},
{
"epoch": 2.9920049200492005,
"grad_norm": 1.15625,
"learning_rate": 3.5294249066475245e-06,
"loss": 0.0997,
"step": 608
},
{
"epoch": 2.9969249692496924,
"grad_norm": 1.1953125,
"learning_rate": 3.5144936079092528e-06,
"loss": 0.0907,
"step": 609
},
{
"epoch": 3.002460024600246,
"grad_norm": 7.90625,
"learning_rate": 3.4995768249982975e-06,
"loss": 0.1999,
"step": 610
},
{
"epoch": 3.007380073800738,
"grad_norm": 1.0546875,
"learning_rate": 3.4846747036760285e-06,
"loss": 0.0815,
"step": 611
},
{
"epoch": 3.01230012300123,
"grad_norm": 1.0234375,
"learning_rate": 3.4697873895605466e-06,
"loss": 0.0754,
"step": 612
},
{
"epoch": 3.01230012300123,
"eval_loss": 0.1035788357257843,
"eval_runtime": 204.6775,
"eval_samples_per_second": 7.993,
"eval_steps_per_second": 1.334,
"step": 612
},
{
"epoch": 3.0172201722017222,
"grad_norm": 1.1796875,
"learning_rate": 3.4549150281252635e-06,
"loss": 0.1022,
"step": 613
},
{
"epoch": 3.022140221402214,
"grad_norm": 0.953125,
"learning_rate": 3.4400577646974766e-06,
"loss": 0.074,
"step": 614
},
{
"epoch": 3.027060270602706,
"grad_norm": 0.98046875,
"learning_rate": 3.4252157444569478e-06,
"loss": 0.0619,
"step": 615
},
{
"epoch": 3.031980319803198,
"grad_norm": 1.2265625,
"learning_rate": 3.410389112434499e-06,
"loss": 0.0887,
"step": 616
},
{
"epoch": 3.03690036900369,
"grad_norm": 0.87109375,
"learning_rate": 3.3955780135105736e-06,
"loss": 0.0831,
"step": 617
},
{
"epoch": 3.041820418204182,
"grad_norm": 1.0625,
"learning_rate": 3.3807825924138356e-06,
"loss": 0.1026,
"step": 618
},
{
"epoch": 3.046740467404674,
"grad_norm": 1.1953125,
"learning_rate": 3.366002993719747e-06,
"loss": 0.0814,
"step": 619
},
{
"epoch": 3.0516605166051662,
"grad_norm": 1.1015625,
"learning_rate": 3.351239361849168e-06,
"loss": 0.1059,
"step": 620
},
{
"epoch": 3.056580565805658,
"grad_norm": 0.82421875,
"learning_rate": 3.336491841066928e-06,
"loss": 0.083,
"step": 621
},
{
"epoch": 3.06150061500615,
"grad_norm": 1.109375,
"learning_rate": 3.3217605754804273e-06,
"loss": 0.0612,
"step": 622
},
{
"epoch": 3.066420664206642,
"grad_norm": 1.1015625,
"learning_rate": 3.307045709038226e-06,
"loss": 0.0833,
"step": 623
},
{
"epoch": 3.071340713407134,
"grad_norm": 1.0078125,
"learning_rate": 3.2923473855286426e-06,
"loss": 0.0717,
"step": 624
},
{
"epoch": 3.076260762607626,
"grad_norm": 0.8671875,
"learning_rate": 3.2776657485783357e-06,
"loss": 0.0752,
"step": 625
},
{
"epoch": 3.081180811808118,
"grad_norm": 1.09375,
"learning_rate": 3.2630009416509167e-06,
"loss": 0.0891,
"step": 626
},
{
"epoch": 3.0861008610086103,
"grad_norm": 1.2109375,
"learning_rate": 3.2483531080455334e-06,
"loss": 0.1059,
"step": 627
},
{
"epoch": 3.091020910209102,
"grad_norm": 1.015625,
"learning_rate": 3.2337223908954834e-06,
"loss": 0.0874,
"step": 628
},
{
"epoch": 3.095940959409594,
"grad_norm": 1.0078125,
"learning_rate": 3.2191089331668036e-06,
"loss": 0.0814,
"step": 629
},
{
"epoch": 3.100861008610086,
"grad_norm": 1.1875,
"learning_rate": 3.2045128776568783e-06,
"loss": 0.0982,
"step": 630
},
{
"epoch": 3.105781057810578,
"grad_norm": 0.7265625,
"learning_rate": 3.1899343669930446e-06,
"loss": 0.0647,
"step": 631
},
{
"epoch": 3.11070110701107,
"grad_norm": 1.2734375,
"learning_rate": 3.1753735436312005e-06,
"loss": 0.1091,
"step": 632
},
{
"epoch": 3.1156211562115623,
"grad_norm": 1.046875,
"learning_rate": 3.1608305498544056e-06,
"loss": 0.0939,
"step": 633
},
{
"epoch": 3.1205412054120543,
"grad_norm": 1.0390625,
"learning_rate": 3.146305527771499e-06,
"loss": 0.074,
"step": 634
},
{
"epoch": 3.125461254612546,
"grad_norm": 1.03125,
"learning_rate": 3.1317986193157023e-06,
"loss": 0.0734,
"step": 635
},
{
"epoch": 3.130381303813038,
"grad_norm": 1.3203125,
"learning_rate": 3.1173099662432426e-06,
"loss": 0.1068,
"step": 636
},
{
"epoch": 3.13530135301353,
"grad_norm": 1.484375,
"learning_rate": 3.1028397101319584e-06,
"loss": 0.0917,
"step": 637
},
{
"epoch": 3.140221402214022,
"grad_norm": 1.2265625,
"learning_rate": 3.0883879923799244e-06,
"loss": 0.0743,
"step": 638
},
{
"epoch": 3.145141451414514,
"grad_norm": 1.2734375,
"learning_rate": 3.0739549542040583e-06,
"loss": 0.0971,
"step": 639
},
{
"epoch": 3.1500615006150063,
"grad_norm": 1.0625,
"learning_rate": 3.059540736638751e-06,
"loss": 0.1033,
"step": 640
},
{
"epoch": 3.1549815498154983,
"grad_norm": 0.94921875,
"learning_rate": 3.0451454805344893e-06,
"loss": 0.0899,
"step": 641
},
{
"epoch": 3.15990159901599,
"grad_norm": 1.0859375,
"learning_rate": 3.0307693265564708e-06,
"loss": 0.0845,
"step": 642
},
{
"epoch": 3.164821648216482,
"grad_norm": 1.1484375,
"learning_rate": 3.016412415183233e-06,
"loss": 0.0914,
"step": 643
},
{
"epoch": 3.169741697416974,
"grad_norm": 1.1796875,
"learning_rate": 3.002074886705284e-06,
"loss": 0.0802,
"step": 644
},
{
"epoch": 3.174661746617466,
"grad_norm": 1.3125,
"learning_rate": 2.9877568812237325e-06,
"loss": 0.0946,
"step": 645
},
{
"epoch": 3.179581795817958,
"grad_norm": 1.2265625,
"learning_rate": 2.9734585386489095e-06,
"loss": 0.0782,
"step": 646
},
{
"epoch": 3.1845018450184504,
"grad_norm": 1.0078125,
"learning_rate": 2.9591799986990098e-06,
"loss": 0.0594,
"step": 647
},
{
"epoch": 3.1894218942189423,
"grad_norm": 1.3515625,
"learning_rate": 2.9449214008987253e-06,
"loss": 0.0794,
"step": 648
},
{
"epoch": 3.1943419434194342,
"grad_norm": 1.1640625,
"learning_rate": 2.93068288457788e-06,
"loss": 0.0818,
"step": 649
},
{
"epoch": 3.199261992619926,
"grad_norm": 1.1796875,
"learning_rate": 2.916464588870067e-06,
"loss": 0.1144,
"step": 650
},
{
"epoch": 3.204182041820418,
"grad_norm": 1.265625,
"learning_rate": 2.9022666527112954e-06,
"loss": 0.093,
"step": 651
},
{
"epoch": 3.20910209102091,
"grad_norm": 0.98828125,
"learning_rate": 2.8880892148386198e-06,
"loss": 0.0725,
"step": 652
},
{
"epoch": 3.2140221402214024,
"grad_norm": 1.1796875,
"learning_rate": 2.873932413788805e-06,
"loss": 0.068,
"step": 653
},
{
"epoch": 3.2189421894218944,
"grad_norm": 0.96484375,
"learning_rate": 2.859796387896949e-06,
"loss": 0.0583,
"step": 654
},
{
"epoch": 3.2238622386223863,
"grad_norm": 1.078125,
"learning_rate": 2.8456812752951483e-06,
"loss": 0.0748,
"step": 655
},
{
"epoch": 3.2287822878228782,
"grad_norm": 1.3984375,
"learning_rate": 2.831587213911142e-06,
"loss": 0.1128,
"step": 656
},
{
"epoch": 3.23370233702337,
"grad_norm": 1.328125,
"learning_rate": 2.817514341466965e-06,
"loss": 0.0812,
"step": 657
},
{
"epoch": 3.238622386223862,
"grad_norm": 1.1875,
"learning_rate": 2.8034627954775993e-06,
"loss": 0.0995,
"step": 658
},
{
"epoch": 3.243542435424354,
"grad_norm": 1.0546875,
"learning_rate": 2.7894327132496324e-06,
"loss": 0.0865,
"step": 659
},
{
"epoch": 3.2484624846248464,
"grad_norm": 1.171875,
"learning_rate": 2.7754242318799174e-06,
"loss": 0.0807,
"step": 660
},
{
"epoch": 3.2533825338253384,
"grad_norm": 1.015625,
"learning_rate": 2.761437488254232e-06,
"loss": 0.1041,
"step": 661
},
{
"epoch": 3.2583025830258303,
"grad_norm": 1.0703125,
"learning_rate": 2.74747261904594e-06,
"loss": 0.074,
"step": 662
},
{
"epoch": 3.2632226322263223,
"grad_norm": 1.203125,
"learning_rate": 2.733529760714655e-06,
"loss": 0.0852,
"step": 663
},
{
"epoch": 3.268142681426814,
"grad_norm": 1.3515625,
"learning_rate": 2.719609049504911e-06,
"loss": 0.1049,
"step": 664
},
{
"epoch": 3.273062730627306,
"grad_norm": 1.2265625,
"learning_rate": 2.7057106214448216e-06,
"loss": 0.0682,
"step": 665
},
{
"epoch": 3.2779827798277985,
"grad_norm": 1.3828125,
"learning_rate": 2.6918346123447708e-06,
"loss": 0.1297,
"step": 666
},
{
"epoch": 3.2829028290282904,
"grad_norm": 1.1328125,
"learning_rate": 2.677981157796059e-06,
"loss": 0.0748,
"step": 667
},
{
"epoch": 3.2878228782287824,
"grad_norm": 1.125,
"learning_rate": 2.6641503931696e-06,
"loss": 0.1069,
"step": 668
},
{
"epoch": 3.2927429274292743,
"grad_norm": 1.0703125,
"learning_rate": 2.650342453614586e-06,
"loss": 0.0802,
"step": 669
},
{
"epoch": 3.2976629766297663,
"grad_norm": 1.1796875,
"learning_rate": 2.636557474057173e-06,
"loss": 0.0908,
"step": 670
},
{
"epoch": 3.302583025830258,
"grad_norm": 0.91015625,
"learning_rate": 2.6227955891991617e-06,
"loss": 0.0663,
"step": 671
},
{
"epoch": 3.30750307503075,
"grad_norm": 1.390625,
"learning_rate": 2.609056933516675e-06,
"loss": 0.0686,
"step": 672
},
{
"epoch": 3.312423124231242,
"grad_norm": 1.0390625,
"learning_rate": 2.5953416412588504e-06,
"loss": 0.0833,
"step": 673
},
{
"epoch": 3.3173431734317345,
"grad_norm": 1.1328125,
"learning_rate": 2.58164984644653e-06,
"loss": 0.0934,
"step": 674
},
{
"epoch": 3.3222632226322264,
"grad_norm": 1.1796875,
"learning_rate": 2.567981682870946e-06,
"loss": 0.0953,
"step": 675
},
{
"epoch": 3.3271832718327183,
"grad_norm": 1.1328125,
"learning_rate": 2.554337284092411e-06,
"loss": 0.1022,
"step": 676
},
{
"epoch": 3.3321033210332103,
"grad_norm": 1.1953125,
"learning_rate": 2.540716783439019e-06,
"loss": 0.0674,
"step": 677
},
{
"epoch": 3.337023370233702,
"grad_norm": 1.28125,
"learning_rate": 2.5271203140053436e-06,
"loss": 0.1064,
"step": 678
},
{
"epoch": 3.341943419434194,
"grad_norm": 1.1796875,
"learning_rate": 2.5135480086511306e-06,
"loss": 0.0989,
"step": 679
},
{
"epoch": 3.3468634686346865,
"grad_norm": 1.25,
"learning_rate": 2.5000000000000015e-06,
"loss": 0.0908,
"step": 680
},
{
"epoch": 3.3517835178351785,
"grad_norm": 1.2421875,
"learning_rate": 2.4864764204381624e-06,
"loss": 0.0834,
"step": 681
},
{
"epoch": 3.3567035670356704,
"grad_norm": 1.4921875,
"learning_rate": 2.472977402113107e-06,
"loss": 0.0902,
"step": 682
},
{
"epoch": 3.3616236162361623,
"grad_norm": 1.1875,
"learning_rate": 2.4595030769323246e-06,
"loss": 0.0718,
"step": 683
},
{
"epoch": 3.3665436654366543,
"grad_norm": 1.609375,
"learning_rate": 2.4460535765620147e-06,
"loss": 0.1142,
"step": 684
},
{
"epoch": 3.3714637146371462,
"grad_norm": 1.1953125,
"learning_rate": 2.4326290324257896e-06,
"loss": 0.0951,
"step": 685
},
{
"epoch": 3.376383763837638,
"grad_norm": 1.328125,
"learning_rate": 2.419229575703411e-06,
"loss": 0.0985,
"step": 686
},
{
"epoch": 3.3813038130381305,
"grad_norm": 1.109375,
"learning_rate": 2.4058553373294846e-06,
"loss": 0.0713,
"step": 687
},
{
"epoch": 3.3862238622386225,
"grad_norm": 1.3359375,
"learning_rate": 2.3925064479921985e-06,
"loss": 0.0882,
"step": 688
},
{
"epoch": 3.3911439114391144,
"grad_norm": 1.2109375,
"learning_rate": 2.379183038132031e-06,
"loss": 0.1156,
"step": 689
},
{
"epoch": 3.3960639606396064,
"grad_norm": 1.171875,
"learning_rate": 2.3658852379404973e-06,
"loss": 0.0914,
"step": 690
},
{
"epoch": 3.4009840098400983,
"grad_norm": 1.1953125,
"learning_rate": 2.352613177358852e-06,
"loss": 0.0794,
"step": 691
},
{
"epoch": 3.4059040590405902,
"grad_norm": 1.0859375,
"learning_rate": 2.3393669860768364e-06,
"loss": 0.0941,
"step": 692
},
{
"epoch": 3.4108241082410826,
"grad_norm": 1.0078125,
"learning_rate": 2.3261467935314097e-06,
"loss": 0.0648,
"step": 693
},
{
"epoch": 3.4157441574415746,
"grad_norm": 1.25,
"learning_rate": 2.3129527289054716e-06,
"loss": 0.0912,
"step": 694
},
{
"epoch": 3.4206642066420665,
"grad_norm": 1.0390625,
"learning_rate": 2.299784921126622e-06,
"loss": 0.0801,
"step": 695
},
{
"epoch": 3.4255842558425584,
"grad_norm": 1.078125,
"learning_rate": 2.286643498865877e-06,
"loss": 0.072,
"step": 696
},
{
"epoch": 3.4305043050430504,
"grad_norm": 0.96875,
"learning_rate": 2.2735285905364307e-06,
"loss": 0.0599,
"step": 697
},
{
"epoch": 3.4354243542435423,
"grad_norm": 1.015625,
"learning_rate": 2.260440324292385e-06,
"loss": 0.0748,
"step": 698
},
{
"epoch": 3.4403444034440342,
"grad_norm": 0.95703125,
"learning_rate": 2.2473788280275164e-06,
"loss": 0.0613,
"step": 699
},
{
"epoch": 3.4452644526445266,
"grad_norm": 1.390625,
"learning_rate": 2.234344229374003e-06,
"loss": 0.107,
"step": 700
},
{
"epoch": 3.4501845018450186,
"grad_norm": 1.109375,
"learning_rate": 2.2213366557011988e-06,
"loss": 0.083,
"step": 701
},
{
"epoch": 3.4551045510455105,
"grad_norm": 0.953125,
"learning_rate": 2.2083562341143695e-06,
"loss": 0.0823,
"step": 702
},
{
"epoch": 3.4600246002460024,
"grad_norm": 1.4609375,
"learning_rate": 2.195403091453473e-06,
"loss": 0.1011,
"step": 703
},
{
"epoch": 3.4649446494464944,
"grad_norm": 1.171875,
"learning_rate": 2.1824773542918957e-06,
"loss": 0.0832,
"step": 704
},
{
"epoch": 3.4698646986469863,
"grad_norm": 1.5859375,
"learning_rate": 2.1695791489352346e-06,
"loss": 0.1078,
"step": 705
},
{
"epoch": 3.4747847478474787,
"grad_norm": 0.8984375,
"learning_rate": 2.156708601420053e-06,
"loss": 0.0529,
"step": 706
},
{
"epoch": 3.4797047970479706,
"grad_norm": 1.0546875,
"learning_rate": 2.1438658375126544e-06,
"loss": 0.0788,
"step": 707
},
{
"epoch": 3.4846248462484626,
"grad_norm": 1.15625,
"learning_rate": 2.131050982707849e-06,
"loss": 0.0749,
"step": 708
},
{
"epoch": 3.4895448954489545,
"grad_norm": 1.296875,
"learning_rate": 2.1182641622277273e-06,
"loss": 0.0666,
"step": 709
},
{
"epoch": 3.4944649446494465,
"grad_norm": 1.0,
"learning_rate": 2.1055055010204427e-06,
"loss": 0.0892,
"step": 710
},
{
"epoch": 3.4993849938499384,
"grad_norm": 1.203125,
"learning_rate": 2.092775123758985e-06,
"loss": 0.1143,
"step": 711
},
{
"epoch": 3.5043050430504303,
"grad_norm": 0.9765625,
"learning_rate": 2.080073154839964e-06,
"loss": 0.0703,
"step": 712
},
{
"epoch": 3.5092250922509223,
"grad_norm": 1.0703125,
"learning_rate": 2.06739971838239e-06,
"loss": 0.083,
"step": 713
},
{
"epoch": 3.5141451414514147,
"grad_norm": 1.09375,
"learning_rate": 2.05475493822647e-06,
"loss": 0.0697,
"step": 714
},
{
"epoch": 3.5141451414514147,
"eval_loss": 0.10337568074464798,
"eval_runtime": 203.8015,
"eval_samples_per_second": 8.027,
"eval_steps_per_second": 1.34,
"step": 714
},
{
"epoch": 3.5190651906519066,
"grad_norm": 1.375,
"learning_rate": 2.042138937932388e-06,
"loss": 0.0886,
"step": 715
},
{
"epoch": 3.5239852398523985,
"grad_norm": 1.21875,
"learning_rate": 2.0295518407791054e-06,
"loss": 0.1,
"step": 716
},
{
"epoch": 3.5289052890528905,
"grad_norm": 1.1484375,
"learning_rate": 2.016993769763147e-06,
"loss": 0.0802,
"step": 717
},
{
"epoch": 3.5338253382533824,
"grad_norm": 0.98828125,
"learning_rate": 2.0044648475974095e-06,
"loss": 0.058,
"step": 718
},
{
"epoch": 3.538745387453875,
"grad_norm": 1.0234375,
"learning_rate": 1.9919651967099556e-06,
"loss": 0.0875,
"step": 719
},
{
"epoch": 3.5436654366543667,
"grad_norm": 1.25,
"learning_rate": 1.979494939242822e-06,
"loss": 0.0656,
"step": 720
},
{
"epoch": 3.5485854858548587,
"grad_norm": 1.09375,
"learning_rate": 1.9670541970508224e-06,
"loss": 0.0676,
"step": 721
},
{
"epoch": 3.5535055350553506,
"grad_norm": 1.0859375,
"learning_rate": 1.954643091700354e-06,
"loss": 0.0672,
"step": 722
},
{
"epoch": 3.5584255842558425,
"grad_norm": 1.3984375,
"learning_rate": 1.9422617444682196e-06,
"loss": 0.0816,
"step": 723
},
{
"epoch": 3.5633456334563345,
"grad_norm": 1.0234375,
"learning_rate": 1.9299102763404335e-06,
"loss": 0.0769,
"step": 724
},
{
"epoch": 3.5682656826568264,
"grad_norm": 0.81640625,
"learning_rate": 1.917588808011045e-06,
"loss": 0.069,
"step": 725
},
{
"epoch": 3.5731857318573184,
"grad_norm": 1.4765625,
"learning_rate": 1.9052974598809505e-06,
"loss": 0.075,
"step": 726
},
{
"epoch": 3.5781057810578107,
"grad_norm": 1.34375,
"learning_rate": 1.893036352056728e-06,
"loss": 0.1036,
"step": 727
},
{
"epoch": 3.5830258302583027,
"grad_norm": 1.2734375,
"learning_rate": 1.880805604349456e-06,
"loss": 0.0658,
"step": 728
},
{
"epoch": 3.5879458794587946,
"grad_norm": 0.8046875,
"learning_rate": 1.8686053362735468e-06,
"loss": 0.052,
"step": 729
},
{
"epoch": 3.5928659286592866,
"grad_norm": 0.828125,
"learning_rate": 1.856435667045577e-06,
"loss": 0.0592,
"step": 730
},
{
"epoch": 3.5977859778597785,
"grad_norm": 1.2265625,
"learning_rate": 1.844296715583117e-06,
"loss": 0.0679,
"step": 731
},
{
"epoch": 3.602706027060271,
"grad_norm": 1.0390625,
"learning_rate": 1.8321886005035812e-06,
"loss": 0.0883,
"step": 732
},
{
"epoch": 3.607626076260763,
"grad_norm": 1.25,
"learning_rate": 1.8201114401230586e-06,
"loss": 0.0875,
"step": 733
},
{
"epoch": 3.6125461254612548,
"grad_norm": 1.1953125,
"learning_rate": 1.8080653524551623e-06,
"loss": 0.0662,
"step": 734
},
{
"epoch": 3.6174661746617467,
"grad_norm": 1.859375,
"learning_rate": 1.796050455209869e-06,
"loss": 0.1162,
"step": 735
},
{
"epoch": 3.6223862238622386,
"grad_norm": 1.21875,
"learning_rate": 1.7840668657923838e-06,
"loss": 0.0753,
"step": 736
},
{
"epoch": 3.6273062730627306,
"grad_norm": 1.1953125,
"learning_rate": 1.772114701301972e-06,
"loss": 0.0767,
"step": 737
},
{
"epoch": 3.6322263222632225,
"grad_norm": 1.1640625,
"learning_rate": 1.7601940785308337e-06,
"loss": 0.1026,
"step": 738
},
{
"epoch": 3.6371463714637144,
"grad_norm": 1.3203125,
"learning_rate": 1.7483051139629482e-06,
"loss": 0.0724,
"step": 739
},
{
"epoch": 3.6420664206642064,
"grad_norm": 1.015625,
"learning_rate": 1.7364479237729526e-06,
"loss": 0.0915,
"step": 740
},
{
"epoch": 3.6469864698646988,
"grad_norm": 1.375,
"learning_rate": 1.7246226238249853e-06,
"loss": 0.057,
"step": 741
},
{
"epoch": 3.6519065190651907,
"grad_norm": 1.3515625,
"learning_rate": 1.7128293296715704e-06,
"loss": 0.0772,
"step": 742
},
{
"epoch": 3.6568265682656826,
"grad_norm": 1.15625,
"learning_rate": 1.7010681565524834e-06,
"loss": 0.0597,
"step": 743
},
{
"epoch": 3.6617466174661746,
"grad_norm": 1.2578125,
"learning_rate": 1.6893392193936231e-06,
"loss": 0.1057,
"step": 744
},
{
"epoch": 3.6666666666666665,
"grad_norm": 1.0390625,
"learning_rate": 1.677642632805892e-06,
"loss": 0.0869,
"step": 745
},
{
"epoch": 3.671586715867159,
"grad_norm": 1.1015625,
"learning_rate": 1.6659785110840704e-06,
"loss": 0.0846,
"step": 746
},
{
"epoch": 3.676506765067651,
"grad_norm": 1.3125,
"learning_rate": 1.6543469682057105e-06,
"loss": 0.0953,
"step": 747
},
{
"epoch": 3.6814268142681428,
"grad_norm": 1.5234375,
"learning_rate": 1.6427481178300064e-06,
"loss": 0.0984,
"step": 748
},
{
"epoch": 3.6863468634686347,
"grad_norm": 1.2265625,
"learning_rate": 1.631182073296706e-06,
"loss": 0.099,
"step": 749
},
{
"epoch": 3.6912669126691267,
"grad_norm": 1.3828125,
"learning_rate": 1.6196489476249777e-06,
"loss": 0.1055,
"step": 750
},
{
"epoch": 3.6961869618696186,
"grad_norm": 1.265625,
"learning_rate": 1.6081488535123274e-06,
"loss": 0.1011,
"step": 751
},
{
"epoch": 3.7011070110701105,
"grad_norm": 0.91015625,
"learning_rate": 1.5966819033334807e-06,
"loss": 0.0697,
"step": 752
},
{
"epoch": 3.7060270602706025,
"grad_norm": 1.203125,
"learning_rate": 1.5852482091393045e-06,
"loss": 0.0792,
"step": 753
},
{
"epoch": 3.710947109471095,
"grad_norm": 0.8671875,
"learning_rate": 1.5738478826556885e-06,
"loss": 0.0589,
"step": 754
},
{
"epoch": 3.715867158671587,
"grad_norm": 1.328125,
"learning_rate": 1.5624810352824709e-06,
"loss": 0.1095,
"step": 755
},
{
"epoch": 3.7207872078720787,
"grad_norm": 1.03125,
"learning_rate": 1.551147778092344e-06,
"loss": 0.0697,
"step": 756
},
{
"epoch": 3.7257072570725707,
"grad_norm": 0.7890625,
"learning_rate": 1.539848221829769e-06,
"loss": 0.0562,
"step": 757
},
{
"epoch": 3.7306273062730626,
"grad_norm": 1.125,
"learning_rate": 1.5285824769098938e-06,
"loss": 0.064,
"step": 758
},
{
"epoch": 3.735547355473555,
"grad_norm": 1.234375,
"learning_rate": 1.517350653417472e-06,
"loss": 0.0805,
"step": 759
},
{
"epoch": 3.740467404674047,
"grad_norm": 1.34375,
"learning_rate": 1.5061528611057917e-06,
"loss": 0.0751,
"step": 760
},
{
"epoch": 3.745387453874539,
"grad_norm": 1.234375,
"learning_rate": 1.4949892093956015e-06,
"loss": 0.0958,
"step": 761
},
{
"epoch": 3.750307503075031,
"grad_norm": 1.1875,
"learning_rate": 1.4838598073740395e-06,
"loss": 0.1045,
"step": 762
},
{
"epoch": 3.7552275522755227,
"grad_norm": 0.984375,
"learning_rate": 1.472764763793565e-06,
"loss": 0.0532,
"step": 763
},
{
"epoch": 3.7601476014760147,
"grad_norm": 1.3046875,
"learning_rate": 1.4617041870709042e-06,
"loss": 0.0961,
"step": 764
},
{
"epoch": 3.7650676506765066,
"grad_norm": 1.2265625,
"learning_rate": 1.4506781852859836e-06,
"loss": 0.0696,
"step": 765
},
{
"epoch": 3.7699876998769986,
"grad_norm": 1.015625,
"learning_rate": 1.4396868661808777e-06,
"loss": 0.076,
"step": 766
},
{
"epoch": 3.774907749077491,
"grad_norm": 1.0078125,
"learning_rate": 1.428730337158749e-06,
"loss": 0.0949,
"step": 767
},
{
"epoch": 3.779827798277983,
"grad_norm": 1.3203125,
"learning_rate": 1.4178087052828098e-06,
"loss": 0.0981,
"step": 768
},
{
"epoch": 3.784747847478475,
"grad_norm": 1.25,
"learning_rate": 1.4069220772752685e-06,
"loss": 0.1072,
"step": 769
},
{
"epoch": 3.7896678966789668,
"grad_norm": 0.9921875,
"learning_rate": 1.3960705595162876e-06,
"loss": 0.063,
"step": 770
},
{
"epoch": 3.7945879458794587,
"grad_norm": 0.86328125,
"learning_rate": 1.385254258042948e-06,
"loss": 0.0707,
"step": 771
},
{
"epoch": 3.799507995079951,
"grad_norm": 1.1484375,
"learning_rate": 1.3744732785482035e-06,
"loss": 0.0804,
"step": 772
},
{
"epoch": 3.804428044280443,
"grad_norm": 1.2109375,
"learning_rate": 1.3637277263798603e-06,
"loss": 0.1245,
"step": 773
},
{
"epoch": 3.809348093480935,
"grad_norm": 1.046875,
"learning_rate": 1.35301770653954e-06,
"loss": 0.0788,
"step": 774
},
{
"epoch": 3.814268142681427,
"grad_norm": 0.85546875,
"learning_rate": 1.3423433236816563e-06,
"loss": 0.0559,
"step": 775
},
{
"epoch": 3.819188191881919,
"grad_norm": 1.265625,
"learning_rate": 1.3317046821123868e-06,
"loss": 0.0958,
"step": 776
},
{
"epoch": 3.8241082410824108,
"grad_norm": 1.3671875,
"learning_rate": 1.3211018857886632e-06,
"loss": 0.0748,
"step": 777
},
{
"epoch": 3.8290282902829027,
"grad_norm": 0.80078125,
"learning_rate": 1.3105350383171484e-06,
"loss": 0.0673,
"step": 778
},
{
"epoch": 3.8339483394833946,
"grad_norm": 1.234375,
"learning_rate": 1.3000042429532267e-06,
"loss": 0.0911,
"step": 779
},
{
"epoch": 3.8388683886838866,
"grad_norm": 0.90625,
"learning_rate": 1.289509602599996e-06,
"loss": 0.0776,
"step": 780
},
{
"epoch": 3.843788437884379,
"grad_norm": 0.9453125,
"learning_rate": 1.2790512198072558e-06,
"loss": 0.0996,
"step": 781
},
{
"epoch": 3.848708487084871,
"grad_norm": 1.1328125,
"learning_rate": 1.268629196770514e-06,
"loss": 0.0685,
"step": 782
},
{
"epoch": 3.853628536285363,
"grad_norm": 1.203125,
"learning_rate": 1.2582436353299832e-06,
"loss": 0.0924,
"step": 783
},
{
"epoch": 3.8585485854858548,
"grad_norm": 1.28125,
"learning_rate": 1.2478946369695882e-06,
"loss": 0.0901,
"step": 784
},
{
"epoch": 3.8634686346863467,
"grad_norm": 1.03125,
"learning_rate": 1.2375823028159667e-06,
"loss": 0.0971,
"step": 785
},
{
"epoch": 3.868388683886839,
"grad_norm": 1.28125,
"learning_rate": 1.2273067336374972e-06,
"loss": 0.087,
"step": 786
},
{
"epoch": 3.873308733087331,
"grad_norm": 1.4609375,
"learning_rate": 1.2170680298432934e-06,
"loss": 0.106,
"step": 787
},
{
"epoch": 3.878228782287823,
"grad_norm": 1.15625,
"learning_rate": 1.2068662914822432e-06,
"loss": 0.0634,
"step": 788
},
{
"epoch": 3.883148831488315,
"grad_norm": 1.734375,
"learning_rate": 1.1967016182420122e-06,
"loss": 0.0983,
"step": 789
},
{
"epoch": 3.888068880688807,
"grad_norm": 1.1484375,
"learning_rate": 1.186574109448091e-06,
"loss": 0.0899,
"step": 790
},
{
"epoch": 3.892988929889299,
"grad_norm": 1.046875,
"learning_rate": 1.1764838640628011e-06,
"loss": 0.0799,
"step": 791
},
{
"epoch": 3.8979089790897907,
"grad_norm": 1.390625,
"learning_rate": 1.1664309806843466e-06,
"loss": 0.0942,
"step": 792
},
{
"epoch": 3.9028290282902827,
"grad_norm": 1.4921875,
"learning_rate": 1.1564155575458414e-06,
"loss": 0.0895,
"step": 793
},
{
"epoch": 3.907749077490775,
"grad_norm": 1.3359375,
"learning_rate": 1.1464376925143528e-06,
"loss": 0.0802,
"step": 794
},
{
"epoch": 3.912669126691267,
"grad_norm": 1.2578125,
"learning_rate": 1.1364974830899438e-06,
"loss": 0.0689,
"step": 795
},
{
"epoch": 3.917589175891759,
"grad_norm": 0.96875,
"learning_rate": 1.1265950264047171e-06,
"loss": 0.0677,
"step": 796
},
{
"epoch": 3.922509225092251,
"grad_norm": 1.1328125,
"learning_rate": 1.1167304192218737e-06,
"loss": 0.1054,
"step": 797
},
{
"epoch": 3.927429274292743,
"grad_norm": 1.203125,
"learning_rate": 1.1069037579347613e-06,
"loss": 0.1051,
"step": 798
},
{
"epoch": 3.932349323493235,
"grad_norm": 1.0,
"learning_rate": 1.0971151385659357e-06,
"loss": 0.0892,
"step": 799
},
{
"epoch": 3.937269372693727,
"grad_norm": 0.8671875,
"learning_rate": 1.0873646567662165e-06,
"loss": 0.0544,
"step": 800
},
{
"epoch": 3.942189421894219,
"grad_norm": 1.328125,
"learning_rate": 1.077652407813764e-06,
"loss": 0.0906,
"step": 801
},
{
"epoch": 3.947109471094711,
"grad_norm": 0.8984375,
"learning_rate": 1.067978486613131e-06,
"loss": 0.056,
"step": 802
},
{
"epoch": 3.952029520295203,
"grad_norm": 1.1328125,
"learning_rate": 1.0583429876943585e-06,
"loss": 0.0857,
"step": 803
},
{
"epoch": 3.956949569495695,
"grad_norm": 1.0078125,
"learning_rate": 1.0487460052120262e-06,
"loss": 0.0857,
"step": 804
},
{
"epoch": 3.961869618696187,
"grad_norm": 1.5859375,
"learning_rate": 1.0391876329443534e-06,
"loss": 0.0975,
"step": 805
},
{
"epoch": 3.9667896678966788,
"grad_norm": 1.0546875,
"learning_rate": 1.0296679642922718e-06,
"loss": 0.0768,
"step": 806
},
{
"epoch": 3.971709717097171,
"grad_norm": 1.484375,
"learning_rate": 1.0201870922785156e-06,
"loss": 0.1113,
"step": 807
},
{
"epoch": 3.976629766297663,
"grad_norm": 1.171875,
"learning_rate": 1.010745109546713e-06,
"loss": 0.0812,
"step": 808
},
{
"epoch": 3.981549815498155,
"grad_norm": 0.82421875,
"learning_rate": 1.0013421083604779e-06,
"loss": 0.062,
"step": 809
},
{
"epoch": 3.986469864698647,
"grad_norm": 0.96875,
"learning_rate": 9.919781806025136e-07,
"loss": 0.0666,
"step": 810
},
{
"epoch": 3.991389913899139,
"grad_norm": 1.0859375,
"learning_rate": 9.826534177737106e-07,
"loss": 0.0616,
"step": 811
},
{
"epoch": 3.9963099630996313,
"grad_norm": 1.140625,
"learning_rate": 9.733679109922567e-07,
"loss": 0.0663,
"step": 812
},
{
"epoch": 4.001230012300123,
"grad_norm": 2.390625,
"learning_rate": 9.641217509927376e-07,
"loss": 0.0926,
"step": 813
},
{
"epoch": 4.006150061500615,
"grad_norm": 1.1875,
"learning_rate": 9.549150281252633e-07,
"loss": 0.0742,
"step": 814
},
{
"epoch": 4.011070110701107,
"grad_norm": 1.0078125,
"learning_rate": 9.457478323545749e-07,
"loss": 0.1039,
"step": 815
},
{
"epoch": 4.015990159901599,
"grad_norm": 1.0859375,
"learning_rate": 9.366202532591717e-07,
"loss": 0.0794,
"step": 816
},
{
"epoch": 4.015990159901599,
"eval_loss": 0.10322786867618561,
"eval_runtime": 204.3584,
"eval_samples_per_second": 8.006,
"eval_steps_per_second": 1.336,
"step": 816
},
{
"epoch": 4.020910209102091,
"grad_norm": 0.78515625,
"learning_rate": 9.275323800304287e-07,
"loss": 0.0628,
"step": 817
},
{
"epoch": 4.025830258302583,
"grad_norm": 0.9296875,
"learning_rate": 9.184843014717337e-07,
"loss": 0.0652,
"step": 818
},
{
"epoch": 4.030750307503075,
"grad_norm": 1.0625,
"learning_rate": 9.094761059976154e-07,
"loss": 0.0884,
"step": 819
},
{
"epoch": 4.035670356703567,
"grad_norm": 1.0234375,
"learning_rate": 9.005078816328772e-07,
"loss": 0.0858,
"step": 820
},
{
"epoch": 4.040590405904059,
"grad_norm": 1.125,
"learning_rate": 8.915797160117423e-07,
"loss": 0.1029,
"step": 821
},
{
"epoch": 4.045510455104551,
"grad_norm": 1.140625,
"learning_rate": 8.826916963769888e-07,
"loss": 0.0981,
"step": 822
},
{
"epoch": 4.0504305043050435,
"grad_norm": 1.0078125,
"learning_rate": 8.738439095791123e-07,
"loss": 0.0688,
"step": 823
},
{
"epoch": 4.055350553505535,
"grad_norm": 0.98828125,
"learning_rate": 8.650364420754581e-07,
"loss": 0.0715,
"step": 824
},
{
"epoch": 4.060270602706027,
"grad_norm": 0.94140625,
"learning_rate": 8.562693799293931e-07,
"loss": 0.0624,
"step": 825
},
{
"epoch": 4.065190651906519,
"grad_norm": 1.328125,
"learning_rate": 8.475428088094517e-07,
"loss": 0.0977,
"step": 826
},
{
"epoch": 4.070110701107011,
"grad_norm": 1.140625,
"learning_rate": 8.388568139885101e-07,
"loss": 0.0827,
"step": 827
},
{
"epoch": 4.075030750307503,
"grad_norm": 1.0078125,
"learning_rate": 8.30211480342945e-07,
"loss": 0.0824,
"step": 828
},
{
"epoch": 4.079950799507995,
"grad_norm": 1.0390625,
"learning_rate": 8.216068923518072e-07,
"loss": 0.0713,
"step": 829
},
{
"epoch": 4.084870848708487,
"grad_norm": 1.171875,
"learning_rate": 8.130431340959982e-07,
"loss": 0.0905,
"step": 830
},
{
"epoch": 4.089790897908979,
"grad_norm": 1.0390625,
"learning_rate": 8.045202892574395e-07,
"loss": 0.0692,
"step": 831
},
{
"epoch": 4.094710947109471,
"grad_norm": 0.9765625,
"learning_rate": 7.960384411182709e-07,
"loss": 0.0713,
"step": 832
},
{
"epoch": 4.099630996309963,
"grad_norm": 1.03125,
"learning_rate": 7.875976725600193e-07,
"loss": 0.0824,
"step": 833
},
{
"epoch": 4.104551045510455,
"grad_norm": 0.76953125,
"learning_rate": 7.791980660628029e-07,
"loss": 0.0579,
"step": 834
},
{
"epoch": 4.109471094710947,
"grad_norm": 0.98828125,
"learning_rate": 7.708397037045129e-07,
"loss": 0.081,
"step": 835
},
{
"epoch": 4.114391143911439,
"grad_norm": 1.0078125,
"learning_rate": 7.625226671600256e-07,
"loss": 0.0884,
"step": 836
},
{
"epoch": 4.1193111931119315,
"grad_norm": 1.0546875,
"learning_rate": 7.542470377003897e-07,
"loss": 0.0809,
"step": 837
},
{
"epoch": 4.124231242312423,
"grad_norm": 1.0546875,
"learning_rate": 7.460128961920432e-07,
"loss": 0.0768,
"step": 838
},
{
"epoch": 4.129151291512915,
"grad_norm": 0.8203125,
"learning_rate": 7.37820323096014e-07,
"loss": 0.0809,
"step": 839
},
{
"epoch": 4.134071340713407,
"grad_norm": 0.8359375,
"learning_rate": 7.296693984671465e-07,
"loss": 0.0748,
"step": 840
},
{
"epoch": 4.138991389913899,
"grad_norm": 1.046875,
"learning_rate": 7.215602019533041e-07,
"loss": 0.1046,
"step": 841
},
{
"epoch": 4.143911439114391,
"grad_norm": 1.1015625,
"learning_rate": 7.134928127946017e-07,
"loss": 0.0804,
"step": 842
},
{
"epoch": 4.148831488314883,
"grad_norm": 0.9375,
"learning_rate": 7.054673098226278e-07,
"loss": 0.0735,
"step": 843
},
{
"epoch": 4.153751537515375,
"grad_norm": 1.1328125,
"learning_rate": 6.974837714596732e-07,
"loss": 0.0918,
"step": 844
},
{
"epoch": 4.158671586715867,
"grad_norm": 1.1328125,
"learning_rate": 6.895422757179682e-07,
"loss": 0.0926,
"step": 845
},
{
"epoch": 4.163591635916359,
"grad_norm": 1.03125,
"learning_rate": 6.816429001989133e-07,
"loss": 0.0887,
"step": 846
},
{
"epoch": 4.168511685116851,
"grad_norm": 1.078125,
"learning_rate": 6.737857220923305e-07,
"loss": 0.0546,
"step": 847
},
{
"epoch": 4.173431734317343,
"grad_norm": 1.015625,
"learning_rate": 6.659708181757013e-07,
"loss": 0.0709,
"step": 848
},
{
"epoch": 4.178351783517835,
"grad_norm": 0.84375,
"learning_rate": 6.581982648134217e-07,
"loss": 0.0709,
"step": 849
},
{
"epoch": 4.183271832718328,
"grad_norm": 1.015625,
"learning_rate": 6.50468137956049e-07,
"loss": 0.0639,
"step": 850
},
{
"epoch": 4.1881918819188195,
"grad_norm": 0.90234375,
"learning_rate": 6.427805131395681e-07,
"loss": 0.0931,
"step": 851
},
{
"epoch": 4.1931119311193115,
"grad_norm": 0.87890625,
"learning_rate": 6.351354654846481e-07,
"loss": 0.0811,
"step": 852
},
{
"epoch": 4.198031980319803,
"grad_norm": 1.171875,
"learning_rate": 6.275330696959109e-07,
"loss": 0.0872,
"step": 853
},
{
"epoch": 4.202952029520295,
"grad_norm": 0.9921875,
"learning_rate": 6.199734000611968e-07,
"loss": 0.0826,
"step": 854
},
{
"epoch": 4.207872078720787,
"grad_norm": 0.81640625,
"learning_rate": 6.12456530450844e-07,
"loss": 0.0732,
"step": 855
},
{
"epoch": 4.212792127921279,
"grad_norm": 0.91796875,
"learning_rate": 6.049825343169652e-07,
"loss": 0.0697,
"step": 856
},
{
"epoch": 4.217712177121771,
"grad_norm": 0.99609375,
"learning_rate": 5.975514846927271e-07,
"loss": 0.0767,
"step": 857
},
{
"epoch": 4.222632226322263,
"grad_norm": 0.921875,
"learning_rate": 5.901634541916406e-07,
"loss": 0.0632,
"step": 858
},
{
"epoch": 4.227552275522755,
"grad_norm": 0.984375,
"learning_rate": 5.828185150068472e-07,
"loss": 0.0798,
"step": 859
},
{
"epoch": 4.232472324723247,
"grad_norm": 0.796875,
"learning_rate": 5.755167389104166e-07,
"loss": 0.0685,
"step": 860
},
{
"epoch": 4.237392373923739,
"grad_norm": 0.87109375,
"learning_rate": 5.682581972526463e-07,
"loss": 0.0645,
"step": 861
},
{
"epoch": 4.242312423124231,
"grad_norm": 0.84375,
"learning_rate": 5.610429609613615e-07,
"loss": 0.0588,
"step": 862
},
{
"epoch": 4.247232472324724,
"grad_norm": 0.7890625,
"learning_rate": 5.538711005412212e-07,
"loss": 0.0581,
"step": 863
},
{
"epoch": 4.252152521525216,
"grad_norm": 0.984375,
"learning_rate": 5.467426860730334e-07,
"loss": 0.082,
"step": 864
},
{
"epoch": 4.2570725707257075,
"grad_norm": 0.86328125,
"learning_rate": 5.396577872130676e-07,
"loss": 0.0584,
"step": 865
},
{
"epoch": 4.2619926199261995,
"grad_norm": 1.0234375,
"learning_rate": 5.32616473192375e-07,
"loss": 0.0822,
"step": 866
},
{
"epoch": 4.266912669126691,
"grad_norm": 1.1328125,
"learning_rate": 5.256188128161116e-07,
"loss": 0.0834,
"step": 867
},
{
"epoch": 4.271832718327183,
"grad_norm": 1.078125,
"learning_rate": 5.186648744628637e-07,
"loss": 0.0703,
"step": 868
},
{
"epoch": 4.276752767527675,
"grad_norm": 1.2734375,
"learning_rate": 5.117547260839845e-07,
"loss": 0.1001,
"step": 869
},
{
"epoch": 4.281672816728167,
"grad_norm": 1.3046875,
"learning_rate": 5.048884352029271e-07,
"loss": 0.1088,
"step": 870
},
{
"epoch": 4.286592865928659,
"grad_norm": 0.80859375,
"learning_rate": 4.980660689145855e-07,
"loss": 0.0635,
"step": 871
},
{
"epoch": 4.291512915129151,
"grad_norm": 1.0546875,
"learning_rate": 4.912876938846345e-07,
"loss": 0.0876,
"step": 872
},
{
"epoch": 4.296432964329643,
"grad_norm": 1.203125,
"learning_rate": 4.845533763488902e-07,
"loss": 0.0893,
"step": 873
},
{
"epoch": 4.301353013530135,
"grad_norm": 1.2421875,
"learning_rate": 4.778631821126473e-07,
"loss": 0.0946,
"step": 874
},
{
"epoch": 4.306273062730627,
"grad_norm": 1.1875,
"learning_rate": 4.712171765500484e-07,
"loss": 0.0911,
"step": 875
},
{
"epoch": 4.31119311193112,
"grad_norm": 1.1171875,
"learning_rate": 4.6461542460343565e-07,
"loss": 0.1092,
"step": 876
},
{
"epoch": 4.316113161131612,
"grad_norm": 1.0546875,
"learning_rate": 4.580579907827287e-07,
"loss": 0.0921,
"step": 877
},
{
"epoch": 4.321033210332104,
"grad_norm": 1.0546875,
"learning_rate": 4.515449391647786e-07,
"loss": 0.0759,
"step": 878
},
{
"epoch": 4.325953259532596,
"grad_norm": 1.109375,
"learning_rate": 4.4507633339275494e-07,
"loss": 0.0808,
"step": 879
},
{
"epoch": 4.3308733087330875,
"grad_norm": 0.98046875,
"learning_rate": 4.386522366755169e-07,
"loss": 0.0842,
"step": 880
},
{
"epoch": 4.3357933579335795,
"grad_norm": 1.078125,
"learning_rate": 4.322727117869951e-07,
"loss": 0.1048,
"step": 881
},
{
"epoch": 4.340713407134071,
"grad_norm": 1.15625,
"learning_rate": 4.2593782106558676e-07,
"loss": 0.0922,
"step": 882
},
{
"epoch": 4.345633456334563,
"grad_norm": 0.9765625,
"learning_rate": 4.1964762641353297e-07,
"loss": 0.0989,
"step": 883
},
{
"epoch": 4.350553505535055,
"grad_norm": 1.171875,
"learning_rate": 4.1340218929632636e-07,
"loss": 0.0922,
"step": 884
},
{
"epoch": 4.355473554735547,
"grad_norm": 1.078125,
"learning_rate": 4.072015707421006e-07,
"loss": 0.0831,
"step": 885
},
{
"epoch": 4.360393603936039,
"grad_norm": 1.2109375,
"learning_rate": 4.0104583134104593e-07,
"loss": 0.092,
"step": 886
},
{
"epoch": 4.365313653136531,
"grad_norm": 0.9921875,
"learning_rate": 3.9493503124480135e-07,
"loss": 0.0722,
"step": 887
},
{
"epoch": 4.370233702337023,
"grad_norm": 1.1796875,
"learning_rate": 3.8886923016588195e-07,
"loss": 0.0792,
"step": 888
},
{
"epoch": 4.375153751537516,
"grad_norm": 1.21875,
"learning_rate": 3.828484873770832e-07,
"loss": 0.0541,
"step": 889
},
{
"epoch": 4.380073800738008,
"grad_norm": 1.140625,
"learning_rate": 3.7687286171091355e-07,
"loss": 0.1026,
"step": 890
},
{
"epoch": 4.3849938499385,
"grad_norm": 1.0,
"learning_rate": 3.709424115590088e-07,
"loss": 0.0849,
"step": 891
},
{
"epoch": 4.389913899138992,
"grad_norm": 1.109375,
"learning_rate": 3.65057194871567e-07,
"loss": 0.098,
"step": 892
},
{
"epoch": 4.394833948339484,
"grad_norm": 0.8359375,
"learning_rate": 3.5921726915678247e-07,
"loss": 0.0659,
"step": 893
},
{
"epoch": 4.3997539975399755,
"grad_norm": 1.09375,
"learning_rate": 3.534226914802813e-07,
"loss": 0.0839,
"step": 894
},
{
"epoch": 4.4046740467404675,
"grad_norm": 0.94921875,
"learning_rate": 3.4767351846456744e-07,
"loss": 0.1014,
"step": 895
},
{
"epoch": 4.409594095940959,
"grad_norm": 1.296875,
"learning_rate": 3.4196980628846297e-07,
"loss": 0.0843,
"step": 896
},
{
"epoch": 4.414514145141451,
"grad_norm": 0.9609375,
"learning_rate": 3.3631161068656604e-07,
"loss": 0.0727,
"step": 897
},
{
"epoch": 4.419434194341943,
"grad_norm": 1.03125,
"learning_rate": 3.3069898694870373e-07,
"loss": 0.0771,
"step": 898
},
{
"epoch": 4.424354243542435,
"grad_norm": 1.21875,
"learning_rate": 3.2513198991939054e-07,
"loss": 0.0975,
"step": 899
},
{
"epoch": 4.429274292742927,
"grad_norm": 0.8984375,
"learning_rate": 3.196106739972926e-07,
"loss": 0.0649,
"step": 900
},
{
"epoch": 4.434194341943419,
"grad_norm": 1.015625,
"learning_rate": 3.1413509313469816e-07,
"loss": 0.0885,
"step": 901
},
{
"epoch": 4.439114391143911,
"grad_norm": 0.88671875,
"learning_rate": 3.087053008369889e-07,
"loss": 0.069,
"step": 902
},
{
"epoch": 4.444034440344403,
"grad_norm": 0.859375,
"learning_rate": 3.0332135016211794e-07,
"loss": 0.0539,
"step": 903
},
{
"epoch": 4.448954489544896,
"grad_norm": 0.97265625,
"learning_rate": 2.979832937200883e-07,
"loss": 0.0869,
"step": 904
},
{
"epoch": 4.453874538745388,
"grad_norm": 1.0,
"learning_rate": 2.9269118367244385e-07,
"loss": 0.0936,
"step": 905
},
{
"epoch": 4.45879458794588,
"grad_norm": 1.109375,
"learning_rate": 2.8744507173175564e-07,
"loss": 0.07,
"step": 906
},
{
"epoch": 4.463714637146372,
"grad_norm": 1.15625,
"learning_rate": 2.822450091611195e-07,
"loss": 0.0865,
"step": 907
},
{
"epoch": 4.468634686346864,
"grad_norm": 1.0078125,
"learning_rate": 2.770910467736532e-07,
"loss": 0.0797,
"step": 908
},
{
"epoch": 4.4735547355473555,
"grad_norm": 0.87890625,
"learning_rate": 2.719832349319973e-07,
"loss": 0.0591,
"step": 909
},
{
"epoch": 4.478474784747847,
"grad_norm": 0.9765625,
"learning_rate": 2.669216235478295e-07,
"loss": 0.0716,
"step": 910
},
{
"epoch": 4.483394833948339,
"grad_norm": 1.390625,
"learning_rate": 2.619062620813728e-07,
"loss": 0.0919,
"step": 911
},
{
"epoch": 4.488314883148831,
"grad_norm": 1.171875,
"learning_rate": 2.5693719954091257e-07,
"loss": 0.073,
"step": 912
},
{
"epoch": 4.493234932349323,
"grad_norm": 1.0,
"learning_rate": 2.520144844823169e-07,
"loss": 0.0683,
"step": 913
},
{
"epoch": 4.498154981549815,
"grad_norm": 1.1328125,
"learning_rate": 2.471381650085647e-07,
"loss": 0.0885,
"step": 914
},
{
"epoch": 4.503075030750307,
"grad_norm": 1.1640625,
"learning_rate": 2.4230828876927293e-07,
"loss": 0.0676,
"step": 915
},
{
"epoch": 4.507995079950799,
"grad_norm": 1.078125,
"learning_rate": 2.375249029602339e-07,
"loss": 0.0829,
"step": 916
},
{
"epoch": 4.512915129151292,
"grad_norm": 1.171875,
"learning_rate": 2.327880543229505e-07,
"loss": 0.0787,
"step": 917
},
{
"epoch": 4.517835178351784,
"grad_norm": 1.1171875,
"learning_rate": 2.2809778914418156e-07,
"loss": 0.0839,
"step": 918
},
{
"epoch": 4.517835178351784,
"eval_loss": 0.10323299467563629,
"eval_runtime": 203.6707,
"eval_samples_per_second": 8.033,
"eval_steps_per_second": 1.34,
"step": 918
},
{
"epoch": 4.522755227552276,
"grad_norm": 1.1875,
"learning_rate": 2.2345415325549125e-07,
"loss": 0.1202,
"step": 919
},
{
"epoch": 4.527675276752768,
"grad_norm": 0.90234375,
"learning_rate": 2.1885719203279587e-07,
"loss": 0.0706,
"step": 920
},
{
"epoch": 4.53259532595326,
"grad_norm": 0.953125,
"learning_rate": 2.143069503959283e-07,
"loss": 0.0835,
"step": 921
},
{
"epoch": 4.537515375153752,
"grad_norm": 1.140625,
"learning_rate": 2.0980347280818935e-07,
"loss": 0.0678,
"step": 922
},
{
"epoch": 4.5424354243542435,
"grad_norm": 0.98046875,
"learning_rate": 2.0534680327592426e-07,
"loss": 0.0662,
"step": 923
},
{
"epoch": 4.5473554735547355,
"grad_norm": 0.93359375,
"learning_rate": 2.009369853480825e-07,
"loss": 0.0582,
"step": 924
},
{
"epoch": 4.552275522755227,
"grad_norm": 1.09375,
"learning_rate": 1.9657406211579966e-07,
"loss": 0.1065,
"step": 925
},
{
"epoch": 4.557195571955719,
"grad_norm": 0.97265625,
"learning_rate": 1.922580762119697e-07,
"loss": 0.0513,
"step": 926
},
{
"epoch": 4.562115621156211,
"grad_norm": 1.1796875,
"learning_rate": 1.8798906981083832e-07,
"loss": 0.0734,
"step": 927
},
{
"epoch": 4.567035670356703,
"grad_norm": 1.03125,
"learning_rate": 1.8376708462757798e-07,
"loss": 0.0901,
"step": 928
},
{
"epoch": 4.571955719557195,
"grad_norm": 1.0703125,
"learning_rate": 1.7959216191789142e-07,
"loss": 0.0806,
"step": 929
},
{
"epoch": 4.576875768757688,
"grad_norm": 1.234375,
"learning_rate": 1.7546434247760147e-07,
"loss": 0.0858,
"step": 930
},
{
"epoch": 4.58179581795818,
"grad_norm": 1.0625,
"learning_rate": 1.713836666422569e-07,
"loss": 0.0808,
"step": 931
},
{
"epoch": 4.586715867158672,
"grad_norm": 0.984375,
"learning_rate": 1.673501742867356e-07,
"loss": 0.0951,
"step": 932
},
{
"epoch": 4.591635916359164,
"grad_norm": 1.0078125,
"learning_rate": 1.633639048248542e-07,
"loss": 0.0835,
"step": 933
},
{
"epoch": 4.596555965559656,
"grad_norm": 0.9140625,
"learning_rate": 1.594248972089879e-07,
"loss": 0.0686,
"step": 934
},
{
"epoch": 4.601476014760148,
"grad_norm": 1.109375,
"learning_rate": 1.555331899296808e-07,
"loss": 0.1139,
"step": 935
},
{
"epoch": 4.60639606396064,
"grad_norm": 1.15625,
"learning_rate": 1.5168882101528282e-07,
"loss": 0.0884,
"step": 936
},
{
"epoch": 4.6113161131611315,
"grad_norm": 1.1171875,
"learning_rate": 1.4789182803156333e-07,
"loss": 0.0855,
"step": 937
},
{
"epoch": 4.6162361623616235,
"grad_norm": 1.2109375,
"learning_rate": 1.44142248081357e-07,
"loss": 0.0854,
"step": 938
},
{
"epoch": 4.621156211562115,
"grad_norm": 1.3125,
"learning_rate": 1.4044011780419032e-07,
"loss": 0.0947,
"step": 939
},
{
"epoch": 4.626076260762607,
"grad_norm": 1.046875,
"learning_rate": 1.3678547337593494e-07,
"loss": 0.0637,
"step": 940
},
{
"epoch": 4.630996309963099,
"grad_norm": 1.140625,
"learning_rate": 1.3317835050844275e-07,
"loss": 0.0743,
"step": 941
},
{
"epoch": 4.635916359163591,
"grad_norm": 1.2265625,
"learning_rate": 1.296187844492053e-07,
"loss": 0.0928,
"step": 942
},
{
"epoch": 4.640836408364084,
"grad_norm": 1.140625,
"learning_rate": 1.2610680998100476e-07,
"loss": 0.0978,
"step": 943
},
{
"epoch": 4.645756457564576,
"grad_norm": 0.94921875,
"learning_rate": 1.2264246142157656e-07,
"loss": 0.0776,
"step": 944
},
{
"epoch": 4.650676506765068,
"grad_norm": 0.91796875,
"learning_rate": 1.1922577262327374e-07,
"loss": 0.0896,
"step": 945
},
{
"epoch": 4.65559655596556,
"grad_norm": 0.95703125,
"learning_rate": 1.1585677697273312e-07,
"loss": 0.0723,
"step": 946
},
{
"epoch": 4.660516605166052,
"grad_norm": 1.1796875,
"learning_rate": 1.1253550739055374e-07,
"loss": 0.0738,
"step": 947
},
{
"epoch": 4.665436654366544,
"grad_norm": 1.0390625,
"learning_rate": 1.0926199633097156e-07,
"loss": 0.0792,
"step": 948
},
{
"epoch": 4.670356703567036,
"grad_norm": 1.265625,
"learning_rate": 1.060362757815453e-07,
"loss": 0.0992,
"step": 949
},
{
"epoch": 4.675276752767528,
"grad_norm": 1.078125,
"learning_rate": 1.0285837726283999e-07,
"loss": 0.0638,
"step": 950
},
{
"epoch": 4.68019680196802,
"grad_norm": 1.0,
"learning_rate": 9.972833182812225e-08,
"loss": 0.0741,
"step": 951
},
{
"epoch": 4.6851168511685115,
"grad_norm": 1.4140625,
"learning_rate": 9.664617006305665e-08,
"loss": 0.1135,
"step": 952
},
{
"epoch": 4.6900369003690034,
"grad_norm": 0.90625,
"learning_rate": 9.361192208540427e-08,
"loss": 0.0603,
"step": 953
},
{
"epoch": 4.694956949569495,
"grad_norm": 1.0,
"learning_rate": 9.062561754473231e-08,
"loss": 0.0811,
"step": 954
},
{
"epoch": 4.699876998769987,
"grad_norm": 1.8984375,
"learning_rate": 8.768728562211948e-08,
"loss": 0.1155,
"step": 955
},
{
"epoch": 4.70479704797048,
"grad_norm": 0.9140625,
"learning_rate": 8.479695502987551e-08,
"loss": 0.0756,
"step": 956
},
{
"epoch": 4.709717097170972,
"grad_norm": 1.265625,
"learning_rate": 8.195465401125812e-08,
"loss": 0.1058,
"step": 957
},
{
"epoch": 4.714637146371464,
"grad_norm": 0.8359375,
"learning_rate": 7.916041034019773e-08,
"loss": 0.0538,
"step": 958
},
{
"epoch": 4.719557195571956,
"grad_norm": 1.0859375,
"learning_rate": 7.64142513210242e-08,
"loss": 0.0974,
"step": 959
},
{
"epoch": 4.724477244772448,
"grad_norm": 1.1484375,
"learning_rate": 7.371620378820555e-08,
"loss": 0.0975,
"step": 960
},
{
"epoch": 4.72939729397294,
"grad_norm": 0.953125,
"learning_rate": 7.10662941060769e-08,
"loss": 0.0675,
"step": 961
},
{
"epoch": 4.734317343173432,
"grad_norm": 1.1796875,
"learning_rate": 6.84645481685925e-08,
"loss": 0.081,
"step": 962
},
{
"epoch": 4.739237392373924,
"grad_norm": 0.9453125,
"learning_rate": 6.59109913990641e-08,
"loss": 0.0836,
"step": 963
},
{
"epoch": 4.744157441574416,
"grad_norm": 1.1640625,
"learning_rate": 6.340564874991906e-08,
"loss": 0.0772,
"step": 964
},
{
"epoch": 4.749077490774908,
"grad_norm": 1.3515625,
"learning_rate": 6.094854470245326e-08,
"loss": 0.1065,
"step": 965
},
{
"epoch": 4.7539975399753995,
"grad_norm": 1.046875,
"learning_rate": 5.853970326659186e-08,
"loss": 0.0749,
"step": 966
},
{
"epoch": 4.7589175891758915,
"grad_norm": 0.97265625,
"learning_rate": 5.6179147980656154e-08,
"loss": 0.0792,
"step": 967
},
{
"epoch": 4.763837638376383,
"grad_norm": 1.0390625,
"learning_rate": 5.3866901911132086e-08,
"loss": 0.0924,
"step": 968
},
{
"epoch": 4.768757687576876,
"grad_norm": 0.9453125,
"learning_rate": 5.160298765244709e-08,
"loss": 0.0824,
"step": 969
},
{
"epoch": 4.773677736777367,
"grad_norm": 0.96484375,
"learning_rate": 4.9387427326745287e-08,
"loss": 0.0622,
"step": 970
},
{
"epoch": 4.77859778597786,
"grad_norm": 0.85546875,
"learning_rate": 4.722024258367597e-08,
"loss": 0.0668,
"step": 971
},
{
"epoch": 4.783517835178352,
"grad_norm": 1.359375,
"learning_rate": 4.5101454600177676e-08,
"loss": 0.1176,
"step": 972
},
{
"epoch": 4.788437884378844,
"grad_norm": 1.2109375,
"learning_rate": 4.303108408027668e-08,
"loss": 0.1058,
"step": 973
},
{
"epoch": 4.793357933579336,
"grad_norm": 0.9921875,
"learning_rate": 4.1009151254878254e-08,
"loss": 0.091,
"step": 974
},
{
"epoch": 4.798277982779828,
"grad_norm": 0.88671875,
"learning_rate": 3.903567588157353e-08,
"loss": 0.0783,
"step": 975
},
{
"epoch": 4.80319803198032,
"grad_norm": 1.015625,
"learning_rate": 3.711067724444517e-08,
"loss": 0.0883,
"step": 976
},
{
"epoch": 4.808118081180812,
"grad_norm": 1.1640625,
"learning_rate": 3.523417415387864e-08,
"loss": 0.0801,
"step": 977
},
{
"epoch": 4.813038130381304,
"grad_norm": 1.0859375,
"learning_rate": 3.340618494637793e-08,
"loss": 0.0695,
"step": 978
},
{
"epoch": 4.817958179581796,
"grad_norm": 1.0234375,
"learning_rate": 3.162672748438844e-08,
"loss": 0.0959,
"step": 979
},
{
"epoch": 4.822878228782288,
"grad_norm": 1.21875,
"learning_rate": 2.989581915611994e-08,
"loss": 0.0944,
"step": 980
},
{
"epoch": 4.8277982779827795,
"grad_norm": 0.890625,
"learning_rate": 2.821347687537834e-08,
"loss": 0.0634,
"step": 981
},
{
"epoch": 4.832718327183272,
"grad_norm": 0.921875,
"learning_rate": 2.657971708139917e-08,
"loss": 0.0954,
"step": 982
},
{
"epoch": 4.837638376383763,
"grad_norm": 1.2265625,
"learning_rate": 2.4994555738688252e-08,
"loss": 0.0935,
"step": 983
},
{
"epoch": 4.842558425584256,
"grad_norm": 1.0859375,
"learning_rate": 2.3458008336864623e-08,
"loss": 0.0965,
"step": 984
},
{
"epoch": 4.847478474784748,
"grad_norm": 1.078125,
"learning_rate": 2.1970089890509527e-08,
"loss": 0.0944,
"step": 985
},
{
"epoch": 4.85239852398524,
"grad_norm": 1.21875,
"learning_rate": 2.0530814939020428e-08,
"loss": 0.0842,
"step": 986
},
{
"epoch": 4.857318573185732,
"grad_norm": 1.140625,
"learning_rate": 1.9140197546467787e-08,
"loss": 0.0857,
"step": 987
},
{
"epoch": 4.862238622386224,
"grad_norm": 1.046875,
"learning_rate": 1.7798251301458512e-08,
"loss": 0.0715,
"step": 988
},
{
"epoch": 4.867158671586716,
"grad_norm": 0.953125,
"learning_rate": 1.6504989317001618e-08,
"loss": 0.0868,
"step": 989
},
{
"epoch": 4.872078720787208,
"grad_norm": 1.2578125,
"learning_rate": 1.5260424230382763e-08,
"loss": 0.0842,
"step": 990
},
{
"epoch": 4.8769987699877,
"grad_norm": 1.1171875,
"learning_rate": 1.4064568203037699e-08,
"loss": 0.0964,
"step": 991
},
{
"epoch": 4.881918819188192,
"grad_norm": 1.203125,
"learning_rate": 1.2917432920437345e-08,
"loss": 0.089,
"step": 992
},
{
"epoch": 4.886838868388684,
"grad_norm": 0.875,
"learning_rate": 1.1819029591968456e-08,
"loss": 0.0584,
"step": 993
},
{
"epoch": 4.891758917589176,
"grad_norm": 1.09375,
"learning_rate": 1.0769368950829252e-08,
"loss": 0.0723,
"step": 994
},
{
"epoch": 4.8966789667896675,
"grad_norm": 1.0390625,
"learning_rate": 9.768461253920614e-09,
"loss": 0.1013,
"step": 995
},
{
"epoch": 4.9015990159901595,
"grad_norm": 0.9921875,
"learning_rate": 8.816316281747839e-09,
"loss": 0.0699,
"step": 996
},
{
"epoch": 4.906519065190652,
"grad_norm": 0.921875,
"learning_rate": 7.912943338324598e-09,
"loss": 0.0806,
"step": 997
},
{
"epoch": 4.911439114391144,
"grad_norm": 0.8359375,
"learning_rate": 7.058351251083007e-09,
"loss": 0.0705,
"step": 998
},
{
"epoch": 4.916359163591636,
"grad_norm": 1.0625,
"learning_rate": 6.252548370784817e-09,
"loss": 0.072,
"step": 999
},
{
"epoch": 4.921279212792128,
"grad_norm": 0.921875,
"learning_rate": 5.495542571443135e-09,
"loss": 0.0715,
"step": 1000
},
{
"epoch": 4.92619926199262,
"grad_norm": 0.9453125,
"learning_rate": 4.787341250241384e-09,
"loss": 0.0838,
"step": 1001
},
{
"epoch": 4.931119311193112,
"grad_norm": 0.890625,
"learning_rate": 4.127951327466684e-09,
"loss": 0.0672,
"step": 1002
},
{
"epoch": 4.936039360393604,
"grad_norm": 1.0703125,
"learning_rate": 3.5173792464360256e-09,
"loss": 0.0753,
"step": 1003
},
{
"epoch": 4.940959409594096,
"grad_norm": 0.94140625,
"learning_rate": 2.955630973437429e-09,
"loss": 0.0858,
"step": 1004
},
{
"epoch": 4.945879458794588,
"grad_norm": 0.83984375,
"learning_rate": 2.442711997670544e-09,
"loss": 0.0674,
"step": 1005
},
{
"epoch": 4.95079950799508,
"grad_norm": 1.0,
"learning_rate": 1.978627331192806e-09,
"loss": 0.0892,
"step": 1006
},
{
"epoch": 4.955719557195572,
"grad_norm": 1.0625,
"learning_rate": 1.5633815088705872e-09,
"loss": 0.074,
"step": 1007
},
{
"epoch": 4.960639606396064,
"grad_norm": 1.109375,
"learning_rate": 1.196978588334785e-09,
"loss": 0.0951,
"step": 1008
},
{
"epoch": 4.9655596555965555,
"grad_norm": 1.1796875,
"learning_rate": 8.794221499408562e-10,
"loss": 0.0977,
"step": 1009
},
{
"epoch": 4.970479704797048,
"grad_norm": 1.2578125,
"learning_rate": 6.107152967349539e-10,
"loss": 0.1068,
"step": 1010
},
{
"epoch": 4.97539975399754,
"grad_norm": 1.0078125,
"learning_rate": 3.908606544228422e-10,
"loss": 0.0803,
"step": 1011
},
{
"epoch": 4.980319803198032,
"grad_norm": 0.88671875,
"learning_rate": 2.1986037134325012e-10,
"loss": 0.0667,
"step": 1012
},
{
"epoch": 4.985239852398524,
"grad_norm": 0.86328125,
"learning_rate": 9.771611844955343e-11,
"loss": 0.0616,
"step": 1013
},
{
"epoch": 4.990159901599016,
"grad_norm": 1.0546875,
"learning_rate": 2.4429089290345375e-11,
"loss": 0.088,
"step": 1014
},
{
"epoch": 4.995079950799508,
"grad_norm": 1.03125,
"learning_rate": 0.0,
"loss": 0.1053,
"step": 1015
}
],
"logging_steps": 1,
"max_steps": 1015,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 51,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.1980182581321662e+19,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}