adammandic87's picture
Training in progress, step 200, checkpoint
8dda745 verified
raw
history blame
36.8 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.13508949679162446,
"eval_steps": 50,
"global_step": 200,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0006754474839581223,
"grad_norm": 0.32079648971557617,
"learning_rate": 2e-05,
"loss": 10.8491,
"step": 1
},
{
"epoch": 0.0006754474839581223,
"eval_loss": 10.850319862365723,
"eval_runtime": 2.6665,
"eval_samples_per_second": 234.019,
"eval_steps_per_second": 117.009,
"step": 1
},
{
"epoch": 0.0013508949679162446,
"grad_norm": 0.32938364148139954,
"learning_rate": 4e-05,
"loss": 10.8557,
"step": 2
},
{
"epoch": 0.002026342451874367,
"grad_norm": 0.3332866430282593,
"learning_rate": 6e-05,
"loss": 10.8487,
"step": 3
},
{
"epoch": 0.002701789935832489,
"grad_norm": 0.3259293735027313,
"learning_rate": 8e-05,
"loss": 10.8485,
"step": 4
},
{
"epoch": 0.003377237419790611,
"grad_norm": 0.33600881695747375,
"learning_rate": 0.0001,
"loss": 10.8459,
"step": 5
},
{
"epoch": 0.004052684903748734,
"grad_norm": 0.3575391173362732,
"learning_rate": 0.00012,
"loss": 10.8517,
"step": 6
},
{
"epoch": 0.004728132387706856,
"grad_norm": 0.33056777715682983,
"learning_rate": 0.00014,
"loss": 10.8459,
"step": 7
},
{
"epoch": 0.005403579871664978,
"grad_norm": 0.35549676418304443,
"learning_rate": 0.00016,
"loss": 10.8443,
"step": 8
},
{
"epoch": 0.0060790273556231,
"grad_norm": 0.34090492129325867,
"learning_rate": 0.00018,
"loss": 10.8468,
"step": 9
},
{
"epoch": 0.006754474839581222,
"grad_norm": 0.34161046147346497,
"learning_rate": 0.0002,
"loss": 10.8432,
"step": 10
},
{
"epoch": 0.007429922323539345,
"grad_norm": 0.3733224868774414,
"learning_rate": 0.0001999863304992469,
"loss": 10.8395,
"step": 11
},
{
"epoch": 0.008105369807497468,
"grad_norm": 0.37448549270629883,
"learning_rate": 0.00019994532573409262,
"loss": 10.8345,
"step": 12
},
{
"epoch": 0.008780817291455589,
"grad_norm": 0.39854905009269714,
"learning_rate": 0.00019987699691483048,
"loss": 10.8358,
"step": 13
},
{
"epoch": 0.009456264775413711,
"grad_norm": 0.4029870629310608,
"learning_rate": 0.00019978136272187747,
"loss": 10.8312,
"step": 14
},
{
"epoch": 0.010131712259371834,
"grad_norm": 0.3861730396747589,
"learning_rate": 0.000199658449300667,
"loss": 10.827,
"step": 15
},
{
"epoch": 0.010807159743329957,
"grad_norm": 0.4259829819202423,
"learning_rate": 0.00019950829025450114,
"loss": 10.8214,
"step": 16
},
{
"epoch": 0.011482607227288078,
"grad_norm": 0.4038204252719879,
"learning_rate": 0.00019933092663536382,
"loss": 10.8221,
"step": 17
},
{
"epoch": 0.0121580547112462,
"grad_norm": 0.4385932385921478,
"learning_rate": 0.00019912640693269752,
"loss": 10.8208,
"step": 18
},
{
"epoch": 0.012833502195204323,
"grad_norm": 0.4669629633426666,
"learning_rate": 0.00019889478706014687,
"loss": 10.8162,
"step": 19
},
{
"epoch": 0.013508949679162444,
"grad_norm": 0.45796719193458557,
"learning_rate": 0.00019863613034027224,
"loss": 10.8103,
"step": 20
},
{
"epoch": 0.014184397163120567,
"grad_norm": 0.46000435948371887,
"learning_rate": 0.00019835050748723824,
"loss": 10.8013,
"step": 21
},
{
"epoch": 0.01485984464707869,
"grad_norm": 0.4528140425682068,
"learning_rate": 0.00019803799658748094,
"loss": 10.7983,
"step": 22
},
{
"epoch": 0.015535292131036813,
"grad_norm": 0.5075253248214722,
"learning_rate": 0.00019769868307835994,
"loss": 10.7945,
"step": 23
},
{
"epoch": 0.016210739614994935,
"grad_norm": 0.49094441533088684,
"learning_rate": 0.0001973326597248006,
"loss": 10.7873,
"step": 24
},
{
"epoch": 0.016886187098953058,
"grad_norm": 0.4418376684188843,
"learning_rate": 0.00019694002659393305,
"loss": 10.7869,
"step": 25
},
{
"epoch": 0.017561634582911177,
"grad_norm": 0.4543623924255371,
"learning_rate": 0.00019652089102773488,
"loss": 10.7771,
"step": 26
},
{
"epoch": 0.0182370820668693,
"grad_norm": 0.514811098575592,
"learning_rate": 0.00019607536761368484,
"loss": 10.7833,
"step": 27
},
{
"epoch": 0.018912529550827423,
"grad_norm": 0.5044733881950378,
"learning_rate": 0.00019560357815343577,
"loss": 10.7702,
"step": 28
},
{
"epoch": 0.019587977034785545,
"grad_norm": 0.5136566162109375,
"learning_rate": 0.00019510565162951537,
"loss": 10.7711,
"step": 29
},
{
"epoch": 0.020263424518743668,
"grad_norm": 0.5506525039672852,
"learning_rate": 0.00019458172417006347,
"loss": 10.7634,
"step": 30
},
{
"epoch": 0.02093887200270179,
"grad_norm": 0.5305942893028259,
"learning_rate": 0.00019403193901161613,
"loss": 10.7524,
"step": 31
},
{
"epoch": 0.021614319486659914,
"grad_norm": 0.5339845418930054,
"learning_rate": 0.0001934564464599461,
"loss": 10.7551,
"step": 32
},
{
"epoch": 0.022289766970618033,
"grad_norm": 0.5734214186668396,
"learning_rate": 0.00019285540384897073,
"loss": 10.7459,
"step": 33
},
{
"epoch": 0.022965214454576156,
"grad_norm": 0.5622735023498535,
"learning_rate": 0.00019222897549773848,
"loss": 10.7351,
"step": 34
},
{
"epoch": 0.02364066193853428,
"grad_norm": 0.5738793015480042,
"learning_rate": 0.00019157733266550575,
"loss": 10.7313,
"step": 35
},
{
"epoch": 0.0243161094224924,
"grad_norm": 0.560500979423523,
"learning_rate": 0.00019090065350491626,
"loss": 10.7251,
"step": 36
},
{
"epoch": 0.024991556906450524,
"grad_norm": 0.5471153259277344,
"learning_rate": 0.00019019912301329592,
"loss": 10.7093,
"step": 37
},
{
"epoch": 0.025667004390408647,
"grad_norm": 0.5753123760223389,
"learning_rate": 0.00018947293298207635,
"loss": 10.7115,
"step": 38
},
{
"epoch": 0.02634245187436677,
"grad_norm": 0.5051413774490356,
"learning_rate": 0.0001887222819443612,
"loss": 10.7161,
"step": 39
},
{
"epoch": 0.02701789935832489,
"grad_norm": 0.5438815355300903,
"learning_rate": 0.0001879473751206489,
"loss": 10.6966,
"step": 40
},
{
"epoch": 0.02769334684228301,
"grad_norm": 0.5222083330154419,
"learning_rate": 0.00018714842436272773,
"loss": 10.6916,
"step": 41
},
{
"epoch": 0.028368794326241134,
"grad_norm": 0.4598117172718048,
"learning_rate": 0.00018632564809575742,
"loss": 10.7039,
"step": 42
},
{
"epoch": 0.029044241810199257,
"grad_norm": 0.517598032951355,
"learning_rate": 0.0001854792712585539,
"loss": 10.6926,
"step": 43
},
{
"epoch": 0.02971968929415738,
"grad_norm": 0.5157806873321533,
"learning_rate": 0.00018460952524209355,
"loss": 10.685,
"step": 44
},
{
"epoch": 0.030395136778115502,
"grad_norm": 0.4841625690460205,
"learning_rate": 0.00018371664782625287,
"loss": 10.6855,
"step": 45
},
{
"epoch": 0.031070584262073625,
"grad_norm": 0.47345930337905884,
"learning_rate": 0.00018280088311480201,
"loss": 10.6667,
"step": 46
},
{
"epoch": 0.031746031746031744,
"grad_norm": 0.3876435458660126,
"learning_rate": 0.00018186248146866927,
"loss": 10.699,
"step": 47
},
{
"epoch": 0.03242147922998987,
"grad_norm": 0.43177780508995056,
"learning_rate": 0.00018090169943749476,
"loss": 10.6633,
"step": 48
},
{
"epoch": 0.03309692671394799,
"grad_norm": 0.4080864489078522,
"learning_rate": 0.0001799187996894925,
"loss": 10.6654,
"step": 49
},
{
"epoch": 0.033772374197906116,
"grad_norm": 0.40777599811553955,
"learning_rate": 0.00017891405093963938,
"loss": 10.659,
"step": 50
},
{
"epoch": 0.033772374197906116,
"eval_loss": 10.658723831176758,
"eval_runtime": 2.2172,
"eval_samples_per_second": 281.442,
"eval_steps_per_second": 140.721,
"step": 50
},
{
"epoch": 0.034447821681864235,
"grad_norm": 0.37587037682533264,
"learning_rate": 0.00017788772787621126,
"loss": 10.6597,
"step": 51
},
{
"epoch": 0.035123269165822354,
"grad_norm": 0.35869449377059937,
"learning_rate": 0.00017684011108568592,
"loss": 10.6479,
"step": 52
},
{
"epoch": 0.03579871664978048,
"grad_norm": 0.3892292380332947,
"learning_rate": 0.0001757714869760335,
"loss": 10.651,
"step": 53
},
{
"epoch": 0.0364741641337386,
"grad_norm": 0.372700572013855,
"learning_rate": 0.0001746821476984154,
"loss": 10.6507,
"step": 54
},
{
"epoch": 0.037149611617696726,
"grad_norm": 0.3261224329471588,
"learning_rate": 0.00017357239106731317,
"loss": 10.6657,
"step": 55
},
{
"epoch": 0.037825059101654845,
"grad_norm": 0.3149575889110565,
"learning_rate": 0.00017244252047910892,
"loss": 10.6366,
"step": 56
},
{
"epoch": 0.03850050658561297,
"grad_norm": 0.3464124798774719,
"learning_rate": 0.00017129284482913972,
"loss": 10.6405,
"step": 57
},
{
"epoch": 0.03917595406957109,
"grad_norm": 0.291887491941452,
"learning_rate": 0.00017012367842724887,
"loss": 10.6374,
"step": 58
},
{
"epoch": 0.03985140155352921,
"grad_norm": 0.2717147767543793,
"learning_rate": 0.0001689353409118566,
"loss": 10.6367,
"step": 59
},
{
"epoch": 0.040526849037487336,
"grad_norm": 0.27471521496772766,
"learning_rate": 0.00016772815716257412,
"loss": 10.6394,
"step": 60
},
{
"epoch": 0.041202296521445456,
"grad_norm": 0.26729482412338257,
"learning_rate": 0.0001665024572113848,
"loss": 10.6301,
"step": 61
},
{
"epoch": 0.04187774400540358,
"grad_norm": 0.25519606471061707,
"learning_rate": 0.00016525857615241687,
"loss": 10.6343,
"step": 62
},
{
"epoch": 0.0425531914893617,
"grad_norm": 0.25052082538604736,
"learning_rate": 0.00016399685405033167,
"loss": 10.6282,
"step": 63
},
{
"epoch": 0.04322863897331983,
"grad_norm": 0.24196283519268036,
"learning_rate": 0.0001627176358473537,
"loss": 10.6253,
"step": 64
},
{
"epoch": 0.04390408645727795,
"grad_norm": 0.24466153979301453,
"learning_rate": 0.0001614212712689668,
"loss": 10.6304,
"step": 65
},
{
"epoch": 0.044579533941236066,
"grad_norm": 0.21384331583976746,
"learning_rate": 0.00016010811472830252,
"loss": 10.6279,
"step": 66
},
{
"epoch": 0.04525498142519419,
"grad_norm": 0.23176230490207672,
"learning_rate": 0.00015877852522924732,
"loss": 10.6254,
"step": 67
},
{
"epoch": 0.04593042890915231,
"grad_norm": 0.21549372375011444,
"learning_rate": 0.00015743286626829437,
"loss": 10.6254,
"step": 68
},
{
"epoch": 0.04660587639311044,
"grad_norm": 0.21212856471538544,
"learning_rate": 0.0001560715057351673,
"loss": 10.6306,
"step": 69
},
{
"epoch": 0.04728132387706856,
"grad_norm": 0.2017771154642105,
"learning_rate": 0.00015469481581224272,
"loss": 10.6216,
"step": 70
},
{
"epoch": 0.04795677136102668,
"grad_norm": 0.21862851083278656,
"learning_rate": 0.0001533031728727994,
"loss": 10.6099,
"step": 71
},
{
"epoch": 0.0486322188449848,
"grad_norm": 0.1980670690536499,
"learning_rate": 0.00015189695737812152,
"loss": 10.6245,
"step": 72
},
{
"epoch": 0.04930766632894292,
"grad_norm": 0.2036397010087967,
"learning_rate": 0.0001504765537734844,
"loss": 10.6232,
"step": 73
},
{
"epoch": 0.04998311381290105,
"grad_norm": 0.17732380330562592,
"learning_rate": 0.00014904235038305083,
"loss": 10.6244,
"step": 74
},
{
"epoch": 0.05065856129685917,
"grad_norm": 0.19181127846240997,
"learning_rate": 0.00014759473930370736,
"loss": 10.6222,
"step": 75
},
{
"epoch": 0.05133400878081729,
"grad_norm": 0.1601181924343109,
"learning_rate": 0.0001461341162978688,
"loss": 10.6218,
"step": 76
},
{
"epoch": 0.05200945626477541,
"grad_norm": 0.18116536736488342,
"learning_rate": 0.00014466088068528068,
"loss": 10.625,
"step": 77
},
{
"epoch": 0.05268490374873354,
"grad_norm": 0.14323946833610535,
"learning_rate": 0.00014317543523384928,
"loss": 10.6489,
"step": 78
},
{
"epoch": 0.05336035123269166,
"grad_norm": 0.1809110790491104,
"learning_rate": 0.00014167818604952906,
"loss": 10.6282,
"step": 79
},
{
"epoch": 0.05403579871664978,
"grad_norm": 0.13735494017601013,
"learning_rate": 0.00014016954246529696,
"loss": 10.6169,
"step": 80
},
{
"epoch": 0.0547112462006079,
"grad_norm": 0.15906408429145813,
"learning_rate": 0.00013864991692924523,
"loss": 10.6212,
"step": 81
},
{
"epoch": 0.05538669368456602,
"grad_norm": 0.17548221349716187,
"learning_rate": 0.00013711972489182208,
"loss": 10.6171,
"step": 82
},
{
"epoch": 0.05606214116852415,
"grad_norm": 0.12418357282876968,
"learning_rate": 0.00013557938469225167,
"loss": 10.6143,
"step": 83
},
{
"epoch": 0.05673758865248227,
"grad_norm": 0.1521899700164795,
"learning_rate": 0.00013402931744416433,
"loss": 10.6199,
"step": 84
},
{
"epoch": 0.057413036136440394,
"grad_norm": 0.14142099022865295,
"learning_rate": 0.00013246994692046836,
"loss": 10.6131,
"step": 85
},
{
"epoch": 0.058088483620398514,
"grad_norm": 0.13722097873687744,
"learning_rate": 0.00013090169943749476,
"loss": 10.6153,
"step": 86
},
{
"epoch": 0.05876393110435663,
"grad_norm": 0.14299722015857697,
"learning_rate": 0.0001293250037384465,
"loss": 10.6094,
"step": 87
},
{
"epoch": 0.05943937858831476,
"grad_norm": 0.13671687245368958,
"learning_rate": 0.00012774029087618446,
"loss": 10.6242,
"step": 88
},
{
"epoch": 0.06011482607227288,
"grad_norm": 0.12905743718147278,
"learning_rate": 0.00012614799409538198,
"loss": 10.6116,
"step": 89
},
{
"epoch": 0.060790273556231005,
"grad_norm": 0.13734190165996552,
"learning_rate": 0.00012454854871407994,
"loss": 10.6142,
"step": 90
},
{
"epoch": 0.061465721040189124,
"grad_norm": 0.13106867671012878,
"learning_rate": 0.00012294239200467516,
"loss": 10.6176,
"step": 91
},
{
"epoch": 0.06214116852414725,
"grad_norm": 0.1443423628807068,
"learning_rate": 0.0001213299630743747,
"loss": 10.6369,
"step": 92
},
{
"epoch": 0.06281661600810537,
"grad_norm": 0.11222351342439651,
"learning_rate": 0.00011971170274514802,
"loss": 10.6164,
"step": 93
},
{
"epoch": 0.06349206349206349,
"grad_norm": 0.11618170142173767,
"learning_rate": 0.000118088053433211,
"loss": 10.6057,
"step": 94
},
{
"epoch": 0.06416751097602161,
"grad_norm": 0.11936385929584503,
"learning_rate": 0.00011645945902807341,
"loss": 10.6146,
"step": 95
},
{
"epoch": 0.06484295845997974,
"grad_norm": 0.12773548066616058,
"learning_rate": 0.0001148263647711842,
"loss": 10.6161,
"step": 96
},
{
"epoch": 0.06551840594393786,
"grad_norm": 0.1389545202255249,
"learning_rate": 0.00011318921713420691,
"loss": 10.6053,
"step": 97
},
{
"epoch": 0.06619385342789598,
"grad_norm": 0.13642248511314392,
"learning_rate": 0.00011154846369695863,
"loss": 10.6153,
"step": 98
},
{
"epoch": 0.0668693009118541,
"grad_norm": 0.11348054558038712,
"learning_rate": 0.0001099045530250463,
"loss": 10.6132,
"step": 99
},
{
"epoch": 0.06754474839581223,
"grad_norm": 0.1491929590702057,
"learning_rate": 0.00010825793454723325,
"loss": 10.6309,
"step": 100
},
{
"epoch": 0.06754474839581223,
"eval_loss": 10.612899780273438,
"eval_runtime": 2.3146,
"eval_samples_per_second": 269.593,
"eval_steps_per_second": 134.797,
"step": 100
},
{
"epoch": 0.06822019587977035,
"grad_norm": 0.11731832474470139,
"learning_rate": 0.00010660905843256994,
"loss": 10.6064,
"step": 101
},
{
"epoch": 0.06889564336372847,
"grad_norm": 0.1060105562210083,
"learning_rate": 0.00010495837546732224,
"loss": 10.6059,
"step": 102
},
{
"epoch": 0.06957109084768659,
"grad_norm": 0.12271567434072495,
"learning_rate": 0.00010330633693173082,
"loss": 10.6046,
"step": 103
},
{
"epoch": 0.07024653833164471,
"grad_norm": 0.1422068178653717,
"learning_rate": 0.00010165339447663587,
"loss": 10.6122,
"step": 104
},
{
"epoch": 0.07092198581560284,
"grad_norm": 0.123976930975914,
"learning_rate": 0.0001,
"loss": 10.6086,
"step": 105
},
{
"epoch": 0.07159743329956096,
"grad_norm": 0.11725173145532608,
"learning_rate": 9.834660552336415e-05,
"loss": 10.6312,
"step": 106
},
{
"epoch": 0.07227288078351908,
"grad_norm": 0.13431864976882935,
"learning_rate": 9.669366306826919e-05,
"loss": 10.6129,
"step": 107
},
{
"epoch": 0.0729483282674772,
"grad_norm": 0.10811913758516312,
"learning_rate": 9.504162453267777e-05,
"loss": 10.6102,
"step": 108
},
{
"epoch": 0.07362377575143532,
"grad_norm": 0.1448288857936859,
"learning_rate": 9.339094156743007e-05,
"loss": 10.6097,
"step": 109
},
{
"epoch": 0.07429922323539345,
"grad_norm": 0.14231544733047485,
"learning_rate": 9.174206545276677e-05,
"loss": 10.6149,
"step": 110
},
{
"epoch": 0.07497467071935157,
"grad_norm": 0.11350355297327042,
"learning_rate": 9.009544697495374e-05,
"loss": 10.6123,
"step": 111
},
{
"epoch": 0.07565011820330969,
"grad_norm": 0.11213658004999161,
"learning_rate": 8.845153630304139e-05,
"loss": 10.6059,
"step": 112
},
{
"epoch": 0.07632556568726781,
"grad_norm": 0.12278730422258377,
"learning_rate": 8.681078286579311e-05,
"loss": 10.6049,
"step": 113
},
{
"epoch": 0.07700101317122594,
"grad_norm": 0.15146122872829437,
"learning_rate": 8.517363522881579e-05,
"loss": 10.6104,
"step": 114
},
{
"epoch": 0.07767646065518406,
"grad_norm": 0.11578516662120819,
"learning_rate": 8.35405409719266e-05,
"loss": 10.6276,
"step": 115
},
{
"epoch": 0.07835190813914218,
"grad_norm": 0.13298995792865753,
"learning_rate": 8.191194656678904e-05,
"loss": 10.6115,
"step": 116
},
{
"epoch": 0.0790273556231003,
"grad_norm": 0.13557101786136627,
"learning_rate": 8.028829725485199e-05,
"loss": 10.6083,
"step": 117
},
{
"epoch": 0.07970280310705842,
"grad_norm": 0.12159471958875656,
"learning_rate": 7.867003692562534e-05,
"loss": 10.6037,
"step": 118
},
{
"epoch": 0.08037825059101655,
"grad_norm": 0.12765274941921234,
"learning_rate": 7.705760799532485e-05,
"loss": 10.6016,
"step": 119
},
{
"epoch": 0.08105369807497467,
"grad_norm": 0.11235444992780685,
"learning_rate": 7.54514512859201e-05,
"loss": 10.6065,
"step": 120
},
{
"epoch": 0.08172914555893279,
"grad_norm": 0.14903466403484344,
"learning_rate": 7.385200590461803e-05,
"loss": 10.5995,
"step": 121
},
{
"epoch": 0.08240459304289091,
"grad_norm": 0.10130941867828369,
"learning_rate": 7.225970912381556e-05,
"loss": 10.6095,
"step": 122
},
{
"epoch": 0.08308004052684904,
"grad_norm": 0.1442113071680069,
"learning_rate": 7.067499626155354e-05,
"loss": 10.6024,
"step": 123
},
{
"epoch": 0.08375548801080716,
"grad_norm": 0.14047802984714508,
"learning_rate": 6.909830056250527e-05,
"loss": 10.6104,
"step": 124
},
{
"epoch": 0.08443093549476528,
"grad_norm": 0.1147422268986702,
"learning_rate": 6.753005307953167e-05,
"loss": 10.6125,
"step": 125
},
{
"epoch": 0.0851063829787234,
"grad_norm": 0.11507318913936615,
"learning_rate": 6.59706825558357e-05,
"loss": 10.6038,
"step": 126
},
{
"epoch": 0.08578183046268152,
"grad_norm": 0.13589414954185486,
"learning_rate": 6.442061530774834e-05,
"loss": 10.6044,
"step": 127
},
{
"epoch": 0.08645727794663965,
"grad_norm": 0.13625499606132507,
"learning_rate": 6.28802751081779e-05,
"loss": 10.6058,
"step": 128
},
{
"epoch": 0.08713272543059777,
"grad_norm": 0.12005976587533951,
"learning_rate": 6.135008307075481e-05,
"loss": 10.62,
"step": 129
},
{
"epoch": 0.0878081729145559,
"grad_norm": 0.11311294883489609,
"learning_rate": 5.983045753470308e-05,
"loss": 10.6092,
"step": 130
},
{
"epoch": 0.08848362039851401,
"grad_norm": 0.13194864988327026,
"learning_rate": 5.832181395047098e-05,
"loss": 10.5994,
"step": 131
},
{
"epoch": 0.08915906788247213,
"grad_norm": 0.12505017220973969,
"learning_rate": 5.6824564766150726e-05,
"loss": 10.6069,
"step": 132
},
{
"epoch": 0.08983451536643026,
"grad_norm": 0.13057582080364227,
"learning_rate": 5.533911931471936e-05,
"loss": 10.5994,
"step": 133
},
{
"epoch": 0.09050996285038838,
"grad_norm": 0.17442747950553894,
"learning_rate": 5.386588370213124e-05,
"loss": 10.6025,
"step": 134
},
{
"epoch": 0.0911854103343465,
"grad_norm": 0.12355850636959076,
"learning_rate": 5.240526069629265e-05,
"loss": 10.6065,
"step": 135
},
{
"epoch": 0.09186085781830462,
"grad_norm": 0.13733559846878052,
"learning_rate": 5.095764961694922e-05,
"loss": 10.5974,
"step": 136
},
{
"epoch": 0.09253630530226276,
"grad_norm": 0.14686278998851776,
"learning_rate": 4.952344622651566e-05,
"loss": 10.5941,
"step": 137
},
{
"epoch": 0.09321175278622088,
"grad_norm": 0.13458283245563507,
"learning_rate": 4.810304262187852e-05,
"loss": 10.5959,
"step": 138
},
{
"epoch": 0.093887200270179,
"grad_norm": 0.1384890228509903,
"learning_rate": 4.669682712720065e-05,
"loss": 10.6081,
"step": 139
},
{
"epoch": 0.09456264775413711,
"grad_norm": 0.15113620460033417,
"learning_rate": 4.530518418775733e-05,
"loss": 10.6194,
"step": 140
},
{
"epoch": 0.09523809523809523,
"grad_norm": 0.14613160490989685,
"learning_rate": 4.392849426483274e-05,
"loss": 10.6113,
"step": 141
},
{
"epoch": 0.09591354272205337,
"grad_norm": 0.12903277575969696,
"learning_rate": 4.256713373170564e-05,
"loss": 10.6009,
"step": 142
},
{
"epoch": 0.09658899020601149,
"grad_norm": 0.15327778458595276,
"learning_rate": 4.12214747707527e-05,
"loss": 10.5978,
"step": 143
},
{
"epoch": 0.0972644376899696,
"grad_norm": 0.14414553344249725,
"learning_rate": 3.9891885271697496e-05,
"loss": 10.5922,
"step": 144
},
{
"epoch": 0.09793988517392772,
"grad_norm": 0.15006524324417114,
"learning_rate": 3.857872873103322e-05,
"loss": 10.6253,
"step": 145
},
{
"epoch": 0.09861533265788584,
"grad_norm": 0.14634265005588531,
"learning_rate": 3.7282364152646297e-05,
"loss": 10.6011,
"step": 146
},
{
"epoch": 0.09929078014184398,
"grad_norm": 0.14320406317710876,
"learning_rate": 3.600314594966834e-05,
"loss": 10.5892,
"step": 147
},
{
"epoch": 0.0999662276258021,
"grad_norm": 0.14602705836296082,
"learning_rate": 3.4741423847583134e-05,
"loss": 10.5927,
"step": 148
},
{
"epoch": 0.10064167510976021,
"grad_norm": 0.14189378917217255,
"learning_rate": 3.349754278861517e-05,
"loss": 10.5985,
"step": 149
},
{
"epoch": 0.10131712259371833,
"grad_norm": 0.15539740025997162,
"learning_rate": 3.227184283742591e-05,
"loss": 10.5968,
"step": 150
},
{
"epoch": 0.10131712259371833,
"eval_loss": 10.599814414978027,
"eval_runtime": 2.2579,
"eval_samples_per_second": 276.357,
"eval_steps_per_second": 138.179,
"step": 150
},
{
"epoch": 0.10199257007767647,
"grad_norm": 0.1630678027868271,
"learning_rate": 3.106465908814342e-05,
"loss": 10.6025,
"step": 151
},
{
"epoch": 0.10266801756163459,
"grad_norm": 0.15854433178901672,
"learning_rate": 2.9876321572751144e-05,
"loss": 10.589,
"step": 152
},
{
"epoch": 0.1033434650455927,
"grad_norm": 0.15652941167354584,
"learning_rate": 2.87071551708603e-05,
"loss": 10.6022,
"step": 153
},
{
"epoch": 0.10401891252955082,
"grad_norm": 0.14386983215808868,
"learning_rate": 2.7557479520891104e-05,
"loss": 10.5974,
"step": 154
},
{
"epoch": 0.10469436001350894,
"grad_norm": 0.17749740183353424,
"learning_rate": 2.6427608932686843e-05,
"loss": 10.6112,
"step": 155
},
{
"epoch": 0.10536980749746708,
"grad_norm": 0.1529720574617386,
"learning_rate": 2.5317852301584643e-05,
"loss": 10.6075,
"step": 156
},
{
"epoch": 0.1060452549814252,
"grad_norm": 0.16465173661708832,
"learning_rate": 2.422851302396655e-05,
"loss": 10.5917,
"step": 157
},
{
"epoch": 0.10672070246538332,
"grad_norm": 0.148993581533432,
"learning_rate": 2.315988891431412e-05,
"loss": 10.5951,
"step": 158
},
{
"epoch": 0.10739614994934144,
"grad_norm": 0.17219752073287964,
"learning_rate": 2.2112272123788768e-05,
"loss": 10.5877,
"step": 159
},
{
"epoch": 0.10807159743329955,
"grad_norm": 0.17853744328022003,
"learning_rate": 2.1085949060360654e-05,
"loss": 10.5945,
"step": 160
},
{
"epoch": 0.10874704491725769,
"grad_norm": 0.16907915472984314,
"learning_rate": 2.008120031050753e-05,
"loss": 10.5967,
"step": 161
},
{
"epoch": 0.1094224924012158,
"grad_norm": 0.18723782896995544,
"learning_rate": 1.9098300562505266e-05,
"loss": 10.5955,
"step": 162
},
{
"epoch": 0.11009793988517393,
"grad_norm": 0.18115279078483582,
"learning_rate": 1.8137518531330767e-05,
"loss": 10.6003,
"step": 163
},
{
"epoch": 0.11077338736913205,
"grad_norm": 0.14582610130310059,
"learning_rate": 1.7199116885197995e-05,
"loss": 10.6084,
"step": 164
},
{
"epoch": 0.11144883485309018,
"grad_norm": 0.16853328049182892,
"learning_rate": 1.6283352173747145e-05,
"loss": 10.6109,
"step": 165
},
{
"epoch": 0.1121242823370483,
"grad_norm": 0.16580811142921448,
"learning_rate": 1.5390474757906446e-05,
"loss": 10.6031,
"step": 166
},
{
"epoch": 0.11279972982100642,
"grad_norm": 0.1782258301973343,
"learning_rate": 1.4520728741446089e-05,
"loss": 10.5978,
"step": 167
},
{
"epoch": 0.11347517730496454,
"grad_norm": 0.16206084191799164,
"learning_rate": 1.3674351904242611e-05,
"loss": 10.5997,
"step": 168
},
{
"epoch": 0.11415062478892266,
"grad_norm": 0.1567702740430832,
"learning_rate": 1.2851575637272262e-05,
"loss": 10.6037,
"step": 169
},
{
"epoch": 0.11482607227288079,
"grad_norm": 0.174870565533638,
"learning_rate": 1.2052624879351104e-05,
"loss": 10.5948,
"step": 170
},
{
"epoch": 0.11550151975683891,
"grad_norm": 0.15422426164150238,
"learning_rate": 1.1277718055638819e-05,
"loss": 10.5984,
"step": 171
},
{
"epoch": 0.11617696724079703,
"grad_norm": 0.17559370398521423,
"learning_rate": 1.0527067017923654e-05,
"loss": 10.5945,
"step": 172
},
{
"epoch": 0.11685241472475515,
"grad_norm": 0.1844927966594696,
"learning_rate": 9.80087698670411e-06,
"loss": 10.5987,
"step": 173
},
{
"epoch": 0.11752786220871327,
"grad_norm": 0.16079100966453552,
"learning_rate": 9.09934649508375e-06,
"loss": 10.6051,
"step": 174
},
{
"epoch": 0.1182033096926714,
"grad_norm": 0.16014216840267181,
"learning_rate": 8.422667334494249e-06,
"loss": 10.596,
"step": 175
},
{
"epoch": 0.11887875717662952,
"grad_norm": 0.16521821916103363,
"learning_rate": 7.771024502261526e-06,
"loss": 10.6092,
"step": 176
},
{
"epoch": 0.11955420466058764,
"grad_norm": 0.1629599630832672,
"learning_rate": 7.144596151029303e-06,
"loss": 10.5984,
"step": 177
},
{
"epoch": 0.12022965214454576,
"grad_norm": 0.16382640600204468,
"learning_rate": 6.543553540053926e-06,
"loss": 10.5851,
"step": 178
},
{
"epoch": 0.12090509962850389,
"grad_norm": 0.15883676707744598,
"learning_rate": 5.968060988383883e-06,
"loss": 10.5956,
"step": 179
},
{
"epoch": 0.12158054711246201,
"grad_norm": 0.184475377202034,
"learning_rate": 5.418275829936537e-06,
"loss": 10.589,
"step": 180
},
{
"epoch": 0.12225599459642013,
"grad_norm": 0.19880411028862,
"learning_rate": 4.8943483704846475e-06,
"loss": 10.581,
"step": 181
},
{
"epoch": 0.12293144208037825,
"grad_norm": 0.16972365975379944,
"learning_rate": 4.3964218465642355e-06,
"loss": 10.5909,
"step": 182
},
{
"epoch": 0.12360688956433637,
"grad_norm": 0.17554166913032532,
"learning_rate": 3.924632386315186e-06,
"loss": 10.5855,
"step": 183
},
{
"epoch": 0.1242823370482945,
"grad_norm": 0.15895400941371918,
"learning_rate": 3.4791089722651436e-06,
"loss": 10.6027,
"step": 184
},
{
"epoch": 0.12495778453225262,
"grad_norm": 0.15381865203380585,
"learning_rate": 3.059973406066963e-06,
"loss": 10.6043,
"step": 185
},
{
"epoch": 0.12563323201621074,
"grad_norm": 0.1588447093963623,
"learning_rate": 2.667340275199426e-06,
"loss": 10.5858,
"step": 186
},
{
"epoch": 0.12630867950016886,
"grad_norm": 0.170160710811615,
"learning_rate": 2.3013169216400733e-06,
"loss": 10.5963,
"step": 187
},
{
"epoch": 0.12698412698412698,
"grad_norm": 0.19386546313762665,
"learning_rate": 1.9620034125190644e-06,
"loss": 10.5906,
"step": 188
},
{
"epoch": 0.1276595744680851,
"grad_norm": 0.15733763575553894,
"learning_rate": 1.6494925127617634e-06,
"loss": 10.5984,
"step": 189
},
{
"epoch": 0.12833502195204322,
"grad_norm": 0.1530665159225464,
"learning_rate": 1.3638696597277679e-06,
"loss": 10.5911,
"step": 190
},
{
"epoch": 0.12901046943600136,
"grad_norm": 0.1876905858516693,
"learning_rate": 1.1052129398531507e-06,
"loss": 10.6023,
"step": 191
},
{
"epoch": 0.12968591691995948,
"grad_norm": 0.1700376272201538,
"learning_rate": 8.735930673024806e-07,
"loss": 10.5979,
"step": 192
},
{
"epoch": 0.1303613644039176,
"grad_norm": 0.18517427146434784,
"learning_rate": 6.690733646361857e-07,
"loss": 10.5894,
"step": 193
},
{
"epoch": 0.13103681188787572,
"grad_norm": 0.16964897513389587,
"learning_rate": 4.917097454988584e-07,
"loss": 10.5996,
"step": 194
},
{
"epoch": 0.13171225937183384,
"grad_norm": 0.1931840479373932,
"learning_rate": 3.415506993330153e-07,
"loss": 10.5854,
"step": 195
},
{
"epoch": 0.13238770685579196,
"grad_norm": 0.15033087134361267,
"learning_rate": 2.1863727812254653e-07,
"loss": 10.6238,
"step": 196
},
{
"epoch": 0.13306315433975008,
"grad_norm": 0.1763419657945633,
"learning_rate": 1.230030851695263e-07,
"loss": 10.5943,
"step": 197
},
{
"epoch": 0.1337386018237082,
"grad_norm": 0.17355754971504211,
"learning_rate": 5.467426590739511e-08,
"loss": 10.5912,
"step": 198
},
{
"epoch": 0.13441404930766632,
"grad_norm": 0.1702238917350769,
"learning_rate": 1.3669500753099585e-08,
"loss": 10.6278,
"step": 199
},
{
"epoch": 0.13508949679162446,
"grad_norm": 0.15678980946540833,
"learning_rate": 0.0,
"loss": 10.5993,
"step": 200
},
{
"epoch": 0.13508949679162446,
"eval_loss": 10.595317840576172,
"eval_runtime": 2.2488,
"eval_samples_per_second": 277.475,
"eval_steps_per_second": 138.738,
"step": 200
}
],
"logging_steps": 1,
"max_steps": 200,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 50,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 32811830476800.0,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}