arco-mini-run-5 / checkpoint-2812 /trainer_state.json
appvoid's picture
Upload folder using huggingface_hub
9257a84 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9998222222222222,
"eval_steps": 5000,
"global_step": 2812,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0035555555555555557,
"grad_norm": 0.4803234934806824,
"learning_rate": 2.0000000000000003e-06,
"loss": 1.5376,
"step": 10
},
{
"epoch": 0.0071111111111111115,
"grad_norm": 0.4609735310077667,
"learning_rate": 4.000000000000001e-06,
"loss": 1.3932,
"step": 20
},
{
"epoch": 0.010666666666666666,
"grad_norm": 0.4800471365451813,
"learning_rate": 6e-06,
"loss": 1.6084,
"step": 30
},
{
"epoch": 0.014222222222222223,
"grad_norm": 0.513812780380249,
"learning_rate": 8.000000000000001e-06,
"loss": 1.5579,
"step": 40
},
{
"epoch": 0.017777777777777778,
"grad_norm": 0.5741678476333618,
"learning_rate": 1e-05,
"loss": 1.5945,
"step": 50
},
{
"epoch": 0.021333333333333333,
"grad_norm": 0.42774268984794617,
"learning_rate": 1.2e-05,
"loss": 1.6154,
"step": 60
},
{
"epoch": 0.024888888888888887,
"grad_norm": 0.37852978706359863,
"learning_rate": 1.4000000000000001e-05,
"loss": 1.5581,
"step": 70
},
{
"epoch": 0.028444444444444446,
"grad_norm": 0.48577743768692017,
"learning_rate": 1.6000000000000003e-05,
"loss": 1.57,
"step": 80
},
{
"epoch": 0.032,
"grad_norm": 0.5572775602340698,
"learning_rate": 1.8e-05,
"loss": 1.6154,
"step": 90
},
{
"epoch": 0.035555555555555556,
"grad_norm": 0.4657267928123474,
"learning_rate": 2e-05,
"loss": 1.5583,
"step": 100
},
{
"epoch": 0.03911111111111111,
"grad_norm": 0.48570355772972107,
"learning_rate": 2.2000000000000003e-05,
"loss": 1.6356,
"step": 110
},
{
"epoch": 0.042666666666666665,
"grad_norm": 0.5374159812927246,
"learning_rate": 2.4e-05,
"loss": 1.5911,
"step": 120
},
{
"epoch": 0.04622222222222222,
"grad_norm": 0.4544757306575775,
"learning_rate": 2.6000000000000002e-05,
"loss": 1.5462,
"step": 130
},
{
"epoch": 0.049777777777777775,
"grad_norm": 0.49096816778182983,
"learning_rate": 2.8000000000000003e-05,
"loss": 1.4696,
"step": 140
},
{
"epoch": 0.05333333333333334,
"grad_norm": 0.45160388946533203,
"learning_rate": 3e-05,
"loss": 1.4397,
"step": 150
},
{
"epoch": 0.05688888888888889,
"grad_norm": 0.4654102325439453,
"learning_rate": 3.2000000000000005e-05,
"loss": 1.61,
"step": 160
},
{
"epoch": 0.060444444444444446,
"grad_norm": 0.5608298182487488,
"learning_rate": 3.4000000000000007e-05,
"loss": 1.5799,
"step": 170
},
{
"epoch": 0.064,
"grad_norm": 0.5087274312973022,
"learning_rate": 3.6e-05,
"loss": 1.6534,
"step": 180
},
{
"epoch": 0.06755555555555555,
"grad_norm": 0.5886797308921814,
"learning_rate": 3.8e-05,
"loss": 1.5561,
"step": 190
},
{
"epoch": 0.07111111111111111,
"grad_norm": 0.5276679396629333,
"learning_rate": 4e-05,
"loss": 1.5705,
"step": 200
},
{
"epoch": 0.07466666666666667,
"grad_norm": 0.4331255257129669,
"learning_rate": 4.2e-05,
"loss": 1.554,
"step": 210
},
{
"epoch": 0.07822222222222222,
"grad_norm": 0.5068897604942322,
"learning_rate": 4.4000000000000006e-05,
"loss": 1.451,
"step": 220
},
{
"epoch": 0.08177777777777778,
"grad_norm": 0.48896610736846924,
"learning_rate": 4.600000000000001e-05,
"loss": 1.6522,
"step": 230
},
{
"epoch": 0.08533333333333333,
"grad_norm": 0.5350568890571594,
"learning_rate": 4.8e-05,
"loss": 1.5571,
"step": 240
},
{
"epoch": 0.08888888888888889,
"grad_norm": 0.6370730400085449,
"learning_rate": 5e-05,
"loss": 1.6112,
"step": 250
},
{
"epoch": 0.09244444444444444,
"grad_norm": 0.5409080386161804,
"learning_rate": 5.2000000000000004e-05,
"loss": 1.5878,
"step": 260
},
{
"epoch": 0.096,
"grad_norm": 0.5505415797233582,
"learning_rate": 5.4000000000000005e-05,
"loss": 1.6938,
"step": 270
},
{
"epoch": 0.09955555555555555,
"grad_norm": 0.557578980922699,
"learning_rate": 5.6000000000000006e-05,
"loss": 1.6776,
"step": 280
},
{
"epoch": 0.10311111111111111,
"grad_norm": 0.5505691170692444,
"learning_rate": 5.8e-05,
"loss": 1.7429,
"step": 290
},
{
"epoch": 0.10666666666666667,
"grad_norm": 0.5328088998794556,
"learning_rate": 6e-05,
"loss": 1.6633,
"step": 300
},
{
"epoch": 0.11022222222222222,
"grad_norm": 0.4911518394947052,
"learning_rate": 6.2e-05,
"loss": 1.4746,
"step": 310
},
{
"epoch": 0.11377777777777778,
"grad_norm": 0.5215829610824585,
"learning_rate": 6.400000000000001e-05,
"loss": 1.4723,
"step": 320
},
{
"epoch": 0.11733333333333333,
"grad_norm": 0.5551868081092834,
"learning_rate": 6.6e-05,
"loss": 1.5322,
"step": 330
},
{
"epoch": 0.12088888888888889,
"grad_norm": 0.607334554195404,
"learning_rate": 6.800000000000001e-05,
"loss": 1.7652,
"step": 340
},
{
"epoch": 0.12444444444444444,
"grad_norm": 0.4849174916744232,
"learning_rate": 7e-05,
"loss": 1.5915,
"step": 350
},
{
"epoch": 0.128,
"grad_norm": 0.5409530997276306,
"learning_rate": 7.2e-05,
"loss": 1.6263,
"step": 360
},
{
"epoch": 0.13155555555555556,
"grad_norm": 0.5955730080604553,
"learning_rate": 7.4e-05,
"loss": 1.598,
"step": 370
},
{
"epoch": 0.1351111111111111,
"grad_norm": 0.5792020559310913,
"learning_rate": 7.6e-05,
"loss": 1.7498,
"step": 380
},
{
"epoch": 0.13866666666666666,
"grad_norm": 0.6078203916549683,
"learning_rate": 7.800000000000001e-05,
"loss": 1.5885,
"step": 390
},
{
"epoch": 0.14222222222222222,
"grad_norm": 0.470548540353775,
"learning_rate": 8e-05,
"loss": 1.4596,
"step": 400
},
{
"epoch": 0.14577777777777778,
"grad_norm": 0.600089967250824,
"learning_rate": 8.2e-05,
"loss": 1.6343,
"step": 410
},
{
"epoch": 0.14933333333333335,
"grad_norm": 0.5364957451820374,
"learning_rate": 8.4e-05,
"loss": 1.4748,
"step": 420
},
{
"epoch": 0.15288888888888888,
"grad_norm": 0.5835921168327332,
"learning_rate": 8.6e-05,
"loss": 1.6809,
"step": 430
},
{
"epoch": 0.15644444444444444,
"grad_norm": 0.5760766863822937,
"learning_rate": 8.800000000000001e-05,
"loss": 1.6276,
"step": 440
},
{
"epoch": 0.16,
"grad_norm": 0.5958523154258728,
"learning_rate": 9e-05,
"loss": 1.4696,
"step": 450
},
{
"epoch": 0.16355555555555557,
"grad_norm": 0.5406598448753357,
"learning_rate": 9.200000000000001e-05,
"loss": 1.625,
"step": 460
},
{
"epoch": 0.1671111111111111,
"grad_norm": 0.5989697575569153,
"learning_rate": 9.4e-05,
"loss": 1.6854,
"step": 470
},
{
"epoch": 0.17066666666666666,
"grad_norm": 0.594978392124176,
"learning_rate": 9.6e-05,
"loss": 1.5774,
"step": 480
},
{
"epoch": 0.17422222222222222,
"grad_norm": 0.5095472931861877,
"learning_rate": 9.8e-05,
"loss": 1.7095,
"step": 490
},
{
"epoch": 0.17777777777777778,
"grad_norm": 0.7304088473320007,
"learning_rate": 0.0001,
"loss": 1.6842,
"step": 500
},
{
"epoch": 0.18133333333333335,
"grad_norm": 0.6202882528305054,
"learning_rate": 9.95674740484429e-05,
"loss": 1.696,
"step": 510
},
{
"epoch": 0.18488888888888888,
"grad_norm": 0.49087047576904297,
"learning_rate": 9.913494809688582e-05,
"loss": 1.6507,
"step": 520
},
{
"epoch": 0.18844444444444444,
"grad_norm": 0.6279426217079163,
"learning_rate": 9.870242214532872e-05,
"loss": 1.6809,
"step": 530
},
{
"epoch": 0.192,
"grad_norm": 0.602878987789154,
"learning_rate": 9.826989619377162e-05,
"loss": 1.6173,
"step": 540
},
{
"epoch": 0.19555555555555557,
"grad_norm": 0.6200081706047058,
"learning_rate": 9.783737024221454e-05,
"loss": 1.6787,
"step": 550
},
{
"epoch": 0.1991111111111111,
"grad_norm": 0.6582353115081787,
"learning_rate": 9.740484429065745e-05,
"loss": 1.7095,
"step": 560
},
{
"epoch": 0.20266666666666666,
"grad_norm": 0.541404664516449,
"learning_rate": 9.697231833910035e-05,
"loss": 1.8418,
"step": 570
},
{
"epoch": 0.20622222222222222,
"grad_norm": 0.5303178429603577,
"learning_rate": 9.653979238754325e-05,
"loss": 1.6112,
"step": 580
},
{
"epoch": 0.20977777777777779,
"grad_norm": 0.527113676071167,
"learning_rate": 9.610726643598617e-05,
"loss": 1.6454,
"step": 590
},
{
"epoch": 0.21333333333333335,
"grad_norm": 0.5398814678192139,
"learning_rate": 9.567474048442907e-05,
"loss": 1.5353,
"step": 600
},
{
"epoch": 0.21688888888888888,
"grad_norm": 0.5357266068458557,
"learning_rate": 9.524221453287197e-05,
"loss": 1.6816,
"step": 610
},
{
"epoch": 0.22044444444444444,
"grad_norm": 0.4769757390022278,
"learning_rate": 9.480968858131488e-05,
"loss": 1.6625,
"step": 620
},
{
"epoch": 0.224,
"grad_norm": 0.5274131298065186,
"learning_rate": 9.437716262975779e-05,
"loss": 1.7129,
"step": 630
},
{
"epoch": 0.22755555555555557,
"grad_norm": 0.6454447507858276,
"learning_rate": 9.394463667820069e-05,
"loss": 1.5577,
"step": 640
},
{
"epoch": 0.2311111111111111,
"grad_norm": 0.5967287421226501,
"learning_rate": 9.35121107266436e-05,
"loss": 1.3905,
"step": 650
},
{
"epoch": 0.23466666666666666,
"grad_norm": 0.57136470079422,
"learning_rate": 9.307958477508652e-05,
"loss": 1.6029,
"step": 660
},
{
"epoch": 0.23822222222222222,
"grad_norm": 0.5300968885421753,
"learning_rate": 9.264705882352942e-05,
"loss": 1.7363,
"step": 670
},
{
"epoch": 0.24177777777777779,
"grad_norm": 0.44372138381004333,
"learning_rate": 9.221453287197233e-05,
"loss": 1.6467,
"step": 680
},
{
"epoch": 0.24533333333333332,
"grad_norm": 0.45939210057258606,
"learning_rate": 9.178200692041523e-05,
"loss": 1.5413,
"step": 690
},
{
"epoch": 0.24888888888888888,
"grad_norm": 0.5719866752624512,
"learning_rate": 9.134948096885813e-05,
"loss": 1.5702,
"step": 700
},
{
"epoch": 0.25244444444444447,
"grad_norm": 0.4535101652145386,
"learning_rate": 9.091695501730105e-05,
"loss": 1.5331,
"step": 710
},
{
"epoch": 0.256,
"grad_norm": 0.5633770227432251,
"learning_rate": 9.048442906574395e-05,
"loss": 1.696,
"step": 720
},
{
"epoch": 0.25955555555555554,
"grad_norm": 0.6633341908454895,
"learning_rate": 9.005190311418685e-05,
"loss": 1.5428,
"step": 730
},
{
"epoch": 0.26311111111111113,
"grad_norm": 0.6005223393440247,
"learning_rate": 8.961937716262977e-05,
"loss": 1.8278,
"step": 740
},
{
"epoch": 0.26666666666666666,
"grad_norm": 0.5239768624305725,
"learning_rate": 8.918685121107267e-05,
"loss": 1.6761,
"step": 750
},
{
"epoch": 0.2702222222222222,
"grad_norm": 0.4788583517074585,
"learning_rate": 8.875432525951558e-05,
"loss": 1.548,
"step": 760
},
{
"epoch": 0.2737777777777778,
"grad_norm": 0.5428124666213989,
"learning_rate": 8.832179930795848e-05,
"loss": 1.6972,
"step": 770
},
{
"epoch": 0.2773333333333333,
"grad_norm": 0.42001649737358093,
"learning_rate": 8.78892733564014e-05,
"loss": 1.6244,
"step": 780
},
{
"epoch": 0.2808888888888889,
"grad_norm": 0.5100952386856079,
"learning_rate": 8.74567474048443e-05,
"loss": 1.6704,
"step": 790
},
{
"epoch": 0.28444444444444444,
"grad_norm": 0.49174782633781433,
"learning_rate": 8.70242214532872e-05,
"loss": 1.6876,
"step": 800
},
{
"epoch": 0.288,
"grad_norm": 0.543880820274353,
"learning_rate": 8.659169550173011e-05,
"loss": 1.599,
"step": 810
},
{
"epoch": 0.29155555555555557,
"grad_norm": 0.4966542422771454,
"learning_rate": 8.615916955017301e-05,
"loss": 1.4723,
"step": 820
},
{
"epoch": 0.2951111111111111,
"grad_norm": 0.44916948676109314,
"learning_rate": 8.572664359861592e-05,
"loss": 1.452,
"step": 830
},
{
"epoch": 0.2986666666666667,
"grad_norm": 0.5791043639183044,
"learning_rate": 8.529411764705883e-05,
"loss": 1.7569,
"step": 840
},
{
"epoch": 0.3022222222222222,
"grad_norm": 0.5880780816078186,
"learning_rate": 8.486159169550173e-05,
"loss": 1.4779,
"step": 850
},
{
"epoch": 0.30577777777777776,
"grad_norm": 0.5578981041908264,
"learning_rate": 8.442906574394463e-05,
"loss": 1.6577,
"step": 860
},
{
"epoch": 0.30933333333333335,
"grad_norm": 0.6132334470748901,
"learning_rate": 8.399653979238755e-05,
"loss": 1.5862,
"step": 870
},
{
"epoch": 0.3128888888888889,
"grad_norm": 0.512484073638916,
"learning_rate": 8.356401384083046e-05,
"loss": 1.4891,
"step": 880
},
{
"epoch": 0.3164444444444444,
"grad_norm": 0.4793022871017456,
"learning_rate": 8.313148788927336e-05,
"loss": 1.601,
"step": 890
},
{
"epoch": 0.32,
"grad_norm": 0.5512364506721497,
"learning_rate": 8.269896193771626e-05,
"loss": 1.6191,
"step": 900
},
{
"epoch": 0.32355555555555554,
"grad_norm": 0.5528599619865417,
"learning_rate": 8.226643598615918e-05,
"loss": 1.6555,
"step": 910
},
{
"epoch": 0.32711111111111113,
"grad_norm": 0.4718135595321655,
"learning_rate": 8.183391003460208e-05,
"loss": 1.3796,
"step": 920
},
{
"epoch": 0.33066666666666666,
"grad_norm": 0.5744655728340149,
"learning_rate": 8.140138408304498e-05,
"loss": 1.7323,
"step": 930
},
{
"epoch": 0.3342222222222222,
"grad_norm": 0.5298731923103333,
"learning_rate": 8.09688581314879e-05,
"loss": 1.6613,
"step": 940
},
{
"epoch": 0.3377777777777778,
"grad_norm": 0.507189154624939,
"learning_rate": 8.05363321799308e-05,
"loss": 1.4932,
"step": 950
},
{
"epoch": 0.3413333333333333,
"grad_norm": 0.5233726501464844,
"learning_rate": 8.01038062283737e-05,
"loss": 1.5989,
"step": 960
},
{
"epoch": 0.3448888888888889,
"grad_norm": 0.5754904747009277,
"learning_rate": 7.967128027681662e-05,
"loss": 1.6657,
"step": 970
},
{
"epoch": 0.34844444444444445,
"grad_norm": 0.6196070909500122,
"learning_rate": 7.923875432525953e-05,
"loss": 1.6808,
"step": 980
},
{
"epoch": 0.352,
"grad_norm": 0.461910605430603,
"learning_rate": 7.880622837370243e-05,
"loss": 1.6012,
"step": 990
},
{
"epoch": 0.35555555555555557,
"grad_norm": 0.4632907807826996,
"learning_rate": 7.837370242214533e-05,
"loss": 1.5879,
"step": 1000
},
{
"epoch": 0.3591111111111111,
"grad_norm": 0.48907166719436646,
"learning_rate": 7.794117647058824e-05,
"loss": 1.6109,
"step": 1010
},
{
"epoch": 0.3626666666666667,
"grad_norm": 0.5417441129684448,
"learning_rate": 7.750865051903114e-05,
"loss": 1.5687,
"step": 1020
},
{
"epoch": 0.3662222222222222,
"grad_norm": 0.519092857837677,
"learning_rate": 7.707612456747404e-05,
"loss": 1.6197,
"step": 1030
},
{
"epoch": 0.36977777777777776,
"grad_norm": 0.45026785135269165,
"learning_rate": 7.664359861591696e-05,
"loss": 1.5194,
"step": 1040
},
{
"epoch": 0.37333333333333335,
"grad_norm": 0.5378026366233826,
"learning_rate": 7.621107266435986e-05,
"loss": 1.419,
"step": 1050
},
{
"epoch": 0.3768888888888889,
"grad_norm": 0.4889166057109833,
"learning_rate": 7.577854671280276e-05,
"loss": 1.6356,
"step": 1060
},
{
"epoch": 0.3804444444444444,
"grad_norm": 0.5200234055519104,
"learning_rate": 7.534602076124569e-05,
"loss": 1.7118,
"step": 1070
},
{
"epoch": 0.384,
"grad_norm": 0.4398309886455536,
"learning_rate": 7.491349480968859e-05,
"loss": 1.525,
"step": 1080
},
{
"epoch": 0.38755555555555554,
"grad_norm": 0.43217843770980835,
"learning_rate": 7.448096885813149e-05,
"loss": 1.5643,
"step": 1090
},
{
"epoch": 0.39111111111111113,
"grad_norm": 0.5189200043678284,
"learning_rate": 7.40484429065744e-05,
"loss": 1.6891,
"step": 1100
},
{
"epoch": 0.39466666666666667,
"grad_norm": 0.5237637758255005,
"learning_rate": 7.36159169550173e-05,
"loss": 1.4445,
"step": 1110
},
{
"epoch": 0.3982222222222222,
"grad_norm": 0.5614046454429626,
"learning_rate": 7.318339100346021e-05,
"loss": 1.6439,
"step": 1120
},
{
"epoch": 0.4017777777777778,
"grad_norm": 0.4191811680793762,
"learning_rate": 7.275086505190312e-05,
"loss": 1.4344,
"step": 1130
},
{
"epoch": 0.4053333333333333,
"grad_norm": 0.5043474435806274,
"learning_rate": 7.231833910034602e-05,
"loss": 1.7017,
"step": 1140
},
{
"epoch": 0.4088888888888889,
"grad_norm": 0.40949752926826477,
"learning_rate": 7.188581314878892e-05,
"loss": 1.5903,
"step": 1150
},
{
"epoch": 0.41244444444444445,
"grad_norm": 0.4599238634109497,
"learning_rate": 7.145328719723184e-05,
"loss": 1.5666,
"step": 1160
},
{
"epoch": 0.416,
"grad_norm": 0.5736513137817383,
"learning_rate": 7.102076124567474e-05,
"loss": 1.5525,
"step": 1170
},
{
"epoch": 0.41955555555555557,
"grad_norm": 0.5437818169593811,
"learning_rate": 7.058823529411765e-05,
"loss": 1.6668,
"step": 1180
},
{
"epoch": 0.4231111111111111,
"grad_norm": 0.4318547248840332,
"learning_rate": 7.015570934256056e-05,
"loss": 1.4075,
"step": 1190
},
{
"epoch": 0.4266666666666667,
"grad_norm": 0.44979867339134216,
"learning_rate": 6.972318339100347e-05,
"loss": 1.3973,
"step": 1200
},
{
"epoch": 0.43022222222222223,
"grad_norm": 0.4653863310813904,
"learning_rate": 6.929065743944637e-05,
"loss": 1.4578,
"step": 1210
},
{
"epoch": 0.43377777777777776,
"grad_norm": 0.5257763266563416,
"learning_rate": 6.885813148788927e-05,
"loss": 1.5667,
"step": 1220
},
{
"epoch": 0.43733333333333335,
"grad_norm": 0.5117728114128113,
"learning_rate": 6.842560553633219e-05,
"loss": 1.6843,
"step": 1230
},
{
"epoch": 0.4408888888888889,
"grad_norm": 0.4611428678035736,
"learning_rate": 6.799307958477509e-05,
"loss": 1.7192,
"step": 1240
},
{
"epoch": 0.4444444444444444,
"grad_norm": 0.586694061756134,
"learning_rate": 6.756055363321799e-05,
"loss": 1.6057,
"step": 1250
},
{
"epoch": 0.448,
"grad_norm": 0.43433499336242676,
"learning_rate": 6.71280276816609e-05,
"loss": 1.6829,
"step": 1260
},
{
"epoch": 0.45155555555555554,
"grad_norm": 0.5851064920425415,
"learning_rate": 6.66955017301038e-05,
"loss": 1.6039,
"step": 1270
},
{
"epoch": 0.45511111111111113,
"grad_norm": 0.5549812912940979,
"learning_rate": 6.626297577854672e-05,
"loss": 1.6076,
"step": 1280
},
{
"epoch": 0.45866666666666667,
"grad_norm": 0.4489051103591919,
"learning_rate": 6.583044982698962e-05,
"loss": 1.6347,
"step": 1290
},
{
"epoch": 0.4622222222222222,
"grad_norm": 0.5636203289031982,
"learning_rate": 6.539792387543253e-05,
"loss": 1.3819,
"step": 1300
},
{
"epoch": 0.4657777777777778,
"grad_norm": 0.5598219037055969,
"learning_rate": 6.496539792387544e-05,
"loss": 1.6914,
"step": 1310
},
{
"epoch": 0.4693333333333333,
"grad_norm": 0.44435766339302063,
"learning_rate": 6.453287197231834e-05,
"loss": 1.5487,
"step": 1320
},
{
"epoch": 0.4728888888888889,
"grad_norm": 0.4664863646030426,
"learning_rate": 6.410034602076125e-05,
"loss": 1.5156,
"step": 1330
},
{
"epoch": 0.47644444444444445,
"grad_norm": 0.5192257761955261,
"learning_rate": 6.366782006920415e-05,
"loss": 1.5507,
"step": 1340
},
{
"epoch": 0.48,
"grad_norm": 0.6014835238456726,
"learning_rate": 6.323529411764705e-05,
"loss": 1.4809,
"step": 1350
},
{
"epoch": 0.48355555555555557,
"grad_norm": 0.5586997270584106,
"learning_rate": 6.280276816608997e-05,
"loss": 1.5619,
"step": 1360
},
{
"epoch": 0.4871111111111111,
"grad_norm": 0.4059390723705292,
"learning_rate": 6.237024221453287e-05,
"loss": 1.4799,
"step": 1370
},
{
"epoch": 0.49066666666666664,
"grad_norm": 0.41877785325050354,
"learning_rate": 6.193771626297578e-05,
"loss": 1.4187,
"step": 1380
},
{
"epoch": 0.49422222222222223,
"grad_norm": 0.4456830620765686,
"learning_rate": 6.15051903114187e-05,
"loss": 1.5507,
"step": 1390
},
{
"epoch": 0.49777777777777776,
"grad_norm": 0.45245644450187683,
"learning_rate": 6.10726643598616e-05,
"loss": 1.5478,
"step": 1400
},
{
"epoch": 0.5013333333333333,
"grad_norm": 0.4327283501625061,
"learning_rate": 6.06401384083045e-05,
"loss": 1.5734,
"step": 1410
},
{
"epoch": 0.5048888888888889,
"grad_norm": 0.6047600507736206,
"learning_rate": 6.020761245674741e-05,
"loss": 1.6916,
"step": 1420
},
{
"epoch": 0.5084444444444445,
"grad_norm": 0.4880015254020691,
"learning_rate": 5.9775086505190316e-05,
"loss": 1.6346,
"step": 1430
},
{
"epoch": 0.512,
"grad_norm": 0.4689754843711853,
"learning_rate": 5.934256055363322e-05,
"loss": 1.5714,
"step": 1440
},
{
"epoch": 0.5155555555555555,
"grad_norm": 0.4588066041469574,
"learning_rate": 5.8910034602076125e-05,
"loss": 1.5547,
"step": 1450
},
{
"epoch": 0.5191111111111111,
"grad_norm": 0.45055222511291504,
"learning_rate": 5.847750865051903e-05,
"loss": 1.4591,
"step": 1460
},
{
"epoch": 0.5226666666666666,
"grad_norm": 0.49660903215408325,
"learning_rate": 5.8044982698961933e-05,
"loss": 1.7409,
"step": 1470
},
{
"epoch": 0.5262222222222223,
"grad_norm": 0.48975858092308044,
"learning_rate": 5.761245674740484e-05,
"loss": 1.5158,
"step": 1480
},
{
"epoch": 0.5297777777777778,
"grad_norm": 0.4958665668964386,
"learning_rate": 5.7179930795847756e-05,
"loss": 1.4612,
"step": 1490
},
{
"epoch": 0.5333333333333333,
"grad_norm": 0.4389738142490387,
"learning_rate": 5.6747404844290664e-05,
"loss": 1.6067,
"step": 1500
},
{
"epoch": 0.5368888888888889,
"grad_norm": 0.44060131907463074,
"learning_rate": 5.631487889273357e-05,
"loss": 1.3391,
"step": 1510
},
{
"epoch": 0.5404444444444444,
"grad_norm": 0.5496039390563965,
"learning_rate": 5.588235294117647e-05,
"loss": 1.5108,
"step": 1520
},
{
"epoch": 0.544,
"grad_norm": 0.5372530221939087,
"learning_rate": 5.544982698961938e-05,
"loss": 1.4777,
"step": 1530
},
{
"epoch": 0.5475555555555556,
"grad_norm": 0.5017176270484924,
"learning_rate": 5.501730103806229e-05,
"loss": 1.5007,
"step": 1540
},
{
"epoch": 0.5511111111111111,
"grad_norm": 0.5397398471832275,
"learning_rate": 5.458477508650519e-05,
"loss": 1.6057,
"step": 1550
},
{
"epoch": 0.5546666666666666,
"grad_norm": 0.5731176733970642,
"learning_rate": 5.41522491349481e-05,
"loss": 1.6644,
"step": 1560
},
{
"epoch": 0.5582222222222222,
"grad_norm": 0.48925384879112244,
"learning_rate": 5.3719723183391005e-05,
"loss": 1.5506,
"step": 1570
},
{
"epoch": 0.5617777777777778,
"grad_norm": 0.47176703810691833,
"learning_rate": 5.3287197231833906e-05,
"loss": 1.5355,
"step": 1580
},
{
"epoch": 0.5653333333333334,
"grad_norm": 0.45709192752838135,
"learning_rate": 5.285467128027683e-05,
"loss": 1.4926,
"step": 1590
},
{
"epoch": 0.5688888888888889,
"grad_norm": 0.5231002569198608,
"learning_rate": 5.242214532871973e-05,
"loss": 1.5645,
"step": 1600
},
{
"epoch": 0.5724444444444444,
"grad_norm": 0.4811376929283142,
"learning_rate": 5.1989619377162636e-05,
"loss": 1.5169,
"step": 1610
},
{
"epoch": 0.576,
"grad_norm": 0.5673519968986511,
"learning_rate": 5.1557093425605544e-05,
"loss": 1.4974,
"step": 1620
},
{
"epoch": 0.5795555555555556,
"grad_norm": 0.5168330669403076,
"learning_rate": 5.1124567474048445e-05,
"loss": 1.4242,
"step": 1630
},
{
"epoch": 0.5831111111111111,
"grad_norm": 0.49934643507003784,
"learning_rate": 5.069204152249135e-05,
"loss": 1.6342,
"step": 1640
},
{
"epoch": 0.5866666666666667,
"grad_norm": 0.44559621810913086,
"learning_rate": 5.0259515570934254e-05,
"loss": 1.5204,
"step": 1650
},
{
"epoch": 0.5902222222222222,
"grad_norm": 0.5243204832077026,
"learning_rate": 4.982698961937716e-05,
"loss": 1.4703,
"step": 1660
},
{
"epoch": 0.5937777777777777,
"grad_norm": 0.5117699503898621,
"learning_rate": 4.9394463667820076e-05,
"loss": 1.7011,
"step": 1670
},
{
"epoch": 0.5973333333333334,
"grad_norm": 0.42527756094932556,
"learning_rate": 4.896193771626298e-05,
"loss": 1.4331,
"step": 1680
},
{
"epoch": 0.6008888888888889,
"grad_norm": 0.5092193484306335,
"learning_rate": 4.8529411764705885e-05,
"loss": 1.4378,
"step": 1690
},
{
"epoch": 0.6044444444444445,
"grad_norm": 0.5100553631782532,
"learning_rate": 4.809688581314879e-05,
"loss": 1.5999,
"step": 1700
},
{
"epoch": 0.608,
"grad_norm": 0.5817019939422607,
"learning_rate": 4.7664359861591694e-05,
"loss": 1.6094,
"step": 1710
},
{
"epoch": 0.6115555555555555,
"grad_norm": 0.41723236441612244,
"learning_rate": 4.723183391003461e-05,
"loss": 1.3555,
"step": 1720
},
{
"epoch": 0.6151111111111112,
"grad_norm": 0.5209750533103943,
"learning_rate": 4.679930795847751e-05,
"loss": 1.56,
"step": 1730
},
{
"epoch": 0.6186666666666667,
"grad_norm": 0.48648136854171753,
"learning_rate": 4.636678200692042e-05,
"loss": 1.48,
"step": 1740
},
{
"epoch": 0.6222222222222222,
"grad_norm": 0.5282444357872009,
"learning_rate": 4.5934256055363325e-05,
"loss": 1.5003,
"step": 1750
},
{
"epoch": 0.6257777777777778,
"grad_norm": 0.4915744364261627,
"learning_rate": 4.5501730103806226e-05,
"loss": 1.4575,
"step": 1760
},
{
"epoch": 0.6293333333333333,
"grad_norm": 0.568629801273346,
"learning_rate": 4.506920415224914e-05,
"loss": 1.4804,
"step": 1770
},
{
"epoch": 0.6328888888888888,
"grad_norm": 0.48679119348526,
"learning_rate": 4.463667820069204e-05,
"loss": 1.5015,
"step": 1780
},
{
"epoch": 0.6364444444444445,
"grad_norm": 0.5196161866188049,
"learning_rate": 4.420415224913495e-05,
"loss": 1.561,
"step": 1790
},
{
"epoch": 0.64,
"grad_norm": 0.4946150779724121,
"learning_rate": 4.377162629757786e-05,
"loss": 1.5151,
"step": 1800
},
{
"epoch": 0.6435555555555555,
"grad_norm": 0.40968433022499084,
"learning_rate": 4.333910034602076e-05,
"loss": 1.5373,
"step": 1810
},
{
"epoch": 0.6471111111111111,
"grad_norm": 0.6429114937782288,
"learning_rate": 4.290657439446367e-05,
"loss": 1.4867,
"step": 1820
},
{
"epoch": 0.6506666666666666,
"grad_norm": 0.474758118391037,
"learning_rate": 4.247404844290658e-05,
"loss": 1.5035,
"step": 1830
},
{
"epoch": 0.6542222222222223,
"grad_norm": 0.51218581199646,
"learning_rate": 4.204152249134948e-05,
"loss": 1.5725,
"step": 1840
},
{
"epoch": 0.6577777777777778,
"grad_norm": 0.49607232213020325,
"learning_rate": 4.160899653979239e-05,
"loss": 1.5352,
"step": 1850
},
{
"epoch": 0.6613333333333333,
"grad_norm": 0.4859633147716522,
"learning_rate": 4.11764705882353e-05,
"loss": 1.4925,
"step": 1860
},
{
"epoch": 0.6648888888888889,
"grad_norm": 0.4592611491680145,
"learning_rate": 4.0743944636678205e-05,
"loss": 1.5156,
"step": 1870
},
{
"epoch": 0.6684444444444444,
"grad_norm": 0.47916117310523987,
"learning_rate": 4.031141868512111e-05,
"loss": 1.545,
"step": 1880
},
{
"epoch": 0.672,
"grad_norm": 0.5305891633033752,
"learning_rate": 3.9878892733564014e-05,
"loss": 1.5384,
"step": 1890
},
{
"epoch": 0.6755555555555556,
"grad_norm": 0.4583602547645569,
"learning_rate": 3.944636678200692e-05,
"loss": 1.5635,
"step": 1900
},
{
"epoch": 0.6791111111111111,
"grad_norm": 0.5024811625480652,
"learning_rate": 3.901384083044983e-05,
"loss": 1.5016,
"step": 1910
},
{
"epoch": 0.6826666666666666,
"grad_norm": 0.5373083353042603,
"learning_rate": 3.858131487889274e-05,
"loss": 1.5399,
"step": 1920
},
{
"epoch": 0.6862222222222222,
"grad_norm": 0.415873646736145,
"learning_rate": 3.8148788927335645e-05,
"loss": 1.6384,
"step": 1930
},
{
"epoch": 0.6897777777777778,
"grad_norm": 0.46044766902923584,
"learning_rate": 3.7716262975778546e-05,
"loss": 1.5588,
"step": 1940
},
{
"epoch": 0.6933333333333334,
"grad_norm": 0.37672215700149536,
"learning_rate": 3.7283737024221454e-05,
"loss": 1.49,
"step": 1950
},
{
"epoch": 0.6968888888888889,
"grad_norm": 0.40345415472984314,
"learning_rate": 3.685121107266436e-05,
"loss": 1.3789,
"step": 1960
},
{
"epoch": 0.7004444444444444,
"grad_norm": 0.49566739797592163,
"learning_rate": 3.641868512110726e-05,
"loss": 1.3318,
"step": 1970
},
{
"epoch": 0.704,
"grad_norm": 0.48587340116500854,
"learning_rate": 3.598615916955018e-05,
"loss": 1.4758,
"step": 1980
},
{
"epoch": 0.7075555555555556,
"grad_norm": 0.596446692943573,
"learning_rate": 3.5553633217993085e-05,
"loss": 1.5347,
"step": 1990
},
{
"epoch": 0.7111111111111111,
"grad_norm": 0.4951972961425781,
"learning_rate": 3.5121107266435986e-05,
"loss": 1.2704,
"step": 2000
},
{
"epoch": 0.7146666666666667,
"grad_norm": 0.4018455743789673,
"learning_rate": 3.4688581314878894e-05,
"loss": 1.5217,
"step": 2010
},
{
"epoch": 0.7182222222222222,
"grad_norm": 0.5961456894874573,
"learning_rate": 3.4256055363321795e-05,
"loss": 1.7622,
"step": 2020
},
{
"epoch": 0.7217777777777777,
"grad_norm": 0.4609779119491577,
"learning_rate": 3.382352941176471e-05,
"loss": 1.5252,
"step": 2030
},
{
"epoch": 0.7253333333333334,
"grad_norm": 0.43142765760421753,
"learning_rate": 3.339100346020762e-05,
"loss": 1.4461,
"step": 2040
},
{
"epoch": 0.7288888888888889,
"grad_norm": 0.4517713785171509,
"learning_rate": 3.295847750865052e-05,
"loss": 1.5962,
"step": 2050
},
{
"epoch": 0.7324444444444445,
"grad_norm": 0.48978036642074585,
"learning_rate": 3.2525951557093426e-05,
"loss": 1.5366,
"step": 2060
},
{
"epoch": 0.736,
"grad_norm": 0.40700387954711914,
"learning_rate": 3.2093425605536334e-05,
"loss": 1.5572,
"step": 2070
},
{
"epoch": 0.7395555555555555,
"grad_norm": 0.5392187237739563,
"learning_rate": 3.166089965397924e-05,
"loss": 1.5552,
"step": 2080
},
{
"epoch": 0.7431111111111111,
"grad_norm": 0.44562795758247375,
"learning_rate": 3.122837370242215e-05,
"loss": 1.5194,
"step": 2090
},
{
"epoch": 0.7466666666666667,
"grad_norm": 0.44013121724128723,
"learning_rate": 3.079584775086505e-05,
"loss": 1.3915,
"step": 2100
},
{
"epoch": 0.7502222222222222,
"grad_norm": 0.4336005747318268,
"learning_rate": 3.036332179930796e-05,
"loss": 1.4715,
"step": 2110
},
{
"epoch": 0.7537777777777778,
"grad_norm": 0.5743895769119263,
"learning_rate": 2.9930795847750863e-05,
"loss": 1.5256,
"step": 2120
},
{
"epoch": 0.7573333333333333,
"grad_norm": 0.45934122800827026,
"learning_rate": 2.9498269896193774e-05,
"loss": 1.5873,
"step": 2130
},
{
"epoch": 0.7608888888888888,
"grad_norm": 0.5065307021141052,
"learning_rate": 2.9065743944636682e-05,
"loss": 1.5582,
"step": 2140
},
{
"epoch": 0.7644444444444445,
"grad_norm": 0.511725664138794,
"learning_rate": 2.8633217993079586e-05,
"loss": 1.4813,
"step": 2150
},
{
"epoch": 0.768,
"grad_norm": 0.5018854737281799,
"learning_rate": 2.820069204152249e-05,
"loss": 1.5786,
"step": 2160
},
{
"epoch": 0.7715555555555556,
"grad_norm": 0.5292341113090515,
"learning_rate": 2.77681660899654e-05,
"loss": 1.6471,
"step": 2170
},
{
"epoch": 0.7751111111111111,
"grad_norm": 0.3638977110385895,
"learning_rate": 2.733564013840831e-05,
"loss": 1.4863,
"step": 2180
},
{
"epoch": 0.7786666666666666,
"grad_norm": 0.49934521317481995,
"learning_rate": 2.6903114186851214e-05,
"loss": 1.4587,
"step": 2190
},
{
"epoch": 0.7822222222222223,
"grad_norm": 0.5519063472747803,
"learning_rate": 2.647058823529412e-05,
"loss": 1.4915,
"step": 2200
},
{
"epoch": 0.7857777777777778,
"grad_norm": 0.5289978384971619,
"learning_rate": 2.6038062283737023e-05,
"loss": 1.5924,
"step": 2210
},
{
"epoch": 0.7893333333333333,
"grad_norm": 0.4768216907978058,
"learning_rate": 2.560553633217993e-05,
"loss": 1.4847,
"step": 2220
},
{
"epoch": 0.7928888888888889,
"grad_norm": 0.5400416254997253,
"learning_rate": 2.5173010380622842e-05,
"loss": 1.5664,
"step": 2230
},
{
"epoch": 0.7964444444444444,
"grad_norm": 0.5170167684555054,
"learning_rate": 2.4740484429065743e-05,
"loss": 1.7191,
"step": 2240
},
{
"epoch": 0.8,
"grad_norm": 0.4661620259284973,
"learning_rate": 2.430795847750865e-05,
"loss": 1.584,
"step": 2250
},
{
"epoch": 0.8035555555555556,
"grad_norm": 0.48751363158226013,
"learning_rate": 2.387543252595156e-05,
"loss": 1.6166,
"step": 2260
},
{
"epoch": 0.8071111111111111,
"grad_norm": 0.48189136385917664,
"learning_rate": 2.3442906574394467e-05,
"loss": 1.412,
"step": 2270
},
{
"epoch": 0.8106666666666666,
"grad_norm": 0.4377134144306183,
"learning_rate": 2.301038062283737e-05,
"loss": 1.4633,
"step": 2280
},
{
"epoch": 0.8142222222222222,
"grad_norm": 0.4900270998477936,
"learning_rate": 2.2577854671280275e-05,
"loss": 1.5968,
"step": 2290
},
{
"epoch": 0.8177777777777778,
"grad_norm": 0.5673394203186035,
"learning_rate": 2.2145328719723187e-05,
"loss": 1.4443,
"step": 2300
},
{
"epoch": 0.8213333333333334,
"grad_norm": 0.5573922395706177,
"learning_rate": 2.171280276816609e-05,
"loss": 1.5951,
"step": 2310
},
{
"epoch": 0.8248888888888889,
"grad_norm": 0.5815632343292236,
"learning_rate": 2.1280276816609e-05,
"loss": 1.54,
"step": 2320
},
{
"epoch": 0.8284444444444444,
"grad_norm": 0.5276745557785034,
"learning_rate": 2.0847750865051903e-05,
"loss": 1.5903,
"step": 2330
},
{
"epoch": 0.832,
"grad_norm": 0.4580680727958679,
"learning_rate": 2.041522491349481e-05,
"loss": 1.6143,
"step": 2340
},
{
"epoch": 0.8355555555555556,
"grad_norm": 0.5015590190887451,
"learning_rate": 1.998269896193772e-05,
"loss": 1.4156,
"step": 2350
},
{
"epoch": 0.8391111111111111,
"grad_norm": 0.508334219455719,
"learning_rate": 1.9550173010380623e-05,
"loss": 1.5272,
"step": 2360
},
{
"epoch": 0.8426666666666667,
"grad_norm": 0.44540607929229736,
"learning_rate": 1.9117647058823528e-05,
"loss": 1.5522,
"step": 2370
},
{
"epoch": 0.8462222222222222,
"grad_norm": 0.5358651876449585,
"learning_rate": 1.868512110726644e-05,
"loss": 1.4527,
"step": 2380
},
{
"epoch": 0.8497777777777777,
"grad_norm": 0.5191397666931152,
"learning_rate": 1.8252595155709343e-05,
"loss": 1.5627,
"step": 2390
},
{
"epoch": 0.8533333333333334,
"grad_norm": 0.4698735773563385,
"learning_rate": 1.782006920415225e-05,
"loss": 1.4434,
"step": 2400
},
{
"epoch": 0.8568888888888889,
"grad_norm": 0.4450030028820038,
"learning_rate": 1.7387543252595156e-05,
"loss": 1.383,
"step": 2410
},
{
"epoch": 0.8604444444444445,
"grad_norm": 0.47164052724838257,
"learning_rate": 1.6955017301038063e-05,
"loss": 1.5499,
"step": 2420
},
{
"epoch": 0.864,
"grad_norm": 0.4113026559352875,
"learning_rate": 1.652249134948097e-05,
"loss": 1.499,
"step": 2430
},
{
"epoch": 0.8675555555555555,
"grad_norm": 0.48519784212112427,
"learning_rate": 1.6089965397923876e-05,
"loss": 1.5236,
"step": 2440
},
{
"epoch": 0.8711111111111111,
"grad_norm": 0.5262069702148438,
"learning_rate": 1.5657439446366783e-05,
"loss": 1.5144,
"step": 2450
},
{
"epoch": 0.8746666666666667,
"grad_norm": 0.42513802647590637,
"learning_rate": 1.522491349480969e-05,
"loss": 1.4932,
"step": 2460
},
{
"epoch": 0.8782222222222222,
"grad_norm": 0.47486934065818787,
"learning_rate": 1.4792387543252596e-05,
"loss": 1.4398,
"step": 2470
},
{
"epoch": 0.8817777777777778,
"grad_norm": 0.5634934902191162,
"learning_rate": 1.4359861591695503e-05,
"loss": 1.65,
"step": 2480
},
{
"epoch": 0.8853333333333333,
"grad_norm": 0.43849506974220276,
"learning_rate": 1.3927335640138408e-05,
"loss": 1.5201,
"step": 2490
},
{
"epoch": 0.8888888888888888,
"grad_norm": 0.4285069704055786,
"learning_rate": 1.3494809688581317e-05,
"loss": 1.492,
"step": 2500
},
{
"epoch": 0.8924444444444445,
"grad_norm": 0.5707066655158997,
"learning_rate": 1.3062283737024222e-05,
"loss": 1.5991,
"step": 2510
},
{
"epoch": 0.896,
"grad_norm": 0.4697231948375702,
"learning_rate": 1.2629757785467128e-05,
"loss": 1.5585,
"step": 2520
},
{
"epoch": 0.8995555555555556,
"grad_norm": 0.4958859384059906,
"learning_rate": 1.2197231833910034e-05,
"loss": 1.5496,
"step": 2530
},
{
"epoch": 0.9031111111111111,
"grad_norm": 0.4845017194747925,
"learning_rate": 1.1764705882352942e-05,
"loss": 1.592,
"step": 2540
},
{
"epoch": 0.9066666666666666,
"grad_norm": 0.4356600344181061,
"learning_rate": 1.1332179930795848e-05,
"loss": 1.5873,
"step": 2550
},
{
"epoch": 0.9102222222222223,
"grad_norm": 0.3917759954929352,
"learning_rate": 1.0899653979238756e-05,
"loss": 1.4984,
"step": 2560
},
{
"epoch": 0.9137777777777778,
"grad_norm": 0.48273682594299316,
"learning_rate": 1.046712802768166e-05,
"loss": 1.4288,
"step": 2570
},
{
"epoch": 0.9173333333333333,
"grad_norm": 0.5662245750427246,
"learning_rate": 1.0034602076124568e-05,
"loss": 1.5123,
"step": 2580
},
{
"epoch": 0.9208888888888889,
"grad_norm": 0.510188639163971,
"learning_rate": 9.602076124567474e-06,
"loss": 1.4915,
"step": 2590
},
{
"epoch": 0.9244444444444444,
"grad_norm": 0.5999191403388977,
"learning_rate": 9.169550173010382e-06,
"loss": 1.7374,
"step": 2600
},
{
"epoch": 0.928,
"grad_norm": 0.5733387470245361,
"learning_rate": 8.737024221453288e-06,
"loss": 1.5551,
"step": 2610
},
{
"epoch": 0.9315555555555556,
"grad_norm": 0.3804311156272888,
"learning_rate": 8.304498269896194e-06,
"loss": 1.5631,
"step": 2620
},
{
"epoch": 0.9351111111111111,
"grad_norm": 0.44998374581336975,
"learning_rate": 7.8719723183391e-06,
"loss": 1.7566,
"step": 2630
},
{
"epoch": 0.9386666666666666,
"grad_norm": 0.4219221770763397,
"learning_rate": 7.439446366782007e-06,
"loss": 1.454,
"step": 2640
},
{
"epoch": 0.9422222222222222,
"grad_norm": 0.48252835869789124,
"learning_rate": 7.006920415224914e-06,
"loss": 1.5107,
"step": 2650
},
{
"epoch": 0.9457777777777778,
"grad_norm": 0.5446690320968628,
"learning_rate": 6.5743944636678194e-06,
"loss": 1.5333,
"step": 2660
},
{
"epoch": 0.9493333333333334,
"grad_norm": 0.4676448106765747,
"learning_rate": 6.141868512110726e-06,
"loss": 1.4822,
"step": 2670
},
{
"epoch": 0.9528888888888889,
"grad_norm": 0.41428953409194946,
"learning_rate": 5.709342560553633e-06,
"loss": 1.398,
"step": 2680
},
{
"epoch": 0.9564444444444444,
"grad_norm": 0.39139533042907715,
"learning_rate": 5.2768166089965395e-06,
"loss": 1.539,
"step": 2690
},
{
"epoch": 0.96,
"grad_norm": 0.5213208794593811,
"learning_rate": 4.8442906574394464e-06,
"loss": 1.535,
"step": 2700
},
{
"epoch": 0.9635555555555556,
"grad_norm": 0.6053948402404785,
"learning_rate": 4.411764705882353e-06,
"loss": 1.484,
"step": 2710
},
{
"epoch": 0.9671111111111111,
"grad_norm": 0.44833922386169434,
"learning_rate": 3.9792387543252595e-06,
"loss": 1.3107,
"step": 2720
},
{
"epoch": 0.9706666666666667,
"grad_norm": 0.4769100844860077,
"learning_rate": 3.5467128027681665e-06,
"loss": 1.5114,
"step": 2730
},
{
"epoch": 0.9742222222222222,
"grad_norm": 0.5112429857254028,
"learning_rate": 3.1141868512110726e-06,
"loss": 1.5357,
"step": 2740
},
{
"epoch": 0.9777777777777777,
"grad_norm": 0.4884449243545532,
"learning_rate": 2.6816608996539796e-06,
"loss": 1.4739,
"step": 2750
},
{
"epoch": 0.9813333333333333,
"grad_norm": 0.5470872521400452,
"learning_rate": 2.249134948096886e-06,
"loss": 1.5557,
"step": 2760
},
{
"epoch": 0.9848888888888889,
"grad_norm": 0.517669141292572,
"learning_rate": 1.8166089965397926e-06,
"loss": 1.5388,
"step": 2770
},
{
"epoch": 0.9884444444444445,
"grad_norm": 0.46402767300605774,
"learning_rate": 1.3840830449826992e-06,
"loss": 1.5509,
"step": 2780
},
{
"epoch": 0.992,
"grad_norm": 0.5298945307731628,
"learning_rate": 9.515570934256056e-07,
"loss": 1.533,
"step": 2790
},
{
"epoch": 0.9955555555555555,
"grad_norm": 0.5506584644317627,
"learning_rate": 5.190311418685121e-07,
"loss": 1.404,
"step": 2800
},
{
"epoch": 0.9991111111111111,
"grad_norm": 0.5148972868919373,
"learning_rate": 8.65051903114187e-08,
"loss": 1.558,
"step": 2810
}
],
"logging_steps": 10,
"max_steps": 2812,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 5000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.1067056282035814e+17,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}