|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 500, |
|
"global_step": 1044, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0009578544061302681, |
|
"grad_norm": 0.6597593171619324, |
|
"learning_rate": 9.523809523809525e-08, |
|
"loss": 1.1525, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.004789272030651341, |
|
"grad_norm": 0.6607550915738474, |
|
"learning_rate": 4.7619047619047623e-07, |
|
"loss": 1.1632, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.009578544061302681, |
|
"grad_norm": 0.6985874969645404, |
|
"learning_rate": 9.523809523809525e-07, |
|
"loss": 1.1753, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.014367816091954023, |
|
"grad_norm": 0.6553768135986224, |
|
"learning_rate": 1.4285714285714286e-06, |
|
"loss": 1.1488, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.019157088122605363, |
|
"grad_norm": 0.6087816631352209, |
|
"learning_rate": 1.904761904761905e-06, |
|
"loss": 1.1384, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.023946360153256706, |
|
"grad_norm": 0.5225466109798246, |
|
"learning_rate": 2.380952380952381e-06, |
|
"loss": 1.122, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.028735632183908046, |
|
"grad_norm": 0.5080052683286845, |
|
"learning_rate": 2.8571428571428573e-06, |
|
"loss": 1.1505, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.033524904214559385, |
|
"grad_norm": 0.4341207873557241, |
|
"learning_rate": 3.3333333333333333e-06, |
|
"loss": 1.1339, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.038314176245210725, |
|
"grad_norm": 0.3985137882747194, |
|
"learning_rate": 3.80952380952381e-06, |
|
"loss": 1.1282, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.04310344827586207, |
|
"grad_norm": 0.36938748261120696, |
|
"learning_rate": 4.2857142857142855e-06, |
|
"loss": 1.1141, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.04789272030651341, |
|
"grad_norm": 0.31186688358132403, |
|
"learning_rate": 4.761904761904762e-06, |
|
"loss": 1.0914, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.05268199233716475, |
|
"grad_norm": 0.3279958963323018, |
|
"learning_rate": 5.2380952380952384e-06, |
|
"loss": 1.1115, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.05747126436781609, |
|
"grad_norm": 0.3107640896000187, |
|
"learning_rate": 5.7142857142857145e-06, |
|
"loss": 1.0897, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.06226053639846743, |
|
"grad_norm": 0.3071723661825106, |
|
"learning_rate": 6.1904761904761914e-06, |
|
"loss": 1.1045, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.06704980842911877, |
|
"grad_norm": 0.2936092764163838, |
|
"learning_rate": 6.666666666666667e-06, |
|
"loss": 1.0661, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.07183908045977011, |
|
"grad_norm": 0.28053978486957076, |
|
"learning_rate": 7.1428571428571436e-06, |
|
"loss": 1.0887, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.07662835249042145, |
|
"grad_norm": 0.2891456438800518, |
|
"learning_rate": 7.61904761904762e-06, |
|
"loss": 1.0756, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.08141762452107279, |
|
"grad_norm": 0.24950135731624706, |
|
"learning_rate": 8.095238095238097e-06, |
|
"loss": 1.0478, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.08620689655172414, |
|
"grad_norm": 0.23670461068088686, |
|
"learning_rate": 8.571428571428571e-06, |
|
"loss": 1.0389, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.09099616858237548, |
|
"grad_norm": 0.2507461508986584, |
|
"learning_rate": 9.047619047619049e-06, |
|
"loss": 1.0509, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.09578544061302682, |
|
"grad_norm": 0.22930855800514297, |
|
"learning_rate": 9.523809523809525e-06, |
|
"loss": 1.0377, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.10057471264367816, |
|
"grad_norm": 0.24445927836597864, |
|
"learning_rate": 1e-05, |
|
"loss": 1.0506, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.1053639846743295, |
|
"grad_norm": 0.22388714842997454, |
|
"learning_rate": 9.999300418283908e-06, |
|
"loss": 1.0377, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.11015325670498084, |
|
"grad_norm": 0.2406057364217601, |
|
"learning_rate": 9.997201868901463e-06, |
|
"loss": 1.0374, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.11494252873563218, |
|
"grad_norm": 0.2275701466092952, |
|
"learning_rate": 9.993704939095376e-06, |
|
"loss": 1.0663, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.11973180076628352, |
|
"grad_norm": 0.22200651495287974, |
|
"learning_rate": 9.988810607420912e-06, |
|
"loss": 1.0448, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.12452107279693486, |
|
"grad_norm": 0.22051254365376602, |
|
"learning_rate": 9.982520243472044e-06, |
|
"loss": 1.0099, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.12931034482758622, |
|
"grad_norm": 0.23279527845715264, |
|
"learning_rate": 9.974835607498224e-06, |
|
"loss": 1.0212, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.13409961685823754, |
|
"grad_norm": 0.22813702069156522, |
|
"learning_rate": 9.965758849911774e-06, |
|
"loss": 1.023, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.1388888888888889, |
|
"grad_norm": 0.23555996785615885, |
|
"learning_rate": 9.955292510686156e-06, |
|
"loss": 0.9997, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.14367816091954022, |
|
"grad_norm": 0.2560040654609538, |
|
"learning_rate": 9.943439518645193e-06, |
|
"loss": 1.0172, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.14846743295019157, |
|
"grad_norm": 0.24096780952464245, |
|
"learning_rate": 9.930203190643491e-06, |
|
"loss": 0.9876, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.1532567049808429, |
|
"grad_norm": 0.24193989881098193, |
|
"learning_rate": 9.915587230638269e-06, |
|
"loss": 1.0417, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.15804597701149425, |
|
"grad_norm": 0.24788380685714667, |
|
"learning_rate": 9.899595728652883e-06, |
|
"loss": 1.0332, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.16283524904214558, |
|
"grad_norm": 0.2570877531631883, |
|
"learning_rate": 9.882233159632297e-06, |
|
"loss": 1.0129, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.16762452107279693, |
|
"grad_norm": 0.26735645578353723, |
|
"learning_rate": 9.863504382190838e-06, |
|
"loss": 1.0255, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.1724137931034483, |
|
"grad_norm": 0.24596191852493296, |
|
"learning_rate": 9.843414637252615e-06, |
|
"loss": 1.0125, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.1772030651340996, |
|
"grad_norm": 0.2510254446995529, |
|
"learning_rate": 9.821969546584922e-06, |
|
"loss": 1.022, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.18199233716475097, |
|
"grad_norm": 0.2626288820781652, |
|
"learning_rate": 9.79917511122509e-06, |
|
"loss": 1.0016, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.1867816091954023, |
|
"grad_norm": 0.2588872211996587, |
|
"learning_rate": 9.775037709801206e-06, |
|
"loss": 1.0292, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.19157088122605365, |
|
"grad_norm": 0.28272889728543066, |
|
"learning_rate": 9.749564096747148e-06, |
|
"loss": 1.0255, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.19636015325670497, |
|
"grad_norm": 0.25691560926098406, |
|
"learning_rate": 9.722761400412496e-06, |
|
"loss": 1.0205, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.20114942528735633, |
|
"grad_norm": 0.31206698507075104, |
|
"learning_rate": 9.694637121067764e-06, |
|
"loss": 1.0018, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.20593869731800765, |
|
"grad_norm": 0.24725309050346184, |
|
"learning_rate": 9.6651991288056e-06, |
|
"loss": 1.013, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.210727969348659, |
|
"grad_norm": 0.2530356632425841, |
|
"learning_rate": 9.63445566133846e-06, |
|
"loss": 0.9921, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.21551724137931033, |
|
"grad_norm": 0.2784122469884546, |
|
"learning_rate": 9.602415321693434e-06, |
|
"loss": 1.0066, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.22030651340996169, |
|
"grad_norm": 0.3098632552557301, |
|
"learning_rate": 9.569087075804842e-06, |
|
"loss": 1.0062, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.22509578544061304, |
|
"grad_norm": 0.26451308039501314, |
|
"learning_rate": 9.534480250005263e-06, |
|
"loss": 0.9951, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.22988505747126436, |
|
"grad_norm": 0.27301629134605937, |
|
"learning_rate": 9.498604528415731e-06, |
|
"loss": 1.0347, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.23467432950191572, |
|
"grad_norm": 0.2891326880180606, |
|
"learning_rate": 9.461469950235795e-06, |
|
"loss": 1.0114, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.23946360153256704, |
|
"grad_norm": 0.26574747916476, |
|
"learning_rate": 9.423086906934228e-06, |
|
"loss": 1.0272, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.2442528735632184, |
|
"grad_norm": 0.26114773270931335, |
|
"learning_rate": 9.38346613934115e-06, |
|
"loss": 1.0039, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.24904214559386972, |
|
"grad_norm": 0.28383740993290196, |
|
"learning_rate": 9.342618734642395e-06, |
|
"loss": 1.0077, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.25383141762452105, |
|
"grad_norm": 0.28127484887890697, |
|
"learning_rate": 9.300556123276955e-06, |
|
"loss": 1.0306, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.25862068965517243, |
|
"grad_norm": 0.2702807344360773, |
|
"learning_rate": 9.257290075738365e-06, |
|
"loss": 0.9924, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.26340996168582376, |
|
"grad_norm": 0.2912366855364206, |
|
"learning_rate": 9.212832699280942e-06, |
|
"loss": 1.026, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.2681992337164751, |
|
"grad_norm": 0.30617702200806085, |
|
"learning_rate": 9.16719643453177e-06, |
|
"loss": 1.0247, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.27298850574712646, |
|
"grad_norm": 0.2590934020046758, |
|
"learning_rate": 9.120394052009412e-06, |
|
"loss": 1.0211, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.2777777777777778, |
|
"grad_norm": 0.27434400868197933, |
|
"learning_rate": 9.072438648550304e-06, |
|
"loss": 1.0243, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.2825670498084291, |
|
"grad_norm": 0.2813695079919259, |
|
"learning_rate": 9.023343643643821e-06, |
|
"loss": 1.0008, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.28735632183908044, |
|
"grad_norm": 0.2764043745947579, |
|
"learning_rate": 8.973122775677078e-06, |
|
"loss": 1.0066, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.2921455938697318, |
|
"grad_norm": 0.2993769190948342, |
|
"learning_rate": 8.921790098090477e-06, |
|
"loss": 1.015, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.29693486590038315, |
|
"grad_norm": 0.28718834695385625, |
|
"learning_rate": 8.869359975445085e-06, |
|
"loss": 1.0212, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.3017241379310345, |
|
"grad_norm": 0.3787907086651271, |
|
"learning_rate": 8.815847079402972e-06, |
|
"loss": 1.0079, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.3065134099616858, |
|
"grad_norm": 0.3012426621088511, |
|
"learning_rate": 8.761266384621599e-06, |
|
"loss": 1.0245, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.3113026819923372, |
|
"grad_norm": 0.545229044151294, |
|
"learning_rate": 8.705633164563413e-06, |
|
"loss": 1.0014, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.3160919540229885, |
|
"grad_norm": 0.3018913245413703, |
|
"learning_rate": 8.648962987221837e-06, |
|
"loss": 1.0035, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.32088122605363983, |
|
"grad_norm": 0.2937330943277994, |
|
"learning_rate": 8.591271710764839e-06, |
|
"loss": 0.9932, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.32567049808429116, |
|
"grad_norm": 0.30415781861083496, |
|
"learning_rate": 8.532575479097294e-06, |
|
"loss": 0.982, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.33045977011494254, |
|
"grad_norm": 0.2726446228184506, |
|
"learning_rate": 8.472890717343391e-06, |
|
"loss": 0.9992, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.33524904214559387, |
|
"grad_norm": 0.2841460417699516, |
|
"learning_rate": 8.412234127250353e-06, |
|
"loss": 1.0007, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.3400383141762452, |
|
"grad_norm": 0.313253280455998, |
|
"learning_rate": 8.350622682514735e-06, |
|
"loss": 0.9951, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.3448275862068966, |
|
"grad_norm": 0.33772538052784323, |
|
"learning_rate": 8.288073624032634e-06, |
|
"loss": 1.0169, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.3496168582375479, |
|
"grad_norm": 0.2810004404857808, |
|
"learning_rate": 8.224604455075115e-06, |
|
"loss": 1.0086, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.3544061302681992, |
|
"grad_norm": 0.2817546952474833, |
|
"learning_rate": 8.160232936390239e-06, |
|
"loss": 0.9888, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.35919540229885055, |
|
"grad_norm": 0.32316453097441, |
|
"learning_rate": 8.094977081233006e-06, |
|
"loss": 0.997, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.36398467432950193, |
|
"grad_norm": 0.2920029553424864, |
|
"learning_rate": 8.02885515032467e-06, |
|
"loss": 1.0172, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.36877394636015326, |
|
"grad_norm": 0.28736749227532504, |
|
"learning_rate": 7.961885646742793e-06, |
|
"loss": 1.0092, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.3735632183908046, |
|
"grad_norm": 0.29552008976856375, |
|
"learning_rate": 7.894087310743468e-06, |
|
"loss": 0.9952, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.3783524904214559, |
|
"grad_norm": 0.28037261402408636, |
|
"learning_rate": 7.825479114517197e-06, |
|
"loss": 1.0148, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.3831417624521073, |
|
"grad_norm": 0.3259544387907705, |
|
"learning_rate": 7.756080256879837e-06, |
|
"loss": 1.0172, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.3879310344827586, |
|
"grad_norm": 0.2866306787997861, |
|
"learning_rate": 7.685910157900158e-06, |
|
"loss": 0.9969, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.39272030651340994, |
|
"grad_norm": 0.27836252924957877, |
|
"learning_rate": 7.614988453465469e-06, |
|
"loss": 0.9981, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.3975095785440613, |
|
"grad_norm": 0.2881265537652179, |
|
"learning_rate": 7.5433349897868445e-06, |
|
"loss": 1.0075, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.40229885057471265, |
|
"grad_norm": 0.286210264814618, |
|
"learning_rate": 7.470969817845518e-06, |
|
"loss": 1.0025, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.407088122605364, |
|
"grad_norm": 0.28784515054514276, |
|
"learning_rate": 7.397913187781962e-06, |
|
"loss": 0.9918, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.4118773946360153, |
|
"grad_norm": 0.28225636678596183, |
|
"learning_rate": 7.324185543229226e-06, |
|
"loss": 1.0164, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.4166666666666667, |
|
"grad_norm": 0.30273570716493303, |
|
"learning_rate": 7.249807515592149e-06, |
|
"loss": 0.991, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.421455938697318, |
|
"grad_norm": 0.29638416230646264, |
|
"learning_rate": 7.174799918274018e-06, |
|
"loss": 1.0103, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.42624521072796934, |
|
"grad_norm": 0.27796285045314706, |
|
"learning_rate": 7.099183740852296e-06, |
|
"loss": 0.9929, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.43103448275862066, |
|
"grad_norm": 0.3024800017009751, |
|
"learning_rate": 7.022980143205046e-06, |
|
"loss": 0.9945, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.43582375478927204, |
|
"grad_norm": 0.308729839004599, |
|
"learning_rate": 6.946210449589714e-06, |
|
"loss": 1.0131, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.44061302681992337, |
|
"grad_norm": 0.29787252415350113, |
|
"learning_rate": 6.868896142675903e-06, |
|
"loss": 1.0053, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.4454022988505747, |
|
"grad_norm": 0.28555796101935293, |
|
"learning_rate": 6.791058857533814e-06, |
|
"loss": 1.0106, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.4501915708812261, |
|
"grad_norm": 0.27754653040108523, |
|
"learning_rate": 6.712720375580057e-06, |
|
"loss": 1.0127, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.4549808429118774, |
|
"grad_norm": 0.3064883256256046, |
|
"learning_rate": 6.633902618482484e-06, |
|
"loss": 1.0137, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.45977011494252873, |
|
"grad_norm": 0.2871757827962217, |
|
"learning_rate": 6.554627642025807e-06, |
|
"loss": 0.9808, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.46455938697318006, |
|
"grad_norm": 0.32016740057261645, |
|
"learning_rate": 6.474917629939652e-06, |
|
"loss": 1.0154, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.46934865900383144, |
|
"grad_norm": 0.28945157272232624, |
|
"learning_rate": 6.394794887690838e-06, |
|
"loss": 0.987, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.47413793103448276, |
|
"grad_norm": 0.45234637623400176, |
|
"learning_rate": 6.314281836241573e-06, |
|
"loss": 1.0072, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.4789272030651341, |
|
"grad_norm": 0.3043561526142259, |
|
"learning_rate": 6.233401005775339e-06, |
|
"loss": 0.9947, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.4837164750957854, |
|
"grad_norm": 0.3021339898048032, |
|
"learning_rate": 6.1521750293922035e-06, |
|
"loss": 1.0168, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.4885057471264368, |
|
"grad_norm": 0.262840671697266, |
|
"learning_rate": 6.070626636775349e-06, |
|
"loss": 0.9854, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.4932950191570881, |
|
"grad_norm": 0.2717384304425024, |
|
"learning_rate": 5.988778647830554e-06, |
|
"loss": 0.9847, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.49808429118773945, |
|
"grad_norm": 0.3005959006799932, |
|
"learning_rate": 5.906653966300444e-06, |
|
"loss": 1.0007, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.5028735632183908, |
|
"grad_norm": 0.2810709946318669, |
|
"learning_rate": 5.824275573355278e-06, |
|
"loss": 0.9891, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.5076628352490421, |
|
"grad_norm": 0.31324328313914135, |
|
"learning_rate": 5.741666521162055e-06, |
|
"loss": 1.0049, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.5124521072796935, |
|
"grad_norm": 0.29241234641844166, |
|
"learning_rate": 5.658849926433774e-06, |
|
"loss": 1.0019, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.5172413793103449, |
|
"grad_norm": 0.29226941844525284, |
|
"learning_rate": 5.575848963960621e-06, |
|
"loss": 0.9964, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.5220306513409961, |
|
"grad_norm": 0.2869562590824223, |
|
"learning_rate": 5.4926868601249e-06, |
|
"loss": 1.003, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.5268199233716475, |
|
"grad_norm": 0.2887086296052449, |
|
"learning_rate": 5.4093868864015405e-06, |
|
"loss": 0.9911, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.5316091954022989, |
|
"grad_norm": 0.2990975733324071, |
|
"learning_rate": 5.325972352845965e-06, |
|
"loss": 0.9961, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.5363984674329502, |
|
"grad_norm": 0.2964344269886325, |
|
"learning_rate": 5.24246660157119e-06, |
|
"loss": 1.0045, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.5411877394636015, |
|
"grad_norm": 0.3080697315686108, |
|
"learning_rate": 5.1588930002159255e-06, |
|
"loss": 0.9897, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.5459770114942529, |
|
"grad_norm": 0.3217970889035437, |
|
"learning_rate": 5.075274935405554e-06, |
|
"loss": 1.0022, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.5507662835249042, |
|
"grad_norm": 0.32311237689799727, |
|
"learning_rate": 4.991635806207788e-06, |
|
"loss": 0.9918, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.5555555555555556, |
|
"grad_norm": 0.3076150767243147, |
|
"learning_rate": 4.90799901758484e-06, |
|
"loss": 1.0156, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.5603448275862069, |
|
"grad_norm": 0.29579117268352634, |
|
"learning_rate": 4.824387973843957e-06, |
|
"loss": 0.9859, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.5651340996168582, |
|
"grad_norm": 0.27808702972070926, |
|
"learning_rate": 4.74082607208812e-06, |
|
"loss": 0.988, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.5699233716475096, |
|
"grad_norm": 0.2772865889207572, |
|
"learning_rate": 4.6573366956687885e-06, |
|
"loss": 1.0042, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.5747126436781609, |
|
"grad_norm": 0.2741988412251713, |
|
"learning_rate": 4.573943207642452e-06, |
|
"loss": 1.018, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.5795019157088123, |
|
"grad_norm": 0.3041843000888798, |
|
"learning_rate": 4.4906689442328935e-06, |
|
"loss": 1.0095, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.5842911877394636, |
|
"grad_norm": 0.3234627396285633, |
|
"learning_rate": 4.407537208300957e-06, |
|
"loss": 0.9981, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.5890804597701149, |
|
"grad_norm": 0.30361837699199173, |
|
"learning_rate": 4.3245712628236356e-06, |
|
"loss": 0.9945, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.5938697318007663, |
|
"grad_norm": 0.3068819956047309, |
|
"learning_rate": 4.241794324384334e-06, |
|
"loss": 0.9829, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.5986590038314177, |
|
"grad_norm": 0.3010857723276443, |
|
"learning_rate": 4.159229556676111e-06, |
|
"loss": 0.9778, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.603448275862069, |
|
"grad_norm": 0.3212249651416007, |
|
"learning_rate": 4.076900064019721e-06, |
|
"loss": 1.007, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.6082375478927203, |
|
"grad_norm": 0.2842660737481042, |
|
"learning_rate": 3.994828884898267e-06, |
|
"loss": 1.0056, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.6130268199233716, |
|
"grad_norm": 0.2887728882458368, |
|
"learning_rate": 3.91303898551028e-06, |
|
"loss": 1.0131, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.617816091954023, |
|
"grad_norm": 0.31374850764975326, |
|
"learning_rate": 3.8315532533430285e-06, |
|
"loss": 0.9979, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.6226053639846744, |
|
"grad_norm": 0.2990812859204113, |
|
"learning_rate": 3.7503944907678543e-06, |
|
"loss": 0.9979, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.6273946360153256, |
|
"grad_norm": 0.31174935012060845, |
|
"learning_rate": 3.6695854086593126e-06, |
|
"loss": 0.9907, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.632183908045977, |
|
"grad_norm": 0.30150077718156976, |
|
"learning_rate": 3.5891486200399413e-06, |
|
"loss": 0.9937, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.6369731800766284, |
|
"grad_norm": 0.304753149818871, |
|
"learning_rate": 3.509106633752387e-06, |
|
"loss": 1.0164, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.6417624521072797, |
|
"grad_norm": 0.2993950641947178, |
|
"learning_rate": 3.429481848160702e-06, |
|
"loss": 1.0093, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.646551724137931, |
|
"grad_norm": 0.3102107207132909, |
|
"learning_rate": 3.350296544882543e-06, |
|
"loss": 0.969, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.6513409961685823, |
|
"grad_norm": 0.27647028696280174, |
|
"learning_rate": 3.2715728825540525e-06, |
|
"loss": 1.0102, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.6561302681992337, |
|
"grad_norm": 0.30329159137172645, |
|
"learning_rate": 3.19333289062915e-06, |
|
"loss": 0.9992, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.6609195402298851, |
|
"grad_norm": 0.3233989260192753, |
|
"learning_rate": 3.1155984632149565e-06, |
|
"loss": 0.9984, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.6657088122605364, |
|
"grad_norm": 0.284592333087901, |
|
"learning_rate": 3.0383913529451286e-06, |
|
"loss": 1.0097, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.6704980842911877, |
|
"grad_norm": 0.284014773245373, |
|
"learning_rate": 2.961733164892744e-06, |
|
"loss": 1.0048, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.6752873563218391, |
|
"grad_norm": 0.29864923395210635, |
|
"learning_rate": 2.8856453505245018e-06, |
|
"loss": 1.008, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.6800766283524904, |
|
"grad_norm": 0.28455877244795186, |
|
"learning_rate": 2.8101492016979027e-06, |
|
"loss": 1.0082, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.6848659003831418, |
|
"grad_norm": 0.29297293441472344, |
|
"learning_rate": 2.7352658447030882e-06, |
|
"loss": 1.0137, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.6896551724137931, |
|
"grad_norm": 0.29774107321754356, |
|
"learning_rate": 2.6610162343510183e-06, |
|
"loss": 0.9878, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.6944444444444444, |
|
"grad_norm": 0.2886426546973218, |
|
"learning_rate": 2.587421148109619e-06, |
|
"loss": 0.9855, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.6992337164750958, |
|
"grad_norm": 0.30340545161406907, |
|
"learning_rate": 2.5145011802895835e-06, |
|
"loss": 1.004, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.7040229885057471, |
|
"grad_norm": 0.28629744556011416, |
|
"learning_rate": 2.4422767362814045e-06, |
|
"loss": 0.9935, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.7088122605363985, |
|
"grad_norm": 0.29480150488982737, |
|
"learning_rate": 2.370768026845276e-06, |
|
"loss": 1.0013, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.7136015325670498, |
|
"grad_norm": 0.30424158130358797, |
|
"learning_rate": 2.299995062455459e-06, |
|
"loss": 0.9932, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.7183908045977011, |
|
"grad_norm": 0.3179663498265742, |
|
"learning_rate": 2.2299776477007073e-06, |
|
"loss": 1.007, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.7231800766283525, |
|
"grad_norm": 0.303515618658323, |
|
"learning_rate": 2.16073537574229e-06, |
|
"loss": 0.9963, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.7279693486590039, |
|
"grad_norm": 0.30307569076630475, |
|
"learning_rate": 2.0922876228311833e-06, |
|
"loss": 0.9772, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.7327586206896551, |
|
"grad_norm": 0.3143565291039063, |
|
"learning_rate": 2.0246535428859652e-06, |
|
"loss": 0.9899, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.7375478927203065, |
|
"grad_norm": 0.28738666111079514, |
|
"learning_rate": 1.957852062132924e-06, |
|
"loss": 0.9848, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.7423371647509579, |
|
"grad_norm": 0.2850271965375348, |
|
"learning_rate": 1.8919018738098704e-06, |
|
"loss": 1.0076, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.7471264367816092, |
|
"grad_norm": 0.29940367717031596, |
|
"learning_rate": 1.8268214329351797e-06, |
|
"loss": 0.9864, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.7519157088122606, |
|
"grad_norm": 0.2957695153675884, |
|
"learning_rate": 1.762628951143454e-06, |
|
"loss": 0.9972, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.7567049808429118, |
|
"grad_norm": 0.2884019240313734, |
|
"learning_rate": 1.6993423915893241e-06, |
|
"loss": 0.9969, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.7614942528735632, |
|
"grad_norm": 0.30882836660635715, |
|
"learning_rate": 1.6369794639207626e-06, |
|
"loss": 1.0005, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.7662835249042146, |
|
"grad_norm": 0.2880996272538283, |
|
"learning_rate": 1.575557619323353e-06, |
|
"loss": 0.9853, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.7710727969348659, |
|
"grad_norm": 0.28083999464104503, |
|
"learning_rate": 1.5150940456368784e-06, |
|
"loss": 0.9579, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.7758620689655172, |
|
"grad_norm": 0.2977795145319976, |
|
"learning_rate": 1.4556056625455922e-06, |
|
"loss": 0.9944, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.7806513409961686, |
|
"grad_norm": 0.3044689990672244, |
|
"learning_rate": 1.3971091168435463e-06, |
|
"loss": 0.997, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.7854406130268199, |
|
"grad_norm": 0.2951506289424605, |
|
"learning_rate": 1.3396207777762732e-06, |
|
"loss": 1.0116, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.7902298850574713, |
|
"grad_norm": 0.284297818009739, |
|
"learning_rate": 1.2831567324601325e-06, |
|
"loss": 0.9792, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.7950191570881227, |
|
"grad_norm": 0.31450502170794314, |
|
"learning_rate": 1.2277327813806123e-06, |
|
"loss": 0.9927, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.7998084291187739, |
|
"grad_norm": 0.2896076523698672, |
|
"learning_rate": 1.173364433970835e-06, |
|
"loss": 0.9795, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.8045977011494253, |
|
"grad_norm": 0.2834770010415917, |
|
"learning_rate": 1.1200669042715163e-06, |
|
"loss": 0.9966, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.8093869731800766, |
|
"grad_norm": 0.34009482243032485, |
|
"learning_rate": 1.0678551066735671e-06, |
|
"loss": 0.9767, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.814176245210728, |
|
"grad_norm": 0.29936688994813515, |
|
"learning_rate": 1.0167436517445777e-06, |
|
"loss": 1.003, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.8189655172413793, |
|
"grad_norm": 0.29389391810900556, |
|
"learning_rate": 9.66746842140287e-07, |
|
"loss": 0.9888, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.8237547892720306, |
|
"grad_norm": 0.29840441886793445, |
|
"learning_rate": 9.178786686022417e-07, |
|
"loss": 1.0011, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.828544061302682, |
|
"grad_norm": 0.3050673773124508, |
|
"learning_rate": 8.701528060427194e-07, |
|
"loss": 0.9867, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.8333333333333334, |
|
"grad_norm": 0.2809885669064277, |
|
"learning_rate": 8.235826097180566e-07, |
|
"loss": 0.9802, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.8381226053639846, |
|
"grad_norm": 0.29288746432276136, |
|
"learning_rate": 7.781811114913995e-07, |
|
"loss": 0.9965, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.842911877394636, |
|
"grad_norm": 0.29110203494522685, |
|
"learning_rate": 7.339610161859618e-07, |
|
"loss": 0.9809, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.8477011494252874, |
|
"grad_norm": 0.30532295542101273, |
|
"learning_rate": 6.909346980298093e-07, |
|
"loss": 1.0039, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.8524904214559387, |
|
"grad_norm": 0.3178843529727934, |
|
"learning_rate": 6.49114197193137e-07, |
|
"loss": 0.9992, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.85727969348659, |
|
"grad_norm": 0.30030807864509074, |
|
"learning_rate": 6.085112164190466e-07, |
|
"loss": 0.9967, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.8620689655172413, |
|
"grad_norm": 0.283276426877559, |
|
"learning_rate": 5.691371177487215e-07, |
|
"loss": 0.9951, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.8668582375478927, |
|
"grad_norm": 0.2771047728402987, |
|
"learning_rate": 5.310029193419697e-07, |
|
"loss": 0.9823, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.8716475095785441, |
|
"grad_norm": 0.30200927801260424, |
|
"learning_rate": 4.941192923939769e-07, |
|
"loss": 0.9944, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.8764367816091954, |
|
"grad_norm": 0.29243200143497705, |
|
"learning_rate": 4.5849655814915683e-07, |
|
"loss": 0.9923, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.8812260536398467, |
|
"grad_norm": 0.38073762499381975, |
|
"learning_rate": 4.2414468501293217e-07, |
|
"loss": 0.9931, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.8860153256704981, |
|
"grad_norm": 0.28457589633356245, |
|
"learning_rate": 3.9107328576224736e-07, |
|
"loss": 0.9879, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.8908045977011494, |
|
"grad_norm": 0.29913915061920765, |
|
"learning_rate": 3.5929161485559694e-07, |
|
"loss": 1.0269, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.8955938697318008, |
|
"grad_norm": 0.28795408054019833, |
|
"learning_rate": 3.2880856584333043e-07, |
|
"loss": 0.984, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.9003831417624522, |
|
"grad_norm": 0.28456219117461123, |
|
"learning_rate": 2.9963266887894526e-07, |
|
"loss": 1.0007, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.9051724137931034, |
|
"grad_norm": 0.30655155945282825, |
|
"learning_rate": 2.717720883320685e-07, |
|
"loss": 1.0093, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.9099616858237548, |
|
"grad_norm": 0.28653578966574017, |
|
"learning_rate": 2.4523462050379864e-07, |
|
"loss": 0.9861, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.9147509578544061, |
|
"grad_norm": 0.30075605446410747, |
|
"learning_rate": 2.2002769144504943e-07, |
|
"loss": 0.997, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 0.9195402298850575, |
|
"grad_norm": 0.28984184857418177, |
|
"learning_rate": 1.9615835487849677e-07, |
|
"loss": 0.9772, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.9243295019157088, |
|
"grad_norm": 0.2971882152661699, |
|
"learning_rate": 1.7363329022471564e-07, |
|
"loss": 1.0125, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 0.9291187739463601, |
|
"grad_norm": 0.30026576782404996, |
|
"learning_rate": 1.5245880073305963e-07, |
|
"loss": 1.0128, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.9339080459770115, |
|
"grad_norm": 0.2971224564720333, |
|
"learning_rate": 1.3264081171780797e-07, |
|
"loss": 1.0114, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.9386973180076629, |
|
"grad_norm": 0.278422035094591, |
|
"learning_rate": 1.1418486890006574e-07, |
|
"loss": 0.982, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.9434865900383141, |
|
"grad_norm": 0.29059990973676236, |
|
"learning_rate": 9.709613685589314e-08, |
|
"loss": 0.998, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 0.9482758620689655, |
|
"grad_norm": 0.29092073403074437, |
|
"learning_rate": 8.137939757108526e-08, |
|
"loss": 1.011, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.9530651340996169, |
|
"grad_norm": 0.2923297366913393, |
|
"learning_rate": 6.703904910301929e-08, |
|
"loss": 0.9656, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 0.9578544061302682, |
|
"grad_norm": 0.2931478943843858, |
|
"learning_rate": 5.4079104349929465e-08, |
|
"loss": 1.0036, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.9626436781609196, |
|
"grad_norm": 0.2923376776086664, |
|
"learning_rate": 4.250318992797375e-08, |
|
"loss": 1.0083, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 0.9674329501915708, |
|
"grad_norm": 0.2863912260830555, |
|
"learning_rate": 3.231454515638221e-08, |
|
"loss": 0.9955, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.9722222222222222, |
|
"grad_norm": 0.2974177186982834, |
|
"learning_rate": 2.351602115099272e-08, |
|
"loss": 0.9865, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 0.9770114942528736, |
|
"grad_norm": 0.29561321400349233, |
|
"learning_rate": 1.6110080026414123e-08, |
|
"loss": 1.0083, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.9818007662835249, |
|
"grad_norm": 0.29932685669861786, |
|
"learning_rate": 1.0098794207047402e-08, |
|
"loss": 1.0118, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.9865900383141762, |
|
"grad_norm": 0.29155813115628504, |
|
"learning_rate": 5.483845847151226e-09, |
|
"loss": 0.9846, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.9913793103448276, |
|
"grad_norm": 0.2910040393119768, |
|
"learning_rate": 2.2665263601240328e-09, |
|
"loss": 0.9812, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 0.9961685823754789, |
|
"grad_norm": 0.28919827046140006, |
|
"learning_rate": 4.4773605712089554e-10, |
|
"loss": 1.0115, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_runtime": 6595.9355, |
|
"eval_samples_per_second": 3.504, |
|
"eval_steps_per_second": 0.876, |
|
"step": 1044 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 1044, |
|
"total_flos": 1940427569627136.0, |
|
"train_loss": 1.0122476654034465, |
|
"train_runtime": 20247.7008, |
|
"train_samples_per_second": 3.298, |
|
"train_steps_per_second": 0.052 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 1044, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1940427569627136.0, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|