|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.9995619798510732, |
|
"eval_steps": 500, |
|
"global_step": 1141, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0008760402978537013, |
|
"grad_norm": 247.023226046774, |
|
"learning_rate": 1.7391304347826088e-07, |
|
"loss": 8.8281, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.004380201489268506, |
|
"grad_norm": 232.10560021267386, |
|
"learning_rate": 8.695652173913044e-07, |
|
"loss": 8.7637, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.008760402978537012, |
|
"grad_norm": 197.45293542519394, |
|
"learning_rate": 1.7391304347826088e-06, |
|
"loss": 8.5789, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.013140604467805518, |
|
"grad_norm": 69.72380158032517, |
|
"learning_rate": 2.6086956521739132e-06, |
|
"loss": 7.607, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.017520805957074025, |
|
"grad_norm": 40.79010585762893, |
|
"learning_rate": 3.4782608695652175e-06, |
|
"loss": 6.5969, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.021901007446342532, |
|
"grad_norm": 31.021399931633685, |
|
"learning_rate": 4.347826086956522e-06, |
|
"loss": 5.3516, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.026281208935611037, |
|
"grad_norm": 15.829514000742273, |
|
"learning_rate": 5.2173913043478265e-06, |
|
"loss": 4.018, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.030661410424879545, |
|
"grad_norm": 12.017912967930332, |
|
"learning_rate": 6.086956521739132e-06, |
|
"loss": 2.9813, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.03504161191414805, |
|
"grad_norm": 2.8893117110423487, |
|
"learning_rate": 6.956521739130435e-06, |
|
"loss": 2.1303, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.03942181340341656, |
|
"grad_norm": 1.4949499944110431, |
|
"learning_rate": 7.82608695652174e-06, |
|
"loss": 1.8031, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.043802014892685065, |
|
"grad_norm": 0.7508052500953509, |
|
"learning_rate": 8.695652173913044e-06, |
|
"loss": 1.5938, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.04818221638195357, |
|
"grad_norm": 0.4902022246315334, |
|
"learning_rate": 9.565217391304349e-06, |
|
"loss": 1.5084, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.052562417871222074, |
|
"grad_norm": 0.42532974187985395, |
|
"learning_rate": 1.0434782608695653e-05, |
|
"loss": 1.4203, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.05694261936049058, |
|
"grad_norm": 0.3463140884611265, |
|
"learning_rate": 1.1304347826086957e-05, |
|
"loss": 1.3941, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.06132282084975909, |
|
"grad_norm": 0.3189213985873913, |
|
"learning_rate": 1.2173913043478263e-05, |
|
"loss": 1.3566, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.0657030223390276, |
|
"grad_norm": 0.2996468490222568, |
|
"learning_rate": 1.3043478260869566e-05, |
|
"loss": 1.3418, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.0700832238282961, |
|
"grad_norm": 0.26712973341018265, |
|
"learning_rate": 1.391304347826087e-05, |
|
"loss": 1.3201, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.07446342531756461, |
|
"grad_norm": 0.2668394062417099, |
|
"learning_rate": 1.4782608695652174e-05, |
|
"loss": 1.3143, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.07884362680683311, |
|
"grad_norm": 0.26274351296407206, |
|
"learning_rate": 1.565217391304348e-05, |
|
"loss": 1.3031, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.08322382829610162, |
|
"grad_norm": 0.25741452098888334, |
|
"learning_rate": 1.6521739130434785e-05, |
|
"loss": 1.2848, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.08760402978537013, |
|
"grad_norm": 0.2441325116229632, |
|
"learning_rate": 1.739130434782609e-05, |
|
"loss": 1.2732, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.09198423127463863, |
|
"grad_norm": 0.24162625924392253, |
|
"learning_rate": 1.8260869565217393e-05, |
|
"loss": 1.2715, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.09636443276390715, |
|
"grad_norm": 0.2519915389487053, |
|
"learning_rate": 1.9130434782608697e-05, |
|
"loss": 1.2711, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.10074463425317565, |
|
"grad_norm": 0.2465793202977791, |
|
"learning_rate": 2e-05, |
|
"loss": 1.2459, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.10512483574244415, |
|
"grad_norm": 0.23497052771629895, |
|
"learning_rate": 1.99988280568259e-05, |
|
"loss": 1.248, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.10950503723171266, |
|
"grad_norm": 0.24553176964100193, |
|
"learning_rate": 1.9995312501993765e-05, |
|
"loss": 1.2238, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.11388523872098116, |
|
"grad_norm": 0.23370960048474732, |
|
"learning_rate": 1.998945415950969e-05, |
|
"loss": 1.248, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.11826544021024968, |
|
"grad_norm": 0.24397346786614016, |
|
"learning_rate": 1.9981254402502568e-05, |
|
"loss": 1.2063, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.12264564169951818, |
|
"grad_norm": 0.27495355893049633, |
|
"learning_rate": 1.9970715152902257e-05, |
|
"loss": 1.2381, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.12702584318878668, |
|
"grad_norm": 0.25282538712904246, |
|
"learning_rate": 1.9957838880989076e-05, |
|
"loss": 1.2281, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.1314060446780552, |
|
"grad_norm": 0.22995731363863653, |
|
"learning_rate": 1.9942628604814827e-05, |
|
"loss": 1.2051, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.1357862461673237, |
|
"grad_norm": 0.23386474611286825, |
|
"learning_rate": 1.9925087889495374e-05, |
|
"loss": 1.2141, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.1401664476565922, |
|
"grad_norm": 0.2309649578398331, |
|
"learning_rate": 1.990522084637503e-05, |
|
"loss": 1.1822, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.1445466491458607, |
|
"grad_norm": 0.24003805632416436, |
|
"learning_rate": 1.9883032132062926e-05, |
|
"loss": 1.2063, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.14892685063512923, |
|
"grad_norm": 0.24338387371805312, |
|
"learning_rate": 1.98585269473415e-05, |
|
"loss": 1.1879, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.1533070521243977, |
|
"grad_norm": 0.23476112374060998, |
|
"learning_rate": 1.9831711035947552e-05, |
|
"loss": 1.2031, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.15768725361366623, |
|
"grad_norm": 0.22915392045241134, |
|
"learning_rate": 1.9802590683225945e-05, |
|
"loss": 1.1943, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.16206745510293474, |
|
"grad_norm": 0.24585515208967154, |
|
"learning_rate": 1.97711727146564e-05, |
|
"loss": 1.1961, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.16644765659220323, |
|
"grad_norm": 0.21471505934692195, |
|
"learning_rate": 1.973746449425368e-05, |
|
"loss": 1.1883, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.17082785808147175, |
|
"grad_norm": 0.22942379356513737, |
|
"learning_rate": 1.970147392284154e-05, |
|
"loss": 1.2193, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.17520805957074026, |
|
"grad_norm": 0.23244007409565076, |
|
"learning_rate": 1.9663209436200887e-05, |
|
"loss": 1.207, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.17958826106000875, |
|
"grad_norm": 0.22422374010740104, |
|
"learning_rate": 1.9622680003092503e-05, |
|
"loss": 1.174, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.18396846254927726, |
|
"grad_norm": 0.2239852274530115, |
|
"learning_rate": 1.957989512315489e-05, |
|
"loss": 1.1766, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.18834866403854578, |
|
"grad_norm": 0.22417592332949882, |
|
"learning_rate": 1.953486482467764e-05, |
|
"loss": 1.2016, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.1927288655278143, |
|
"grad_norm": 0.23718560975504316, |
|
"learning_rate": 1.9487599662250945e-05, |
|
"loss": 1.1836, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.19710906701708278, |
|
"grad_norm": 0.24831579992264016, |
|
"learning_rate": 1.9438110714291697e-05, |
|
"loss": 1.1885, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.2014892685063513, |
|
"grad_norm": 0.22705440610322433, |
|
"learning_rate": 1.9386409580446846e-05, |
|
"loss": 1.1768, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.2058694699956198, |
|
"grad_norm": 0.22232082846727982, |
|
"learning_rate": 1.933250837887457e-05, |
|
"loss": 1.1695, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.2102496714848883, |
|
"grad_norm": 0.20849807972480366, |
|
"learning_rate": 1.9276419743403934e-05, |
|
"loss": 1.1664, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.2146298729741568, |
|
"grad_norm": 0.22884959889218245, |
|
"learning_rate": 1.9218156820573618e-05, |
|
"loss": 1.1715, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.21901007446342532, |
|
"grad_norm": 0.22251029198025443, |
|
"learning_rate": 1.9157733266550577e-05, |
|
"loss": 1.1664, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.2233902759526938, |
|
"grad_norm": 0.21607317255636946, |
|
"learning_rate": 1.9095163243929143e-05, |
|
"loss": 1.1758, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.22777047744196233, |
|
"grad_norm": 0.2275376818247485, |
|
"learning_rate": 1.9030461418411498e-05, |
|
"loss": 1.173, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.23215067893123084, |
|
"grad_norm": 0.2289840446661591, |
|
"learning_rate": 1.8963642955370203e-05, |
|
"loss": 1.1852, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.23653088042049936, |
|
"grad_norm": 0.23768451520157288, |
|
"learning_rate": 1.889472351629358e-05, |
|
"loss": 1.1551, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.24091108190976784, |
|
"grad_norm": 0.22459484609481822, |
|
"learning_rate": 1.882371925511488e-05, |
|
"loss": 1.1564, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.24529128339903636, |
|
"grad_norm": 0.22828454066610995, |
|
"learning_rate": 1.875064681442594e-05, |
|
"loss": 1.1838, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.24967148488830487, |
|
"grad_norm": 0.23814414474493634, |
|
"learning_rate": 1.867552332157637e-05, |
|
"loss": 1.165, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.25405168637757336, |
|
"grad_norm": 0.23695582890788727, |
|
"learning_rate": 1.8598366384659113e-05, |
|
"loss": 1.1496, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.25843188786684185, |
|
"grad_norm": 0.21285197719762172, |
|
"learning_rate": 1.851919408838327e-05, |
|
"loss": 1.1671, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.2628120893561104, |
|
"grad_norm": 0.23402098820515907, |
|
"learning_rate": 1.843802498983529e-05, |
|
"loss": 1.1662, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.2671922908453789, |
|
"grad_norm": 0.21951316193224032, |
|
"learning_rate": 1.8354878114129368e-05, |
|
"loss": 1.1623, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.2715724923346474, |
|
"grad_norm": 0.21226630310736813, |
|
"learning_rate": 1.8269772949948185e-05, |
|
"loss": 1.1571, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.2759526938239159, |
|
"grad_norm": 0.21300490616956957, |
|
"learning_rate": 1.8182729444974993e-05, |
|
"loss": 1.1729, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.2803328953131844, |
|
"grad_norm": 0.2298100232081548, |
|
"learning_rate": 1.8093768001218096e-05, |
|
"loss": 1.1648, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.28471309680245294, |
|
"grad_norm": 0.20939381292392953, |
|
"learning_rate": 1.800290947022884e-05, |
|
"loss": 1.1633, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.2890932982917214, |
|
"grad_norm": 0.21652397528724918, |
|
"learning_rate": 1.7910175148214274e-05, |
|
"loss": 1.1703, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.2934734997809899, |
|
"grad_norm": 0.23693799112996852, |
|
"learning_rate": 1.7815586771045535e-05, |
|
"loss": 1.1557, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.29785370127025845, |
|
"grad_norm": 0.23875037210600733, |
|
"learning_rate": 1.771916650916321e-05, |
|
"loss": 1.1546, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.30223390275952694, |
|
"grad_norm": 0.2086246080725189, |
|
"learning_rate": 1.762093696238086e-05, |
|
"loss": 1.1414, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.3066141042487954, |
|
"grad_norm": 0.2032809446181252, |
|
"learning_rate": 1.752092115458784e-05, |
|
"loss": 1.1557, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.31099430573806397, |
|
"grad_norm": 0.2299671919872682, |
|
"learning_rate": 1.7419142528352815e-05, |
|
"loss": 1.1584, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.31537450722733246, |
|
"grad_norm": 0.2208475255216932, |
|
"learning_rate": 1.731562493942904e-05, |
|
"loss": 1.1436, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.31975470871660094, |
|
"grad_norm": 0.21527409406555012, |
|
"learning_rate": 1.721039265116285e-05, |
|
"loss": 1.1592, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.3241349102058695, |
|
"grad_norm": 0.2075937240152717, |
|
"learning_rate": 1.710347032880664e-05, |
|
"loss": 1.1332, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.328515111695138, |
|
"grad_norm": 0.2155228439936621, |
|
"learning_rate": 1.6994883033737582e-05, |
|
"loss": 1.1395, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.33289531318440646, |
|
"grad_norm": 0.24834287218683154, |
|
"learning_rate": 1.688465621758352e-05, |
|
"loss": 1.1566, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.337275514673675, |
|
"grad_norm": 0.2401994560404927, |
|
"learning_rate": 1.6772815716257414e-05, |
|
"loss": 1.1582, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.3416557161629435, |
|
"grad_norm": 0.22010411419127637, |
|
"learning_rate": 1.6659387743901688e-05, |
|
"loss": 1.1479, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.346035917652212, |
|
"grad_norm": 0.2141467823787157, |
|
"learning_rate": 1.6544398886743934e-05, |
|
"loss": 1.1424, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.3504161191414805, |
|
"grad_norm": 0.20902762397068814, |
|
"learning_rate": 1.6427876096865394e-05, |
|
"loss": 1.1401, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.354796320630749, |
|
"grad_norm": 0.20864014440590625, |
|
"learning_rate": 1.6309846685883726e-05, |
|
"loss": 1.1404, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.3591765221200175, |
|
"grad_norm": 0.22432181453053213, |
|
"learning_rate": 1.6190338318551426e-05, |
|
"loss": 1.1539, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.36355672360928604, |
|
"grad_norm": 0.20427723583571802, |
|
"learning_rate": 1.606937900627157e-05, |
|
"loss": 1.1402, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.3679369250985545, |
|
"grad_norm": 0.20345261931074785, |
|
"learning_rate": 1.594699710053223e-05, |
|
"loss": 1.1454, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.37231712658782307, |
|
"grad_norm": 0.20907937476304458, |
|
"learning_rate": 1.5823221286261217e-05, |
|
"loss": 1.1282, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.37669732807709155, |
|
"grad_norm": 0.2065934957421514, |
|
"learning_rate": 1.5698080575102662e-05, |
|
"loss": 1.1334, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.38107752956636004, |
|
"grad_norm": 0.2161916996533843, |
|
"learning_rate": 1.557160429861702e-05, |
|
"loss": 1.152, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.3854577310556286, |
|
"grad_norm": 0.21804609517909665, |
|
"learning_rate": 1.5443822101406066e-05, |
|
"loss": 1.1379, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.38983793254489707, |
|
"grad_norm": 0.20636895330872523, |
|
"learning_rate": 1.531476393416456e-05, |
|
"loss": 1.1592, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.39421813403416556, |
|
"grad_norm": 0.20437911994063385, |
|
"learning_rate": 1.5184460046660139e-05, |
|
"loss": 1.1277, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.3985983355234341, |
|
"grad_norm": 0.2060999428500134, |
|
"learning_rate": 1.50529409806431e-05, |
|
"loss": 1.1453, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.4029785370127026, |
|
"grad_norm": 0.22533191644861675, |
|
"learning_rate": 1.4920237562687784e-05, |
|
"loss": 1.1502, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.4073587385019711, |
|
"grad_norm": 0.23260594557785327, |
|
"learning_rate": 1.478638089696716e-05, |
|
"loss": 1.1248, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.4117389399912396, |
|
"grad_norm": 0.22947787094553646, |
|
"learning_rate": 1.4651402357962368e-05, |
|
"loss": 1.1318, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.4161191414805081, |
|
"grad_norm": 0.2009264544016685, |
|
"learning_rate": 1.4515333583108896e-05, |
|
"loss": 1.1543, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.4204993429697766, |
|
"grad_norm": 0.20022135464643717, |
|
"learning_rate": 1.4378206465381122e-05, |
|
"loss": 1.1405, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.42487954445904513, |
|
"grad_norm": 0.20457278775552865, |
|
"learning_rate": 1.4240053145816968e-05, |
|
"loss": 1.1434, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.4292597459483136, |
|
"grad_norm": 0.20324616627736056, |
|
"learning_rate": 1.4100906005984404e-05, |
|
"loss": 1.1479, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.4336399474375821, |
|
"grad_norm": 0.21908018025065346, |
|
"learning_rate": 1.396079766039157e-05, |
|
"loss": 1.1449, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.43802014892685065, |
|
"grad_norm": 0.210033035128553, |
|
"learning_rate": 1.381976094884232e-05, |
|
"loss": 1.132, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.44240035041611914, |
|
"grad_norm": 0.2013972756453919, |
|
"learning_rate": 1.3677828928738934e-05, |
|
"loss": 1.1473, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.4467805519053876, |
|
"grad_norm": 0.20820144402446275, |
|
"learning_rate": 1.3535034867333838e-05, |
|
"loss": 1.1398, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.45116075339465617, |
|
"grad_norm": 0.21960485445473304, |
|
"learning_rate": 1.3391412233932148e-05, |
|
"loss": 1.1527, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.45554095488392465, |
|
"grad_norm": 0.21395914429214383, |
|
"learning_rate": 1.3246994692046837e-05, |
|
"loss": 1.1486, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.45992115637319314, |
|
"grad_norm": 0.21230485993764248, |
|
"learning_rate": 1.3101816091508389e-05, |
|
"loss": 1.1537, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.4643013578624617, |
|
"grad_norm": 0.21651837542900804, |
|
"learning_rate": 1.2955910460530787e-05, |
|
"loss": 1.1496, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.46868155935173017, |
|
"grad_norm": 0.20750629246921734, |
|
"learning_rate": 1.2809311997735697e-05, |
|
"loss": 1.1355, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.4730617608409987, |
|
"grad_norm": 0.2039566928608635, |
|
"learning_rate": 1.266205506413667e-05, |
|
"loss": 1.1375, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.4774419623302672, |
|
"grad_norm": 0.20670703932671342, |
|
"learning_rate": 1.2514174175085346e-05, |
|
"loss": 1.1441, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.4818221638195357, |
|
"grad_norm": 0.20394796072197643, |
|
"learning_rate": 1.2365703992181425e-05, |
|
"loss": 1.1192, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.48620236530880423, |
|
"grad_norm": 0.19785902077082673, |
|
"learning_rate": 1.2216679315148388e-05, |
|
"loss": 1.1303, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.4905825667980727, |
|
"grad_norm": 0.20446955932426367, |
|
"learning_rate": 1.2067135073676841e-05, |
|
"loss": 1.1327, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.4949627682873412, |
|
"grad_norm": 0.20819551033939337, |
|
"learning_rate": 1.1917106319237386e-05, |
|
"loss": 1.1262, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.49934296977660975, |
|
"grad_norm": 0.202514436311041, |
|
"learning_rate": 1.1766628216864961e-05, |
|
"loss": 1.1451, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.5037231712658782, |
|
"grad_norm": 0.20270248880014877, |
|
"learning_rate": 1.161573603691655e-05, |
|
"loss": 1.1096, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.5081033727551467, |
|
"grad_norm": 0.19902608440066913, |
|
"learning_rate": 1.1464465146804218e-05, |
|
"loss": 1.1429, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.5124835742444153, |
|
"grad_norm": 0.21488103670405032, |
|
"learning_rate": 1.1312851002705383e-05, |
|
"loss": 1.1258, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.5168637757336837, |
|
"grad_norm": 0.2079533279978145, |
|
"learning_rate": 1.1160929141252303e-05, |
|
"loss": 1.1326, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.5212439772229522, |
|
"grad_norm": 0.2024904670673594, |
|
"learning_rate": 1.1008735171202685e-05, |
|
"loss": 1.1516, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.5256241787122208, |
|
"grad_norm": 0.2009798683547318, |
|
"learning_rate": 1.0856304765093391e-05, |
|
"loss": 1.126, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.5300043802014893, |
|
"grad_norm": 0.20247591682210167, |
|
"learning_rate": 1.0703673650879219e-05, |
|
"loss": 1.1182, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.5343845816907578, |
|
"grad_norm": 0.2040201994565877, |
|
"learning_rate": 1.0550877603558656e-05, |
|
"loss": 1.1406, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.5387647831800263, |
|
"grad_norm": 0.2015454994775207, |
|
"learning_rate": 1.0397952436788643e-05, |
|
"loss": 1.1166, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.5431449846692948, |
|
"grad_norm": 0.19828586396121736, |
|
"learning_rate": 1.024493399449025e-05, |
|
"loss": 1.1271, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.5475251861585633, |
|
"grad_norm": 0.202990185981543, |
|
"learning_rate": 1.0091858142447266e-05, |
|
"loss": 1.1404, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.5519053876478318, |
|
"grad_norm": 0.1953396887915411, |
|
"learning_rate": 9.938760759899674e-06, |
|
"loss": 1.1282, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.5562855891371004, |
|
"grad_norm": 0.19897584069638927, |
|
"learning_rate": 9.785677731133972e-06, |
|
"loss": 1.116, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.5606657906263688, |
|
"grad_norm": 0.19579937491086502, |
|
"learning_rate": 9.632644937072277e-06, |
|
"loss": 1.1211, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.5650459921156373, |
|
"grad_norm": 0.19292547383921327, |
|
"learning_rate": 9.479698246862277e-06, |
|
"loss": 1.1249, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.5694261936049059, |
|
"grad_norm": 0.2035897790907176, |
|
"learning_rate": 9.326873509469887e-06, |
|
"loss": 1.1222, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.5738063950941743, |
|
"grad_norm": 0.20245065608346366, |
|
"learning_rate": 9.174206545276678e-06, |
|
"loss": 1.1354, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.5781865965834428, |
|
"grad_norm": 0.19934269310500388, |
|
"learning_rate": 9.021733137683963e-06, |
|
"loss": 1.1443, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.5825667980727114, |
|
"grad_norm": 0.19598319523008984, |
|
"learning_rate": 8.869489024725595e-06, |
|
"loss": 1.1326, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.5869469995619798, |
|
"grad_norm": 0.1945281080977394, |
|
"learning_rate": 8.717509890691369e-06, |
|
"loss": 1.1239, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.5913272010512484, |
|
"grad_norm": 0.20806507918308362, |
|
"learning_rate": 8.565831357763039e-06, |
|
"loss": 1.1294, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.5957074025405169, |
|
"grad_norm": 0.20251893056570133, |
|
"learning_rate": 8.414488977664858e-06, |
|
"loss": 1.1239, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.6000876040297853, |
|
"grad_norm": 0.1991936009015444, |
|
"learning_rate": 8.263518223330698e-06, |
|
"loss": 1.1322, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.6044678055190539, |
|
"grad_norm": 0.1961163554964441, |
|
"learning_rate": 8.112954480589558e-06, |
|
"loss": 1.127, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.6088480070083224, |
|
"grad_norm": 0.1939932601906298, |
|
"learning_rate": 7.962833039871562e-06, |
|
"loss": 1.1275, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.6132282084975909, |
|
"grad_norm": 0.19487292794675937, |
|
"learning_rate": 7.813189087936243e-06, |
|
"loss": 1.1222, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.6176084099868594, |
|
"grad_norm": 0.19843142347965212, |
|
"learning_rate": 7.664057699625215e-06, |
|
"loss": 1.1509, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.6219886114761279, |
|
"grad_norm": 0.20098071367810383, |
|
"learning_rate": 7.515473829640987e-06, |
|
"loss": 1.1228, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.6263688129653964, |
|
"grad_norm": 0.1972827327140634, |
|
"learning_rate": 7.367472304354011e-06, |
|
"loss": 1.1268, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.6307490144546649, |
|
"grad_norm": 0.20576577812522315, |
|
"learning_rate": 7.2200878136397355e-06, |
|
"loss": 1.115, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.6351292159439335, |
|
"grad_norm": 0.1972932881390162, |
|
"learning_rate": 7.073354902747742e-06, |
|
"loss": 1.1104, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.6395094174332019, |
|
"grad_norm": 0.20208492611746368, |
|
"learning_rate": 6.927307964204695e-06, |
|
"loss": 1.1356, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.6438896189224704, |
|
"grad_norm": 0.19600730585362314, |
|
"learning_rate": 6.781981229753145e-06, |
|
"loss": 1.1225, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.648269820411739, |
|
"grad_norm": 0.19661439451248647, |
|
"learning_rate": 6.637408762327972e-06, |
|
"loss": 1.1175, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.6526500219010074, |
|
"grad_norm": 0.20253732148210266, |
|
"learning_rate": 6.4936244480724575e-06, |
|
"loss": 1.1224, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.657030223390276, |
|
"grad_norm": 0.197391894393827, |
|
"learning_rate": 6.350661988395723e-06, |
|
"loss": 1.1369, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.6614104248795445, |
|
"grad_norm": 0.19396216640485134, |
|
"learning_rate": 6.208554892073528e-06, |
|
"loss": 1.1332, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.6657906263688129, |
|
"grad_norm": 0.19950066877821537, |
|
"learning_rate": 6.067336467394169e-06, |
|
"loss": 1.1199, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.6701708278580815, |
|
"grad_norm": 0.20943729884482393, |
|
"learning_rate": 5.927039814351426e-06, |
|
"loss": 1.1201, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.67455102934735, |
|
"grad_norm": 0.19237438150257924, |
|
"learning_rate": 5.787697816886273e-06, |
|
"loss": 1.1267, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.6789312308366184, |
|
"grad_norm": 0.19252386131062685, |
|
"learning_rate": 5.649343135179271e-06, |
|
"loss": 1.11, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.683311432325887, |
|
"grad_norm": 0.19638246872011564, |
|
"learning_rate": 5.512008197995379e-06, |
|
"loss": 1.1164, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.6876916338151555, |
|
"grad_norm": 0.1877681729558122, |
|
"learning_rate": 5.375725195083046e-06, |
|
"loss": 1.1257, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.692071835304424, |
|
"grad_norm": 0.19133651997309503, |
|
"learning_rate": 5.240526069629265e-06, |
|
"loss": 1.134, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.6964520367936925, |
|
"grad_norm": 0.19745307442946453, |
|
"learning_rate": 5.106442510772489e-06, |
|
"loss": 1.1412, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.700832238282961, |
|
"grad_norm": 0.18826987382961585, |
|
"learning_rate": 4.97350594617502e-06, |
|
"loss": 1.12, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.7052124397722295, |
|
"grad_norm": 0.20198435756216934, |
|
"learning_rate": 4.8417475346567635e-06, |
|
"loss": 1.1343, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.709592641261498, |
|
"grad_norm": 0.19941557715707345, |
|
"learning_rate": 4.711198158891909e-06, |
|
"loss": 1.1229, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.7139728427507666, |
|
"grad_norm": 0.20442006319949774, |
|
"learning_rate": 4.581888418170429e-06, |
|
"loss": 1.1353, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.718353044240035, |
|
"grad_norm": 0.2001055609078045, |
|
"learning_rate": 4.453848621225913e-06, |
|
"loss": 1.1149, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.7227332457293035, |
|
"grad_norm": 0.18925787454519752, |
|
"learning_rate": 4.327108779131573e-06, |
|
"loss": 1.133, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.7271134472185721, |
|
"grad_norm": 0.1863143936478909, |
|
"learning_rate": 4.201698598265973e-06, |
|
"loss": 1.1162, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.7314936487078406, |
|
"grad_norm": 0.1879694577872159, |
|
"learning_rate": 4.077647473350201e-06, |
|
"loss": 1.126, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.735873850197109, |
|
"grad_norm": 0.1919436380192233, |
|
"learning_rate": 3.954984480558071e-06, |
|
"loss": 1.1195, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.7402540516863776, |
|
"grad_norm": 0.20358611131495935, |
|
"learning_rate": 3.83373837070101e-06, |
|
"loss": 1.1255, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.7446342531756461, |
|
"grad_norm": 0.20008378915370192, |
|
"learning_rate": 3.7139375624891795e-06, |
|
"loss": 1.1412, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.7490144546649146, |
|
"grad_norm": 0.19290096180375577, |
|
"learning_rate": 3.595610135870472e-06, |
|
"loss": 1.1207, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.7533946561541831, |
|
"grad_norm": 0.196072459641094, |
|
"learning_rate": 3.478783825448869e-06, |
|
"loss": 1.1125, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.7577748576434516, |
|
"grad_norm": 0.1928642029524879, |
|
"learning_rate": 3.3634860139837877e-06, |
|
"loss": 1.1229, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.7621550591327201, |
|
"grad_norm": 0.20130571120198174, |
|
"learning_rate": 3.249743725971849e-06, |
|
"loss": 1.1312, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.7665352606219886, |
|
"grad_norm": 0.20961643180726597, |
|
"learning_rate": 3.1375836213126653e-06, |
|
"loss": 1.1232, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.7709154621112572, |
|
"grad_norm": 0.18941668671789638, |
|
"learning_rate": 3.0270319890600465e-06, |
|
"loss": 1.1434, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.7752956636005256, |
|
"grad_norm": 0.19381237131051324, |
|
"learning_rate": 2.918114741260156e-06, |
|
"loss": 1.1439, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.7796758650897941, |
|
"grad_norm": 0.1945291975485229, |
|
"learning_rate": 2.8108574068780093e-06, |
|
"loss": 1.1158, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.7840560665790627, |
|
"grad_norm": 0.18927206178983724, |
|
"learning_rate": 2.7052851258137936e-06, |
|
"loss": 1.1241, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.7884362680683311, |
|
"grad_norm": 0.19223522737541443, |
|
"learning_rate": 2.601422643010335e-06, |
|
"loss": 1.1056, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.7928164695575997, |
|
"grad_norm": 0.20045874340313333, |
|
"learning_rate": 2.4992943026531935e-06, |
|
"loss": 1.1239, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.7971966710468682, |
|
"grad_norm": 0.19018510655063894, |
|
"learning_rate": 2.3989240424646355e-06, |
|
"loss": 1.1189, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.8015768725361366, |
|
"grad_norm": 0.19034937418838804, |
|
"learning_rate": 2.300335388092929e-06, |
|
"loss": 1.1152, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.8059570740254052, |
|
"grad_norm": 0.19645900089237364, |
|
"learning_rate": 2.2035514475981756e-06, |
|
"loss": 1.1313, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.8103372755146737, |
|
"grad_norm": 0.18847516841511264, |
|
"learning_rate": 2.1085949060360654e-06, |
|
"loss": 1.1135, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.8147174770039421, |
|
"grad_norm": 0.19132781030558177, |
|
"learning_rate": 2.015488020140737e-06, |
|
"loss": 1.1045, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.8190976784932107, |
|
"grad_norm": 0.18904026009165054, |
|
"learning_rate": 1.924252613108073e-06, |
|
"loss": 1.1318, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.8234778799824792, |
|
"grad_norm": 0.19251546970491693, |
|
"learning_rate": 1.8349100694805711e-06, |
|
"loss": 1.1243, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.8278580814717477, |
|
"grad_norm": 0.19006518497007893, |
|
"learning_rate": 1.7474813301350668e-06, |
|
"loss": 1.123, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.8322382829610162, |
|
"grad_norm": 0.19325212813265874, |
|
"learning_rate": 1.661986887374415e-06, |
|
"loss": 1.1129, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.8366184844502847, |
|
"grad_norm": 0.1889419224748978, |
|
"learning_rate": 1.578446780124344e-06, |
|
"loss": 1.1103, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 0.8409986859395532, |
|
"grad_norm": 0.19033279856504365, |
|
"learning_rate": 1.49688058923654e-06, |
|
"loss": 1.1098, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.8453788874288217, |
|
"grad_norm": 0.18633846336335902, |
|
"learning_rate": 1.4173074328991376e-06, |
|
"loss": 1.1287, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 0.8497590889180903, |
|
"grad_norm": 0.19006030021409406, |
|
"learning_rate": 1.339745962155613e-06, |
|
"loss": 1.1329, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.8541392904073587, |
|
"grad_norm": 0.18903779583859964, |
|
"learning_rate": 1.2642143565332154e-06, |
|
"loss": 1.1316, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.8585194918966272, |
|
"grad_norm": 0.19272564813632415, |
|
"learning_rate": 1.1907303197818665e-06, |
|
"loss": 1.1303, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.8628996933858958, |
|
"grad_norm": 0.18878052764179085, |
|
"learning_rate": 1.1193110757246251e-06, |
|
"loss": 1.1205, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 0.8672798948751642, |
|
"grad_norm": 0.19218035241765347, |
|
"learning_rate": 1.0499733642206034e-06, |
|
"loss": 1.124, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.8716600963644328, |
|
"grad_norm": 0.18852107182367117, |
|
"learning_rate": 9.827334372413444e-07, |
|
"loss": 1.1301, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 0.8760402978537013, |
|
"grad_norm": 0.19264462588904008, |
|
"learning_rate": 9.176070550615379e-07, |
|
"loss": 1.1254, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.8804204993429697, |
|
"grad_norm": 0.18440224651228512, |
|
"learning_rate": 8.546094825649909e-07, |
|
"loss": 1.1189, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 0.8848007008322383, |
|
"grad_norm": 0.18637643466062778, |
|
"learning_rate": 7.937554856667196e-07, |
|
"loss": 1.1386, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.8891809023215068, |
|
"grad_norm": 0.19217636415097236, |
|
"learning_rate": 7.350593278519824e-07, |
|
"loss": 1.1512, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 0.8935611038107752, |
|
"grad_norm": 0.19868031904477793, |
|
"learning_rate": 6.785347668330777e-07, |
|
"loss": 1.1385, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.8979413053000438, |
|
"grad_norm": 0.18822804076907482, |
|
"learning_rate": 6.241950513246931e-07, |
|
"loss": 1.1236, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.9023215067893123, |
|
"grad_norm": 0.18768750481368643, |
|
"learning_rate": 5.720529179385659e-07, |
|
"loss": 1.1096, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.9067017082785808, |
|
"grad_norm": 0.18565141772239874, |
|
"learning_rate": 5.221205881981594e-07, |
|
"loss": 1.1197, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 0.9110819097678493, |
|
"grad_norm": 0.2034853308480849, |
|
"learning_rate": 4.7440976567407096e-07, |
|
"loss": 1.1371, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.9154621112571178, |
|
"grad_norm": 0.1998255782368389, |
|
"learning_rate": 4.2893163324085886e-07, |
|
"loss": 1.1115, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 0.9198423127463863, |
|
"grad_norm": 0.19439905516574385, |
|
"learning_rate": 3.856968504558989e-07, |
|
"loss": 1.1156, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.9242225142356548, |
|
"grad_norm": 0.18518716914861474, |
|
"learning_rate": 3.4471555106090573e-07, |
|
"loss": 1.1316, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 0.9286027157249234, |
|
"grad_norm": 0.18819573805967685, |
|
"learning_rate": 3.059973406066963e-07, |
|
"loss": 1.1215, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.9329829172141918, |
|
"grad_norm": 0.18672133869392946, |
|
"learning_rate": 2.6955129420176193e-07, |
|
"loss": 1.1264, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 0.9373631187034603, |
|
"grad_norm": 0.18848502790832367, |
|
"learning_rate": 2.3538595438516442e-07, |
|
"loss": 1.1215, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.9417433201927289, |
|
"grad_norm": 0.18909475354670285, |
|
"learning_rate": 2.035093291242607e-07, |
|
"loss": 1.1324, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.9461235216819974, |
|
"grad_norm": 0.18392034150187955, |
|
"learning_rate": 1.7392888993773005e-07, |
|
"loss": 1.0989, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.9505037231712659, |
|
"grad_norm": 0.18422655351430084, |
|
"learning_rate": 1.466515701443294e-07, |
|
"loss": 1.149, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 0.9548839246605344, |
|
"grad_norm": 0.18472721089921113, |
|
"learning_rate": 1.2168376323780652e-07, |
|
"loss": 1.1338, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.9592641261498029, |
|
"grad_norm": 0.18618589859950402, |
|
"learning_rate": 9.90313213883376e-08, |
|
"loss": 1.1488, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 0.9636443276390714, |
|
"grad_norm": 0.1841956829600993, |
|
"learning_rate": 7.86995540708424e-08, |
|
"loss": 1.1058, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.9680245291283399, |
|
"grad_norm": 0.18737860582967023, |
|
"learning_rate": 6.069322682050516e-08, |
|
"loss": 1.1347, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 0.9724047306176085, |
|
"grad_norm": 0.18327423894477116, |
|
"learning_rate": 4.501656011579037e-08, |
|
"loss": 1.1223, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.9767849321068769, |
|
"grad_norm": 0.18664413828439025, |
|
"learning_rate": 3.167322838920406e-08, |
|
"loss": 1.1036, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 0.9811651335961454, |
|
"grad_norm": 0.18891847587451832, |
|
"learning_rate": 2.066635916605386e-08, |
|
"loss": 1.1229, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.985545335085414, |
|
"grad_norm": 0.19664758693417972, |
|
"learning_rate": 1.1998532331389812e-08, |
|
"loss": 1.156, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.9899255365746824, |
|
"grad_norm": 0.18716549597123566, |
|
"learning_rate": 5.671779525311394e-09, |
|
"loss": 1.1146, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.994305738063951, |
|
"grad_norm": 0.1847881495731737, |
|
"learning_rate": 1.6875836667729073e-09, |
|
"loss": 1.1207, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 0.9986859395532195, |
|
"grad_norm": 0.18567353384044155, |
|
"learning_rate": 4.687860599927874e-11, |
|
"loss": 1.1244, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.9995619798510732, |
|
"step": 1141, |
|
"total_flos": 215836199485440.0, |
|
"train_loss": 1.3199829850460123, |
|
"train_runtime": 8951.0623, |
|
"train_samples_per_second": 16.317, |
|
"train_steps_per_second": 0.127 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 1141, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 215836199485440.0, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|