|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 500, |
|
"global_step": 21487, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0046539768231954205, |
|
"grad_norm": 0.2191164344549179, |
|
"learning_rate": 9.31098696461825e-07, |
|
"loss": 1.8284, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.009307953646390841, |
|
"grad_norm": 0.23232442140579224, |
|
"learning_rate": 1.86219739292365e-06, |
|
"loss": 1.8324, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.013961930469586261, |
|
"grad_norm": 0.27771520614624023, |
|
"learning_rate": 2.793296089385475e-06, |
|
"loss": 1.8126, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.018615907292781682, |
|
"grad_norm": 0.46955519914627075, |
|
"learning_rate": 3.7243947858473e-06, |
|
"loss": 1.7894, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.023269884115977104, |
|
"grad_norm": 0.4999992847442627, |
|
"learning_rate": 4.655493482309125e-06, |
|
"loss": 1.7438, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.027923860939172523, |
|
"grad_norm": 0.6398751139640808, |
|
"learning_rate": 5.58659217877095e-06, |
|
"loss": 1.6969, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.032577837762367945, |
|
"grad_norm": 0.7858927249908447, |
|
"learning_rate": 6.517690875232775e-06, |
|
"loss": 1.7084, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.037231814585563364, |
|
"grad_norm": 0.7117326259613037, |
|
"learning_rate": 7.4487895716946e-06, |
|
"loss": 1.709, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.04188579140875878, |
|
"grad_norm": 0.7050628066062927, |
|
"learning_rate": 8.379888268156426e-06, |
|
"loss": 1.6655, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.04653976823195421, |
|
"grad_norm": 0.9369891285896301, |
|
"learning_rate": 9.31098696461825e-06, |
|
"loss": 1.6765, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.05119374505514963, |
|
"grad_norm": 0.7342795729637146, |
|
"learning_rate": 1.0242085661080074e-05, |
|
"loss": 1.6398, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.055847721878345045, |
|
"grad_norm": 1.5446929931640625, |
|
"learning_rate": 1.11731843575419e-05, |
|
"loss": 1.6714, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.060501698701540464, |
|
"grad_norm": 0.8926479816436768, |
|
"learning_rate": 1.2104283054003724e-05, |
|
"loss": 1.6355, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.06515567552473589, |
|
"grad_norm": 0.9157938361167908, |
|
"learning_rate": 1.303538175046555e-05, |
|
"loss": 1.5984, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.06980965234793131, |
|
"grad_norm": 0.9163572192192078, |
|
"learning_rate": 1.3966480446927374e-05, |
|
"loss": 1.6372, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.07446362917112673, |
|
"grad_norm": 1.3793407678604126, |
|
"learning_rate": 1.48975791433892e-05, |
|
"loss": 1.6025, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.07911760599432215, |
|
"grad_norm": 1.2203880548477173, |
|
"learning_rate": 1.5828677839851026e-05, |
|
"loss": 1.6359, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.08377158281751756, |
|
"grad_norm": 1.18509840965271, |
|
"learning_rate": 1.6759776536312852e-05, |
|
"loss": 1.6273, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.08842555964071298, |
|
"grad_norm": 1.1019757986068726, |
|
"learning_rate": 1.7690875232774675e-05, |
|
"loss": 1.6286, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.09307953646390842, |
|
"grad_norm": 0.6310858726501465, |
|
"learning_rate": 1.86219739292365e-05, |
|
"loss": 1.6132, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.09773351328710383, |
|
"grad_norm": 1.2102677822113037, |
|
"learning_rate": 1.9553072625698326e-05, |
|
"loss": 1.6125, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.10238749011029925, |
|
"grad_norm": 1.16068696975708, |
|
"learning_rate": 1.999964321562415e-05, |
|
"loss": 1.618, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.10704146693349467, |
|
"grad_norm": 1.1368638277053833, |
|
"learning_rate": 1.999695163593931e-05, |
|
"loss": 1.5991, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.11169544375669009, |
|
"grad_norm": 0.9747415781021118, |
|
"learning_rate": 1.9991621965335746e-05, |
|
"loss": 1.5878, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.11634942057988551, |
|
"grad_norm": 0.9197840690612793, |
|
"learning_rate": 1.998365561025775e-05, |
|
"loss": 1.6189, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.12100339740308093, |
|
"grad_norm": 0.9155735373497009, |
|
"learning_rate": 1.997305467294307e-05, |
|
"loss": 1.587, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.12565737422627635, |
|
"grad_norm": 0.7722734212875366, |
|
"learning_rate": 1.9959821950868117e-05, |
|
"loss": 1.5681, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.13031135104947178, |
|
"grad_norm": 1.044062852859497, |
|
"learning_rate": 1.994396093600977e-05, |
|
"loss": 1.5845, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.13496532787266718, |
|
"grad_norm": 0.7750725150108337, |
|
"learning_rate": 1.9925475813923858e-05, |
|
"loss": 1.5936, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.13961930469586262, |
|
"grad_norm": 0.9063745141029358, |
|
"learning_rate": 1.990437146264066e-05, |
|
"loss": 1.5788, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.14427328151905802, |
|
"grad_norm": 0.655881404876709, |
|
"learning_rate": 1.988065345137762e-05, |
|
"loss": 1.5927, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.14892725834225345, |
|
"grad_norm": 1.1087632179260254, |
|
"learning_rate": 1.98543280390697e-05, |
|
"loss": 1.5607, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.1535812351654489, |
|
"grad_norm": 0.7348114848136902, |
|
"learning_rate": 1.982540217271772e-05, |
|
"loss": 1.5344, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.1582352119886443, |
|
"grad_norm": 0.8793938755989075, |
|
"learning_rate": 1.979388348555509e-05, |
|
"loss": 1.5576, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.16288918881183972, |
|
"grad_norm": 0.9496509432792664, |
|
"learning_rate": 1.975978029503352e-05, |
|
"loss": 1.5439, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.16754316563503513, |
|
"grad_norm": 0.9296020269393921, |
|
"learning_rate": 1.9723101600628095e-05, |
|
"loss": 1.6044, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.17219714245823056, |
|
"grad_norm": 0.7520443797111511, |
|
"learning_rate": 1.9683857081462423e-05, |
|
"loss": 1.5319, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.17685111928142597, |
|
"grad_norm": 0.9139896631240845, |
|
"learning_rate": 1.9642057093754426e-05, |
|
"loss": 1.5457, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.1815050961046214, |
|
"grad_norm": 0.781461775302887, |
|
"learning_rate": 1.9597712668083413e-05, |
|
"loss": 1.5604, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.18615907292781683, |
|
"grad_norm": 0.9205677509307861, |
|
"learning_rate": 1.9550835506479266e-05, |
|
"loss": 1.5255, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.19081304975101224, |
|
"grad_norm": 0.7341404557228088, |
|
"learning_rate": 1.9501437979334358e-05, |
|
"loss": 1.5876, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.19546702657420767, |
|
"grad_norm": 0.8709468245506287, |
|
"learning_rate": 1.9449533122139157e-05, |
|
"loss": 1.5469, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.20012100339740307, |
|
"grad_norm": 0.9768949151039124, |
|
"learning_rate": 1.939513463204229e-05, |
|
"loss": 1.5417, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.2047749802205985, |
|
"grad_norm": 0.7858554720878601, |
|
"learning_rate": 1.9338256864236024e-05, |
|
"loss": 1.5672, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.2094289570437939, |
|
"grad_norm": 0.7643204927444458, |
|
"learning_rate": 1.927891482816805e-05, |
|
"loss": 1.5032, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.21408293386698934, |
|
"grad_norm": 1.2041771411895752, |
|
"learning_rate": 1.9217124183580674e-05, |
|
"loss": 1.51, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.21873691069018475, |
|
"grad_norm": 0.9453679323196411, |
|
"learning_rate": 1.9152901236378364e-05, |
|
"loss": 1.5662, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.22339088751338018, |
|
"grad_norm": 0.6478219628334045, |
|
"learning_rate": 1.908626293432479e-05, |
|
"loss": 1.5932, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.22804486433657561, |
|
"grad_norm": 0.838474690914154, |
|
"learning_rate": 1.901722686257051e-05, |
|
"loss": 1.567, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.23269884115977102, |
|
"grad_norm": 1.1010560989379883, |
|
"learning_rate": 1.894581123901238e-05, |
|
"loss": 1.4849, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.23735281798296645, |
|
"grad_norm": 0.8271942734718323, |
|
"learning_rate": 1.88720349094861e-05, |
|
"loss": 1.56, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.24200679480616186, |
|
"grad_norm": 0.9592469334602356, |
|
"learning_rate": 1.8795917342792967e-05, |
|
"loss": 1.6055, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.2466607716293573, |
|
"grad_norm": 0.7423321008682251, |
|
"learning_rate": 1.8717478625562256e-05, |
|
"loss": 1.5268, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.2513147484525527, |
|
"grad_norm": 1.0170488357543945, |
|
"learning_rate": 1.8636739456950593e-05, |
|
"loss": 1.5765, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.2559687252757481, |
|
"grad_norm": 0.7664517760276794, |
|
"learning_rate": 1.8553721143179626e-05, |
|
"loss": 1.5625, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.26062270209894356, |
|
"grad_norm": 0.712568461894989, |
|
"learning_rate": 1.8468445591913587e-05, |
|
"loss": 1.5853, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.26527667892213896, |
|
"grad_norm": 0.9137941002845764, |
|
"learning_rate": 1.8380935306478036e-05, |
|
"loss": 1.5628, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.26993065574533437, |
|
"grad_norm": 0.9478745460510254, |
|
"learning_rate": 1.829121337992151e-05, |
|
"loss": 1.4895, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.27458463256852983, |
|
"grad_norm": 0.5866535305976868, |
|
"learning_rate": 1.8199303488921493e-05, |
|
"loss": 1.5484, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.27923860939172523, |
|
"grad_norm": 0.9393543004989624, |
|
"learning_rate": 1.810522988753638e-05, |
|
"loss": 1.5307, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.28389258621492064, |
|
"grad_norm": 0.6128756403923035, |
|
"learning_rate": 1.8009017400805094e-05, |
|
"loss": 1.5821, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.28854656303811604, |
|
"grad_norm": 0.8191483020782471, |
|
"learning_rate": 1.7910691418196022e-05, |
|
"loss": 1.5206, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.2932005398613115, |
|
"grad_norm": 0.8590126633644104, |
|
"learning_rate": 1.7810277886906998e-05, |
|
"loss": 1.5488, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.2978545166845069, |
|
"grad_norm": 0.6708762645721436, |
|
"learning_rate": 1.7707803305018097e-05, |
|
"loss": 1.5383, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.3025084935077023, |
|
"grad_norm": 0.6897620558738708, |
|
"learning_rate": 1.760329471449909e-05, |
|
"loss": 1.5497, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.3071624703308978, |
|
"grad_norm": 0.7875939607620239, |
|
"learning_rate": 1.7496779694073333e-05, |
|
"loss": 1.5251, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.3118164471540932, |
|
"grad_norm": 0.9739288091659546, |
|
"learning_rate": 1.738828635194003e-05, |
|
"loss": 1.5235, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.3164704239772886, |
|
"grad_norm": 0.9243332743644714, |
|
"learning_rate": 1.7277843318356798e-05, |
|
"loss": 1.5256, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.321124400800484, |
|
"grad_norm": 1.0979468822479248, |
|
"learning_rate": 1.7165479738084405e-05, |
|
"loss": 1.5841, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.32577837762367945, |
|
"grad_norm": 0.931398868560791, |
|
"learning_rate": 1.7051225262695812e-05, |
|
"loss": 1.5408, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.33043235444687485, |
|
"grad_norm": 0.7612720131874084, |
|
"learning_rate": 1.693511004275139e-05, |
|
"loss": 1.4606, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.33508633127007026, |
|
"grad_norm": 1.0624606609344482, |
|
"learning_rate": 1.6817164719842548e-05, |
|
"loss": 1.5322, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.3397403080932657, |
|
"grad_norm": 0.6526899933815002, |
|
"learning_rate": 1.66974204185057e-05, |
|
"loss": 1.5801, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.3443942849164611, |
|
"grad_norm": 0.7007707953453064, |
|
"learning_rate": 1.6575908738008847e-05, |
|
"loss": 1.4867, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.34904826173965653, |
|
"grad_norm": 0.8433778882026672, |
|
"learning_rate": 1.645266174401285e-05, |
|
"loss": 1.5193, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.35370223856285193, |
|
"grad_norm": 1.0169028043746948, |
|
"learning_rate": 1.632771196010965e-05, |
|
"loss": 1.53, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.3583562153860474, |
|
"grad_norm": 0.9545369744300842, |
|
"learning_rate": 1.6201092359239624e-05, |
|
"loss": 1.4862, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.3630101922092428, |
|
"grad_norm": 1.2505319118499756, |
|
"learning_rate": 1.607283635499036e-05, |
|
"loss": 1.5349, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.3676641690324382, |
|
"grad_norm": 0.8437191843986511, |
|
"learning_rate": 1.5942977792779193e-05, |
|
"loss": 1.5245, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.37231814585563366, |
|
"grad_norm": 0.8392757773399353, |
|
"learning_rate": 1.581155094092171e-05, |
|
"loss": 1.519, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.37697212267882907, |
|
"grad_norm": 0.7851161956787109, |
|
"learning_rate": 1.5678590481588743e-05, |
|
"loss": 1.5301, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.3816260995020245, |
|
"grad_norm": 0.8554267883300781, |
|
"learning_rate": 1.5544131501654078e-05, |
|
"loss": 1.5143, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.3862800763252199, |
|
"grad_norm": 0.9319170117378235, |
|
"learning_rate": 1.5408209483435402e-05, |
|
"loss": 1.5079, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.39093405314841534, |
|
"grad_norm": 0.9828823804855347, |
|
"learning_rate": 1.5270860295330912e-05, |
|
"loss": 1.5248, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.39558802997161074, |
|
"grad_norm": 0.8951559662818909, |
|
"learning_rate": 1.5132120182353984e-05, |
|
"loss": 1.4728, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.40024200679480615, |
|
"grad_norm": 0.8046073317527771, |
|
"learning_rate": 1.499202575656852e-05, |
|
"loss": 1.5096, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.4048959836180016, |
|
"grad_norm": 0.7216867208480835, |
|
"learning_rate": 1.4850613987427398e-05, |
|
"loss": 1.538, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.409549960441197, |
|
"grad_norm": 0.5751676559448242, |
|
"learning_rate": 1.470792219201662e-05, |
|
"loss": 1.508, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.4142039372643924, |
|
"grad_norm": 0.7514901161193848, |
|
"learning_rate": 1.4563988025207726e-05, |
|
"loss": 1.5365, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 0.4188579140875878, |
|
"grad_norm": 0.8872545957565308, |
|
"learning_rate": 1.4418849469721073e-05, |
|
"loss": 1.4705, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.4235118909107833, |
|
"grad_norm": 0.772324800491333, |
|
"learning_rate": 1.4272544826102576e-05, |
|
"loss": 1.521, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 0.4281658677339787, |
|
"grad_norm": 0.8108826279640198, |
|
"learning_rate": 1.4125112702616607e-05, |
|
"loss": 1.4958, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 0.4328198445571741, |
|
"grad_norm": 0.8969531655311584, |
|
"learning_rate": 1.3976592005057684e-05, |
|
"loss": 1.4596, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 0.4374738213803695, |
|
"grad_norm": 0.93816739320755, |
|
"learning_rate": 1.3827021926483622e-05, |
|
"loss": 1.5404, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 0.44212779820356496, |
|
"grad_norm": 0.5850194096565247, |
|
"learning_rate": 1.3676441936872904e-05, |
|
"loss": 1.5342, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.44678177502676036, |
|
"grad_norm": 0.903218686580658, |
|
"learning_rate": 1.352489177270898e-05, |
|
"loss": 1.5315, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 0.45143575184995577, |
|
"grad_norm": 0.9102717041969299, |
|
"learning_rate": 1.337241142649419e-05, |
|
"loss": 1.5105, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 0.45608972867315123, |
|
"grad_norm": 0.5675022006034851, |
|
"learning_rate": 1.3219041136196202e-05, |
|
"loss": 1.465, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 0.46074370549634663, |
|
"grad_norm": 1.007855772972107, |
|
"learning_rate": 1.3064821374629606e-05, |
|
"loss": 1.5131, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 0.46539768231954204, |
|
"grad_norm": 0.8553173542022705, |
|
"learning_rate": 1.2909792838775571e-05, |
|
"loss": 1.5054, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.47005165914273744, |
|
"grad_norm": 0.9074007868766785, |
|
"learning_rate": 1.2755558069133832e-05, |
|
"loss": 1.5582, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 0.4747056359659329, |
|
"grad_norm": 0.7873587608337402, |
|
"learning_rate": 1.2599041981920994e-05, |
|
"loss": 1.5109, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 0.4793596127891283, |
|
"grad_norm": 0.9822585582733154, |
|
"learning_rate": 1.2441840034728076e-05, |
|
"loss": 1.5067, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 0.4840135896123237, |
|
"grad_norm": 1.0133826732635498, |
|
"learning_rate": 1.2283993711503824e-05, |
|
"loss": 1.4886, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 0.4886675664355192, |
|
"grad_norm": 1.1799577474594116, |
|
"learning_rate": 1.2125544666241078e-05, |
|
"loss": 1.4655, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.4933215432587146, |
|
"grad_norm": 0.927507221698761, |
|
"learning_rate": 1.1966534711984717e-05, |
|
"loss": 1.4748, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 0.49797552008191, |
|
"grad_norm": 0.8527470827102661, |
|
"learning_rate": 1.1808603529490986e-05, |
|
"loss": 1.4607, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 0.5026294969051054, |
|
"grad_norm": 0.8991233706474304, |
|
"learning_rate": 1.1648602337072668e-05, |
|
"loss": 1.5432, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 0.5072834737283008, |
|
"grad_norm": 0.7587746381759644, |
|
"learning_rate": 1.1488166095750207e-05, |
|
"loss": 1.5053, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 0.5119374505514962, |
|
"grad_norm": 0.6698244214057922, |
|
"learning_rate": 1.132733714296872e-05, |
|
"loss": 1.4281, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.5165914273746917, |
|
"grad_norm": 0.848193883895874, |
|
"learning_rate": 1.1166157919805772e-05, |
|
"loss": 1.5043, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 0.5212454041978871, |
|
"grad_norm": 1.0735975503921509, |
|
"learning_rate": 1.1004670959771624e-05, |
|
"loss": 1.5072, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 0.5258993810210825, |
|
"grad_norm": 1.1171404123306274, |
|
"learning_rate": 1.0842918877585058e-05, |
|
"loss": 1.4864, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 0.5305533578442779, |
|
"grad_norm": 0.8309624195098877, |
|
"learning_rate": 1.0680944357927836e-05, |
|
"loss": 1.5373, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 0.5352073346674734, |
|
"grad_norm": 0.8913981318473816, |
|
"learning_rate": 1.0518790144180635e-05, |
|
"loss": 1.4432, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.5398613114906687, |
|
"grad_norm": 0.8174418807029724, |
|
"learning_rate": 1.0356499027143519e-05, |
|
"loss": 1.4852, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 0.5445152883138642, |
|
"grad_norm": 0.9098989367485046, |
|
"learning_rate": 1.0194113833743897e-05, |
|
"loss": 1.5405, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 0.5491692651370597, |
|
"grad_norm": 0.983168363571167, |
|
"learning_rate": 1.0031677415734946e-05, |
|
"loss": 1.5014, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 0.553823241960255, |
|
"grad_norm": 0.793428361415863, |
|
"learning_rate": 9.869232638387492e-06, |
|
"loss": 1.5047, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 0.5584772187834505, |
|
"grad_norm": 0.9047326445579529, |
|
"learning_rate": 9.706822369178289e-06, |
|
"loss": 1.4991, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.5631311956066459, |
|
"grad_norm": 0.825706422328949, |
|
"learning_rate": 9.54448946647775e-06, |
|
"loss": 1.5175, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 0.5677851724298413, |
|
"grad_norm": 0.9917203187942505, |
|
"learning_rate": 9.382276768240104e-06, |
|
"loss": 1.4916, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 0.5724391492530367, |
|
"grad_norm": 0.8516975045204163, |
|
"learning_rate": 9.220227080698847e-06, |
|
"loss": 1.5003, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 0.5770931260762321, |
|
"grad_norm": 0.7885426878929138, |
|
"learning_rate": 9.058383167070676e-06, |
|
"loss": 1.5427, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 0.5817471028994275, |
|
"grad_norm": 0.9626107215881348, |
|
"learning_rate": 8.89678773627071e-06, |
|
"loss": 1.4672, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.586401079722623, |
|
"grad_norm": 1.023768663406372, |
|
"learning_rate": 8.735483431642033e-06, |
|
"loss": 1.5236, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 0.5910550565458184, |
|
"grad_norm": 0.9998337626457214, |
|
"learning_rate": 8.574512819702594e-06, |
|
"loss": 1.4932, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 0.5957090333690138, |
|
"grad_norm": 0.9876322746276855, |
|
"learning_rate": 8.413918378912325e-06, |
|
"loss": 1.4849, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 0.6003630101922093, |
|
"grad_norm": 0.9000487923622131, |
|
"learning_rate": 8.253742488463531e-06, |
|
"loss": 1.4532, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 0.6050169870154046, |
|
"grad_norm": 0.9810521006584167, |
|
"learning_rate": 8.094027417097468e-06, |
|
"loss": 1.4818, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.6096709638386001, |
|
"grad_norm": 0.8786018490791321, |
|
"learning_rate": 7.936404805248885e-06, |
|
"loss": 1.524, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 0.6143249406617955, |
|
"grad_norm": 0.8514164686203003, |
|
"learning_rate": 7.777732023413952e-06, |
|
"loss": 1.4763, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 0.6189789174849909, |
|
"grad_norm": 0.603767991065979, |
|
"learning_rate": 7.619645674842026e-06, |
|
"loss": 1.4977, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 0.6236328943081864, |
|
"grad_norm": 0.8721168637275696, |
|
"learning_rate": 7.462187476866129e-06, |
|
"loss": 1.4339, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 0.6282868711313818, |
|
"grad_norm": 0.8393872380256653, |
|
"learning_rate": 7.305398981056899e-06, |
|
"loss": 1.4888, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.6329408479545772, |
|
"grad_norm": 1.0900713205337524, |
|
"learning_rate": 7.149321562257598e-06, |
|
"loss": 1.4612, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 0.6375948247777726, |
|
"grad_norm": 0.857639491558075, |
|
"learning_rate": 6.9939964076657e-06, |
|
"loss": 1.4866, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 0.642248801600968, |
|
"grad_norm": 0.7077500820159912, |
|
"learning_rate": 6.839464505964017e-06, |
|
"loss": 1.4587, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 0.6469027784241634, |
|
"grad_norm": 0.9254826307296753, |
|
"learning_rate": 6.6857666365042354e-06, |
|
"loss": 1.463, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 0.6515567552473589, |
|
"grad_norm": 0.8199195265769958, |
|
"learning_rate": 6.5329433585456625e-06, |
|
"loss": 1.5347, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.6562107320705542, |
|
"grad_norm": 1.0524635314941406, |
|
"learning_rate": 6.381035000552058e-06, |
|
"loss": 1.5253, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 0.6608647088937497, |
|
"grad_norm": 0.9598236083984375, |
|
"learning_rate": 6.230081649549371e-06, |
|
"loss": 1.5064, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 0.6655186857169452, |
|
"grad_norm": 0.8717284202575684, |
|
"learning_rate": 6.080123140547213e-06, |
|
"loss": 1.5284, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 0.6701726625401405, |
|
"grad_norm": 0.7019008994102478, |
|
"learning_rate": 5.931199046026791e-06, |
|
"loss": 1.5057, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 0.674826639363336, |
|
"grad_norm": 0.7632008790969849, |
|
"learning_rate": 5.78334866549816e-06, |
|
"loss": 1.4382, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.6794806161865314, |
|
"grad_norm": 0.9148026704788208, |
|
"learning_rate": 5.63661101512949e-06, |
|
"loss": 1.4656, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 0.6841345930097268, |
|
"grad_norm": 1.0867414474487305, |
|
"learning_rate": 5.491024817451108e-06, |
|
"loss": 1.4611, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 0.6887885698329222, |
|
"grad_norm": 0.9941554665565491, |
|
"learning_rate": 5.346628491137016e-06, |
|
"loss": 1.471, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 0.6934425466561176, |
|
"grad_norm": 0.756461501121521, |
|
"learning_rate": 5.203460140866625e-06, |
|
"loss": 1.4691, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 0.6980965234793131, |
|
"grad_norm": 0.9088025093078613, |
|
"learning_rate": 5.061557547269322e-06, |
|
"loss": 1.4712, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.7027505003025085, |
|
"grad_norm": 0.850143313407898, |
|
"learning_rate": 4.920958156954544e-06, |
|
"loss": 1.4434, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 0.7074044771257039, |
|
"grad_norm": 1.072135090827942, |
|
"learning_rate": 4.781699072630018e-06, |
|
"loss": 1.5095, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 0.7120584539488993, |
|
"grad_norm": 0.8175047039985657, |
|
"learning_rate": 4.643817043310759e-06, |
|
"loss": 1.455, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 0.7167124307720948, |
|
"grad_norm": 0.8658547401428223, |
|
"learning_rate": 4.5073484546213705e-06, |
|
"loss": 1.4545, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 0.7213664075952901, |
|
"grad_norm": 0.9122436046600342, |
|
"learning_rate": 4.372329319194262e-06, |
|
"loss": 1.5157, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.7260203844184856, |
|
"grad_norm": 1.0552470684051514, |
|
"learning_rate": 4.238795267166298e-06, |
|
"loss": 1.4664, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 0.7306743612416811, |
|
"grad_norm": 0.8493475914001465, |
|
"learning_rate": 4.106781536776383e-06, |
|
"loss": 1.5032, |
|
"step": 15700 |
|
}, |
|
{ |
|
"epoch": 0.7353283380648764, |
|
"grad_norm": 0.9446743130683899, |
|
"learning_rate": 3.976322965066436e-06, |
|
"loss": 1.4747, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 0.7399823148880719, |
|
"grad_norm": 0.7051513195037842, |
|
"learning_rate": 3.847453978688287e-06, |
|
"loss": 1.4649, |
|
"step": 15900 |
|
}, |
|
{ |
|
"epoch": 0.7446362917112673, |
|
"grad_norm": 0.941286563873291, |
|
"learning_rate": 3.720208584818833e-06, |
|
"loss": 1.4755, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.7492902685344627, |
|
"grad_norm": 0.9868441820144653, |
|
"learning_rate": 3.5946203621858867e-06, |
|
"loss": 1.5239, |
|
"step": 16100 |
|
}, |
|
{ |
|
"epoch": 0.7539442453576581, |
|
"grad_norm": 1.0158665180206299, |
|
"learning_rate": 3.4707224522071294e-06, |
|
"loss": 1.4625, |
|
"step": 16200 |
|
}, |
|
{ |
|
"epoch": 0.7585982221808535, |
|
"grad_norm": 1.0703004598617554, |
|
"learning_rate": 3.3485475502444055e-06, |
|
"loss": 1.4627, |
|
"step": 16300 |
|
}, |
|
{ |
|
"epoch": 0.763252199004049, |
|
"grad_norm": 0.9671617746353149, |
|
"learning_rate": 3.2281278969757913e-06, |
|
"loss": 1.4777, |
|
"step": 16400 |
|
}, |
|
{ |
|
"epoch": 0.7679061758272444, |
|
"grad_norm": 1.0113691091537476, |
|
"learning_rate": 3.109495269887587e-06, |
|
"loss": 1.4954, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 0.7725601526504398, |
|
"grad_norm": 1.0288777351379395, |
|
"learning_rate": 2.9926809748885964e-06, |
|
"loss": 1.5105, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 0.7772141294736352, |
|
"grad_norm": 0.851613461971283, |
|
"learning_rate": 2.877715838048811e-06, |
|
"loss": 1.4307, |
|
"step": 16700 |
|
}, |
|
{ |
|
"epoch": 0.7818681062968307, |
|
"grad_norm": 0.7523232102394104, |
|
"learning_rate": 2.764630197464729e-06, |
|
"loss": 1.4429, |
|
"step": 16800 |
|
}, |
|
{ |
|
"epoch": 0.786522083120026, |
|
"grad_norm": 1.0177912712097168, |
|
"learning_rate": 2.653453895253466e-06, |
|
"loss": 1.4585, |
|
"step": 16900 |
|
}, |
|
{ |
|
"epoch": 0.7911760599432215, |
|
"grad_norm": 0.7985556721687317, |
|
"learning_rate": 2.5442162696777095e-06, |
|
"loss": 1.4912, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.795830036766417, |
|
"grad_norm": 1.0856115818023682, |
|
"learning_rate": 2.4369461474036825e-06, |
|
"loss": 1.4824, |
|
"step": 17100 |
|
}, |
|
{ |
|
"epoch": 0.8004840135896123, |
|
"grad_norm": 1.0469813346862793, |
|
"learning_rate": 2.332714608087351e-06, |
|
"loss": 1.4442, |
|
"step": 17200 |
|
}, |
|
{ |
|
"epoch": 0.8051379904128078, |
|
"grad_norm": 0.7078351378440857, |
|
"learning_rate": 2.2294435164482266e-06, |
|
"loss": 1.5291, |
|
"step": 17300 |
|
}, |
|
{ |
|
"epoch": 0.8097919672360032, |
|
"grad_norm": 0.9295824766159058, |
|
"learning_rate": 2.1282229933463116e-06, |
|
"loss": 1.4083, |
|
"step": 17400 |
|
}, |
|
{ |
|
"epoch": 0.8144459440591986, |
|
"grad_norm": 0.7632860541343689, |
|
"learning_rate": 2.0290797498183858e-06, |
|
"loss": 1.4367, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 0.819099920882394, |
|
"grad_norm": 0.9528563022613525, |
|
"learning_rate": 1.9320399487288832e-06, |
|
"loss": 1.445, |
|
"step": 17600 |
|
}, |
|
{ |
|
"epoch": 0.8237538977055894, |
|
"grad_norm": 0.8745253682136536, |
|
"learning_rate": 1.8371291978657857e-06, |
|
"loss": 1.4694, |
|
"step": 17700 |
|
}, |
|
{ |
|
"epoch": 0.8284078745287848, |
|
"grad_norm": 0.9276142120361328, |
|
"learning_rate": 1.7443725431830127e-06, |
|
"loss": 1.4682, |
|
"step": 17800 |
|
}, |
|
{ |
|
"epoch": 0.8330618513519803, |
|
"grad_norm": 0.8589171171188354, |
|
"learning_rate": 1.6537944621910307e-06, |
|
"loss": 1.4999, |
|
"step": 17900 |
|
}, |
|
{ |
|
"epoch": 0.8377158281751756, |
|
"grad_norm": 0.8758667707443237, |
|
"learning_rate": 1.5654188574975215e-06, |
|
"loss": 1.4621, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.8423698049983711, |
|
"grad_norm": 0.9397442936897278, |
|
"learning_rate": 1.4792690504997087e-06, |
|
"loss": 1.4598, |
|
"step": 18100 |
|
}, |
|
{ |
|
"epoch": 0.8470237818215666, |
|
"grad_norm": 1.0151734352111816, |
|
"learning_rate": 1.3953677752301032e-06, |
|
"loss": 1.4698, |
|
"step": 18200 |
|
}, |
|
{ |
|
"epoch": 0.8516777586447619, |
|
"grad_norm": 0.9862604737281799, |
|
"learning_rate": 1.3137371723572257e-06, |
|
"loss": 1.4478, |
|
"step": 18300 |
|
}, |
|
{ |
|
"epoch": 0.8563317354679574, |
|
"grad_norm": 1.16762113571167, |
|
"learning_rate": 1.2343987833429062e-06, |
|
"loss": 1.4806, |
|
"step": 18400 |
|
}, |
|
{ |
|
"epoch": 0.8609857122911528, |
|
"grad_norm": 1.06341552734375, |
|
"learning_rate": 1.1573735447577327e-06, |
|
"loss": 1.4555, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 0.8656396891143482, |
|
"grad_norm": 1.0032464265823364, |
|
"learning_rate": 1.082681782756102e-06, |
|
"loss": 1.4471, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 0.8702936659375436, |
|
"grad_norm": 0.9144641160964966, |
|
"learning_rate": 1.0103432077123532e-06, |
|
"loss": 1.4721, |
|
"step": 18700 |
|
}, |
|
{ |
|
"epoch": 0.874947642760739, |
|
"grad_norm": 1.0318922996520996, |
|
"learning_rate": 9.403769090194192e-07, |
|
"loss": 1.4542, |
|
"step": 18800 |
|
}, |
|
{ |
|
"epoch": 0.8796016195839345, |
|
"grad_norm": 1.0590611696243286, |
|
"learning_rate": 8.72801350051331e-07, |
|
"loss": 1.497, |
|
"step": 18900 |
|
}, |
|
{ |
|
"epoch": 0.8842555964071299, |
|
"grad_norm": 0.5952686667442322, |
|
"learning_rate": 8.076343632909211e-07, |
|
"loss": 1.4613, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.8889095732303253, |
|
"grad_norm": 0.8993837237358093, |
|
"learning_rate": 7.448931456240205e-07, |
|
"loss": 1.5027, |
|
"step": 19100 |
|
}, |
|
{ |
|
"epoch": 0.8935635500535207, |
|
"grad_norm": 0.8999748229980469, |
|
"learning_rate": 6.845942538013894e-07, |
|
"loss": 1.4677, |
|
"step": 19200 |
|
}, |
|
{ |
|
"epoch": 0.8982175268767162, |
|
"grad_norm": 1.1852108240127563, |
|
"learning_rate": 6.273197876706216e-07, |
|
"loss": 1.4537, |
|
"step": 19300 |
|
}, |
|
{ |
|
"epoch": 0.9028715036999115, |
|
"grad_norm": 0.8120535016059875, |
|
"learning_rate": 5.71927827141906e-07, |
|
"loss": 1.4965, |
|
"step": 19400 |
|
}, |
|
{ |
|
"epoch": 0.907525480523107, |
|
"grad_norm": 0.8027576804161072, |
|
"learning_rate": 5.190238361949229e-07, |
|
"loss": 1.5109, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 0.9121794573463025, |
|
"grad_norm": 1.2424169778823853, |
|
"learning_rate": 4.686217756392464e-07, |
|
"loss": 1.4614, |
|
"step": 19600 |
|
}, |
|
{ |
|
"epoch": 0.9168334341694978, |
|
"grad_norm": 1.248057246208191, |
|
"learning_rate": 4.207349460512022e-07, |
|
"loss": 1.4786, |
|
"step": 19700 |
|
}, |
|
{ |
|
"epoch": 0.9214874109926933, |
|
"grad_norm": 0.8463609218597412, |
|
"learning_rate": 3.7537598426397925e-07, |
|
"loss": 1.4714, |
|
"step": 19800 |
|
}, |
|
{ |
|
"epoch": 0.9261413878158887, |
|
"grad_norm": 0.6483246684074402, |
|
"learning_rate": 3.325568600329143e-07, |
|
"loss": 1.5209, |
|
"step": 19900 |
|
}, |
|
{ |
|
"epoch": 0.9307953646390841, |
|
"grad_norm": 0.6988236904144287, |
|
"learning_rate": 2.922888728767903e-07, |
|
"loss": 1.485, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.9354493414622795, |
|
"grad_norm": 0.9654485583305359, |
|
"learning_rate": 2.545826490960146e-07, |
|
"loss": 1.4532, |
|
"step": 20100 |
|
}, |
|
{ |
|
"epoch": 0.9401033182854749, |
|
"grad_norm": 1.0685368776321411, |
|
"learning_rate": 2.194481389684433e-07, |
|
"loss": 1.4925, |
|
"step": 20200 |
|
}, |
|
{ |
|
"epoch": 0.9447572951086703, |
|
"grad_norm": 1.161594033241272, |
|
"learning_rate": 1.8689461412362164e-07, |
|
"loss": 1.4376, |
|
"step": 20300 |
|
}, |
|
{ |
|
"epoch": 0.9494112719318658, |
|
"grad_norm": 1.1302499771118164, |
|
"learning_rate": 1.5693066509608333e-07, |
|
"loss": 1.4776, |
|
"step": 20400 |
|
}, |
|
{ |
|
"epoch": 0.9540652487550612, |
|
"grad_norm": 1.0777382850646973, |
|
"learning_rate": 1.295641990584007e-07, |
|
"loss": 1.4702, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 0.9587192255782566, |
|
"grad_norm": 1.027956247329712, |
|
"learning_rate": 1.0503714003311738e-07, |
|
"loss": 1.4699, |
|
"step": 20600 |
|
}, |
|
{ |
|
"epoch": 0.9633732024014521, |
|
"grad_norm": 0.7537710666656494, |
|
"learning_rate": 8.286047531811392e-08, |
|
"loss": 1.489, |
|
"step": 20700 |
|
}, |
|
{ |
|
"epoch": 0.9680271792246474, |
|
"grad_norm": 0.7544810175895691, |
|
"learning_rate": 6.33008399409063e-08, |
|
"loss": 1.4563, |
|
"step": 20800 |
|
}, |
|
{ |
|
"epoch": 0.9726811560478429, |
|
"grad_norm": 0.9680439829826355, |
|
"learning_rate": 4.636339548457991e-08, |
|
"loss": 1.4345, |
|
"step": 20900 |
|
}, |
|
{ |
|
"epoch": 0.9773351328710383, |
|
"grad_norm": 0.773059606552124, |
|
"learning_rate": 3.205261156347783e-08, |
|
"loss": 1.4591, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.9819891096942337, |
|
"grad_norm": 0.8769505023956299, |
|
"learning_rate": 2.0372264643713223e-08, |
|
"loss": 1.4756, |
|
"step": 21100 |
|
}, |
|
{ |
|
"epoch": 0.9866430865174292, |
|
"grad_norm": 0.9209620356559753, |
|
"learning_rate": 1.132543704660316e-08, |
|
"loss": 1.4957, |
|
"step": 21200 |
|
}, |
|
{ |
|
"epoch": 0.9912970633406246, |
|
"grad_norm": 0.9970301389694214, |
|
"learning_rate": 4.914516135275937e-09, |
|
"loss": 1.4321, |
|
"step": 21300 |
|
}, |
|
{ |
|
"epoch": 0.99595104016382, |
|
"grad_norm": 0.6997882127761841, |
|
"learning_rate": 1.1411936846705828e-09, |
|
"loss": 1.4656, |
|
"step": 21400 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 21487, |
|
"total_flos": 3.101747713212678e+18, |
|
"train_loss": 1.5223615952344096, |
|
"train_runtime": 9293.396, |
|
"train_samples_per_second": 4.624, |
|
"train_steps_per_second": 2.312 |
|
} |
|
], |
|
"logging_steps": 100, |
|
"max_steps": 21487, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3.101747713212678e+18, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|