|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.034174125305126, |
|
"eval_steps": 1000, |
|
"global_step": 5000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01017087062652563, |
|
"grad_norm": 16.184080123901367, |
|
"learning_rate": 4.800000000000001e-07, |
|
"loss": 1.6215, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.02034174125305126, |
|
"grad_norm": 14.08956241607666, |
|
"learning_rate": 9.800000000000001e-07, |
|
"loss": 1.2721, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.030512611879576892, |
|
"grad_norm": 11.49959945678711, |
|
"learning_rate": 1.48e-06, |
|
"loss": 0.9884, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.04068348250610252, |
|
"grad_norm": 8.895913124084473, |
|
"learning_rate": 1.98e-06, |
|
"loss": 0.8726, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.050854353132628156, |
|
"grad_norm": 9.52476692199707, |
|
"learning_rate": 2.4800000000000004e-06, |
|
"loss": 0.8062, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.061025223759153785, |
|
"grad_norm": 9.508084297180176, |
|
"learning_rate": 2.9800000000000003e-06, |
|
"loss": 0.7792, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.07119609438567942, |
|
"grad_norm": 9.352806091308594, |
|
"learning_rate": 3.48e-06, |
|
"loss": 0.7541, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.08136696501220504, |
|
"grad_norm": 10.18264102935791, |
|
"learning_rate": 3.980000000000001e-06, |
|
"loss": 0.6983, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.09153783563873068, |
|
"grad_norm": 10.840176582336426, |
|
"learning_rate": 4.48e-06, |
|
"loss": 0.7416, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.10170870626525631, |
|
"grad_norm": 9.403505325317383, |
|
"learning_rate": 4.980000000000001e-06, |
|
"loss": 0.7154, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.11187957689178193, |
|
"grad_norm": 8.170647621154785, |
|
"learning_rate": 5.480000000000001e-06, |
|
"loss": 0.6564, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.12205044751830757, |
|
"grad_norm": 9.974289894104004, |
|
"learning_rate": 5.98e-06, |
|
"loss": 0.6401, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.1322213181448332, |
|
"grad_norm": 10.3145112991333, |
|
"learning_rate": 6.480000000000001e-06, |
|
"loss": 0.6564, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.14239218877135884, |
|
"grad_norm": 8.214014053344727, |
|
"learning_rate": 6.98e-06, |
|
"loss": 0.6191, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.15256305939788445, |
|
"grad_norm": 8.800423622131348, |
|
"learning_rate": 7.48e-06, |
|
"loss": 0.6492, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.16273393002441008, |
|
"grad_norm": 8.481072425842285, |
|
"learning_rate": 7.980000000000002e-06, |
|
"loss": 0.6295, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.17290480065093572, |
|
"grad_norm": 8.669546127319336, |
|
"learning_rate": 8.48e-06, |
|
"loss": 0.6167, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.18307567127746135, |
|
"grad_norm": 7.645444869995117, |
|
"learning_rate": 8.98e-06, |
|
"loss": 0.6021, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.193246541903987, |
|
"grad_norm": 11.719562530517578, |
|
"learning_rate": 9.48e-06, |
|
"loss": 0.6006, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.20341741253051263, |
|
"grad_norm": 9.42712116241455, |
|
"learning_rate": 9.980000000000001e-06, |
|
"loss": 0.5777, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.21358828315703823, |
|
"grad_norm": 7.9745988845825195, |
|
"learning_rate": 9.946666666666667e-06, |
|
"loss": 0.5928, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.22375915378356387, |
|
"grad_norm": 7.762495994567871, |
|
"learning_rate": 9.891111111111113e-06, |
|
"loss": 0.5735, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.2339300244100895, |
|
"grad_norm": 8.57645320892334, |
|
"learning_rate": 9.835555555555556e-06, |
|
"loss": 0.5613, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.24410089503661514, |
|
"grad_norm": 7.481997489929199, |
|
"learning_rate": 9.780000000000001e-06, |
|
"loss": 0.5496, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.25427176566314075, |
|
"grad_norm": 6.589779376983643, |
|
"learning_rate": 9.724444444444445e-06, |
|
"loss": 0.5746, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.2644426362896664, |
|
"grad_norm": 8.590812683105469, |
|
"learning_rate": 9.66888888888889e-06, |
|
"loss": 0.5385, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.274613506916192, |
|
"grad_norm": 7.343937397003174, |
|
"learning_rate": 9.613333333333335e-06, |
|
"loss": 0.5432, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.2847843775427177, |
|
"grad_norm": 7.952889919281006, |
|
"learning_rate": 9.557777777777777e-06, |
|
"loss": 0.5213, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.2949552481692433, |
|
"grad_norm": 8.083431243896484, |
|
"learning_rate": 9.502222222222223e-06, |
|
"loss": 0.5367, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.3051261187957689, |
|
"grad_norm": 7.172378063201904, |
|
"learning_rate": 9.446666666666667e-06, |
|
"loss": 0.5317, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.31529698942229456, |
|
"grad_norm": 7.955025672912598, |
|
"learning_rate": 9.391111111111111e-06, |
|
"loss": 0.5208, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.32546786004882017, |
|
"grad_norm": 7.041996479034424, |
|
"learning_rate": 9.335555555555557e-06, |
|
"loss": 0.5123, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.33563873067534583, |
|
"grad_norm": 7.7593607902526855, |
|
"learning_rate": 9.280000000000001e-06, |
|
"loss": 0.5019, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.34580960130187144, |
|
"grad_norm": 9.114460945129395, |
|
"learning_rate": 9.224444444444445e-06, |
|
"loss": 0.5189, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.35598047192839705, |
|
"grad_norm": 8.199577331542969, |
|
"learning_rate": 9.171111111111112e-06, |
|
"loss": 0.4844, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.3661513425549227, |
|
"grad_norm": 7.192209243774414, |
|
"learning_rate": 9.115555555555556e-06, |
|
"loss": 0.4867, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.3763222131814483, |
|
"grad_norm": 7.112555980682373, |
|
"learning_rate": 9.060000000000001e-06, |
|
"loss": 0.4673, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.386493083807974, |
|
"grad_norm": 6.723269462585449, |
|
"learning_rate": 9.004444444444445e-06, |
|
"loss": 0.4886, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.3966639544344996, |
|
"grad_norm": 6.772889137268066, |
|
"learning_rate": 8.94888888888889e-06, |
|
"loss": 0.5116, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.40683482506102525, |
|
"grad_norm": 7.985610485076904, |
|
"learning_rate": 8.893333333333333e-06, |
|
"loss": 0.5072, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.40683482506102525, |
|
"eval_loss": 0.5020337700843811, |
|
"eval_runtime": 1789.7876, |
|
"eval_samples_per_second": 2.18, |
|
"eval_steps_per_second": 0.136, |
|
"eval_wer": 0.35670820052158797, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.41700569568755086, |
|
"grad_norm": 10.103032112121582, |
|
"learning_rate": 8.83777777777778e-06, |
|
"loss": 0.4951, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.42717656631407647, |
|
"grad_norm": 6.815390586853027, |
|
"learning_rate": 8.782222222222223e-06, |
|
"loss": 0.4682, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.43734743694060213, |
|
"grad_norm": 7.63904333114624, |
|
"learning_rate": 8.726666666666667e-06, |
|
"loss": 0.4842, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.44751830756712774, |
|
"grad_norm": 7.750236988067627, |
|
"learning_rate": 8.671111111111113e-06, |
|
"loss": 0.4823, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.4576891781936534, |
|
"grad_norm": 6.178125858306885, |
|
"learning_rate": 8.615555555555555e-06, |
|
"loss": 0.4764, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.467860048820179, |
|
"grad_norm": 6.7011284828186035, |
|
"learning_rate": 8.560000000000001e-06, |
|
"loss": 0.487, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.4780309194467046, |
|
"grad_norm": 6.630568504333496, |
|
"learning_rate": 8.504444444444445e-06, |
|
"loss": 0.444, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.4882017900732303, |
|
"grad_norm": 7.386712074279785, |
|
"learning_rate": 8.448888888888889e-06, |
|
"loss": 0.4372, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.4983726606997559, |
|
"grad_norm": 6.448077201843262, |
|
"learning_rate": 8.393333333333335e-06, |
|
"loss": 0.4474, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.5085435313262815, |
|
"grad_norm": 6.787522315979004, |
|
"learning_rate": 8.337777777777777e-06, |
|
"loss": 0.4692, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.5187144019528072, |
|
"grad_norm": 6.614894866943359, |
|
"learning_rate": 8.282222222222223e-06, |
|
"loss": 0.4401, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.5288852725793328, |
|
"grad_norm": 7.09251594543457, |
|
"learning_rate": 8.226666666666667e-06, |
|
"loss": 0.4419, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.5390561432058584, |
|
"grad_norm": 7.047774791717529, |
|
"learning_rate": 8.171111111111113e-06, |
|
"loss": 0.4323, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 0.549227013832384, |
|
"grad_norm": 6.616914749145508, |
|
"learning_rate": 8.115555555555557e-06, |
|
"loss": 0.4354, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.5593978844589097, |
|
"grad_norm": 8.16896915435791, |
|
"learning_rate": 8.06e-06, |
|
"loss": 0.4299, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 0.5695687550854354, |
|
"grad_norm": 7.04884672164917, |
|
"learning_rate": 8.004444444444445e-06, |
|
"loss": 0.4477, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.5797396257119609, |
|
"grad_norm": 6.052219867706299, |
|
"learning_rate": 7.948888888888889e-06, |
|
"loss": 0.4561, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 0.5899104963384866, |
|
"grad_norm": 6.257965087890625, |
|
"learning_rate": 7.893333333333335e-06, |
|
"loss": 0.4343, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.6000813669650122, |
|
"grad_norm": 7.182107925415039, |
|
"learning_rate": 7.837777777777779e-06, |
|
"loss": 0.4215, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 0.6102522375915378, |
|
"grad_norm": 6.276819705963135, |
|
"learning_rate": 7.782222222222223e-06, |
|
"loss": 0.4188, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.6204231082180635, |
|
"grad_norm": 6.533692359924316, |
|
"learning_rate": 7.726666666666667e-06, |
|
"loss": 0.4855, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 0.6305939788445891, |
|
"grad_norm": 6.108315467834473, |
|
"learning_rate": 7.67111111111111e-06, |
|
"loss": 0.4255, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.6407648494711147, |
|
"grad_norm": 6.741276741027832, |
|
"learning_rate": 7.6155555555555564e-06, |
|
"loss": 0.4529, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 0.6509357200976403, |
|
"grad_norm": 6.165073871612549, |
|
"learning_rate": 7.5600000000000005e-06, |
|
"loss": 0.4541, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.661106590724166, |
|
"grad_norm": 7.246621608734131, |
|
"learning_rate": 7.504444444444445e-06, |
|
"loss": 0.4073, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 0.6712774613506917, |
|
"grad_norm": 7.50054407119751, |
|
"learning_rate": 7.44888888888889e-06, |
|
"loss": 0.4143, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.6814483319772172, |
|
"grad_norm": 6.225817680358887, |
|
"learning_rate": 7.393333333333333e-06, |
|
"loss": 0.438, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 0.6916192026037429, |
|
"grad_norm": 6.704942226409912, |
|
"learning_rate": 7.337777777777778e-06, |
|
"loss": 0.4209, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.7017900732302685, |
|
"grad_norm": 6.276001930236816, |
|
"learning_rate": 7.282222222222222e-06, |
|
"loss": 0.4033, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 0.7119609438567941, |
|
"grad_norm": 7.256411075592041, |
|
"learning_rate": 7.226666666666667e-06, |
|
"loss": 0.416, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.7221318144833198, |
|
"grad_norm": 6.347288608551025, |
|
"learning_rate": 7.171111111111112e-06, |
|
"loss": 0.4373, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 0.7323026851098454, |
|
"grad_norm": 6.241955757141113, |
|
"learning_rate": 7.115555555555557e-06, |
|
"loss": 0.403, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.7424735557363711, |
|
"grad_norm": 5.548867702484131, |
|
"learning_rate": 7.06e-06, |
|
"loss": 0.4034, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 0.7526444263628966, |
|
"grad_norm": 7.0312113761901855, |
|
"learning_rate": 7.004444444444445e-06, |
|
"loss": 0.4037, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.7628152969894223, |
|
"grad_norm": 7.7608256340026855, |
|
"learning_rate": 6.948888888888889e-06, |
|
"loss": 0.3939, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 0.772986167615948, |
|
"grad_norm": 6.072271823883057, |
|
"learning_rate": 6.893333333333334e-06, |
|
"loss": 0.4067, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.7831570382424735, |
|
"grad_norm": 6.9571123123168945, |
|
"learning_rate": 6.837777777777779e-06, |
|
"loss": 0.4107, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 0.7933279088689992, |
|
"grad_norm": 5.884681224822998, |
|
"learning_rate": 6.782222222222222e-06, |
|
"loss": 0.3995, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.8034987794955248, |
|
"grad_norm": 5.850441932678223, |
|
"learning_rate": 6.726666666666667e-06, |
|
"loss": 0.4017, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 0.8136696501220505, |
|
"grad_norm": 6.021271705627441, |
|
"learning_rate": 6.671111111111112e-06, |
|
"loss": 0.3708, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.8136696501220505, |
|
"eval_loss": 0.42604002356529236, |
|
"eval_runtime": 1798.3084, |
|
"eval_samples_per_second": 2.169, |
|
"eval_steps_per_second": 0.136, |
|
"eval_wer": 0.32584758041147494, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.823840520748576, |
|
"grad_norm": 6.479699611663818, |
|
"learning_rate": 6.615555555555556e-06, |
|
"loss": 0.3914, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 0.8340113913751017, |
|
"grad_norm": 6.589993476867676, |
|
"learning_rate": 6.560000000000001e-06, |
|
"loss": 0.401, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.8441822620016274, |
|
"grad_norm": 5.026104927062988, |
|
"learning_rate": 6.504444444444446e-06, |
|
"loss": 0.38, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 0.8543531326281529, |
|
"grad_norm": 6.458845138549805, |
|
"learning_rate": 6.448888888888889e-06, |
|
"loss": 0.3885, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.8645240032546786, |
|
"grad_norm": 6.570550441741943, |
|
"learning_rate": 6.393333333333334e-06, |
|
"loss": 0.3992, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 0.8746948738812043, |
|
"grad_norm": 6.160183906555176, |
|
"learning_rate": 6.3377777777777786e-06, |
|
"loss": 0.385, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.8848657445077298, |
|
"grad_norm": 5.925890922546387, |
|
"learning_rate": 6.282222222222223e-06, |
|
"loss": 0.4082, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 0.8950366151342555, |
|
"grad_norm": 6.508133888244629, |
|
"learning_rate": 6.2266666666666675e-06, |
|
"loss": 0.4043, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.9052074857607811, |
|
"grad_norm": 6.300657272338867, |
|
"learning_rate": 6.171111111111112e-06, |
|
"loss": 0.3729, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 0.9153783563873068, |
|
"grad_norm": 5.284305572509766, |
|
"learning_rate": 6.1155555555555555e-06, |
|
"loss": 0.4046, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.9255492270138324, |
|
"grad_norm": 5.809286594390869, |
|
"learning_rate": 6.0600000000000004e-06, |
|
"loss": 0.3864, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 0.935720097640358, |
|
"grad_norm": 6.667562007904053, |
|
"learning_rate": 6.004444444444445e-06, |
|
"loss": 0.3929, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.9458909682668837, |
|
"grad_norm": 6.4545392990112305, |
|
"learning_rate": 5.948888888888889e-06, |
|
"loss": 0.3975, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 0.9560618388934092, |
|
"grad_norm": 5.876199245452881, |
|
"learning_rate": 5.893333333333334e-06, |
|
"loss": 0.3769, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.9662327095199349, |
|
"grad_norm": 5.008429527282715, |
|
"learning_rate": 5.837777777777777e-06, |
|
"loss": 0.4031, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 0.9764035801464606, |
|
"grad_norm": 7.304080963134766, |
|
"learning_rate": 5.782222222222222e-06, |
|
"loss": 0.3804, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.9865744507729862, |
|
"grad_norm": 7.11293888092041, |
|
"learning_rate": 5.726666666666667e-06, |
|
"loss": 0.3849, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 0.9967453213995118, |
|
"grad_norm": 6.156993389129639, |
|
"learning_rate": 5.671111111111112e-06, |
|
"loss": 0.3804, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 1.0069161920260374, |
|
"grad_norm": 4.09175968170166, |
|
"learning_rate": 5.615555555555556e-06, |
|
"loss": 0.3138, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 1.017087062652563, |
|
"grad_norm": 5.104438304901123, |
|
"learning_rate": 5.560000000000001e-06, |
|
"loss": 0.2747, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.0272579332790888, |
|
"grad_norm": 5.911495685577393, |
|
"learning_rate": 5.504444444444444e-06, |
|
"loss": 0.2825, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 1.0374288039056143, |
|
"grad_norm": 5.619592189788818, |
|
"learning_rate": 5.448888888888889e-06, |
|
"loss": 0.2991, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 1.0475996745321399, |
|
"grad_norm": 5.471043109893799, |
|
"learning_rate": 5.393333333333334e-06, |
|
"loss": 0.2766, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 1.0577705451586656, |
|
"grad_norm": 4.353448867797852, |
|
"learning_rate": 5.337777777777779e-06, |
|
"loss": 0.2883, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 1.0679414157851912, |
|
"grad_norm": 4.982877731323242, |
|
"learning_rate": 5.282222222222223e-06, |
|
"loss": 0.2888, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 1.0781122864117167, |
|
"grad_norm": 5.228238582611084, |
|
"learning_rate": 5.226666666666667e-06, |
|
"loss": 0.2802, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 1.0882831570382425, |
|
"grad_norm": 5.883535385131836, |
|
"learning_rate": 5.171111111111111e-06, |
|
"loss": 0.2779, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 1.098454027664768, |
|
"grad_norm": 4.599557876586914, |
|
"learning_rate": 5.115555555555556e-06, |
|
"loss": 0.2817, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 1.1086248982912936, |
|
"grad_norm": 5.081583023071289, |
|
"learning_rate": 5.060000000000001e-06, |
|
"loss": 0.2804, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 1.1187957689178194, |
|
"grad_norm": 5.387673854827881, |
|
"learning_rate": 5.004444444444445e-06, |
|
"loss": 0.2676, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 1.128966639544345, |
|
"grad_norm": 5.010275840759277, |
|
"learning_rate": 4.94888888888889e-06, |
|
"loss": 0.2752, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 1.1391375101708707, |
|
"grad_norm": 5.828437805175781, |
|
"learning_rate": 4.893333333333334e-06, |
|
"loss": 0.2795, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 1.1493083807973963, |
|
"grad_norm": 5.313021183013916, |
|
"learning_rate": 4.837777777777778e-06, |
|
"loss": 0.2876, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 1.1594792514239218, |
|
"grad_norm": 5.790019512176514, |
|
"learning_rate": 4.7822222222222226e-06, |
|
"loss": 0.283, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 1.1696501220504476, |
|
"grad_norm": 5.2214250564575195, |
|
"learning_rate": 4.7266666666666674e-06, |
|
"loss": 0.2762, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 1.1798209926769732, |
|
"grad_norm": 6.338456630706787, |
|
"learning_rate": 4.6711111111111115e-06, |
|
"loss": 0.2588, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 1.1899918633034987, |
|
"grad_norm": 6.858678340911865, |
|
"learning_rate": 4.6155555555555555e-06, |
|
"loss": 0.2724, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 1.2001627339300245, |
|
"grad_norm": 5.54526948928833, |
|
"learning_rate": 4.56e-06, |
|
"loss": 0.2893, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 1.21033360455655, |
|
"grad_norm": 5.165172100067139, |
|
"learning_rate": 4.504444444444444e-06, |
|
"loss": 0.2751, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 1.2205044751830756, |
|
"grad_norm": 6.359204292297363, |
|
"learning_rate": 4.448888888888889e-06, |
|
"loss": 0.2599, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.2205044751830756, |
|
"eval_loss": 0.39731866121292114, |
|
"eval_runtime": 1796.0978, |
|
"eval_samples_per_second": 2.172, |
|
"eval_steps_per_second": 0.136, |
|
"eval_wer": 0.30034772529701537, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.2306753458096014, |
|
"grad_norm": 4.958346366882324, |
|
"learning_rate": 4.393333333333334e-06, |
|
"loss": 0.2754, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 1.240846216436127, |
|
"grad_norm": 7.284126281738281, |
|
"learning_rate": 4.337777777777778e-06, |
|
"loss": 0.2764, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 1.2510170870626527, |
|
"grad_norm": 5.106224536895752, |
|
"learning_rate": 4.282222222222222e-06, |
|
"loss": 0.2679, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 1.2611879576891782, |
|
"grad_norm": 4.638324737548828, |
|
"learning_rate": 4.226666666666667e-06, |
|
"loss": 0.2678, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 1.2713588283157038, |
|
"grad_norm": 5.0809245109558105, |
|
"learning_rate": 4.171111111111111e-06, |
|
"loss": 0.2679, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 1.2815296989422293, |
|
"grad_norm": 5.177508354187012, |
|
"learning_rate": 4.115555555555556e-06, |
|
"loss": 0.2603, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 1.2917005695687551, |
|
"grad_norm": 4.664310932159424, |
|
"learning_rate": 4.060000000000001e-06, |
|
"loss": 0.2659, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 1.3018714401952807, |
|
"grad_norm": 5.230463981628418, |
|
"learning_rate": 4.004444444444445e-06, |
|
"loss": 0.271, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 1.3120423108218064, |
|
"grad_norm": 5.396975994110107, |
|
"learning_rate": 3.948888888888889e-06, |
|
"loss": 0.2653, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 1.322213181448332, |
|
"grad_norm": 4.997494220733643, |
|
"learning_rate": 3.893333333333333e-06, |
|
"loss": 0.2678, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 1.3323840520748575, |
|
"grad_norm": 6.467133045196533, |
|
"learning_rate": 3.837777777777778e-06, |
|
"loss": 0.2628, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 1.342554922701383, |
|
"grad_norm": 4.6464338302612305, |
|
"learning_rate": 3.782222222222223e-06, |
|
"loss": 0.2819, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 1.3527257933279089, |
|
"grad_norm": 5.477710723876953, |
|
"learning_rate": 3.726666666666667e-06, |
|
"loss": 0.2694, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 1.3628966639544344, |
|
"grad_norm": 5.745607376098633, |
|
"learning_rate": 3.6711111111111113e-06, |
|
"loss": 0.2731, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 1.3730675345809602, |
|
"grad_norm": 4.110312461853027, |
|
"learning_rate": 3.615555555555556e-06, |
|
"loss": 0.2726, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 1.3832384052074858, |
|
"grad_norm": 5.377216815948486, |
|
"learning_rate": 3.5600000000000002e-06, |
|
"loss": 0.2677, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 1.3934092758340113, |
|
"grad_norm": 5.4512763023376465, |
|
"learning_rate": 3.5044444444444447e-06, |
|
"loss": 0.2621, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 1.403580146460537, |
|
"grad_norm": 5.338203430175781, |
|
"learning_rate": 3.4488888888888896e-06, |
|
"loss": 0.2731, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 1.4137510170870626, |
|
"grad_norm": 5.551027774810791, |
|
"learning_rate": 3.3933333333333336e-06, |
|
"loss": 0.2821, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 1.4239218877135884, |
|
"grad_norm": 6.804419040679932, |
|
"learning_rate": 3.337777777777778e-06, |
|
"loss": 0.2652, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 1.434092758340114, |
|
"grad_norm": 5.15053653717041, |
|
"learning_rate": 3.282222222222223e-06, |
|
"loss": 0.2626, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 1.4442636289666395, |
|
"grad_norm": 4.987846851348877, |
|
"learning_rate": 3.226666666666667e-06, |
|
"loss": 0.2558, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 1.454434499593165, |
|
"grad_norm": 5.460192680358887, |
|
"learning_rate": 3.1711111111111114e-06, |
|
"loss": 0.2688, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 1.4646053702196908, |
|
"grad_norm": 5.20620059967041, |
|
"learning_rate": 3.1155555555555555e-06, |
|
"loss": 0.2541, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 1.4747762408462164, |
|
"grad_norm": 5.177892208099365, |
|
"learning_rate": 3.0600000000000003e-06, |
|
"loss": 0.2546, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 1.4849471114727422, |
|
"grad_norm": 5.275447845458984, |
|
"learning_rate": 3.004444444444445e-06, |
|
"loss": 0.2487, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 1.4951179820992677, |
|
"grad_norm": 4.815367698669434, |
|
"learning_rate": 2.948888888888889e-06, |
|
"loss": 0.2505, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 1.5052888527257933, |
|
"grad_norm": 4.586023330688477, |
|
"learning_rate": 2.8933333333333337e-06, |
|
"loss": 0.2385, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 1.5154597233523188, |
|
"grad_norm": 5.920341491699219, |
|
"learning_rate": 2.837777777777778e-06, |
|
"loss": 0.2502, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 1.5256305939788446, |
|
"grad_norm": 5.697859764099121, |
|
"learning_rate": 2.7822222222222222e-06, |
|
"loss": 0.2574, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 1.5358014646053704, |
|
"grad_norm": 4.68679666519165, |
|
"learning_rate": 2.726666666666667e-06, |
|
"loss": 0.2543, |
|
"step": 3775 |
|
}, |
|
{ |
|
"epoch": 1.545972335231896, |
|
"grad_norm": 5.064055442810059, |
|
"learning_rate": 2.6711111111111116e-06, |
|
"loss": 0.2639, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 1.5561432058584215, |
|
"grad_norm": 5.656905651092529, |
|
"learning_rate": 2.6155555555555556e-06, |
|
"loss": 0.2639, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 1.566314076484947, |
|
"grad_norm": 5.397951126098633, |
|
"learning_rate": 2.56e-06, |
|
"loss": 0.271, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 1.5764849471114726, |
|
"grad_norm": 6.378788948059082, |
|
"learning_rate": 2.504444444444445e-06, |
|
"loss": 0.2542, |
|
"step": 3875 |
|
}, |
|
{ |
|
"epoch": 1.5866558177379984, |
|
"grad_norm": 6.6806960105896, |
|
"learning_rate": 2.448888888888889e-06, |
|
"loss": 0.2471, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 1.5968266883645241, |
|
"grad_norm": 4.837161064147949, |
|
"learning_rate": 2.3933333333333334e-06, |
|
"loss": 0.2497, |
|
"step": 3925 |
|
}, |
|
{ |
|
"epoch": 1.6069975589910497, |
|
"grad_norm": 5.817267894744873, |
|
"learning_rate": 2.337777777777778e-06, |
|
"loss": 0.2583, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 1.6171684296175752, |
|
"grad_norm": 5.383573532104492, |
|
"learning_rate": 2.2822222222222223e-06, |
|
"loss": 0.2535, |
|
"step": 3975 |
|
}, |
|
{ |
|
"epoch": 1.6273393002441008, |
|
"grad_norm": 4.556544780731201, |
|
"learning_rate": 2.226666666666667e-06, |
|
"loss": 0.2618, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.6273393002441008, |
|
"eval_loss": 0.37833017110824585, |
|
"eval_runtime": 1788.9834, |
|
"eval_samples_per_second": 2.181, |
|
"eval_steps_per_second": 0.136, |
|
"eval_wer": 0.2904713609581764, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.6375101708706266, |
|
"grad_norm": 5.119606018066406, |
|
"learning_rate": 2.1711111111111113e-06, |
|
"loss": 0.2613, |
|
"step": 4025 |
|
}, |
|
{ |
|
"epoch": 1.647681041497152, |
|
"grad_norm": 5.435574054718018, |
|
"learning_rate": 2.1155555555555557e-06, |
|
"loss": 0.243, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 1.6578519121236779, |
|
"grad_norm": 5.2191643714904785, |
|
"learning_rate": 2.06e-06, |
|
"loss": 0.2563, |
|
"step": 4075 |
|
}, |
|
{ |
|
"epoch": 1.6680227827502034, |
|
"grad_norm": 4.851512908935547, |
|
"learning_rate": 2.0044444444444446e-06, |
|
"loss": 0.2608, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 1.678193653376729, |
|
"grad_norm": 5.570041179656982, |
|
"learning_rate": 1.948888888888889e-06, |
|
"loss": 0.2572, |
|
"step": 4125 |
|
}, |
|
{ |
|
"epoch": 1.6883645240032545, |
|
"grad_norm": 4.914021968841553, |
|
"learning_rate": 1.8933333333333333e-06, |
|
"loss": 0.2502, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 1.6985353946297803, |
|
"grad_norm": 4.91426944732666, |
|
"learning_rate": 1.837777777777778e-06, |
|
"loss": 0.2484, |
|
"step": 4175 |
|
}, |
|
{ |
|
"epoch": 1.708706265256306, |
|
"grad_norm": 5.054934501647949, |
|
"learning_rate": 1.7822222222222225e-06, |
|
"loss": 0.2543, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 1.7188771358828316, |
|
"grad_norm": 5.295421600341797, |
|
"learning_rate": 1.7266666666666667e-06, |
|
"loss": 0.2479, |
|
"step": 4225 |
|
}, |
|
{ |
|
"epoch": 1.7290480065093572, |
|
"grad_norm": 4.494225978851318, |
|
"learning_rate": 1.6711111111111112e-06, |
|
"loss": 0.2708, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 1.7392188771358827, |
|
"grad_norm": 5.637074947357178, |
|
"learning_rate": 1.6155555555555559e-06, |
|
"loss": 0.2536, |
|
"step": 4275 |
|
}, |
|
{ |
|
"epoch": 1.7493897477624083, |
|
"grad_norm": 4.7007222175598145, |
|
"learning_rate": 1.56e-06, |
|
"loss": 0.2445, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 1.759560618388934, |
|
"grad_norm": 5.922358989715576, |
|
"learning_rate": 1.5044444444444446e-06, |
|
"loss": 0.2629, |
|
"step": 4325 |
|
}, |
|
{ |
|
"epoch": 1.7697314890154598, |
|
"grad_norm": 5.818343162536621, |
|
"learning_rate": 1.4488888888888892e-06, |
|
"loss": 0.2462, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 1.7799023596419854, |
|
"grad_norm": 4.40110969543457, |
|
"learning_rate": 1.3933333333333335e-06, |
|
"loss": 0.2389, |
|
"step": 4375 |
|
}, |
|
{ |
|
"epoch": 1.790073230268511, |
|
"grad_norm": 6.067351818084717, |
|
"learning_rate": 1.337777777777778e-06, |
|
"loss": 0.2662, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 1.8002441008950365, |
|
"grad_norm": 4.228072166442871, |
|
"learning_rate": 1.2822222222222222e-06, |
|
"loss": 0.2611, |
|
"step": 4425 |
|
}, |
|
{ |
|
"epoch": 1.8104149715215623, |
|
"grad_norm": 5.135393142700195, |
|
"learning_rate": 1.2266666666666666e-06, |
|
"loss": 0.2607, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 1.8205858421480878, |
|
"grad_norm": 5.195592403411865, |
|
"learning_rate": 1.171111111111111e-06, |
|
"loss": 0.2534, |
|
"step": 4475 |
|
}, |
|
{ |
|
"epoch": 1.8307567127746136, |
|
"grad_norm": 4.900667667388916, |
|
"learning_rate": 1.1155555555555558e-06, |
|
"loss": 0.2558, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 1.8409275834011392, |
|
"grad_norm": 4.837711811065674, |
|
"learning_rate": 1.06e-06, |
|
"loss": 0.2476, |
|
"step": 4525 |
|
}, |
|
{ |
|
"epoch": 1.8510984540276647, |
|
"grad_norm": 5.072052955627441, |
|
"learning_rate": 1.0044444444444445e-06, |
|
"loss": 0.2333, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 1.8612693246541903, |
|
"grad_norm": 5.385496139526367, |
|
"learning_rate": 9.488888888888889e-07, |
|
"loss": 0.2453, |
|
"step": 4575 |
|
}, |
|
{ |
|
"epoch": 1.871440195280716, |
|
"grad_norm": 5.36069393157959, |
|
"learning_rate": 8.933333333333334e-07, |
|
"loss": 0.2477, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 1.8816110659072418, |
|
"grad_norm": 4.781914710998535, |
|
"learning_rate": 8.37777777777778e-07, |
|
"loss": 0.2494, |
|
"step": 4625 |
|
}, |
|
{ |
|
"epoch": 1.8917819365337674, |
|
"grad_norm": 5.366285800933838, |
|
"learning_rate": 7.822222222222223e-07, |
|
"loss": 0.2519, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 1.901952807160293, |
|
"grad_norm": 4.479983806610107, |
|
"learning_rate": 7.266666666666668e-07, |
|
"loss": 0.241, |
|
"step": 4675 |
|
}, |
|
{ |
|
"epoch": 1.9121236777868185, |
|
"grad_norm": 5.053223133087158, |
|
"learning_rate": 6.711111111111111e-07, |
|
"loss": 0.2462, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 1.922294548413344, |
|
"grad_norm": 4.833332061767578, |
|
"learning_rate": 6.155555555555556e-07, |
|
"loss": 0.2698, |
|
"step": 4725 |
|
}, |
|
{ |
|
"epoch": 1.9324654190398698, |
|
"grad_norm": 3.715808629989624, |
|
"learning_rate": 5.6e-07, |
|
"loss": 0.2338, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 1.9426362896663956, |
|
"grad_norm": 5.295703887939453, |
|
"learning_rate": 5.044444444444445e-07, |
|
"loss": 0.2508, |
|
"step": 4775 |
|
}, |
|
{ |
|
"epoch": 1.9528071602929211, |
|
"grad_norm": 5.4455342292785645, |
|
"learning_rate": 4.488888888888889e-07, |
|
"loss": 0.2468, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 1.9629780309194467, |
|
"grad_norm": 4.946452617645264, |
|
"learning_rate": 3.9333333333333336e-07, |
|
"loss": 0.2459, |
|
"step": 4825 |
|
}, |
|
{ |
|
"epoch": 1.9731489015459722, |
|
"grad_norm": 4.054887294769287, |
|
"learning_rate": 3.3777777777777777e-07, |
|
"loss": 0.2488, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 1.983319772172498, |
|
"grad_norm": 5.577613830566406, |
|
"learning_rate": 2.822222222222222e-07, |
|
"loss": 0.2435, |
|
"step": 4875 |
|
}, |
|
{ |
|
"epoch": 1.9934906427990235, |
|
"grad_norm": 4.222945213317871, |
|
"learning_rate": 2.266666666666667e-07, |
|
"loss": 0.2399, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 2.0036615134255493, |
|
"grad_norm": 4.715478897094727, |
|
"learning_rate": 1.7111111111111114e-07, |
|
"loss": 0.2385, |
|
"step": 4925 |
|
}, |
|
{ |
|
"epoch": 2.013832384052075, |
|
"grad_norm": 4.2399678230285645, |
|
"learning_rate": 1.1555555555555556e-07, |
|
"loss": 0.181, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 2.0240032546786004, |
|
"grad_norm": 4.25973653793335, |
|
"learning_rate": 6.000000000000001e-08, |
|
"loss": 0.1792, |
|
"step": 4975 |
|
}, |
|
{ |
|
"epoch": 2.034174125305126, |
|
"grad_norm": 4.661466598510742, |
|
"learning_rate": 4.444444444444445e-09, |
|
"loss": 0.1846, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 2.034174125305126, |
|
"eval_loss": 0.36655664443969727, |
|
"eval_runtime": 1783.5562, |
|
"eval_samples_per_second": 2.187, |
|
"eval_steps_per_second": 0.137, |
|
"eval_wer": 0.27733507195981844, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 2.034174125305126, |
|
"step": 5000, |
|
"total_flos": 5.435045990970163e+20, |
|
"train_loss": 0.38478662071228026, |
|
"train_runtime": 58003.8059, |
|
"train_samples_per_second": 2.758, |
|
"train_steps_per_second": 0.086 |
|
} |
|
], |
|
"logging_steps": 25, |
|
"max_steps": 5000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 1000, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 5.435045990970163e+20, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|