|
{ |
|
"best_metric": 1.0734283924102783, |
|
"best_model_checkpoint": "ai-light-dance_stepmania_ft_wav2vec2-large-xlsr-53-v3/checkpoint-752", |
|
"epoch": 10.0, |
|
"global_step": 1880, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 4e-07, |
|
"loss": 0.9557, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 8e-07, |
|
"loss": 0.9283, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.2e-06, |
|
"loss": 0.9279, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.6e-06, |
|
"loss": 0.9274, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 2e-06, |
|
"loss": 0.9421, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 2.4e-06, |
|
"loss": 0.9436, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 2.8e-06, |
|
"loss": 0.9287, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 3.2e-06, |
|
"loss": 0.9219, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 3.6e-06, |
|
"loss": 0.9221, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4e-06, |
|
"loss": 0.9576, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.97752808988764e-06, |
|
"loss": 0.9228, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.955056179775281e-06, |
|
"loss": 0.9266, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.932584269662921e-06, |
|
"loss": 0.9177, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.910112359550562e-06, |
|
"loss": 0.9282, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.887640449438202e-06, |
|
"loss": 0.9431, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.865168539325843e-06, |
|
"loss": 0.9176, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 3.842696629213483e-06, |
|
"loss": 0.9226, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 3.820224719101124e-06, |
|
"loss": 0.9189, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 1.0770223140716553, |
|
"eval_runtime": 406.6878, |
|
"eval_samples_per_second": 6.575, |
|
"eval_steps_per_second": 0.824, |
|
"eval_wer": 0.7002255465619452, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 3.797752808988764e-06, |
|
"loss": 0.9312, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 3.7752808988764043e-06, |
|
"loss": 0.9402, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 3.7528089887640448e-06, |
|
"loss": 0.9272, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 3.730337078651685e-06, |
|
"loss": 0.9265, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 3.7078651685393257e-06, |
|
"loss": 0.922, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 3.6853932584269662e-06, |
|
"loss": 0.9516, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 3.6629213483146067e-06, |
|
"loss": 0.9271, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 3.6404494382022472e-06, |
|
"loss": 0.91, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 3.6179775280898873e-06, |
|
"loss": 0.9152, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 3.5955056179775278e-06, |
|
"loss": 0.9169, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 3.5730337078651683e-06, |
|
"loss": 0.9475, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 3.5505617977528087e-06, |
|
"loss": 0.9178, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 3.5280898876404497e-06, |
|
"loss": 0.9237, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 3.5056179775280897e-06, |
|
"loss": 0.9286, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 3.48314606741573e-06, |
|
"loss": 0.9439, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 3.4606741573033707e-06, |
|
"loss": 0.9344, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 3.438202247191011e-06, |
|
"loss": 0.9231, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 3.4157303370786513e-06, |
|
"loss": 0.9249, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 3.3932584269662917e-06, |
|
"loss": 0.9172, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 1.0780019760131836, |
|
"eval_runtime": 403.0307, |
|
"eval_samples_per_second": 6.635, |
|
"eval_steps_per_second": 0.831, |
|
"eval_wer": 0.695516850070395, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 3.3707865168539322e-06, |
|
"loss": 0.9465, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 3.348314606741573e-06, |
|
"loss": 0.9361, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 3.3258426966292136e-06, |
|
"loss": 0.9194, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 3.3033707865168537e-06, |
|
"loss": 0.9194, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 3.280898876404494e-06, |
|
"loss": 0.9167, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 3.2584269662921347e-06, |
|
"loss": 0.9519, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 3.235955056179775e-06, |
|
"loss": 0.9196, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 3.2134831460674156e-06, |
|
"loss": 0.9202, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 3.1910112359550557e-06, |
|
"loss": 0.915, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 3.1685393258426966e-06, |
|
"loss": 0.9288, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 3.146067415730337e-06, |
|
"loss": 0.938, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 3.1235955056179776e-06, |
|
"loss": 0.91, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 3.1011235955056177e-06, |
|
"loss": 0.9186, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 3.078651685393258e-06, |
|
"loss": 0.9166, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 3.0561797752808986e-06, |
|
"loss": 0.9487, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 3.033707865168539e-06, |
|
"loss": 0.9291, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 3.0112359550561796e-06, |
|
"loss": 0.9237, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 2.98876404494382e-06, |
|
"loss": 0.9119, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 2.9662921348314606e-06, |
|
"loss": 0.9177, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 1.082428216934204, |
|
"eval_runtime": 407.3898, |
|
"eval_samples_per_second": 6.564, |
|
"eval_steps_per_second": 0.822, |
|
"eval_wer": 0.6916180493753914, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 2.943820224719101e-06, |
|
"loss": 0.9457, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 2.9213483146067416e-06, |
|
"loss": 0.9231, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 2.898876404494382e-06, |
|
"loss": 0.9216, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 2.876404494382022e-06, |
|
"loss": 0.9139, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 2.8539325842696626e-06, |
|
"loss": 0.9202, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 2.831460674157303e-06, |
|
"loss": 0.9445, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 2.808988764044944e-06, |
|
"loss": 0.9131, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 2.786516853932584e-06, |
|
"loss": 0.9159, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 2.7640449438202246e-06, |
|
"loss": 0.9138, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"learning_rate": 2.741573033707865e-06, |
|
"loss": 0.9378, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"learning_rate": 2.7191011235955055e-06, |
|
"loss": 0.933, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 3.62, |
|
"learning_rate": 2.696629213483146e-06, |
|
"loss": 0.9162, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"learning_rate": 2.674157303370786e-06, |
|
"loss": 0.9162, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 3.72, |
|
"learning_rate": 2.6516853932584266e-06, |
|
"loss": 0.9117, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"learning_rate": 2.6292134831460675e-06, |
|
"loss": 0.949, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"learning_rate": 2.606741573033708e-06, |
|
"loss": 0.916, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 2.5842696629213485e-06, |
|
"loss": 0.9146, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 3.94, |
|
"learning_rate": 2.5617977528089885e-06, |
|
"loss": 0.9144, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"learning_rate": 2.539325842696629e-06, |
|
"loss": 0.9184, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 1.0734283924102783, |
|
"eval_runtime": 412.0378, |
|
"eval_samples_per_second": 6.49, |
|
"eval_steps_per_second": 0.813, |
|
"eval_wer": 0.6927999321947705, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 2.5168539325842695e-06, |
|
"loss": 0.945, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 4.1, |
|
"learning_rate": 2.49438202247191e-06, |
|
"loss": 0.918, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 4.15, |
|
"learning_rate": 2.47191011235955e-06, |
|
"loss": 0.9163, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 4.2, |
|
"learning_rate": 2.449438202247191e-06, |
|
"loss": 0.9101, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 4.26, |
|
"learning_rate": 2.4269662921348315e-06, |
|
"loss": 0.9238, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 4.31, |
|
"learning_rate": 2.404494382022472e-06, |
|
"loss": 0.9434, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 4.36, |
|
"learning_rate": 2.3820224719101125e-06, |
|
"loss": 0.9172, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 4.41, |
|
"learning_rate": 2.3595505617977525e-06, |
|
"loss": 0.9107, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 4.47, |
|
"learning_rate": 2.337078651685393e-06, |
|
"loss": 0.9141, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 4.52, |
|
"learning_rate": 2.3146067415730335e-06, |
|
"loss": 0.938, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 4.57, |
|
"learning_rate": 2.292134831460674e-06, |
|
"loss": 0.9264, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 4.63, |
|
"learning_rate": 2.269662921348315e-06, |
|
"loss": 0.9124, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 4.68, |
|
"learning_rate": 2.247191011235955e-06, |
|
"loss": 0.9117, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 4.73, |
|
"learning_rate": 2.2247191011235954e-06, |
|
"loss": 0.9139, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 4.79, |
|
"learning_rate": 2.202247191011236e-06, |
|
"loss": 0.9502, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 4.84, |
|
"learning_rate": 2.1797752808988764e-06, |
|
"loss": 0.9181, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 4.89, |
|
"learning_rate": 2.1573033707865165e-06, |
|
"loss": 0.9138, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"learning_rate": 2.134831460674157e-06, |
|
"loss": 0.9056, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 2.1123595505617975e-06, |
|
"loss": 0.9072, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 1.0841398239135742, |
|
"eval_runtime": 407.688, |
|
"eval_samples_per_second": 6.559, |
|
"eval_steps_per_second": 0.822, |
|
"eval_wer": 0.6896592316349065, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 5.05, |
|
"learning_rate": 2.0898876404494384e-06, |
|
"loss": 0.9556, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 5.11, |
|
"learning_rate": 2.067415730337079e-06, |
|
"loss": 0.9087, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 5.16, |
|
"learning_rate": 2.044943820224719e-06, |
|
"loss": 0.9166, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 5.21, |
|
"learning_rate": 2.0224719101123594e-06, |
|
"loss": 0.9154, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 5.27, |
|
"learning_rate": 2e-06, |
|
"loss": 0.9276, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 5.32, |
|
"learning_rate": 1.9775280898876404e-06, |
|
"loss": 0.9279, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 5.37, |
|
"learning_rate": 1.955056179775281e-06, |
|
"loss": 0.9055, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 5.43, |
|
"learning_rate": 1.9325842696629214e-06, |
|
"loss": 0.9113, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 5.48, |
|
"learning_rate": 1.910112359550562e-06, |
|
"loss": 0.9022, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 5.53, |
|
"learning_rate": 1.8876404494382021e-06, |
|
"loss": 0.941, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 5.59, |
|
"learning_rate": 1.8651685393258424e-06, |
|
"loss": 0.9111, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 5.64, |
|
"learning_rate": 1.8426966292134831e-06, |
|
"loss": 0.9163, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 5.69, |
|
"learning_rate": 1.8202247191011236e-06, |
|
"loss": 0.9057, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 5.74, |
|
"learning_rate": 1.7977528089887639e-06, |
|
"loss": 0.9226, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 5.8, |
|
"learning_rate": 1.7752808988764044e-06, |
|
"loss": 0.9304, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 5.85, |
|
"learning_rate": 1.7528089887640449e-06, |
|
"loss": 0.9174, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 5.9, |
|
"learning_rate": 1.7303370786516853e-06, |
|
"loss": 0.9082, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 5.96, |
|
"learning_rate": 1.7078651685393256e-06, |
|
"loss": 0.9089, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 1.0787887573242188, |
|
"eval_runtime": 410.2973, |
|
"eval_samples_per_second": 6.517, |
|
"eval_steps_per_second": 0.816, |
|
"eval_wer": 0.6869799833312145, |
|
"step": 1128 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 1.6853932584269661e-06, |
|
"loss": 0.9236, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 6.06, |
|
"learning_rate": 1.6629213483146068e-06, |
|
"loss": 0.9356, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 6.12, |
|
"learning_rate": 1.640449438202247e-06, |
|
"loss": 0.9073, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 6.17, |
|
"learning_rate": 1.6179775280898876e-06, |
|
"loss": 0.9165, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 6.22, |
|
"learning_rate": 1.5955056179775279e-06, |
|
"loss": 0.9087, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 6.28, |
|
"learning_rate": 1.5730337078651686e-06, |
|
"loss": 0.9378, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 6.33, |
|
"learning_rate": 1.5505617977528088e-06, |
|
"loss": 0.9121, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 6.38, |
|
"learning_rate": 1.5280898876404493e-06, |
|
"loss": 0.9097, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 6.44, |
|
"learning_rate": 1.5056179775280898e-06, |
|
"loss": 0.9109, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 6.49, |
|
"learning_rate": 1.4831460674157303e-06, |
|
"loss": 0.9015, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 6.54, |
|
"learning_rate": 1.4606741573033708e-06, |
|
"loss": 0.9426, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 6.6, |
|
"learning_rate": 1.438202247191011e-06, |
|
"loss": 0.9049, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 6.65, |
|
"learning_rate": 1.4157303370786516e-06, |
|
"loss": 0.9145, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 6.7, |
|
"learning_rate": 1.393258426966292e-06, |
|
"loss": 0.9066, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 6.76, |
|
"learning_rate": 1.3707865168539325e-06, |
|
"loss": 0.9222, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 6.81, |
|
"learning_rate": 1.348314606741573e-06, |
|
"loss": 0.9158, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 6.86, |
|
"learning_rate": 1.3258426966292133e-06, |
|
"loss": 0.9054, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 6.91, |
|
"learning_rate": 1.303370786516854e-06, |
|
"loss": 0.9136, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 6.97, |
|
"learning_rate": 1.28314606741573e-06, |
|
"loss": 0.9174, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_loss": 1.0761058330535889, |
|
"eval_runtime": 408.984, |
|
"eval_samples_per_second": 6.538, |
|
"eval_steps_per_second": 0.819, |
|
"eval_wer": 0.6855862091697156, |
|
"step": 1316 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 1.2606741573033708e-06, |
|
"loss": 0.9324, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 7.07, |
|
"learning_rate": 1.2382022471910113e-06, |
|
"loss": 0.9281, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 7.13, |
|
"learning_rate": 1.2157303370786516e-06, |
|
"loss": 0.9043, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 7.18, |
|
"learning_rate": 1.193258426966292e-06, |
|
"loss": 0.9073, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 7.23, |
|
"learning_rate": 1.1707865168539325e-06, |
|
"loss": 0.9089, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 7.29, |
|
"learning_rate": 1.148314606741573e-06, |
|
"loss": 0.9397, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 7.34, |
|
"learning_rate": 1.1258426966292133e-06, |
|
"loss": 0.9072, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 7.39, |
|
"learning_rate": 1.1033707865168538e-06, |
|
"loss": 0.9105, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 7.45, |
|
"learning_rate": 1.0808988764044945e-06, |
|
"loss": 0.9114, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"learning_rate": 1.0584269662921348e-06, |
|
"loss": 0.9192, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 7.55, |
|
"learning_rate": 1.0359550561797753e-06, |
|
"loss": 0.9257, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 7.61, |
|
"learning_rate": 1.0134831460674155e-06, |
|
"loss": 0.9171, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 7.66, |
|
"learning_rate": 9.91011235955056e-07, |
|
"loss": 0.9076, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 7.71, |
|
"learning_rate": 9.685393258426965e-07, |
|
"loss": 0.9096, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 7.77, |
|
"learning_rate": 9.46067415730337e-07, |
|
"loss": 0.9349, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 7.82, |
|
"learning_rate": 9.235955056179775e-07, |
|
"loss": 0.9065, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 7.87, |
|
"learning_rate": 9.011235955056179e-07, |
|
"loss": 0.9021, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 7.93, |
|
"learning_rate": 8.786516853932584e-07, |
|
"loss": 0.9036, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 7.98, |
|
"learning_rate": 8.561797752808988e-07, |
|
"loss": 0.9072, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 1.0776124000549316, |
|
"eval_runtime": 410.1618, |
|
"eval_samples_per_second": 6.519, |
|
"eval_steps_per_second": 0.817, |
|
"eval_wer": 0.6849505351433562, |
|
"step": 1504 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 8.337078651685393e-07, |
|
"loss": 0.9369, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 8.09, |
|
"learning_rate": 8.112359550561797e-07, |
|
"loss": 0.9093, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 8.14, |
|
"learning_rate": 7.887640449438202e-07, |
|
"loss": 0.9067, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 8.19, |
|
"learning_rate": 7.662921348314606e-07, |
|
"loss": 0.914, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 8.24, |
|
"learning_rate": 7.438202247191011e-07, |
|
"loss": 0.9081, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 8.3, |
|
"learning_rate": 7.213483146067415e-07, |
|
"loss": 0.9415, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 8.35, |
|
"learning_rate": 6.988764044943821e-07, |
|
"loss": 0.9051, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 8.4, |
|
"learning_rate": 6.764044943820224e-07, |
|
"loss": 0.9077, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 8.46, |
|
"learning_rate": 6.539325842696629e-07, |
|
"loss": 0.8991, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 8.51, |
|
"learning_rate": 6.314606741573033e-07, |
|
"loss": 0.9333, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 8.56, |
|
"learning_rate": 6.089887640449438e-07, |
|
"loss": 0.9244, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 8.62, |
|
"learning_rate": 5.865168539325842e-07, |
|
"loss": 0.9063, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 8.67, |
|
"learning_rate": 5.640449438202247e-07, |
|
"loss": 0.9131, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 8.72, |
|
"learning_rate": 5.415730337078652e-07, |
|
"loss": 0.8977, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 8.78, |
|
"learning_rate": 5.191011235955056e-07, |
|
"loss": 0.9377, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 8.83, |
|
"learning_rate": 4.96629213483146e-07, |
|
"loss": 0.9043, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 8.88, |
|
"learning_rate": 4.741573033707865e-07, |
|
"loss": 0.9058, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 8.94, |
|
"learning_rate": 4.5168539325842695e-07, |
|
"loss": 0.907, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 8.99, |
|
"learning_rate": 4.2921348314606744e-07, |
|
"loss": 0.9079, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_loss": 1.079460620880127, |
|
"eval_runtime": 405.6922, |
|
"eval_samples_per_second": 6.591, |
|
"eval_steps_per_second": 0.826, |
|
"eval_wer": 0.6851671351819676, |
|
"step": 1692 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 4.067415730337079e-07, |
|
"loss": 0.9416, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 9.1, |
|
"learning_rate": 3.8426966292134826e-07, |
|
"loss": 0.9146, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 9.15, |
|
"learning_rate": 3.617977528089887e-07, |
|
"loss": 0.9127, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 9.2, |
|
"learning_rate": 3.393258426966292e-07, |
|
"loss": 0.9077, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 9.26, |
|
"learning_rate": 3.168539325842696e-07, |
|
"loss": 0.9105, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 9.31, |
|
"learning_rate": 2.9438202247191006e-07, |
|
"loss": 0.9242, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 9.36, |
|
"learning_rate": 2.7191011235955054e-07, |
|
"loss": 0.9059, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 9.41, |
|
"learning_rate": 2.49438202247191e-07, |
|
"loss": 0.9075, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 9.47, |
|
"learning_rate": 2.2696629213483147e-07, |
|
"loss": 0.905, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 9.52, |
|
"learning_rate": 2.044943820224719e-07, |
|
"loss": 0.9415, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 9.57, |
|
"learning_rate": 1.8202247191011237e-07, |
|
"loss": 0.9149, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 9.63, |
|
"learning_rate": 1.595505617977528e-07, |
|
"loss": 0.8975, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 9.68, |
|
"learning_rate": 1.3707865168539326e-07, |
|
"loss": 0.9074, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 9.73, |
|
"learning_rate": 1.146067415730337e-07, |
|
"loss": 0.9077, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 9.79, |
|
"learning_rate": 9.213483146067415e-08, |
|
"loss": 0.9307, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 9.84, |
|
"learning_rate": 6.96629213483146e-08, |
|
"loss": 0.9027, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 9.89, |
|
"learning_rate": 4.719101123595506e-08, |
|
"loss": 0.9091, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 9.95, |
|
"learning_rate": 2.4719101123595504e-08, |
|
"loss": 0.9042, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 2.247191011235955e-09, |
|
"loss": 0.9016, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 1.0816693305969238, |
|
"eval_runtime": 405.3322, |
|
"eval_samples_per_second": 6.597, |
|
"eval_steps_per_second": 0.826, |
|
"eval_wer": 0.6850023308047634, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"step": 1880, |
|
"total_flos": 7.290614938338836e+19, |
|
"train_loss": 0.9204230947697416, |
|
"train_runtime": 82261.6684, |
|
"train_samples_per_second": 2.925, |
|
"train_steps_per_second": 0.023 |
|
} |
|
], |
|
"max_steps": 1880, |
|
"num_train_epochs": 10, |
|
"total_flos": 7.290614938338836e+19, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|