|
{ |
|
"best_metric": 1.0393410921096802, |
|
"best_model_checkpoint": "checkpoints/instrucode/text_models/llama-2-7b/checkpoint-6000", |
|
"epoch": 1.9198464122870171, |
|
"eval_steps": 200, |
|
"global_step": 6000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 9e-06, |
|
"loss": 2.0232, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.9e-05, |
|
"loss": 1.8837, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.8000000000000003e-05, |
|
"loss": 1.9127, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 3.8e-05, |
|
"loss": 1.8061, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.8e-05, |
|
"loss": 1.5632, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 5.8e-05, |
|
"loss": 1.3373, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 6.800000000000001e-05, |
|
"loss": 1.2455, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 7.800000000000001e-05, |
|
"loss": 1.2518, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 8.800000000000001e-05, |
|
"loss": 1.2229, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.8e-05, |
|
"loss": 1.1142, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.994847020933978e-05, |
|
"loss": 1.1445, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.989049919484702e-05, |
|
"loss": 1.0731, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.982608695652175e-05, |
|
"loss": 1.1097, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.976167471819646e-05, |
|
"loss": 1.1207, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.969726247987118e-05, |
|
"loss": 1.0919, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.96328502415459e-05, |
|
"loss": 1.0958, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.956843800322063e-05, |
|
"loss": 1.1041, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.950402576489533e-05, |
|
"loss": 1.1192, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.943961352657005e-05, |
|
"loss": 1.0308, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.937520128824477e-05, |
|
"loss": 1.0693, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"eval_loss": 1.0856528282165527, |
|
"eval_runtime": 118.5535, |
|
"eval_samples_per_second": 16.87, |
|
"eval_steps_per_second": 2.109, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.931078904991949e-05, |
|
"loss": 1.0917, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.924637681159421e-05, |
|
"loss": 1.0982, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.918196457326893e-05, |
|
"loss": 1.0737, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.911755233494364e-05, |
|
"loss": 1.0501, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.905314009661836e-05, |
|
"loss": 1.0305, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.898872785829308e-05, |
|
"loss": 1.0801, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.89243156199678e-05, |
|
"loss": 1.0732, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.885990338164252e-05, |
|
"loss": 1.0894, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.879549114331724e-05, |
|
"loss": 1.0716, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.873107890499196e-05, |
|
"loss": 1.0448, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.866666666666668e-05, |
|
"loss": 1.0576, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.86022544283414e-05, |
|
"loss": 1.086, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.85378421900161e-05, |
|
"loss": 1.0878, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.847342995169082e-05, |
|
"loss": 1.1065, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.840901771336555e-05, |
|
"loss": 1.0576, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.834460547504027e-05, |
|
"loss": 1.0438, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.828019323671497e-05, |
|
"loss": 1.0566, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.821578099838969e-05, |
|
"loss": 1.026, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.815136876006442e-05, |
|
"loss": 1.047, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.808695652173914e-05, |
|
"loss": 1.0785, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"eval_loss": 1.0718845129013062, |
|
"eval_runtime": 118.4414, |
|
"eval_samples_per_second": 16.886, |
|
"eval_steps_per_second": 2.111, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.802254428341385e-05, |
|
"loss": 1.0367, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.795813204508857e-05, |
|
"loss": 1.0285, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.789371980676329e-05, |
|
"loss": 1.1112, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.7829307568438e-05, |
|
"loss": 1.0778, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.776489533011272e-05, |
|
"loss": 1.0484, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.770048309178744e-05, |
|
"loss": 1.0761, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.763607085346216e-05, |
|
"loss": 1.0681, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.757165861513688e-05, |
|
"loss": 1.0956, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.75072463768116e-05, |
|
"loss": 1.0765, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.744283413848632e-05, |
|
"loss": 1.0752, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.737842190016104e-05, |
|
"loss": 1.0817, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.731400966183575e-05, |
|
"loss": 1.0408, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.724959742351047e-05, |
|
"loss": 1.0893, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.718518518518519e-05, |
|
"loss": 1.0742, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.712077294685991e-05, |
|
"loss": 1.1052, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.705636070853462e-05, |
|
"loss": 1.017, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.699194847020935e-05, |
|
"loss": 1.0808, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.692753623188407e-05, |
|
"loss": 1.048, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.686312399355879e-05, |
|
"loss": 1.0592, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.679871175523349e-05, |
|
"loss": 1.0441, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"eval_loss": 1.0663692951202393, |
|
"eval_runtime": 118.6583, |
|
"eval_samples_per_second": 16.855, |
|
"eval_steps_per_second": 2.107, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.673429951690822e-05, |
|
"loss": 1.0924, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.666988727858294e-05, |
|
"loss": 1.088, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.660547504025766e-05, |
|
"loss": 1.0763, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.654106280193237e-05, |
|
"loss": 1.042, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.647665056360708e-05, |
|
"loss": 1.0745, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.641223832528182e-05, |
|
"loss": 1.0777, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.634782608695652e-05, |
|
"loss": 1.0841, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.628341384863124e-05, |
|
"loss": 1.0393, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.621900161030596e-05, |
|
"loss": 1.0496, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.615458937198069e-05, |
|
"loss": 1.0676, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.60901771336554e-05, |
|
"loss": 1.0929, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.602576489533011e-05, |
|
"loss": 1.08, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.596135265700483e-05, |
|
"loss": 1.0723, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.589694041867955e-05, |
|
"loss": 1.0565, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.583252818035427e-05, |
|
"loss": 1.0756, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.576811594202899e-05, |
|
"loss": 1.0197, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.570370370370371e-05, |
|
"loss": 0.9768, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.563929146537843e-05, |
|
"loss": 1.0742, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.557487922705315e-05, |
|
"loss": 1.0732, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.551046698872786e-05, |
|
"loss": 1.0664, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"eval_loss": 1.062617301940918, |
|
"eval_runtime": 118.4976, |
|
"eval_samples_per_second": 16.878, |
|
"eval_steps_per_second": 2.11, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.544605475040258e-05, |
|
"loss": 1.0821, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.53816425120773e-05, |
|
"loss": 1.0551, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.531723027375202e-05, |
|
"loss": 1.0281, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.525281803542674e-05, |
|
"loss": 1.0604, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.518840579710146e-05, |
|
"loss": 1.0327, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.512399355877616e-05, |
|
"loss": 1.0879, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.505958132045088e-05, |
|
"loss": 1.0594, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.499516908212561e-05, |
|
"loss": 1.0701, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.493075684380033e-05, |
|
"loss": 1.0606, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.486634460547504e-05, |
|
"loss": 1.079, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.480193236714976e-05, |
|
"loss": 1.0503, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.473752012882449e-05, |
|
"loss": 1.048, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.467310789049921e-05, |
|
"loss": 1.0758, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.460869565217391e-05, |
|
"loss": 1.0308, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.454428341384863e-05, |
|
"loss": 1.0528, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.447987117552335e-05, |
|
"loss": 1.0377, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.441545893719808e-05, |
|
"loss": 1.0504, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.435104669887279e-05, |
|
"loss": 1.0575, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.42866344605475e-05, |
|
"loss": 1.0842, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.422222222222223e-05, |
|
"loss": 1.099, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"eval_loss": 1.061242699623108, |
|
"eval_runtime": 118.614, |
|
"eval_samples_per_second": 16.861, |
|
"eval_steps_per_second": 2.108, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.415780998389694e-05, |
|
"loss": 1.0378, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.409339774557166e-05, |
|
"loss": 1.0768, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.402898550724638e-05, |
|
"loss": 1.0809, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.39645732689211e-05, |
|
"loss": 1.0501, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.390016103059582e-05, |
|
"loss": 1.0512, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.383574879227054e-05, |
|
"loss": 1.048, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.377133655394526e-05, |
|
"loss": 1.1142, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.370692431561997e-05, |
|
"loss": 1.0862, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.364251207729468e-05, |
|
"loss": 1.0654, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.357809983896941e-05, |
|
"loss": 1.0423, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.351368760064413e-05, |
|
"loss": 1.0398, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.344927536231885e-05, |
|
"loss": 1.0779, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.338486312399355e-05, |
|
"loss": 1.0806, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.332045088566829e-05, |
|
"loss": 1.0668, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.3256038647343e-05, |
|
"loss": 1.0262, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.319162640901772e-05, |
|
"loss": 1.0334, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.312721417069243e-05, |
|
"loss": 1.0793, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.306280193236715e-05, |
|
"loss": 1.0412, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.299838969404188e-05, |
|
"loss": 1.079, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.293397745571659e-05, |
|
"loss": 0.9992, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"eval_loss": 1.0581145286560059, |
|
"eval_runtime": 124.535, |
|
"eval_samples_per_second": 16.06, |
|
"eval_steps_per_second": 2.007, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.28695652173913e-05, |
|
"loss": 1.0414, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.280515297906602e-05, |
|
"loss": 1.0812, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.274074074074076e-05, |
|
"loss": 1.0552, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.267632850241546e-05, |
|
"loss": 1.0574, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.261191626409018e-05, |
|
"loss": 1.0875, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.25475040257649e-05, |
|
"loss": 1.0563, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.248309178743962e-05, |
|
"loss": 1.0867, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.241867954911434e-05, |
|
"loss": 1.0314, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.235426731078905e-05, |
|
"loss": 1.0841, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.228985507246377e-05, |
|
"loss": 1.0493, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.222544283413849e-05, |
|
"loss": 1.1141, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.216103059581321e-05, |
|
"loss": 1.0405, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.209661835748793e-05, |
|
"loss": 1.0577, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.203220611916265e-05, |
|
"loss": 1.0747, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.196779388083737e-05, |
|
"loss": 1.0313, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.190338164251207e-05, |
|
"loss": 1.0848, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.18389694041868e-05, |
|
"loss": 1.0198, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.177455716586152e-05, |
|
"loss": 1.059, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.171014492753624e-05, |
|
"loss": 1.013, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.164573268921095e-05, |
|
"loss": 1.0563, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"eval_loss": 1.055400013923645, |
|
"eval_runtime": 118.7688, |
|
"eval_samples_per_second": 16.839, |
|
"eval_steps_per_second": 2.105, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.158132045088568e-05, |
|
"loss": 1.06, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.15169082125604e-05, |
|
"loss": 1.0426, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.14524959742351e-05, |
|
"loss": 1.0642, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.138808373590982e-05, |
|
"loss": 1.0871, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.132367149758455e-05, |
|
"loss": 1.0328, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.125925925925927e-05, |
|
"loss": 1.0137, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.119484702093398e-05, |
|
"loss": 1.0605, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.11304347826087e-05, |
|
"loss": 1.0404, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.106602254428341e-05, |
|
"loss": 1.0132, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.100161030595815e-05, |
|
"loss": 0.9935, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.093719806763285e-05, |
|
"loss": 1.0954, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.087278582930757e-05, |
|
"loss": 1.0853, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.080837359098229e-05, |
|
"loss": 1.0688, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.074396135265701e-05, |
|
"loss": 1.0509, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.067954911433173e-05, |
|
"loss": 1.0424, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.061513687600645e-05, |
|
"loss": 1.0711, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.055072463768116e-05, |
|
"loss": 1.0537, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.048631239935588e-05, |
|
"loss": 1.1012, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.04219001610306e-05, |
|
"loss": 1.0725, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.035748792270532e-05, |
|
"loss": 1.0444, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"eval_loss": 1.0541496276855469, |
|
"eval_runtime": 118.5768, |
|
"eval_samples_per_second": 16.867, |
|
"eval_steps_per_second": 2.108, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.029307568438004e-05, |
|
"loss": 1.0641, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.022866344605474e-05, |
|
"loss": 1.0323, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.016425120772948e-05, |
|
"loss": 1.0538, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.00998389694042e-05, |
|
"loss": 1.0588, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.003542673107891e-05, |
|
"loss": 1.0026, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 8.997101449275362e-05, |
|
"loss": 1.0614, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 8.990660225442834e-05, |
|
"loss": 1.0737, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 8.984219001610307e-05, |
|
"loss": 1.058, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 8.977777777777779e-05, |
|
"loss": 1.0653, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 8.97133655394525e-05, |
|
"loss": 0.9893, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.964895330112721e-05, |
|
"loss": 1.0182, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.958454106280194e-05, |
|
"loss": 1.0247, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.952012882447666e-05, |
|
"loss": 1.016, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.945571658615137e-05, |
|
"loss": 1.019, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.939130434782609e-05, |
|
"loss": 1.0312, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.932689210950082e-05, |
|
"loss": 1.0646, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.926247987117552e-05, |
|
"loss": 1.0485, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.919806763285024e-05, |
|
"loss": 1.1001, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.913365539452496e-05, |
|
"loss": 1.0379, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.906924315619968e-05, |
|
"loss": 1.0264, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"eval_loss": 1.05194890499115, |
|
"eval_runtime": 119.277, |
|
"eval_samples_per_second": 16.768, |
|
"eval_steps_per_second": 2.096, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.90048309178744e-05, |
|
"loss": 1.0574, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.894041867954912e-05, |
|
"loss": 1.0484, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.887600644122384e-05, |
|
"loss": 1.0297, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.881159420289856e-05, |
|
"loss": 1.0167, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.874718196457327e-05, |
|
"loss": 1.0421, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.868276972624799e-05, |
|
"loss": 1.0276, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.861835748792271e-05, |
|
"loss": 1.0477, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.855394524959743e-05, |
|
"loss": 1.0671, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 8.848953301127214e-05, |
|
"loss": 0.9882, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 8.842512077294687e-05, |
|
"loss": 1.006, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 8.836070853462159e-05, |
|
"loss": 1.018, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 8.82962962962963e-05, |
|
"loss": 1.0346, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 8.823188405797101e-05, |
|
"loss": 1.0879, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 8.816747181964574e-05, |
|
"loss": 1.0687, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 8.810305958132046e-05, |
|
"loss": 1.0331, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 8.803864734299517e-05, |
|
"loss": 1.0115, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 8.797423510466989e-05, |
|
"loss": 1.0119, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 8.79098228663446e-05, |
|
"loss": 1.0575, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 8.784541062801934e-05, |
|
"loss": 1.0744, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 8.778099838969404e-05, |
|
"loss": 1.0341, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"eval_loss": 1.0501959323883057, |
|
"eval_runtime": 120.6722, |
|
"eval_samples_per_second": 16.574, |
|
"eval_steps_per_second": 2.072, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 8.771658615136876e-05, |
|
"loss": 1.0206, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 8.765217391304348e-05, |
|
"loss": 1.0297, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 8.758776167471821e-05, |
|
"loss": 1.0848, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 8.752334943639292e-05, |
|
"loss": 1.0528, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 8.745893719806763e-05, |
|
"loss": 1.0074, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 8.739452495974235e-05, |
|
"loss": 1.0934, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 8.733011272141707e-05, |
|
"loss": 1.034, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 8.726570048309179e-05, |
|
"loss": 1.0866, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 8.720128824476651e-05, |
|
"loss": 1.0517, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 8.713687600644123e-05, |
|
"loss": 1.0339, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 8.707246376811595e-05, |
|
"loss": 1.0164, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 8.700805152979067e-05, |
|
"loss": 1.0597, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 8.694363929146538e-05, |
|
"loss": 1.0539, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 8.68792270531401e-05, |
|
"loss": 0.9942, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 8.681481481481482e-05, |
|
"loss": 1.0192, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 8.675040257648954e-05, |
|
"loss": 1.0685, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 8.668599033816426e-05, |
|
"loss": 1.0198, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 8.662157809983898e-05, |
|
"loss": 1.0623, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 8.655716586151368e-05, |
|
"loss": 1.0441, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 8.64927536231884e-05, |
|
"loss": 1.0155, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"eval_loss": 1.0501713752746582, |
|
"eval_runtime": 122.07, |
|
"eval_samples_per_second": 16.384, |
|
"eval_steps_per_second": 2.048, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 8.642834138486313e-05, |
|
"loss": 1.0347, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 8.636392914653785e-05, |
|
"loss": 1.0301, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 8.629951690821256e-05, |
|
"loss": 1.085, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 8.623510466988728e-05, |
|
"loss": 1.0661, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 8.617069243156201e-05, |
|
"loss": 1.0196, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 8.610628019323673e-05, |
|
"loss": 1.0601, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 8.604186795491143e-05, |
|
"loss": 1.0719, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 8.597745571658615e-05, |
|
"loss": 1.0123, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 8.591304347826087e-05, |
|
"loss": 1.0766, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 8.584863123993559e-05, |
|
"loss": 1.0474, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 8.578421900161031e-05, |
|
"loss": 1.0737, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 8.571980676328503e-05, |
|
"loss": 1.0482, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 8.565539452495974e-05, |
|
"loss": 1.0543, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 8.559098228663446e-05, |
|
"loss": 1.0549, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 8.552657004830918e-05, |
|
"loss": 1.0869, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 8.54621578099839e-05, |
|
"loss": 1.0821, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 8.539774557165862e-05, |
|
"loss": 1.0302, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 8.533333333333334e-05, |
|
"loss": 1.0653, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 8.526892109500806e-05, |
|
"loss": 1.0328, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 8.520450885668278e-05, |
|
"loss": 1.0685, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"eval_loss": 1.0498476028442383, |
|
"eval_runtime": 118.8004, |
|
"eval_samples_per_second": 16.835, |
|
"eval_steps_per_second": 2.104, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 8.51400966183575e-05, |
|
"loss": 1.0243, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 8.50756843800322e-05, |
|
"loss": 1.075, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 8.501127214170693e-05, |
|
"loss": 1.0475, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 8.494685990338165e-05, |
|
"loss": 1.0269, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 8.488244766505637e-05, |
|
"loss": 1.0431, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 8.481803542673107e-05, |
|
"loss": 1.0412, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 8.475362318840581e-05, |
|
"loss": 1.029, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 8.468921095008053e-05, |
|
"loss": 1.0884, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 8.462479871175524e-05, |
|
"loss": 1.0268, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 8.456038647342995e-05, |
|
"loss": 1.0475, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 8.449597423510467e-05, |
|
"loss": 1.0258, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.44315619967794e-05, |
|
"loss": 1.0797, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.43671497584541e-05, |
|
"loss": 1.0914, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.430273752012882e-05, |
|
"loss": 1.0445, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 8.423832528180354e-05, |
|
"loss": 1.016, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 8.417391304347828e-05, |
|
"loss": 1.0463, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 8.410950080515298e-05, |
|
"loss": 0.9827, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 8.40450885668277e-05, |
|
"loss": 1.0417, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 8.398067632850242e-05, |
|
"loss": 1.0501, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 8.391626409017714e-05, |
|
"loss": 1.0037, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"eval_loss": 1.0493232011795044, |
|
"eval_runtime": 118.6245, |
|
"eval_samples_per_second": 16.86, |
|
"eval_steps_per_second": 2.107, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 8.385185185185186e-05, |
|
"loss": 1.0176, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 8.378743961352657e-05, |
|
"loss": 1.0409, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 8.372302737520129e-05, |
|
"loss": 1.0449, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 8.365861513687601e-05, |
|
"loss": 1.0731, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 8.359420289855073e-05, |
|
"loss": 1.1184, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 8.352979066022545e-05, |
|
"loss": 1.0286, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 8.346537842190017e-05, |
|
"loss": 1.0409, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 8.340096618357489e-05, |
|
"loss": 1.0627, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 8.33365539452496e-05, |
|
"loss": 1.0625, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 8.327214170692432e-05, |
|
"loss": 1.0324, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 8.320772946859904e-05, |
|
"loss": 1.0755, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 8.314331723027375e-05, |
|
"loss": 1.0663, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 8.307890499194847e-05, |
|
"loss": 1.0512, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 8.30144927536232e-05, |
|
"loss": 1.0399, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 8.295008051529792e-05, |
|
"loss": 1.0016, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 8.288566827697262e-05, |
|
"loss": 1.0499, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 8.282125603864734e-05, |
|
"loss": 1.0846, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 8.275684380032207e-05, |
|
"loss": 1.0197, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 8.269243156199679e-05, |
|
"loss": 1.0626, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 8.26280193236715e-05, |
|
"loss": 1.0752, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"eval_loss": 1.0474530458450317, |
|
"eval_runtime": 118.8942, |
|
"eval_samples_per_second": 16.822, |
|
"eval_steps_per_second": 2.103, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 8.256360708534622e-05, |
|
"loss": 1.0803, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 8.249919484702093e-05, |
|
"loss": 0.9713, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 8.243478260869565e-05, |
|
"loss": 1.0177, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 8.237037037037037e-05, |
|
"loss": 1.0667, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 8.230595813204509e-05, |
|
"loss": 1.0138, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 8.224154589371981e-05, |
|
"loss": 1.0347, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 8.217713365539453e-05, |
|
"loss": 1.0308, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 8.211272141706925e-05, |
|
"loss": 1.0261, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 8.204830917874397e-05, |
|
"loss": 1.0751, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 8.198389694041868e-05, |
|
"loss": 0.9864, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 8.19194847020934e-05, |
|
"loss": 1.0282, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 8.185507246376812e-05, |
|
"loss": 1.0381, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 8.179066022544284e-05, |
|
"loss": 1.0368, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 8.172624798711756e-05, |
|
"loss": 1.0216, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 8.166183574879226e-05, |
|
"loss": 1.0745, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 8.1597423510467e-05, |
|
"loss": 1.0904, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 8.153301127214171e-05, |
|
"loss": 1.0684, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 8.146859903381643e-05, |
|
"loss": 1.04, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 8.140418679549114e-05, |
|
"loss": 0.9984, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 8.133977455716587e-05, |
|
"loss": 1.0787, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"eval_loss": 1.046216607093811, |
|
"eval_runtime": 118.7583, |
|
"eval_samples_per_second": 16.841, |
|
"eval_steps_per_second": 2.105, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 8.127536231884059e-05, |
|
"loss": 1.0192, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 8.121095008051531e-05, |
|
"loss": 1.0692, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 8.114653784219001e-05, |
|
"loss": 1.035, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 8.108212560386473e-05, |
|
"loss": 1.0113, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 8.101771336553946e-05, |
|
"loss": 1.0744, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 8.095330112721417e-05, |
|
"loss": 1.0348, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 8.088888888888889e-05, |
|
"loss": 1.059, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 8.082447665056361e-05, |
|
"loss": 1.0587, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 8.076006441223834e-05, |
|
"loss": 1.0221, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 8.069565217391304e-05, |
|
"loss": 1.0645, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 8.063123993558776e-05, |
|
"loss": 1.1149, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 8.056682769726248e-05, |
|
"loss": 1.0484, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 8.05024154589372e-05, |
|
"loss": 1.0725, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 8.043800322061192e-05, |
|
"loss": 1.0554, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 8.037359098228664e-05, |
|
"loss": 1.0015, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 8.030917874396136e-05, |
|
"loss": 1.0234, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 8.024476650563608e-05, |
|
"loss": 1.0617, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 8.01803542673108e-05, |
|
"loss": 1.0285, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 8.011594202898551e-05, |
|
"loss": 1.0525, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 8.005152979066023e-05, |
|
"loss": 0.9877, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"eval_loss": 1.0453153848648071, |
|
"eval_runtime": 118.7099, |
|
"eval_samples_per_second": 16.848, |
|
"eval_steps_per_second": 2.106, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 7.998711755233495e-05, |
|
"loss": 1.0163, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 7.992270531400966e-05, |
|
"loss": 1.0191, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 7.985829307568439e-05, |
|
"loss": 1.0528, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 7.97938808373591e-05, |
|
"loss": 1.0427, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 7.972946859903383e-05, |
|
"loss": 1.0345, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 7.966505636070853e-05, |
|
"loss": 0.9728, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 7.960064412238326e-05, |
|
"loss": 1.0709, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 7.953623188405798e-05, |
|
"loss": 1.0216, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 7.947181964573269e-05, |
|
"loss": 1.0422, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 7.94074074074074e-05, |
|
"loss": 1.0559, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 7.934299516908214e-05, |
|
"loss": 1.033, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 7.927858293075686e-05, |
|
"loss": 1.0307, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 7.921417069243156e-05, |
|
"loss": 1.0239, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 7.914975845410628e-05, |
|
"loss": 1.022, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 7.9085346215781e-05, |
|
"loss": 1.0435, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 7.902093397745573e-05, |
|
"loss": 0.9938, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 7.895652173913044e-05, |
|
"loss": 1.0168, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 7.889210950080515e-05, |
|
"loss": 1.0455, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 7.882769726247987e-05, |
|
"loss": 0.99, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 7.876328502415459e-05, |
|
"loss": 1.0541, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"eval_loss": 1.0455336570739746, |
|
"eval_runtime": 118.6315, |
|
"eval_samples_per_second": 16.859, |
|
"eval_steps_per_second": 2.107, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 7.869887278582931e-05, |
|
"loss": 0.994, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 7.863446054750403e-05, |
|
"loss": 1.0331, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 7.857004830917875e-05, |
|
"loss": 1.0379, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 7.850563607085347e-05, |
|
"loss": 1.0344, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 7.844122383252819e-05, |
|
"loss": 0.9944, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 7.83768115942029e-05, |
|
"loss": 1.0271, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 7.831239935587762e-05, |
|
"loss": 1.0045, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 7.824798711755233e-05, |
|
"loss": 1.0009, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 7.818357487922706e-05, |
|
"loss": 1.0289, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 7.811916264090178e-05, |
|
"loss": 1.0096, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 7.80547504025765e-05, |
|
"loss": 1.0594, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 7.79903381642512e-05, |
|
"loss": 1.0305, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 7.792592592592592e-05, |
|
"loss": 1.0664, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 7.786151368760065e-05, |
|
"loss": 1.0089, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 7.779710144927537e-05, |
|
"loss": 1.0596, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 7.773268921095008e-05, |
|
"loss": 1.0075, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 7.76682769726248e-05, |
|
"loss": 1.0613, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 7.760386473429953e-05, |
|
"loss": 1.033, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 7.753945249597425e-05, |
|
"loss": 1.0108, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 7.747504025764895e-05, |
|
"loss": 1.0601, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"eval_loss": 1.0447208881378174, |
|
"eval_runtime": 118.6536, |
|
"eval_samples_per_second": 16.856, |
|
"eval_steps_per_second": 2.107, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 7.741062801932367e-05, |
|
"loss": 1.0148, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 7.73462157809984e-05, |
|
"loss": 1.0283, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 7.728180354267311e-05, |
|
"loss": 1.0018, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 7.721739130434783e-05, |
|
"loss": 1.0525, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 7.715297906602255e-05, |
|
"loss": 1.001, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 7.708856682769726e-05, |
|
"loss": 1.0189, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 7.702415458937198e-05, |
|
"loss": 1.036, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 7.69597423510467e-05, |
|
"loss": 0.9648, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 7.689533011272142e-05, |
|
"loss": 0.9938, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 7.683091787439614e-05, |
|
"loss": 0.9936, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 7.676650563607086e-05, |
|
"loss": 1.0395, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 7.670209339774558e-05, |
|
"loss": 1.0354, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 7.66376811594203e-05, |
|
"loss": 0.9744, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 7.657326892109501e-05, |
|
"loss": 1.0472, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 7.650885668276972e-05, |
|
"loss": 0.9919, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 7.644444444444445e-05, |
|
"loss": 1.0487, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 7.638003220611917e-05, |
|
"loss": 0.9969, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 7.631561996779389e-05, |
|
"loss": 0.9966, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 7.62512077294686e-05, |
|
"loss": 1.0435, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 7.618679549114333e-05, |
|
"loss": 0.9847, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"eval_loss": 1.0448102951049805, |
|
"eval_runtime": 118.8166, |
|
"eval_samples_per_second": 16.833, |
|
"eval_steps_per_second": 2.104, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 7.612238325281805e-05, |
|
"loss": 1.0465, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 7.605797101449275e-05, |
|
"loss": 1.0431, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 7.599355877616747e-05, |
|
"loss": 1.0613, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 7.592914653784219e-05, |
|
"loss": 1.0558, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 7.586473429951692e-05, |
|
"loss": 1.0588, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 7.580032206119163e-05, |
|
"loss": 1.0227, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 7.573590982286634e-05, |
|
"loss": 1.0089, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 7.567149758454106e-05, |
|
"loss": 1.0016, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 7.56070853462158e-05, |
|
"loss": 1.0205, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 7.55426731078905e-05, |
|
"loss": 1.029, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 7.547826086956522e-05, |
|
"loss": 1.0509, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 7.541384863123994e-05, |
|
"loss": 1.0208, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 7.534943639291466e-05, |
|
"loss": 0.982, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 7.528502415458938e-05, |
|
"loss": 1.0413, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 7.52206119162641e-05, |
|
"loss": 1.0252, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 7.515619967793881e-05, |
|
"loss": 1.0495, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 7.509178743961353e-05, |
|
"loss": 1.0006, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 7.502737520128825e-05, |
|
"loss": 0.9674, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 7.496296296296297e-05, |
|
"loss": 0.9765, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 7.489855072463769e-05, |
|
"loss": 0.9765, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"eval_loss": 1.0446877479553223, |
|
"eval_runtime": 119.0017, |
|
"eval_samples_per_second": 16.806, |
|
"eval_steps_per_second": 2.101, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 7.48341384863124e-05, |
|
"loss": 1.0418, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 7.476972624798712e-05, |
|
"loss": 0.9666, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 7.470531400966184e-05, |
|
"loss": 1.0535, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 7.464090177133656e-05, |
|
"loss": 0.9831, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 7.457648953301127e-05, |
|
"loss": 1.0021, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 7.451207729468599e-05, |
|
"loss": 1.0304, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 7.444766505636072e-05, |
|
"loss": 1.0077, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 7.438325281803544e-05, |
|
"loss": 1.0215, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 7.431884057971014e-05, |
|
"loss": 1.0181, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 7.425442834138486e-05, |
|
"loss": 1.047, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 7.419001610305959e-05, |
|
"loss": 1.0137, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 7.412560386473431e-05, |
|
"loss": 0.9946, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 7.406119162640902e-05, |
|
"loss": 1.0567, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 7.399677938808374e-05, |
|
"loss": 1.0348, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 7.393236714975845e-05, |
|
"loss": 1.0314, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 7.386795491143317e-05, |
|
"loss": 1.0446, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 7.380354267310789e-05, |
|
"loss": 0.9648, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 7.373913043478261e-05, |
|
"loss": 1.0185, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 7.367471819645733e-05, |
|
"loss": 1.1056, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 7.361030595813205e-05, |
|
"loss": 1.008, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"eval_loss": 1.0431263446807861, |
|
"eval_runtime": 119.6999, |
|
"eval_samples_per_second": 16.708, |
|
"eval_steps_per_second": 2.089, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 7.354589371980677e-05, |
|
"loss": 1.0346, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 7.348148148148149e-05, |
|
"loss": 1.0152, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 7.34170692431562e-05, |
|
"loss": 1.0362, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 7.335265700483092e-05, |
|
"loss": 1.0163, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 7.328824476650564e-05, |
|
"loss": 1.0121, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 7.322383252818036e-05, |
|
"loss": 1.0123, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 7.315942028985508e-05, |
|
"loss": 1.0301, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 7.309500805152978e-05, |
|
"loss": 1.0268, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 7.303059581320452e-05, |
|
"loss": 0.9811, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 7.296618357487923e-05, |
|
"loss": 1.0304, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 7.290177133655395e-05, |
|
"loss": 1.0348, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 7.283735909822866e-05, |
|
"loss": 1.0158, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 7.277294685990339e-05, |
|
"loss": 0.9808, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 7.270853462157811e-05, |
|
"loss": 1.0387, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 7.264412238325283e-05, |
|
"loss": 1.0669, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 7.258615136876007e-05, |
|
"loss": 1.0417, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 7.252173913043479e-05, |
|
"loss": 1.0617, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 7.24573268921095e-05, |
|
"loss": 1.0563, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 7.239291465378422e-05, |
|
"loss": 1.0697, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 7.232850241545894e-05, |
|
"loss": 1.0392, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"eval_loss": 1.0430742502212524, |
|
"eval_runtime": 119.6968, |
|
"eval_samples_per_second": 16.709, |
|
"eval_steps_per_second": 2.089, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 7.226409017713366e-05, |
|
"loss": 1.0449, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 7.219967793880838e-05, |
|
"loss": 1.0561, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 7.213526570048309e-05, |
|
"loss": 0.9966, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 7.207085346215782e-05, |
|
"loss": 1.0508, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 7.200644122383254e-05, |
|
"loss": 1.0333, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 7.194202898550724e-05, |
|
"loss": 1.022, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 7.187761674718196e-05, |
|
"loss": 0.993, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 7.181320450885669e-05, |
|
"loss": 0.9847, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 7.174879227053141e-05, |
|
"loss": 1.0315, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 7.168438003220612e-05, |
|
"loss": 1.0533, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 7.161996779388084e-05, |
|
"loss": 1.009, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 7.155555555555555e-05, |
|
"loss": 1.0426, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 7.149114331723029e-05, |
|
"loss": 0.9935, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 7.142673107890499e-05, |
|
"loss": 1.0453, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 7.136231884057971e-05, |
|
"loss": 1.0236, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 7.129790660225443e-05, |
|
"loss": 1.0324, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 7.123349436392915e-05, |
|
"loss": 0.9708, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 7.116908212560387e-05, |
|
"loss": 0.9613, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 7.110466988727859e-05, |
|
"loss": 0.9964, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 7.10402576489533e-05, |
|
"loss": 1.01, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"eval_loss": 1.0430107116699219, |
|
"eval_runtime": 118.876, |
|
"eval_samples_per_second": 16.824, |
|
"eval_steps_per_second": 2.103, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 7.097584541062802e-05, |
|
"loss": 1.0325, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 7.091143317230274e-05, |
|
"loss": 1.0201, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 7.084702093397746e-05, |
|
"loss": 1.0044, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 7.078260869565218e-05, |
|
"loss": 1.0172, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 7.07181964573269e-05, |
|
"loss": 0.9837, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 7.065378421900162e-05, |
|
"loss": 0.9713, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 7.058937198067633e-05, |
|
"loss": 1.0316, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 7.052495974235105e-05, |
|
"loss": 1.0329, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 7.046054750402576e-05, |
|
"loss": 1.0332, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 7.039613526570049e-05, |
|
"loss": 1.0352, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 7.033172302737521e-05, |
|
"loss": 1.0147, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 7.026731078904993e-05, |
|
"loss": 1.0397, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 7.020289855072463e-05, |
|
"loss": 0.9909, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 7.013848631239935e-05, |
|
"loss": 1.0525, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 7.007407407407408e-05, |
|
"loss": 1.049, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 7.00096618357488e-05, |
|
"loss": 0.9986, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 6.994524959742351e-05, |
|
"loss": 1.0441, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 6.988083735909823e-05, |
|
"loss": 1.0286, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 6.981642512077296e-05, |
|
"loss": 1.0155, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 6.975201288244766e-05, |
|
"loss": 1.0787, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"eval_loss": 1.0421252250671387, |
|
"eval_runtime": 118.8803, |
|
"eval_samples_per_second": 16.824, |
|
"eval_steps_per_second": 2.103, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 6.968760064412238e-05, |
|
"loss": 1.0397, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 6.96231884057971e-05, |
|
"loss": 1.0473, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 6.955877616747182e-05, |
|
"loss": 1.0341, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 6.949436392914654e-05, |
|
"loss": 1.0107, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 6.942995169082126e-05, |
|
"loss": 1.0231, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 6.936553945249598e-05, |
|
"loss": 1.0472, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 6.93011272141707e-05, |
|
"loss": 1.0225, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 6.923671497584541e-05, |
|
"loss": 1.0247, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 6.917230273752013e-05, |
|
"loss": 1.0302, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 6.910789049919485e-05, |
|
"loss": 1.007, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 6.904347826086957e-05, |
|
"loss": 0.9973, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 6.897906602254429e-05, |
|
"loss": 1.0316, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 6.891465378421901e-05, |
|
"loss": 1.01, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 6.885024154589373e-05, |
|
"loss": 1.0023, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 6.878582930756844e-05, |
|
"loss": 0.9919, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 6.872141706924315e-05, |
|
"loss": 1.0537, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 6.865700483091788e-05, |
|
"loss": 0.9987, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 6.85925925925926e-05, |
|
"loss": 0.991, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 6.85281803542673e-05, |
|
"loss": 1.043, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 6.846376811594202e-05, |
|
"loss": 1.0538, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"eval_loss": 1.0414823293685913, |
|
"eval_runtime": 118.8401, |
|
"eval_samples_per_second": 16.829, |
|
"eval_steps_per_second": 2.104, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 6.839935587761676e-05, |
|
"loss": 1.0458, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 6.833494363929148e-05, |
|
"loss": 1.0063, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 6.827053140096618e-05, |
|
"loss": 1.011, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 6.82061191626409e-05, |
|
"loss": 1.05, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 6.814170692431562e-05, |
|
"loss": 1.048, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 6.807729468599035e-05, |
|
"loss": 0.9938, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 6.801288244766506e-05, |
|
"loss": 1.0101, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 6.794847020933977e-05, |
|
"loss": 1.0036, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 6.788405797101449e-05, |
|
"loss": 1.0539, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 6.781964573268923e-05, |
|
"loss": 1.0127, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 6.775523349436393e-05, |
|
"loss": 1.0426, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 6.769082125603865e-05, |
|
"loss": 1.0214, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 6.762640901771337e-05, |
|
"loss": 1.0157, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 6.756199677938809e-05, |
|
"loss": 1.0077, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 6.74975845410628e-05, |
|
"loss": 1.0412, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 6.743317230273752e-05, |
|
"loss": 1.0017, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 6.736876006441224e-05, |
|
"loss": 0.9827, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 6.730434782608696e-05, |
|
"loss": 1.0302, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 6.723993558776168e-05, |
|
"loss": 1.08, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 6.71755233494364e-05, |
|
"loss": 1.0497, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"eval_loss": 1.0406599044799805, |
|
"eval_runtime": 118.7412, |
|
"eval_samples_per_second": 16.843, |
|
"eval_steps_per_second": 2.105, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 6.711111111111112e-05, |
|
"loss": 1.0289, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 6.704669887278582e-05, |
|
"loss": 0.9878, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 6.698228663446055e-05, |
|
"loss": 1.0198, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 6.691787439613527e-05, |
|
"loss": 1.0086, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 6.685346215780999e-05, |
|
"loss": 1.0732, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 6.67890499194847e-05, |
|
"loss": 1.0408, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 6.672463768115942e-05, |
|
"loss": 1.0185, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 6.666022544283415e-05, |
|
"loss": 1.0486, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 6.659581320450887e-05, |
|
"loss": 1.0021, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 6.653140096618357e-05, |
|
"loss": 1.0195, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 6.646698872785829e-05, |
|
"loss": 1.0199, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 6.640257648953302e-05, |
|
"loss": 0.995, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 6.633816425120773e-05, |
|
"loss": 0.9822, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 6.627375201288245e-05, |
|
"loss": 1.0525, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 6.620933977455717e-05, |
|
"loss": 1.0204, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 6.614492753623188e-05, |
|
"loss": 1.0089, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 6.60805152979066e-05, |
|
"loss": 1.0226, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 6.601610305958132e-05, |
|
"loss": 1.0142, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 6.595169082125604e-05, |
|
"loss": 0.9809, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 6.588727858293076e-05, |
|
"loss": 1.0352, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"eval_loss": 1.0410431623458862, |
|
"eval_runtime": 118.8272, |
|
"eval_samples_per_second": 16.831, |
|
"eval_steps_per_second": 2.104, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 6.582286634460548e-05, |
|
"loss": 1.0318, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 6.57584541062802e-05, |
|
"loss": 1.0316, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 6.569404186795492e-05, |
|
"loss": 1.0124, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 6.562962962962963e-05, |
|
"loss": 1.0544, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 6.556521739130434e-05, |
|
"loss": 1.0078, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 6.550080515297907e-05, |
|
"loss": 1.0065, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 6.543639291465379e-05, |
|
"loss": 1.0265, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 6.537198067632851e-05, |
|
"loss": 1.0032, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 6.530756843800321e-05, |
|
"loss": 1.023, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 6.524315619967795e-05, |
|
"loss": 1.0446, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 6.517874396135267e-05, |
|
"loss": 0.9889, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 6.511433172302738e-05, |
|
"loss": 1.0353, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 6.504991948470209e-05, |
|
"loss": 1.0278, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 6.498550724637682e-05, |
|
"loss": 1.0253, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 6.492109500805154e-05, |
|
"loss": 1.0311, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 6.485668276972625e-05, |
|
"loss": 0.9975, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 6.479227053140096e-05, |
|
"loss": 1.0454, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 6.472785829307568e-05, |
|
"loss": 1.0552, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 6.466344605475041e-05, |
|
"loss": 1.025, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 6.459903381642512e-05, |
|
"loss": 1.01, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"eval_loss": 1.041631817817688, |
|
"eval_runtime": 118.7407, |
|
"eval_samples_per_second": 16.843, |
|
"eval_steps_per_second": 2.105, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 6.453462157809984e-05, |
|
"loss": 1.0156, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 6.447020933977456e-05, |
|
"loss": 1.0479, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 6.440579710144929e-05, |
|
"loss": 1.0229, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 6.4341384863124e-05, |
|
"loss": 0.9508, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 6.427697262479871e-05, |
|
"loss": 1.0362, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 6.421256038647343e-05, |
|
"loss": 1.0037, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 6.414814814814815e-05, |
|
"loss": 1.0062, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 6.408373590982287e-05, |
|
"loss": 1.045, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 6.401932367149759e-05, |
|
"loss": 0.9951, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 6.395491143317231e-05, |
|
"loss": 0.9898, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 6.389049919484703e-05, |
|
"loss": 1.0397, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 6.382608695652174e-05, |
|
"loss": 1.0571, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 6.376167471819646e-05, |
|
"loss": 1.0252, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 6.369726247987118e-05, |
|
"loss": 0.9697, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 6.36328502415459e-05, |
|
"loss": 1.0087, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 6.35684380032206e-05, |
|
"loss": 1.0284, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 6.350402576489534e-05, |
|
"loss": 0.9931, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 6.343961352657006e-05, |
|
"loss": 1.043, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 6.337520128824476e-05, |
|
"loss": 1.0397, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 6.331078904991948e-05, |
|
"loss": 1.0297, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"eval_loss": 1.0399410724639893, |
|
"eval_runtime": 118.6923, |
|
"eval_samples_per_second": 16.85, |
|
"eval_steps_per_second": 2.106, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 6.324637681159421e-05, |
|
"loss": 1.032, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 6.318196457326893e-05, |
|
"loss": 1.0077, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 6.311755233494364e-05, |
|
"loss": 0.9986, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 6.305314009661836e-05, |
|
"loss": 1.0348, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 6.298872785829309e-05, |
|
"loss": 1.0383, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 6.29243156199678e-05, |
|
"loss": 1.0266, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 6.285990338164251e-05, |
|
"loss": 1.0014, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 6.279549114331723e-05, |
|
"loss": 1.043, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 6.273107890499195e-05, |
|
"loss": 1.0, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 6.266666666666667e-05, |
|
"loss": 1.0062, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 6.260225442834139e-05, |
|
"loss": 1.0426, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 6.25378421900161e-05, |
|
"loss": 0.9844, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 6.247342995169082e-05, |
|
"loss": 1.024, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 6.240901771336554e-05, |
|
"loss": 0.9982, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 6.234460547504026e-05, |
|
"loss": 1.0251, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 6.228019323671498e-05, |
|
"loss": 0.9999, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 6.22157809983897e-05, |
|
"loss": 1.0115, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 6.21513687600644e-05, |
|
"loss": 1.0179, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 6.208695652173914e-05, |
|
"loss": 0.9518, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 6.202254428341385e-05, |
|
"loss": 0.9966, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"eval_loss": 1.0393410921096802, |
|
"eval_runtime": 118.6864, |
|
"eval_samples_per_second": 16.851, |
|
"eval_steps_per_second": 2.106, |
|
"step": 6000 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 15625, |
|
"num_train_epochs": 5, |
|
"save_steps": 200, |
|
"total_flos": 6.625191289926451e+17, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|