|
{ |
|
"best_metric": 0.44935256242752075, |
|
"best_model_checkpoint": "checkpoints/instrucode/with_input/1b_52k/checkpoint-4600", |
|
"epoch": 2.958199356913183, |
|
"eval_steps": 200, |
|
"global_step": 4600, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.6999999999999996e-05, |
|
"loss": 1.27, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 5.6999999999999996e-05, |
|
"loss": 1.2423, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 8.699999999999999e-05, |
|
"loss": 1.0757, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.000117, |
|
"loss": 0.8238, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.000147, |
|
"loss": 0.7379, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00017699999999999997, |
|
"loss": 0.6487, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00020699999999999996, |
|
"loss": 0.5623, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.000237, |
|
"loss": 0.5591, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.000267, |
|
"loss": 0.5358, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00029699999999999996, |
|
"loss": 0.5251, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00029964820846905535, |
|
"loss": 0.5371, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.000299257328990228, |
|
"loss": 0.5192, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00029886644951140066, |
|
"loss": 0.4993, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00029847557003257326, |
|
"loss": 0.4993, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0002980846905537459, |
|
"loss": 0.5304, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00029769381107491857, |
|
"loss": 0.5094, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00029730293159609117, |
|
"loss": 0.492, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0002969120521172638, |
|
"loss": 0.5123, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0002965211726384364, |
|
"loss": 0.5018, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0002961302931596091, |
|
"loss": 0.4899, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"eval_loss": 0.49921754002571106, |
|
"eval_runtime": 220.0251, |
|
"eval_samples_per_second": 9.09, |
|
"eval_steps_per_second": 1.136, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00029573941368078174, |
|
"loss": 0.5064, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0002953485342019544, |
|
"loss": 0.4656, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00029495765472312705, |
|
"loss": 0.4961, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00029456677524429965, |
|
"loss": 0.4909, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0002941758957654723, |
|
"loss": 0.484, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0002937850162866449, |
|
"loss": 0.5019, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00029339413680781756, |
|
"loss": 0.4887, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0002930032573289902, |
|
"loss": 0.4761, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0002926123778501628, |
|
"loss": 0.4874, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00029222149837133547, |
|
"loss": 0.479, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0002918306188925081, |
|
"loss": 0.5078, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0002914397394136808, |
|
"loss": 0.4875, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00029108794788273616, |
|
"loss": 0.4858, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0002906970684039088, |
|
"loss": 0.4633, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.0002903061889250814, |
|
"loss": 0.4804, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00028991530944625407, |
|
"loss": 0.4981, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00028952442996742667, |
|
"loss": 0.4784, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0002891335504885993, |
|
"loss": 0.4923, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.000288742671009772, |
|
"loss": 0.4892, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0002883517915309446, |
|
"loss": 0.4902, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"eval_loss": 0.48341748118400574, |
|
"eval_runtime": 220.2998, |
|
"eval_samples_per_second": 9.079, |
|
"eval_steps_per_second": 1.135, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00028796091205211723, |
|
"loss": 0.4725, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0002875700325732899, |
|
"loss": 0.4873, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00028717915309446255, |
|
"loss": 0.483, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00028678827361563515, |
|
"loss": 0.4587, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0002863973941368078, |
|
"loss": 0.4735, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0002860065146579804, |
|
"loss": 0.5109, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00028561563517915306, |
|
"loss": 0.478, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0002852247557003257, |
|
"loss": 0.4645, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00028483387622149837, |
|
"loss": 0.4932, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00028444299674267097, |
|
"loss": 0.4663, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0002840521172638436, |
|
"loss": 0.478, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0002836612377850163, |
|
"loss": 0.4567, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0002832703583061889, |
|
"loss": 0.4811, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00028287947882736154, |
|
"loss": 0.4678, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00028248859934853414, |
|
"loss": 0.4714, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002820977198697068, |
|
"loss": 0.4714, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00028170684039087945, |
|
"loss": 0.4583, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002813159609120521, |
|
"loss": 0.4493, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00028092508143322476, |
|
"loss": 0.4551, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00028053420195439736, |
|
"loss": 0.5053, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"eval_loss": 0.4742714762687683, |
|
"eval_runtime": 220.0995, |
|
"eval_samples_per_second": 9.087, |
|
"eval_steps_per_second": 1.136, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00028014332247557, |
|
"loss": 0.4556, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00027975244299674267, |
|
"loss": 0.4819, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00027936156351791527, |
|
"loss": 0.4729, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002789706840390879, |
|
"loss": 0.4798, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002785798045602606, |
|
"loss": 0.4804, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002781889250814332, |
|
"loss": 0.4661, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00027779804560260584, |
|
"loss": 0.465, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0002774071661237785, |
|
"loss": 0.4698, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00027701628664495115, |
|
"loss": 0.4824, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00027662540716612375, |
|
"loss": 0.4575, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.0002762345276872964, |
|
"loss": 0.4747, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.000275843648208469, |
|
"loss": 0.461, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00027545276872964166, |
|
"loss": 0.4588, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.0002750618892508143, |
|
"loss": 0.464, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00027467100977198697, |
|
"loss": 0.4829, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00027428013029315957, |
|
"loss": 0.4703, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00027388925081433223, |
|
"loss": 0.4697, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.0002734983713355049, |
|
"loss": 0.4774, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.0002731074918566775, |
|
"loss": 0.467, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00027271661237785014, |
|
"loss": 0.4924, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"eval_loss": 0.4694923460483551, |
|
"eval_runtime": 220.2001, |
|
"eval_samples_per_second": 9.083, |
|
"eval_steps_per_second": 1.135, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00027232573289902274, |
|
"loss": 0.4788, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.0002719348534201954, |
|
"loss": 0.4635, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00027154397394136805, |
|
"loss": 0.4816, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.0002711530944625407, |
|
"loss": 0.4582, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00027076221498371336, |
|
"loss": 0.4541, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00027037133550488596, |
|
"loss": 0.4748, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.0002699804560260586, |
|
"loss": 0.4762, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 0.0002695895765472312, |
|
"loss": 0.4678, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 0.0002691986970684039, |
|
"loss": 0.4581, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 0.00026880781758957653, |
|
"loss": 0.4439, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 0.00026841693811074913, |
|
"loss": 0.4881, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 0.0002680260586319218, |
|
"loss": 0.4636, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 0.00026763517915309444, |
|
"loss": 0.4649, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 0.0002672442996742671, |
|
"loss": 0.4769, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 0.00026685342019543975, |
|
"loss": 0.4568, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 0.00026646254071661235, |
|
"loss": 0.4673, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 0.000266071661237785, |
|
"loss": 0.4485, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 0.0002656807817589576, |
|
"loss": 0.471, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 0.00026528990228013026, |
|
"loss": 0.4631, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 0.0002648990228013029, |
|
"loss": 0.4724, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"eval_loss": 0.46559515595436096, |
|
"eval_runtime": 220.2191, |
|
"eval_samples_per_second": 9.082, |
|
"eval_steps_per_second": 1.135, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 0.0002645081433224756, |
|
"loss": 0.465, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 0.0002641172638436482, |
|
"loss": 0.4504, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 0.00026372638436482083, |
|
"loss": 0.4583, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 0.0002633355048859935, |
|
"loss": 0.471, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 0.0002629446254071661, |
|
"loss": 0.4779, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 0.00026255374592833874, |
|
"loss": 0.4543, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 0.00026216286644951134, |
|
"loss": 0.4629, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 0.000261771986970684, |
|
"loss": 0.45, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 0.00026138110749185665, |
|
"loss": 0.4493, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 0.0002609902280130293, |
|
"loss": 0.4695, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 0.00026059934853420196, |
|
"loss": 0.4538, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 0.00026020846905537456, |
|
"loss": 0.4672, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 0.0002598175895765472, |
|
"loss": 0.4802, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 0.0002594267100977198, |
|
"loss": 0.4656, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 0.0002590358306188925, |
|
"loss": 0.4466, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 0.00025864495114006513, |
|
"loss": 0.475, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 0.00025825407166123773, |
|
"loss": 0.4857, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 0.0002578631921824104, |
|
"loss": 0.4479, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 0.00025747231270358304, |
|
"loss": 0.4488, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 0.0002570814332247557, |
|
"loss": 0.4646, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"eval_loss": 0.46300560235977173, |
|
"eval_runtime": 220.2387, |
|
"eval_samples_per_second": 9.081, |
|
"eval_steps_per_second": 1.135, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 0.00025669055374592835, |
|
"loss": 0.4722, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 0.00025629967426710095, |
|
"loss": 0.4661, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 0.0002559087947882736, |
|
"loss": 0.4489, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 0.0002555179153094462, |
|
"loss": 0.4769, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 0.00025512703583061887, |
|
"loss": 0.4707, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 0.0002547361563517915, |
|
"loss": 0.4448, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 0.0002543452768729641, |
|
"loss": 0.4647, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 0.0002539543973941368, |
|
"loss": 0.4694, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 0.00025356351791530943, |
|
"loss": 0.475, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 0.0002531726384364821, |
|
"loss": 0.4589, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 0.0002527817589576547, |
|
"loss": 0.483, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 0.00025239087947882734, |
|
"loss": 0.449, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 0.00025199999999999995, |
|
"loss": 0.4533, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 0.0002516091205211726, |
|
"loss": 0.4429, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 0.00025121824104234526, |
|
"loss": 0.4623, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 0.0002508273615635179, |
|
"loss": 0.4767, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 0.00025043648208469057, |
|
"loss": 0.4803, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 0.00025004560260586317, |
|
"loss": 0.4499, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 0.0002496547231270358, |
|
"loss": 0.444, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 0.0002492638436482084, |
|
"loss": 0.477, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"eval_loss": 0.46069058775901794, |
|
"eval_runtime": 220.249, |
|
"eval_samples_per_second": 9.081, |
|
"eval_steps_per_second": 1.135, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 0.0002488729641693811, |
|
"loss": 0.4628, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 0.00024848208469055373, |
|
"loss": 0.4528, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 0.00024809120521172634, |
|
"loss": 0.4669, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 0.000247700325732899, |
|
"loss": 0.4607, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 0.00024730944625407165, |
|
"loss": 0.4703, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 0.0002469185667752443, |
|
"loss": 0.4613, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 0.00024652768729641696, |
|
"loss": 0.4593, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 0.00024613680781758956, |
|
"loss": 0.4515, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 0.0002457459283387622, |
|
"loss": 0.4493, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 0.0002453550488599348, |
|
"loss": 0.4572, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 0.00024496416938110747, |
|
"loss": 0.4396, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 0.0002445732899022801, |
|
"loss": 0.4572, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 0.0002441824104234527, |
|
"loss": 0.4652, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 0.00024379153094462538, |
|
"loss": 0.4449, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 0.00024340065146579804, |
|
"loss": 0.4574, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 0.00024300977198697066, |
|
"loss": 0.4336, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 0.00024261889250814332, |
|
"loss": 0.4368, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 0.00024222801302931592, |
|
"loss": 0.45, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 0.00024183713355048858, |
|
"loss": 0.4272, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 0.0002414462540716612, |
|
"loss": 0.4475, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"eval_loss": 0.4588911533355713, |
|
"eval_runtime": 220.3526, |
|
"eval_samples_per_second": 9.076, |
|
"eval_steps_per_second": 1.135, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 0.00024105537459283386, |
|
"loss": 0.4423, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 0.00024066449511400651, |
|
"loss": 0.442, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 0.00024027361563517911, |
|
"loss": 0.4412, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 0.00023988273615635177, |
|
"loss": 0.4403, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 0.0002394918566775244, |
|
"loss": 0.4397, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 0.00023910097719869705, |
|
"loss": 0.44, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 0.0002387100977198697, |
|
"loss": 0.4295, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 0.0002383192182410423, |
|
"loss": 0.4305, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 0.00023792833876221497, |
|
"loss": 0.433, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 0.0002375374592833876, |
|
"loss": 0.4514, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 0.00023714657980456025, |
|
"loss": 0.4696, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 0.00023675570032573288, |
|
"loss": 0.4497, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 0.0002363648208469055, |
|
"loss": 0.433, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 0.00023597394136807813, |
|
"loss": 0.4441, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 0.0002355830618892508, |
|
"loss": 0.4283, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 0.00023519218241042344, |
|
"loss": 0.46, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 0.00023480130293159607, |
|
"loss": 0.4386, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 0.00023441042345276873, |
|
"loss": 0.4548, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 0.00023401954397394133, |
|
"loss": 0.4546, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 0.00023362866449511398, |
|
"loss": 0.4482, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"eval_loss": 0.4585164189338684, |
|
"eval_runtime": 220.4297, |
|
"eval_samples_per_second": 9.073, |
|
"eval_steps_per_second": 1.134, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 0.00023323778501628664, |
|
"loss": 0.4383, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 0.00023284690553745927, |
|
"loss": 0.4464, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 0.00023245602605863192, |
|
"loss": 0.4354, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 0.00023206514657980452, |
|
"loss": 0.4403, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 0.00023167426710097718, |
|
"loss": 0.4367, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 0.0002312833876221498, |
|
"loss": 0.4485, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 0.00023089250814332246, |
|
"loss": 0.4553, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 0.00023050162866449512, |
|
"loss": 0.4262, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 0.00023011074918566772, |
|
"loss": 0.4394, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 0.00022971986970684037, |
|
"loss": 0.4441, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 0.000229328990228013, |
|
"loss": 0.4405, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 0.00022893811074918566, |
|
"loss": 0.434, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 0.0002285472312703583, |
|
"loss": 0.4326, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 0.0002281563517915309, |
|
"loss": 0.4401, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 0.00022776547231270357, |
|
"loss": 0.4394, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 0.0002273745928338762, |
|
"loss": 0.4404, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 0.00022698371335504885, |
|
"loss": 0.4437, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 0.00022659283387622148, |
|
"loss": 0.4598, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 0.0002262019543973941, |
|
"loss": 0.4308, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 0.00022581107491856674, |
|
"loss": 0.4498, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"eval_loss": 0.456809401512146, |
|
"eval_runtime": 220.2549, |
|
"eval_samples_per_second": 9.08, |
|
"eval_steps_per_second": 1.135, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 0.0002254201954397394, |
|
"loss": 0.4555, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 0.00022502931596091205, |
|
"loss": 0.4397, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 0.00022463843648208467, |
|
"loss": 0.4459, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 0.0002242475570032573, |
|
"loss": 0.4373, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 0.00022385667752442993, |
|
"loss": 0.4346, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 0.00022346579804560259, |
|
"loss": 0.4224, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 0.00022307491856677524, |
|
"loss": 0.4539, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 0.00022268403908794787, |
|
"loss": 0.4431, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 0.0002222931596091205, |
|
"loss": 0.4436, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 0.00022190228013029313, |
|
"loss": 0.4453, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 0.00022151140065146578, |
|
"loss": 0.4268, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 0.0002211205211726384, |
|
"loss": 0.4121, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 0.00022072964169381106, |
|
"loss": 0.4614, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 0.00022033876221498372, |
|
"loss": 0.452, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 0.00021994788273615632, |
|
"loss": 0.4472, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 0.00021955700325732898, |
|
"loss": 0.4318, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 0.0002191661237785016, |
|
"loss": 0.4476, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 0.00021877524429967426, |
|
"loss": 0.4378, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 0.00021838436482084691, |
|
"loss": 0.425, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 0.00021799348534201952, |
|
"loss": 0.4392, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"eval_loss": 0.45531219244003296, |
|
"eval_runtime": 220.3519, |
|
"eval_samples_per_second": 9.076, |
|
"eval_steps_per_second": 1.135, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 0.00021760260586319217, |
|
"loss": 0.4422, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 0.0002172117263843648, |
|
"loss": 0.4403, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 0.00021682084690553745, |
|
"loss": 0.45, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 0.00021642996742671008, |
|
"loss": 0.4282, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 0.0002160390879478827, |
|
"loss": 0.4319, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 0.00021564820846905534, |
|
"loss": 0.4438, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 0.000215257328990228, |
|
"loss": 0.4509, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 0.00021486644951140065, |
|
"loss": 0.4454, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 0.00021447557003257328, |
|
"loss": 0.4176, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 0.0002140846905537459, |
|
"loss": 0.4554, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 0.00021369381107491853, |
|
"loss": 0.4388, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 0.0002133029315960912, |
|
"loss": 0.456, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 0.00021291205211726384, |
|
"loss": 0.4257, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 0.00021252117263843647, |
|
"loss": 0.4428, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 0.0002121302931596091, |
|
"loss": 0.4266, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 0.00021173941368078173, |
|
"loss": 0.4591, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 0.00021134853420195438, |
|
"loss": 0.453, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 0.000210957654723127, |
|
"loss": 0.4414, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 0.00021056677524429967, |
|
"loss": 0.4619, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 0.00021017589576547227, |
|
"loss": 0.4285, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"eval_loss": 0.45426830649375916, |
|
"eval_runtime": 220.2316, |
|
"eval_samples_per_second": 9.081, |
|
"eval_steps_per_second": 1.135, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 0.00020978501628664492, |
|
"loss": 0.4476, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 0.00020939413680781758, |
|
"loss": 0.4361, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 0.0002090032573289902, |
|
"loss": 0.452, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 0.00020861237785016286, |
|
"loss": 0.4358, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 0.00020822149837133546, |
|
"loss": 0.447, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 0.00020783061889250812, |
|
"loss": 0.4592, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 0.00020743973941368077, |
|
"loss": 0.4529, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 0.0002070488599348534, |
|
"loss": 0.4545, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 0.00020665798045602606, |
|
"loss": 0.4303, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 0.00020626710097719866, |
|
"loss": 0.4489, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 0.0002058762214983713, |
|
"loss": 0.4306, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 0.00020548534201954394, |
|
"loss": 0.4441, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 0.0002050944625407166, |
|
"loss": 0.4132, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 0.00020470358306188925, |
|
"loss": 0.428, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 0.00020431270358306188, |
|
"loss": 0.4242, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 0.0002039218241042345, |
|
"loss": 0.4372, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 0.00020353094462540714, |
|
"loss": 0.4339, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 0.0002031400651465798, |
|
"loss": 0.434, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 0.00020274918566775245, |
|
"loss": 0.4282, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 0.00020235830618892507, |
|
"loss": 0.4352, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"eval_loss": 0.4528008699417114, |
|
"eval_runtime": 220.0694, |
|
"eval_samples_per_second": 9.088, |
|
"eval_steps_per_second": 1.136, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 0.0002019674267100977, |
|
"loss": 0.4478, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 0.00020157654723127033, |
|
"loss": 0.4317, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 0.00020118566775244299, |
|
"loss": 0.4447, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 0.00020079478827361561, |
|
"loss": 0.4454, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 0.00020040390879478827, |
|
"loss": 0.4422, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 0.00020001302931596087, |
|
"loss": 0.4504, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 0.00019962214983713353, |
|
"loss": 0.4322, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 0.00019923127035830618, |
|
"loss": 0.4387, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 0.0001988403908794788, |
|
"loss": 0.4398, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 0.00019844951140065146, |
|
"loss": 0.4384, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 0.00019805863192182407, |
|
"loss": 0.4412, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 0.00019766775244299672, |
|
"loss": 0.4535, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 0.00019727687296416938, |
|
"loss": 0.437, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 0.000196885993485342, |
|
"loss": 0.4403, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 0.00019649511400651466, |
|
"loss": 0.4343, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 0.00019610423452768726, |
|
"loss": 0.4281, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 0.00019571335504885992, |
|
"loss": 0.4364, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 0.00019532247557003254, |
|
"loss": 0.4473, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 0.0001949315960912052, |
|
"loss": 0.4421, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 0.00019454071661237785, |
|
"loss": 0.4427, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"eval_loss": 0.4516027867794037, |
|
"eval_runtime": 220.0573, |
|
"eval_samples_per_second": 9.089, |
|
"eval_steps_per_second": 1.136, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 0.00019414983713355046, |
|
"loss": 0.449, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 0.0001937589576547231, |
|
"loss": 0.433, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 0.00019336807817589574, |
|
"loss": 0.4368, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 0.0001929771986970684, |
|
"loss": 0.4432, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 0.00019258631921824105, |
|
"loss": 0.4394, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 0.00019219543973941365, |
|
"loss": 0.4421, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 0.0001918045602605863, |
|
"loss": 0.4526, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 0.00019141368078175893, |
|
"loss": 0.4322, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 0.0001910228013029316, |
|
"loss": 0.4376, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 0.00019063192182410422, |
|
"loss": 0.4496, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 0.00019024104234527687, |
|
"loss": 0.4429, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 0.00018985016286644947, |
|
"loss": 0.4269, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 0.00018945928338762213, |
|
"loss": 0.4409, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 0.00018906840390879478, |
|
"loss": 0.4328, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 0.0001886775244299674, |
|
"loss": 0.4438, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 0.00018828664495114007, |
|
"loss": 0.4238, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 0.00018789576547231267, |
|
"loss": 0.4288, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 0.00018750488599348532, |
|
"loss": 0.4356, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 0.00018711400651465798, |
|
"loss": 0.453, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 0.0001867231270358306, |
|
"loss": 0.4574, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"eval_loss": 0.4507245421409607, |
|
"eval_runtime": 219.8338, |
|
"eval_samples_per_second": 9.098, |
|
"eval_steps_per_second": 1.137, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 0.00018633224755700326, |
|
"loss": 0.4401, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 0.00018594136807817586, |
|
"loss": 0.4562, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 0.00018555048859934852, |
|
"loss": 0.4458, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 0.00018515960912052115, |
|
"loss": 0.4391, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 0.0001847687296416938, |
|
"loss": 0.4328, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 0.00018437785016286646, |
|
"loss": 0.4397, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 0.00018398697068403906, |
|
"loss": 0.4199, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 0.0001835960912052117, |
|
"loss": 0.439, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 0.00018320521172638434, |
|
"loss": 0.4354, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 0.000182814332247557, |
|
"loss": 0.4265, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 0.00018242345276872965, |
|
"loss": 0.4338, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 0.00018203257328990225, |
|
"loss": 0.4107, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 0.0001816416938110749, |
|
"loss": 0.4254, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 0.00018125081433224754, |
|
"loss": 0.4312, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 0.0001808599348534202, |
|
"loss": 0.4049, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 0.00018046905537459282, |
|
"loss": 0.4226, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 0.00018007817589576545, |
|
"loss": 0.4172, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 0.00017968729641693808, |
|
"loss": 0.4047, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 0.00017929641693811073, |
|
"loss": 0.4269, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 0.00017890553745928339, |
|
"loss": 0.4027, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"eval_loss": 0.45280253887176514, |
|
"eval_runtime": 219.8269, |
|
"eval_samples_per_second": 9.098, |
|
"eval_steps_per_second": 1.137, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 0.00017851465798045601, |
|
"loss": 0.4149, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 0.00017812377850162864, |
|
"loss": 0.4317, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 0.00017773289902280127, |
|
"loss": 0.4243, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 0.00017734201954397393, |
|
"loss": 0.4036, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 0.00017695114006514658, |
|
"loss": 0.4113, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 0.0001765602605863192, |
|
"loss": 0.4206, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 0.00017616938110749184, |
|
"loss": 0.4194, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 0.00017577850162866447, |
|
"loss": 0.4203, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 0.00017538762214983712, |
|
"loss": 0.4149, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 0.00017499674267100975, |
|
"loss": 0.3972, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 0.0001746058631921824, |
|
"loss": 0.3971, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 0.00017421498371335506, |
|
"loss": 0.4309, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 0.00017382410423452766, |
|
"loss": 0.4216, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 0.00017343322475570032, |
|
"loss": 0.4291, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 0.00017304234527687294, |
|
"loss": 0.4186, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 0.0001726514657980456, |
|
"loss": 0.4258, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 0.00017226058631921825, |
|
"loss": 0.4114, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 0.00017186970684039086, |
|
"loss": 0.4166, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 0.0001714788273615635, |
|
"loss": 0.4157, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 0.00017108794788273614, |
|
"loss": 0.4302, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"eval_loss": 0.4515763223171234, |
|
"eval_runtime": 219.868, |
|
"eval_samples_per_second": 9.096, |
|
"eval_steps_per_second": 1.137, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 0.0001706970684039088, |
|
"loss": 0.4098, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 0.00017030618892508142, |
|
"loss": 0.4326, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 0.00016991530944625405, |
|
"loss": 0.4303, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 0.00016952442996742668, |
|
"loss": 0.4325, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 0.00016913355048859933, |
|
"loss": 0.4164, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 0.000168742671009772, |
|
"loss": 0.4195, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 0.00016835179153094462, |
|
"loss": 0.408, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 0.00016796091205211725, |
|
"loss": 0.4185, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 0.00016757003257328987, |
|
"loss": 0.4155, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 0.00016717915309446253, |
|
"loss": 0.4004, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 0.00016678827361563518, |
|
"loss": 0.4078, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 0.0001663973941368078, |
|
"loss": 0.411, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 0.00016600651465798044, |
|
"loss": 0.4242, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 0.00016561563517915307, |
|
"loss": 0.4092, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 0.00016522475570032572, |
|
"loss": 0.4193, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 0.00016483387622149835, |
|
"loss": 0.419, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 0.000164442996742671, |
|
"loss": 0.4173, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 0.0001640521172638436, |
|
"loss": 0.414, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 0.00016366123778501626, |
|
"loss": 0.4115, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 0.00016327035830618892, |
|
"loss": 0.4227, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"eval_loss": 0.45233967900276184, |
|
"eval_runtime": 220.0403, |
|
"eval_samples_per_second": 9.089, |
|
"eval_steps_per_second": 1.136, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 0.00016287947882736155, |
|
"loss": 0.4238, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 0.0001624885993485342, |
|
"loss": 0.4066, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 0.0001620977198697068, |
|
"loss": 0.4226, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 0.00016170684039087946, |
|
"loss": 0.429, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 0.0001613159609120521, |
|
"loss": 0.4181, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 0.00016092508143322474, |
|
"loss": 0.4276, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 0.0001605342019543974, |
|
"loss": 0.4169, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 0.00016014332247557003, |
|
"loss": 0.4223, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 0.00015975244299674265, |
|
"loss": 0.4136, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 0.00015936156351791528, |
|
"loss": 0.4137, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 0.00015897068403908794, |
|
"loss": 0.4247, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 0.0001585798045602606, |
|
"loss": 0.4052, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 0.00015818892508143322, |
|
"loss": 0.4308, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 0.00015779804560260585, |
|
"loss": 0.4115, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 0.00015740716612377848, |
|
"loss": 0.423, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 0.00015701628664495113, |
|
"loss": 0.4181, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 0.00015662540716612376, |
|
"loss": 0.4069, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 0.00015623452768729641, |
|
"loss": 0.4112, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 0.00015584364820846904, |
|
"loss": 0.4221, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 0.00015545276872964167, |
|
"loss": 0.4318, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"eval_loss": 0.4517545998096466, |
|
"eval_runtime": 220.1454, |
|
"eval_samples_per_second": 9.085, |
|
"eval_steps_per_second": 1.136, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 0.00015506188925081433, |
|
"loss": 0.419, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 0.00015467100977198695, |
|
"loss": 0.4357, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 0.0001542801302931596, |
|
"loss": 0.4197, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 0.0001538892508143322, |
|
"loss": 0.4064, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 0.00015349837133550487, |
|
"loss": 0.421, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 0.00015310749185667752, |
|
"loss": 0.416, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 0.00015271661237785015, |
|
"loss": 0.3995, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 0.0001523257328990228, |
|
"loss": 0.4222, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 0.0001519348534201954, |
|
"loss": 0.3989, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 0.00015154397394136806, |
|
"loss": 0.3986, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 0.0001511530944625407, |
|
"loss": 0.4256, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 0.00015076221498371334, |
|
"loss": 0.4289, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 0.000150371335504886, |
|
"loss": 0.4138, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 0.00014998045602605863, |
|
"loss": 0.4125, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 0.00014958957654723126, |
|
"loss": 0.4236, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 0.00014919869706840388, |
|
"loss": 0.4189, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 0.00014880781758957654, |
|
"loss": 0.4155, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 0.00014841693811074917, |
|
"loss": 0.4144, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 0.00014802605863192182, |
|
"loss": 0.4295, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 0.00014763517915309445, |
|
"loss": 0.3985, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"eval_loss": 0.4515686333179474, |
|
"eval_runtime": 220.1634, |
|
"eval_samples_per_second": 9.084, |
|
"eval_steps_per_second": 1.136, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 0.00014724429967426708, |
|
"loss": 0.418, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 0.00014685342019543973, |
|
"loss": 0.4267, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 0.00014646254071661236, |
|
"loss": 0.4231, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 0.000146071661237785, |
|
"loss": 0.4153, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 0.00014568078175895765, |
|
"loss": 0.4228, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 0.00014528990228013027, |
|
"loss": 0.4041, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 0.00014489902280130293, |
|
"loss": 0.4361, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 0.00014450814332247556, |
|
"loss": 0.4158, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 0.00014411726384364819, |
|
"loss": 0.4235, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 0.00014372638436482084, |
|
"loss": 0.4157, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 0.00014333550488599347, |
|
"loss": 0.4233, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 0.00014294462540716612, |
|
"loss": 0.4138, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 0.00014255374592833875, |
|
"loss": 0.422, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 0.00014216286644951138, |
|
"loss": 0.4234, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 0.00014177198697068404, |
|
"loss": 0.4224, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 0.00014138110749185666, |
|
"loss": 0.4256, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 0.0001409902280130293, |
|
"loss": 0.4216, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 0.00014059934853420195, |
|
"loss": 0.4139, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 0.00014020846905537458, |
|
"loss": 0.4112, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 0.00013981758957654723, |
|
"loss": 0.4203, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"eval_loss": 0.45124533772468567, |
|
"eval_runtime": 220.1249, |
|
"eval_samples_per_second": 9.086, |
|
"eval_steps_per_second": 1.136, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 0.00013942671009771986, |
|
"loss": 0.4072, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 0.0001390358306188925, |
|
"loss": 0.4134, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 0.00013864495114006514, |
|
"loss": 0.4175, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 0.00013825407166123777, |
|
"loss": 0.4215, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 0.00013786319218241043, |
|
"loss": 0.4226, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 0.00013747231270358305, |
|
"loss": 0.4003, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 0.00013708143322475568, |
|
"loss": 0.4198, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 0.00013669055374592834, |
|
"loss": 0.4225, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 0.00013629967426710097, |
|
"loss": 0.418, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 0.0001359087947882736, |
|
"loss": 0.4234, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 0.00013551791530944622, |
|
"loss": 0.421, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 0.00013512703583061888, |
|
"loss": 0.4162, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 0.00013473615635179153, |
|
"loss": 0.4103, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 0.00013434527687296416, |
|
"loss": 0.4157, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 0.0001339543973941368, |
|
"loss": 0.4144, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 0.00013356351791530944, |
|
"loss": 0.4096, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 0.00013317263843648207, |
|
"loss": 0.4134, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 0.00013278175895765473, |
|
"loss": 0.42, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 0.00013239087947882735, |
|
"loss": 0.4187, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 0.00013199999999999998, |
|
"loss": 0.4194, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"eval_loss": 0.4500805735588074, |
|
"eval_runtime": 220.1289, |
|
"eval_samples_per_second": 9.086, |
|
"eval_steps_per_second": 1.136, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 0.00013160912052117264, |
|
"loss": 0.4193, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 0.00013121824104234527, |
|
"loss": 0.4249, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 0.0001308273615635179, |
|
"loss": 0.4187, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 0.00013043648208469052, |
|
"loss": 0.4215, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 0.00013004560260586318, |
|
"loss": 0.4075, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 0.00012965472312703583, |
|
"loss": 0.4165, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 0.00012926384364820846, |
|
"loss": 0.4197, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 0.0001288729641693811, |
|
"loss": 0.4392, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 0.00012848208469055372, |
|
"loss": 0.4064, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 0.00012809120521172637, |
|
"loss": 0.4176, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 0.00012770032573289903, |
|
"loss": 0.4116, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 0.00012730944625407166, |
|
"loss": 0.4177, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 0.00012691856677524428, |
|
"loss": 0.4109, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 0.00012652768729641694, |
|
"loss": 0.4072, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 0.00012613680781758957, |
|
"loss": 0.435, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 0.0001257459283387622, |
|
"loss": 0.401, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 0.00012535504885993482, |
|
"loss": 0.4175, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 0.00012496416938110748, |
|
"loss": 0.4245, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 0.00012457328990228013, |
|
"loss": 0.415, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 0.00012418241042345276, |
|
"loss": 0.4247, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"eval_loss": 0.44935256242752075, |
|
"eval_runtime": 220.1783, |
|
"eval_samples_per_second": 9.084, |
|
"eval_steps_per_second": 1.135, |
|
"step": 4600 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 7775, |
|
"num_train_epochs": 5, |
|
"save_steps": 200, |
|
"total_flos": 2.3487809765824266e+18, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|