|
{ |
|
"best_metric": 0.7669394612312317, |
|
"best_model_checkpoint": "/mnt/bn/qingyi-bn-lq/llama/saved-alpaca-cot7b/checkpoint-2600", |
|
"epoch": 2.6680348896870187, |
|
"global_step": 2600, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 5.9999999999999995e-05, |
|
"loss": 2.2274, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00011999999999999999, |
|
"loss": 1.8266, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00017999999999999998, |
|
"loss": 1.2659, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00023999999999999998, |
|
"loss": 1.0441, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003, |
|
"loss": 0.9422, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0002978738483345145, |
|
"loss": 0.8938, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.000295747696669029, |
|
"loss": 0.8698, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00029362154500354353, |
|
"loss": 0.8406, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0002914953933380581, |
|
"loss": 0.835, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0002893692416725726, |
|
"loss": 0.8199, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"eval_loss": 0.8377243876457214, |
|
"eval_runtime": 16.2911, |
|
"eval_samples_per_second": 122.767, |
|
"eval_steps_per_second": 1.964, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00028724309000708714, |
|
"loss": 0.8119, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00028511693834160166, |
|
"loss": 0.808, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0002829907866761162, |
|
"loss": 0.8067, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00028086463501063075, |
|
"loss": 0.8058, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00027873848334514527, |
|
"loss": 0.7993, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0002766123316796598, |
|
"loss": 0.802, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0002744861800141743, |
|
"loss": 0.7932, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002723600283486889, |
|
"loss": 0.7873, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0002702338766832034, |
|
"loss": 0.7864, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002681077250177179, |
|
"loss": 0.7925, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"eval_loss": 0.8077966570854187, |
|
"eval_runtime": 16.2968, |
|
"eval_samples_per_second": 122.724, |
|
"eval_steps_per_second": 1.964, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00026598157335223243, |
|
"loss": 0.7862, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00026385542168674695, |
|
"loss": 0.7825, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.0002617292700212615, |
|
"loss": 0.7914, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00025960311835577604, |
|
"loss": 0.7951, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00025747696669029056, |
|
"loss": 0.7824, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.0002553508150248051, |
|
"loss": 0.7808, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.0002532246633593196, |
|
"loss": 0.7811, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 0.0002510985116938341, |
|
"loss": 0.7826, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 0.0002489723600283487, |
|
"loss": 0.7758, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 0.0002468462083628632, |
|
"loss": 0.7827, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"eval_loss": 0.796688973903656, |
|
"eval_runtime": 16.4193, |
|
"eval_samples_per_second": 121.808, |
|
"eval_steps_per_second": 1.949, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 0.0002447200566973777, |
|
"loss": 0.7732, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 0.00024259390503189224, |
|
"loss": 0.7784, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 0.0002404677533664068, |
|
"loss": 0.78, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 0.0002383416017009213, |
|
"loss": 0.7789, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 0.00023621545003543583, |
|
"loss": 0.7762, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 0.0002340892983699504, |
|
"loss": 0.7802, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 0.00023196314670446492, |
|
"loss": 0.7749, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 0.00022983699503897943, |
|
"loss": 0.7643, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 0.00022771084337349395, |
|
"loss": 0.7648, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 0.00022558469170800847, |
|
"loss": 0.767, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"eval_loss": 0.7896291613578796, |
|
"eval_runtime": 16.2986, |
|
"eval_samples_per_second": 122.71, |
|
"eval_steps_per_second": 1.963, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 0.00022345854004252302, |
|
"loss": 0.7707, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 0.00022133238837703754, |
|
"loss": 0.76, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 0.00021920623671155208, |
|
"loss": 0.7644, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 0.0002170800850460666, |
|
"loss": 0.7741, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 0.00021495393338058114, |
|
"loss": 0.7678, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 0.00021282778171509566, |
|
"loss": 0.76, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 0.00021070163004961018, |
|
"loss": 0.779, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 0.0002085754783841247, |
|
"loss": 0.7605, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 0.00020644932671863922, |
|
"loss": 0.7571, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 0.0002043231750531538, |
|
"loss": 0.7595, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"eval_loss": 0.7847340106964111, |
|
"eval_runtime": 16.2989, |
|
"eval_samples_per_second": 122.707, |
|
"eval_steps_per_second": 1.963, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 0.0002021970233876683, |
|
"loss": 0.7569, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 0.00020007087172218283, |
|
"loss": 0.7523, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 0.00019794472005669735, |
|
"loss": 0.7621, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 0.0001958185683912119, |
|
"loss": 0.7556, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 0.0001936924167257264, |
|
"loss": 0.7594, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 0.00019156626506024093, |
|
"loss": 0.7612, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 0.0001894401133947555, |
|
"loss": 0.7587, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 0.00018731396172927002, |
|
"loss": 0.7533, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 0.00018518781006378454, |
|
"loss": 0.7573, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 0.00018306165839829906, |
|
"loss": 0.761, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"eval_loss": 0.7821407318115234, |
|
"eval_runtime": 16.2951, |
|
"eval_samples_per_second": 122.736, |
|
"eval_steps_per_second": 1.964, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 0.00018093550673281358, |
|
"loss": 0.7548, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 0.00017880935506732812, |
|
"loss": 0.7556, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 0.00017668320340184267, |
|
"loss": 0.7605, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 0.00017455705173635719, |
|
"loss": 0.7528, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 0.0001724309000708717, |
|
"loss": 0.7584, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 0.00017030474840538625, |
|
"loss": 0.7596, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 0.00016817859673990077, |
|
"loss": 0.7514, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 0.00016605244507441529, |
|
"loss": 0.746, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 0.0001639262934089298, |
|
"loss": 0.7581, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 0.00016180014174344438, |
|
"loss": 0.7619, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"eval_loss": 0.777686357498169, |
|
"eval_runtime": 16.3364, |
|
"eval_samples_per_second": 122.426, |
|
"eval_steps_per_second": 1.959, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 0.0001596739900779589, |
|
"loss": 0.7421, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 0.00015754783841247341, |
|
"loss": 0.745, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 0.00015542168674698793, |
|
"loss": 0.7631, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 0.00015329553508150245, |
|
"loss": 0.7532, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 0.000151169383416017, |
|
"loss": 0.7491, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 0.00014904323175053151, |
|
"loss": 0.743, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 0.00014691708008504606, |
|
"loss": 0.7575, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 0.00014479092841956058, |
|
"loss": 0.7519, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 0.00014266477675407512, |
|
"loss": 0.7504, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 0.00014053862508858964, |
|
"loss": 0.748, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"eval_loss": 0.7752982378005981, |
|
"eval_runtime": 16.3209, |
|
"eval_samples_per_second": 122.542, |
|
"eval_steps_per_second": 1.961, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 0.00013841247342310416, |
|
"loss": 0.7536, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 0.0001362863217576187, |
|
"loss": 0.7455, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 0.00013416017009213323, |
|
"loss": 0.7509, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 0.00013203401842664774, |
|
"loss": 0.7529, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 0.0001299078667611623, |
|
"loss": 0.7505, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 0.0001277817150956768, |
|
"loss": 0.7384, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 0.00012565556343019135, |
|
"loss": 0.7396, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 0.00012352941176470587, |
|
"loss": 0.7552, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 0.0001214032600992204, |
|
"loss": 0.7459, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 0.00011927710843373494, |
|
"loss": 0.7494, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"eval_loss": 0.7731354832649231, |
|
"eval_runtime": 16.3191, |
|
"eval_samples_per_second": 122.556, |
|
"eval_steps_per_second": 1.961, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 0.00011715095676824945, |
|
"loss": 0.7498, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 0.000115024805102764, |
|
"loss": 0.7544, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 0.00011289865343727852, |
|
"loss": 0.751, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 0.00011077250177179304, |
|
"loss": 0.745, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 0.00010864635010630757, |
|
"loss": 0.742, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 0.00010652019844082211, |
|
"loss": 0.745, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 0.00010439404677533663, |
|
"loss": 0.7457, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 0.00010226789510985115, |
|
"loss": 0.7475, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 0.0001001417434443657, |
|
"loss": 0.742, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 9.801559177888021e-05, |
|
"loss": 0.7365, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"eval_loss": 0.7715820074081421, |
|
"eval_runtime": 16.3224, |
|
"eval_samples_per_second": 122.531, |
|
"eval_steps_per_second": 1.96, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 9.588944011339475e-05, |
|
"loss": 0.7431, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 9.376328844790927e-05, |
|
"loss": 0.7515, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 9.163713678242381e-05, |
|
"loss": 0.7483, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 8.951098511693833e-05, |
|
"loss": 0.7355, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 8.738483345145286e-05, |
|
"loss": 0.7404, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 8.525868178596739e-05, |
|
"loss": 0.7403, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 8.313253012048193e-05, |
|
"loss": 0.7435, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 8.100637845499644e-05, |
|
"loss": 0.7503, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 7.888022678951099e-05, |
|
"loss": 0.7335, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 7.675407512402551e-05, |
|
"loss": 0.7481, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"eval_loss": 0.769282341003418, |
|
"eval_runtime": 16.3527, |
|
"eval_samples_per_second": 122.304, |
|
"eval_steps_per_second": 1.957, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 7.462792345854004e-05, |
|
"loss": 0.7396, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 7.250177179305457e-05, |
|
"loss": 0.736, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 7.037562012756909e-05, |
|
"loss": 0.7443, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 6.824946846208362e-05, |
|
"loss": 0.7501, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 6.612331679659815e-05, |
|
"loss": 0.7335, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 6.399716513111269e-05, |
|
"loss": 0.7507, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 6.187101346562722e-05, |
|
"loss": 0.7449, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 5.9744861800141736e-05, |
|
"loss": 0.7397, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 5.761871013465627e-05, |
|
"loss": 0.743, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 5.5492558469170794e-05, |
|
"loss": 0.7416, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"eval_loss": 0.7678119540214539, |
|
"eval_runtime": 16.3147, |
|
"eval_samples_per_second": 122.589, |
|
"eval_steps_per_second": 1.961, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 5.3366406803685326e-05, |
|
"loss": 0.7457, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 5.124025513819986e-05, |
|
"loss": 0.7405, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 4.911410347271438e-05, |
|
"loss": 0.7381, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 4.6987951807228915e-05, |
|
"loss": 0.736, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 4.486180014174344e-05, |
|
"loss": 0.7437, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 4.273564847625797e-05, |
|
"loss": 0.7485, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 4.0609496810772504e-05, |
|
"loss": 0.7383, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 3.848334514528702e-05, |
|
"loss": 0.7362, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 3.635719347980156e-05, |
|
"loss": 0.7387, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 3.4231041814316086e-05, |
|
"loss": 0.7419, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"eval_loss": 0.7669394612312317, |
|
"eval_runtime": 16.3329, |
|
"eval_samples_per_second": 122.452, |
|
"eval_steps_per_second": 1.959, |
|
"step": 2600 |
|
} |
|
], |
|
"max_steps": 2922, |
|
"num_train_epochs": 3, |
|
"total_flos": 6.758732816713777e+18, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|