|
{ |
|
"best_metric": 10.34882926940918, |
|
"best_model_checkpoint": "miner_id_24/checkpoint-100", |
|
"epoch": 0.01864193503285641, |
|
"eval_steps": 50, |
|
"global_step": 100, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0001864193503285641, |
|
"grad_norm": 0.029469789937138557, |
|
"learning_rate": 3.3333333333333333e-06, |
|
"loss": 10.3815, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0001864193503285641, |
|
"eval_loss": 10.381146430969238, |
|
"eval_runtime": 27.3606, |
|
"eval_samples_per_second": 330.22, |
|
"eval_steps_per_second": 165.128, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0003728387006571282, |
|
"grad_norm": 0.03511423617601395, |
|
"learning_rate": 6.666666666666667e-06, |
|
"loss": 10.3815, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.0005592580509856923, |
|
"grad_norm": 0.03218838572502136, |
|
"learning_rate": 1e-05, |
|
"loss": 10.3826, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0007456774013142564, |
|
"grad_norm": 0.027689678594470024, |
|
"learning_rate": 1.3333333333333333e-05, |
|
"loss": 10.3792, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.0009320967516428206, |
|
"grad_norm": 0.03155677765607834, |
|
"learning_rate": 1.6666666666666667e-05, |
|
"loss": 10.3816, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.0011185161019713846, |
|
"grad_norm": 0.03052655979990959, |
|
"learning_rate": 2e-05, |
|
"loss": 10.38, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.0013049354522999487, |
|
"grad_norm": 0.0268377885222435, |
|
"learning_rate": 2.3333333333333336e-05, |
|
"loss": 10.3798, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.0014913548026285128, |
|
"grad_norm": 0.03491060808300972, |
|
"learning_rate": 2.6666666666666667e-05, |
|
"loss": 10.3801, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.0016777741529570768, |
|
"grad_norm": 0.03049907088279724, |
|
"learning_rate": 3e-05, |
|
"loss": 10.3793, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.0018641935032856411, |
|
"grad_norm": 0.03284202516078949, |
|
"learning_rate": 3.3333333333333335e-05, |
|
"loss": 10.3808, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.002050612853614205, |
|
"grad_norm": 0.03127828985452652, |
|
"learning_rate": 3.6666666666666666e-05, |
|
"loss": 10.3818, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.0022370322039427693, |
|
"grad_norm": 0.029472503811120987, |
|
"learning_rate": 4e-05, |
|
"loss": 10.3824, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.0024234515542713333, |
|
"grad_norm": 0.03045397251844406, |
|
"learning_rate": 4.3333333333333334e-05, |
|
"loss": 10.3807, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.0026098709045998974, |
|
"grad_norm": 0.030138535425066948, |
|
"learning_rate": 4.666666666666667e-05, |
|
"loss": 10.3783, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.0027962902549284615, |
|
"grad_norm": 0.033551353961229324, |
|
"learning_rate": 5e-05, |
|
"loss": 10.3795, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.0029827096052570255, |
|
"grad_norm": 0.04115981236100197, |
|
"learning_rate": 5.333333333333333e-05, |
|
"loss": 10.3785, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.0031691289555855896, |
|
"grad_norm": 0.036658305674791336, |
|
"learning_rate": 5.666666666666667e-05, |
|
"loss": 10.3802, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.0033555483059141537, |
|
"grad_norm": 0.03716996684670448, |
|
"learning_rate": 6e-05, |
|
"loss": 10.3815, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.003541967656242718, |
|
"grad_norm": 0.03936820104718208, |
|
"learning_rate": 6.333333333333333e-05, |
|
"loss": 10.3802, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.0037283870065712823, |
|
"grad_norm": 0.03620583936572075, |
|
"learning_rate": 6.666666666666667e-05, |
|
"loss": 10.3805, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.003914806356899846, |
|
"grad_norm": 0.03842416778206825, |
|
"learning_rate": 7e-05, |
|
"loss": 10.3793, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.00410122570722841, |
|
"grad_norm": 0.039217282086610794, |
|
"learning_rate": 7.333333333333333e-05, |
|
"loss": 10.3797, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.0042876450575569745, |
|
"grad_norm": 0.04393612593412399, |
|
"learning_rate": 7.666666666666667e-05, |
|
"loss": 10.3784, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.0044740644078855385, |
|
"grad_norm": 0.03723832592368126, |
|
"learning_rate": 8e-05, |
|
"loss": 10.3764, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.004660483758214103, |
|
"grad_norm": 0.05864546820521355, |
|
"learning_rate": 8.333333333333334e-05, |
|
"loss": 10.3784, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.004846903108542667, |
|
"grad_norm": 0.05707390606403351, |
|
"learning_rate": 8.666666666666667e-05, |
|
"loss": 10.3796, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.005033322458871231, |
|
"grad_norm": 0.05543961003422737, |
|
"learning_rate": 9e-05, |
|
"loss": 10.38, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.005219741809199795, |
|
"grad_norm": 0.06495416909456253, |
|
"learning_rate": 9.333333333333334e-05, |
|
"loss": 10.3752, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.005406161159528359, |
|
"grad_norm": 0.07095163315534592, |
|
"learning_rate": 9.666666666666667e-05, |
|
"loss": 10.376, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.005592580509856923, |
|
"grad_norm": 0.07251051068305969, |
|
"learning_rate": 0.0001, |
|
"loss": 10.3783, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.005778999860185487, |
|
"grad_norm": 0.0717591941356659, |
|
"learning_rate": 9.994965332706573e-05, |
|
"loss": 10.3761, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.005965419210514051, |
|
"grad_norm": 0.08555017411708832, |
|
"learning_rate": 9.979871469976196e-05, |
|
"loss": 10.3753, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.006151838560842615, |
|
"grad_norm": 0.11320396512746811, |
|
"learning_rate": 9.954748808839674e-05, |
|
"loss": 10.3745, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.006338257911171179, |
|
"grad_norm": 0.09341420978307724, |
|
"learning_rate": 9.919647942993148e-05, |
|
"loss": 10.3745, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.006524677261499743, |
|
"grad_norm": 0.11599099636077881, |
|
"learning_rate": 9.874639560909117e-05, |
|
"loss": 10.3704, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.006711096611828307, |
|
"grad_norm": 0.10898492485284805, |
|
"learning_rate": 9.819814303479267e-05, |
|
"loss": 10.3727, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.006897515962156872, |
|
"grad_norm": 0.14422820508480072, |
|
"learning_rate": 9.755282581475769e-05, |
|
"loss": 10.3704, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.007083935312485436, |
|
"grad_norm": 0.14940433204174042, |
|
"learning_rate": 9.681174353198687e-05, |
|
"loss": 10.3691, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.0072703546628140004, |
|
"grad_norm": 0.14902304112911224, |
|
"learning_rate": 9.597638862757255e-05, |
|
"loss": 10.3665, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.0074567740131425645, |
|
"grad_norm": 0.16568797826766968, |
|
"learning_rate": 9.504844339512095e-05, |
|
"loss": 10.3681, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.007643193363471129, |
|
"grad_norm": 0.17017991840839386, |
|
"learning_rate": 9.40297765928369e-05, |
|
"loss": 10.3653, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.007829612713799693, |
|
"grad_norm": 0.18572087585926056, |
|
"learning_rate": 9.292243968009331e-05, |
|
"loss": 10.3632, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.008016032064128256, |
|
"grad_norm": 0.22373846173286438, |
|
"learning_rate": 9.172866268606513e-05, |
|
"loss": 10.3638, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.00820245141445682, |
|
"grad_norm": 0.19732271134853363, |
|
"learning_rate": 9.045084971874738e-05, |
|
"loss": 10.3615, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.008388870764785384, |
|
"grad_norm": 0.19585733115673065, |
|
"learning_rate": 8.90915741234015e-05, |
|
"loss": 10.3552, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.008575290115113949, |
|
"grad_norm": 0.20070262253284454, |
|
"learning_rate": 8.765357330018056e-05, |
|
"loss": 10.361, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.008761709465442512, |
|
"grad_norm": 0.18280582129955292, |
|
"learning_rate": 8.613974319136958e-05, |
|
"loss": 10.358, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.008948128815771077, |
|
"grad_norm": 0.19768880307674408, |
|
"learning_rate": 8.455313244934324e-05, |
|
"loss": 10.3532, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.009134548166099642, |
|
"grad_norm": 0.1908486932516098, |
|
"learning_rate": 8.289693629698564e-05, |
|
"loss": 10.3526, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.009320967516428205, |
|
"grad_norm": 0.24415874481201172, |
|
"learning_rate": 8.117449009293668e-05, |
|
"loss": 10.3496, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.009320967516428205, |
|
"eval_loss": 10.3571138381958, |
|
"eval_runtime": 27.4265, |
|
"eval_samples_per_second": 329.426, |
|
"eval_steps_per_second": 164.731, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.00950738686675677, |
|
"grad_norm": 0.1561504453420639, |
|
"learning_rate": 7.938926261462366e-05, |
|
"loss": 10.3609, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.009693806217085333, |
|
"grad_norm": 0.14776375889778137, |
|
"learning_rate": 7.754484907260513e-05, |
|
"loss": 10.3593, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.009880225567413898, |
|
"grad_norm": 0.14880362153053284, |
|
"learning_rate": 7.564496387029532e-05, |
|
"loss": 10.3593, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.010066644917742461, |
|
"grad_norm": 0.12313321232795715, |
|
"learning_rate": 7.369343312364993e-05, |
|
"loss": 10.3598, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.010253064268071026, |
|
"grad_norm": 0.14246243238449097, |
|
"learning_rate": 7.169418695587791e-05, |
|
"loss": 10.3582, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.01043948361839959, |
|
"grad_norm": 0.13358600437641144, |
|
"learning_rate": 6.965125158269619e-05, |
|
"loss": 10.3575, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.010625902968728155, |
|
"grad_norm": 0.10732542723417282, |
|
"learning_rate": 6.756874120406714e-05, |
|
"loss": 10.3576, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.010812322319056718, |
|
"grad_norm": 0.10971201956272125, |
|
"learning_rate": 6.545084971874738e-05, |
|
"loss": 10.3567, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.010998741669385283, |
|
"grad_norm": 0.09025903791189194, |
|
"learning_rate": 6.330184227833376e-05, |
|
"loss": 10.3538, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.011185161019713846, |
|
"grad_norm": 0.10398943722248077, |
|
"learning_rate": 6.112604669781572e-05, |
|
"loss": 10.3545, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.01137158037004241, |
|
"grad_norm": 0.10011833906173706, |
|
"learning_rate": 5.8927844739931834e-05, |
|
"loss": 10.3557, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.011557999720370974, |
|
"grad_norm": 0.10548334568738937, |
|
"learning_rate": 5.6711663290882776e-05, |
|
"loss": 10.3532, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.011744419070699539, |
|
"grad_norm": 0.08827032148838043, |
|
"learning_rate": 5.448196544517168e-05, |
|
"loss": 10.3518, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.011930838421028102, |
|
"grad_norm": 0.05861261487007141, |
|
"learning_rate": 5.2243241517525754e-05, |
|
"loss": 10.353, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.012117257771356667, |
|
"grad_norm": 0.08971337974071503, |
|
"learning_rate": 5e-05, |
|
"loss": 10.355, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.01230367712168523, |
|
"grad_norm": 0.08864837139844894, |
|
"learning_rate": 4.775675848247427e-05, |
|
"loss": 10.3548, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.012490096472013795, |
|
"grad_norm": 0.08446395397186279, |
|
"learning_rate": 4.551803455482833e-05, |
|
"loss": 10.3526, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.012676515822342358, |
|
"grad_norm": 0.06572958827018738, |
|
"learning_rate": 4.328833670911724e-05, |
|
"loss": 10.3523, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.012862935172670923, |
|
"grad_norm": 0.08103340864181519, |
|
"learning_rate": 4.107215526006817e-05, |
|
"loss": 10.3508, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.013049354522999487, |
|
"grad_norm": 0.06975308805704117, |
|
"learning_rate": 3.887395330218429e-05, |
|
"loss": 10.3515, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.013235773873328052, |
|
"grad_norm": 0.055503133684396744, |
|
"learning_rate": 3.6698157721666246e-05, |
|
"loss": 10.3512, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.013422193223656615, |
|
"grad_norm": 0.06876011937856674, |
|
"learning_rate": 3.4549150281252636e-05, |
|
"loss": 10.3516, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.01360861257398518, |
|
"grad_norm": 0.06682562828063965, |
|
"learning_rate": 3.243125879593286e-05, |
|
"loss": 10.3509, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.013795031924313745, |
|
"grad_norm": 0.0667678713798523, |
|
"learning_rate": 3.0348748417303823e-05, |
|
"loss": 10.3496, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.013981451274642308, |
|
"grad_norm": 0.05310027301311493, |
|
"learning_rate": 2.8305813044122097e-05, |
|
"loss": 10.3485, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.014167870624970873, |
|
"grad_norm": 0.07601341605186462, |
|
"learning_rate": 2.630656687635007e-05, |
|
"loss": 10.3489, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.014354289975299436, |
|
"grad_norm": 0.06028437986969948, |
|
"learning_rate": 2.43550361297047e-05, |
|
"loss": 10.3492, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.014540709325628001, |
|
"grad_norm": 0.054510705173015594, |
|
"learning_rate": 2.245515092739488e-05, |
|
"loss": 10.3496, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.014727128675956564, |
|
"grad_norm": 0.08019602298736572, |
|
"learning_rate": 2.061073738537635e-05, |
|
"loss": 10.3497, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.014913548026285129, |
|
"grad_norm": 0.0804019346833229, |
|
"learning_rate": 1.8825509907063327e-05, |
|
"loss": 10.3537, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.015099967376613692, |
|
"grad_norm": 0.07119160145521164, |
|
"learning_rate": 1.7103063703014372e-05, |
|
"loss": 10.3479, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.015286386726942257, |
|
"grad_norm": 0.0820232704281807, |
|
"learning_rate": 1.544686755065677e-05, |
|
"loss": 10.3519, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.01547280607727082, |
|
"grad_norm": 0.055926691740751266, |
|
"learning_rate": 1.3860256808630428e-05, |
|
"loss": 10.3495, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.015659225427599385, |
|
"grad_norm": 0.06931938230991364, |
|
"learning_rate": 1.2346426699819458e-05, |
|
"loss": 10.3468, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.01584564477792795, |
|
"grad_norm": 0.08054453134536743, |
|
"learning_rate": 1.090842587659851e-05, |
|
"loss": 10.3472, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.01603206412825651, |
|
"grad_norm": 0.07081463932991028, |
|
"learning_rate": 9.549150281252633e-06, |
|
"loss": 10.3477, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.016218483478585077, |
|
"grad_norm": 0.1176007091999054, |
|
"learning_rate": 8.271337313934869e-06, |
|
"loss": 10.3496, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.01640490282891364, |
|
"grad_norm": 0.06712072342634201, |
|
"learning_rate": 7.077560319906695e-06, |
|
"loss": 10.3454, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.016591322179242207, |
|
"grad_norm": 0.06814035773277283, |
|
"learning_rate": 5.9702234071631e-06, |
|
"loss": 10.3486, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.016777741529570768, |
|
"grad_norm": 0.09200986474752426, |
|
"learning_rate": 4.951556604879048e-06, |
|
"loss": 10.3454, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.016964160879899333, |
|
"grad_norm": 0.08182810246944427, |
|
"learning_rate": 4.023611372427471e-06, |
|
"loss": 10.3469, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.017150580230227898, |
|
"grad_norm": 0.0823381245136261, |
|
"learning_rate": 3.18825646801314e-06, |
|
"loss": 10.3436, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.017336999580556463, |
|
"grad_norm": 0.12770026922225952, |
|
"learning_rate": 2.4471741852423237e-06, |
|
"loss": 10.3455, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.017523418930885024, |
|
"grad_norm": 0.09665341675281525, |
|
"learning_rate": 1.8018569652073381e-06, |
|
"loss": 10.345, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.01770983828121359, |
|
"grad_norm": 0.08394477516412735, |
|
"learning_rate": 1.2536043909088191e-06, |
|
"loss": 10.342, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.017896257631542154, |
|
"grad_norm": 0.15145687758922577, |
|
"learning_rate": 8.035205700685167e-07, |
|
"loss": 10.3465, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.01808267698187072, |
|
"grad_norm": 0.11746706068515778, |
|
"learning_rate": 4.52511911603265e-07, |
|
"loss": 10.3434, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.018269096332199284, |
|
"grad_norm": 0.09703138470649719, |
|
"learning_rate": 2.012853002380466e-07, |
|
"loss": 10.3448, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.018455515682527845, |
|
"grad_norm": 0.15032333135604858, |
|
"learning_rate": 5.0346672934270534e-08, |
|
"loss": 10.3436, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.01864193503285641, |
|
"grad_norm": 0.18205277621746063, |
|
"learning_rate": 0.0, |
|
"loss": 10.3399, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.01864193503285641, |
|
"eval_loss": 10.34882926940918, |
|
"eval_runtime": 27.4349, |
|
"eval_samples_per_second": 329.326, |
|
"eval_steps_per_second": 164.681, |
|
"step": 100 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 100, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 50, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 5, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 0 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 11740707028992.0, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|