|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 5.0, |
|
"eval_steps": 500, |
|
"global_step": 45500, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.054945054945054944, |
|
"grad_norm": 1.780437707901001, |
|
"learning_rate": 0.0001978021978021978, |
|
"loss": 1.5906, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.10989010989010989, |
|
"grad_norm": 2.4972989559173584, |
|
"learning_rate": 0.00019560439560439562, |
|
"loss": 1.5616, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.16483516483516483, |
|
"grad_norm": 1.6102287769317627, |
|
"learning_rate": 0.00019340659340659342, |
|
"loss": 1.5549, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.21978021978021978, |
|
"grad_norm": 1.6323072910308838, |
|
"learning_rate": 0.00019120879120879122, |
|
"loss": 1.5446, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.27472527472527475, |
|
"grad_norm": 1.6228586435317993, |
|
"learning_rate": 0.00018901098901098903, |
|
"loss": 1.5549, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.32967032967032966, |
|
"grad_norm": 1.433160662651062, |
|
"learning_rate": 0.00018681318681318683, |
|
"loss": 1.5521, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.38461538461538464, |
|
"grad_norm": 1.4724050760269165, |
|
"learning_rate": 0.00018461538461538463, |
|
"loss": 1.5335, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.43956043956043955, |
|
"grad_norm": 1.2267966270446777, |
|
"learning_rate": 0.0001824175824175824, |
|
"loss": 1.5359, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.4945054945054945, |
|
"grad_norm": 1.3739681243896484, |
|
"learning_rate": 0.00018021978021978024, |
|
"loss": 1.5362, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.5494505494505495, |
|
"grad_norm": 1.3636854887008667, |
|
"learning_rate": 0.00017802197802197802, |
|
"loss": 1.5257, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.6043956043956044, |
|
"grad_norm": 1.9858155250549316, |
|
"learning_rate": 0.00017582417582417582, |
|
"loss": 1.5378, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.6593406593406593, |
|
"grad_norm": 1.6531416177749634, |
|
"learning_rate": 0.00017362637362637365, |
|
"loss": 1.5239, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.7142857142857143, |
|
"grad_norm": 1.638098120689392, |
|
"learning_rate": 0.00017142857142857143, |
|
"loss": 1.5331, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.7692307692307693, |
|
"grad_norm": 1.8147164583206177, |
|
"learning_rate": 0.00016923076923076923, |
|
"loss": 1.5291, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.8241758241758241, |
|
"grad_norm": 1.4330575466156006, |
|
"learning_rate": 0.00016703296703296706, |
|
"loss": 1.5328, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.8791208791208791, |
|
"grad_norm": 1.3865350484848022, |
|
"learning_rate": 0.00016483516483516484, |
|
"loss": 1.5266, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.9340659340659341, |
|
"grad_norm": 1.8921843767166138, |
|
"learning_rate": 0.00016263736263736264, |
|
"loss": 1.5323, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.989010989010989, |
|
"grad_norm": 2.120455026626587, |
|
"learning_rate": 0.00016043956043956044, |
|
"loss": 1.5254, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 1.043956043956044, |
|
"grad_norm": 1.6023536920547485, |
|
"learning_rate": 0.00015824175824175824, |
|
"loss": 1.4525, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 1.098901098901099, |
|
"grad_norm": 1.4707773923873901, |
|
"learning_rate": 0.00015604395604395605, |
|
"loss": 1.4538, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 1.1538461538461537, |
|
"grad_norm": 1.7984991073608398, |
|
"learning_rate": 0.00015384615384615385, |
|
"loss": 1.4421, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 1.2087912087912087, |
|
"grad_norm": 1.70187509059906, |
|
"learning_rate": 0.00015164835164835165, |
|
"loss": 1.4454, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 1.2637362637362637, |
|
"grad_norm": 2.5541532039642334, |
|
"learning_rate": 0.00014945054945054946, |
|
"loss": 1.4512, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 1.3186813186813187, |
|
"grad_norm": 1.5920054912567139, |
|
"learning_rate": 0.00014725274725274726, |
|
"loss": 1.4557, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 1.3736263736263736, |
|
"grad_norm": 2.0435898303985596, |
|
"learning_rate": 0.00014505494505494506, |
|
"loss": 1.4496, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 1.4285714285714286, |
|
"grad_norm": 2.3308217525482178, |
|
"learning_rate": 0.00014285714285714287, |
|
"loss": 1.437, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 1.4835164835164836, |
|
"grad_norm": 2.7433712482452393, |
|
"learning_rate": 0.00014065934065934067, |
|
"loss": 1.437, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 1.5384615384615383, |
|
"grad_norm": 2.6499412059783936, |
|
"learning_rate": 0.00013846153846153847, |
|
"loss": 1.4419, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 1.5934065934065935, |
|
"grad_norm": 2.6177027225494385, |
|
"learning_rate": 0.00013626373626373628, |
|
"loss": 1.4408, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 1.6483516483516483, |
|
"grad_norm": 1.9484443664550781, |
|
"learning_rate": 0.00013406593406593405, |
|
"loss": 1.4617, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 1.7032967032967035, |
|
"grad_norm": 2.649055242538452, |
|
"learning_rate": 0.00013186813186813188, |
|
"loss": 1.4482, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 1.7582417582417582, |
|
"grad_norm": 1.9473294019699097, |
|
"learning_rate": 0.0001296703296703297, |
|
"loss": 1.4358, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 1.8131868131868132, |
|
"grad_norm": 2.2678744792938232, |
|
"learning_rate": 0.00012747252747252746, |
|
"loss": 1.4333, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 1.8681318681318682, |
|
"grad_norm": 1.629854440689087, |
|
"learning_rate": 0.00012527472527472527, |
|
"loss": 1.4305, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 1.9230769230769231, |
|
"grad_norm": 2.318514823913574, |
|
"learning_rate": 0.0001230769230769231, |
|
"loss": 1.4292, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 1.978021978021978, |
|
"grad_norm": 2.1276426315307617, |
|
"learning_rate": 0.00012087912087912087, |
|
"loss": 1.435, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 2.032967032967033, |
|
"grad_norm": 2.3333852291107178, |
|
"learning_rate": 0.00011868131868131869, |
|
"loss": 1.3761, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 2.087912087912088, |
|
"grad_norm": 3.305957078933716, |
|
"learning_rate": 0.0001164835164835165, |
|
"loss": 1.3318, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 2.142857142857143, |
|
"grad_norm": 5.382218837738037, |
|
"learning_rate": 0.00011428571428571428, |
|
"loss": 1.3463, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 2.197802197802198, |
|
"grad_norm": 3.35638427734375, |
|
"learning_rate": 0.0001120879120879121, |
|
"loss": 1.3418, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 2.2527472527472527, |
|
"grad_norm": 3.04825496673584, |
|
"learning_rate": 0.0001098901098901099, |
|
"loss": 1.3568, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 2.3076923076923075, |
|
"grad_norm": 2.9136993885040283, |
|
"learning_rate": 0.0001076923076923077, |
|
"loss": 1.3507, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 2.3626373626373627, |
|
"grad_norm": 2.5451395511627197, |
|
"learning_rate": 0.0001054945054945055, |
|
"loss": 1.353, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 2.4175824175824174, |
|
"grad_norm": 2.779684066772461, |
|
"learning_rate": 0.00010329670329670331, |
|
"loss": 1.3456, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 2.4725274725274726, |
|
"grad_norm": 3.8211276531219482, |
|
"learning_rate": 0.0001010989010989011, |
|
"loss": 1.3519, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 2.5274725274725274, |
|
"grad_norm": 2.877223253250122, |
|
"learning_rate": 9.89010989010989e-05, |
|
"loss": 1.3506, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 2.5824175824175826, |
|
"grad_norm": 2.53200101852417, |
|
"learning_rate": 9.670329670329671e-05, |
|
"loss": 1.3564, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 2.6373626373626373, |
|
"grad_norm": 3.249028205871582, |
|
"learning_rate": 9.450549450549451e-05, |
|
"loss": 1.3393, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 2.6923076923076925, |
|
"grad_norm": 2.819939613342285, |
|
"learning_rate": 9.230769230769232e-05, |
|
"loss": 1.3413, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 2.7472527472527473, |
|
"grad_norm": 3.4391469955444336, |
|
"learning_rate": 9.010989010989012e-05, |
|
"loss": 1.3517, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 2.802197802197802, |
|
"grad_norm": 2.9319329261779785, |
|
"learning_rate": 8.791208791208791e-05, |
|
"loss": 1.3362, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 2.857142857142857, |
|
"grad_norm": 2.257335901260376, |
|
"learning_rate": 8.571428571428571e-05, |
|
"loss": 1.3402, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 2.912087912087912, |
|
"grad_norm": 2.6121294498443604, |
|
"learning_rate": 8.351648351648353e-05, |
|
"loss": 1.3615, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 2.967032967032967, |
|
"grad_norm": 2.716127395629883, |
|
"learning_rate": 8.131868131868132e-05, |
|
"loss": 1.3418, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 3.021978021978022, |
|
"grad_norm": 2.227154493331909, |
|
"learning_rate": 7.912087912087912e-05, |
|
"loss": 1.3199, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 3.076923076923077, |
|
"grad_norm": 3.0606658458709717, |
|
"learning_rate": 7.692307692307693e-05, |
|
"loss": 1.2553, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 3.131868131868132, |
|
"grad_norm": 2.8261947631835938, |
|
"learning_rate": 7.472527472527473e-05, |
|
"loss": 1.2688, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 3.186813186813187, |
|
"grad_norm": 3.738393783569336, |
|
"learning_rate": 7.252747252747253e-05, |
|
"loss": 1.2513, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 3.241758241758242, |
|
"grad_norm": 2.5947108268737793, |
|
"learning_rate": 7.032967032967034e-05, |
|
"loss": 1.2757, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 3.2967032967032965, |
|
"grad_norm": 2.707387924194336, |
|
"learning_rate": 6.813186813186814e-05, |
|
"loss": 1.274, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 3.3516483516483517, |
|
"grad_norm": 3.4624383449554443, |
|
"learning_rate": 6.593406593406594e-05, |
|
"loss": 1.2702, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 3.4065934065934065, |
|
"grad_norm": 3.7795355319976807, |
|
"learning_rate": 6.373626373626373e-05, |
|
"loss": 1.2409, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 3.4615384615384617, |
|
"grad_norm": 3.521811008453369, |
|
"learning_rate": 6.153846153846155e-05, |
|
"loss": 1.267, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 3.5164835164835164, |
|
"grad_norm": 2.7402503490448, |
|
"learning_rate": 5.9340659340659345e-05, |
|
"loss": 1.2641, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 3.571428571428571, |
|
"grad_norm": 2.937671661376953, |
|
"learning_rate": 5.714285714285714e-05, |
|
"loss": 1.2626, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 3.6263736263736264, |
|
"grad_norm": 2.886845350265503, |
|
"learning_rate": 5.494505494505495e-05, |
|
"loss": 1.269, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 3.6813186813186816, |
|
"grad_norm": 3.9152579307556152, |
|
"learning_rate": 5.274725274725275e-05, |
|
"loss": 1.2659, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 3.7362637362637363, |
|
"grad_norm": 2.7805376052856445, |
|
"learning_rate": 5.054945054945055e-05, |
|
"loss": 1.2523, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 3.791208791208791, |
|
"grad_norm": 3.7305562496185303, |
|
"learning_rate": 4.8351648351648355e-05, |
|
"loss": 1.2528, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 3.8461538461538463, |
|
"grad_norm": 3.0859031677246094, |
|
"learning_rate": 4.615384615384616e-05, |
|
"loss": 1.273, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 3.901098901098901, |
|
"grad_norm": 4.31486177444458, |
|
"learning_rate": 4.3956043956043955e-05, |
|
"loss": 1.2739, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 3.956043956043956, |
|
"grad_norm": 2.46838116645813, |
|
"learning_rate": 4.1758241758241765e-05, |
|
"loss": 1.2555, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 4.010989010989011, |
|
"grad_norm": 3.3113248348236084, |
|
"learning_rate": 3.956043956043956e-05, |
|
"loss": 1.2388, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 4.065934065934066, |
|
"grad_norm": 5.830134868621826, |
|
"learning_rate": 3.7362637362637365e-05, |
|
"loss": 1.1952, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 4.1208791208791204, |
|
"grad_norm": 2.97209095954895, |
|
"learning_rate": 3.516483516483517e-05, |
|
"loss": 1.1902, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 4.175824175824176, |
|
"grad_norm": 3.9511055946350098, |
|
"learning_rate": 3.296703296703297e-05, |
|
"loss": 1.1837, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 4.230769230769231, |
|
"grad_norm": 6.446171760559082, |
|
"learning_rate": 3.0769230769230774e-05, |
|
"loss": 1.1888, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 4.285714285714286, |
|
"grad_norm": 4.0232744216918945, |
|
"learning_rate": 2.857142857142857e-05, |
|
"loss": 1.1752, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 4.34065934065934, |
|
"grad_norm": 4.632090091705322, |
|
"learning_rate": 2.6373626373626374e-05, |
|
"loss": 1.1964, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 4.395604395604396, |
|
"grad_norm": 3.9423835277557373, |
|
"learning_rate": 2.4175824175824177e-05, |
|
"loss": 1.1861, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 4.450549450549451, |
|
"grad_norm": 4.528479099273682, |
|
"learning_rate": 2.1978021978021977e-05, |
|
"loss": 1.1999, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 4.5054945054945055, |
|
"grad_norm": 4.554440021514893, |
|
"learning_rate": 1.978021978021978e-05, |
|
"loss": 1.2045, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 4.56043956043956, |
|
"grad_norm": 4.417350769042969, |
|
"learning_rate": 1.7582417582417584e-05, |
|
"loss": 1.2074, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 4.615384615384615, |
|
"grad_norm": 4.048484802246094, |
|
"learning_rate": 1.5384615384615387e-05, |
|
"loss": 1.203, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 4.670329670329671, |
|
"grad_norm": 4.659012317657471, |
|
"learning_rate": 1.3186813186813187e-05, |
|
"loss": 1.1917, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 4.725274725274725, |
|
"grad_norm": 3.7058045864105225, |
|
"learning_rate": 1.0989010989010989e-05, |
|
"loss": 1.1981, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 4.78021978021978, |
|
"grad_norm": 3.214338779449463, |
|
"learning_rate": 8.791208791208792e-06, |
|
"loss": 1.1868, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 4.835164835164835, |
|
"grad_norm": 3.2473630905151367, |
|
"learning_rate": 6.5934065934065935e-06, |
|
"loss": 1.1979, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 4.8901098901098905, |
|
"grad_norm": 3.6867451667785645, |
|
"learning_rate": 4.395604395604396e-06, |
|
"loss": 1.1873, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 4.945054945054945, |
|
"grad_norm": 4.085856914520264, |
|
"learning_rate": 2.197802197802198e-06, |
|
"loss": 1.1997, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"grad_norm": 3.7809948921203613, |
|
"learning_rate": 0.0, |
|
"loss": 1.1797, |
|
"step": 45500 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 45500, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 5, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3.6063170257526784e+18, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|