|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.8585365853658536, |
|
"eval_steps": 500, |
|
"global_step": 440, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01951219512195122, |
|
"grad_norm": 0.14488260447978973, |
|
"learning_rate": 0.00019934683213585893, |
|
"loss": 0.6356, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.03902439024390244, |
|
"grad_norm": 0.1634250283241272, |
|
"learning_rate": 0.00019804049640757677, |
|
"loss": 0.4368, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.05853658536585366, |
|
"grad_norm": 0.1504327356815338, |
|
"learning_rate": 0.0001967341606792946, |
|
"loss": 0.407, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.07804878048780488, |
|
"grad_norm": 0.16282524168491364, |
|
"learning_rate": 0.00019542782495101242, |
|
"loss": 0.3788, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.0975609756097561, |
|
"grad_norm": 0.17505913972854614, |
|
"learning_rate": 0.00019412148922273026, |
|
"loss": 0.3705, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.11707317073170732, |
|
"grad_norm": 0.18271876871585846, |
|
"learning_rate": 0.00019281515349444807, |
|
"loss": 0.3654, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.13658536585365855, |
|
"grad_norm": 0.1547907590866089, |
|
"learning_rate": 0.0001915088177661659, |
|
"loss": 0.3845, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.15609756097560976, |
|
"grad_norm": 0.1806977242231369, |
|
"learning_rate": 0.00019020248203788375, |
|
"loss": 0.362, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.17560975609756097, |
|
"grad_norm": 0.16502083837985992, |
|
"learning_rate": 0.00018889614630960156, |
|
"loss": 0.3436, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.1951219512195122, |
|
"grad_norm": 0.15947739779949188, |
|
"learning_rate": 0.0001875898105813194, |
|
"loss": 0.3221, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.2146341463414634, |
|
"grad_norm": 0.16028478741645813, |
|
"learning_rate": 0.00018628347485303724, |
|
"loss": 0.3291, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.23414634146341465, |
|
"grad_norm": 0.16755621135234833, |
|
"learning_rate": 0.00018497713912475508, |
|
"loss": 0.3454, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.25365853658536586, |
|
"grad_norm": 0.17687192559242249, |
|
"learning_rate": 0.00018367080339647292, |
|
"loss": 0.3265, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.2731707317073171, |
|
"grad_norm": 0.12853658199310303, |
|
"learning_rate": 0.00018236446766819073, |
|
"loss": 0.3136, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.2926829268292683, |
|
"grad_norm": 0.15641653537750244, |
|
"learning_rate": 0.00018105813193990857, |
|
"loss": 0.3124, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.3121951219512195, |
|
"grad_norm": 0.1840222179889679, |
|
"learning_rate": 0.00017975179621162638, |
|
"loss": 0.3308, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.33170731707317075, |
|
"grad_norm": 0.16983111202716827, |
|
"learning_rate": 0.00017844546048334422, |
|
"loss": 0.3116, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.35121951219512193, |
|
"grad_norm": 0.18679502606391907, |
|
"learning_rate": 0.00017713912475506206, |
|
"loss": 0.3458, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.37073170731707317, |
|
"grad_norm": 0.1655397266149521, |
|
"learning_rate": 0.0001758327890267799, |
|
"loss": 0.3041, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.3902439024390244, |
|
"grad_norm": 0.16899655759334564, |
|
"learning_rate": 0.00017452645329849774, |
|
"loss": 0.3179, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.4097560975609756, |
|
"grad_norm": 0.14320065081119537, |
|
"learning_rate": 0.00017322011757021555, |
|
"loss": 0.2896, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.4292682926829268, |
|
"grad_norm": 0.18079160153865814, |
|
"learning_rate": 0.0001719137818419334, |
|
"loss": 0.3121, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.44878048780487806, |
|
"grad_norm": 0.15252567827701569, |
|
"learning_rate": 0.00017060744611365123, |
|
"loss": 0.3054, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.4682926829268293, |
|
"grad_norm": 0.1688212901353836, |
|
"learning_rate": 0.00016930111038536904, |
|
"loss": 0.316, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.4878048780487805, |
|
"grad_norm": 0.1604667603969574, |
|
"learning_rate": 0.00016799477465708688, |
|
"loss": 0.3174, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.5073170731707317, |
|
"grad_norm": 0.19493722915649414, |
|
"learning_rate": 0.0001666884389288047, |
|
"loss": 0.307, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.526829268292683, |
|
"grad_norm": 0.13511119782924652, |
|
"learning_rate": 0.00016538210320052253, |
|
"loss": 0.2918, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.5463414634146342, |
|
"grad_norm": 0.1683417409658432, |
|
"learning_rate": 0.00016407576747224037, |
|
"loss": 0.2704, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.5658536585365853, |
|
"grad_norm": 0.16002771258354187, |
|
"learning_rate": 0.0001627694317439582, |
|
"loss": 0.3401, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.5853658536585366, |
|
"grad_norm": 0.1851789504289627, |
|
"learning_rate": 0.00016146309601567605, |
|
"loss": 0.3134, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.6048780487804878, |
|
"grad_norm": 0.18187113106250763, |
|
"learning_rate": 0.00016015676028739386, |
|
"loss": 0.2732, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.624390243902439, |
|
"grad_norm": 0.13298119604587555, |
|
"learning_rate": 0.0001588504245591117, |
|
"loss": 0.3317, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.6439024390243903, |
|
"grad_norm": 0.17413659393787384, |
|
"learning_rate": 0.00015754408883082954, |
|
"loss": 0.3176, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.6634146341463415, |
|
"grad_norm": 0.17591014504432678, |
|
"learning_rate": 0.00015623775310254735, |
|
"loss": 0.3074, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.6829268292682927, |
|
"grad_norm": 0.17484629154205322, |
|
"learning_rate": 0.0001549314173742652, |
|
"loss": 0.3033, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.7024390243902439, |
|
"grad_norm": 0.17957189679145813, |
|
"learning_rate": 0.000153625081645983, |
|
"loss": 0.2995, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.7219512195121951, |
|
"grad_norm": 0.1616777628660202, |
|
"learning_rate": 0.00015231874591770084, |
|
"loss": 0.2955, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.7414634146341463, |
|
"grad_norm": 0.165989488363266, |
|
"learning_rate": 0.00015101241018941868, |
|
"loss": 0.278, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.7609756097560976, |
|
"grad_norm": 0.15299169719219208, |
|
"learning_rate": 0.00014970607446113652, |
|
"loss": 0.2877, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.7804878048780488, |
|
"grad_norm": 0.17317935824394226, |
|
"learning_rate": 0.00014839973873285436, |
|
"loss": 0.2888, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.17615154385566711, |
|
"learning_rate": 0.00014709340300457217, |
|
"loss": 0.298, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.8195121951219512, |
|
"grad_norm": 0.15918245911598206, |
|
"learning_rate": 0.00014578706727629, |
|
"loss": 0.3118, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.8390243902439024, |
|
"grad_norm": 0.1735045462846756, |
|
"learning_rate": 0.00014448073154800785, |
|
"loss": 0.2968, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.8585365853658536, |
|
"grad_norm": 0.16569170355796814, |
|
"learning_rate": 0.00014317439581972566, |
|
"loss": 0.2902, |
|
"step": 440 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 1536, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 20, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2.956730879299584e+16, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|