|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.0, |
|
"global_step": 26990, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.9073731011485735e-05, |
|
"loss": 1.4116, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.8147462022971475e-05, |
|
"loss": 1.4254, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.722119303445721e-05, |
|
"loss": 1.4358, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 4.629492404594295e-05, |
|
"loss": 1.4448, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.536865505742868e-05, |
|
"loss": 1.4453, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 4.444238606891441e-05, |
|
"loss": 1.4351, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.3516117080400145e-05, |
|
"loss": 1.4568, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.2589848091885885e-05, |
|
"loss": 1.4584, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.166357910337162e-05, |
|
"loss": 1.4641, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.073731011485736e-05, |
|
"loss": 1.4731, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 3.981104112634309e-05, |
|
"loss": 1.4893, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 3.888477213782883e-05, |
|
"loss": 1.4762, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 3.795850314931456e-05, |
|
"loss": 1.4931, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 3.70322341608003e-05, |
|
"loss": 1.5, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 3.6105965172286034e-05, |
|
"loss": 1.493, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.5179696183771774e-05, |
|
"loss": 1.4961, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.425342719525751e-05, |
|
"loss": 1.5016, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.332715820674324e-05, |
|
"loss": 1.5055, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.240088921822897e-05, |
|
"loss": 1.5153, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.147462022971471e-05, |
|
"loss": 1.5185, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 3.0548351241200445e-05, |
|
"loss": 1.527, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 2.9622082252686184e-05, |
|
"loss": 1.5339, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 2.8695813264171917e-05, |
|
"loss": 1.5337, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.776954427565765e-05, |
|
"loss": 1.5377, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.684327528714339e-05, |
|
"loss": 1.5431, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.5917006298629122e-05, |
|
"loss": 1.5574, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.4990737310114858e-05, |
|
"loss": 1.5592, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 2.4064468321600594e-05, |
|
"loss": 1.3231, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 2.313819933308633e-05, |
|
"loss": 1.3354, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 2.2211930344572066e-05, |
|
"loss": 1.3405, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 2.1285661356057803e-05, |
|
"loss": 1.3607, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 2.0359392367543535e-05, |
|
"loss": 1.3663, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.943312337902927e-05, |
|
"loss": 1.3626, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.8506854390515008e-05, |
|
"loss": 1.3191, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.7580585402000744e-05, |
|
"loss": 1.3267, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.6654316413486477e-05, |
|
"loss": 1.3336, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 1.5728047424972213e-05, |
|
"loss": 1.3735, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 1.4801778436457947e-05, |
|
"loss": 1.3832, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 1.3875509447943683e-05, |
|
"loss": 1.389, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 1.2949240459429418e-05, |
|
"loss": 1.4002, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 1.2022971470915156e-05, |
|
"loss": 1.4074, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 1.109670248240089e-05, |
|
"loss": 1.4229, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 1.0170433493886624e-05, |
|
"loss": 1.4368, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.24416450537236e-06, |
|
"loss": 1.4427, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 8.317895516858095e-06, |
|
"loss": 1.4402, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 7.391626528343831e-06, |
|
"loss": 1.4554, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 6.4653575398295665e-06, |
|
"loss": 1.4611, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 5.539088551315303e-06, |
|
"loss": 1.4801, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 4.612819562801038e-06, |
|
"loss": 1.4792, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 3.686550574286773e-06, |
|
"loss": 1.4974, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 2.7602815857725085e-06, |
|
"loss": 1.4973, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.8340125972582438e-06, |
|
"loss": 1.5153, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 9.077436087439793e-07, |
|
"loss": 1.5293, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"step": 26990, |
|
"total_flos": 2.5064145824199475e+17, |
|
"train_loss": 0.05643905973204775, |
|
"train_runtime": 943.9109, |
|
"train_samples_per_second": 285.921, |
|
"train_steps_per_second": 28.594 |
|
} |
|
], |
|
"max_steps": 26990, |
|
"num_train_epochs": 2, |
|
"total_flos": 2.5064145824199475e+17, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|