lawbuddy-7b / trainer_state.json
monshinawatra's picture
Upload folder using huggingface_hub
6adc7f5 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.986666666666667,
"eval_steps": 500,
"global_step": 336,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.08888888888888889,
"grad_norm": 1.792281150817871,
"kl": 0.15646514296531677,
"learning_rate": 1.4705882352941177e-06,
"logps/chosen": -164.1169189453125,
"loss": 0.5043,
"rewards/chosen": -0.0013721466064453125,
"step": 10
},
{
"epoch": 0.17777777777777778,
"grad_norm": 1.7645841836929321,
"kl": 0.23952817916870117,
"learning_rate": 2.9411764705882355e-06,
"logps/chosen": -161.21431884765624,
"loss": 0.506,
"rewards/chosen": -0.000222191633656621,
"step": 20
},
{
"epoch": 0.26666666666666666,
"grad_norm": 1.6355303525924683,
"kl": 0.2789602279663086,
"learning_rate": 4.411764705882353e-06,
"logps/chosen": -161.8250244140625,
"loss": 0.5046,
"rewards/chosen": 0.009497338533401489,
"step": 30
},
{
"epoch": 0.35555555555555557,
"grad_norm": 1.8716518878936768,
"kl": 0.5364601016044617,
"learning_rate": 4.995131923687488e-06,
"logps/chosen": -160.8272705078125,
"loss": 0.5011,
"rewards/chosen": 0.04936442375183105,
"step": 40
},
{
"epoch": 0.4444444444444444,
"grad_norm": 1.7810463905334473,
"kl": 1.5138229131698608,
"learning_rate": 4.965451197130373e-06,
"logps/chosen": -159.2620361328125,
"loss": 0.5081,
"rewards/chosen": 0.11891223192214966,
"step": 50
},
{
"epoch": 0.5333333333333333,
"grad_norm": 2.2072041034698486,
"kl": 3.1553397178649902,
"learning_rate": 4.90911473983908e-06,
"logps/chosen": -163.2381103515625,
"loss": 0.5049,
"rewards/chosen": 0.295930004119873,
"step": 60
},
{
"epoch": 0.6222222222222222,
"grad_norm": 1.858336329460144,
"kl": 6.15082311630249,
"learning_rate": 4.826731644963705e-06,
"logps/chosen": -160.2790771484375,
"loss": 0.5238,
"rewards/chosen": 0.517311429977417,
"step": 70
},
{
"epoch": 0.7111111111111111,
"grad_norm": 2.1484944820404053,
"kl": 9.094244003295898,
"learning_rate": 4.71919261421297e-06,
"logps/chosen": -165.65592041015626,
"loss": 0.4949,
"rewards/chosen": 0.9306423187255859,
"step": 80
},
{
"epoch": 0.8,
"grad_norm": 1.9223041534423828,
"kl": 12.533430099487305,
"learning_rate": 4.587660327850203e-06,
"logps/chosen": -147.84144287109376,
"loss": 0.5327,
"rewards/chosen": 1.1160342216491699,
"step": 90
},
{
"epoch": 0.8888888888888888,
"grad_norm": 2.052889823913574,
"kl": 15.455533981323242,
"learning_rate": 4.43355687413747e-06,
"logps/chosen": -141.32998046875,
"loss": 0.5258,
"rewards/chosen": 1.4277078628540039,
"step": 100
},
{
"epoch": 0.9777777777777777,
"grad_norm": 1.8008378744125366,
"kl": 18.05435562133789,
"learning_rate": 4.258548374136976e-06,
"logps/chosen": -142.55394287109374,
"loss": 0.5491,
"rewards/chosen": 1.5885238647460938,
"step": 110
},
{
"epoch": 1.0666666666666667,
"grad_norm": 1.9117740392684937,
"kl": 20.216676712036133,
"learning_rate": 4.064526968101844e-06,
"logps/chosen": -138.30635986328124,
"loss": 0.5418,
"rewards/chosen": 1.8262062072753906,
"step": 120
},
{
"epoch": 1.1555555555555554,
"grad_norm": 2.0452499389648438,
"kl": 22.620920181274414,
"learning_rate": 3.853590358214119e-06,
"logps/chosen": -142.5436767578125,
"loss": 0.5414,
"rewards/chosen": 2.02716178894043,
"step": 130
},
{
"epoch": 1.2444444444444445,
"grad_norm": 1.5861475467681885,
"kl": 21.474929809570312,
"learning_rate": 3.6280191288478437e-06,
"logps/chosen": -139.66346435546876,
"loss": 0.5436,
"rewards/chosen": 1.9418670654296875,
"step": 140
},
{
"epoch": 1.3333333333333333,
"grad_norm": 2.040761947631836,
"kl": 24.884174346923828,
"learning_rate": 3.3902520895638674e-06,
"logps/chosen": -144.4580078125,
"loss": 0.5551,
"rewards/chosen": 2.2067359924316405,
"step": 150
},
{
"epoch": 1.4222222222222223,
"grad_norm": 1.4137938022613525,
"kl": 24.920988082885742,
"learning_rate": 3.142859907420615e-06,
"logps/chosen": -134.41856689453124,
"loss": 0.5452,
"rewards/chosen": 2.2827281951904297,
"step": 160
},
{
"epoch": 1.511111111111111,
"grad_norm": 1.9247097969055176,
"kl": 27.40714454650879,
"learning_rate": 2.8885173136805126e-06,
"logps/chosen": -134.56304931640625,
"loss": 0.5554,
"rewards/chosen": 2.478318786621094,
"step": 170
},
{
"epoch": 1.6,
"grad_norm": 1.7178384065628052,
"kl": 26.36468505859375,
"learning_rate": 2.629974185404951e-06,
"logps/chosen": -140.7825927734375,
"loss": 0.5482,
"rewards/chosen": 2.3276464462280275,
"step": 180
},
{
"epoch": 1.6888888888888889,
"grad_norm": 1.899584174156189,
"kl": 27.777027130126953,
"learning_rate": 2.3700258145950495e-06,
"logps/chosen": -140.53466796875,
"loss": 0.5618,
"rewards/chosen": 2.424651336669922,
"step": 190
},
{
"epoch": 1.7777777777777777,
"grad_norm": 1.7362196445465088,
"kl": 27.96476173400879,
"learning_rate": 2.1114826863194882e-06,
"logps/chosen": -132.751318359375,
"loss": 0.5549,
"rewards/chosen": 2.5185680389404297,
"step": 200
},
{
"epoch": 1.8666666666666667,
"grad_norm": 1.5568628311157227,
"kl": 25.686458587646484,
"learning_rate": 1.8571400925793855e-06,
"logps/chosen": -141.460986328125,
"loss": 0.4799,
"rewards/chosen": 2.678207778930664,
"step": 210
},
{
"epoch": 1.9555555555555557,
"grad_norm": 1.4741252660751343,
"kl": 27.350994110107422,
"learning_rate": 1.6097479104361328e-06,
"logps/chosen": -138.36417236328126,
"loss": 0.5164,
"rewards/chosen": 2.5984352111816404,
"step": 220
},
{
"epoch": 2.0444444444444443,
"grad_norm": 1.4206268787384033,
"kl": 27.860248565673828,
"learning_rate": 1.3719808711521573e-06,
"logps/chosen": -140.159765625,
"loss": 0.5315,
"rewards/chosen": 2.6533029556274412,
"step": 230
},
{
"epoch": 2.1333333333333333,
"grad_norm": 1.964991569519043,
"kl": 28.185623168945312,
"learning_rate": 1.1464096417858821e-06,
"logps/chosen": -136.175439453125,
"loss": 0.5194,
"rewards/chosen": 2.697865676879883,
"step": 240
},
{
"epoch": 2.2222222222222223,
"grad_norm": 2.0303757190704346,
"kl": 29.453914642333984,
"learning_rate": 9.354730318981561e-07,
"logps/chosen": -124.939111328125,
"loss": 0.5565,
"rewards/chosen": 2.6518789291381837,
"step": 250
},
{
"epoch": 2.311111111111111,
"grad_norm": 1.9753375053405762,
"kl": 28.224218368530273,
"learning_rate": 7.414516258630245e-07,
"logps/chosen": -143.79691162109376,
"loss": 0.4645,
"rewards/chosen": 2.974603271484375,
"step": 260
},
{
"epoch": 2.4,
"grad_norm": 1.5573641061782837,
"kl": 28.22760581970215,
"learning_rate": 5.664431258625305e-07,
"logps/chosen": -135.832470703125,
"loss": 0.5039,
"rewards/chosen": 2.787263107299805,
"step": 270
},
{
"epoch": 2.488888888888889,
"grad_norm": 1.8748925924301147,
"kl": 29.364704132080078,
"learning_rate": 4.123396721497977e-07,
"logps/chosen": -132.555029296875,
"loss": 0.5659,
"rewards/chosen": 2.5855674743652344,
"step": 280
},
{
"epoch": 2.5777777777777775,
"grad_norm": 1.9802337884902954,
"kl": 29.822818756103516,
"learning_rate": 2.8080738578703054e-07,
"logps/chosen": -132.623876953125,
"loss": 0.5254,
"rewards/chosen": 2.7944366455078127,
"step": 290
},
{
"epoch": 2.6666666666666665,
"grad_norm": 1.6388640403747559,
"kl": 29.55389404296875,
"learning_rate": 1.7326835503629542e-07,
"logps/chosen": -132.94588623046874,
"loss": 0.5661,
"rewards/chosen": 2.6015438079833983,
"step": 300
},
{
"epoch": 2.7555555555555555,
"grad_norm": 2.059163808822632,
"kl": 28.346817016601562,
"learning_rate": 9.088526016092142e-08,
"logps/chosen": -142.41676025390626,
"loss": 0.4982,
"rewards/chosen": 2.8422063827514648,
"step": 310
},
{
"epoch": 2.8444444444444446,
"grad_norm": 1.8374332189559937,
"kl": 28.879934310913086,
"learning_rate": 3.4548802869627806e-08,
"logps/chosen": -133.9089599609375,
"loss": 0.508,
"rewards/chosen": 2.8151947021484376,
"step": 320
},
{
"epoch": 2.9333333333333336,
"grad_norm": 1.6760042905807495,
"kl": 28.577600479125977,
"learning_rate": 4.868076312512515e-09,
"logps/chosen": -133.6552001953125,
"loss": 0.5269,
"rewards/chosen": 2.694635200500488,
"step": 330
},
{
"epoch": 2.986666666666667,
"step": 336,
"total_flos": 3.5905450080116736e+16,
"train_loss": 0.5264021051781518,
"train_runtime": 1249.11,
"train_samples_per_second": 2.162,
"train_steps_per_second": 0.269
}
],
"logging_steps": 10,
"max_steps": 336,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 3.5905450080116736e+16,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}