nash_dpo_doff_no_golden_iter_3 / trainer_state.json
YYYYYYibo's picture
Model save
8d6f77f verified
raw
history blame
9.34 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 100,
"global_step": 162,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.01,
"learning_rate": 2.9411764705882356e-07,
"logits/chosen": -2.640578269958496,
"logits/rejected": -2.6619861125946045,
"logps/chosen": -410.817138671875,
"logps/rejected": -784.9041137695312,
"loss": 0.6931,
"rewards/accuracies": 0.0,
"rewards/chosen": 0.0,
"rewards/margins": 0.0,
"rewards/rejected": 0.0,
"step": 1
},
{
"epoch": 0.06,
"learning_rate": 2.9411764705882355e-06,
"logits/chosen": -2.580946683883667,
"logits/rejected": -2.562483549118042,
"logps/chosen": -348.2420959472656,
"logps/rejected": -529.5997924804688,
"loss": 0.6857,
"rewards/accuracies": 0.5416666865348816,
"rewards/chosen": 0.047004420310258865,
"rewards/margins": 0.018747717142105103,
"rewards/rejected": 0.028256705030798912,
"step": 10
},
{
"epoch": 0.12,
"learning_rate": 4.994720857837211e-06,
"logits/chosen": -2.563664197921753,
"logits/rejected": -2.5549397468566895,
"logps/chosen": -303.5776062011719,
"logps/rejected": -509.2774353027344,
"loss": 0.6578,
"rewards/accuracies": 0.6625000238418579,
"rewards/chosen": 0.23815405368804932,
"rewards/margins": 0.07588066160678864,
"rewards/rejected": 0.1622733771800995,
"step": 20
},
{
"epoch": 0.19,
"learning_rate": 4.901488388458247e-06,
"logits/chosen": -2.526702404022217,
"logits/rejected": -2.5158963203430176,
"logps/chosen": -335.10333251953125,
"logps/rejected": -552.9954223632812,
"loss": 0.6058,
"rewards/accuracies": 0.7875000238418579,
"rewards/chosen": 0.2050032615661621,
"rewards/margins": 0.20939965546131134,
"rewards/rejected": -0.004396387841552496,
"step": 30
},
{
"epoch": 0.25,
"learning_rate": 4.6959649910976165e-06,
"logits/chosen": -2.532031536102295,
"logits/rejected": -2.4838037490844727,
"logps/chosen": -300.42132568359375,
"logps/rejected": -509.403564453125,
"loss": 0.5705,
"rewards/accuracies": 0.7875000238418579,
"rewards/chosen": 0.23074543476104736,
"rewards/margins": 0.2837901711463928,
"rewards/rejected": -0.053044695407152176,
"step": 40
},
{
"epoch": 0.31,
"learning_rate": 4.387760711393052e-06,
"logits/chosen": -2.5283217430114746,
"logits/rejected": -2.46999192237854,
"logps/chosen": -327.3263244628906,
"logps/rejected": -512.6666870117188,
"loss": 0.5487,
"rewards/accuracies": 0.800000011920929,
"rewards/chosen": 0.2256808578968048,
"rewards/margins": 0.36225491762161255,
"rewards/rejected": -0.13657405972480774,
"step": 50
},
{
"epoch": 0.37,
"learning_rate": 3.991286838919086e-06,
"logits/chosen": -2.508136034011841,
"logits/rejected": -2.471802234649658,
"logps/chosen": -333.85711669921875,
"logps/rejected": -523.9845581054688,
"loss": 0.5393,
"rewards/accuracies": 0.784375011920929,
"rewards/chosen": 0.1835666298866272,
"rewards/margins": 0.38937950134277344,
"rewards/rejected": -0.20581285655498505,
"step": 60
},
{
"epoch": 0.43,
"learning_rate": 3.5250820513035403e-06,
"logits/chosen": -2.492502450942993,
"logits/rejected": -2.484013080596924,
"logps/chosen": -314.8898620605469,
"logps/rejected": -589.8179931640625,
"loss": 0.5165,
"rewards/accuracies": 0.84375,
"rewards/chosen": 0.19387896358966827,
"rewards/margins": 0.5265167951583862,
"rewards/rejected": -0.33263787627220154,
"step": 70
},
{
"epoch": 0.49,
"learning_rate": 3.0109455662659126e-06,
"logits/chosen": -2.527844190597534,
"logits/rejected": -2.502004623413086,
"logps/chosen": -320.5412292480469,
"logps/rejected": -579.2618408203125,
"loss": 0.5083,
"rewards/accuracies": 0.815625011920929,
"rewards/chosen": 0.12712647020816803,
"rewards/margins": 0.5316451787948608,
"rewards/rejected": -0.4045187532901764,
"step": 80
},
{
"epoch": 0.56,
"learning_rate": 2.4729178344249007e-06,
"logits/chosen": -2.5523924827575684,
"logits/rejected": -2.5245110988616943,
"logps/chosen": -341.2242431640625,
"logps/rejected": -596.2203369140625,
"loss": 0.496,
"rewards/accuracies": 0.8125,
"rewards/chosen": 0.06856798380613327,
"rewards/margins": 0.590112030506134,
"rewards/rejected": -0.5215439796447754,
"step": 90
},
{
"epoch": 0.62,
"learning_rate": 1.936156434546515e-06,
"logits/chosen": -2.5596001148223877,
"logits/rejected": -2.568671703338623,
"logps/chosen": -333.087158203125,
"logps/rejected": -611.3602905273438,
"loss": 0.4768,
"rewards/accuracies": 0.8187500238418579,
"rewards/chosen": 0.12124671787023544,
"rewards/margins": 0.6403753757476807,
"rewards/rejected": -0.5191286206245422,
"step": 100
},
{
"epoch": 0.62,
"eval_logits/chosen": -2.585242509841919,
"eval_logits/rejected": -2.479508399963379,
"eval_logps/chosen": -293.3752136230469,
"eval_logps/rejected": -292.1757507324219,
"eval_loss": 0.6442223191261292,
"eval_rewards/accuracies": 0.5659999847412109,
"eval_rewards/chosen": 0.18966493010520935,
"eval_rewards/margins": 0.15400375425815582,
"eval_rewards/rejected": 0.035661179572343826,
"eval_runtime": 398.0482,
"eval_samples_per_second": 5.025,
"eval_steps_per_second": 0.628,
"step": 100
},
{
"epoch": 0.68,
"learning_rate": 1.4257597331216211e-06,
"logits/chosen": -2.5242629051208496,
"logits/rejected": -2.484691619873047,
"logps/chosen": -325.24761962890625,
"logps/rejected": -541.4208984375,
"loss": 0.4856,
"rewards/accuracies": 0.8031250238418579,
"rewards/chosen": 0.11120424419641495,
"rewards/margins": 0.6161486506462097,
"rewards/rejected": -0.5049443244934082,
"step": 110
},
{
"epoch": 0.74,
"learning_rate": 9.655933126436565e-07,
"logits/chosen": -2.5429892539978027,
"logits/rejected": -2.520012855529785,
"logps/chosen": -329.37347412109375,
"logps/rejected": -596.545654296875,
"loss": 0.4791,
"rewards/accuracies": 0.800000011920929,
"rewards/chosen": 0.06060720607638359,
"rewards/margins": 0.6603037714958191,
"rewards/rejected": -0.5996966361999512,
"step": 120
},
{
"epoch": 0.8,
"learning_rate": 5.771740434959278e-07,
"logits/chosen": -2.5939841270446777,
"logits/rejected": -2.543733835220337,
"logps/chosen": -326.51824951171875,
"logps/rejected": -585.2608032226562,
"loss": 0.4705,
"rewards/accuracies": 0.831250011920929,
"rewards/chosen": 0.0589924156665802,
"rewards/margins": 0.7284678220748901,
"rewards/rejected": -0.6694754362106323,
"step": 130
},
{
"epoch": 0.86,
"learning_rate": 2.786639790067719e-07,
"logits/chosen": -2.538792848587036,
"logits/rejected": -2.572946548461914,
"logps/chosen": -327.40155029296875,
"logps/rejected": -593.7877197265625,
"loss": 0.4828,
"rewards/accuracies": 0.828125,
"rewards/chosen": 0.006606454961001873,
"rewards/margins": 0.6702617406845093,
"rewards/rejected": -0.6636553406715393,
"step": 140
},
{
"epoch": 0.93,
"learning_rate": 8.402111802159413e-08,
"logits/chosen": -2.60109281539917,
"logits/rejected": -2.500326633453369,
"logps/chosen": -350.4434814453125,
"logps/rejected": -545.3392333984375,
"loss": 0.4729,
"rewards/accuracies": 0.800000011920929,
"rewards/chosen": 0.03271277993917465,
"rewards/margins": 0.6479501128196716,
"rewards/rejected": -0.615237295627594,
"step": 150
},
{
"epoch": 0.99,
"learning_rate": 2.34674439005822e-09,
"logits/chosen": -2.6234354972839355,
"logits/rejected": -2.563753366470337,
"logps/chosen": -370.4894104003906,
"logps/rejected": -656.5492553710938,
"loss": 0.4684,
"rewards/accuracies": 0.784375011920929,
"rewards/chosen": 0.009518811479210854,
"rewards/margins": 0.7224863171577454,
"rewards/rejected": -0.7129674553871155,
"step": 160
},
{
"epoch": 1.0,
"step": 162,
"total_flos": 0.0,
"train_loss": 0.5280830058050744,
"train_runtime": 8024.3189,
"train_samples_per_second": 2.584,
"train_steps_per_second": 0.02
}
],
"logging_steps": 10,
"max_steps": 162,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"total_flos": 0.0,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}