|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.0, |
|
"eval_steps": 500, |
|
"global_step": 930, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0032258064516129032, |
|
"grad_norm": 115.22052507274148, |
|
"learning_rate": 5.3763440860215056e-09, |
|
"logits/chosen": -4.125, |
|
"logits/rejected": -4.0625, |
|
"logps/chosen": -185.0, |
|
"logps/rejected": -196.0, |
|
"loss": 0.6914, |
|
"rewards/accuracies": 0.0, |
|
"rewards/chosen": 0.0, |
|
"rewards/margins": 0.0, |
|
"rewards/rejected": 0.0, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.03225806451612903, |
|
"grad_norm": 94.47396535248316, |
|
"learning_rate": 5.3763440860215054e-08, |
|
"logits/chosen": -4.125, |
|
"logits/rejected": -4.15625, |
|
"logps/chosen": -186.0, |
|
"logps/rejected": -195.0, |
|
"loss": 0.6886, |
|
"rewards/accuracies": 0.2222222238779068, |
|
"rewards/chosen": 1.3589859008789062e-05, |
|
"rewards/margins": 0.009765625, |
|
"rewards/rejected": -0.009765625, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.06451612903225806, |
|
"grad_norm": 94.33148772179948, |
|
"learning_rate": 1.0752688172043011e-07, |
|
"logits/chosen": -4.15625, |
|
"logits/rejected": -4.0625, |
|
"logps/chosen": -188.0, |
|
"logps/rejected": -200.0, |
|
"loss": 0.6872, |
|
"rewards/accuracies": 0.2874999940395355, |
|
"rewards/chosen": -0.005645751953125, |
|
"rewards/margins": 0.01312255859375, |
|
"rewards/rejected": -0.018798828125, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0967741935483871, |
|
"grad_norm": 92.49633755921806, |
|
"learning_rate": 1.6129032258064515e-07, |
|
"logits/chosen": -4.0625, |
|
"logits/rejected": -4.03125, |
|
"logps/chosen": -185.0, |
|
"logps/rejected": -191.0, |
|
"loss": 0.669, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.016845703125, |
|
"rewards/margins": 0.054443359375, |
|
"rewards/rejected": -0.03759765625, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.12903225806451613, |
|
"grad_norm": 82.30619665044928, |
|
"learning_rate": 2.1505376344086022e-07, |
|
"logits/chosen": -4.03125, |
|
"logits/rejected": -4.03125, |
|
"logps/chosen": -189.0, |
|
"logps/rejected": -200.0, |
|
"loss": 0.6341, |
|
"rewards/accuracies": 0.6625000238418579, |
|
"rewards/chosen": 0.038818359375, |
|
"rewards/margins": 0.13671875, |
|
"rewards/rejected": -0.09765625, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.16129032258064516, |
|
"grad_norm": 73.60890561659406, |
|
"learning_rate": 2.6881720430107523e-07, |
|
"logits/chosen": -4.03125, |
|
"logits/rejected": -3.890625, |
|
"logps/chosen": -189.0, |
|
"logps/rejected": -201.0, |
|
"loss": 0.5644, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.029296875, |
|
"rewards/margins": 0.328125, |
|
"rewards/rejected": -0.357421875, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.1935483870967742, |
|
"grad_norm": 83.24582572373353, |
|
"learning_rate": 3.225806451612903e-07, |
|
"logits/chosen": -3.890625, |
|
"logits/rejected": -3.828125, |
|
"logps/chosen": -191.0, |
|
"logps/rejected": -204.0, |
|
"loss": 0.5283, |
|
"rewards/accuracies": 0.675000011920929, |
|
"rewards/chosen": -0.25390625, |
|
"rewards/margins": 0.484375, |
|
"rewards/rejected": -0.73828125, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.22580645161290322, |
|
"grad_norm": 70.93826710868849, |
|
"learning_rate": 3.7634408602150537e-07, |
|
"logits/chosen": -3.9375, |
|
"logits/rejected": -3.890625, |
|
"logps/chosen": -187.0, |
|
"logps/rejected": -206.0, |
|
"loss": 0.4758, |
|
"rewards/accuracies": 0.887499988079071, |
|
"rewards/chosen": -0.1494140625, |
|
"rewards/margins": 0.83984375, |
|
"rewards/rejected": -0.98828125, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.25806451612903225, |
|
"grad_norm": 67.89432749815619, |
|
"learning_rate": 4.3010752688172043e-07, |
|
"logits/chosen": -3.875, |
|
"logits/rejected": -3.84375, |
|
"logps/chosen": -188.0, |
|
"logps/rejected": -215.0, |
|
"loss": 0.4225, |
|
"rewards/accuracies": 0.824999988079071, |
|
"rewards/chosen": -0.06494140625, |
|
"rewards/margins": 1.3359375, |
|
"rewards/rejected": -1.3984375, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.2903225806451613, |
|
"grad_norm": 56.57247809818219, |
|
"learning_rate": 4.838709677419355e-07, |
|
"logits/chosen": -3.9375, |
|
"logits/rejected": -3.921875, |
|
"logps/chosen": -195.0, |
|
"logps/rejected": -220.0, |
|
"loss": 0.3769, |
|
"rewards/accuracies": 0.862500011920929, |
|
"rewards/chosen": -0.8125, |
|
"rewards/margins": 1.390625, |
|
"rewards/rejected": -2.203125, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.3225806451612903, |
|
"grad_norm": 54.52791585727774, |
|
"learning_rate": 4.958183990442055e-07, |
|
"logits/chosen": -3.90625, |
|
"logits/rejected": -3.890625, |
|
"logps/chosen": -198.0, |
|
"logps/rejected": -225.0, |
|
"loss": 0.3733, |
|
"rewards/accuracies": 0.7749999761581421, |
|
"rewards/chosen": -1.03125, |
|
"rewards/margins": 1.6015625, |
|
"rewards/rejected": -2.625, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.3548387096774194, |
|
"grad_norm": 70.14660020921133, |
|
"learning_rate": 4.898446833930704e-07, |
|
"logits/chosen": -3.96875, |
|
"logits/rejected": -3.90625, |
|
"logps/chosen": -195.0, |
|
"logps/rejected": -225.0, |
|
"loss": 0.3651, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": -1.0390625, |
|
"rewards/margins": 1.7578125, |
|
"rewards/rejected": -2.796875, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.3870967741935484, |
|
"grad_norm": 70.71484436398696, |
|
"learning_rate": 4.838709677419355e-07, |
|
"logits/chosen": -3.9375, |
|
"logits/rejected": -3.90625, |
|
"logps/chosen": -196.0, |
|
"logps/rejected": -224.0, |
|
"loss": 0.4252, |
|
"rewards/accuracies": 0.8500000238418579, |
|
"rewards/chosen": -0.9609375, |
|
"rewards/margins": 1.7421875, |
|
"rewards/rejected": -2.703125, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.41935483870967744, |
|
"grad_norm": 60.790615448301835, |
|
"learning_rate": 4.778972520908004e-07, |
|
"logits/chosen": -3.9375, |
|
"logits/rejected": -3.9375, |
|
"logps/chosen": -197.0, |
|
"logps/rejected": -231.0, |
|
"loss": 0.3266, |
|
"rewards/accuracies": 0.9125000238418579, |
|
"rewards/chosen": -0.95703125, |
|
"rewards/margins": 2.21875, |
|
"rewards/rejected": -3.1875, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.45161290322580644, |
|
"grad_norm": 69.89033127044969, |
|
"learning_rate": 4.7192353643966544e-07, |
|
"logits/chosen": -3.9375, |
|
"logits/rejected": -3.890625, |
|
"logps/chosen": -201.0, |
|
"logps/rejected": -230.0, |
|
"loss": 0.3487, |
|
"rewards/accuracies": 0.8999999761581421, |
|
"rewards/chosen": -1.171875, |
|
"rewards/margins": 1.8359375, |
|
"rewards/rejected": -3.0, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.4838709677419355, |
|
"grad_norm": 70.95003315189264, |
|
"learning_rate": 4.6594982078853044e-07, |
|
"logits/chosen": -3.90625, |
|
"logits/rejected": -3.828125, |
|
"logps/chosen": -196.0, |
|
"logps/rejected": -229.0, |
|
"loss": 0.3011, |
|
"rewards/accuracies": 0.824999988079071, |
|
"rewards/chosen": -1.140625, |
|
"rewards/margins": 2.078125, |
|
"rewards/rejected": -3.21875, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.5161290322580645, |
|
"grad_norm": 78.12294251323976, |
|
"learning_rate": 4.5997610513739544e-07, |
|
"logits/chosen": -3.90625, |
|
"logits/rejected": -3.859375, |
|
"logps/chosen": -194.0, |
|
"logps/rejected": -229.0, |
|
"loss": 0.2904, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": -0.80859375, |
|
"rewards/margins": 2.453125, |
|
"rewards/rejected": -3.265625, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.5483870967741935, |
|
"grad_norm": 58.17040697560013, |
|
"learning_rate": 4.540023894862604e-07, |
|
"logits/chosen": -3.984375, |
|
"logits/rejected": -3.890625, |
|
"logps/chosen": -202.0, |
|
"logps/rejected": -232.0, |
|
"loss": 0.286, |
|
"rewards/accuracies": 0.8374999761581421, |
|
"rewards/chosen": -1.53125, |
|
"rewards/margins": 2.484375, |
|
"rewards/rejected": -4.0, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.5806451612903226, |
|
"grad_norm": 75.00483410106833, |
|
"learning_rate": 4.4802867383512544e-07, |
|
"logits/chosen": -3.8125, |
|
"logits/rejected": -3.765625, |
|
"logps/chosen": -200.0, |
|
"logps/rejected": -239.0, |
|
"loss": 0.3374, |
|
"rewards/accuracies": 0.8125, |
|
"rewards/chosen": -1.5625, |
|
"rewards/margins": 2.484375, |
|
"rewards/rejected": -4.03125, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.6129032258064516, |
|
"grad_norm": 72.167130272584, |
|
"learning_rate": 4.4205495818399044e-07, |
|
"logits/chosen": -3.90625, |
|
"logits/rejected": -3.875, |
|
"logps/chosen": -200.0, |
|
"logps/rejected": -239.0, |
|
"loss": 0.3004, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": -1.0703125, |
|
"rewards/margins": 2.890625, |
|
"rewards/rejected": -3.96875, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.6451612903225806, |
|
"grad_norm": 61.76892460773258, |
|
"learning_rate": 4.3608124253285543e-07, |
|
"logits/chosen": -3.90625, |
|
"logits/rejected": -3.90625, |
|
"logps/chosen": -199.0, |
|
"logps/rejected": -230.0, |
|
"loss": 0.301, |
|
"rewards/accuracies": 0.800000011920929, |
|
"rewards/chosen": -1.375, |
|
"rewards/margins": 2.015625, |
|
"rewards/rejected": -3.390625, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.6774193548387096, |
|
"grad_norm": 52.49765267963524, |
|
"learning_rate": 4.3010752688172043e-07, |
|
"logits/chosen": -3.890625, |
|
"logits/rejected": -3.78125, |
|
"logps/chosen": -196.0, |
|
"logps/rejected": -233.0, |
|
"loss": 0.258, |
|
"rewards/accuracies": 0.887499988079071, |
|
"rewards/chosen": -0.8828125, |
|
"rewards/margins": 2.84375, |
|
"rewards/rejected": -3.734375, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.7096774193548387, |
|
"grad_norm": 58.816815163007185, |
|
"learning_rate": 4.241338112305854e-07, |
|
"logits/chosen": -3.859375, |
|
"logits/rejected": -3.84375, |
|
"logps/chosen": -194.0, |
|
"logps/rejected": -237.0, |
|
"loss": 0.278, |
|
"rewards/accuracies": 0.824999988079071, |
|
"rewards/chosen": -1.0390625, |
|
"rewards/margins": 2.6875, |
|
"rewards/rejected": -3.734375, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.7419354838709677, |
|
"grad_norm": 77.22842288731327, |
|
"learning_rate": 4.1816009557945043e-07, |
|
"logits/chosen": -3.859375, |
|
"logits/rejected": -3.734375, |
|
"logps/chosen": -192.0, |
|
"logps/rejected": -234.0, |
|
"loss": 0.2537, |
|
"rewards/accuracies": 0.949999988079071, |
|
"rewards/chosen": -0.7890625, |
|
"rewards/margins": 3.046875, |
|
"rewards/rejected": -3.84375, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.7741935483870968, |
|
"grad_norm": 55.73484353697734, |
|
"learning_rate": 4.121863799283154e-07, |
|
"logits/chosen": -3.875, |
|
"logits/rejected": -3.796875, |
|
"logps/chosen": -199.0, |
|
"logps/rejected": -235.0, |
|
"loss": 0.3267, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": -1.5078125, |
|
"rewards/margins": 2.8125, |
|
"rewards/rejected": -4.3125, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.8064516129032258, |
|
"grad_norm": 73.3120879364035, |
|
"learning_rate": 4.0621266427718037e-07, |
|
"logits/chosen": -3.953125, |
|
"logits/rejected": -3.890625, |
|
"logps/chosen": -194.0, |
|
"logps/rejected": -233.0, |
|
"loss": 0.2759, |
|
"rewards/accuracies": 0.887499988079071, |
|
"rewards/chosen": -0.85546875, |
|
"rewards/margins": 2.96875, |
|
"rewards/rejected": -3.8125, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.8387096774193549, |
|
"grad_norm": 53.82592061402359, |
|
"learning_rate": 4.002389486260454e-07, |
|
"logits/chosen": -3.953125, |
|
"logits/rejected": -3.890625, |
|
"logps/chosen": -195.0, |
|
"logps/rejected": -245.0, |
|
"loss": 0.2255, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": -0.59765625, |
|
"rewards/margins": 3.875, |
|
"rewards/rejected": -4.46875, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.8709677419354839, |
|
"grad_norm": 73.21099669158846, |
|
"learning_rate": 3.9426523297491037e-07, |
|
"logits/chosen": -3.953125, |
|
"logits/rejected": -3.921875, |
|
"logps/chosen": -195.0, |
|
"logps/rejected": -244.0, |
|
"loss": 0.2336, |
|
"rewards/accuracies": 0.949999988079071, |
|
"rewards/chosen": -1.0703125, |
|
"rewards/margins": 3.53125, |
|
"rewards/rejected": -4.59375, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.9032258064516129, |
|
"grad_norm": 61.8854330576029, |
|
"learning_rate": 3.8829151732377537e-07, |
|
"logits/chosen": -3.9375, |
|
"logits/rejected": -3.90625, |
|
"logps/chosen": -199.0, |
|
"logps/rejected": -232.0, |
|
"loss": 0.2834, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": -1.15625, |
|
"rewards/margins": 2.890625, |
|
"rewards/rejected": -4.03125, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.9354838709677419, |
|
"grad_norm": 58.02390241120245, |
|
"learning_rate": 3.8231780167264037e-07, |
|
"logits/chosen": -4.03125, |
|
"logits/rejected": -3.9375, |
|
"logps/chosen": -196.0, |
|
"logps/rejected": -230.0, |
|
"loss": 0.2844, |
|
"rewards/accuracies": 0.8374999761581421, |
|
"rewards/chosen": -0.7734375, |
|
"rewards/margins": 2.9375, |
|
"rewards/rejected": -3.71875, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.967741935483871, |
|
"grad_norm": 78.57587657653356, |
|
"learning_rate": 3.7634408602150537e-07, |
|
"logits/chosen": -3.984375, |
|
"logits/rejected": -3.875, |
|
"logps/chosen": -189.0, |
|
"logps/rejected": -236.0, |
|
"loss": 0.2828, |
|
"rewards/accuracies": 0.925000011920929, |
|
"rewards/chosen": -0.6171875, |
|
"rewards/margins": 3.234375, |
|
"rewards/rejected": -3.859375, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 67.82380071078845, |
|
"learning_rate": 3.703703703703703e-07, |
|
"logits/chosen": -3.890625, |
|
"logits/rejected": -3.859375, |
|
"logps/chosen": -196.0, |
|
"logps/rejected": -230.0, |
|
"loss": 0.271, |
|
"rewards/accuracies": 0.887499988079071, |
|
"rewards/chosen": -0.9921875, |
|
"rewards/margins": 3.0625, |
|
"rewards/rejected": -4.0625, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_logits/chosen": -3.953125, |
|
"eval_logits/rejected": -3.890625, |
|
"eval_logps/chosen": -217.0, |
|
"eval_logps/rejected": -255.0, |
|
"eval_loss": 0.20095214247703552, |
|
"eval_rewards/accuracies": 0.9285714030265808, |
|
"eval_rewards/chosen": -1.7890625, |
|
"eval_rewards/margins": 3.21875, |
|
"eval_rewards/rejected": -5.0, |
|
"eval_runtime": 13.4206, |
|
"eval_samples_per_second": 14.902, |
|
"eval_steps_per_second": 0.522, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.032258064516129, |
|
"grad_norm": 14.7763712358255, |
|
"learning_rate": 3.6439665471923536e-07, |
|
"logits/chosen": -4.03125, |
|
"logits/rejected": -3.90625, |
|
"logps/chosen": -197.0, |
|
"logps/rejected": -268.0, |
|
"loss": 0.032, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -1.0234375, |
|
"rewards/margins": 5.96875, |
|
"rewards/rejected": -7.0, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.064516129032258, |
|
"grad_norm": 6.092608827300573, |
|
"learning_rate": 3.5842293906810036e-07, |
|
"logits/chosen": -3.9375, |
|
"logits/rejected": -3.859375, |
|
"logps/chosen": -193.0, |
|
"logps/rejected": -268.0, |
|
"loss": 0.0204, |
|
"rewards/accuracies": 0.9750000238418579, |
|
"rewards/chosen": -0.349609375, |
|
"rewards/margins": 6.0, |
|
"rewards/rejected": -6.34375, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.096774193548387, |
|
"grad_norm": 13.413597095885587, |
|
"learning_rate": 3.524492234169653e-07, |
|
"logits/chosen": -3.859375, |
|
"logits/rejected": -3.75, |
|
"logps/chosen": -192.0, |
|
"logps/rejected": -270.0, |
|
"loss": 0.0397, |
|
"rewards/accuracies": 0.9750000238418579, |
|
"rewards/chosen": -0.123046875, |
|
"rewards/margins": 6.40625, |
|
"rewards/rejected": -6.53125, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.129032258064516, |
|
"grad_norm": 23.482097218429256, |
|
"learning_rate": 3.4647550776583036e-07, |
|
"logits/chosen": -3.875, |
|
"logits/rejected": -3.84375, |
|
"logps/chosen": -193.0, |
|
"logps/rejected": -272.0, |
|
"loss": 0.0236, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -0.376953125, |
|
"rewards/margins": 7.0, |
|
"rewards/rejected": -7.375, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.1612903225806452, |
|
"grad_norm": 23.358340694800074, |
|
"learning_rate": 3.405017921146953e-07, |
|
"logits/chosen": -3.828125, |
|
"logits/rejected": -3.8125, |
|
"logps/chosen": -183.0, |
|
"logps/rejected": -260.0, |
|
"loss": 0.0311, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": 0.154296875, |
|
"rewards/margins": 6.3125, |
|
"rewards/rejected": -6.1875, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.1935483870967742, |
|
"grad_norm": 3.914331549525913, |
|
"learning_rate": 3.345280764635603e-07, |
|
"logits/chosen": -3.890625, |
|
"logits/rejected": -3.8125, |
|
"logps/chosen": -193.0, |
|
"logps/rejected": -274.0, |
|
"loss": 0.0253, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.0299072265625, |
|
"rewards/margins": 7.03125, |
|
"rewards/rejected": -7.0625, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.2258064516129032, |
|
"grad_norm": 7.274490128378863, |
|
"learning_rate": 3.285543608124253e-07, |
|
"logits/chosen": -3.96875, |
|
"logits/rejected": -3.84375, |
|
"logps/chosen": -190.0, |
|
"logps/rejected": -270.0, |
|
"loss": 0.0083, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.08349609375, |
|
"rewards/margins": 7.3125, |
|
"rewards/rejected": -7.21875, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.2580645161290323, |
|
"grad_norm": 3.5384348394162375, |
|
"learning_rate": 3.225806451612903e-07, |
|
"logits/chosen": -3.8125, |
|
"logits/rejected": -3.828125, |
|
"logps/chosen": -189.0, |
|
"logps/rejected": -276.0, |
|
"loss": 0.0157, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -0.119140625, |
|
"rewards/margins": 7.09375, |
|
"rewards/rejected": -7.1875, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.2903225806451613, |
|
"grad_norm": 27.375764382001474, |
|
"learning_rate": 3.1660692951015535e-07, |
|
"logits/chosen": -3.75, |
|
"logits/rejected": -3.78125, |
|
"logps/chosen": -194.0, |
|
"logps/rejected": -274.0, |
|
"loss": 0.0192, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.42578125, |
|
"rewards/margins": 7.4375, |
|
"rewards/rejected": -7.875, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.3225806451612903, |
|
"grad_norm": 21.54322951685049, |
|
"learning_rate": 3.106332138590203e-07, |
|
"logits/chosen": -3.859375, |
|
"logits/rejected": -3.859375, |
|
"logps/chosen": -186.0, |
|
"logps/rejected": -274.0, |
|
"loss": 0.0143, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.150390625, |
|
"rewards/margins": 8.0, |
|
"rewards/rejected": -8.125, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.3548387096774195, |
|
"grad_norm": 25.577249022826813, |
|
"learning_rate": 3.046594982078853e-07, |
|
"logits/chosen": -3.78125, |
|
"logits/rejected": -3.78125, |
|
"logps/chosen": -193.0, |
|
"logps/rejected": -276.0, |
|
"loss": 0.0173, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.37890625, |
|
"rewards/margins": 7.28125, |
|
"rewards/rejected": -7.65625, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.3870967741935485, |
|
"grad_norm": 0.8297796576977869, |
|
"learning_rate": 2.986857825567503e-07, |
|
"logits/chosen": -3.828125, |
|
"logits/rejected": -3.765625, |
|
"logps/chosen": -190.0, |
|
"logps/rejected": -274.0, |
|
"loss": 0.0179, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.2333984375, |
|
"rewards/margins": 7.53125, |
|
"rewards/rejected": -7.75, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.4193548387096775, |
|
"grad_norm": 37.60882100528414, |
|
"learning_rate": 2.927120669056153e-07, |
|
"logits/chosen": -3.71875, |
|
"logits/rejected": -3.765625, |
|
"logps/chosen": -189.0, |
|
"logps/rejected": -278.0, |
|
"loss": 0.0169, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.3671875, |
|
"rewards/margins": 7.84375, |
|
"rewards/rejected": -8.1875, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.4516129032258065, |
|
"grad_norm": 9.165654729466173, |
|
"learning_rate": 2.8673835125448024e-07, |
|
"logits/chosen": -3.75, |
|
"logits/rejected": -3.6875, |
|
"logps/chosen": -196.0, |
|
"logps/rejected": -284.0, |
|
"loss": 0.0261, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -0.1904296875, |
|
"rewards/margins": 7.90625, |
|
"rewards/rejected": -8.0625, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.4838709677419355, |
|
"grad_norm": 2.112943941209443, |
|
"learning_rate": 2.807646356033453e-07, |
|
"logits/chosen": -3.640625, |
|
"logits/rejected": -3.65625, |
|
"logps/chosen": -190.0, |
|
"logps/rejected": -280.0, |
|
"loss": 0.0271, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -0.66796875, |
|
"rewards/margins": 7.8125, |
|
"rewards/rejected": -8.5, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.5161290322580645, |
|
"grad_norm": 53.66978666939312, |
|
"learning_rate": 2.747909199522103e-07, |
|
"logits/chosen": -3.6875, |
|
"logits/rejected": -3.6875, |
|
"logps/chosen": -198.0, |
|
"logps/rejected": -276.0, |
|
"loss": 0.024, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -0.890625, |
|
"rewards/margins": 7.5, |
|
"rewards/rejected": -8.375, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.5483870967741935, |
|
"grad_norm": 13.967124404944409, |
|
"learning_rate": 2.6881720430107523e-07, |
|
"logits/chosen": -3.671875, |
|
"logits/rejected": -3.671875, |
|
"logps/chosen": -198.0, |
|
"logps/rejected": -284.0, |
|
"loss": 0.0212, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.66796875, |
|
"rewards/margins": 7.8125, |
|
"rewards/rejected": -8.5, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.5806451612903225, |
|
"grad_norm": 1.505490222650231, |
|
"learning_rate": 2.628434886499403e-07, |
|
"logits/chosen": -3.59375, |
|
"logits/rejected": -3.625, |
|
"logps/chosen": -180.0, |
|
"logps/rejected": -266.0, |
|
"loss": 0.0126, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.1103515625, |
|
"rewards/margins": 7.96875, |
|
"rewards/rejected": -7.875, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.6129032258064515, |
|
"grad_norm": 4.552495049762715, |
|
"learning_rate": 2.5686977299880523e-07, |
|
"logits/chosen": -3.65625, |
|
"logits/rejected": -3.703125, |
|
"logps/chosen": -190.0, |
|
"logps/rejected": -282.0, |
|
"loss": 0.0185, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.5234375, |
|
"rewards/margins": 8.125, |
|
"rewards/rejected": -8.6875, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.6451612903225805, |
|
"grad_norm": 7.1849195987738925, |
|
"learning_rate": 2.508960573476702e-07, |
|
"logits/chosen": -3.625, |
|
"logits/rejected": -3.6875, |
|
"logps/chosen": -196.0, |
|
"logps/rejected": -280.0, |
|
"loss": 0.0084, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.63671875, |
|
"rewards/margins": 7.53125, |
|
"rewards/rejected": -8.1875, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.6774193548387095, |
|
"grad_norm": 1.599699862303645, |
|
"learning_rate": 2.449223416965352e-07, |
|
"logits/chosen": -3.53125, |
|
"logits/rejected": -3.578125, |
|
"logps/chosen": -188.0, |
|
"logps/rejected": -282.0, |
|
"loss": 0.0167, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -0.388671875, |
|
"rewards/margins": 8.1875, |
|
"rewards/rejected": -8.625, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.7096774193548387, |
|
"grad_norm": 2.623858535249452, |
|
"learning_rate": 2.389486260454002e-07, |
|
"logits/chosen": -3.734375, |
|
"logits/rejected": -3.6875, |
|
"logps/chosen": -212.0, |
|
"logps/rejected": -302.0, |
|
"loss": 0.0408, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.8046875, |
|
"rewards/margins": 8.1875, |
|
"rewards/rejected": -10.0, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.7419354838709677, |
|
"grad_norm": 72.63368145693447, |
|
"learning_rate": 2.3297491039426522e-07, |
|
"logits/chosen": -3.65625, |
|
"logits/rejected": -3.640625, |
|
"logps/chosen": -202.0, |
|
"logps/rejected": -296.0, |
|
"loss": 0.0512, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -1.2421875, |
|
"rewards/margins": 8.4375, |
|
"rewards/rejected": -9.6875, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.7741935483870968, |
|
"grad_norm": 3.1206520900533667, |
|
"learning_rate": 2.270011947431302e-07, |
|
"logits/chosen": -3.640625, |
|
"logits/rejected": -3.6875, |
|
"logps/chosen": -180.0, |
|
"logps/rejected": -266.0, |
|
"loss": 0.04, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.1953125, |
|
"rewards/margins": 7.46875, |
|
"rewards/rejected": -7.28125, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.8064516129032258, |
|
"grad_norm": 41.020674776975646, |
|
"learning_rate": 2.2102747909199522e-07, |
|
"logits/chosen": -3.6875, |
|
"logits/rejected": -3.6875, |
|
"logps/chosen": -194.0, |
|
"logps/rejected": -290.0, |
|
"loss": 0.0124, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.03125, |
|
"rewards/margins": 8.0625, |
|
"rewards/rejected": -9.125, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.838709677419355, |
|
"grad_norm": 4.010133968357106, |
|
"learning_rate": 2.1505376344086022e-07, |
|
"logits/chosen": -3.59375, |
|
"logits/rejected": -3.546875, |
|
"logps/chosen": -212.0, |
|
"logps/rejected": -316.0, |
|
"loss": 0.0198, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -2.328125, |
|
"rewards/margins": 9.25, |
|
"rewards/rejected": -11.625, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.870967741935484, |
|
"grad_norm": 7.069397017435533, |
|
"learning_rate": 2.0908004778972521e-07, |
|
"logits/chosen": -3.65625, |
|
"logits/rejected": -3.6875, |
|
"logps/chosen": -188.0, |
|
"logps/rejected": -282.0, |
|
"loss": 0.0133, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.1337890625, |
|
"rewards/margins": 8.375, |
|
"rewards/rejected": -8.5625, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.903225806451613, |
|
"grad_norm": 8.936652736646373, |
|
"learning_rate": 2.0310633213859019e-07, |
|
"logits/chosen": -3.625, |
|
"logits/rejected": -3.546875, |
|
"logps/chosen": -196.0, |
|
"logps/rejected": -300.0, |
|
"loss": 0.0118, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -1.171875, |
|
"rewards/margins": 8.875, |
|
"rewards/rejected": -10.0625, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.935483870967742, |
|
"grad_norm": 7.8141795170212465, |
|
"learning_rate": 1.9713261648745518e-07, |
|
"logits/chosen": -3.546875, |
|
"logits/rejected": -3.640625, |
|
"logps/chosen": -190.0, |
|
"logps/rejected": -280.0, |
|
"loss": 0.0089, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.8515625, |
|
"rewards/margins": 8.125, |
|
"rewards/rejected": -9.0, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.967741935483871, |
|
"grad_norm": 5.395939675702438, |
|
"learning_rate": 1.9115890083632018e-07, |
|
"logits/chosen": -3.6875, |
|
"logits/rejected": -3.703125, |
|
"logps/chosen": -199.0, |
|
"logps/rejected": -292.0, |
|
"loss": 0.0136, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.86328125, |
|
"rewards/margins": 8.75, |
|
"rewards/rejected": -9.625, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 17.221158878981882, |
|
"learning_rate": 1.8518518518518516e-07, |
|
"logits/chosen": -3.6875, |
|
"logits/rejected": -3.6875, |
|
"logps/chosen": -197.0, |
|
"logps/rejected": -286.0, |
|
"loss": 0.0135, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.76953125, |
|
"rewards/margins": 8.3125, |
|
"rewards/rejected": -9.125, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_logits/chosen": -3.796875, |
|
"eval_logits/rejected": -3.765625, |
|
"eval_logps/chosen": -222.0, |
|
"eval_logps/rejected": -280.0, |
|
"eval_loss": 0.191436767578125, |
|
"eval_rewards/accuracies": 0.8928571343421936, |
|
"eval_rewards/chosen": -2.21875, |
|
"eval_rewards/margins": 5.40625, |
|
"eval_rewards/rejected": -7.625, |
|
"eval_runtime": 13.2451, |
|
"eval_samples_per_second": 15.1, |
|
"eval_steps_per_second": 0.528, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 2.032258064516129, |
|
"grad_norm": 0.45384363049400467, |
|
"learning_rate": 1.7921146953405018e-07, |
|
"logits/chosen": -3.5625, |
|
"logits/rejected": -3.578125, |
|
"logps/chosen": -191.0, |
|
"logps/rejected": -300.0, |
|
"loss": 0.0012, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.28125, |
|
"rewards/margins": 10.1875, |
|
"rewards/rejected": -10.4375, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 2.064516129032258, |
|
"grad_norm": 2.0283023878654722, |
|
"learning_rate": 1.7323775388291518e-07, |
|
"logits/chosen": -3.609375, |
|
"logits/rejected": -3.640625, |
|
"logps/chosen": -191.0, |
|
"logps/rejected": -298.0, |
|
"loss": 0.0019, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.302734375, |
|
"rewards/margins": 10.25, |
|
"rewards/rejected": -10.5625, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 2.096774193548387, |
|
"grad_norm": 0.9217664266560652, |
|
"learning_rate": 1.6726403823178015e-07, |
|
"logits/chosen": -3.6875, |
|
"logits/rejected": -3.75, |
|
"logps/chosen": -193.0, |
|
"logps/rejected": -298.0, |
|
"loss": 0.0013, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.93359375, |
|
"rewards/margins": 9.625, |
|
"rewards/rejected": -10.5625, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 2.129032258064516, |
|
"grad_norm": 0.44858819876518125, |
|
"learning_rate": 1.6129032258064515e-07, |
|
"logits/chosen": -3.6875, |
|
"logits/rejected": -3.703125, |
|
"logps/chosen": -207.0, |
|
"logps/rejected": -316.0, |
|
"loss": 0.0012, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.3046875, |
|
"rewards/margins": 10.125, |
|
"rewards/rejected": -11.4375, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 2.161290322580645, |
|
"grad_norm": 9.792046659236782, |
|
"learning_rate": 1.5531660692951015e-07, |
|
"logits/chosen": -3.578125, |
|
"logits/rejected": -3.59375, |
|
"logps/chosen": -199.0, |
|
"logps/rejected": -310.0, |
|
"loss": 0.0019, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.1875, |
|
"rewards/margins": 9.875, |
|
"rewards/rejected": -11.0625, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 2.193548387096774, |
|
"grad_norm": 0.3983602269153926, |
|
"learning_rate": 1.4934289127837515e-07, |
|
"logits/chosen": -3.5, |
|
"logits/rejected": -3.703125, |
|
"logps/chosen": -188.0, |
|
"logps/rejected": -302.0, |
|
"loss": 0.0024, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.59375, |
|
"rewards/margins": 10.25, |
|
"rewards/rejected": -10.875, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 2.225806451612903, |
|
"grad_norm": 1.2472683997191045, |
|
"learning_rate": 1.4336917562724012e-07, |
|
"logits/chosen": -3.59375, |
|
"logits/rejected": -3.65625, |
|
"logps/chosen": -202.0, |
|
"logps/rejected": -312.0, |
|
"loss": 0.0011, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.0546875, |
|
"rewards/margins": 10.25, |
|
"rewards/rejected": -11.3125, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 2.258064516129032, |
|
"grad_norm": 0.9104056031581937, |
|
"learning_rate": 1.3739545997610514e-07, |
|
"logits/chosen": -3.453125, |
|
"logits/rejected": -3.53125, |
|
"logps/chosen": -192.0, |
|
"logps/rejected": -308.0, |
|
"loss": 0.0008, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.90625, |
|
"rewards/margins": 10.625, |
|
"rewards/rejected": -11.5625, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 2.2903225806451615, |
|
"grad_norm": 0.5005896742747854, |
|
"learning_rate": 1.3142174432497014e-07, |
|
"logits/chosen": -3.453125, |
|
"logits/rejected": -3.5625, |
|
"logps/chosen": -196.0, |
|
"logps/rejected": -316.0, |
|
"loss": 0.0015, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.51953125, |
|
"rewards/margins": 11.4375, |
|
"rewards/rejected": -11.9375, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 2.3225806451612905, |
|
"grad_norm": 2.0961727669442247, |
|
"learning_rate": 1.254480286738351e-07, |
|
"logits/chosen": -3.53125, |
|
"logits/rejected": -3.546875, |
|
"logps/chosen": -206.0, |
|
"logps/rejected": -318.0, |
|
"loss": 0.0011, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.515625, |
|
"rewards/margins": 10.5, |
|
"rewards/rejected": -12.0, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 2.3548387096774195, |
|
"grad_norm": 0.33710671811391674, |
|
"learning_rate": 1.194743130227001e-07, |
|
"logits/chosen": -3.53125, |
|
"logits/rejected": -3.578125, |
|
"logps/chosen": -201.0, |
|
"logps/rejected": -318.0, |
|
"loss": 0.0008, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.625, |
|
"rewards/margins": 10.375, |
|
"rewards/rejected": -12.0, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 2.3870967741935485, |
|
"grad_norm": 3.7387698089394634, |
|
"learning_rate": 1.135005973715651e-07, |
|
"logits/chosen": -3.59375, |
|
"logits/rejected": -3.625, |
|
"logps/chosen": -202.0, |
|
"logps/rejected": -314.0, |
|
"loss": 0.0014, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.3671875, |
|
"rewards/margins": 10.375, |
|
"rewards/rejected": -11.75, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 2.4193548387096775, |
|
"grad_norm": 3.5076436110564426, |
|
"learning_rate": 1.0752688172043011e-07, |
|
"logits/chosen": -3.5625, |
|
"logits/rejected": -3.6875, |
|
"logps/chosen": -192.0, |
|
"logps/rejected": -306.0, |
|
"loss": 0.0019, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.53125, |
|
"rewards/margins": 10.4375, |
|
"rewards/rejected": -11.0, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 2.4516129032258065, |
|
"grad_norm": 0.3051005963902528, |
|
"learning_rate": 1.0155316606929509e-07, |
|
"logits/chosen": -3.46875, |
|
"logits/rejected": -3.53125, |
|
"logps/chosen": -194.0, |
|
"logps/rejected": -312.0, |
|
"loss": 0.0009, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.94140625, |
|
"rewards/margins": 10.625, |
|
"rewards/rejected": -11.5625, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 2.4838709677419355, |
|
"grad_norm": 0.4055555200925197, |
|
"learning_rate": 9.557945041816009e-08, |
|
"logits/chosen": -3.40625, |
|
"logits/rejected": -3.546875, |
|
"logps/chosen": -190.0, |
|
"logps/rejected": -314.0, |
|
"loss": 0.0014, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.76953125, |
|
"rewards/margins": 11.0, |
|
"rewards/rejected": -11.75, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 2.5161290322580645, |
|
"grad_norm": 0.4423005388531537, |
|
"learning_rate": 8.960573476702509e-08, |
|
"logits/chosen": -3.5625, |
|
"logits/rejected": -3.5625, |
|
"logps/chosen": -211.0, |
|
"logps/rejected": -336.0, |
|
"loss": 0.0013, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.640625, |
|
"rewards/margins": 11.5, |
|
"rewards/rejected": -13.1875, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 2.5483870967741935, |
|
"grad_norm": 0.34201712686499286, |
|
"learning_rate": 8.363201911589008e-08, |
|
"logits/chosen": -3.4375, |
|
"logits/rejected": -3.515625, |
|
"logps/chosen": -197.0, |
|
"logps/rejected": -312.0, |
|
"loss": 0.001, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.81640625, |
|
"rewards/margins": 10.9375, |
|
"rewards/rejected": -11.75, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 2.5806451612903225, |
|
"grad_norm": 0.08324339497209432, |
|
"learning_rate": 7.765830346475507e-08, |
|
"logits/chosen": -3.421875, |
|
"logits/rejected": -3.59375, |
|
"logps/chosen": -207.0, |
|
"logps/rejected": -330.0, |
|
"loss": 0.0005, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.609375, |
|
"rewards/margins": 11.5, |
|
"rewards/rejected": -13.125, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 2.6129032258064515, |
|
"grad_norm": 0.35882151597975387, |
|
"learning_rate": 7.168458781362006e-08, |
|
"logits/chosen": -3.59375, |
|
"logits/rejected": -3.6875, |
|
"logps/chosen": -202.0, |
|
"logps/rejected": -328.0, |
|
"loss": 0.0045, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.03125, |
|
"rewards/margins": 11.625, |
|
"rewards/rejected": -12.6875, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 2.6451612903225805, |
|
"grad_norm": 0.3006723416131319, |
|
"learning_rate": 6.571087216248507e-08, |
|
"logits/chosen": -3.4375, |
|
"logits/rejected": -3.609375, |
|
"logps/chosen": -197.0, |
|
"logps/rejected": -322.0, |
|
"loss": 0.0026, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.40625, |
|
"rewards/margins": 11.0, |
|
"rewards/rejected": -12.4375, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 2.6774193548387095, |
|
"grad_norm": 0.9135709315414355, |
|
"learning_rate": 5.973715651135006e-08, |
|
"logits/chosen": -3.40625, |
|
"logits/rejected": -3.546875, |
|
"logps/chosen": -194.0, |
|
"logps/rejected": -312.0, |
|
"loss": 0.0026, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.046875, |
|
"rewards/margins": 11.3125, |
|
"rewards/rejected": -12.3125, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 2.709677419354839, |
|
"grad_norm": 0.11799710974470022, |
|
"learning_rate": 5.3763440860215054e-08, |
|
"logits/chosen": -3.453125, |
|
"logits/rejected": -3.609375, |
|
"logps/chosen": -193.0, |
|
"logps/rejected": -330.0, |
|
"loss": 0.001, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.79296875, |
|
"rewards/margins": 12.8125, |
|
"rewards/rejected": -13.625, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 2.741935483870968, |
|
"grad_norm": 0.24029168708895726, |
|
"learning_rate": 4.7789725209080046e-08, |
|
"logits/chosen": -3.484375, |
|
"logits/rejected": -3.5625, |
|
"logps/chosen": -200.0, |
|
"logps/rejected": -320.0, |
|
"loss": 0.0006, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.2734375, |
|
"rewards/margins": 10.875, |
|
"rewards/rejected": -12.1875, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 2.774193548387097, |
|
"grad_norm": 1.5601788534872705, |
|
"learning_rate": 4.181600955794504e-08, |
|
"logits/chosen": -3.484375, |
|
"logits/rejected": -3.625, |
|
"logps/chosen": -206.0, |
|
"logps/rejected": -326.0, |
|
"loss": 0.0009, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.265625, |
|
"rewards/margins": 11.1875, |
|
"rewards/rejected": -12.4375, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 2.806451612903226, |
|
"grad_norm": 2.7780970090627104, |
|
"learning_rate": 3.584229390681003e-08, |
|
"logits/chosen": -3.53125, |
|
"logits/rejected": -3.65625, |
|
"logps/chosen": -201.0, |
|
"logps/rejected": -320.0, |
|
"loss": 0.0025, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.0546875, |
|
"rewards/margins": 10.8125, |
|
"rewards/rejected": -11.875, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 2.838709677419355, |
|
"grad_norm": 0.36034720795454384, |
|
"learning_rate": 2.986857825567503e-08, |
|
"logits/chosen": -3.546875, |
|
"logits/rejected": -3.578125, |
|
"logps/chosen": -198.0, |
|
"logps/rejected": -318.0, |
|
"loss": 0.0008, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.86328125, |
|
"rewards/margins": 11.25, |
|
"rewards/rejected": -12.125, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 2.870967741935484, |
|
"grad_norm": 0.2764304820275906, |
|
"learning_rate": 2.3894862604540023e-08, |
|
"logits/chosen": -3.59375, |
|
"logits/rejected": -3.671875, |
|
"logps/chosen": -203.0, |
|
"logps/rejected": -316.0, |
|
"loss": 0.0037, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.03125, |
|
"rewards/margins": 10.6875, |
|
"rewards/rejected": -11.75, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 2.903225806451613, |
|
"grad_norm": 0.4293826404732342, |
|
"learning_rate": 1.7921146953405015e-08, |
|
"logits/chosen": -3.4375, |
|
"logits/rejected": -3.546875, |
|
"logps/chosen": -208.0, |
|
"logps/rejected": -322.0, |
|
"loss": 0.0011, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.859375, |
|
"rewards/margins": 10.875, |
|
"rewards/rejected": -12.6875, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 2.935483870967742, |
|
"grad_norm": 0.34603080271508346, |
|
"learning_rate": 1.1947431302270011e-08, |
|
"logits/chosen": -3.546875, |
|
"logits/rejected": -3.703125, |
|
"logps/chosen": -207.0, |
|
"logps/rejected": -320.0, |
|
"loss": 0.0007, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.265625, |
|
"rewards/margins": 11.0, |
|
"rewards/rejected": -12.25, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 2.967741935483871, |
|
"grad_norm": 2.3868143014919885, |
|
"learning_rate": 5.973715651135006e-09, |
|
"logits/chosen": -3.4375, |
|
"logits/rejected": -3.65625, |
|
"logps/chosen": -188.0, |
|
"logps/rejected": -304.0, |
|
"loss": 0.002, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.71484375, |
|
"rewards/margins": 10.375, |
|
"rewards/rejected": -11.125, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"grad_norm": 0.40772402916251804, |
|
"learning_rate": 0.0, |
|
"logits/chosen": -3.5, |
|
"logits/rejected": -3.609375, |
|
"logps/chosen": -194.0, |
|
"logps/rejected": -320.0, |
|
"loss": 0.0006, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.1796875, |
|
"rewards/margins": 10.9375, |
|
"rewards/rejected": -12.125, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_logits/chosen": -3.734375, |
|
"eval_logits/rejected": -3.75, |
|
"eval_logps/chosen": -226.0, |
|
"eval_logps/rejected": -294.0, |
|
"eval_loss": 0.19125854969024658, |
|
"eval_rewards/accuracies": 0.8928571343421936, |
|
"eval_rewards/chosen": -2.703125, |
|
"eval_rewards/margins": 6.3125, |
|
"eval_rewards/rejected": -9.0, |
|
"eval_runtime": 16.0795, |
|
"eval_samples_per_second": 12.438, |
|
"eval_steps_per_second": 0.435, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 930, |
|
"total_flos": 0.0, |
|
"train_loss": 0.13316144410480735, |
|
"train_runtime": 5554.5038, |
|
"train_samples_per_second": 5.355, |
|
"train_steps_per_second": 0.167 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 930, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 0.0, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|