|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0262989095574087, |
|
"eval_steps": 100, |
|
"global_step": 900, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.002280664243460908, |
|
"grad_norm": 104.30809234123367, |
|
"learning_rate": 4e-09, |
|
"logits/chosen": -1.353676199913025, |
|
"logits/rejected": -1.4426417350769043, |
|
"logps/chosen": -169.59426879882812, |
|
"logps/rejected": -212.90069580078125, |
|
"loss": 0.7312, |
|
"rewards/accuracies": 0.4375, |
|
"rewards/chosen": 0.23574542999267578, |
|
"rewards/margins": -0.06179435923695564, |
|
"rewards/rejected": 0.29753977060317993, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.004561328486921816, |
|
"grad_norm": 92.96399840391902, |
|
"learning_rate": 8e-09, |
|
"logits/chosen": -1.2849147319793701, |
|
"logits/rejected": -1.3181504011154175, |
|
"logps/chosen": -156.07391357421875, |
|
"logps/rejected": -180.74073791503906, |
|
"loss": 0.6935, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.32024019956588745, |
|
"rewards/margins": 0.08485272526741028, |
|
"rewards/rejected": 0.23538745939731598, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.006841992730382724, |
|
"grad_norm": 98.95959882761899, |
|
"learning_rate": 1.1999999999999998e-08, |
|
"logits/chosen": -1.4182077646255493, |
|
"logits/rejected": -1.4656966924667358, |
|
"logps/chosen": -162.13934326171875, |
|
"logps/rejected": -188.761474609375, |
|
"loss": 0.6766, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.4807170629501343, |
|
"rewards/margins": 0.1233808621764183, |
|
"rewards/rejected": 0.35733622312545776, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.009122656973843632, |
|
"grad_norm": 100.15592083854503, |
|
"learning_rate": 1.6e-08, |
|
"logits/chosen": -1.3926836252212524, |
|
"logits/rejected": -1.401774287223816, |
|
"logps/chosen": -170.7425079345703, |
|
"logps/rejected": -190.57833862304688, |
|
"loss": 0.7039, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": 0.4466077983379364, |
|
"rewards/margins": 0.06900040060281754, |
|
"rewards/rejected": 0.37760740518569946, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.01140332121730454, |
|
"grad_norm": 97.0889776355223, |
|
"learning_rate": 2e-08, |
|
"logits/chosen": -1.376230239868164, |
|
"logits/rejected": -1.409171223640442, |
|
"logps/chosen": -169.57582092285156, |
|
"logps/rejected": -185.82571411132812, |
|
"loss": 0.7113, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.39193734526634216, |
|
"rewards/margins": 0.3205062747001648, |
|
"rewards/rejected": 0.07143110036849976, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.013683985460765448, |
|
"grad_norm": 92.66648488760384, |
|
"learning_rate": 2.3999999999999997e-08, |
|
"logits/chosen": -1.3613837957382202, |
|
"logits/rejected": -1.4228042364120483, |
|
"logps/chosen": -142.47850036621094, |
|
"logps/rejected": -195.29649353027344, |
|
"loss": 0.6867, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.19295203685760498, |
|
"rewards/margins": -0.01336541399359703, |
|
"rewards/rejected": 0.20631742477416992, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.015964649704226355, |
|
"grad_norm": 88.1654009314199, |
|
"learning_rate": 2.8000000000000003e-08, |
|
"logits/chosen": -1.5071678161621094, |
|
"logits/rejected": -1.5190939903259277, |
|
"logps/chosen": -128.34288024902344, |
|
"logps/rejected": -131.07789611816406, |
|
"loss": 0.6954, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": 0.23906439542770386, |
|
"rewards/margins": 0.07392804324626923, |
|
"rewards/rejected": 0.16513636708259583, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.018245313947687265, |
|
"grad_norm": 83.0913005177034, |
|
"learning_rate": 3.2e-08, |
|
"logits/chosen": -1.2882771492004395, |
|
"logits/rejected": -1.3209519386291504, |
|
"logps/chosen": -125.15751647949219, |
|
"logps/rejected": -132.21255493164062, |
|
"loss": 0.6652, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": 0.18403516709804535, |
|
"rewards/margins": -0.032971642911434174, |
|
"rewards/rejected": 0.21700681746006012, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.02052597819114817, |
|
"grad_norm": 92.77604284067941, |
|
"learning_rate": 3.6e-08, |
|
"logits/chosen": -1.3992947340011597, |
|
"logits/rejected": -1.473150610923767, |
|
"logps/chosen": -173.55206298828125, |
|
"logps/rejected": -202.7044677734375, |
|
"loss": 0.7212, |
|
"rewards/accuracies": 0.34375, |
|
"rewards/chosen": 0.4242628216743469, |
|
"rewards/margins": -0.0976889505982399, |
|
"rewards/rejected": 0.5219517350196838, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.02280664243460908, |
|
"grad_norm": 88.66990299818134, |
|
"learning_rate": 4e-08, |
|
"logits/chosen": -1.432981014251709, |
|
"logits/rejected": -1.5539088249206543, |
|
"logps/chosen": -182.9879913330078, |
|
"logps/rejected": -203.99020385742188, |
|
"loss": 0.7184, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": 0.4513806104660034, |
|
"rewards/margins": 0.20164981484413147, |
|
"rewards/rejected": 0.24973079562187195, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.02508730667806999, |
|
"grad_norm": 77.91111396247051, |
|
"learning_rate": 4.4e-08, |
|
"logits/chosen": -1.3197808265686035, |
|
"logits/rejected": -1.3996787071228027, |
|
"logps/chosen": -154.55003356933594, |
|
"logps/rejected": -232.81744384765625, |
|
"loss": 0.6479, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.3307289481163025, |
|
"rewards/margins": 0.18726389110088348, |
|
"rewards/rejected": 0.14346502721309662, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.027367970921530895, |
|
"grad_norm": 99.08871355118758, |
|
"learning_rate": 4.799999999999999e-08, |
|
"logits/chosen": -1.3024228811264038, |
|
"logits/rejected": -1.3258510828018188, |
|
"logps/chosen": -131.18356323242188, |
|
"logps/rejected": -148.7147216796875, |
|
"loss": 0.7046, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.421053409576416, |
|
"rewards/margins": 0.3122374713420868, |
|
"rewards/rejected": 0.10881592333316803, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.029648635164991805, |
|
"grad_norm": 89.36423052851278, |
|
"learning_rate": 5.2e-08, |
|
"logits/chosen": -1.329531192779541, |
|
"logits/rejected": -1.4354490041732788, |
|
"logps/chosen": -168.1116180419922, |
|
"logps/rejected": -207.8105010986328, |
|
"loss": 0.6619, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.40395674109458923, |
|
"rewards/margins": 0.26189538836479187, |
|
"rewards/rejected": 0.14206132292747498, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.03192929940845271, |
|
"grad_norm": 99.33427508045129, |
|
"learning_rate": 5.6000000000000005e-08, |
|
"logits/chosen": -1.3796099424362183, |
|
"logits/rejected": -1.3620996475219727, |
|
"logps/chosen": -131.89183044433594, |
|
"logps/rejected": -133.18844604492188, |
|
"loss": 0.6782, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.37585678696632385, |
|
"rewards/margins": 0.27094465494155884, |
|
"rewards/rejected": 0.10491211712360382, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.03420996365191362, |
|
"grad_norm": 98.1725487266389, |
|
"learning_rate": 6e-08, |
|
"logits/chosen": -1.4092419147491455, |
|
"logits/rejected": -1.4259543418884277, |
|
"logps/chosen": -228.17083740234375, |
|
"logps/rejected": -249.2948455810547, |
|
"loss": 0.6923, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.4407646358013153, |
|
"rewards/margins": 0.335997998714447, |
|
"rewards/rejected": 0.10476663708686829, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.03649062789537453, |
|
"grad_norm": 75.77879619863296, |
|
"learning_rate": 6.4e-08, |
|
"logits/chosen": -1.364654541015625, |
|
"logits/rejected": -1.4574649333953857, |
|
"logps/chosen": -134.63072204589844, |
|
"logps/rejected": -159.65719604492188, |
|
"loss": 0.6454, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": 0.28305765986442566, |
|
"rewards/margins": 0.14592355489730835, |
|
"rewards/rejected": 0.1371341049671173, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.038771292138835435, |
|
"grad_norm": 100.55666520099189, |
|
"learning_rate": 6.8e-08, |
|
"logits/chosen": -1.3500841856002808, |
|
"logits/rejected": -1.4460691213607788, |
|
"logps/chosen": -134.4729461669922, |
|
"logps/rejected": -159.3612823486328, |
|
"loss": 0.716, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.44891929626464844, |
|
"rewards/margins": 0.12445079535245895, |
|
"rewards/rejected": 0.3244684934616089, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.04105195638229634, |
|
"grad_norm": 104.71963688579538, |
|
"learning_rate": 7.2e-08, |
|
"logits/chosen": -1.3853086233139038, |
|
"logits/rejected": -1.434361457824707, |
|
"logps/chosen": -139.8603057861328, |
|
"logps/rejected": -171.86366271972656, |
|
"loss": 0.6988, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.3810410499572754, |
|
"rewards/margins": 0.2534639537334442, |
|
"rewards/rejected": 0.12757712602615356, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.043332620625757254, |
|
"grad_norm": 77.00258209889141, |
|
"learning_rate": 7.599999999999999e-08, |
|
"logits/chosen": -1.3689879179000854, |
|
"logits/rejected": -1.3514069318771362, |
|
"logps/chosen": -208.98460388183594, |
|
"logps/rejected": -216.9095458984375, |
|
"loss": 0.6691, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.5943199396133423, |
|
"rewards/margins": 0.26101234555244446, |
|
"rewards/rejected": 0.33330756425857544, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.04561328486921816, |
|
"grad_norm": 83.25872912559916, |
|
"learning_rate": 8e-08, |
|
"logits/chosen": -1.3037664890289307, |
|
"logits/rejected": -1.3465715646743774, |
|
"logps/chosen": -160.97340393066406, |
|
"logps/rejected": -178.60267639160156, |
|
"loss": 0.6319, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.4052881896495819, |
|
"rewards/margins": 0.16212955117225647, |
|
"rewards/rejected": 0.24315865337848663, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.047893949112679066, |
|
"grad_norm": 113.55879280714245, |
|
"learning_rate": 8.4e-08, |
|
"logits/chosen": -1.3739519119262695, |
|
"logits/rejected": -1.442929744720459, |
|
"logps/chosen": -163.52279663085938, |
|
"logps/rejected": -217.51824951171875, |
|
"loss": 0.7171, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.13579119741916656, |
|
"rewards/margins": 0.14747555553913116, |
|
"rewards/rejected": -0.011684387922286987, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.05017461335613998, |
|
"grad_norm": 81.14122988211874, |
|
"learning_rate": 8.8e-08, |
|
"logits/chosen": -1.373199701309204, |
|
"logits/rejected": -1.4750025272369385, |
|
"logps/chosen": -132.52114868164062, |
|
"logps/rejected": -154.50648498535156, |
|
"loss": 0.6947, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.4790118932723999, |
|
"rewards/margins": 0.16499680280685425, |
|
"rewards/rejected": 0.31401512026786804, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.052455277599600884, |
|
"grad_norm": 89.9603968710281, |
|
"learning_rate": 9.2e-08, |
|
"logits/chosen": -1.3388586044311523, |
|
"logits/rejected": -1.4220290184020996, |
|
"logps/chosen": -181.95082092285156, |
|
"logps/rejected": -226.30923461914062, |
|
"loss": 0.6941, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": 0.4439522325992584, |
|
"rewards/margins": 0.17398414015769958, |
|
"rewards/rejected": 0.26996806263923645, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.05473594184306179, |
|
"grad_norm": 104.88400068525443, |
|
"learning_rate": 9.599999999999999e-08, |
|
"logits/chosen": -1.456312894821167, |
|
"logits/rejected": -1.451519250869751, |
|
"logps/chosen": -199.40493774414062, |
|
"logps/rejected": -216.68386840820312, |
|
"loss": 0.7537, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.4792614281177521, |
|
"rewards/margins": 0.112982377409935, |
|
"rewards/rejected": 0.36627906560897827, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.0570166060865227, |
|
"grad_norm": 73.96994315968016, |
|
"learning_rate": 1e-07, |
|
"logits/chosen": -1.3086615800857544, |
|
"logits/rejected": -1.3628607988357544, |
|
"logps/chosen": -151.69989013671875, |
|
"logps/rejected": -171.46029663085938, |
|
"loss": 0.6314, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.1983395218849182, |
|
"rewards/margins": 0.22184070944786072, |
|
"rewards/rejected": -0.023501206189393997, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.05929727032998361, |
|
"grad_norm": 100.64321021539473, |
|
"learning_rate": 1.04e-07, |
|
"logits/chosen": -1.5143202543258667, |
|
"logits/rejected": -1.5794754028320312, |
|
"logps/chosen": -118.7998046875, |
|
"logps/rejected": -149.1954803466797, |
|
"loss": 0.6704, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.34698718786239624, |
|
"rewards/margins": 0.06618239730596542, |
|
"rewards/rejected": 0.2808048129081726, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.061577934573444515, |
|
"grad_norm": 96.37635530030595, |
|
"learning_rate": 1.08e-07, |
|
"logits/chosen": -1.254115104675293, |
|
"logits/rejected": -1.3123385906219482, |
|
"logps/chosen": -152.57516479492188, |
|
"logps/rejected": -172.66200256347656, |
|
"loss": 0.6966, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": 0.28736555576324463, |
|
"rewards/margins": 0.0115619245916605, |
|
"rewards/rejected": 0.2758035957813263, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.06385859881690542, |
|
"grad_norm": 99.11153586277212, |
|
"learning_rate": 1.1200000000000001e-07, |
|
"logits/chosen": -1.3405977487564087, |
|
"logits/rejected": -1.4054450988769531, |
|
"logps/chosen": -194.3892822265625, |
|
"logps/rejected": -219.24171447753906, |
|
"loss": 0.6625, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.382589727640152, |
|
"rewards/margins": 0.24182234704494476, |
|
"rewards/rejected": 0.14076738059520721, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.06613926306036633, |
|
"grad_norm": 82.95051663248452, |
|
"learning_rate": 1.1599999999999999e-07, |
|
"logits/chosen": -1.4561753273010254, |
|
"logits/rejected": -1.4183000326156616, |
|
"logps/chosen": -189.18923950195312, |
|
"logps/rejected": -186.99264526367188, |
|
"loss": 0.6206, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.4602745771408081, |
|
"rewards/margins": 0.18983443081378937, |
|
"rewards/rejected": 0.27044013142585754, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.06841992730382725, |
|
"grad_norm": 85.63947077517764, |
|
"learning_rate": 1.2e-07, |
|
"logits/chosen": -1.4239516258239746, |
|
"logits/rejected": -1.5685184001922607, |
|
"logps/chosen": -140.7595977783203, |
|
"logps/rejected": -184.33067321777344, |
|
"loss": 0.7255, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": 0.2823754847049713, |
|
"rewards/margins": 0.02492373436689377, |
|
"rewards/rejected": 0.25745177268981934, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.07070059154728815, |
|
"grad_norm": 97.67365684595, |
|
"learning_rate": 1.24e-07, |
|
"logits/chosen": -1.350022554397583, |
|
"logits/rejected": -1.3478128910064697, |
|
"logps/chosen": -171.09677124023438, |
|
"logps/rejected": -165.71694946289062, |
|
"loss": 0.6819, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": 0.3082874119281769, |
|
"rewards/margins": 0.11261623352766037, |
|
"rewards/rejected": 0.1956711858510971, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.07298125579074906, |
|
"grad_norm": 99.3804737289846, |
|
"learning_rate": 1.28e-07, |
|
"logits/chosen": -1.4815456867218018, |
|
"logits/rejected": -1.4830272197723389, |
|
"logps/chosen": -107.2784194946289, |
|
"logps/rejected": -121.1039047241211, |
|
"loss": 0.7165, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.31323567032814026, |
|
"rewards/margins": 0.11968746036291122, |
|
"rewards/rejected": 0.19354820251464844, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.07526192003420996, |
|
"grad_norm": 94.44141503559203, |
|
"learning_rate": 1.32e-07, |
|
"logits/chosen": -1.493153691291809, |
|
"logits/rejected": -1.4850858449935913, |
|
"logps/chosen": -164.4472198486328, |
|
"logps/rejected": -169.9815673828125, |
|
"loss": 0.6558, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.5161038637161255, |
|
"rewards/margins": 0.30025073885917664, |
|
"rewards/rejected": 0.21585312485694885, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.07754258427767087, |
|
"grad_norm": 86.21425469616315, |
|
"learning_rate": 1.36e-07, |
|
"logits/chosen": -1.333409070968628, |
|
"logits/rejected": -1.3404844999313354, |
|
"logps/chosen": -153.38101196289062, |
|
"logps/rejected": -159.85690307617188, |
|
"loss": 0.6781, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.3244006931781769, |
|
"rewards/margins": 0.22184959053993225, |
|
"rewards/rejected": 0.10255111753940582, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.07982324852113178, |
|
"grad_norm": 90.03565465587579, |
|
"learning_rate": 1.3999999999999998e-07, |
|
"logits/chosen": -1.5341051816940308, |
|
"logits/rejected": -1.5601752996444702, |
|
"logps/chosen": -149.2415008544922, |
|
"logps/rejected": -156.4012908935547, |
|
"loss": 0.6936, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": 0.27590540051460266, |
|
"rewards/margins": 0.12058861553668976, |
|
"rewards/rejected": 0.1553167849779129, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.08210391276459268, |
|
"grad_norm": 80.65697099385187, |
|
"learning_rate": 1.44e-07, |
|
"logits/chosen": -1.3111834526062012, |
|
"logits/rejected": -1.3080787658691406, |
|
"logps/chosen": -152.25611877441406, |
|
"logps/rejected": -169.93919372558594, |
|
"loss": 0.6471, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": 0.34636569023132324, |
|
"rewards/margins": 0.3154309093952179, |
|
"rewards/rejected": 0.03093479573726654, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.0843845770080536, |
|
"grad_norm": 116.72716177672243, |
|
"learning_rate": 1.48e-07, |
|
"logits/chosen": -1.3597153425216675, |
|
"logits/rejected": -1.4166897535324097, |
|
"logps/chosen": -174.33534240722656, |
|
"logps/rejected": -200.9926300048828, |
|
"loss": 0.6831, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.2202744334936142, |
|
"rewards/margins": 0.13681186735630035, |
|
"rewards/rejected": 0.08346255123615265, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.08666524125151451, |
|
"grad_norm": 99.62367819711024, |
|
"learning_rate": 1.5199999999999998e-07, |
|
"logits/chosen": -1.3973196744918823, |
|
"logits/rejected": -1.4653477668762207, |
|
"logps/chosen": -189.6744384765625, |
|
"logps/rejected": -224.8593292236328, |
|
"loss": 0.7086, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.33062875270843506, |
|
"rewards/margins": 0.11676961928606033, |
|
"rewards/rejected": 0.21385914087295532, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.08894590549497541, |
|
"grad_norm": 86.47188266565705, |
|
"learning_rate": 1.56e-07, |
|
"logits/chosen": -1.3268086910247803, |
|
"logits/rejected": -1.302788257598877, |
|
"logps/chosen": -135.62147521972656, |
|
"logps/rejected": -153.5860137939453, |
|
"loss": 0.6835, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": 0.2682960629463196, |
|
"rewards/margins": 0.05024400353431702, |
|
"rewards/rejected": 0.21805202960968018, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.09122656973843632, |
|
"grad_norm": 125.16786139574279, |
|
"learning_rate": 1.6e-07, |
|
"logits/chosen": -1.2104482650756836, |
|
"logits/rejected": -1.3274182081222534, |
|
"logps/chosen": -161.164794921875, |
|
"logps/rejected": -210.53590393066406, |
|
"loss": 0.7174, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": 0.2355586439371109, |
|
"rewards/margins": 0.07326290011405945, |
|
"rewards/rejected": 0.16229577362537384, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.09350723398189723, |
|
"grad_norm": 98.85832957481998, |
|
"learning_rate": 1.6399999999999999e-07, |
|
"logits/chosen": -1.3579671382904053, |
|
"logits/rejected": -1.4671956300735474, |
|
"logps/chosen": -153.21591186523438, |
|
"logps/rejected": -173.60589599609375, |
|
"loss": 0.6862, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": 0.2675905227661133, |
|
"rewards/margins": 0.046474047005176544, |
|
"rewards/rejected": 0.22111651301383972, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.09578789822535813, |
|
"grad_norm": 83.11445245790541, |
|
"learning_rate": 1.68e-07, |
|
"logits/chosen": -1.2888550758361816, |
|
"logits/rejected": -1.3688700199127197, |
|
"logps/chosen": -145.24754333496094, |
|
"logps/rejected": -174.91700744628906, |
|
"loss": 0.637, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.36274462938308716, |
|
"rewards/margins": 0.15029877424240112, |
|
"rewards/rejected": 0.21244585514068604, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.09806856246881904, |
|
"grad_norm": 82.49239750970325, |
|
"learning_rate": 1.7199999999999998e-07, |
|
"logits/chosen": -1.3739551305770874, |
|
"logits/rejected": -1.3799883127212524, |
|
"logps/chosen": -178.80386352539062, |
|
"logps/rejected": -215.7256317138672, |
|
"loss": 0.6907, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.32459211349487305, |
|
"rewards/margins": 0.16312535107135773, |
|
"rewards/rejected": 0.1614667773246765, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.10034922671227996, |
|
"grad_norm": 105.64550946139919, |
|
"learning_rate": 1.76e-07, |
|
"logits/chosen": -1.4653171300888062, |
|
"logits/rejected": -1.508003830909729, |
|
"logps/chosen": -192.55154418945312, |
|
"logps/rejected": -214.61546325683594, |
|
"loss": 0.7401, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.43679386377334595, |
|
"rewards/margins": 0.1878117173910141, |
|
"rewards/rejected": 0.24898216128349304, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.10262989095574086, |
|
"grad_norm": 106.99677192741333, |
|
"learning_rate": 1.8e-07, |
|
"logits/chosen": -1.3307772874832153, |
|
"logits/rejected": -1.4872570037841797, |
|
"logps/chosen": -226.3917236328125, |
|
"logps/rejected": -265.8712158203125, |
|
"loss": 0.6758, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": 0.42067086696624756, |
|
"rewards/margins": 0.07930833846330643, |
|
"rewards/rejected": 0.34136250615119934, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.10491055519920177, |
|
"grad_norm": 104.67741510645097, |
|
"learning_rate": 1.84e-07, |
|
"logits/chosen": -1.3470810651779175, |
|
"logits/rejected": -1.3835241794586182, |
|
"logps/chosen": -162.22958374023438, |
|
"logps/rejected": -184.17697143554688, |
|
"loss": 0.6993, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.25471583008766174, |
|
"rewards/margins": 0.08957283943891525, |
|
"rewards/rejected": 0.1651429831981659, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.10719121944266267, |
|
"grad_norm": 82.2596269922547, |
|
"learning_rate": 1.88e-07, |
|
"logits/chosen": -1.3317389488220215, |
|
"logits/rejected": -1.367332935333252, |
|
"logps/chosen": -138.1045684814453, |
|
"logps/rejected": -213.2273712158203, |
|
"loss": 0.5933, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": 0.27782779932022095, |
|
"rewards/margins": 0.2610364258289337, |
|
"rewards/rejected": 0.01679137721657753, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.10947188368612358, |
|
"grad_norm": 106.46518450026352, |
|
"learning_rate": 1.9199999999999997e-07, |
|
"logits/chosen": -1.3722081184387207, |
|
"logits/rejected": -1.4072837829589844, |
|
"logps/chosen": -185.40281677246094, |
|
"logps/rejected": -205.63037109375, |
|
"loss": 0.7079, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": 0.4706501066684723, |
|
"rewards/margins": 0.29465728998184204, |
|
"rewards/rejected": 0.17599281668663025, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.11175254792958449, |
|
"grad_norm": 79.05608727628055, |
|
"learning_rate": 1.9599999999999998e-07, |
|
"logits/chosen": -1.2723050117492676, |
|
"logits/rejected": -1.2712621688842773, |
|
"logps/chosen": -170.6747589111328, |
|
"logps/rejected": -172.70376586914062, |
|
"loss": 0.6713, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.4158933460712433, |
|
"rewards/margins": 0.22599810361862183, |
|
"rewards/rejected": 0.18989527225494385, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.1140332121730454, |
|
"grad_norm": 102.50537690642588, |
|
"learning_rate": 2e-07, |
|
"logits/chosen": -1.2609540224075317, |
|
"logits/rejected": -1.3507779836654663, |
|
"logps/chosen": -126.81893157958984, |
|
"logps/rejected": -172.69830322265625, |
|
"loss": 0.7344, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.2087049037218094, |
|
"rewards/margins": 0.05775710940361023, |
|
"rewards/rejected": 0.15094780921936035, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.1140332121730454, |
|
"eval_logits/chosen": -1.465891718864441, |
|
"eval_logits/rejected": -1.452298879623413, |
|
"eval_logps/chosen": -126.79114532470703, |
|
"eval_logps/rejected": -128.49049377441406, |
|
"eval_loss": 0.7120834589004517, |
|
"eval_rewards/accuracies": 0.4000000059604645, |
|
"eval_rewards/chosen": 0.3106829524040222, |
|
"eval_rewards/margins": -0.04654194042086601, |
|
"eval_rewards/rejected": 0.3572249114513397, |
|
"eval_runtime": 22.2628, |
|
"eval_samples_per_second": 4.492, |
|
"eval_steps_per_second": 1.123, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.11631387641650631, |
|
"grad_norm": 115.65869747987765, |
|
"learning_rate": 1.9999927671554446e-07, |
|
"logits/chosen": -1.4759191274642944, |
|
"logits/rejected": -1.458261251449585, |
|
"logps/chosen": -200.50059509277344, |
|
"logps/rejected": -213.1103973388672, |
|
"loss": 0.6618, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": 0.3923037648200989, |
|
"rewards/margins": 0.1664755940437317, |
|
"rewards/rejected": 0.2258281409740448, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.11859454065996722, |
|
"grad_norm": 93.12582037555201, |
|
"learning_rate": 1.9999710687264071e-07, |
|
"logits/chosen": -1.311631441116333, |
|
"logits/rejected": -1.3278183937072754, |
|
"logps/chosen": -132.20401000976562, |
|
"logps/rejected": -157.69627380371094, |
|
"loss": 0.7135, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.19503216445446014, |
|
"rewards/margins": -0.017576567828655243, |
|
"rewards/rejected": 0.21260873973369598, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.12087520490342812, |
|
"grad_norm": 76.40018414480076, |
|
"learning_rate": 1.9999349050267698e-07, |
|
"logits/chosen": -1.291077733039856, |
|
"logits/rejected": -1.3958327770233154, |
|
"logps/chosen": -158.81573486328125, |
|
"logps/rejected": -199.284423828125, |
|
"loss": 0.6252, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.25081124901771545, |
|
"rewards/margins": 0.25090575218200684, |
|
"rewards/rejected": -9.45068895816803e-05, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.12315586914688903, |
|
"grad_norm": 90.49004280394047, |
|
"learning_rate": 1.9998842765796658e-07, |
|
"logits/chosen": -1.4403332471847534, |
|
"logits/rejected": -1.4392634630203247, |
|
"logps/chosen": -143.04052734375, |
|
"logps/rejected": -156.4816436767578, |
|
"loss": 0.6816, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.10938511788845062, |
|
"rewards/margins": 0.17133575677871704, |
|
"rewards/rejected": -0.06195063889026642, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.12543653339034994, |
|
"grad_norm": 81.5018944914688, |
|
"learning_rate": 1.9998191841174702e-07, |
|
"logits/chosen": -1.4619545936584473, |
|
"logits/rejected": -1.483961582183838, |
|
"logps/chosen": -120.11573028564453, |
|
"logps/rejected": -121.2934341430664, |
|
"loss": 0.6783, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.4725874662399292, |
|
"rewards/margins": 0.37433862686157227, |
|
"rewards/rejected": 0.09824882447719574, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.12771719763381084, |
|
"grad_norm": 91.38630396004429, |
|
"learning_rate": 1.9997396285817904e-07, |
|
"logits/chosen": -1.152282953262329, |
|
"logits/rejected": -1.252787470817566, |
|
"logps/chosen": -145.3370361328125, |
|
"logps/rejected": -174.6603240966797, |
|
"loss": 0.6269, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": 0.40825897455215454, |
|
"rewards/margins": 0.25122010707855225, |
|
"rewards/rejected": 0.1570388376712799, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.12999786187727175, |
|
"grad_norm": 97.62629025517847, |
|
"learning_rate": 1.9996456111234526e-07, |
|
"logits/chosen": -1.393836259841919, |
|
"logits/rejected": -1.5118399858474731, |
|
"logps/chosen": -173.96847534179688, |
|
"logps/rejected": -229.94232177734375, |
|
"loss": 0.6697, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.21259565651416779, |
|
"rewards/margins": 0.18360112607479095, |
|
"rewards/rejected": 0.028994524851441383, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.13227852612073265, |
|
"grad_norm": 96.19706483741392, |
|
"learning_rate": 1.9995371331024833e-07, |
|
"logits/chosen": -1.299769401550293, |
|
"logits/rejected": -1.3910255432128906, |
|
"logps/chosen": -112.2242660522461, |
|
"logps/rejected": -130.05230712890625, |
|
"loss": 0.7283, |
|
"rewards/accuracies": 0.4375, |
|
"rewards/chosen": 0.14800208806991577, |
|
"rewards/margins": -0.08275800198316574, |
|
"rewards/rejected": 0.23076008260250092, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.13455919036419356, |
|
"grad_norm": 114.37970019595372, |
|
"learning_rate": 1.999414196088092e-07, |
|
"logits/chosen": -1.3472720384597778, |
|
"logits/rejected": -1.4912211894989014, |
|
"logps/chosen": -202.3058319091797, |
|
"logps/rejected": -244.92904663085938, |
|
"loss": 0.7194, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": 0.3503240644931793, |
|
"rewards/margins": 0.024903126060962677, |
|
"rewards/rejected": 0.32542091608047485, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.1368398546076545, |
|
"grad_norm": 87.71566928589093, |
|
"learning_rate": 1.9992768018586478e-07, |
|
"logits/chosen": -1.3490660190582275, |
|
"logits/rejected": -1.3772616386413574, |
|
"logps/chosen": -146.59048461914062, |
|
"logps/rejected": -159.0959930419922, |
|
"loss": 0.6736, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": 0.3947565257549286, |
|
"rewards/margins": 0.15041925013065338, |
|
"rewards/rejected": 0.244337260723114, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.1391205188511154, |
|
"grad_norm": 109.98286067474862, |
|
"learning_rate": 1.9991249524016528e-07, |
|
"logits/chosen": -1.2736543416976929, |
|
"logits/rejected": -1.4115746021270752, |
|
"logps/chosen": -150.3526611328125, |
|
"logps/rejected": -189.20545959472656, |
|
"loss": 0.761, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": 0.28761598467826843, |
|
"rewards/margins": -0.15278732776641846, |
|
"rewards/rejected": 0.44040337204933167, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.1414011830945763, |
|
"grad_norm": 108.93431507211488, |
|
"learning_rate": 1.9989586499137135e-07, |
|
"logits/chosen": -1.2978495359420776, |
|
"logits/rejected": -1.3472024202346802, |
|
"logps/chosen": -179.5664825439453, |
|
"logps/rejected": -203.53968811035156, |
|
"loss": 0.7215, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": 0.39709562063217163, |
|
"rewards/margins": 0.20268934965133667, |
|
"rewards/rejected": 0.19440627098083496, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.1436818473380372, |
|
"grad_norm": 99.31026397357329, |
|
"learning_rate": 1.9987778968005106e-07, |
|
"logits/chosen": -1.2714462280273438, |
|
"logits/rejected": -1.4020087718963623, |
|
"logps/chosen": -142.49758911132812, |
|
"logps/rejected": -262.75775146484375, |
|
"loss": 0.7111, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.1395024061203003, |
|
"rewards/margins": -0.11353214085102081, |
|
"rewards/rejected": 0.2530345320701599, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.14596251158149812, |
|
"grad_norm": 92.0460328359884, |
|
"learning_rate": 1.9985826956767617e-07, |
|
"logits/chosen": -1.474685788154602, |
|
"logits/rejected": -1.5739442110061646, |
|
"logps/chosen": -135.28616333007812, |
|
"logps/rejected": -165.1414031982422, |
|
"loss": 0.6874, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": 0.14916732907295227, |
|
"rewards/margins": 0.06242816895246506, |
|
"rewards/rejected": 0.0867391973733902, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.14824317582495902, |
|
"grad_norm": 96.22054918626448, |
|
"learning_rate": 1.9983730493661865e-07, |
|
"logits/chosen": -1.276041030883789, |
|
"logits/rejected": -1.3194851875305176, |
|
"logps/chosen": -221.15992736816406, |
|
"logps/rejected": -242.77647399902344, |
|
"loss": 0.6457, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.260553777217865, |
|
"rewards/margins": 0.26679596304893494, |
|
"rewards/rejected": -0.006242218893021345, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.15052384006841993, |
|
"grad_norm": 78.01858280488834, |
|
"learning_rate": 1.998148960901463e-07, |
|
"logits/chosen": -1.393154263496399, |
|
"logits/rejected": -1.423519492149353, |
|
"logps/chosen": -201.9635467529297, |
|
"logps/rejected": -235.25816345214844, |
|
"loss": 0.6893, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.382313072681427, |
|
"rewards/margins": 0.3356338441371918, |
|
"rewards/rejected": 0.04667920991778374, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.15280450431188083, |
|
"grad_norm": 103.00857915327929, |
|
"learning_rate": 1.997910433524185e-07, |
|
"logits/chosen": -1.3205957412719727, |
|
"logits/rejected": -1.3766120672225952, |
|
"logps/chosen": -189.32142639160156, |
|
"logps/rejected": -217.68544006347656, |
|
"loss": 0.724, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.25681746006011963, |
|
"rewards/margins": 0.15590888261795044, |
|
"rewards/rejected": 0.10090853273868561, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.15508516855534174, |
|
"grad_norm": 90.52690719042177, |
|
"learning_rate": 1.9976574706848153e-07, |
|
"logits/chosen": -1.4535361528396606, |
|
"logits/rejected": -1.4828790426254272, |
|
"logps/chosen": -171.7893829345703, |
|
"logps/rejected": -213.03846740722656, |
|
"loss": 0.6788, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": 0.33188748359680176, |
|
"rewards/margins": 0.07647443562746048, |
|
"rewards/rejected": 0.2554129958152771, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.15736583279880265, |
|
"grad_norm": 101.69492728925366, |
|
"learning_rate": 1.9973900760426363e-07, |
|
"logits/chosen": -1.4226858615875244, |
|
"logits/rejected": -1.4269222021102905, |
|
"logps/chosen": -166.29103088378906, |
|
"logps/rejected": -268.772705078125, |
|
"loss": 0.678, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": 0.3539034128189087, |
|
"rewards/margins": 0.052625708281993866, |
|
"rewards/rejected": 0.3012777268886566, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.15964649704226355, |
|
"grad_norm": 83.07835169069685, |
|
"learning_rate": 1.9971082534656955e-07, |
|
"logits/chosen": -1.37977933883667, |
|
"logits/rejected": -1.466927409172058, |
|
"logps/chosen": -113.37686157226562, |
|
"logps/rejected": -142.82138061523438, |
|
"loss": 0.6442, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.15869390964508057, |
|
"rewards/margins": 0.20419231057167053, |
|
"rewards/rejected": -0.045498400926589966, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.16192716128572446, |
|
"grad_norm": 103.54665420816978, |
|
"learning_rate": 1.99681200703075e-07, |
|
"logits/chosen": -1.4178937673568726, |
|
"logits/rejected": -1.4681955575942993, |
|
"logps/chosen": -158.346435546875, |
|
"logps/rejected": -180.1523895263672, |
|
"loss": 0.7127, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": 0.2817228436470032, |
|
"rewards/margins": 0.11103209853172302, |
|
"rewards/rejected": 0.17069074511528015, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.16420782552918536, |
|
"grad_norm": 77.71797198863521, |
|
"learning_rate": 1.9965013410232097e-07, |
|
"logits/chosen": -1.40109384059906, |
|
"logits/rejected": -1.4553627967834473, |
|
"logps/chosen": -172.07899475097656, |
|
"logps/rejected": -222.21438598632812, |
|
"loss": 0.6728, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": 0.27610713243484497, |
|
"rewards/margins": 0.35402336716651917, |
|
"rewards/rejected": -0.077916219830513, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.16648848977264627, |
|
"grad_norm": 91.27829192980697, |
|
"learning_rate": 1.996176259937072e-07, |
|
"logits/chosen": -1.3127285242080688, |
|
"logits/rejected": -1.3604404926300049, |
|
"logps/chosen": -168.33004760742188, |
|
"logps/rejected": -193.3489227294922, |
|
"loss": 0.679, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.18943265080451965, |
|
"rewards/margins": 0.11781211942434311, |
|
"rewards/rejected": 0.07162053138017654, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.1687691540161072, |
|
"grad_norm": 105.10949942036537, |
|
"learning_rate": 1.9958367684748586e-07, |
|
"logits/chosen": -1.301593542098999, |
|
"logits/rejected": -1.3123453855514526, |
|
"logps/chosen": -153.8624267578125, |
|
"logps/rejected": -172.3568115234375, |
|
"loss": 0.6824, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": 0.022907603532075882, |
|
"rewards/margins": 0.08112253248691559, |
|
"rewards/rejected": -0.058214932680130005, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.1710498182595681, |
|
"grad_norm": 99.17231684606652, |
|
"learning_rate": 1.995482871547548e-07, |
|
"logits/chosen": -1.2876176834106445, |
|
"logits/rejected": -1.3324649333953857, |
|
"logps/chosen": -127.41693115234375, |
|
"logps/rejected": -158.8002166748047, |
|
"loss": 0.6995, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": 0.21076691150665283, |
|
"rewards/margins": 0.24259139597415924, |
|
"rewards/rejected": -0.031824491918087006, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.17333048250302902, |
|
"grad_norm": 91.41364046696481, |
|
"learning_rate": 1.9951145742745024e-07, |
|
"logits/chosen": -1.2649379968643188, |
|
"logits/rejected": -1.4087891578674316, |
|
"logps/chosen": -130.08468627929688, |
|
"logps/rejected": -158.69525146484375, |
|
"loss": 0.7016, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.23989444971084595, |
|
"rewards/margins": 0.030172180384397507, |
|
"rewards/rejected": 0.20972229540348053, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.17561114674648992, |
|
"grad_norm": 75.7202165191751, |
|
"learning_rate": 1.9947318819833962e-07, |
|
"logits/chosen": -1.3358944654464722, |
|
"logits/rejected": -1.3563066720962524, |
|
"logps/chosen": -106.91460418701172, |
|
"logps/rejected": -113.2139892578125, |
|
"loss": 0.6607, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": 0.14014360308647156, |
|
"rewards/margins": -0.031036917120218277, |
|
"rewards/rejected": 0.17118054628372192, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.17789181098995083, |
|
"grad_norm": 109.3316094106793, |
|
"learning_rate": 1.994334800210137e-07, |
|
"logits/chosen": -1.3152189254760742, |
|
"logits/rejected": -1.3776334524154663, |
|
"logps/chosen": -160.3906707763672, |
|
"logps/rejected": -232.65725708007812, |
|
"loss": 0.7105, |
|
"rewards/accuracies": 0.4375, |
|
"rewards/chosen": 0.10319292545318604, |
|
"rewards/margins": -0.037254467606544495, |
|
"rewards/rejected": 0.14044740796089172, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.18017247523341173, |
|
"grad_norm": 84.0665837850057, |
|
"learning_rate": 1.9939233346987862e-07, |
|
"logits/chosen": -1.4214388132095337, |
|
"logits/rejected": -1.437930703163147, |
|
"logps/chosen": -172.41867065429688, |
|
"logps/rejected": -182.8797607421875, |
|
"loss": 0.6807, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.16843558847904205, |
|
"rewards/margins": 0.16182081401348114, |
|
"rewards/rejected": 0.0066147902980446815, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.18245313947687264, |
|
"grad_norm": 93.91811474012694, |
|
"learning_rate": 1.9934974914014762e-07, |
|
"logits/chosen": -1.3304110765457153, |
|
"logits/rejected": -1.386664628982544, |
|
"logps/chosen": -112.3897933959961, |
|
"logps/rejected": -133.53810119628906, |
|
"loss": 0.7031, |
|
"rewards/accuracies": 0.40625, |
|
"rewards/chosen": 0.19350233674049377, |
|
"rewards/margins": -0.08095596730709076, |
|
"rewards/rejected": 0.27445828914642334, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.18473380372033354, |
|
"grad_norm": 90.17492690496479, |
|
"learning_rate": 1.9930572764783236e-07, |
|
"logits/chosen": -1.2432975769042969, |
|
"logits/rejected": -1.2546998262405396, |
|
"logps/chosen": -167.08847045898438, |
|
"logps/rejected": -183.3931884765625, |
|
"loss": 0.6979, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.05589159578084946, |
|
"rewards/margins": 0.17001007497310638, |
|
"rewards/rejected": -0.11411847919225693, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.18701446796379445, |
|
"grad_norm": 95.44656254255437, |
|
"learning_rate": 1.9926026962973403e-07, |
|
"logits/chosen": -1.326186180114746, |
|
"logits/rejected": -1.4297109842300415, |
|
"logps/chosen": -183.224853515625, |
|
"logps/rejected": -239.97793579101562, |
|
"loss": 0.6436, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.16740243136882782, |
|
"rewards/margins": 0.23012831807136536, |
|
"rewards/rejected": -0.06272587180137634, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.18929513220725536, |
|
"grad_norm": 92.54409107962749, |
|
"learning_rate": 1.992133757434342e-07, |
|
"logits/chosen": -1.3230640888214111, |
|
"logits/rejected": -1.31871497631073, |
|
"logps/chosen": -226.2008819580078, |
|
"logps/rejected": -243.1304931640625, |
|
"loss": 0.6795, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.1608559638261795, |
|
"rewards/margins": 0.35546940565109253, |
|
"rewards/rejected": -0.1946134865283966, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.19157579645071626, |
|
"grad_norm": 93.78690688865437, |
|
"learning_rate": 1.991650466672853e-07, |
|
"logits/chosen": -1.2681585550308228, |
|
"logits/rejected": -1.27677321434021, |
|
"logps/chosen": -151.71714782714844, |
|
"logps/rejected": -171.88832092285156, |
|
"loss": 0.6772, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.21696476638317108, |
|
"rewards/margins": 0.30003517866134644, |
|
"rewards/rejected": -0.08307043462991714, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.19385646069417717, |
|
"grad_norm": 79.03039382866096, |
|
"learning_rate": 1.991152831004007e-07, |
|
"logits/chosen": -1.4594199657440186, |
|
"logits/rejected": -1.456141710281372, |
|
"logps/chosen": -169.4336700439453, |
|
"logps/rejected": -180.23756408691406, |
|
"loss": 0.695, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.091593436896801, |
|
"rewards/margins": 0.008565463125705719, |
|
"rewards/rejected": 0.08302795141935349, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.19613712493763807, |
|
"grad_norm": 84.01165418355862, |
|
"learning_rate": 1.9906408576264467e-07, |
|
"logits/chosen": -1.272233247756958, |
|
"logits/rejected": -1.309181809425354, |
|
"logps/chosen": -182.87887573242188, |
|
"logps/rejected": -218.7847442626953, |
|
"loss": 0.6868, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.13641594350337982, |
|
"rewards/margins": 0.20986737310886383, |
|
"rewards/rejected": -0.07345142215490341, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.198417789181099, |
|
"grad_norm": 95.73538074987611, |
|
"learning_rate": 1.9901145539462197e-07, |
|
"logits/chosen": -1.2831114530563354, |
|
"logits/rejected": -1.302902340888977, |
|
"logps/chosen": -147.5471954345703, |
|
"logps/rejected": -178.70321655273438, |
|
"loss": 0.6013, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.2684887647628784, |
|
"rewards/margins": 0.47333818674087524, |
|
"rewards/rejected": -0.20484942197799683, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.20069845342455991, |
|
"grad_norm": 102.90991917593935, |
|
"learning_rate": 1.9895739275766715e-07, |
|
"logits/chosen": -1.3818506002426147, |
|
"logits/rejected": -1.4092711210250854, |
|
"logps/chosen": -227.61679077148438, |
|
"logps/rejected": -226.88558959960938, |
|
"loss": 0.737, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.13025416433811188, |
|
"rewards/margins": 0.26329731941223145, |
|
"rewards/rejected": -0.13304319977760315, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.20297911766802082, |
|
"grad_norm": 105.89850456514735, |
|
"learning_rate": 1.9890189863383353e-07, |
|
"logits/chosen": -1.4667214155197144, |
|
"logits/rejected": -1.5101617574691772, |
|
"logps/chosen": -185.08575439453125, |
|
"logps/rejected": -237.65597534179688, |
|
"loss": 0.7025, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.2359275072813034, |
|
"rewards/margins": 0.08144809305667877, |
|
"rewards/rejected": 0.15447941422462463, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.20525978191148173, |
|
"grad_norm": 82.46406340773291, |
|
"learning_rate": 1.9884497382588183e-07, |
|
"logits/chosen": -1.2853760719299316, |
|
"logits/rejected": -1.3631393909454346, |
|
"logps/chosen": -140.35952758789062, |
|
"logps/rejected": -153.50978088378906, |
|
"loss": 0.6683, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.1511276513338089, |
|
"rewards/margins": 0.06823855638504028, |
|
"rewards/rejected": 0.08288908749818802, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.20754044615494263, |
|
"grad_norm": 90.32390874613019, |
|
"learning_rate": 1.9878661915726865e-07, |
|
"logits/chosen": -1.4392482042312622, |
|
"logits/rejected": -1.4409726858139038, |
|
"logps/chosen": -171.09303283691406, |
|
"logps/rejected": -182.13232421875, |
|
"loss": 0.6691, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": 0.20515748858451843, |
|
"rewards/margins": 0.10885492712259293, |
|
"rewards/rejected": 0.0963025838136673, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.20982111039840354, |
|
"grad_norm": 75.10429123148643, |
|
"learning_rate": 1.9872683547213446e-07, |
|
"logits/chosen": -1.3081107139587402, |
|
"logits/rejected": -1.317389965057373, |
|
"logps/chosen": -174.84463500976562, |
|
"logps/rejected": -210.18130493164062, |
|
"loss": 0.6158, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.18892019987106323, |
|
"rewards/margins": 0.4061851501464844, |
|
"rewards/rejected": -0.21726495027542114, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.21210177464186444, |
|
"grad_norm": 107.2216966034497, |
|
"learning_rate": 1.9866562363529144e-07, |
|
"logits/chosen": -1.2954977750778198, |
|
"logits/rejected": -1.3127189874649048, |
|
"logps/chosen": -197.97561645507812, |
|
"logps/rejected": -204.29013061523438, |
|
"loss": 0.7154, |
|
"rewards/accuracies": 0.4375, |
|
"rewards/chosen": 0.17145520448684692, |
|
"rewards/margins": -0.03724336624145508, |
|
"rewards/rejected": 0.208698570728302, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.21438243888532535, |
|
"grad_norm": 96.60657963150389, |
|
"learning_rate": 1.9860298453221107e-07, |
|
"logits/chosen": -1.3616856336593628, |
|
"logits/rejected": -1.3934649229049683, |
|
"logps/chosen": -179.2027130126953, |
|
"logps/rejected": -187.0948028564453, |
|
"loss": 0.6701, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": 0.0954272449016571, |
|
"rewards/margins": -0.021092543378472328, |
|
"rewards/rejected": 0.11651977896690369, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.21666310312878626, |
|
"grad_norm": 87.8142315385882, |
|
"learning_rate": 1.9853891906901108e-07, |
|
"logits/chosen": -1.3980423212051392, |
|
"logits/rejected": -1.4947185516357422, |
|
"logps/chosen": -142.8896942138672, |
|
"logps/rejected": -180.8528594970703, |
|
"loss": 0.6415, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.13048064708709717, |
|
"rewards/margins": 0.3490259051322937, |
|
"rewards/rejected": -0.21854525804519653, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.21894376737224716, |
|
"grad_norm": 91.22189567457308, |
|
"learning_rate": 1.9847342817244254e-07, |
|
"logits/chosen": -1.3708425760269165, |
|
"logits/rejected": -1.3748908042907715, |
|
"logps/chosen": -155.70945739746094, |
|
"logps/rejected": -175.20013427734375, |
|
"loss": 0.6507, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": 0.3499588370323181, |
|
"rewards/margins": 0.12689019739627838, |
|
"rewards/rejected": 0.22306862473487854, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.22122443161570807, |
|
"grad_norm": 103.90331869438803, |
|
"learning_rate": 1.9840651278987642e-07, |
|
"logits/chosen": -1.3758065700531006, |
|
"logits/rejected": -1.3824987411499023, |
|
"logps/chosen": -181.644775390625, |
|
"logps/rejected": -178.72109985351562, |
|
"loss": 0.6563, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": 0.29924964904785156, |
|
"rewards/margins": 0.3290708065032959, |
|
"rewards/rejected": -0.02982119843363762, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.22350509585916897, |
|
"grad_norm": 103.8705512922574, |
|
"learning_rate": 1.9833817388928984e-07, |
|
"logits/chosen": -1.4562551975250244, |
|
"logits/rejected": -1.511305809020996, |
|
"logps/chosen": -168.09715270996094, |
|
"logps/rejected": -189.35787963867188, |
|
"loss": 0.6926, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": 0.31778958439826965, |
|
"rewards/margins": 0.04199263080954552, |
|
"rewards/rejected": 0.27579694986343384, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.22578576010262988, |
|
"grad_norm": 103.40331674623164, |
|
"learning_rate": 1.982684124592521e-07, |
|
"logits/chosen": -1.3754979372024536, |
|
"logits/rejected": -1.347690463066101, |
|
"logps/chosen": -168.61148071289062, |
|
"logps/rejected": -186.8193359375, |
|
"loss": 0.7315, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": 0.14702042937278748, |
|
"rewards/margins": 0.059913113713264465, |
|
"rewards/rejected": 0.08710730820894241, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.2280664243460908, |
|
"grad_norm": 102.65405433185128, |
|
"learning_rate": 1.981972295089103e-07, |
|
"logits/chosen": -1.4264771938323975, |
|
"logits/rejected": -1.5102254152297974, |
|
"logps/chosen": -160.4220428466797, |
|
"logps/rejected": -173.62269592285156, |
|
"loss": 0.6553, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.3021972179412842, |
|
"rewards/margins": 0.10177014023065567, |
|
"rewards/rejected": 0.2004270851612091, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.2280664243460908, |
|
"eval_logits/chosen": -1.4570631980895996, |
|
"eval_logits/rejected": -1.442795753479004, |
|
"eval_logps/chosen": -127.66089630126953, |
|
"eval_logps/rejected": -129.5493621826172, |
|
"eval_loss": 0.6965904235839844, |
|
"eval_rewards/accuracies": 0.5199999809265137, |
|
"eval_rewards/chosen": 0.22370710968971252, |
|
"eval_rewards/margins": -0.02763114497065544, |
|
"eval_rewards/rejected": 0.2513382136821747, |
|
"eval_runtime": 22.1102, |
|
"eval_samples_per_second": 4.523, |
|
"eval_steps_per_second": 1.131, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.23034708858955172, |
|
"grad_norm": 70.85461032661505, |
|
"learning_rate": 1.9812462606797498e-07, |
|
"logits/chosen": -1.1571446657180786, |
|
"logits/rejected": -1.2567036151885986, |
|
"logps/chosen": -170.15008544921875, |
|
"logps/rejected": -181.82806396484375, |
|
"loss": 0.6236, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.13917526602745056, |
|
"rewards/margins": 0.30568423867225647, |
|
"rewards/rejected": -0.1665089875459671, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.23262775283301262, |
|
"grad_norm": 101.20833615282186, |
|
"learning_rate": 1.9805060318670485e-07, |
|
"logits/chosen": -1.3708992004394531, |
|
"logits/rejected": -1.3186278343200684, |
|
"logps/chosen": -154.29226684570312, |
|
"logps/rejected": -148.5332794189453, |
|
"loss": 0.6319, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": 0.09488646686077118, |
|
"rewards/margins": 0.11530336737632751, |
|
"rewards/rejected": -0.020416898652911186, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.23490841707647353, |
|
"grad_norm": 96.32887019136864, |
|
"learning_rate": 1.9797516193589192e-07, |
|
"logits/chosen": -1.3597060441970825, |
|
"logits/rejected": -1.3876497745513916, |
|
"logps/chosen": -101.66006469726562, |
|
"logps/rejected": -131.19041442871094, |
|
"loss": 0.7312, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": 0.20645485818386078, |
|
"rewards/margins": 0.06082191318273544, |
|
"rewards/rejected": 0.14563296735286713, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.23718908131993444, |
|
"grad_norm": 91.31654055574913, |
|
"learning_rate": 1.978983034068459e-07, |
|
"logits/chosen": -1.3642208576202393, |
|
"logits/rejected": -1.405060052871704, |
|
"logps/chosen": -188.59420776367188, |
|
"logps/rejected": -199.42637634277344, |
|
"loss": 0.6922, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.13877716660499573, |
|
"rewards/margins": 0.17159023880958557, |
|
"rewards/rejected": -0.03281306475400925, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.23946974556339534, |
|
"grad_norm": 99.31042679769261, |
|
"learning_rate": 1.9782002871137832e-07, |
|
"logits/chosen": -1.3214318752288818, |
|
"logits/rejected": -1.3617138862609863, |
|
"logps/chosen": -195.82009887695312, |
|
"logps/rejected": -214.06614685058594, |
|
"loss": 0.6673, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.16799718141555786, |
|
"rewards/margins": 0.25821763277053833, |
|
"rewards/rejected": -0.09022042155265808, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.24175040980685625, |
|
"grad_norm": 103.3617201342953, |
|
"learning_rate": 1.9774033898178666e-07, |
|
"logits/chosen": -1.4174686670303345, |
|
"logits/rejected": -1.420326590538025, |
|
"logps/chosen": -129.22848510742188, |
|
"logps/rejected": -145.9871826171875, |
|
"loss": 0.6716, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.3314517140388489, |
|
"rewards/margins": 0.1635245978832245, |
|
"rewards/rejected": 0.167927086353302, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.24403107405031715, |
|
"grad_norm": 85.24728778878888, |
|
"learning_rate": 1.9765923537083774e-07, |
|
"logits/chosen": -1.243302583694458, |
|
"logits/rejected": -1.2585264444351196, |
|
"logps/chosen": -181.84434509277344, |
|
"logps/rejected": -238.97323608398438, |
|
"loss": 0.6159, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.0770760029554367, |
|
"rewards/margins": 0.2519676089286804, |
|
"rewards/rejected": -0.17489157617092133, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.24631173829377806, |
|
"grad_norm": 96.37551956943565, |
|
"learning_rate": 1.9757671905175115e-07, |
|
"logits/chosen": -1.3951995372772217, |
|
"logits/rejected": -1.4096182584762573, |
|
"logps/chosen": -119.3514404296875, |
|
"logps/rejected": -131.9920196533203, |
|
"loss": 0.6021, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.2738375961780548, |
|
"rewards/margins": 0.2932060658931732, |
|
"rewards/rejected": -0.01936846226453781, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.24859240253723897, |
|
"grad_norm": 90.01951430126202, |
|
"learning_rate": 1.9749279121818234e-07, |
|
"logits/chosen": -1.3084944486618042, |
|
"logits/rejected": -1.3836894035339355, |
|
"logps/chosen": -155.07765197753906, |
|
"logps/rejected": -191.67471313476562, |
|
"loss": 0.6953, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.12750835716724396, |
|
"rewards/margins": 0.31851378083229065, |
|
"rewards/rejected": -0.1910054087638855, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.25087306678069987, |
|
"grad_norm": 107.56947817547281, |
|
"learning_rate": 1.9740745308420528e-07, |
|
"logits/chosen": -1.4428541660308838, |
|
"logits/rejected": -1.4522430896759033, |
|
"logps/chosen": -159.34136962890625, |
|
"logps/rejected": -201.03256225585938, |
|
"loss": 0.7085, |
|
"rewards/accuracies": 0.4375, |
|
"rewards/chosen": 0.08158425986766815, |
|
"rewards/margins": 0.062202394008636475, |
|
"rewards/rejected": 0.019381869584321976, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.2531537310241608, |
|
"grad_norm": 83.84650346626758, |
|
"learning_rate": 1.9732070588429487e-07, |
|
"logits/chosen": -1.4249211549758911, |
|
"logits/rejected": -1.4040560722351074, |
|
"logps/chosen": -139.1720733642578, |
|
"logps/rejected": -150.09683227539062, |
|
"loss": 0.6528, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.25594449043273926, |
|
"rewards/margins": 0.27077823877334595, |
|
"rewards/rejected": -0.014833783730864525, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.2554343952676217, |
|
"grad_norm": 95.84521399355921, |
|
"learning_rate": 1.972325508733091e-07, |
|
"logits/chosen": -1.4068944454193115, |
|
"logits/rejected": -1.4884459972381592, |
|
"logps/chosen": -178.6918182373047, |
|
"logps/rejected": -178.89633178710938, |
|
"loss": 0.6725, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.17113548517227173, |
|
"rewards/margins": 0.35012179613113403, |
|
"rewards/rejected": -0.1789863258600235, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.2577150595110826, |
|
"grad_norm": 90.48418130257966, |
|
"learning_rate": 1.9714298932647098e-07, |
|
"logits/chosen": -1.3818295001983643, |
|
"logits/rejected": -1.4686387777328491, |
|
"logps/chosen": -203.4779510498047, |
|
"logps/rejected": -231.25755310058594, |
|
"loss": 0.6145, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.2646425664424896, |
|
"rewards/margins": 0.30655670166015625, |
|
"rewards/rejected": -0.04191412031650543, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.2599957237545435, |
|
"grad_norm": 96.02415026722944, |
|
"learning_rate": 1.9705202253935004e-07, |
|
"logits/chosen": -1.459298849105835, |
|
"logits/rejected": -1.4874234199523926, |
|
"logps/chosen": -190.43370056152344, |
|
"logps/rejected": -208.5654296875, |
|
"loss": 0.6259, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.2924637198448181, |
|
"rewards/margins": 0.3736059367656708, |
|
"rewards/rejected": -0.08114223182201385, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.2622763879980044, |
|
"grad_norm": 94.75504400212658, |
|
"learning_rate": 1.9695965182784343e-07, |
|
"logits/chosen": -1.37540864944458, |
|
"logits/rejected": -1.4327397346496582, |
|
"logps/chosen": -158.9178924560547, |
|
"logps/rejected": -187.00572204589844, |
|
"loss": 0.6413, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.031393878161907196, |
|
"rewards/margins": 0.2354445904493332, |
|
"rewards/rejected": -0.2040507048368454, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.2645570522414653, |
|
"grad_norm": 88.68651480780905, |
|
"learning_rate": 1.968658785281573e-07, |
|
"logits/chosen": -1.4091265201568604, |
|
"logits/rejected": -1.3750542402267456, |
|
"logps/chosen": -159.5924530029297, |
|
"logps/rejected": -157.25701904296875, |
|
"loss": 0.6956, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.05787750333547592, |
|
"rewards/margins": 0.07036175578832626, |
|
"rewards/rejected": -0.012484237551689148, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.2668377164849262, |
|
"grad_norm": 87.80878700459785, |
|
"learning_rate": 1.9677070399678693e-07, |
|
"logits/chosen": -1.3906813859939575, |
|
"logits/rejected": -1.4185580015182495, |
|
"logps/chosen": -170.7882843017578, |
|
"logps/rejected": -192.6103515625, |
|
"loss": 0.6498, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.026271583512425423, |
|
"rewards/margins": 0.08472327888011932, |
|
"rewards/rejected": -0.058451712131500244, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.2691183807283871, |
|
"grad_norm": 95.5482100990242, |
|
"learning_rate": 1.9667412961049754e-07, |
|
"logits/chosen": -1.4015161991119385, |
|
"logits/rejected": -1.4669052362442017, |
|
"logps/chosen": -180.34584045410156, |
|
"logps/rejected": -205.79164123535156, |
|
"loss": 0.6991, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": 0.12791365385055542, |
|
"rewards/margins": 0.28303825855255127, |
|
"rewards/rejected": -0.15512457489967346, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.271399044971848, |
|
"grad_norm": 88.45630335400043, |
|
"learning_rate": 1.9657615676630418e-07, |
|
"logits/chosen": -1.2741804122924805, |
|
"logits/rejected": -1.2581013441085815, |
|
"logps/chosen": -184.23312377929688, |
|
"logps/rejected": -201.00103759765625, |
|
"loss": 0.6514, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.036131352186203, |
|
"rewards/margins": 0.508301317691803, |
|
"rewards/rejected": -0.4721699059009552, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.273679709215309, |
|
"grad_norm": 99.3871661426225, |
|
"learning_rate": 1.9647678688145159e-07, |
|
"logits/chosen": -1.327433466911316, |
|
"logits/rejected": -1.4025075435638428, |
|
"logps/chosen": -167.82005310058594, |
|
"logps/rejected": -194.77706909179688, |
|
"loss": 0.6544, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.01046083215624094, |
|
"rewards/margins": 0.2960438132286072, |
|
"rewards/rejected": -0.30650466680526733, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.2759603734587699, |
|
"grad_norm": 129.18874275194366, |
|
"learning_rate": 1.9637602139339355e-07, |
|
"logits/chosen": -1.1904761791229248, |
|
"logits/rejected": -1.2964377403259277, |
|
"logps/chosen": -198.83749389648438, |
|
"logps/rejected": -213.1676025390625, |
|
"loss": 0.7065, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.018888656049966812, |
|
"rewards/margins": 0.0817704051733017, |
|
"rewards/rejected": -0.06288175284862518, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.2782410377022308, |
|
"grad_norm": 78.46019369342993, |
|
"learning_rate": 1.9627386175977238e-07, |
|
"logits/chosen": -1.363703966140747, |
|
"logits/rejected": -1.385565161705017, |
|
"logps/chosen": -158.67543029785156, |
|
"logps/rejected": -171.18373107910156, |
|
"loss": 0.6328, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.0427667461335659, |
|
"rewards/margins": 0.24535968899726868, |
|
"rewards/rejected": -0.20259293913841248, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.2805217019456917, |
|
"grad_norm": 93.42637414011723, |
|
"learning_rate": 1.9617030945839747e-07, |
|
"logits/chosen": -1.4135775566101074, |
|
"logits/rejected": -1.482433557510376, |
|
"logps/chosen": -201.14869689941406, |
|
"logps/rejected": -240.91909790039062, |
|
"loss": 0.6572, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": 0.14201030135154724, |
|
"rewards/margins": 0.3642757833003998, |
|
"rewards/rejected": -0.22226551175117493, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.2828023661891526, |
|
"grad_norm": 94.8034192613356, |
|
"learning_rate": 1.9606536598722431e-07, |
|
"logits/chosen": -1.3897498846054077, |
|
"logits/rejected": -1.4480516910552979, |
|
"logps/chosen": -211.97976684570312, |
|
"logps/rejected": -213.59344482421875, |
|
"loss": 0.6947, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.1269560605287552, |
|
"rewards/margins": 0.18549999594688416, |
|
"rewards/rejected": -0.05854398384690285, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.2850830304326135, |
|
"grad_norm": 107.48956854367718, |
|
"learning_rate": 1.9595903286433253e-07, |
|
"logits/chosen": -1.3192163705825806, |
|
"logits/rejected": -1.3453290462493896, |
|
"logps/chosen": -166.43765258789062, |
|
"logps/rejected": -185.8678436279297, |
|
"loss": 0.6446, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.22617535293102264, |
|
"rewards/margins": 0.19618834555149078, |
|
"rewards/rejected": -0.4223636984825134, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.2873636946760744, |
|
"grad_norm": 97.18344743194635, |
|
"learning_rate": 1.9585131162790395e-07, |
|
"logits/chosen": -1.2871501445770264, |
|
"logits/rejected": -1.3303256034851074, |
|
"logps/chosen": -169.388671875, |
|
"logps/rejected": -222.89968872070312, |
|
"loss": 0.6597, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.016080807894468307, |
|
"rewards/margins": 0.22609557211399078, |
|
"rewards/rejected": -0.21001477539539337, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.2896443589195353, |
|
"grad_norm": 102.3447949736702, |
|
"learning_rate": 1.9574220383620054e-07, |
|
"logits/chosen": -1.332809329032898, |
|
"logits/rejected": -1.3781144618988037, |
|
"logps/chosen": -192.14796447753906, |
|
"logps/rejected": -229.9279327392578, |
|
"loss": 0.6849, |
|
"rewards/accuracies": 0.4375, |
|
"rewards/chosen": -0.12542986869812012, |
|
"rewards/margins": 0.2946632206439972, |
|
"rewards/rejected": -0.4200930893421173, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.29192502316299623, |
|
"grad_norm": 95.89945385250834, |
|
"learning_rate": 1.956317110675417e-07, |
|
"logits/chosen": -1.3515238761901855, |
|
"logits/rejected": -1.3907783031463623, |
|
"logps/chosen": -135.85316467285156, |
|
"logps/rejected": -168.83544921875, |
|
"loss": 0.6599, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.2081318199634552, |
|
"rewards/margins": 0.3642672896385193, |
|
"rewards/rejected": -0.15613549947738647, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.29420568740645714, |
|
"grad_norm": 104.38674521939737, |
|
"learning_rate": 1.9551983492028144e-07, |
|
"logits/chosen": -1.3251550197601318, |
|
"logits/rejected": -1.418371319770813, |
|
"logps/chosen": -172.90769958496094, |
|
"logps/rejected": -200.890380859375, |
|
"loss": 0.6987, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.051881127059459686, |
|
"rewards/margins": 0.18907570838928223, |
|
"rewards/rejected": -0.24095678329467773, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.29648635164991805, |
|
"grad_norm": 110.83873967012946, |
|
"learning_rate": 1.9540657701278533e-07, |
|
"logits/chosen": -1.4099152088165283, |
|
"logits/rejected": -1.4348920583724976, |
|
"logps/chosen": -196.08592224121094, |
|
"logps/rejected": -237.6608428955078, |
|
"loss": 0.6704, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.22441606223583221, |
|
"rewards/margins": 0.2391555905342102, |
|
"rewards/rejected": -0.4635716676712036, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.29876701589337895, |
|
"grad_norm": 103.49844878423605, |
|
"learning_rate": 1.952919389834071e-07, |
|
"logits/chosen": -1.2992289066314697, |
|
"logits/rejected": -1.3971199989318848, |
|
"logps/chosen": -173.28128051757812, |
|
"logps/rejected": -213.29893493652344, |
|
"loss": 0.6847, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.19963645935058594, |
|
"rewards/margins": 0.11320970952510834, |
|
"rewards/rejected": 0.0864267572760582, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.30104768013683986, |
|
"grad_norm": 99.13994519459824, |
|
"learning_rate": 1.9517592249046475e-07, |
|
"logits/chosen": -1.2953108549118042, |
|
"logits/rejected": -1.3212617635726929, |
|
"logps/chosen": -133.38534545898438, |
|
"logps/rejected": -153.64303588867188, |
|
"loss": 0.6807, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": 0.2558751106262207, |
|
"rewards/margins": 0.31515762209892273, |
|
"rewards/rejected": -0.059282511472702026, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.30332834438030076, |
|
"grad_norm": 105.16807067574956, |
|
"learning_rate": 1.950585292122169e-07, |
|
"logits/chosen": -1.380258560180664, |
|
"logits/rejected": -1.381796956062317, |
|
"logps/chosen": -265.73822021484375, |
|
"logps/rejected": -279.2653503417969, |
|
"loss": 0.6065, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.020402366295456886, |
|
"rewards/margins": 0.3811405599117279, |
|
"rewards/rejected": -0.36073821783065796, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.30560900862376167, |
|
"grad_norm": 83.84465669976083, |
|
"learning_rate": 1.9493976084683813e-07, |
|
"logits/chosen": -1.3013784885406494, |
|
"logits/rejected": -1.3473682403564453, |
|
"logps/chosen": -143.68994140625, |
|
"logps/rejected": -161.75987243652344, |
|
"loss": 0.6565, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.09173958003520966, |
|
"rewards/margins": 0.3339024484157562, |
|
"rewards/rejected": -0.4256420433521271, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.3078896728672226, |
|
"grad_norm": 97.4032150220662, |
|
"learning_rate": 1.9481961911239475e-07, |
|
"logits/chosen": -1.3531267642974854, |
|
"logits/rejected": -1.3918657302856445, |
|
"logps/chosen": -158.61968994140625, |
|
"logps/rejected": -196.80203247070312, |
|
"loss": 0.6872, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.018977202475070953, |
|
"rewards/margins": 0.31147533655166626, |
|
"rewards/rejected": -0.330452561378479, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.3101703371106835, |
|
"grad_norm": 102.05098535340969, |
|
"learning_rate": 1.9469810574681968e-07, |
|
"logits/chosen": -1.2518314123153687, |
|
"logits/rejected": -1.373193383216858, |
|
"logps/chosen": -208.13992309570312, |
|
"logps/rejected": -265.2911376953125, |
|
"loss": 0.6599, |
|
"rewards/accuracies": 0.78125, |
|
"rewards/chosen": -0.10957180708646774, |
|
"rewards/margins": 0.4111281931400299, |
|
"rewards/rejected": -0.5207000374794006, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.3124510013541444, |
|
"grad_norm": 101.2119365968094, |
|
"learning_rate": 1.9457522250788756e-07, |
|
"logits/chosen": -1.4733508825302124, |
|
"logits/rejected": -1.4335081577301025, |
|
"logps/chosen": -152.27394104003906, |
|
"logps/rejected": -149.08934020996094, |
|
"loss": 0.6512, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": 0.0012259185314178467, |
|
"rewards/margins": 0.06475882232189178, |
|
"rewards/rejected": -0.06353290379047394, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.3147316655976053, |
|
"grad_norm": 102.70187300991479, |
|
"learning_rate": 1.9445097117318912e-07, |
|
"logits/chosen": -1.2339739799499512, |
|
"logits/rejected": -1.2878998517990112, |
|
"logps/chosen": -154.39866638183594, |
|
"logps/rejected": -169.81304931640625, |
|
"loss": 0.6069, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.030073221772909164, |
|
"rewards/margins": 0.06755417585372925, |
|
"rewards/rejected": -0.09762738645076752, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.3170123298410662, |
|
"grad_norm": 92.89132088481786, |
|
"learning_rate": 1.9432535354010545e-07, |
|
"logits/chosen": -1.381090521812439, |
|
"logits/rejected": -1.408646583557129, |
|
"logps/chosen": -193.48159790039062, |
|
"logps/rejected": -218.6549530029297, |
|
"loss": 0.721, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.018713481724262238, |
|
"rewards/margins": 0.14033427834510803, |
|
"rewards/rejected": -0.15904778242111206, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.3192929940845271, |
|
"grad_norm": 92.57611013262112, |
|
"learning_rate": 1.9419837142578226e-07, |
|
"logits/chosen": -1.3027143478393555, |
|
"logits/rejected": -1.2925523519515991, |
|
"logps/chosen": -159.863525390625, |
|
"logps/rejected": -182.41249084472656, |
|
"loss": 0.6453, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": 0.16124297678470612, |
|
"rewards/margins": 0.37421250343322754, |
|
"rewards/rejected": -0.21296954154968262, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.321573658327988, |
|
"grad_norm": 88.34893118769534, |
|
"learning_rate": 1.9407002666710333e-07, |
|
"logits/chosen": -1.2514938116073608, |
|
"logits/rejected": -1.2840971946716309, |
|
"logps/chosen": -167.23568725585938, |
|
"logps/rejected": -185.19515991210938, |
|
"loss": 0.6776, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.0749707818031311, |
|
"rewards/margins": 0.2606170177459717, |
|
"rewards/rejected": -0.18564626574516296, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.3238543225714489, |
|
"grad_norm": 87.77224347782207, |
|
"learning_rate": 1.93940321120664e-07, |
|
"logits/chosen": -1.234694242477417, |
|
"logits/rejected": -1.245900273323059, |
|
"logps/chosen": -181.2189178466797, |
|
"logps/rejected": -201.54295349121094, |
|
"loss": 0.656, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.03384766727685928, |
|
"rewards/margins": 0.15053534507751465, |
|
"rewards/rejected": -0.18438303470611572, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.3261349868149098, |
|
"grad_norm": 102.48592883969225, |
|
"learning_rate": 1.9380925666274443e-07, |
|
"logits/chosen": -1.29789400100708, |
|
"logits/rejected": -1.3402178287506104, |
|
"logps/chosen": -200.9523468017578, |
|
"logps/rejected": -274.8287658691406, |
|
"loss": 0.6817, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.21383914351463318, |
|
"rewards/margins": 0.0331435352563858, |
|
"rewards/rejected": -0.24698270857334137, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.32841565105837073, |
|
"grad_norm": 98.03735612419584, |
|
"learning_rate": 1.9367683518928225e-07, |
|
"logits/chosen": -1.3940753936767578, |
|
"logits/rejected": -1.464486837387085, |
|
"logps/chosen": -124.9957504272461, |
|
"logps/rejected": -153.5186767578125, |
|
"loss": 0.6812, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.005818367004394531, |
|
"rewards/margins": 0.20061178505420685, |
|
"rewards/rejected": -0.20643013715744019, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.33069631530183163, |
|
"grad_norm": 103.78616077769101, |
|
"learning_rate": 1.935430586158454e-07, |
|
"logits/chosen": -1.1263610124588013, |
|
"logits/rejected": -1.1101529598236084, |
|
"logps/chosen": -117.66938018798828, |
|
"logps/rejected": -164.22933959960938, |
|
"loss": 0.6884, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.0009657144546508789, |
|
"rewards/margins": 0.10708783566951752, |
|
"rewards/rejected": -0.1080535426735878, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.33297697954529254, |
|
"grad_norm": 110.5063902175955, |
|
"learning_rate": 1.9340792887760412e-07, |
|
"logits/chosen": -1.324884295463562, |
|
"logits/rejected": -1.3702059984207153, |
|
"logps/chosen": -194.0547332763672, |
|
"logps/rejected": -223.9933624267578, |
|
"loss": 0.6334, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.0599634051322937, |
|
"rewards/margins": 0.17343786358833313, |
|
"rewards/rejected": -0.23340126872062683, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.3352576437887535, |
|
"grad_norm": 85.87924361401205, |
|
"learning_rate": 1.932714479293033e-07, |
|
"logits/chosen": -1.2896180152893066, |
|
"logits/rejected": -1.3805060386657715, |
|
"logps/chosen": -182.65460205078125, |
|
"logps/rejected": -212.5992889404297, |
|
"loss": 0.6151, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.10547913610935211, |
|
"rewards/margins": 0.20131757855415344, |
|
"rewards/rejected": -0.30679669976234436, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.3375383080322144, |
|
"grad_norm": 108.47177643694172, |
|
"learning_rate": 1.9313361774523386e-07, |
|
"logits/chosen": -1.4138604402542114, |
|
"logits/rejected": -1.402896523475647, |
|
"logps/chosen": -147.80795288085938, |
|
"logps/rejected": -165.24566650390625, |
|
"loss": 0.7377, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.21476852893829346, |
|
"rewards/margins": 0.08000420778989792, |
|
"rewards/rejected": -0.2947727143764496, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.3398189722756753, |
|
"grad_norm": 111.8896748854085, |
|
"learning_rate": 1.9299444031920436e-07, |
|
"logits/chosen": -1.2892051935195923, |
|
"logits/rejected": -1.380386233329773, |
|
"logps/chosen": -190.3002471923828, |
|
"logps/rejected": -246.29263305664062, |
|
"loss": 0.6347, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.08067283034324646, |
|
"rewards/margins": 0.4462183713912964, |
|
"rewards/rejected": -0.5268911719322205, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.3420996365191362, |
|
"grad_norm": 96.80869969762502, |
|
"learning_rate": 1.9285391766451217e-07, |
|
"logits/chosen": -1.3364348411560059, |
|
"logits/rejected": -1.3174083232879639, |
|
"logps/chosen": -161.74612426757812, |
|
"logps/rejected": -179.8140106201172, |
|
"loss": 0.669, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.1118321567773819, |
|
"rewards/margins": 0.23199215531349182, |
|
"rewards/rejected": -0.34382426738739014, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.3420996365191362, |
|
"eval_logits/chosen": -1.4281266927719116, |
|
"eval_logits/rejected": -1.4122414588928223, |
|
"eval_logps/chosen": -128.86663818359375, |
|
"eval_logps/rejected": -131.4814910888672, |
|
"eval_loss": 0.6797733306884766, |
|
"eval_rewards/accuracies": 0.5199999809265137, |
|
"eval_rewards/chosen": 0.10313291847705841, |
|
"eval_rewards/margins": 0.045006848871707916, |
|
"eval_rewards/rejected": 0.05812607705593109, |
|
"eval_runtime": 21.8566, |
|
"eval_samples_per_second": 4.575, |
|
"eval_steps_per_second": 1.144, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.3443803007625971, |
|
"grad_norm": 84.25578372291146, |
|
"learning_rate": 1.927120518139144e-07, |
|
"logits/chosen": -1.3597999811172485, |
|
"logits/rejected": -1.4241136312484741, |
|
"logps/chosen": -180.1940460205078, |
|
"logps/rejected": -199.74412536621094, |
|
"loss": 0.6421, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.04787690192461014, |
|
"rewards/margins": 0.13789615035057068, |
|
"rewards/rejected": -0.18577302992343903, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.34666096500605803, |
|
"grad_norm": 89.24528063714898, |
|
"learning_rate": 1.925688448195983e-07, |
|
"logits/chosen": -1.1900302171707153, |
|
"logits/rejected": -1.3024637699127197, |
|
"logps/chosen": -130.7202606201172, |
|
"logps/rejected": -170.3651123046875, |
|
"loss": 0.6326, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": 0.08976805210113525, |
|
"rewards/margins": 0.30762431025505066, |
|
"rewards/rejected": -0.2178562879562378, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.34894162924951894, |
|
"grad_norm": 86.12361588594867, |
|
"learning_rate": 1.924242987531517e-07, |
|
"logits/chosen": -1.3129158020019531, |
|
"logits/rejected": -1.2686495780944824, |
|
"logps/chosen": -170.01808166503906, |
|
"logps/rejected": -172.82835388183594, |
|
"loss": 0.6621, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.010694226250052452, |
|
"rewards/margins": 0.25143933296203613, |
|
"rewards/rejected": -0.26213353872299194, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.35122229349297984, |
|
"grad_norm": 80.23753458776699, |
|
"learning_rate": 1.922784157055331e-07, |
|
"logits/chosen": -1.3715269565582275, |
|
"logits/rejected": -1.3273383378982544, |
|
"logps/chosen": -190.35610961914062, |
|
"logps/rejected": -180.68020629882812, |
|
"loss": 0.6522, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.09160695225000381, |
|
"rewards/margins": 0.13991902768611908, |
|
"rewards/rejected": -0.2315259724855423, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.35350295773644075, |
|
"grad_norm": 98.71267017165292, |
|
"learning_rate": 1.9213119778704127e-07, |
|
"logits/chosen": -1.3962262868881226, |
|
"logits/rejected": -1.4428459405899048, |
|
"logps/chosen": -222.09481811523438, |
|
"logps/rejected": -248.44686889648438, |
|
"loss": 0.6602, |
|
"rewards/accuracies": 0.78125, |
|
"rewards/chosen": -0.001766495406627655, |
|
"rewards/margins": 0.569767415523529, |
|
"rewards/rejected": -0.5715339183807373, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.35578362197990165, |
|
"grad_norm": 88.65548046490639, |
|
"learning_rate": 1.919826471272849e-07, |
|
"logits/chosen": -1.3152942657470703, |
|
"logits/rejected": -1.365240216255188, |
|
"logps/chosen": -128.6586151123047, |
|
"logps/rejected": -141.93685913085938, |
|
"loss": 0.6252, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.05175452306866646, |
|
"rewards/margins": 0.1226281225681305, |
|
"rewards/rejected": -0.17438265681266785, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.35806428622336256, |
|
"grad_norm": 84.35369248301531, |
|
"learning_rate": 1.9183276587515165e-07, |
|
"logits/chosen": -1.2944692373275757, |
|
"logits/rejected": -1.3323063850402832, |
|
"logps/chosen": -165.85975646972656, |
|
"logps/rejected": -191.24041748046875, |
|
"loss": 0.6487, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.07458823919296265, |
|
"rewards/margins": 0.3784411549568176, |
|
"rewards/rejected": -0.4530293941497803, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.36034495046682347, |
|
"grad_norm": 86.49341842550122, |
|
"learning_rate": 1.9168155619877707e-07, |
|
"logits/chosen": -1.2946043014526367, |
|
"logits/rejected": -1.4176356792449951, |
|
"logps/chosen": -130.90985107421875, |
|
"logps/rejected": -162.9954071044922, |
|
"loss": 0.6467, |
|
"rewards/accuracies": 0.78125, |
|
"rewards/chosen": 0.011317873373627663, |
|
"rewards/margins": 0.4069710373878479, |
|
"rewards/rejected": -0.39565321803092957, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.36262561471028437, |
|
"grad_norm": 102.96541053849566, |
|
"learning_rate": 1.9152902028551335e-07, |
|
"logits/chosen": -1.3063163757324219, |
|
"logits/rejected": -1.3172858953475952, |
|
"logps/chosen": -157.0022735595703, |
|
"logps/rejected": -165.20347595214844, |
|
"loss": 0.6669, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.07298893481492996, |
|
"rewards/margins": 0.17434003949165344, |
|
"rewards/rejected": -0.2473289668560028, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.3649062789537453, |
|
"grad_norm": 76.39528252287091, |
|
"learning_rate": 1.9137516034189765e-07, |
|
"logits/chosen": -1.3794448375701904, |
|
"logits/rejected": -1.4233444929122925, |
|
"logps/chosen": -164.51431274414062, |
|
"logps/rejected": -191.4024658203125, |
|
"loss": 0.629, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": 0.07598446309566498, |
|
"rewards/margins": 0.355991005897522, |
|
"rewards/rejected": -0.2800065577030182, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.3671869431972062, |
|
"grad_norm": 90.58917607125439, |
|
"learning_rate": 1.9121997859361997e-07, |
|
"logits/chosen": -1.3726584911346436, |
|
"logits/rejected": -1.4218438863754272, |
|
"logps/chosen": -161.8759765625, |
|
"logps/rejected": -191.62745666503906, |
|
"loss": 0.6777, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.03798535838723183, |
|
"rewards/margins": 0.2353454828262329, |
|
"rewards/rejected": -0.27333083748817444, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.3694676074406671, |
|
"grad_norm": 88.70829690205912, |
|
"learning_rate": 1.9106347728549132e-07, |
|
"logits/chosen": -1.4393202066421509, |
|
"logits/rejected": -1.4652281999588013, |
|
"logps/chosen": -159.0114288330078, |
|
"logps/rejected": -169.2324981689453, |
|
"loss": 0.6457, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.03970104455947876, |
|
"rewards/margins": 0.19649550318717957, |
|
"rewards/rejected": -0.23619654774665833, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.371748271684128, |
|
"grad_norm": 97.10274345850128, |
|
"learning_rate": 1.9090565868141093e-07, |
|
"logits/chosen": -1.415880560874939, |
|
"logits/rejected": -1.4168412685394287, |
|
"logps/chosen": -200.85858154296875, |
|
"logps/rejected": -214.74472045898438, |
|
"loss": 0.6936, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.0973915159702301, |
|
"rewards/margins": -0.0053062885999679565, |
|
"rewards/rejected": -0.09208524972200394, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.3740289359275889, |
|
"grad_norm": 79.53686855879208, |
|
"learning_rate": 1.9074652506433367e-07, |
|
"logits/chosen": -1.2665050029754639, |
|
"logits/rejected": -1.2990330457687378, |
|
"logps/chosen": -188.47377014160156, |
|
"logps/rejected": -196.28854370117188, |
|
"loss": 0.6894, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.2655767500400543, |
|
"rewards/margins": 0.0786442756652832, |
|
"rewards/rejected": -0.3442210257053375, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.3763096001710498, |
|
"grad_norm": 92.25716890573669, |
|
"learning_rate": 1.9058607873623696e-07, |
|
"logits/chosen": -1.2212181091308594, |
|
"logits/rejected": -1.2750239372253418, |
|
"logps/chosen": -136.28663635253906, |
|
"logps/rejected": -182.32699584960938, |
|
"loss": 0.6309, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.14970597624778748, |
|
"rewards/margins": 0.13012784719467163, |
|
"rewards/rejected": -0.2798338234424591, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.3785902644145107, |
|
"grad_norm": 96.38565738306772, |
|
"learning_rate": 1.9042432201808753e-07, |
|
"logits/chosen": -1.3811233043670654, |
|
"logits/rejected": -1.3999402523040771, |
|
"logps/chosen": -153.0084686279297, |
|
"logps/rejected": -174.6513671875, |
|
"loss": 0.6334, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.06558946520090103, |
|
"rewards/margins": 0.20774412155151367, |
|
"rewards/rejected": -0.2733335494995117, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.3808709286579716, |
|
"grad_norm": 101.76003883417016, |
|
"learning_rate": 1.9026125724980777e-07, |
|
"logits/chosen": -1.3316706418991089, |
|
"logits/rejected": -1.2979425191879272, |
|
"logps/chosen": -172.97792053222656, |
|
"logps/rejected": -187.61904907226562, |
|
"loss": 0.635, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.022555537521839142, |
|
"rewards/margins": 0.31808096170425415, |
|
"rewards/rejected": -0.2955254316329956, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.3831515929014325, |
|
"grad_norm": 89.57559954819274, |
|
"learning_rate": 1.900968867902419e-07, |
|
"logits/chosen": -1.392217993736267, |
|
"logits/rejected": -1.4426984786987305, |
|
"logps/chosen": -171.40127563476562, |
|
"logps/rejected": -171.21238708496094, |
|
"loss": 0.6236, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.11336915194988251, |
|
"rewards/margins": 0.3153633177280426, |
|
"rewards/rejected": -0.2019941657781601, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.38543225714489343, |
|
"grad_norm": 95.74770834081339, |
|
"learning_rate": 1.899312130171219e-07, |
|
"logits/chosen": -1.410531759262085, |
|
"logits/rejected": -1.4363325834274292, |
|
"logps/chosen": -176.63296508789062, |
|
"logps/rejected": -194.51480102539062, |
|
"loss": 0.6307, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.0749322921037674, |
|
"rewards/margins": 0.2967851161956787, |
|
"rewards/rejected": -0.3717173933982849, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.38771292138835434, |
|
"grad_norm": 94.33644084218925, |
|
"learning_rate": 1.897642383270331e-07, |
|
"logits/chosen": -1.4411017894744873, |
|
"logits/rejected": -1.5141332149505615, |
|
"logps/chosen": -220.8849639892578, |
|
"logps/rejected": -249.29640197753906, |
|
"loss": 0.6396, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.1280645728111267, |
|
"rewards/margins": 0.4015355110168457, |
|
"rewards/rejected": -0.5296000838279724, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.38999358563181524, |
|
"grad_norm": 93.41769187789713, |
|
"learning_rate": 1.8959596513537937e-07, |
|
"logits/chosen": -1.4379459619522095, |
|
"logits/rejected": -1.463719367980957, |
|
"logps/chosen": -136.88717651367188, |
|
"logps/rejected": -146.65672302246094, |
|
"loss": 0.6624, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.024333830922842026, |
|
"rewards/margins": 0.15197159349918365, |
|
"rewards/rejected": -0.12763777375221252, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.39227424987527615, |
|
"grad_norm": 110.48154958430904, |
|
"learning_rate": 1.894263958763485e-07, |
|
"logits/chosen": -1.477798581123352, |
|
"logits/rejected": -1.4197784662246704, |
|
"logps/chosen": -192.24853515625, |
|
"logps/rejected": -182.48162841796875, |
|
"loss": 0.6866, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.029505489394068718, |
|
"rewards/margins": 0.263489693403244, |
|
"rewards/rejected": -0.23398420214653015, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.39455491411873705, |
|
"grad_norm": 120.46352804291743, |
|
"learning_rate": 1.892555330028766e-07, |
|
"logits/chosen": -1.2521153688430786, |
|
"logits/rejected": -1.2826614379882812, |
|
"logps/chosen": -217.5478515625, |
|
"logps/rejected": -273.1899719238281, |
|
"loss": 0.6418, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.2787259817123413, |
|
"rewards/margins": 0.3908158242702484, |
|
"rewards/rejected": -0.6695418357849121, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.396835578362198, |
|
"grad_norm": 94.24036369589392, |
|
"learning_rate": 1.8908337898661285e-07, |
|
"logits/chosen": -1.3746845722198486, |
|
"logits/rejected": -1.46570885181427, |
|
"logps/chosen": -179.17138671875, |
|
"logps/rejected": -217.95236206054688, |
|
"loss": 0.6201, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.12253744155168533, |
|
"rewards/margins": 0.39310359954833984, |
|
"rewards/rejected": -0.5156410336494446, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.3991162426056589, |
|
"grad_norm": 88.9093668290791, |
|
"learning_rate": 1.889099363178838e-07, |
|
"logits/chosen": -1.445816993713379, |
|
"logits/rejected": -1.4381108283996582, |
|
"logps/chosen": -132.04539489746094, |
|
"logps/rejected": -139.59835815429688, |
|
"loss": 0.6502, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.10454928129911423, |
|
"rewards/margins": 0.15687233209609985, |
|
"rewards/rejected": -0.2614216208457947, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.40139690684911983, |
|
"grad_norm": 118.96237042903994, |
|
"learning_rate": 1.8873520750565714e-07, |
|
"logits/chosen": -1.3293050527572632, |
|
"logits/rejected": -1.3804844617843628, |
|
"logps/chosen": -172.6136474609375, |
|
"logps/rejected": -184.08212280273438, |
|
"loss": 0.6857, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.09095679968595505, |
|
"rewards/margins": 0.1739913374185562, |
|
"rewards/rejected": -0.26494812965393066, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.40367757109258073, |
|
"grad_norm": 123.21030025798191, |
|
"learning_rate": 1.8855919507750556e-07, |
|
"logits/chosen": -1.299286127090454, |
|
"logits/rejected": -1.2867302894592285, |
|
"logps/chosen": -228.6288604736328, |
|
"logps/rejected": -253.30572509765625, |
|
"loss": 0.6819, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.27850618958473206, |
|
"rewards/margins": 0.12168295681476593, |
|
"rewards/rejected": -0.40018919110298157, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.40595823533604164, |
|
"grad_norm": 93.75103343413197, |
|
"learning_rate": 1.883819015795701e-07, |
|
"logits/chosen": -1.3721405267715454, |
|
"logits/rejected": -1.4811532497406006, |
|
"logps/chosen": -158.45986938476562, |
|
"logps/rejected": -191.99571228027344, |
|
"loss": 0.6742, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": 0.06369534879922867, |
|
"rewards/margins": 0.13870595395565033, |
|
"rewards/rejected": -0.07501061260700226, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.40823889957950255, |
|
"grad_norm": 102.06521296332116, |
|
"learning_rate": 1.8820332957652343e-07, |
|
"logits/chosen": -1.3636136054992676, |
|
"logits/rejected": -1.4187768697738647, |
|
"logps/chosen": -145.48468017578125, |
|
"logps/rejected": -148.94664001464844, |
|
"loss": 0.6414, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.0663696825504303, |
|
"rewards/margins": 0.29655539989471436, |
|
"rewards/rejected": -0.23018570244312286, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.41051956382296345, |
|
"grad_norm": 80.21120922967526, |
|
"learning_rate": 1.8802348165153257e-07, |
|
"logits/chosen": -1.0907145738601685, |
|
"logits/rejected": -1.2451905012130737, |
|
"logps/chosen": -140.98895263671875, |
|
"logps/rejected": -169.56390380859375, |
|
"loss": 0.6267, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.1324671506881714, |
|
"rewards/margins": 0.2304615080356598, |
|
"rewards/rejected": -0.0979943573474884, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.41280022806642436, |
|
"grad_norm": 90.74971301077287, |
|
"learning_rate": 1.8784236040622173e-07, |
|
"logits/chosen": -1.4046244621276855, |
|
"logits/rejected": -1.5294733047485352, |
|
"logps/chosen": -220.828125, |
|
"logps/rejected": -252.51620483398438, |
|
"loss": 0.6485, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.029995199292898178, |
|
"rewards/margins": 0.10216841101646423, |
|
"rewards/rejected": -0.1321636140346527, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.41508089230988526, |
|
"grad_norm": 116.10619062475209, |
|
"learning_rate": 1.8765996846063453e-07, |
|
"logits/chosen": -1.3053789138793945, |
|
"logits/rejected": -1.331100344657898, |
|
"logps/chosen": -145.552978515625, |
|
"logps/rejected": -175.96580505371094, |
|
"loss": 0.7355, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": -0.08064378052949905, |
|
"rewards/margins": 0.18256865441799164, |
|
"rewards/rejected": -0.2632124125957489, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.41736155655334617, |
|
"grad_norm": 85.92321632812764, |
|
"learning_rate": 1.8747630845319612e-07, |
|
"logits/chosen": -1.3430732488632202, |
|
"logits/rejected": -1.321002721786499, |
|
"logps/chosen": -149.8716278076172, |
|
"logps/rejected": -168.82972717285156, |
|
"loss": 0.6414, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.08440528810024261, |
|
"rewards/margins": 0.3168274164199829, |
|
"rewards/rejected": -0.23242215812206268, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.4196422207968071, |
|
"grad_norm": 89.28728181459165, |
|
"learning_rate": 1.8729138304067512e-07, |
|
"logits/chosen": -1.464853286743164, |
|
"logits/rejected": -1.478898286819458, |
|
"logps/chosen": -157.5703887939453, |
|
"logps/rejected": -171.20423889160156, |
|
"loss": 0.6186, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.0049234069883823395, |
|
"rewards/margins": 0.2588205635547638, |
|
"rewards/rejected": -0.25389716029167175, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.421922885040268, |
|
"grad_norm": 86.83502819065235, |
|
"learning_rate": 1.8710519489814503e-07, |
|
"logits/chosen": -1.2088496685028076, |
|
"logits/rejected": -1.238593578338623, |
|
"logps/chosen": -156.05690002441406, |
|
"logps/rejected": -176.14474487304688, |
|
"loss": 0.7105, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.01236521452665329, |
|
"rewards/margins": 0.17280086874961853, |
|
"rewards/rejected": -0.18516604602336884, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.4242035492837289, |
|
"grad_norm": 101.40852416620743, |
|
"learning_rate": 1.869177467189456e-07, |
|
"logits/chosen": -1.4426904916763306, |
|
"logits/rejected": -1.409334659576416, |
|
"logps/chosen": -182.2246551513672, |
|
"logps/rejected": -194.63980102539062, |
|
"loss": 0.6937, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.13504472374916077, |
|
"rewards/margins": 0.12622104585170746, |
|
"rewards/rejected": -0.2612657845020294, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.4264842135271898, |
|
"grad_norm": 78.92812775165392, |
|
"learning_rate": 1.8672904121464402e-07, |
|
"logits/chosen": -1.3653035163879395, |
|
"logits/rejected": -1.3275575637817383, |
|
"logps/chosen": -180.02243041992188, |
|
"logps/rejected": -195.43106079101562, |
|
"loss": 0.6208, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.13813042640686035, |
|
"rewards/margins": 0.25349733233451843, |
|
"rewards/rejected": -0.3916277587413788, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.4287648777706507, |
|
"grad_norm": 85.95171696022348, |
|
"learning_rate": 1.8653908111499531e-07, |
|
"logits/chosen": -1.4379856586456299, |
|
"logits/rejected": -1.5062627792358398, |
|
"logps/chosen": -123.3829345703125, |
|
"logps/rejected": -155.55897521972656, |
|
"loss": 0.6752, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.02849392406642437, |
|
"rewards/margins": 0.1816612333059311, |
|
"rewards/rejected": -0.2101551592350006, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.4310455420141116, |
|
"grad_norm": 74.55363107506629, |
|
"learning_rate": 1.8634786916790331e-07, |
|
"logits/chosen": -1.306863784790039, |
|
"logits/rejected": -1.4182804822921753, |
|
"logps/chosen": -185.36831665039062, |
|
"logps/rejected": -205.2194366455078, |
|
"loss": 0.602, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.23371680080890656, |
|
"rewards/margins": 0.4712272882461548, |
|
"rewards/rejected": -0.23751050233840942, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.4333262062575725, |
|
"grad_norm": 98.09376209675088, |
|
"learning_rate": 1.861554081393806e-07, |
|
"logits/chosen": -1.298234462738037, |
|
"logits/rejected": -1.320231556892395, |
|
"logps/chosen": -170.1300506591797, |
|
"logps/rejected": -201.2412567138672, |
|
"loss": 0.6166, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.11016674339771271, |
|
"rewards/margins": 0.3684804141521454, |
|
"rewards/rejected": -0.25831368565559387, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.4356068705010334, |
|
"grad_norm": 95.75031333883096, |
|
"learning_rate": 1.8596170081350855e-07, |
|
"logits/chosen": -1.4295341968536377, |
|
"logits/rejected": -1.4480311870574951, |
|
"logps/chosen": -158.36959838867188, |
|
"logps/rejected": -185.86700439453125, |
|
"loss": 0.6506, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": 0.11881276965141296, |
|
"rewards/margins": 0.2408534288406372, |
|
"rewards/rejected": -0.12204068899154663, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.4378875347444943, |
|
"grad_norm": 89.97741935605185, |
|
"learning_rate": 1.8576674999239713e-07, |
|
"logits/chosen": -1.2293694019317627, |
|
"logits/rejected": -1.319061040878296, |
|
"logps/chosen": -168.48960876464844, |
|
"logps/rejected": -213.40200805664062, |
|
"loss": 0.6379, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": 0.07465538382530212, |
|
"rewards/margins": 0.39287707209587097, |
|
"rewards/rejected": -0.31822171807289124, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.44016819898795523, |
|
"grad_norm": 84.57348322808303, |
|
"learning_rate": 1.8557055849614428e-07, |
|
"logits/chosen": -1.2857288122177124, |
|
"logits/rejected": -1.317692518234253, |
|
"logps/chosen": -195.7303009033203, |
|
"logps/rejected": -239.7130126953125, |
|
"loss": 0.5716, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.07659934461116791, |
|
"rewards/margins": 0.5603799819946289, |
|
"rewards/rejected": -0.636979341506958, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.44244886323141613, |
|
"grad_norm": 85.15633806373397, |
|
"learning_rate": 1.8537312916279523e-07, |
|
"logits/chosen": -1.3817092180252075, |
|
"logits/rejected": -1.501309871673584, |
|
"logps/chosen": -146.41033935546875, |
|
"logps/rejected": -171.60218811035156, |
|
"loss": 0.6673, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.07926470041275024, |
|
"rewards/margins": 0.26870349049568176, |
|
"rewards/rejected": -0.347968190908432, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.44472952747487704, |
|
"grad_norm": 89.75323963715459, |
|
"learning_rate": 1.8517446484830136e-07, |
|
"logits/chosen": -1.4567725658416748, |
|
"logits/rejected": -1.5358428955078125, |
|
"logps/chosen": -172.92111206054688, |
|
"logps/rejected": -192.05184936523438, |
|
"loss": 0.6553, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.1497884839773178, |
|
"rewards/margins": 0.06301924586296082, |
|
"rewards/rejected": -0.21280772984027863, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.44701019171833795, |
|
"grad_norm": 105.51876340320463, |
|
"learning_rate": 1.8497456842647878e-07, |
|
"logits/chosen": -1.2944555282592773, |
|
"logits/rejected": -1.3529748916625977, |
|
"logps/chosen": -157.77642822265625, |
|
"logps/rejected": -176.93093872070312, |
|
"loss": 0.682, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.07403340190649033, |
|
"rewards/margins": 0.10720521211624146, |
|
"rewards/rejected": -0.18123860657215118, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.44929085596179885, |
|
"grad_norm": 85.63092920252892, |
|
"learning_rate": 1.8477344278896706e-07, |
|
"logits/chosen": -1.4767736196517944, |
|
"logits/rejected": -1.5019617080688477, |
|
"logps/chosen": -173.24639892578125, |
|
"logps/rejected": -193.271240234375, |
|
"loss": 0.6935, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.13190394639968872, |
|
"rewards/margins": 0.48726946115493774, |
|
"rewards/rejected": -0.355365514755249, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.45157152020525976, |
|
"grad_norm": 100.4356452029139, |
|
"learning_rate": 1.8457109084518718e-07, |
|
"logits/chosen": -1.4190278053283691, |
|
"logits/rejected": -1.3654460906982422, |
|
"logps/chosen": -227.8990020751953, |
|
"logps/rejected": -215.39312744140625, |
|
"loss": 0.6697, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.15163551270961761, |
|
"rewards/margins": 0.19186191260814667, |
|
"rewards/rejected": -0.3434974253177643, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.45385218444872066, |
|
"grad_norm": 89.70522326638444, |
|
"learning_rate": 1.8436751552229938e-07, |
|
"logits/chosen": -1.3089743852615356, |
|
"logits/rejected": -1.340443730354309, |
|
"logps/chosen": -145.11398315429688, |
|
"logps/rejected": -172.60635375976562, |
|
"loss": 0.6571, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.20591212809085846, |
|
"rewards/margins": 0.34871169924736023, |
|
"rewards/rejected": -0.5546237230300903, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.4561328486921816, |
|
"grad_norm": 84.89862446594006, |
|
"learning_rate": 1.84162719765161e-07, |
|
"logits/chosen": -1.2627918720245361, |
|
"logits/rejected": -1.2893693447113037, |
|
"logps/chosen": -130.3240966796875, |
|
"logps/rejected": -160.49227905273438, |
|
"loss": 0.6402, |
|
"rewards/accuracies": 0.78125, |
|
"rewards/chosen": 0.0029626190662384033, |
|
"rewards/margins": 0.4357815384864807, |
|
"rewards/rejected": -0.4328188896179199, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.4561328486921816, |
|
"eval_logits/chosen": -1.4406086206436157, |
|
"eval_logits/rejected": -1.4253712892532349, |
|
"eval_logps/chosen": -129.20408630371094, |
|
"eval_logps/rejected": -132.1772003173828, |
|
"eval_loss": 0.659533679485321, |
|
"eval_rewards/accuracies": 0.6399999856948853, |
|
"eval_rewards/chosen": 0.06938672810792923, |
|
"eval_rewards/margins": 0.08083352446556091, |
|
"eval_rewards/rejected": -0.011446798220276833, |
|
"eval_runtime": 21.3693, |
|
"eval_samples_per_second": 4.68, |
|
"eval_steps_per_second": 1.17, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.45841351293564253, |
|
"grad_norm": 86.51322437777803, |
|
"learning_rate": 1.839567065362838e-07, |
|
"logits/chosen": -1.368638038635254, |
|
"logits/rejected": -1.3851830959320068, |
|
"logps/chosen": -169.3848876953125, |
|
"logps/rejected": -175.4376678466797, |
|
"loss": 0.635, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.0501415841281414, |
|
"rewards/margins": 0.1427716761827469, |
|
"rewards/rejected": -0.1929132342338562, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.46069417717910344, |
|
"grad_norm": 101.82366291792587, |
|
"learning_rate": 1.8374947881579112e-07, |
|
"logits/chosen": -1.201027512550354, |
|
"logits/rejected": -1.2725883722305298, |
|
"logps/chosen": -191.0504150390625, |
|
"logps/rejected": -233.1793212890625, |
|
"loss": 0.6195, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.2396811544895172, |
|
"rewards/margins": 0.2062177062034607, |
|
"rewards/rejected": -0.4458988606929779, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.46297484142256434, |
|
"grad_norm": 108.6063244513385, |
|
"learning_rate": 1.8354103960137473e-07, |
|
"logits/chosen": -1.4036638736724854, |
|
"logits/rejected": -1.4488118886947632, |
|
"logps/chosen": -230.92767333984375, |
|
"logps/rejected": -239.439208984375, |
|
"loss": 0.6795, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.32087260484695435, |
|
"rewards/margins": 0.10698382556438446, |
|
"rewards/rejected": -0.4278564453125, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.46525550566602525, |
|
"grad_norm": 86.59174830548812, |
|
"learning_rate": 1.833313919082515e-07, |
|
"logits/chosen": -1.1800165176391602, |
|
"logits/rejected": -1.3113179206848145, |
|
"logps/chosen": -161.62985229492188, |
|
"logps/rejected": -202.1416473388672, |
|
"loss": 0.6378, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.14624057710170746, |
|
"rewards/margins": 0.3915923237800598, |
|
"rewards/rejected": -0.5378329157829285, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.46753616990948615, |
|
"grad_norm": 92.36447587751275, |
|
"learning_rate": 1.8312053876911977e-07, |
|
"logits/chosen": -1.360296368598938, |
|
"logits/rejected": -1.3766343593597412, |
|
"logps/chosen": -203.7003173828125, |
|
"logps/rejected": -233.6885223388672, |
|
"loss": 0.6525, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.13416332006454468, |
|
"rewards/margins": 0.36563166975975037, |
|
"rewards/rejected": -0.49979501962661743, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.46981683415294706, |
|
"grad_norm": 79.86005120772512, |
|
"learning_rate": 1.8290848323411553e-07, |
|
"logits/chosen": -1.4597082138061523, |
|
"logits/rejected": -1.4648618698120117, |
|
"logps/chosen": -151.07449340820312, |
|
"logps/rejected": -151.00494384765625, |
|
"loss": 0.7101, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.05570954829454422, |
|
"rewards/margins": 0.08499263972043991, |
|
"rewards/rejected": -0.02928308956325054, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.47209749839640797, |
|
"grad_norm": 70.72545574881394, |
|
"learning_rate": 1.8269522837076817e-07, |
|
"logits/chosen": -1.4345017671585083, |
|
"logits/rejected": -1.498835802078247, |
|
"logps/chosen": -180.71951293945312, |
|
"logps/rejected": -195.18490600585938, |
|
"loss": 0.6126, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.060558006167411804, |
|
"rewards/margins": 0.23918625712394714, |
|
"rewards/rejected": -0.29974424839019775, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.47437816263986887, |
|
"grad_norm": 91.45944584413142, |
|
"learning_rate": 1.8248077726395631e-07, |
|
"logits/chosen": -1.381393313407898, |
|
"logits/rejected": -1.3962711095809937, |
|
"logps/chosen": -139.09548950195312, |
|
"logps/rejected": -154.32943725585938, |
|
"loss": 0.644, |
|
"rewards/accuracies": 0.78125, |
|
"rewards/chosen": 0.08523928374052048, |
|
"rewards/margins": 0.4487174153327942, |
|
"rewards/rejected": -0.3634781539440155, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.4766588268833298, |
|
"grad_norm": 118.15308317787861, |
|
"learning_rate": 1.8226513301586297e-07, |
|
"logits/chosen": -1.1643140316009521, |
|
"logits/rejected": -1.1523492336273193, |
|
"logps/chosen": -181.3363494873047, |
|
"logps/rejected": -208.14474487304688, |
|
"loss": 0.6631, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.18151137232780457, |
|
"rewards/margins": 0.11719075590372086, |
|
"rewards/rejected": -0.2987021207809448, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.4789394911267907, |
|
"grad_norm": 87.18607024470172, |
|
"learning_rate": 1.820482987459308e-07, |
|
"logits/chosen": -1.2820664644241333, |
|
"logits/rejected": -1.340864658355713, |
|
"logps/chosen": -152.3217315673828, |
|
"logps/rejected": -166.40199279785156, |
|
"loss": 0.6306, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.20804628729820251, |
|
"rewards/margins": 0.08607995510101318, |
|
"rewards/rejected": -0.2941262423992157, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.4812201553702516, |
|
"grad_norm": 101.2197264458815, |
|
"learning_rate": 1.818302775908169e-07, |
|
"logits/chosen": -1.3642746210098267, |
|
"logits/rejected": -1.4530210494995117, |
|
"logps/chosen": -177.27783203125, |
|
"logps/rejected": -209.4615478515625, |
|
"loss": 0.6002, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.06609566509723663, |
|
"rewards/margins": 0.24875135719776154, |
|
"rewards/rejected": -0.31484702229499817, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.4835008196137125, |
|
"grad_norm": 102.24009472626234, |
|
"learning_rate": 1.8161107270434757e-07, |
|
"logits/chosen": -1.322772741317749, |
|
"logits/rejected": -1.3549699783325195, |
|
"logps/chosen": -175.10890197753906, |
|
"logps/rejected": -189.1114959716797, |
|
"loss": 0.7052, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.2034468799829483, |
|
"rewards/margins": -0.003699667751789093, |
|
"rewards/rejected": -0.199747234582901, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.4857814838571734, |
|
"grad_norm": 88.40359207083654, |
|
"learning_rate": 1.8139068725747251e-07, |
|
"logits/chosen": -1.2980220317840576, |
|
"logits/rejected": -1.328129768371582, |
|
"logps/chosen": -127.48295593261719, |
|
"logps/rejected": -145.63137817382812, |
|
"loss": 0.6353, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.17486554384231567, |
|
"rewards/margins": 0.16773542761802673, |
|
"rewards/rejected": -0.3426010012626648, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.4880621481006343, |
|
"grad_norm": 85.01546232524993, |
|
"learning_rate": 1.811691244382191e-07, |
|
"logits/chosen": -1.240834355354309, |
|
"logits/rejected": -1.315454125404358, |
|
"logps/chosen": -171.84417724609375, |
|
"logps/rejected": -196.60919189453125, |
|
"loss": 0.6332, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.043825387954711914, |
|
"rewards/margins": 0.35121482610702515, |
|
"rewards/rejected": -0.39504021406173706, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.4903428123440952, |
|
"grad_norm": 92.50180646153169, |
|
"learning_rate": 1.8094638745164619e-07, |
|
"logits/chosen": -1.368233561515808, |
|
"logits/rejected": -1.4737908840179443, |
|
"logps/chosen": -231.7982177734375, |
|
"logps/rejected": -257.540771484375, |
|
"loss": 0.6393, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.41820961236953735, |
|
"rewards/margins": 0.03227778524160385, |
|
"rewards/rejected": -0.4504873752593994, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.4926234765875561, |
|
"grad_norm": 93.17182791936958, |
|
"learning_rate": 1.8072247951979782e-07, |
|
"logits/chosen": -1.218642234802246, |
|
"logits/rejected": -1.3522933721542358, |
|
"logps/chosen": -161.9176025390625, |
|
"logps/rejected": -191.59205627441406, |
|
"loss": 0.6532, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": -0.2410704791545868, |
|
"rewards/margins": 0.09214940667152405, |
|
"rewards/rejected": -0.33321988582611084, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.494904140831017, |
|
"grad_norm": 85.22671128905367, |
|
"learning_rate": 1.8049740388165646e-07, |
|
"logits/chosen": -1.3610548973083496, |
|
"logits/rejected": -1.369373083114624, |
|
"logps/chosen": -170.4285888671875, |
|
"logps/rejected": -203.44435119628906, |
|
"loss": 0.6788, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.12429852783679962, |
|
"rewards/margins": 0.06445370614528656, |
|
"rewards/rejected": -0.18875223398208618, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.49718480507447793, |
|
"grad_norm": 106.2899671913355, |
|
"learning_rate": 1.8027116379309635e-07, |
|
"logits/chosen": -1.3293408155441284, |
|
"logits/rejected": -1.3227120637893677, |
|
"logps/chosen": -189.62835693359375, |
|
"logps/rejected": -196.80776977539062, |
|
"loss": 0.6658, |
|
"rewards/accuracies": 0.4375, |
|
"rewards/chosen": -0.36924922466278076, |
|
"rewards/margins": 0.04584325850009918, |
|
"rewards/rejected": -0.41509246826171875, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.49946546931793884, |
|
"grad_norm": 97.4627918918024, |
|
"learning_rate": 1.8004376252683629e-07, |
|
"logits/chosen": -1.3211669921875, |
|
"logits/rejected": -1.318382740020752, |
|
"logps/chosen": -170.44007873535156, |
|
"logps/rejected": -177.15567016601562, |
|
"loss": 0.652, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.0650164932012558, |
|
"rewards/margins": 0.2474796026945114, |
|
"rewards/rejected": -0.3124960958957672, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.5017461335613997, |
|
"grad_norm": 94.35984756124336, |
|
"learning_rate": 1.7981520337239229e-07, |
|
"logits/chosen": -1.4195128679275513, |
|
"logits/rejected": -1.517584204673767, |
|
"logps/chosen": -124.39069366455078, |
|
"logps/rejected": -148.39462280273438, |
|
"loss": 0.6071, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.07836849987506866, |
|
"rewards/margins": 0.2909252941608429, |
|
"rewards/rejected": -0.21255679428577423, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.5040267978048607, |
|
"grad_norm": 105.60849785169918, |
|
"learning_rate": 1.7958548963603e-07, |
|
"logits/chosen": -1.3688621520996094, |
|
"logits/rejected": -1.3791152238845825, |
|
"logps/chosen": -193.5460968017578, |
|
"logps/rejected": -198.70162963867188, |
|
"loss": 0.6155, |
|
"rewards/accuracies": 0.8125, |
|
"rewards/chosen": 0.006075154058635235, |
|
"rewards/margins": 0.4186677634716034, |
|
"rewards/rejected": -0.41259267926216125, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.5063074620483216, |
|
"grad_norm": 105.19505214300226, |
|
"learning_rate": 1.7935462464071694e-07, |
|
"logits/chosen": -1.4063529968261719, |
|
"logits/rejected": -1.4609131813049316, |
|
"logps/chosen": -136.6985321044922, |
|
"logps/rejected": -142.0517120361328, |
|
"loss": 0.6638, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -0.021208282560110092, |
|
"rewards/margins": 0.23011060059070587, |
|
"rewards/rejected": -0.25131887197494507, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.5085881262917825, |
|
"grad_norm": 80.98126219811707, |
|
"learning_rate": 1.7912261172607434e-07, |
|
"logits/chosen": -1.4103684425354004, |
|
"logits/rejected": -1.4659305810928345, |
|
"logps/chosen": -181.66671752929688, |
|
"logps/rejected": -211.0421142578125, |
|
"loss": 0.667, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.0669156014919281, |
|
"rewards/margins": 0.23970045149326324, |
|
"rewards/rejected": -0.30661600828170776, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.5108687905352434, |
|
"grad_norm": 86.7187808268374, |
|
"learning_rate": 1.7888945424832892e-07, |
|
"logits/chosen": -1.2415688037872314, |
|
"logits/rejected": -1.2903611660003662, |
|
"logps/chosen": -195.07174682617188, |
|
"logps/rejected": -223.18804931640625, |
|
"loss": 0.6344, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.26416096091270447, |
|
"rewards/margins": 0.23690298199653625, |
|
"rewards/rejected": -0.5010639429092407, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.5131494547787043, |
|
"grad_norm": 91.51916447980646, |
|
"learning_rate": 1.7865515558026425e-07, |
|
"logits/chosen": -1.2545722723007202, |
|
"logits/rejected": -1.320673942565918, |
|
"logps/chosen": -190.4950714111328, |
|
"logps/rejected": -196.8434600830078, |
|
"loss": 0.645, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.12865591049194336, |
|
"rewards/margins": 0.2827560007572174, |
|
"rewards/rejected": -0.41141194105148315, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.5154301190221652, |
|
"grad_norm": 78.95115222899102, |
|
"learning_rate": 1.78419719111172e-07, |
|
"logits/chosen": -1.2161290645599365, |
|
"logits/rejected": -1.2955366373062134, |
|
"logps/chosen": -133.3023223876953, |
|
"logps/rejected": -173.38735961914062, |
|
"loss": 0.6202, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.031168239191174507, |
|
"rewards/margins": 0.3425666093826294, |
|
"rewards/rejected": -0.37373483180999756, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 0.5177107832656261, |
|
"grad_norm": 85.03550287122602, |
|
"learning_rate": 1.78183148246803e-07, |
|
"logits/chosen": -1.3149534463882446, |
|
"logits/rejected": -1.3734480142593384, |
|
"logps/chosen": -181.44749450683594, |
|
"logps/rejected": -190.66098022460938, |
|
"loss": 0.6579, |
|
"rewards/accuracies": 0.4375, |
|
"rewards/chosen": -0.02150813303887844, |
|
"rewards/margins": 0.09824002534151077, |
|
"rewards/rejected": -0.11974816769361496, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 0.519991447509087, |
|
"grad_norm": 101.82433211173363, |
|
"learning_rate": 1.779454464093177e-07, |
|
"logits/chosen": -1.2327746152877808, |
|
"logits/rejected": -1.2479290962219238, |
|
"logps/chosen": -123.62596130371094, |
|
"logps/rejected": -145.822021484375, |
|
"loss": 0.71, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.008450210094451904, |
|
"rewards/margins": 0.1642961949110031, |
|
"rewards/rejected": -0.1558459848165512, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 0.522272111752548, |
|
"grad_norm": 86.25249319253086, |
|
"learning_rate": 1.7770661703723716e-07, |
|
"logits/chosen": -1.4179208278656006, |
|
"logits/rejected": -1.4704089164733887, |
|
"logps/chosen": -155.395263671875, |
|
"logps/rejected": -241.91619873046875, |
|
"loss": 0.6544, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.0192805714905262, |
|
"rewards/margins": 0.4996611773967743, |
|
"rewards/rejected": -0.5189418792724609, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 0.5245527759960088, |
|
"grad_norm": 95.73185772448069, |
|
"learning_rate": 1.7746666358539268e-07, |
|
"logits/chosen": -1.324285864830017, |
|
"logits/rejected": -1.3773202896118164, |
|
"logps/chosen": -138.01214599609375, |
|
"logps/rejected": -187.76605224609375, |
|
"loss": 0.6781, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.11806146800518036, |
|
"rewards/margins": 0.12408170104026794, |
|
"rewards/rejected": -0.2421431541442871, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.5268334402394698, |
|
"grad_norm": 83.04802100256194, |
|
"learning_rate": 1.7722558952487637e-07, |
|
"logits/chosen": -1.2252494096755981, |
|
"logits/rejected": -1.2402747869491577, |
|
"logps/chosen": -145.02134704589844, |
|
"logps/rejected": -157.798095703125, |
|
"loss": 0.6658, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.16228270530700684, |
|
"rewards/margins": -0.006807304918766022, |
|
"rewards/rejected": -0.1554754078388214, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 0.5291141044829306, |
|
"grad_norm": 106.4784679324614, |
|
"learning_rate": 1.7698339834299062e-07, |
|
"logits/chosen": -1.2572526931762695, |
|
"logits/rejected": -1.2763482332229614, |
|
"logps/chosen": -205.0850067138672, |
|
"logps/rejected": -216.61383056640625, |
|
"loss": 0.6758, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -0.10254046320915222, |
|
"rewards/margins": 0.47703590989112854, |
|
"rewards/rejected": -0.5795763731002808, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 0.5313947687263916, |
|
"grad_norm": 88.23039919938233, |
|
"learning_rate": 1.7674009354319776e-07, |
|
"logits/chosen": -1.2906126976013184, |
|
"logits/rejected": -1.3812601566314697, |
|
"logps/chosen": -182.90606689453125, |
|
"logps/rejected": -192.94520568847656, |
|
"loss": 0.5967, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.12573322653770447, |
|
"rewards/margins": 0.15527519583702087, |
|
"rewards/rejected": -0.28100842237472534, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 0.5336754329698524, |
|
"grad_norm": 96.24054074341065, |
|
"learning_rate": 1.764956786450694e-07, |
|
"logits/chosen": -1.2798047065734863, |
|
"logits/rejected": -1.3678182363510132, |
|
"logps/chosen": -160.85719299316406, |
|
"logps/rejected": -189.01951599121094, |
|
"loss": 0.6321, |
|
"rewards/accuracies": 0.78125, |
|
"rewards/chosen": -0.01916433498263359, |
|
"rewards/margins": 0.328861802816391, |
|
"rewards/rejected": -0.3480261266231537, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 0.5359560972133134, |
|
"grad_norm": 87.51604676838089, |
|
"learning_rate": 1.7625015718423548e-07, |
|
"logits/chosen": -1.3382031917572021, |
|
"logits/rejected": -1.4446802139282227, |
|
"logps/chosen": -152.4910125732422, |
|
"logps/rejected": -192.94979858398438, |
|
"loss": 0.6364, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.021129202097654343, |
|
"rewards/margins": 0.3612152636051178, |
|
"rewards/rejected": -0.38234445452690125, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.5382367614567742, |
|
"grad_norm": 118.40842425272356, |
|
"learning_rate": 1.7600353271233312e-07, |
|
"logits/chosen": -1.319710373878479, |
|
"logits/rejected": -1.284549355506897, |
|
"logps/chosen": -152.95933532714844, |
|
"logps/rejected": -156.8072052001953, |
|
"loss": 0.7236, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.10047149658203125, |
|
"rewards/margins": 0.010637946426868439, |
|
"rewards/rejected": -0.11110943555831909, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 0.5405174257002352, |
|
"grad_norm": 91.83509455780727, |
|
"learning_rate": 1.7575580879695522e-07, |
|
"logits/chosen": -1.445102572441101, |
|
"logits/rejected": -1.4954876899719238, |
|
"logps/chosen": -196.539306640625, |
|
"logps/rejected": -211.00616455078125, |
|
"loss": 0.6409, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.04536067694425583, |
|
"rewards/margins": 0.1365879625082016, |
|
"rewards/rejected": -0.18194863200187683, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 0.542798089943696, |
|
"grad_norm": 90.74830800218245, |
|
"learning_rate": 1.7550698902159894e-07, |
|
"logits/chosen": -1.3519353866577148, |
|
"logits/rejected": -1.4368261098861694, |
|
"logps/chosen": -228.4932098388672, |
|
"logps/rejected": -261.87109375, |
|
"loss": 0.6373, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": -0.1739276647567749, |
|
"rewards/margins": 0.011357128620147705, |
|
"rewards/rejected": -0.1852847784757614, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 0.545078754187157, |
|
"grad_norm": 114.14083612276099, |
|
"learning_rate": 1.7525707698561384e-07, |
|
"logits/chosen": -1.296964168548584, |
|
"logits/rejected": -1.3777580261230469, |
|
"logps/chosen": -178.81484985351562, |
|
"logps/rejected": -220.63436889648438, |
|
"loss": 0.6515, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.10972025990486145, |
|
"rewards/margins": 0.45839735865592957, |
|
"rewards/rejected": -0.5681176781654358, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 0.547359418430618, |
|
"grad_norm": 106.70099483754956, |
|
"learning_rate": 1.750060763041497e-07, |
|
"logits/chosen": -1.3242031335830688, |
|
"logits/rejected": -1.3594613075256348, |
|
"logps/chosen": -139.518310546875, |
|
"logps/rejected": -171.3057403564453, |
|
"loss": 0.6932, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.09948215633630753, |
|
"rewards/margins": 0.13577090203762054, |
|
"rewards/rejected": -0.23525305092334747, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.5496400826740788, |
|
"grad_norm": 93.76871490458784, |
|
"learning_rate": 1.7475399060810435e-07, |
|
"logits/chosen": -1.5048158168792725, |
|
"logits/rejected": -1.567376971244812, |
|
"logps/chosen": -175.3671112060547, |
|
"logps/rejected": -195.3438262939453, |
|
"loss": 0.6284, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.08378006517887115, |
|
"rewards/margins": 0.32175135612487793, |
|
"rewards/rejected": -0.23797132074832916, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 0.5519207469175398, |
|
"grad_norm": 87.30478432077047, |
|
"learning_rate": 1.7450082354407107e-07, |
|
"logits/chosen": -1.3454846143722534, |
|
"logits/rejected": -1.3264262676239014, |
|
"logps/chosen": -150.41058349609375, |
|
"logps/rejected": -158.28414916992188, |
|
"loss": 0.6392, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": 0.06661003828048706, |
|
"rewards/margins": 0.05708365514874458, |
|
"rewards/rejected": 0.009526383131742477, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 0.5542014111610006, |
|
"grad_norm": 97.70119982177076, |
|
"learning_rate": 1.7424657877428594e-07, |
|
"logits/chosen": -1.4438061714172363, |
|
"logits/rejected": -1.5046687126159668, |
|
"logps/chosen": -210.83221435546875, |
|
"logps/rejected": -220.98056030273438, |
|
"loss": 0.6711, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.02250128984451294, |
|
"rewards/margins": 0.25783565640449524, |
|
"rewards/rejected": -0.23533441126346588, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 0.5564820754044616, |
|
"grad_norm": 92.7470076491593, |
|
"learning_rate": 1.7399125997657475e-07, |
|
"logits/chosen": -1.4578030109405518, |
|
"logits/rejected": -1.5047590732574463, |
|
"logps/chosen": -173.8478546142578, |
|
"logps/rejected": -218.70567321777344, |
|
"loss": 0.6464, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.005423933267593384, |
|
"rewards/margins": 0.22333881258964539, |
|
"rewards/rejected": -0.22876277565956116, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 0.5587627396479224, |
|
"grad_norm": 80.43868813641923, |
|
"learning_rate": 1.7373487084429986e-07, |
|
"logits/chosen": -1.4293080568313599, |
|
"logits/rejected": -1.3778735399246216, |
|
"logps/chosen": -137.72915649414062, |
|
"logps/rejected": -144.8123779296875, |
|
"loss": 0.6175, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.04381662234663963, |
|
"rewards/margins": 0.17386358976364136, |
|
"rewards/rejected": -0.2176802158355713, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.5610434038913834, |
|
"grad_norm": 88.32928179136898, |
|
"learning_rate": 1.734774150863067e-07, |
|
"logits/chosen": -1.4234795570373535, |
|
"logits/rejected": -1.4454480409622192, |
|
"logps/chosen": -168.4951171875, |
|
"logps/rejected": -183.8735809326172, |
|
"loss": 0.6522, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.008489616215229034, |
|
"rewards/margins": 0.20173315703868866, |
|
"rewards/rejected": -0.19324351847171783, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 0.5633240681348443, |
|
"grad_norm": 93.76675043932717, |
|
"learning_rate": 1.732188964268703e-07, |
|
"logits/chosen": -1.2950657606124878, |
|
"logits/rejected": -1.3514856100082397, |
|
"logps/chosen": -174.49368286132812, |
|
"logps/rejected": -198.4128875732422, |
|
"loss": 0.6429, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.17520295083522797, |
|
"rewards/margins": 0.17245244979858398, |
|
"rewards/rejected": -0.34765538573265076, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 0.5656047323783052, |
|
"grad_norm": 124.66844580791373, |
|
"learning_rate": 1.7295931860564118e-07, |
|
"logits/chosen": -1.3973047733306885, |
|
"logits/rejected": -1.3848158121109009, |
|
"logps/chosen": -187.01031494140625, |
|
"logps/rejected": -201.88232421875, |
|
"loss": 0.661, |
|
"rewards/accuracies": 0.4375, |
|
"rewards/chosen": -0.4809916317462921, |
|
"rewards/margins": -0.037112899124622345, |
|
"rewards/rejected": -0.443878710269928, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 0.5678853966217661, |
|
"grad_norm": 98.27416452810739, |
|
"learning_rate": 1.7269868537759138e-07, |
|
"logits/chosen": -1.4414607286453247, |
|
"logits/rejected": -1.4525971412658691, |
|
"logps/chosen": -186.87220764160156, |
|
"logps/rejected": -201.6339874267578, |
|
"loss": 0.6551, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.048535339534282684, |
|
"rewards/margins": 0.2541671097278595, |
|
"rewards/rejected": -0.20563176274299622, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 0.570166060865227, |
|
"grad_norm": 81.20755107084783, |
|
"learning_rate": 1.7243700051296016e-07, |
|
"logits/chosen": -1.4566954374313354, |
|
"logits/rejected": -1.458418369293213, |
|
"logps/chosen": -186.07911682128906, |
|
"logps/rejected": -181.54051208496094, |
|
"loss": 0.6716, |
|
"rewards/accuracies": 0.4375, |
|
"rewards/chosen": -0.08042742311954498, |
|
"rewards/margins": 0.003591485321521759, |
|
"rewards/rejected": -0.08401892334222794, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.570166060865227, |
|
"eval_logits/chosen": -1.4689228534698486, |
|
"eval_logits/rejected": -1.4550352096557617, |
|
"eval_logps/chosen": -128.87640380859375, |
|
"eval_logps/rejected": -132.28384399414062, |
|
"eval_loss": 0.6351009011268616, |
|
"eval_rewards/accuracies": 0.6399999856948853, |
|
"eval_rewards/chosen": 0.10215496271848679, |
|
"eval_rewards/margins": 0.12426460534334183, |
|
"eval_rewards/rejected": -0.022109635174274445, |
|
"eval_runtime": 20.6735, |
|
"eval_samples_per_second": 4.837, |
|
"eval_steps_per_second": 1.209, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.5724467251086879, |
|
"grad_norm": 87.15380241514247, |
|
"learning_rate": 1.7217426779719944e-07, |
|
"logits/chosen": -1.461560606956482, |
|
"logits/rejected": -1.4978594779968262, |
|
"logps/chosen": -145.8336639404297, |
|
"logps/rejected": -164.410888671875, |
|
"loss": 0.623, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.030693676322698593, |
|
"rewards/margins": 0.31371259689331055, |
|
"rewards/rejected": -0.34440624713897705, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 0.5747273893521488, |
|
"grad_norm": 109.6734083472702, |
|
"learning_rate": 1.71910491030919e-07, |
|
"logits/chosen": -1.4200553894042969, |
|
"logits/rejected": -1.4383351802825928, |
|
"logps/chosen": -185.2432861328125, |
|
"logps/rejected": -198.3214569091797, |
|
"loss": 0.6592, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.09178508818149567, |
|
"rewards/margins": 0.1750696301460266, |
|
"rewards/rejected": -0.2668547034263611, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 0.5770080535956097, |
|
"grad_norm": 95.33984559143799, |
|
"learning_rate": 1.716456740298315e-07, |
|
"logits/chosen": -1.356040120124817, |
|
"logits/rejected": -1.3635413646697998, |
|
"logps/chosen": -221.06320190429688, |
|
"logps/rejected": -246.34091186523438, |
|
"loss": 0.6466, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.224991112947464, |
|
"rewards/margins": 0.12949411571025848, |
|
"rewards/rejected": -0.35448524355888367, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 0.5792887178390707, |
|
"grad_norm": 104.51344570968739, |
|
"learning_rate": 1.7137982062469739e-07, |
|
"logits/chosen": -1.199120044708252, |
|
"logits/rejected": -1.2254160642623901, |
|
"logps/chosen": -182.35678100585938, |
|
"logps/rejected": -205.9598388671875, |
|
"loss": 0.6302, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.01045935694128275, |
|
"rewards/margins": 0.4567835330963135, |
|
"rewards/rejected": -0.46724286675453186, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 0.5815693820825315, |
|
"grad_norm": 80.61166146639357, |
|
"learning_rate": 1.7111293466126936e-07, |
|
"logits/chosen": -1.4290285110473633, |
|
"logits/rejected": -1.465309977531433, |
|
"logps/chosen": -183.41546630859375, |
|
"logps/rejected": -190.38209533691406, |
|
"loss": 0.6303, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.04882081598043442, |
|
"rewards/margins": 0.2510834336280823, |
|
"rewards/rejected": -0.2999042272567749, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.5838500463259925, |
|
"grad_norm": 81.58673928063743, |
|
"learning_rate": 1.7084502000023678e-07, |
|
"logits/chosen": -1.3187546730041504, |
|
"logits/rejected": -1.3721165657043457, |
|
"logps/chosen": -166.35317993164062, |
|
"logps/rejected": -205.14459228515625, |
|
"loss": 0.6609, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.06919605284929276, |
|
"rewards/margins": 0.25993865728378296, |
|
"rewards/rejected": -0.1907426118850708, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 0.5861307105694533, |
|
"grad_norm": 107.99959199152349, |
|
"learning_rate": 1.7057608051716987e-07, |
|
"logits/chosen": -1.2958874702453613, |
|
"logits/rejected": -1.3409713506698608, |
|
"logps/chosen": -189.3512420654297, |
|
"logps/rejected": -229.98101806640625, |
|
"loss": 0.6805, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.1412532776594162, |
|
"rewards/margins": 0.1916051059961319, |
|
"rewards/rejected": -0.3328583836555481, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 0.5884113748129143, |
|
"grad_norm": 103.31085588981476, |
|
"learning_rate": 1.7030612010246357e-07, |
|
"logits/chosen": -1.4542144536972046, |
|
"logits/rejected": -1.4537440538406372, |
|
"logps/chosen": -187.4440155029297, |
|
"logps/rejected": -197.0413818359375, |
|
"loss": 0.6901, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.20216982066631317, |
|
"rewards/margins": -0.05732645466923714, |
|
"rewards/rejected": -0.14484335482120514, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 0.5906920390563751, |
|
"grad_norm": 93.7004047268541, |
|
"learning_rate": 1.7003514266128128e-07, |
|
"logits/chosen": -1.3705557584762573, |
|
"logits/rejected": -1.4125196933746338, |
|
"logps/chosen": -176.84820556640625, |
|
"logps/rejected": -218.4021453857422, |
|
"loss": 0.688, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.13926447927951813, |
|
"rewards/margins": 0.1602565199136734, |
|
"rewards/rejected": -0.2995210289955139, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 0.5929727032998361, |
|
"grad_norm": 76.7210943847488, |
|
"learning_rate": 1.6976315211349848e-07, |
|
"logits/chosen": -1.402063250541687, |
|
"logits/rejected": -1.5004804134368896, |
|
"logps/chosen": -183.45787048339844, |
|
"logps/rejected": -226.31532287597656, |
|
"loss": 0.6405, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": 0.08890336751937866, |
|
"rewards/margins": 0.2809264659881592, |
|
"rewards/rejected": -0.19202306866645813, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.5952533675432969, |
|
"grad_norm": 79.99952128240997, |
|
"learning_rate": 1.694901523936458e-07, |
|
"logits/chosen": -1.3303985595703125, |
|
"logits/rejected": -1.4111953973770142, |
|
"logps/chosen": -161.44390869140625, |
|
"logps/rejected": -201.76593017578125, |
|
"loss": 0.601, |
|
"rewards/accuracies": 0.78125, |
|
"rewards/chosen": 0.12671524286270142, |
|
"rewards/margins": 0.4966307282447815, |
|
"rewards/rejected": -0.3699154853820801, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 0.5975340317867579, |
|
"grad_norm": 89.60709297255305, |
|
"learning_rate": 1.6921614745085235e-07, |
|
"logits/chosen": -1.4100148677825928, |
|
"logits/rejected": -1.4378656148910522, |
|
"logps/chosen": -193.54835510253906, |
|
"logps/rejected": -202.9269256591797, |
|
"loss": 0.6182, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -0.035577308386564255, |
|
"rewards/margins": 0.3006408214569092, |
|
"rewards/rejected": -0.33621811866760254, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 0.5998146960302188, |
|
"grad_norm": 104.27911734761277, |
|
"learning_rate": 1.689411412487885e-07, |
|
"logits/chosen": -1.3611271381378174, |
|
"logits/rejected": -1.3725131750106812, |
|
"logps/chosen": -166.821533203125, |
|
"logps/rejected": -197.65394592285156, |
|
"loss": 0.6274, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.09458069503307343, |
|
"rewards/margins": 0.400374174118042, |
|
"rewards/rejected": -0.30579352378845215, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 0.6020953602736797, |
|
"grad_norm": 77.9183795766502, |
|
"learning_rate": 1.6866513776560841e-07, |
|
"logits/chosen": -1.2477551698684692, |
|
"logits/rejected": -1.2647314071655273, |
|
"logps/chosen": -143.03448486328125, |
|
"logps/rejected": -149.70594787597656, |
|
"loss": 0.6035, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": 0.14375856518745422, |
|
"rewards/margins": 0.17092323303222656, |
|
"rewards/rejected": -0.02716466784477234, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 0.6043760245171406, |
|
"grad_norm": 79.08140854679833, |
|
"learning_rate": 1.6838814099389265e-07, |
|
"logits/chosen": -1.379310965538025, |
|
"logits/rejected": -1.49821138381958, |
|
"logps/chosen": -176.78890991210938, |
|
"logps/rejected": -228.0948028564453, |
|
"loss": 0.6114, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.18548306822776794, |
|
"rewards/margins": 0.3985109329223633, |
|
"rewards/rejected": -0.5839939117431641, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.6066566887606015, |
|
"grad_norm": 86.55652744872312, |
|
"learning_rate": 1.6811015494059045e-07, |
|
"logits/chosen": -1.340354323387146, |
|
"logits/rejected": -1.372084617614746, |
|
"logps/chosen": -139.92453002929688, |
|
"logps/rejected": -170.04629516601562, |
|
"loss": 0.5944, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.09279327094554901, |
|
"rewards/margins": 0.36332762241363525, |
|
"rewards/rejected": -0.27053436636924744, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 0.6089373530040625, |
|
"grad_norm": 86.16548448796566, |
|
"learning_rate": 1.678311836269616e-07, |
|
"logits/chosen": -1.2494535446166992, |
|
"logits/rejected": -1.3522768020629883, |
|
"logps/chosen": -133.36306762695312, |
|
"logps/rejected": -140.9103240966797, |
|
"loss": 0.626, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.22486190497875214, |
|
"rewards/margins": 0.10757801681756973, |
|
"rewards/rejected": -0.3324398994445801, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 0.6112180172475233, |
|
"grad_norm": 78.5036790731728, |
|
"learning_rate": 1.6755123108851842e-07, |
|
"logits/chosen": -1.2597969770431519, |
|
"logits/rejected": -1.3842471837997437, |
|
"logps/chosen": -102.91565704345703, |
|
"logps/rejected": -149.77996826171875, |
|
"loss": 0.6558, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.049101147800683975, |
|
"rewards/margins": 0.2616952359676361, |
|
"rewards/rejected": -0.21259410679340363, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 0.6134986814909843, |
|
"grad_norm": 97.00034287566903, |
|
"learning_rate": 1.6727030137496727e-07, |
|
"logits/chosen": -1.333500862121582, |
|
"logits/rejected": -1.407701015472412, |
|
"logps/chosen": -168.34678649902344, |
|
"logps/rejected": -212.6487579345703, |
|
"loss": 0.6333, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.23587258160114288, |
|
"rewards/margins": 0.40141719579696655, |
|
"rewards/rejected": -0.637289822101593, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 0.6157793457344451, |
|
"grad_norm": 93.14382617410926, |
|
"learning_rate": 1.6698839855015007e-07, |
|
"logits/chosen": -1.3826831579208374, |
|
"logits/rejected": -1.3780534267425537, |
|
"logps/chosen": -197.9063262939453, |
|
"logps/rejected": -201.8406982421875, |
|
"loss": 0.653, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": -0.19411635398864746, |
|
"rewards/margins": 0.06094612553715706, |
|
"rewards/rejected": -0.2550624907016754, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.6180600099779061, |
|
"grad_norm": 84.06073173280129, |
|
"learning_rate": 1.6670552669198546e-07, |
|
"logits/chosen": -1.459622859954834, |
|
"logits/rejected": -1.5475270748138428, |
|
"logps/chosen": -149.38450622558594, |
|
"logps/rejected": -174.00827026367188, |
|
"loss": 0.6878, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.22387176752090454, |
|
"rewards/margins": 0.04707152396440506, |
|
"rewards/rejected": -0.2709433138370514, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 0.620340674221367, |
|
"grad_norm": 75.86401555746195, |
|
"learning_rate": 1.6642168989240974e-07, |
|
"logits/chosen": -1.3617055416107178, |
|
"logits/rejected": -1.3725388050079346, |
|
"logps/chosen": -177.85269165039062, |
|
"logps/rejected": -204.47271728515625, |
|
"loss": 0.6393, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.2369232326745987, |
|
"rewards/margins": 0.4420955181121826, |
|
"rewards/rejected": -0.6790187358856201, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 0.6226213384648279, |
|
"grad_norm": 99.08946862124856, |
|
"learning_rate": 1.6613689225731787e-07, |
|
"logits/chosen": -1.437178611755371, |
|
"logits/rejected": -1.4497863054275513, |
|
"logps/chosen": -152.61231994628906, |
|
"logps/rejected": -163.36196899414062, |
|
"loss": 0.6514, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.24256393313407898, |
|
"rewards/margins": 0.14335207641124725, |
|
"rewards/rejected": -0.3859160244464874, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 0.6249020027082888, |
|
"grad_norm": 82.17381709045306, |
|
"learning_rate": 1.6585113790650386e-07, |
|
"logits/chosen": -1.3418258428573608, |
|
"logits/rejected": -1.357533574104309, |
|
"logps/chosen": -150.5135955810547, |
|
"logps/rejected": -168.92135620117188, |
|
"loss": 0.6404, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.1530638188123703, |
|
"rewards/margins": 0.16317017376422882, |
|
"rewards/rejected": -0.3162340223789215, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 0.6271826669517497, |
|
"grad_norm": 98.6032079855375, |
|
"learning_rate": 1.6556443097360133e-07, |
|
"logits/chosen": -1.305914282798767, |
|
"logits/rejected": -1.3782196044921875, |
|
"logps/chosen": -187.4833526611328, |
|
"logps/rejected": -213.8814697265625, |
|
"loss": 0.6415, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.03956926241517067, |
|
"rewards/margins": 0.45714423060417175, |
|
"rewards/rejected": -0.4967134892940521, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.6294633311952106, |
|
"grad_norm": 89.5606898436777, |
|
"learning_rate": 1.6527677560602363e-07, |
|
"logits/chosen": -1.4830090999603271, |
|
"logits/rejected": -1.4891180992126465, |
|
"logps/chosen": -178.29112243652344, |
|
"logps/rejected": -182.18019104003906, |
|
"loss": 0.658, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.10583019256591797, |
|
"rewards/margins": 0.16508930921554565, |
|
"rewards/rejected": -0.270919531583786, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 0.6317439954386715, |
|
"grad_norm": 94.29213072448552, |
|
"learning_rate": 1.6498817596490388e-07, |
|
"logits/chosen": -1.3679401874542236, |
|
"logits/rejected": -1.397881031036377, |
|
"logps/chosen": -170.22218322753906, |
|
"logps/rejected": -190.22608947753906, |
|
"loss": 0.6632, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.07908223569393158, |
|
"rewards/margins": 0.2489246428012848, |
|
"rewards/rejected": -0.32800689339637756, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 0.6340246596821324, |
|
"grad_norm": 97.6468147380264, |
|
"learning_rate": 1.6469863622503475e-07, |
|
"logits/chosen": -1.3809480667114258, |
|
"logits/rejected": -1.3862886428833008, |
|
"logps/chosen": -161.6693878173828, |
|
"logps/rejected": -186.37608337402344, |
|
"loss": 0.6762, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.08833756297826767, |
|
"rewards/margins": 0.3514275848865509, |
|
"rewards/rejected": -0.4397651255130768, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 0.6363053239255934, |
|
"grad_norm": 90.63822694085678, |
|
"learning_rate": 1.6440816057480812e-07, |
|
"logits/chosen": -1.4811036586761475, |
|
"logits/rejected": -1.554297924041748, |
|
"logps/chosen": -214.304931640625, |
|
"logps/rejected": -267.9236145019531, |
|
"loss": 0.638, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.08694766461849213, |
|
"rewards/margins": 0.4463149309158325, |
|
"rewards/rejected": -0.359367311000824, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 0.6385859881690542, |
|
"grad_norm": 109.09749536253133, |
|
"learning_rate": 1.641167532161545e-07, |
|
"logits/chosen": -1.3816839456558228, |
|
"logits/rejected": -1.3854830265045166, |
|
"logps/chosen": -146.0878448486328, |
|
"logps/rejected": -151.40838623046875, |
|
"loss": 0.6199, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.2251240313053131, |
|
"rewards/margins": 0.2835414409637451, |
|
"rewards/rejected": -0.5086655020713806, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.6408666524125152, |
|
"grad_norm": 85.56418756561474, |
|
"learning_rate": 1.63824418364482e-07, |
|
"logits/chosen": -1.3435323238372803, |
|
"logits/rejected": -1.4011439085006714, |
|
"logps/chosen": -171.78982543945312, |
|
"logps/rejected": -188.12899780273438, |
|
"loss": 0.6342, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.03885346278548241, |
|
"rewards/margins": 0.2555912137031555, |
|
"rewards/rejected": -0.29444462060928345, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 0.643147316655976, |
|
"grad_norm": 91.83203136203487, |
|
"learning_rate": 1.6353116024861583e-07, |
|
"logits/chosen": -1.3896362781524658, |
|
"logits/rejected": -1.3640623092651367, |
|
"logps/chosen": -186.40646362304688, |
|
"logps/rejected": -194.1944580078125, |
|
"loss": 0.6166, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -0.14515984058380127, |
|
"rewards/margins": 0.4773363471031189, |
|
"rewards/rejected": -0.6224961876869202, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 0.645427980899437, |
|
"grad_norm": 81.53572929971928, |
|
"learning_rate": 1.6323698311073665e-07, |
|
"logits/chosen": -1.2487475872039795, |
|
"logits/rejected": -1.298531413078308, |
|
"logps/chosen": -157.927734375, |
|
"logps/rejected": -172.3856658935547, |
|
"loss": 0.6697, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.2098311483860016, |
|
"rewards/margins": 0.2927185297012329, |
|
"rewards/rejected": -0.5025496482849121, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 0.6477086451428978, |
|
"grad_norm": 114.13313398743495, |
|
"learning_rate": 1.6294189120631955e-07, |
|
"logits/chosen": -1.4274886846542358, |
|
"logits/rejected": -1.5344736576080322, |
|
"logps/chosen": -182.54502868652344, |
|
"logps/rejected": -216.51434326171875, |
|
"loss": 0.6837, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -0.11003254354000092, |
|
"rewards/margins": 0.4595148265361786, |
|
"rewards/rejected": -0.5695474147796631, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 0.6499893093863588, |
|
"grad_norm": 91.72506895488603, |
|
"learning_rate": 1.6264588880407218e-07, |
|
"logits/chosen": -1.3987553119659424, |
|
"logits/rejected": -1.375367283821106, |
|
"logps/chosen": -150.62136840820312, |
|
"logps/rejected": -147.3133087158203, |
|
"loss": 0.655, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.012118123471736908, |
|
"rewards/margins": 0.23711130023002625, |
|
"rewards/rejected": -0.22499315440654755, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.6522699736298196, |
|
"grad_norm": 107.88986883450896, |
|
"learning_rate": 1.6234898018587336e-07, |
|
"logits/chosen": -1.2692500352859497, |
|
"logits/rejected": -1.369061827659607, |
|
"logps/chosen": -139.50912475585938, |
|
"logps/rejected": -160.3131561279297, |
|
"loss": 0.6783, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": -0.21677826344966888, |
|
"rewards/margins": 0.14994385838508606, |
|
"rewards/rejected": -0.36672213673591614, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 0.6545506378732806, |
|
"grad_norm": 78.69743235560887, |
|
"learning_rate": 1.620511696467108e-07, |
|
"logits/chosen": -1.4847999811172485, |
|
"logits/rejected": -1.4455361366271973, |
|
"logps/chosen": -143.21315002441406, |
|
"logps/rejected": -142.0272674560547, |
|
"loss": 0.5931, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.023255709558725357, |
|
"rewards/margins": 0.20145481824874878, |
|
"rewards/rejected": -0.17819911241531372, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 0.6568313021167415, |
|
"grad_norm": 88.95142783400384, |
|
"learning_rate": 1.6175246149461918e-07, |
|
"logits/chosen": -1.2740530967712402, |
|
"logits/rejected": -1.454386830329895, |
|
"logps/chosen": -151.84759521484375, |
|
"logps/rejected": -200.50332641601562, |
|
"loss": 0.6686, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.14912785589694977, |
|
"rewards/margins": 0.19120272994041443, |
|
"rewards/rejected": -0.3403306007385254, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 0.6591119663602024, |
|
"grad_norm": 106.08082954445281, |
|
"learning_rate": 1.614528600506178e-07, |
|
"logits/chosen": -1.3865227699279785, |
|
"logits/rejected": -1.375576138496399, |
|
"logps/chosen": -190.82093811035156, |
|
"logps/rejected": -221.74673461914062, |
|
"loss": 0.6078, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.02210266701877117, |
|
"rewards/margins": 0.2864355444908142, |
|
"rewards/rejected": -0.2643328607082367, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 0.6613926306036633, |
|
"grad_norm": 87.3395501088531, |
|
"learning_rate": 1.6115236964864796e-07, |
|
"logits/chosen": -1.4399878978729248, |
|
"logits/rejected": -1.4395971298217773, |
|
"logps/chosen": -173.22372436523438, |
|
"logps/rejected": -176.63436889648438, |
|
"loss": 0.6855, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.08596701920032501, |
|
"rewards/margins": 0.2504124939441681, |
|
"rewards/rejected": -0.3363794982433319, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.6636732948471242, |
|
"grad_norm": 106.76668867973154, |
|
"learning_rate": 1.6085099463551042e-07, |
|
"logits/chosen": -1.4452544450759888, |
|
"logits/rejected": -1.5641038417816162, |
|
"logps/chosen": -168.6453094482422, |
|
"logps/rejected": -173.80926513671875, |
|
"loss": 0.6725, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.173700749874115, |
|
"rewards/margins": 0.12216023355722427, |
|
"rewards/rejected": -0.29586100578308105, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 0.6659539590905851, |
|
"grad_norm": 87.05178703169928, |
|
"learning_rate": 1.6054873937080243e-07, |
|
"logits/chosen": -1.304762601852417, |
|
"logits/rejected": -1.368096113204956, |
|
"logps/chosen": -172.9827423095703, |
|
"logps/rejected": -219.77606201171875, |
|
"loss": 0.6104, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.15153777599334717, |
|
"rewards/margins": 0.4469776451587677, |
|
"rewards/rejected": -0.5985154509544373, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 0.668234623334046, |
|
"grad_norm": 110.06770845486187, |
|
"learning_rate": 1.6024560822685464e-07, |
|
"logits/chosen": -1.2189353704452515, |
|
"logits/rejected": -1.2109813690185547, |
|
"logps/chosen": -124.88150024414062, |
|
"logps/rejected": -134.57229614257812, |
|
"loss": 0.6398, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.04388846457004547, |
|
"rewards/margins": 0.09341102838516235, |
|
"rewards/rejected": -0.13729947805404663, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 0.670515287577507, |
|
"grad_norm": 96.86048631567245, |
|
"learning_rate": 1.5994160558866803e-07, |
|
"logits/chosen": -1.4283928871154785, |
|
"logits/rejected": -1.454673171043396, |
|
"logps/chosen": -141.67201232910156, |
|
"logps/rejected": -165.86553955078125, |
|
"loss": 0.6099, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.005323350429534912, |
|
"rewards/margins": 0.37266457080841064, |
|
"rewards/rejected": -0.37798795104026794, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 0.6727959518209679, |
|
"grad_norm": 87.33194066270343, |
|
"learning_rate": 1.5963673585385014e-07, |
|
"logits/chosen": -1.3119065761566162, |
|
"logits/rejected": -1.3611897230148315, |
|
"logps/chosen": -149.3885498046875, |
|
"logps/rejected": -233.38589477539062, |
|
"loss": 0.6465, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -0.1451117843389511, |
|
"rewards/margins": 0.5241594314575195, |
|
"rewards/rejected": -0.6692711710929871, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.6750766160644288, |
|
"grad_norm": 89.80632782962597, |
|
"learning_rate": 1.5933100343255182e-07, |
|
"logits/chosen": -1.478175401687622, |
|
"logits/rejected": -1.5088427066802979, |
|
"logps/chosen": -160.82077026367188, |
|
"logps/rejected": -175.51266479492188, |
|
"loss": 0.6521, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.026576366275548935, |
|
"rewards/margins": 0.050725750625133514, |
|
"rewards/rejected": -0.07730211317539215, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 0.6773572803078897, |
|
"grad_norm": 74.35318626181845, |
|
"learning_rate": 1.5902441274740328e-07, |
|
"logits/chosen": -1.3951764106750488, |
|
"logits/rejected": -1.5363534688949585, |
|
"logps/chosen": -199.1707000732422, |
|
"logps/rejected": -223.73233032226562, |
|
"loss": 0.6397, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -0.12182176113128662, |
|
"rewards/margins": 0.23543262481689453, |
|
"rewards/rejected": -0.35725438594818115, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 0.6796379445513506, |
|
"grad_norm": 94.79924897563647, |
|
"learning_rate": 1.5871696823344998e-07, |
|
"logits/chosen": -1.296017050743103, |
|
"logits/rejected": -1.3113356828689575, |
|
"logps/chosen": -217.41751098632812, |
|
"logps/rejected": -248.5911865234375, |
|
"loss": 0.6955, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.08776041865348816, |
|
"rewards/margins": -0.076215460896492, |
|
"rewards/rejected": -0.011544965207576752, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 0.6819186087948115, |
|
"grad_norm": 92.59481105091255, |
|
"learning_rate": 1.584086743380887e-07, |
|
"logits/chosen": -1.4293127059936523, |
|
"logits/rejected": -1.4433321952819824, |
|
"logps/chosen": -118.54342651367188, |
|
"logps/rejected": -134.85447692871094, |
|
"loss": 0.6368, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.17175720632076263, |
|
"rewards/margins": 0.14412283897399902, |
|
"rewards/rejected": -0.31588006019592285, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 0.6841992730382724, |
|
"grad_norm": 89.66618993893586, |
|
"learning_rate": 1.580995355210031e-07, |
|
"logits/chosen": -1.2930192947387695, |
|
"logits/rejected": -1.368417501449585, |
|
"logps/chosen": -156.90586853027344, |
|
"logps/rejected": -204.53976440429688, |
|
"loss": 0.655, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.10334792733192444, |
|
"rewards/margins": 0.27799686789512634, |
|
"rewards/rejected": -0.3813447952270508, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.6841992730382724, |
|
"eval_logits/chosen": -1.4766055345535278, |
|
"eval_logits/rejected": -1.4625221490859985, |
|
"eval_logps/chosen": -128.85867309570312, |
|
"eval_logps/rejected": -132.3487091064453, |
|
"eval_loss": 0.6277530789375305, |
|
"eval_rewards/accuracies": 0.6000000238418579, |
|
"eval_rewards/chosen": 0.10392862558364868, |
|
"eval_rewards/margins": 0.13252468407154083, |
|
"eval_rewards/rejected": -0.028596054762601852, |
|
"eval_runtime": 20.497, |
|
"eval_samples_per_second": 4.879, |
|
"eval_steps_per_second": 1.22, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.6864799372817333, |
|
"grad_norm": 99.81191522440201, |
|
"learning_rate": 1.577895562540992e-07, |
|
"logits/chosen": -1.5043866634368896, |
|
"logits/rejected": -1.5639445781707764, |
|
"logps/chosen": -176.97195434570312, |
|
"logps/rejected": -200.05955505371094, |
|
"loss": 0.6298, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.08353392779827118, |
|
"rewards/margins": 0.1179167851805687, |
|
"rewards/rejected": -0.20145073533058167, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 0.6887606015251942, |
|
"grad_norm": 76.70830865796096, |
|
"learning_rate": 1.574787410214407e-07, |
|
"logits/chosen": -1.409602165222168, |
|
"logits/rejected": -1.4615368843078613, |
|
"logps/chosen": -155.466064453125, |
|
"logps/rejected": -179.92701721191406, |
|
"loss": 0.5929, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.06697350740432739, |
|
"rewards/margins": 0.43057531118392944, |
|
"rewards/rejected": -0.4975488483905792, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 0.6910412657686551, |
|
"grad_norm": 94.3888285857028, |
|
"learning_rate": 1.571670943191841e-07, |
|
"logits/chosen": -1.4215627908706665, |
|
"logits/rejected": -1.4542145729064941, |
|
"logps/chosen": -179.4984130859375, |
|
"logps/rejected": -194.54330444335938, |
|
"loss": 0.6238, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.06045013293623924, |
|
"rewards/margins": 0.22754909098148346, |
|
"rewards/rejected": -0.2879992127418518, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 0.6933219300121161, |
|
"grad_norm": 85.0198174251867, |
|
"learning_rate": 1.5685462065551372e-07, |
|
"logits/chosen": -1.3846269845962524, |
|
"logits/rejected": -1.403570532798767, |
|
"logps/chosen": -119.62281799316406, |
|
"logps/rejected": -122.60442352294922, |
|
"loss": 0.6716, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": 0.05273966118693352, |
|
"rewards/margins": 0.23132431507110596, |
|
"rewards/rejected": -0.17858465015888214, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 0.6956025942555769, |
|
"grad_norm": 106.39373760782738, |
|
"learning_rate": 1.5654132455057648e-07, |
|
"logits/chosen": -1.4965307712554932, |
|
"logits/rejected": -1.5246574878692627, |
|
"logps/chosen": -172.1771240234375, |
|
"logps/rejected": -178.22027587890625, |
|
"loss": 0.6898, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": -0.24939650297164917, |
|
"rewards/margins": -0.02339550480246544, |
|
"rewards/rejected": -0.22600099444389343, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.6978832584990379, |
|
"grad_norm": 90.59146954300502, |
|
"learning_rate": 1.562272105364164e-07, |
|
"logits/chosen": -1.2977298498153687, |
|
"logits/rejected": -1.356859803199768, |
|
"logps/chosen": -184.52252197265625, |
|
"logps/rejected": -211.69007873535156, |
|
"loss": 0.6426, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.14194564521312714, |
|
"rewards/margins": 0.10268077999353409, |
|
"rewards/rejected": -0.24462641775608063, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 0.7001639227424987, |
|
"grad_norm": 132.7396200312291, |
|
"learning_rate": 1.5591228315690912e-07, |
|
"logits/chosen": -1.312849760055542, |
|
"logits/rejected": -1.3643529415130615, |
|
"logps/chosen": -183.94964599609375, |
|
"logps/rejected": -215.9215087890625, |
|
"loss": 0.6904, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.048665039241313934, |
|
"rewards/margins": 0.34853097796440125, |
|
"rewards/rejected": -0.3971959948539734, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 0.7024445869859597, |
|
"grad_norm": 85.32152495878141, |
|
"learning_rate": 1.5559654696769627e-07, |
|
"logits/chosen": -1.3156163692474365, |
|
"logits/rejected": -1.3534281253814697, |
|
"logps/chosen": -135.82110595703125, |
|
"logps/rejected": -146.7247314453125, |
|
"loss": 0.6404, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.19835428893566132, |
|
"rewards/margins": 0.1589125692844391, |
|
"rewards/rejected": -0.3572668731212616, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 0.7047252512294205, |
|
"grad_norm": 97.72254578285961, |
|
"learning_rate": 1.5528000653611933e-07, |
|
"logits/chosen": -1.3361620903015137, |
|
"logits/rejected": -1.3171348571777344, |
|
"logps/chosen": -133.0448760986328, |
|
"logps/rejected": -157.213623046875, |
|
"loss": 0.6296, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.2093026041984558, |
|
"rewards/margins": 0.04571903124451637, |
|
"rewards/rejected": -0.2550216317176819, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 0.7070059154728815, |
|
"grad_norm": 81.04476758820772, |
|
"learning_rate": 1.5496266644115383e-07, |
|
"logits/chosen": -1.335296869277954, |
|
"logits/rejected": -1.4067720174789429, |
|
"logps/chosen": -166.6826934814453, |
|
"logps/rejected": -201.6751251220703, |
|
"loss": 0.5956, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.02005193382501602, |
|
"rewards/margins": 0.4108864665031433, |
|
"rewards/rejected": -0.3908345103263855, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.7092865797163423, |
|
"grad_norm": 117.48862378162204, |
|
"learning_rate": 1.5464453127334292e-07, |
|
"logits/chosen": -1.3234319686889648, |
|
"logits/rejected": -1.3931816816329956, |
|
"logps/chosen": -211.87327575683594, |
|
"logps/rejected": -244.9843292236328, |
|
"loss": 0.6435, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.23849377036094666, |
|
"rewards/margins": 0.13237711787223816, |
|
"rewards/rejected": -0.3708708584308624, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 0.7115672439598033, |
|
"grad_norm": 97.28790849798257, |
|
"learning_rate": 1.5432560563473102e-07, |
|
"logits/chosen": -1.3715243339538574, |
|
"logits/rejected": -1.4115769863128662, |
|
"logps/chosen": -167.96871948242188, |
|
"logps/rejected": -195.65823364257812, |
|
"loss": 0.6435, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.18149994313716888, |
|
"rewards/margins": 0.21998891234397888, |
|
"rewards/rejected": -0.4014888405799866, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 0.7138479082032642, |
|
"grad_norm": 74.77329414265908, |
|
"learning_rate": 1.5400589413879728e-07, |
|
"logits/chosen": -1.3941603899002075, |
|
"logits/rejected": -1.4356334209442139, |
|
"logps/chosen": -146.8162841796875, |
|
"logps/rejected": -166.09603881835938, |
|
"loss": 0.6115, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": 0.0008982233703136444, |
|
"rewards/margins": 0.2602536678314209, |
|
"rewards/rejected": -0.25935542583465576, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 0.7161285724467251, |
|
"grad_norm": 102.40110691278403, |
|
"learning_rate": 1.5368540141038877e-07, |
|
"logits/chosen": -1.33980393409729, |
|
"logits/rejected": -1.3720303773880005, |
|
"logps/chosen": -177.0899658203125, |
|
"logps/rejected": -177.4815216064453, |
|
"loss": 0.6489, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.17625108361244202, |
|
"rewards/margins": 0.22537095844745636, |
|
"rewards/rejected": -0.40162205696105957, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 0.718409236690186, |
|
"grad_norm": 77.60752469183791, |
|
"learning_rate": 1.5336413208565372e-07, |
|
"logits/chosen": -1.424604058265686, |
|
"logits/rejected": -1.4951881170272827, |
|
"logps/chosen": -217.2845458984375, |
|
"logps/rejected": -298.6929016113281, |
|
"loss": 0.5522, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.31031134724617004, |
|
"rewards/margins": 0.769706130027771, |
|
"rewards/rejected": -1.0800174474716187, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.7206899009336469, |
|
"grad_norm": 86.60572317721935, |
|
"learning_rate": 1.5304209081197424e-07, |
|
"logits/chosen": -1.2321975231170654, |
|
"logits/rejected": -1.1972179412841797, |
|
"logps/chosen": -162.22976684570312, |
|
"logps/rejected": -173.07118225097656, |
|
"loss": 0.651, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.0614384189248085, |
|
"rewards/margins": 0.27045127749443054, |
|
"rewards/rejected": -0.33188968896865845, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 0.7229705651771078, |
|
"grad_norm": 105.40040114011227, |
|
"learning_rate": 1.5271928224789933e-07, |
|
"logits/chosen": -1.3975963592529297, |
|
"logits/rejected": -1.4231470823287964, |
|
"logps/chosen": -202.03836059570312, |
|
"logps/rejected": -236.25344848632812, |
|
"loss": 0.6755, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.32155218720436096, |
|
"rewards/margins": 0.17959336936473846, |
|
"rewards/rejected": -0.5011455416679382, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 0.7252512294205687, |
|
"grad_norm": 81.36310946045379, |
|
"learning_rate": 1.5239571106307727e-07, |
|
"logits/chosen": -1.3148503303527832, |
|
"logits/rejected": -1.3296858072280884, |
|
"logps/chosen": -97.38423919677734, |
|
"logps/rejected": -119.96273803710938, |
|
"loss": 0.632, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.0010671375785022974, |
|
"rewards/margins": 0.2470298409461975, |
|
"rewards/rejected": -0.2459626942873001, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 0.7275318936640296, |
|
"grad_norm": 96.83510386866747, |
|
"learning_rate": 1.5207138193818824e-07, |
|
"logits/chosen": -1.4544883966445923, |
|
"logits/rejected": -1.4232516288757324, |
|
"logps/chosen": -173.56967163085938, |
|
"logps/rejected": -202.63575744628906, |
|
"loss": 0.618, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.11890554428100586, |
|
"rewards/margins": 0.4641531705856323, |
|
"rewards/rejected": -0.5830587148666382, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 0.7298125579074906, |
|
"grad_norm": 98.36905363836429, |
|
"learning_rate": 1.5174629956487657e-07, |
|
"logits/chosen": -1.3621355295181274, |
|
"logits/rejected": -1.470801591873169, |
|
"logps/chosen": -152.83348083496094, |
|
"logps/rejected": -180.09095764160156, |
|
"loss": 0.6283, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.1331956833600998, |
|
"rewards/margins": 0.12644681334495544, |
|
"rewards/rejected": -0.25964248180389404, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.7320932221509515, |
|
"grad_norm": 99.21748796643946, |
|
"learning_rate": 1.514204686456828e-07, |
|
"logits/chosen": -1.4174845218658447, |
|
"logits/rejected": -1.5134220123291016, |
|
"logps/chosen": -199.45867919921875, |
|
"logps/rejected": -233.53509521484375, |
|
"loss": 0.6424, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": 0.029536306858062744, |
|
"rewards/margins": 0.40678656101226807, |
|
"rewards/rejected": -0.3772502541542053, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 0.7343738863944124, |
|
"grad_norm": 104.85775879115107, |
|
"learning_rate": 1.5109389389397564e-07, |
|
"logits/chosen": -1.309433102607727, |
|
"logits/rejected": -1.3806891441345215, |
|
"logps/chosen": -198.72012329101562, |
|
"logps/rejected": -229.50213623046875, |
|
"loss": 0.636, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.2979825735092163, |
|
"rewards/margins": 0.2780360281467438, |
|
"rewards/rejected": -0.5760185718536377, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 0.7366545506378733, |
|
"grad_norm": 83.70012784480859, |
|
"learning_rate": 1.50766580033884e-07, |
|
"logits/chosen": -1.3813220262527466, |
|
"logits/rejected": -1.4268507957458496, |
|
"logps/chosen": -144.8968048095703, |
|
"logps/rejected": -168.29891967773438, |
|
"loss": 0.6301, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.2506921887397766, |
|
"rewards/margins": 0.17526662349700928, |
|
"rewards/rejected": -0.4259588122367859, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 0.7389352148813342, |
|
"grad_norm": 76.15531374501609, |
|
"learning_rate": 1.5043853180022838e-07, |
|
"logits/chosen": -1.3109767436981201, |
|
"logits/rejected": -1.393604040145874, |
|
"logps/chosen": -123.48320770263672, |
|
"logps/rejected": -146.57994079589844, |
|
"loss": 0.627, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.030930444598197937, |
|
"rewards/margins": 0.237480029463768, |
|
"rewards/rejected": -0.26841047406196594, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 0.7412158791247951, |
|
"grad_norm": 80.96947358910683, |
|
"learning_rate": 1.5010975393845257e-07, |
|
"logits/chosen": -1.249889612197876, |
|
"logits/rejected": -1.2592819929122925, |
|
"logps/chosen": -210.37823486328125, |
|
"logps/rejected": -225.4598388671875, |
|
"loss": 0.6471, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.39817529916763306, |
|
"rewards/margins": 0.08833743631839752, |
|
"rewards/rejected": -0.4865127205848694, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.743496543368256, |
|
"grad_norm": 87.1921118015952, |
|
"learning_rate": 1.4978025120455482e-07, |
|
"logits/chosen": -1.3601963520050049, |
|
"logits/rejected": -1.376826286315918, |
|
"logps/chosen": -167.55612182617188, |
|
"logps/rejected": -175.45620727539062, |
|
"loss": 0.6323, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.03152412921190262, |
|
"rewards/margins": 0.33548736572265625, |
|
"rewards/rejected": -0.30396321415901184, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 0.745777207611717, |
|
"grad_norm": 101.45108780703592, |
|
"learning_rate": 1.4945002836501935e-07, |
|
"logits/chosen": -1.4448864459991455, |
|
"logits/rejected": -1.5169280767440796, |
|
"logps/chosen": -251.17156982421875, |
|
"logps/rejected": -303.9085693359375, |
|
"loss": 0.6505, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.29235783219337463, |
|
"rewards/margins": 0.28852301836013794, |
|
"rewards/rejected": -0.580880880355835, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 0.7480578718551778, |
|
"grad_norm": 78.1814307088951, |
|
"learning_rate": 1.4911909019674702e-07, |
|
"logits/chosen": -1.3950117826461792, |
|
"logits/rejected": -1.3936371803283691, |
|
"logps/chosen": -221.95086669921875, |
|
"logps/rejected": -246.12425231933594, |
|
"loss": 0.7095, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.40981170535087585, |
|
"rewards/margins": 0.04004380851984024, |
|
"rewards/rejected": -0.4498555362224579, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 0.7503385360986388, |
|
"grad_norm": 86.3290682570179, |
|
"learning_rate": 1.4878744148698654e-07, |
|
"logits/chosen": -1.4538379907608032, |
|
"logits/rejected": -1.4509309530258179, |
|
"logps/chosen": -240.70957946777344, |
|
"logps/rejected": -277.9938049316406, |
|
"loss": 0.6274, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -0.10867351293563843, |
|
"rewards/margins": 0.7040268182754517, |
|
"rewards/rejected": -0.8127003908157349, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 0.7526192003420996, |
|
"grad_norm": 76.663224249533, |
|
"learning_rate": 1.48455087033265e-07, |
|
"logits/chosen": -1.347740888595581, |
|
"logits/rejected": -1.4315268993377686, |
|
"logps/chosen": -175.90298461914062, |
|
"logps/rejected": -204.00741577148438, |
|
"loss": 0.589, |
|
"rewards/accuracies": 0.8125, |
|
"rewards/chosen": 0.0198584645986557, |
|
"rewards/margins": 0.3997357189655304, |
|
"rewards/rejected": -0.3798772692680359, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.7548998645855606, |
|
"grad_norm": 89.45613833312, |
|
"learning_rate": 1.4812203164331865e-07, |
|
"logits/chosen": -1.3782517910003662, |
|
"logits/rejected": -1.3579304218292236, |
|
"logps/chosen": -106.40101623535156, |
|
"logps/rejected": -115.47900390625, |
|
"loss": 0.6276, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.002452205866575241, |
|
"rewards/margins": 0.2672392427921295, |
|
"rewards/rejected": -0.2647870182991028, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 0.7571805288290214, |
|
"grad_norm": 90.62698292022132, |
|
"learning_rate": 1.4778828013502315e-07, |
|
"logits/chosen": -1.3181474208831787, |
|
"logits/rejected": -1.2984850406646729, |
|
"logps/chosen": -121.110595703125, |
|
"logps/rejected": -144.5844268798828, |
|
"loss": 0.6345, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": -0.14049722254276276, |
|
"rewards/margins": 0.16425339877605438, |
|
"rewards/rejected": -0.30475062131881714, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 0.7594611930724824, |
|
"grad_norm": 90.46387262398376, |
|
"learning_rate": 1.474538373363241e-07, |
|
"logits/chosen": -1.2038509845733643, |
|
"logits/rejected": -1.1771519184112549, |
|
"logps/chosen": -123.54405212402344, |
|
"logps/rejected": -128.04415893554688, |
|
"loss": 0.6513, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.0023752544075250626, |
|
"rewards/margins": 0.20540934801101685, |
|
"rewards/rejected": -0.20778462290763855, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 0.7617418573159432, |
|
"grad_norm": 98.8814238604518, |
|
"learning_rate": 1.4711870808516705e-07, |
|
"logits/chosen": -1.3800256252288818, |
|
"logits/rejected": -1.4212613105773926, |
|
"logps/chosen": -143.5724639892578, |
|
"logps/rejected": -149.91879272460938, |
|
"loss": 0.6367, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -0.25999924540519714, |
|
"rewards/margins": 0.30095675587654114, |
|
"rewards/rejected": -0.5609559416770935, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 0.7640225215594042, |
|
"grad_norm": 91.14237505783925, |
|
"learning_rate": 1.4678289722942755e-07, |
|
"logits/chosen": -1.4362088441848755, |
|
"logits/rejected": -1.4652812480926514, |
|
"logps/chosen": -238.3679962158203, |
|
"logps/rejected": -245.81390380859375, |
|
"loss": 0.6248, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.2839330732822418, |
|
"rewards/margins": 0.23081044852733612, |
|
"rewards/rejected": -0.5147435069084167, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.766303185802865, |
|
"grad_norm": 92.40909256739194, |
|
"learning_rate": 1.4644640962684106e-07, |
|
"logits/chosen": -1.2180765867233276, |
|
"logits/rejected": -1.2128483057022095, |
|
"logps/chosen": -139.85572814941406, |
|
"logps/rejected": -155.68014526367188, |
|
"loss": 0.6594, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.31525713205337524, |
|
"rewards/margins": 0.2499125599861145, |
|
"rewards/rejected": -0.5651696920394897, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 0.768583850046326, |
|
"grad_norm": 85.52183536423232, |
|
"learning_rate": 1.461092501449326e-07, |
|
"logits/chosen": -1.432080864906311, |
|
"logits/rejected": -1.4694151878356934, |
|
"logps/chosen": -192.5015411376953, |
|
"logps/rejected": -206.3730926513672, |
|
"loss": 0.6026, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.06770830601453781, |
|
"rewards/margins": 0.42051053047180176, |
|
"rewards/rejected": -0.48821884393692017, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 0.7708645142897869, |
|
"grad_norm": 90.10321239819011, |
|
"learning_rate": 1.4577142366094641e-07, |
|
"logits/chosen": -1.357291579246521, |
|
"logits/rejected": -1.4703514575958252, |
|
"logps/chosen": -159.84365844726562, |
|
"logps/rejected": -179.19932556152344, |
|
"loss": 0.6417, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.26335883140563965, |
|
"rewards/margins": 0.19603171944618225, |
|
"rewards/rejected": -0.4593905210494995, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 0.7731451785332478, |
|
"grad_norm": 89.71495801422127, |
|
"learning_rate": 1.454329350617754e-07, |
|
"logits/chosen": -1.2440226078033447, |
|
"logits/rejected": -1.2795155048370361, |
|
"logps/chosen": -142.0009765625, |
|
"logps/rejected": -174.45877075195312, |
|
"loss": 0.8425, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.18753905594348907, |
|
"rewards/margins": 0.41140902042388916, |
|
"rewards/rejected": -0.598948061466217, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 0.7754258427767087, |
|
"grad_norm": 86.78570750881197, |
|
"learning_rate": 1.4509378924389042e-07, |
|
"logits/chosen": -1.3585668802261353, |
|
"logits/rejected": -1.4263851642608643, |
|
"logps/chosen": -138.0133056640625, |
|
"logps/rejected": -149.3664093017578, |
|
"loss": 0.6293, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.08084426820278168, |
|
"rewards/margins": 0.16473475098609924, |
|
"rewards/rejected": -0.24557898938655853, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.7777065070201696, |
|
"grad_norm": 107.27075654983396, |
|
"learning_rate": 1.4475399111326942e-07, |
|
"logits/chosen": -1.3274388313293457, |
|
"logits/rejected": -1.4092543125152588, |
|
"logps/chosen": -205.84066772460938, |
|
"logps/rejected": -229.88511657714844, |
|
"loss": 0.6449, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.28960123658180237, |
|
"rewards/margins": 0.23577138781547546, |
|
"rewards/rejected": -0.5253726243972778, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 0.7799871712636305, |
|
"grad_norm": 78.77386252056206, |
|
"learning_rate": 1.4441354558532653e-07, |
|
"logits/chosen": -1.3599066734313965, |
|
"logits/rejected": -1.4156947135925293, |
|
"logps/chosen": -178.33787536621094, |
|
"logps/rejected": -178.77232360839844, |
|
"loss": 0.644, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.20014306902885437, |
|
"rewards/margins": 0.3175792396068573, |
|
"rewards/rejected": -0.5177222490310669, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 0.7822678355070914, |
|
"grad_norm": 96.56956615564326, |
|
"learning_rate": 1.4407245758484092e-07, |
|
"logits/chosen": -1.4600579738616943, |
|
"logits/rejected": -1.4812790155410767, |
|
"logps/chosen": -200.6736602783203, |
|
"logps/rejected": -215.11302185058594, |
|
"loss": 0.6408, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.31072649359703064, |
|
"rewards/margins": 0.2427929937839508, |
|
"rewards/rejected": -0.5535194873809814, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 0.7845484997505523, |
|
"grad_norm": 93.7282766427954, |
|
"learning_rate": 1.4373073204588556e-07, |
|
"logits/chosen": -1.392540454864502, |
|
"logits/rejected": -1.5576982498168945, |
|
"logps/chosen": -173.74102783203125, |
|
"logps/rejected": -209.58255004882812, |
|
"loss": 0.6396, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.07411661744117737, |
|
"rewards/margins": 0.2528524100780487, |
|
"rewards/rejected": -0.32696905732154846, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 0.7868291639940133, |
|
"grad_norm": 78.03441039978401, |
|
"learning_rate": 1.433883739117558e-07, |
|
"logits/chosen": -1.4191378355026245, |
|
"logits/rejected": -1.376889705657959, |
|
"logps/chosen": -167.5023651123047, |
|
"logps/rejected": -175.2780303955078, |
|
"loss": 0.6232, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.10266311466693878, |
|
"rewards/margins": 0.33145037293434143, |
|
"rewards/rejected": -0.22878730297088623, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.7891098282374741, |
|
"grad_norm": 102.68741637989442, |
|
"learning_rate": 1.4304538813489807e-07, |
|
"logits/chosen": -1.365789771080017, |
|
"logits/rejected": -1.4465529918670654, |
|
"logps/chosen": -241.91314697265625, |
|
"logps/rejected": -287.3153381347656, |
|
"loss": 0.626, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -0.22250030934810638, |
|
"rewards/margins": 0.3011859655380249, |
|
"rewards/rejected": -0.5236862301826477, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 0.7913904924809351, |
|
"grad_norm": 126.01863649654494, |
|
"learning_rate": 1.4270177967683795e-07, |
|
"logits/chosen": -1.1870087385177612, |
|
"logits/rejected": -1.2576615810394287, |
|
"logps/chosen": -193.8482666015625, |
|
"logps/rejected": -214.00135803222656, |
|
"loss": 0.683, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.6029391884803772, |
|
"rewards/margins": 0.08865418285131454, |
|
"rewards/rejected": -0.69159334897995, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 0.793671156724396, |
|
"grad_norm": 105.78445666178517, |
|
"learning_rate": 1.4235755350810852e-07, |
|
"logits/chosen": -1.4039561748504639, |
|
"logits/rejected": -1.483564853668213, |
|
"logps/chosen": -131.8105926513672, |
|
"logps/rejected": -161.2996368408203, |
|
"loss": 0.5968, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -0.06916500627994537, |
|
"rewards/margins": 0.4792707562446594, |
|
"rewards/rejected": -0.5484358072280884, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 0.7959518209678569, |
|
"grad_norm": 111.82639934813994, |
|
"learning_rate": 1.420127146081786e-07, |
|
"logits/chosen": -1.182243824005127, |
|
"logits/rejected": -1.2272214889526367, |
|
"logps/chosen": -142.89569091796875, |
|
"logps/rejected": -167.72943115234375, |
|
"loss": 0.7077, |
|
"rewards/accuracies": 0.4375, |
|
"rewards/chosen": -0.33521801233291626, |
|
"rewards/margins": 0.031463623046875, |
|
"rewards/rejected": -0.36668163537979126, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 0.7982324852113178, |
|
"grad_norm": 93.0448822586256, |
|
"learning_rate": 1.4166726796538043e-07, |
|
"logits/chosen": -1.337456464767456, |
|
"logits/rejected": -1.38818359375, |
|
"logps/chosen": -161.9310760498047, |
|
"logps/rejected": -177.00717163085938, |
|
"loss": 0.5943, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.11526203155517578, |
|
"rewards/margins": 0.27307793498039246, |
|
"rewards/rejected": -0.38833993673324585, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.7982324852113178, |
|
"eval_logits/chosen": -1.4622485637664795, |
|
"eval_logits/rejected": -1.4484994411468506, |
|
"eval_logps/chosen": -129.2548065185547, |
|
"eval_logps/rejected": -133.13601684570312, |
|
"eval_loss": 0.6083559989929199, |
|
"eval_rewards/accuracies": 0.6399999856948853, |
|
"eval_rewards/chosen": 0.06431641429662704, |
|
"eval_rewards/margins": 0.17164374887943268, |
|
"eval_rewards/rejected": -0.10732734948396683, |
|
"eval_runtime": 19.822, |
|
"eval_samples_per_second": 5.045, |
|
"eval_steps_per_second": 1.261, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.8005131494547787, |
|
"grad_norm": 81.54591894736863, |
|
"learning_rate": 1.413212185768378e-07, |
|
"logits/chosen": -1.3647618293762207, |
|
"logits/rejected": -1.4832388162612915, |
|
"logps/chosen": -163.0651092529297, |
|
"logps/rejected": -199.4841766357422, |
|
"loss": 0.6247, |
|
"rewards/accuracies": 0.8125, |
|
"rewards/chosen": 0.14452999830245972, |
|
"rewards/margins": 0.4548623561859131, |
|
"rewards/rejected": -0.31033238768577576, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 0.8027938136982397, |
|
"grad_norm": 88.3455225094938, |
|
"learning_rate": 1.409745714483936e-07, |
|
"logits/chosen": -1.3221628665924072, |
|
"logits/rejected": -1.379399299621582, |
|
"logps/chosen": -124.75774383544922, |
|
"logps/rejected": -142.49942016601562, |
|
"loss": 0.6477, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.004875157028436661, |
|
"rewards/margins": 0.23325148224830627, |
|
"rewards/rejected": -0.23812663555145264, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 0.8050744779417005, |
|
"grad_norm": 104.14495646431855, |
|
"learning_rate": 1.406273315945374e-07, |
|
"logits/chosen": -1.4262161254882812, |
|
"logits/rejected": -1.460605502128601, |
|
"logps/chosen": -188.75218200683594, |
|
"logps/rejected": -196.2736358642578, |
|
"loss": 0.6307, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.09237577021121979, |
|
"rewards/margins": 0.26734021306037903, |
|
"rewards/rejected": -0.35971593856811523, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 0.8073551421851615, |
|
"grad_norm": 88.27883050256283, |
|
"learning_rate": 1.4027950403833294e-07, |
|
"logits/chosen": -1.3447182178497314, |
|
"logits/rejected": -1.3849260807037354, |
|
"logps/chosen": -194.97628784179688, |
|
"logps/rejected": -231.31488037109375, |
|
"loss": 0.608, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.018238365650177002, |
|
"rewards/margins": 0.4262167513370514, |
|
"rewards/rejected": -0.444455087184906, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 0.8096358064286223, |
|
"grad_norm": 83.81720367834012, |
|
"learning_rate": 1.3993109381134552e-07, |
|
"logits/chosen": -1.458938717842102, |
|
"logits/rejected": -1.4975433349609375, |
|
"logps/chosen": -217.40234375, |
|
"logps/rejected": -230.56849670410156, |
|
"loss": 0.5997, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.06927414983510971, |
|
"rewards/margins": 0.34377092123031616, |
|
"rewards/rejected": -0.41304510831832886, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.8119164706720833, |
|
"grad_norm": 83.92843092918012, |
|
"learning_rate": 1.3958210595356923e-07, |
|
"logits/chosen": -1.295701265335083, |
|
"logits/rejected": -1.3689717054367065, |
|
"logps/chosen": -184.71035766601562, |
|
"logps/rejected": -231.5858154296875, |
|
"loss": 0.6496, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.07217526435852051, |
|
"rewards/margins": 0.5205077528953552, |
|
"rewards/rejected": -0.5926830768585205, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 0.8141971349155441, |
|
"grad_norm": 64.10905806570788, |
|
"learning_rate": 1.3923254551335385e-07, |
|
"logits/chosen": -1.3552354574203491, |
|
"logits/rejected": -1.5121649503707886, |
|
"logps/chosen": -162.65443420410156, |
|
"logps/rejected": -202.2783203125, |
|
"loss": 0.6038, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.21346963942050934, |
|
"rewards/margins": 0.2602553367614746, |
|
"rewards/rejected": -0.47372499108314514, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 0.8164777991590051, |
|
"grad_norm": 96.65072480066111, |
|
"learning_rate": 1.388824175473321e-07, |
|
"logits/chosen": -1.4826998710632324, |
|
"logits/rejected": -1.525536060333252, |
|
"logps/chosen": -157.91253662109375, |
|
"logps/rejected": -189.42970275878906, |
|
"loss": 0.6418, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.15659984946250916, |
|
"rewards/margins": 0.5801843404769897, |
|
"rewards/rejected": -0.7367842197418213, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 0.8187584634024659, |
|
"grad_norm": 133.38988345223416, |
|
"learning_rate": 1.3853172712034618e-07, |
|
"logits/chosen": -1.453382968902588, |
|
"logits/rejected": -1.4554516077041626, |
|
"logps/chosen": -123.51729583740234, |
|
"logps/rejected": -133.98687744140625, |
|
"loss": 0.6316, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.17259877920150757, |
|
"rewards/margins": 0.09528855979442596, |
|
"rewards/rejected": -0.26788732409477234, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 0.8210391276459269, |
|
"grad_norm": 103.6833810750344, |
|
"learning_rate": 1.3818047930537489e-07, |
|
"logits/chosen": -1.4238344430923462, |
|
"logits/rejected": -1.4430863857269287, |
|
"logps/chosen": -185.29090881347656, |
|
"logps/rejected": -190.6168212890625, |
|
"loss": 0.6817, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.1492576152086258, |
|
"rewards/margins": 0.1548936814069748, |
|
"rewards/rejected": -0.304151326417923, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.8233197918893878, |
|
"grad_norm": 104.78480311225195, |
|
"learning_rate": 1.3782867918345986e-07, |
|
"logits/chosen": -1.4426988363265991, |
|
"logits/rejected": -1.474928855895996, |
|
"logps/chosen": -182.13693237304688, |
|
"logps/rejected": -207.8540802001953, |
|
"loss": 0.688, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.28544071316719055, |
|
"rewards/margins": 0.010512780398130417, |
|
"rewards/rejected": -0.2959534525871277, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 0.8256004561328487, |
|
"grad_norm": 83.99721091366573, |
|
"learning_rate": 1.374763318436323e-07, |
|
"logits/chosen": -1.3846015930175781, |
|
"logits/rejected": -1.374580979347229, |
|
"logps/chosen": -188.639892578125, |
|
"logps/rejected": -203.4746856689453, |
|
"loss": 0.6455, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.09084916859865189, |
|
"rewards/margins": 0.13614203035831451, |
|
"rewards/rejected": -0.2269912213087082, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 0.8278811203763096, |
|
"grad_norm": 100.80568473474456, |
|
"learning_rate": 1.371234423828393e-07, |
|
"logits/chosen": -1.3973044157028198, |
|
"logits/rejected": -1.4167506694793701, |
|
"logps/chosen": -221.90057373046875, |
|
"logps/rejected": -223.55625915527344, |
|
"loss": 0.6358, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.5376750826835632, |
|
"rewards/margins": 0.1100195124745369, |
|
"rewards/rejected": -0.6476945281028748, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 0.8301617846197705, |
|
"grad_norm": 95.2695276072168, |
|
"learning_rate": 1.367700159058701e-07, |
|
"logits/chosen": -1.304479718208313, |
|
"logits/rejected": -1.3024879693984985, |
|
"logps/chosen": -132.629150390625, |
|
"logps/rejected": -157.943115234375, |
|
"loss": 0.6385, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.17815637588500977, |
|
"rewards/margins": 0.16010738909244537, |
|
"rewards/rejected": -0.33826377987861633, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 0.8324424488632314, |
|
"grad_norm": 84.57675708069675, |
|
"learning_rate": 1.3641605752528223e-07, |
|
"logits/chosen": -1.345677137374878, |
|
"logits/rejected": -1.369272232055664, |
|
"logps/chosen": -174.67979431152344, |
|
"logps/rejected": -177.95265197753906, |
|
"loss": 0.6631, |
|
"rewards/accuracies": 0.375, |
|
"rewards/chosen": -0.33565065264701843, |
|
"rewards/margins": -0.10611479729413986, |
|
"rewards/rejected": -0.22953587770462036, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.8347231131066923, |
|
"grad_norm": 95.98437315557469, |
|
"learning_rate": 1.3606157236132753e-07, |
|
"logits/chosen": -1.3388769626617432, |
|
"logits/rejected": -1.4168953895568848, |
|
"logps/chosen": -128.86917114257812, |
|
"logps/rejected": -155.26756286621094, |
|
"loss": 0.5995, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.12645135819911957, |
|
"rewards/margins": 0.4350703954696655, |
|
"rewards/rejected": -0.5615217685699463, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 0.8370037773501532, |
|
"grad_norm": 98.72975210848782, |
|
"learning_rate": 1.3570656554187823e-07, |
|
"logits/chosen": -1.293025255203247, |
|
"logits/rejected": -1.3117451667785645, |
|
"logps/chosen": -198.599609375, |
|
"logps/rejected": -195.0568084716797, |
|
"loss": 0.6664, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.3894595205783844, |
|
"rewards/margins": 0.01842033676803112, |
|
"rewards/rejected": -0.40787985920906067, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 0.8392844415936142, |
|
"grad_norm": 83.44989797339352, |
|
"learning_rate": 1.353510422023526e-07, |
|
"logits/chosen": -1.3290454149246216, |
|
"logits/rejected": -1.3762670755386353, |
|
"logps/chosen": -108.33480834960938, |
|
"logps/rejected": -132.43887329101562, |
|
"loss": 0.6699, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.07590903341770172, |
|
"rewards/margins": 0.31655895709991455, |
|
"rewards/rejected": -0.3924679458141327, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 0.841565105837075, |
|
"grad_norm": 94.8310467886239, |
|
"learning_rate": 1.3499500748564075e-07, |
|
"logits/chosen": -1.4289618730545044, |
|
"logits/rejected": -1.4254275560379028, |
|
"logps/chosen": -111.20790100097656, |
|
"logps/rejected": -120.1327896118164, |
|
"loss": 0.6651, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.06255226582288742, |
|
"rewards/margins": 0.16382303833961487, |
|
"rewards/rejected": -0.2263752967119217, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 0.843845770080536, |
|
"grad_norm": 106.55945054665669, |
|
"learning_rate": 1.346384665420302e-07, |
|
"logits/chosen": -1.2947108745574951, |
|
"logits/rejected": -1.3571364879608154, |
|
"logps/chosen": -157.76895141601562, |
|
"logps/rejected": -187.6517333984375, |
|
"loss": 0.6321, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.14570750296115875, |
|
"rewards/margins": 0.39257413148880005, |
|
"rewards/rejected": -0.5382816791534424, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.8461264343239968, |
|
"grad_norm": 112.76083171183551, |
|
"learning_rate": 1.3428142452913133e-07, |
|
"logits/chosen": -1.4128669500350952, |
|
"logits/rejected": -1.4613770246505737, |
|
"logps/chosen": -202.3125457763672, |
|
"logps/rejected": -242.2345428466797, |
|
"loss": 0.6542, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.27778810262680054, |
|
"rewards/margins": 0.3144941031932831, |
|
"rewards/rejected": -0.592282235622406, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 0.8484070985674578, |
|
"grad_norm": 100.11503458436644, |
|
"learning_rate": 1.3392388661180302e-07, |
|
"logits/chosen": -1.3573894500732422, |
|
"logits/rejected": -1.344806432723999, |
|
"logps/chosen": -186.35108947753906, |
|
"logps/rejected": -188.49490356445312, |
|
"loss": 0.6701, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.16072505712509155, |
|
"rewards/margins": 0.19017082452774048, |
|
"rewards/rejected": -0.3508959114551544, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 0.8506877628109187, |
|
"grad_norm": 91.60219394383267, |
|
"learning_rate": 1.3356585796207755e-07, |
|
"logits/chosen": -1.3965177536010742, |
|
"logits/rejected": -1.4244239330291748, |
|
"logps/chosen": -201.50318908691406, |
|
"logps/rejected": -217.1870574951172, |
|
"loss": 0.6803, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.007203985005617142, |
|
"rewards/margins": 0.4475135803222656, |
|
"rewards/rejected": -0.44030967354774475, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 0.8529684270543796, |
|
"grad_norm": 93.27202711374441, |
|
"learning_rate": 1.3320734375908606e-07, |
|
"logits/chosen": -1.4266613721847534, |
|
"logits/rejected": -1.4168901443481445, |
|
"logps/chosen": -208.13624572753906, |
|
"logps/rejected": -238.91502380371094, |
|
"loss": 0.6581, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.2992842197418213, |
|
"rewards/margins": 0.2575336992740631, |
|
"rewards/rejected": -0.556817889213562, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 0.8552490912978405, |
|
"grad_norm": 99.8766408873739, |
|
"learning_rate": 1.328483491889836e-07, |
|
"logits/chosen": -1.427262306213379, |
|
"logits/rejected": -1.4984745979309082, |
|
"logps/chosen": -215.9974365234375, |
|
"logps/rejected": -236.14759826660156, |
|
"loss": 0.6592, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.048569947481155396, |
|
"rewards/margins": 0.15008097887039185, |
|
"rewards/rejected": -0.19865092635154724, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.8575297555413014, |
|
"grad_norm": 82.7550556411054, |
|
"learning_rate": 1.3248887944487393e-07, |
|
"logits/chosen": -1.410501480102539, |
|
"logits/rejected": -1.452683687210083, |
|
"logps/chosen": -122.24147033691406, |
|
"logps/rejected": -137.63116455078125, |
|
"loss": 0.6134, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.11725883185863495, |
|
"rewards/margins": 0.1699836403131485, |
|
"rewards/rejected": -0.05272480472922325, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 0.8598104197847624, |
|
"grad_norm": 88.9952584850963, |
|
"learning_rate": 1.321289397267347e-07, |
|
"logits/chosen": -1.4281740188598633, |
|
"logits/rejected": -1.445003628730774, |
|
"logps/chosen": -177.8570556640625, |
|
"logps/rejected": -181.70663452148438, |
|
"loss": 0.6919, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.2182515561580658, |
|
"rewards/margins": 0.0669536367058754, |
|
"rewards/rejected": -0.28520524501800537, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 0.8620910840282232, |
|
"grad_norm": 100.22335380917765, |
|
"learning_rate": 1.3176853524134198e-07, |
|
"logits/chosen": -1.391309142112732, |
|
"logits/rejected": -1.4018808603286743, |
|
"logps/chosen": -164.9562530517578, |
|
"logps/rejected": -178.225341796875, |
|
"loss": 0.6706, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.04197298735380173, |
|
"rewards/margins": 0.15188831090927124, |
|
"rewards/rejected": -0.19386129081249237, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 0.8643717482716842, |
|
"grad_norm": 100.1727346446802, |
|
"learning_rate": 1.314076712021949e-07, |
|
"logits/chosen": -1.35850989818573, |
|
"logits/rejected": -1.420215129852295, |
|
"logps/chosen": -157.33843994140625, |
|
"logps/rejected": -192.61769104003906, |
|
"loss": 0.6228, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.10978934168815613, |
|
"rewards/margins": 0.24940092861652374, |
|
"rewards/rejected": -0.3591902256011963, |
|
"step": 758 |
|
}, |
|
{ |
|
"epoch": 0.866652412515145, |
|
"grad_norm": 89.59203114100742, |
|
"learning_rate": 1.3104635282944052e-07, |
|
"logits/chosen": -1.4790459871292114, |
|
"logits/rejected": -1.4666205644607544, |
|
"logps/chosen": -161.86483764648438, |
|
"logps/rejected": -211.35423278808594, |
|
"loss": 0.6404, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": 0.004301717504858971, |
|
"rewards/margins": 0.35839876532554626, |
|
"rewards/rejected": -0.35409703850746155, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.868933076758606, |
|
"grad_norm": 94.6581914253334, |
|
"learning_rate": 1.3068458534979812e-07, |
|
"logits/chosen": -1.3811688423156738, |
|
"logits/rejected": -1.4190725088119507, |
|
"logps/chosen": -80.00721740722656, |
|
"logps/rejected": -101.73347473144531, |
|
"loss": 0.6688, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.02035554125905037, |
|
"rewards/margins": 0.07091270387172699, |
|
"rewards/rejected": -0.09126824140548706, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 0.8712137410020668, |
|
"grad_norm": 84.05822092329534, |
|
"learning_rate": 1.3032237399648357e-07, |
|
"logits/chosen": -1.3661377429962158, |
|
"logits/rejected": -1.3815851211547852, |
|
"logps/chosen": -106.82656860351562, |
|
"logps/rejected": -128.75656127929688, |
|
"loss": 0.6309, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.03877873718738556, |
|
"rewards/margins": 0.08257640898227692, |
|
"rewards/rejected": -0.12135513871908188, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 0.8734944052455278, |
|
"grad_norm": 99.58060667623182, |
|
"learning_rate": 1.2995972400913367e-07, |
|
"logits/chosen": -1.4144562482833862, |
|
"logits/rejected": -1.4412004947662354, |
|
"logps/chosen": -136.15481567382812, |
|
"logps/rejected": -141.8118438720703, |
|
"loss": 0.6589, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.0782175287604332, |
|
"rewards/margins": 0.17598725855350494, |
|
"rewards/rejected": -0.25420477986335754, |
|
"step": 766 |
|
}, |
|
{ |
|
"epoch": 0.8757750694889886, |
|
"grad_norm": 80.17753876582776, |
|
"learning_rate": 1.2959664063373042e-07, |
|
"logits/chosen": -1.3012042045593262, |
|
"logits/rejected": -1.2482383251190186, |
|
"logps/chosen": -100.65060424804688, |
|
"logps/rejected": -112.65821838378906, |
|
"loss": 0.5368, |
|
"rewards/accuracies": 0.78125, |
|
"rewards/chosen": 0.04776890203356743, |
|
"rewards/margins": 0.4462125897407532, |
|
"rewards/rejected": -0.39844369888305664, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 0.8780557337324496, |
|
"grad_norm": 88.16104383702894, |
|
"learning_rate": 1.2923312912252506e-07, |
|
"logits/chosen": -1.4704458713531494, |
|
"logits/rejected": -1.5337769985198975, |
|
"logps/chosen": -118.58712768554688, |
|
"logps/rejected": -128.90980529785156, |
|
"loss": 0.6971, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": 0.005523152183741331, |
|
"rewards/margins": 0.031648553907871246, |
|
"rewards/rejected": -0.02612539939582348, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.8803363979759105, |
|
"grad_norm": 88.57684727206627, |
|
"learning_rate": 1.288691947339621e-07, |
|
"logits/chosen": -1.4641995429992676, |
|
"logits/rejected": -1.4404255151748657, |
|
"logps/chosen": -188.07818603515625, |
|
"logps/rejected": -189.4072265625, |
|
"loss": 0.6435, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.11661610007286072, |
|
"rewards/margins": 0.22115452587604523, |
|
"rewards/rejected": -0.33777061104774475, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 0.8826170622193714, |
|
"grad_norm": 88.03404600894119, |
|
"learning_rate": 1.2850484273260325e-07, |
|
"logits/chosen": -1.2959342002868652, |
|
"logits/rejected": -1.3783100843429565, |
|
"logps/chosen": -205.9984130859375, |
|
"logps/rejected": -227.6925506591797, |
|
"loss": 0.6664, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.16467887163162231, |
|
"rewards/margins": 0.33329781889915466, |
|
"rewards/rejected": -0.497976690530777, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 0.8848977264628323, |
|
"grad_norm": 85.2906054415725, |
|
"learning_rate": 1.2814007838905129e-07, |
|
"logits/chosen": -1.3777484893798828, |
|
"logits/rejected": -1.4851555824279785, |
|
"logps/chosen": -169.7268524169922, |
|
"logps/rejected": -183.84872436523438, |
|
"loss": 0.5976, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.050482071936130524, |
|
"rewards/margins": 0.21894827485084534, |
|
"rewards/rejected": -0.26943036913871765, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 0.8871783907062932, |
|
"grad_norm": 113.67510179823725, |
|
"learning_rate": 1.2777490697987375e-07, |
|
"logits/chosen": -1.361803412437439, |
|
"logits/rejected": -1.4689048528671265, |
|
"logps/chosen": -188.6908416748047, |
|
"logps/rejected": -239.35243225097656, |
|
"loss": 0.6593, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": 0.024178601801395416, |
|
"rewards/margins": 0.5260721445083618, |
|
"rewards/rejected": -0.5018935203552246, |
|
"step": 778 |
|
}, |
|
{ |
|
"epoch": 0.8894590549497541, |
|
"grad_norm": 77.93754881822308, |
|
"learning_rate": 1.2740933378752683e-07, |
|
"logits/chosen": -1.4177353382110596, |
|
"logits/rejected": -1.4752013683319092, |
|
"logps/chosen": -196.52670288085938, |
|
"logps/rejected": -216.70066833496094, |
|
"loss": 0.5885, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.18388278782367706, |
|
"rewards/margins": 0.34562817215919495, |
|
"rewards/rejected": -0.5295109748840332, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.891739719193215, |
|
"grad_norm": 85.24959090845118, |
|
"learning_rate": 1.2704336410027862e-07, |
|
"logits/chosen": -1.4634852409362793, |
|
"logits/rejected": -1.5205847024917603, |
|
"logps/chosen": -250.54759216308594, |
|
"logps/rejected": -285.6510009765625, |
|
"loss": 0.6189, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.32170242071151733, |
|
"rewards/margins": 0.21656639873981476, |
|
"rewards/rejected": -0.5382688045501709, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 0.8940203834366759, |
|
"grad_norm": 94.62957856064564, |
|
"learning_rate": 1.2667700321213279e-07, |
|
"logits/chosen": -1.216140627861023, |
|
"logits/rejected": -1.3199231624603271, |
|
"logps/chosen": -163.63653564453125, |
|
"logps/rejected": -182.559814453125, |
|
"loss": 0.6221, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.21417541801929474, |
|
"rewards/margins": 0.07622166723012924, |
|
"rewards/rejected": -0.29039713740348816, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 0.8963010476801369, |
|
"grad_norm": 84.13982267155843, |
|
"learning_rate": 1.263102564227521e-07, |
|
"logits/chosen": -1.3924446105957031, |
|
"logits/rejected": -1.3973028659820557, |
|
"logps/chosen": -164.72738647460938, |
|
"logps/rejected": -185.91358947753906, |
|
"loss": 0.6378, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.017229687422513962, |
|
"rewards/margins": 0.29563748836517334, |
|
"rewards/rejected": -0.3128671646118164, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 0.8985817119235977, |
|
"grad_norm": 92.91613247214089, |
|
"learning_rate": 1.2594312903738162e-07, |
|
"logits/chosen": -1.4232394695281982, |
|
"logits/rejected": -1.5436615943908691, |
|
"logps/chosen": -137.60520935058594, |
|
"logps/rejected": -159.5936737060547, |
|
"loss": 0.6099, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.14493076503276825, |
|
"rewards/margins": 0.23380112648010254, |
|
"rewards/rejected": -0.3787318468093872, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 0.9008623761670587, |
|
"grad_norm": 109.24475281169839, |
|
"learning_rate": 1.2557562636677192e-07, |
|
"logits/chosen": -1.3665200471878052, |
|
"logits/rejected": -1.4350488185882568, |
|
"logps/chosen": -191.39637756347656, |
|
"logps/rejected": -255.76321411132812, |
|
"loss": 0.6655, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.16933086514472961, |
|
"rewards/margins": 0.5223888158798218, |
|
"rewards/rejected": -0.691719651222229, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.9031430404105195, |
|
"grad_norm": 66.209268910187, |
|
"learning_rate": 1.252077537271024e-07, |
|
"logits/chosen": -1.318701982498169, |
|
"logits/rejected": -1.3234184980392456, |
|
"logps/chosen": -180.39767456054688, |
|
"logps/rejected": -182.42808532714844, |
|
"loss": 0.5986, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -0.012874465435743332, |
|
"rewards/margins": 0.4246137738227844, |
|
"rewards/rejected": -0.43748825788497925, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 0.9054237046539805, |
|
"grad_norm": 110.64800895876068, |
|
"learning_rate": 1.2483951643990425e-07, |
|
"logits/chosen": -1.427907943725586, |
|
"logits/rejected": -1.4352999925613403, |
|
"logps/chosen": -234.02430725097656, |
|
"logps/rejected": -252.70291137695312, |
|
"loss": 0.6882, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.23474670946598053, |
|
"rewards/margins": 0.3372371196746826, |
|
"rewards/rejected": -0.571983814239502, |
|
"step": 794 |
|
}, |
|
{ |
|
"epoch": 0.9077043688974413, |
|
"grad_norm": 71.22024881783034, |
|
"learning_rate": 1.2447091983198366e-07, |
|
"logits/chosen": -1.4595139026641846, |
|
"logits/rejected": -1.494476556777954, |
|
"logps/chosen": -177.576904296875, |
|
"logps/rejected": -200.62152099609375, |
|
"loss": 0.622, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.10567592829465866, |
|
"rewards/margins": 0.4641418755054474, |
|
"rewards/rejected": -0.5698177814483643, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 0.9099850331409023, |
|
"grad_norm": 77.66730114746076, |
|
"learning_rate": 1.2410196923534453e-07, |
|
"logits/chosen": -1.3751342296600342, |
|
"logits/rejected": -1.4557826519012451, |
|
"logps/chosen": -156.71981811523438, |
|
"logps/rejected": -175.78079223632812, |
|
"loss": 0.6241, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.17370270192623138, |
|
"rewards/margins": 0.5142669081687927, |
|
"rewards/rejected": -0.6879696249961853, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 0.9122656973843632, |
|
"grad_norm": 79.01703720689792, |
|
"learning_rate": 1.237326699871115e-07, |
|
"logits/chosen": -1.3648015260696411, |
|
"logits/rejected": -1.4242901802062988, |
|
"logps/chosen": -139.64234924316406, |
|
"logps/rejected": -162.11544799804688, |
|
"loss": 0.6048, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.2178276777267456, |
|
"rewards/margins": 0.2655281722545624, |
|
"rewards/rejected": -0.48335587978363037, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.9122656973843632, |
|
"eval_logits/chosen": -1.4735329151153564, |
|
"eval_logits/rejected": -1.460737705230713, |
|
"eval_logps/chosen": -128.9962158203125, |
|
"eval_logps/rejected": -133.23785400390625, |
|
"eval_loss": 0.6001502871513367, |
|
"eval_rewards/accuracies": 0.6800000071525574, |
|
"eval_rewards/chosen": 0.09017454832792282, |
|
"eval_rewards/margins": 0.20768418908119202, |
|
"eval_rewards/rejected": -0.11750967055559158, |
|
"eval_runtime": 19.4284, |
|
"eval_samples_per_second": 5.147, |
|
"eval_steps_per_second": 1.287, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.9145463616278241, |
|
"grad_norm": 114.49291402716334, |
|
"learning_rate": 1.2336302742945265e-07, |
|
"logits/chosen": -1.1926456689834595, |
|
"logits/rejected": -1.1694424152374268, |
|
"logps/chosen": -151.49020385742188, |
|
"logps/rejected": -221.17051696777344, |
|
"loss": 0.6541, |
|
"rewards/accuracies": 0.4375, |
|
"rewards/chosen": -0.3034975230693817, |
|
"rewards/margins": 0.10761985182762146, |
|
"rewards/rejected": -0.41111740469932556, |
|
"step": 802 |
|
}, |
|
{ |
|
"epoch": 0.9168270258712851, |
|
"grad_norm": 87.25759889144325, |
|
"learning_rate": 1.2299304690950234e-07, |
|
"logits/chosen": -1.314978837966919, |
|
"logits/rejected": -1.3282618522644043, |
|
"logps/chosen": -130.7468719482422, |
|
"logps/rejected": -169.4188690185547, |
|
"loss": 0.6349, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.07509980350732803, |
|
"rewards/margins": 0.23689687252044678, |
|
"rewards/rejected": -0.3119966685771942, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 0.9191076901147459, |
|
"grad_norm": 90.2755897487053, |
|
"learning_rate": 1.2262273377928374e-07, |
|
"logits/chosen": -1.4546884298324585, |
|
"logits/rejected": -1.5022200345993042, |
|
"logps/chosen": -155.28553771972656, |
|
"logps/rejected": -186.28756713867188, |
|
"loss": 0.6376, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.20369568467140198, |
|
"rewards/margins": 0.15442761778831482, |
|
"rewards/rejected": -0.3581232726573944, |
|
"step": 806 |
|
}, |
|
{ |
|
"epoch": 0.9213883543582069, |
|
"grad_norm": 87.61957852766155, |
|
"learning_rate": 1.2225209339563143e-07, |
|
"logits/chosen": -1.1901319026947021, |
|
"logits/rejected": -1.2732369899749756, |
|
"logps/chosen": -126.2178726196289, |
|
"logps/rejected": -131.4507598876953, |
|
"loss": 0.6119, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -0.06662334501743317, |
|
"rewards/margins": 0.1810164749622345, |
|
"rewards/rejected": -0.24763983488082886, |
|
"step": 808 |
|
}, |
|
{ |
|
"epoch": 0.9236690186016677, |
|
"grad_norm": 123.86797719697702, |
|
"learning_rate": 1.2188113112011404e-07, |
|
"logits/chosen": -1.260641098022461, |
|
"logits/rejected": -1.3186612129211426, |
|
"logps/chosen": -247.8528594970703, |
|
"logps/rejected": -279.6898498535156, |
|
"loss": 0.6643, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.3332889676094055, |
|
"rewards/margins": 0.2615528106689453, |
|
"rewards/rejected": -0.594841718673706, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.9259496828451287, |
|
"grad_norm": 92.05739920896082, |
|
"learning_rate": 1.2150985231895645e-07, |
|
"logits/chosen": -1.5352187156677246, |
|
"logits/rejected": -1.5089805126190186, |
|
"logps/chosen": -168.19351196289062, |
|
"logps/rejected": -155.74728393554688, |
|
"loss": 0.6464, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.05093672126531601, |
|
"rewards/margins": 0.23030051589012146, |
|
"rewards/rejected": -0.2812372148036957, |
|
"step": 812 |
|
}, |
|
{ |
|
"epoch": 0.9282303470885895, |
|
"grad_norm": 93.01772334051867, |
|
"learning_rate": 1.2113826236296244e-07, |
|
"logits/chosen": -1.4838998317718506, |
|
"logits/rejected": -1.4913954734802246, |
|
"logps/chosen": -145.08023071289062, |
|
"logps/rejected": -237.1237030029297, |
|
"loss": 0.63, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.10176398605108261, |
|
"rewards/margins": 0.37156644463539124, |
|
"rewards/rejected": -0.47333040833473206, |
|
"step": 814 |
|
}, |
|
{ |
|
"epoch": 0.9305110113320505, |
|
"grad_norm": 104.77223934596387, |
|
"learning_rate": 1.207663666274367e-07, |
|
"logits/chosen": -1.3763738870620728, |
|
"logits/rejected": -1.3552130460739136, |
|
"logps/chosen": -177.25376892089844, |
|
"logps/rejected": -221.85498046875, |
|
"loss": 0.6857, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.18679024279117584, |
|
"rewards/margins": 0.05987125262618065, |
|
"rewards/rejected": -0.2466614842414856, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 0.9327916755755113, |
|
"grad_norm": 103.0413802693114, |
|
"learning_rate": 1.2039417049210742e-07, |
|
"logits/chosen": -1.3318116664886475, |
|
"logits/rejected": -1.3942524194717407, |
|
"logps/chosen": -142.11920166015625, |
|
"logps/rejected": -182.98092651367188, |
|
"loss": 0.6606, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.08502069115638733, |
|
"rewards/margins": 0.2950421869754791, |
|
"rewards/rejected": -0.38006284832954407, |
|
"step": 818 |
|
}, |
|
{ |
|
"epoch": 0.9350723398189723, |
|
"grad_norm": 85.38211010208585, |
|
"learning_rate": 1.2002167934104814e-07, |
|
"logits/chosen": -1.3738641738891602, |
|
"logits/rejected": -1.3845763206481934, |
|
"logps/chosen": -120.14466857910156, |
|
"logps/rejected": -148.59951782226562, |
|
"loss": 0.6025, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.09485463052988052, |
|
"rewards/margins": 0.2216353416442871, |
|
"rewards/rejected": -0.31648993492126465, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.9373530040624332, |
|
"grad_norm": 78.59698161262021, |
|
"learning_rate": 1.1964889856260001e-07, |
|
"logits/chosen": -1.405772089958191, |
|
"logits/rejected": -1.4993829727172852, |
|
"logps/chosen": -135.0908203125, |
|
"logps/rejected": -158.73516845703125, |
|
"loss": 0.6449, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.057033490389585495, |
|
"rewards/margins": 0.19890281558036804, |
|
"rewards/rejected": -0.14186930656433105, |
|
"step": 822 |
|
}, |
|
{ |
|
"epoch": 0.9396336683058941, |
|
"grad_norm": 77.09101408818802, |
|
"learning_rate": 1.1927583354929392e-07, |
|
"logits/chosen": -1.2669503688812256, |
|
"logits/rejected": -1.3981753587722778, |
|
"logps/chosen": -185.62281799316406, |
|
"logps/rejected": -229.72349548339844, |
|
"loss": 0.5898, |
|
"rewards/accuracies": 0.78125, |
|
"rewards/chosen": -0.05814466252923012, |
|
"rewards/margins": 0.5641728043556213, |
|
"rewards/rejected": -0.6223174333572388, |
|
"step": 824 |
|
}, |
|
{ |
|
"epoch": 0.941914332549355, |
|
"grad_norm": 80.62233528509246, |
|
"learning_rate": 1.1890248969777239e-07, |
|
"logits/chosen": -1.4591572284698486, |
|
"logits/rejected": -1.4886600971221924, |
|
"logps/chosen": -133.4588165283203, |
|
"logps/rejected": -161.96484375, |
|
"loss": 0.6355, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -0.1525450050830841, |
|
"rewards/margins": 0.35945460200309753, |
|
"rewards/rejected": -0.5119996666908264, |
|
"step": 826 |
|
}, |
|
{ |
|
"epoch": 0.9441949967928159, |
|
"grad_norm": 82.91104665388441, |
|
"learning_rate": 1.1852887240871144e-07, |
|
"logits/chosen": -1.3768235445022583, |
|
"logits/rejected": -1.4797062873840332, |
|
"logps/chosen": -168.19825744628906, |
|
"logps/rejected": -195.83607482910156, |
|
"loss": 0.6328, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.021572627127170563, |
|
"rewards/margins": 0.4793768525123596, |
|
"rewards/rejected": -0.45780429244041443, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 0.9464756610362768, |
|
"grad_norm": 99.28432995782116, |
|
"learning_rate": 1.1815498708674265e-07, |
|
"logits/chosen": -1.3641839027404785, |
|
"logits/rejected": -1.5057623386383057, |
|
"logps/chosen": -188.45806884765625, |
|
"logps/rejected": -205.60987854003906, |
|
"loss": 0.6396, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.27040964365005493, |
|
"rewards/margins": 0.16016271710395813, |
|
"rewards/rejected": -0.4305723309516907, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.9487563252797377, |
|
"grad_norm": 91.59735539302851, |
|
"learning_rate": 1.1778083914037489e-07, |
|
"logits/chosen": -1.4061131477355957, |
|
"logits/rejected": -1.425171136856079, |
|
"logps/chosen": -127.95852661132812, |
|
"logps/rejected": -161.05148315429688, |
|
"loss": 0.6552, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.016702737659215927, |
|
"rewards/margins": 0.2845785915851593, |
|
"rewards/rejected": -0.2678758502006531, |
|
"step": 832 |
|
}, |
|
{ |
|
"epoch": 0.9510369895231986, |
|
"grad_norm": 86.74397487981402, |
|
"learning_rate": 1.17406433981916e-07, |
|
"logits/chosen": -1.3160350322723389, |
|
"logits/rejected": -1.3871701955795288, |
|
"logps/chosen": -176.73638916015625, |
|
"logps/rejected": -211.8169403076172, |
|
"loss": 0.5804, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.43320897221565247, |
|
"rewards/margins": 0.319485068321228, |
|
"rewards/rejected": -0.7526940107345581, |
|
"step": 834 |
|
}, |
|
{ |
|
"epoch": 0.9533176537666596, |
|
"grad_norm": 95.1172952955302, |
|
"learning_rate": 1.1703177702739459e-07, |
|
"logits/chosen": -1.4391237497329712, |
|
"logits/rejected": -1.4496339559555054, |
|
"logps/chosen": -170.07687377929688, |
|
"logps/rejected": -187.2700958251953, |
|
"loss": 0.6198, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.08456218987703323, |
|
"rewards/margins": 0.2940000295639038, |
|
"rewards/rejected": -0.37856221199035645, |
|
"step": 836 |
|
}, |
|
{ |
|
"epoch": 0.9555983180101204, |
|
"grad_norm": 88.76345502432684, |
|
"learning_rate": 1.1665687369648172e-07, |
|
"logits/chosen": -1.323454737663269, |
|
"logits/rejected": -1.330315351486206, |
|
"logps/chosen": -124.01346588134766, |
|
"logps/rejected": -141.37887573242188, |
|
"loss": 0.6466, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.2537275552749634, |
|
"rewards/margins": -0.010971667245030403, |
|
"rewards/rejected": -0.24275588989257812, |
|
"step": 838 |
|
}, |
|
{ |
|
"epoch": 0.9578789822535814, |
|
"grad_norm": 88.84157969865495, |
|
"learning_rate": 1.1628172941241239e-07, |
|
"logits/chosen": -1.2515794038772583, |
|
"logits/rejected": -1.2858678102493286, |
|
"logps/chosen": -161.89578247070312, |
|
"logps/rejected": -192.5193328857422, |
|
"loss": 0.5897, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.062142424285411835, |
|
"rewards/margins": 0.38304537534713745, |
|
"rewards/rejected": -0.44518783688545227, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.9601596464970422, |
|
"grad_norm": 77.39513907682372, |
|
"learning_rate": 1.159063496019072e-07, |
|
"logits/chosen": -1.3906288146972656, |
|
"logits/rejected": -1.529634952545166, |
|
"logps/chosen": -149.05487060546875, |
|
"logps/rejected": -175.32589721679688, |
|
"loss": 0.607, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": -0.14833964407444, |
|
"rewards/margins": 0.27842655777931213, |
|
"rewards/rejected": -0.42676615715026855, |
|
"step": 842 |
|
}, |
|
{ |
|
"epoch": 0.9624403107405032, |
|
"grad_norm": 95.00453640521121, |
|
"learning_rate": 1.1553073969509382e-07, |
|
"logits/chosen": -1.3770264387130737, |
|
"logits/rejected": -1.4574774503707886, |
|
"logps/chosen": -154.23617553710938, |
|
"logps/rejected": -165.87744140625, |
|
"loss": 0.6221, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.23750467598438263, |
|
"rewards/margins": 0.23638057708740234, |
|
"rewards/rejected": -0.47388529777526855, |
|
"step": 844 |
|
}, |
|
{ |
|
"epoch": 0.964720974983964, |
|
"grad_norm": 106.04741673208737, |
|
"learning_rate": 1.1515490512542831e-07, |
|
"logits/chosen": -1.467265248298645, |
|
"logits/rejected": -1.4767037630081177, |
|
"logps/chosen": -154.89556884765625, |
|
"logps/rejected": -169.7006072998047, |
|
"loss": 0.6862, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.292231023311615, |
|
"rewards/margins": 0.32214629650115967, |
|
"rewards/rejected": -0.6143773198127747, |
|
"step": 846 |
|
}, |
|
{ |
|
"epoch": 0.967001639227425, |
|
"grad_norm": 105.71509066302269, |
|
"learning_rate": 1.1477885132961677e-07, |
|
"logits/chosen": -1.3954505920410156, |
|
"logits/rejected": -1.4496076107025146, |
|
"logps/chosen": -135.10621643066406, |
|
"logps/rejected": -155.632080078125, |
|
"loss": 0.6695, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -0.11768642067909241, |
|
"rewards/margins": 0.2924567759037018, |
|
"rewards/rejected": -0.4101432263851166, |
|
"step": 848 |
|
}, |
|
{ |
|
"epoch": 0.9692823034708858, |
|
"grad_norm": 83.14192239326704, |
|
"learning_rate": 1.1440258374753649e-07, |
|
"logits/chosen": -1.2167763710021973, |
|
"logits/rejected": -1.2719136476516724, |
|
"logps/chosen": -161.74093627929688, |
|
"logps/rejected": -186.17892456054688, |
|
"loss": 0.6215, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.2962597608566284, |
|
"rewards/margins": 0.19979730248451233, |
|
"rewards/rejected": -0.49605703353881836, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.9715629677143468, |
|
"grad_norm": 91.00056492305369, |
|
"learning_rate": 1.1402610782215733e-07, |
|
"logits/chosen": -1.338948369026184, |
|
"logits/rejected": -1.4312294721603394, |
|
"logps/chosen": -184.9549560546875, |
|
"logps/rejected": -219.67489624023438, |
|
"loss": 0.6323, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.3477857708930969, |
|
"rewards/margins": 0.11345378309488297, |
|
"rewards/rejected": -0.46123960614204407, |
|
"step": 852 |
|
}, |
|
{ |
|
"epoch": 0.9738436319578078, |
|
"grad_norm": 83.43376640938686, |
|
"learning_rate": 1.1364942899946299e-07, |
|
"logits/chosen": -1.4227657318115234, |
|
"logits/rejected": -1.4756063222885132, |
|
"logps/chosen": -176.06297302246094, |
|
"logps/rejected": -190.72410583496094, |
|
"loss": 0.6332, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.2725405991077423, |
|
"rewards/margins": 0.23061612248420715, |
|
"rewards/rejected": -0.5031567215919495, |
|
"step": 854 |
|
}, |
|
{ |
|
"epoch": 0.9761242962012686, |
|
"grad_norm": 80.98108181650579, |
|
"learning_rate": 1.132725527283722e-07, |
|
"logits/chosen": -1.390743613243103, |
|
"logits/rejected": -1.4082766771316528, |
|
"logps/chosen": -148.54055786132812, |
|
"logps/rejected": -182.93699645996094, |
|
"loss": 0.6133, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.07731722295284271, |
|
"rewards/margins": 0.25885114073753357, |
|
"rewards/rejected": -0.3361683785915375, |
|
"step": 856 |
|
}, |
|
{ |
|
"epoch": 0.9784049604447296, |
|
"grad_norm": 84.02405944942352, |
|
"learning_rate": 1.1289548446065993e-07, |
|
"logits/chosen": -1.3017364740371704, |
|
"logits/rejected": -1.3097370862960815, |
|
"logps/chosen": -162.28753662109375, |
|
"logps/rejected": -187.27439880371094, |
|
"loss": 0.6506, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.13367247581481934, |
|
"rewards/margins": 0.2926619052886963, |
|
"rewards/rejected": -0.426334410905838, |
|
"step": 858 |
|
}, |
|
{ |
|
"epoch": 0.9806856246881904, |
|
"grad_norm": 92.76881750126093, |
|
"learning_rate": 1.1251822965087854e-07, |
|
"logits/chosen": -1.2643218040466309, |
|
"logits/rejected": -1.4101805686950684, |
|
"logps/chosen": -162.29969787597656, |
|
"logps/rejected": -204.03713989257812, |
|
"loss": 0.6025, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.28711238503456116, |
|
"rewards/margins": 0.36070847511291504, |
|
"rewards/rejected": -0.6478208899497986, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.9829662889316514, |
|
"grad_norm": 118.85388390167415, |
|
"learning_rate": 1.1214079375627883e-07, |
|
"logits/chosen": -1.4487006664276123, |
|
"logits/rejected": -1.474176049232483, |
|
"logps/chosen": -271.73126220703125, |
|
"logps/rejected": -296.800537109375, |
|
"loss": 0.6742, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.5793529152870178, |
|
"rewards/margins": 0.24643933773040771, |
|
"rewards/rejected": -0.8257922530174255, |
|
"step": 862 |
|
}, |
|
{ |
|
"epoch": 0.9852469531751122, |
|
"grad_norm": 86.67252914799577, |
|
"learning_rate": 1.1176318223673105e-07, |
|
"logits/chosen": -1.4337862730026245, |
|
"logits/rejected": -1.4527971744537354, |
|
"logps/chosen": -200.96485900878906, |
|
"logps/rejected": -221.7783660888672, |
|
"loss": 0.6498, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.2869877815246582, |
|
"rewards/margins": 0.2341936230659485, |
|
"rewards/rejected": -0.5211814045906067, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 0.9875276174185732, |
|
"grad_norm": 113.98036843821767, |
|
"learning_rate": 1.1138540055464609e-07, |
|
"logits/chosen": -1.3671294450759888, |
|
"logits/rejected": -1.368064522743225, |
|
"logps/chosen": -163.43055725097656, |
|
"logps/rejected": -175.2020721435547, |
|
"loss": 0.5793, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.022854965180158615, |
|
"rewards/margins": 0.4233308434486389, |
|
"rewards/rejected": -0.44618579745292664, |
|
"step": 866 |
|
}, |
|
{ |
|
"epoch": 0.989808281662034, |
|
"grad_norm": 80.29058110520482, |
|
"learning_rate": 1.110074541748963e-07, |
|
"logits/chosen": -1.346540093421936, |
|
"logits/rejected": -1.3472204208374023, |
|
"logps/chosen": -186.08084106445312, |
|
"logps/rejected": -207.60470581054688, |
|
"loss": 0.6321, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.25492382049560547, |
|
"rewards/margins": 0.35832715034484863, |
|
"rewards/rejected": -0.6132509708404541, |
|
"step": 868 |
|
}, |
|
{ |
|
"epoch": 0.992088945905495, |
|
"grad_norm": 90.95115779695014, |
|
"learning_rate": 1.1062934856473653e-07, |
|
"logits/chosen": -1.4387091398239136, |
|
"logits/rejected": -1.455345630645752, |
|
"logps/chosen": -98.49878692626953, |
|
"logps/rejected": -172.24810791015625, |
|
"loss": 0.6444, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.03821319341659546, |
|
"rewards/margins": 0.8005204796791077, |
|
"rewards/rejected": -0.8387336730957031, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.9943696101489559, |
|
"grad_norm": 100.52505053286734, |
|
"learning_rate": 1.1025108919372499e-07, |
|
"logits/chosen": -1.3714473247528076, |
|
"logits/rejected": -1.3371658325195312, |
|
"logps/chosen": -167.58447265625, |
|
"logps/rejected": -186.04751586914062, |
|
"loss": 0.6675, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.16858822107315063, |
|
"rewards/margins": 0.19196297228336334, |
|
"rewards/rejected": -0.36055123805999756, |
|
"step": 872 |
|
}, |
|
{ |
|
"epoch": 0.9966502743924168, |
|
"grad_norm": 110.94600706114623, |
|
"learning_rate": 1.0987268153364411e-07, |
|
"logits/chosen": -1.3899271488189697, |
|
"logits/rejected": -1.3658580780029297, |
|
"logps/chosen": -130.26220703125, |
|
"logps/rejected": -134.71063232421875, |
|
"loss": 0.66, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.10100418329238892, |
|
"rewards/margins": 0.2500309348106384, |
|
"rewards/rejected": -0.35103511810302734, |
|
"step": 874 |
|
}, |
|
{ |
|
"epoch": 0.9989309386358777, |
|
"grad_norm": 105.88720383741304, |
|
"learning_rate": 1.0949413105842146e-07, |
|
"logits/chosen": -1.4607656002044678, |
|
"logits/rejected": -1.5963387489318848, |
|
"logps/chosen": -196.40066528320312, |
|
"logps/rejected": -225.38638305664062, |
|
"loss": 0.6348, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.15517286956310272, |
|
"rewards/margins": 0.15176478028297424, |
|
"rewards/rejected": -0.30693763494491577, |
|
"step": 876 |
|
}, |
|
{ |
|
"epoch": 1.0012116028793385, |
|
"grad_norm": 56.3694410581267, |
|
"learning_rate": 1.091154432440506e-07, |
|
"logits/chosen": -1.389705777168274, |
|
"logits/rejected": -1.4425851106643677, |
|
"logps/chosen": -184.08709716796875, |
|
"logps/rejected": -215.4456024169922, |
|
"loss": 0.5323, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": 0.004449841566383839, |
|
"rewards/margins": 0.635086178779602, |
|
"rewards/rejected": -0.6306363344192505, |
|
"step": 878 |
|
}, |
|
{ |
|
"epoch": 1.0034922671227995, |
|
"grad_norm": 59.14871364307595, |
|
"learning_rate": 1.0873662356851164e-07, |
|
"logits/chosen": -1.2264068126678467, |
|
"logits/rejected": -1.2039250135421753, |
|
"logps/chosen": -143.2238311767578, |
|
"logps/rejected": -148.1532745361328, |
|
"loss": 0.4487, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": 0.14279066026210785, |
|
"rewards/margins": 0.6712049841880798, |
|
"rewards/rejected": -0.528414249420166, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 1.0057729313662604, |
|
"grad_norm": 56.08399590075636, |
|
"learning_rate": 1.0835767751169225e-07, |
|
"logits/chosen": -1.3238468170166016, |
|
"logits/rejected": -1.4208531379699707, |
|
"logps/chosen": -172.80247497558594, |
|
"logps/rejected": -221.4132843017578, |
|
"loss": 0.4898, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.04028752073645592, |
|
"rewards/margins": 0.7049790024757385, |
|
"rewards/rejected": -0.7452664375305176, |
|
"step": 882 |
|
}, |
|
{ |
|
"epoch": 1.0080535956097214, |
|
"grad_norm": 52.855939708340124, |
|
"learning_rate": 1.0797861055530831e-07, |
|
"logits/chosen": -1.3633054494857788, |
|
"logits/rejected": -1.3413807153701782, |
|
"logps/chosen": -146.97352600097656, |
|
"logps/rejected": -183.86334228515625, |
|
"loss": 0.4712, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": -0.06403888016939163, |
|
"rewards/margins": 0.8986248970031738, |
|
"rewards/rejected": -0.9626636505126953, |
|
"step": 884 |
|
}, |
|
{ |
|
"epoch": 1.0103342598531821, |
|
"grad_norm": 66.20782299721449, |
|
"learning_rate": 1.0759942818282453e-07, |
|
"logits/chosen": -1.2454776763916016, |
|
"logits/rejected": -1.3409074544906616, |
|
"logps/chosen": -165.26669311523438, |
|
"logps/rejected": -199.02630615234375, |
|
"loss": 0.4379, |
|
"rewards/accuracies": 0.90625, |
|
"rewards/chosen": -0.18418747186660767, |
|
"rewards/margins": 0.8542241454124451, |
|
"rewards/rejected": -1.0384116172790527, |
|
"step": 886 |
|
}, |
|
{ |
|
"epoch": 1.012614924096643, |
|
"grad_norm": 61.538116919835716, |
|
"learning_rate": 1.0722013587937526e-07, |
|
"logits/chosen": -1.345977783203125, |
|
"logits/rejected": -1.3574477434158325, |
|
"logps/chosen": -140.93460083007812, |
|
"logps/rejected": -170.11268615722656, |
|
"loss": 0.4964, |
|
"rewards/accuracies": 0.8125, |
|
"rewards/chosen": 0.001187225803732872, |
|
"rewards/margins": 0.6773791313171387, |
|
"rewards/rejected": -0.6761919260025024, |
|
"step": 888 |
|
}, |
|
{ |
|
"epoch": 1.014895588340104, |
|
"grad_norm": 66.18054031071868, |
|
"learning_rate": 1.0684073913168501e-07, |
|
"logits/chosen": -1.506268858909607, |
|
"logits/rejected": -1.5550968647003174, |
|
"logps/chosen": -227.16659545898438, |
|
"logps/rejected": -231.74356079101562, |
|
"loss": 0.4649, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": -0.04133976250886917, |
|
"rewards/margins": 0.7655045986175537, |
|
"rewards/rejected": -0.8068442940711975, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 1.017176252583565, |
|
"grad_norm": 63.53563886574213, |
|
"learning_rate": 1.0646124342798919e-07, |
|
"logits/chosen": -1.2807819843292236, |
|
"logits/rejected": -1.3626489639282227, |
|
"logps/chosen": -149.43206787109375, |
|
"logps/rejected": -194.97207641601562, |
|
"loss": 0.4679, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": 0.026193831115961075, |
|
"rewards/margins": 0.7646716237068176, |
|
"rewards/rejected": -0.7384778261184692, |
|
"step": 892 |
|
}, |
|
{ |
|
"epoch": 1.0194569168270258, |
|
"grad_norm": 63.988279055249954, |
|
"learning_rate": 1.0608165425795468e-07, |
|
"logits/chosen": -1.5072122812271118, |
|
"logits/rejected": -1.5509088039398193, |
|
"logps/chosen": -185.84625244140625, |
|
"logps/rejected": -213.9545440673828, |
|
"loss": 0.5205, |
|
"rewards/accuracies": 0.84375, |
|
"rewards/chosen": -0.03688034415245056, |
|
"rewards/margins": 0.8168615102767944, |
|
"rewards/rejected": -0.8537418842315674, |
|
"step": 894 |
|
}, |
|
{ |
|
"epoch": 1.0217375810704867, |
|
"grad_norm": 58.69101964281229, |
|
"learning_rate": 1.0570197711260038e-07, |
|
"logits/chosen": -1.388396978378296, |
|
"logits/rejected": -1.3820849657058716, |
|
"logps/chosen": -127.9742431640625, |
|
"logps/rejected": -137.78054809570312, |
|
"loss": 0.4494, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": 0.10387776046991348, |
|
"rewards/margins": 0.5781891942024231, |
|
"rewards/rejected": -0.47431135177612305, |
|
"step": 896 |
|
}, |
|
{ |
|
"epoch": 1.0240182453139477, |
|
"grad_norm": 74.70102903392454, |
|
"learning_rate": 1.0532221748421786e-07, |
|
"logits/chosen": -1.3617676496505737, |
|
"logits/rejected": -1.4110198020935059, |
|
"logps/chosen": -160.6496124267578, |
|
"logps/rejected": -182.05860900878906, |
|
"loss": 0.4881, |
|
"rewards/accuracies": 0.78125, |
|
"rewards/chosen": -0.21052153408527374, |
|
"rewards/margins": 0.5095985531806946, |
|
"rewards/rejected": -0.7201201319694519, |
|
"step": 898 |
|
}, |
|
{ |
|
"epoch": 1.0262989095574087, |
|
"grad_norm": 61.26905190831476, |
|
"learning_rate": 1.0494238086629183e-07, |
|
"logits/chosen": -1.4342904090881348, |
|
"logits/rejected": -1.4664863348007202, |
|
"logps/chosen": -173.26043701171875, |
|
"logps/rejected": -202.07130432128906, |
|
"loss": 0.4934, |
|
"rewards/accuracies": 0.78125, |
|
"rewards/chosen": -0.21440905332565308, |
|
"rewards/margins": 0.7449226379394531, |
|
"rewards/rejected": -0.9593316316604614, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.0262989095574087, |
|
"eval_logits/chosen": -1.4491240978240967, |
|
"eval_logits/rejected": -1.4348580837249756, |
|
"eval_logps/chosen": -129.59962463378906, |
|
"eval_logps/rejected": -134.80783081054688, |
|
"eval_loss": 0.5797997713088989, |
|
"eval_rewards/accuracies": 0.7200000286102295, |
|
"eval_rewards/chosen": 0.029834765940904617, |
|
"eval_rewards/margins": 0.3043439984321594, |
|
"eval_rewards/rejected": -0.2745092511177063, |
|
"eval_runtime": 20.081, |
|
"eval_samples_per_second": 4.98, |
|
"eval_steps_per_second": 1.245, |
|
"step": 900 |
|
} |
|
], |
|
"logging_steps": 2, |
|
"max_steps": 1752, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 300, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 0.0, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|