|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.48685491723466406, |
|
"eval_steps": 500, |
|
"global_step": 250, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0019474196689386564, |
|
"grad_norm": 0.6627452969551086, |
|
"learning_rate": 2.5e-05, |
|
"loss": 1.8562, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0038948393378773127, |
|
"grad_norm": 0.9325484037399292, |
|
"learning_rate": 5e-05, |
|
"loss": 2.0532, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.005842259006815969, |
|
"grad_norm": 0.8353978991508484, |
|
"learning_rate": 7.500000000000001e-05, |
|
"loss": 1.8732, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.007789678675754625, |
|
"grad_norm": 0.6053460836410522, |
|
"learning_rate": 0.0001, |
|
"loss": 1.942, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.009737098344693282, |
|
"grad_norm": 0.5565560460090637, |
|
"learning_rate": 0.000125, |
|
"loss": 1.6511, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.011684518013631937, |
|
"grad_norm": 0.9751281142234802, |
|
"learning_rate": 0.00015000000000000001, |
|
"loss": 1.9445, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.013631937682570594, |
|
"grad_norm": 0.730299711227417, |
|
"learning_rate": 0.000175, |
|
"loss": 1.3815, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.01557935735150925, |
|
"grad_norm": 0.753030002117157, |
|
"learning_rate": 0.0002, |
|
"loss": 1.1519, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.017526777020447908, |
|
"grad_norm": 0.8523620367050171, |
|
"learning_rate": 0.00019917355371900828, |
|
"loss": 1.2913, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.019474196689386564, |
|
"grad_norm": 1.002160906791687, |
|
"learning_rate": 0.00019834710743801655, |
|
"loss": 0.8663, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.021421616358325218, |
|
"grad_norm": 1.114942193031311, |
|
"learning_rate": 0.00019752066115702482, |
|
"loss": 0.7233, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.023369036027263874, |
|
"grad_norm": 1.1557896137237549, |
|
"learning_rate": 0.0001966942148760331, |
|
"loss": 0.9432, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.02531645569620253, |
|
"grad_norm": 2.1235923767089844, |
|
"learning_rate": 0.00019586776859504133, |
|
"loss": 0.6598, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.027263875365141188, |
|
"grad_norm": 1.2073708772659302, |
|
"learning_rate": 0.0001950413223140496, |
|
"loss": 1.0136, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.029211295034079845, |
|
"grad_norm": 1.1807072162628174, |
|
"learning_rate": 0.00019421487603305787, |
|
"loss": 1.0244, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.0311587147030185, |
|
"grad_norm": 0.9304102659225464, |
|
"learning_rate": 0.0001933884297520661, |
|
"loss": 0.5349, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.033106134371957155, |
|
"grad_norm": 0.7814887762069702, |
|
"learning_rate": 0.00019256198347107438, |
|
"loss": 0.5058, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.035053554040895815, |
|
"grad_norm": 0.8314672708511353, |
|
"learning_rate": 0.00019173553719008265, |
|
"loss": 0.3814, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.03700097370983447, |
|
"grad_norm": 0.916557788848877, |
|
"learning_rate": 0.00019090909090909092, |
|
"loss": 0.4953, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.03894839337877313, |
|
"grad_norm": 0.9891867637634277, |
|
"learning_rate": 0.0001900826446280992, |
|
"loss": 0.932, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.04089581304771178, |
|
"grad_norm": 0.9992872476577759, |
|
"learning_rate": 0.00018925619834710743, |
|
"loss": 0.8751, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.042843232716650435, |
|
"grad_norm": 0.5688315033912659, |
|
"learning_rate": 0.0001884297520661157, |
|
"loss": 0.3556, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.044790652385589096, |
|
"grad_norm": 0.5576704144477844, |
|
"learning_rate": 0.00018760330578512397, |
|
"loss": 0.2855, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.04673807205452775, |
|
"grad_norm": 0.8203558921813965, |
|
"learning_rate": 0.00018677685950413224, |
|
"loss": 0.9009, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.04868549172346641, |
|
"grad_norm": 0.7509244680404663, |
|
"learning_rate": 0.0001859504132231405, |
|
"loss": 0.5587, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.05063291139240506, |
|
"grad_norm": 0.9155460596084595, |
|
"learning_rate": 0.00018512396694214878, |
|
"loss": 0.7122, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.05258033106134372, |
|
"grad_norm": 0.5604537129402161, |
|
"learning_rate": 0.00018429752066115705, |
|
"loss": 0.4308, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.054527750730282376, |
|
"grad_norm": 0.902310311794281, |
|
"learning_rate": 0.00018347107438016532, |
|
"loss": 0.7039, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.05647517039922103, |
|
"grad_norm": 0.6863579154014587, |
|
"learning_rate": 0.00018264462809917356, |
|
"loss": 0.7063, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.05842259006815969, |
|
"grad_norm": 0.6778506636619568, |
|
"learning_rate": 0.00018181818181818183, |
|
"loss": 0.6357, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.06037000973709834, |
|
"grad_norm": 0.5920571684837341, |
|
"learning_rate": 0.00018099173553719008, |
|
"loss": 0.4388, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.062317429406037, |
|
"grad_norm": 0.5040196776390076, |
|
"learning_rate": 0.00018016528925619835, |
|
"loss": 0.4409, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.06426484907497566, |
|
"grad_norm": 0.5752211809158325, |
|
"learning_rate": 0.00017933884297520662, |
|
"loss": 0.4516, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.06621226874391431, |
|
"grad_norm": 0.7298389673233032, |
|
"learning_rate": 0.00017851239669421489, |
|
"loss": 0.571, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.06815968841285297, |
|
"grad_norm": 0.7012993693351746, |
|
"learning_rate": 0.00017768595041322316, |
|
"loss": 0.4716, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.07010710808179163, |
|
"grad_norm": 0.7051262855529785, |
|
"learning_rate": 0.00017685950413223143, |
|
"loss": 0.6295, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.07205452775073028, |
|
"grad_norm": 0.7011151909828186, |
|
"learning_rate": 0.00017603305785123967, |
|
"loss": 0.4605, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.07400194741966894, |
|
"grad_norm": 1.0121755599975586, |
|
"learning_rate": 0.00017520661157024794, |
|
"loss": 0.7132, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.0759493670886076, |
|
"grad_norm": 1.274100422859192, |
|
"learning_rate": 0.0001743801652892562, |
|
"loss": 1.0475, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.07789678675754626, |
|
"grad_norm": 1.2915902137756348, |
|
"learning_rate": 0.00017355371900826448, |
|
"loss": 0.9126, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.0798442064264849, |
|
"grad_norm": 1.1296463012695312, |
|
"learning_rate": 0.00017272727272727275, |
|
"loss": 1.0518, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.08179162609542356, |
|
"grad_norm": 0.7432142496109009, |
|
"learning_rate": 0.00017190082644628102, |
|
"loss": 0.5967, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.08373904576436222, |
|
"grad_norm": 0.878311812877655, |
|
"learning_rate": 0.00017107438016528926, |
|
"loss": 0.6643, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.08568646543330087, |
|
"grad_norm": 1.0814746618270874, |
|
"learning_rate": 0.00017024793388429753, |
|
"loss": 0.7934, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.08763388510223953, |
|
"grad_norm": 1.0332353115081787, |
|
"learning_rate": 0.00016942148760330577, |
|
"loss": 0.7566, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.08958130477117819, |
|
"grad_norm": 0.952813446521759, |
|
"learning_rate": 0.00016859504132231404, |
|
"loss": 0.6161, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.09152872444011685, |
|
"grad_norm": 0.9768489003181458, |
|
"learning_rate": 0.0001677685950413223, |
|
"loss": 0.8178, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.0934761441090555, |
|
"grad_norm": 0.7100651860237122, |
|
"learning_rate": 0.00016694214876033058, |
|
"loss": 0.6014, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.09542356377799416, |
|
"grad_norm": 0.906511127948761, |
|
"learning_rate": 0.00016611570247933885, |
|
"loss": 0.952, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.09737098344693282, |
|
"grad_norm": 1.0189696550369263, |
|
"learning_rate": 0.00016528925619834712, |
|
"loss": 0.8181, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.09931840311587146, |
|
"grad_norm": 0.8355782628059387, |
|
"learning_rate": 0.0001644628099173554, |
|
"loss": 0.6442, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.10126582278481013, |
|
"grad_norm": 0.747944176197052, |
|
"learning_rate": 0.00016363636363636366, |
|
"loss": 0.5864, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.10321324245374879, |
|
"grad_norm": 0.6874829530715942, |
|
"learning_rate": 0.0001628099173553719, |
|
"loss": 0.5721, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.10516066212268745, |
|
"grad_norm": 0.8766732811927795, |
|
"learning_rate": 0.00016198347107438017, |
|
"loss": 1.1118, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.10710808179162609, |
|
"grad_norm": 0.503648579120636, |
|
"learning_rate": 0.00016115702479338844, |
|
"loss": 0.4632, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.10905550146056475, |
|
"grad_norm": 0.5597368478775024, |
|
"learning_rate": 0.0001603305785123967, |
|
"loss": 0.6327, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.11100292112950341, |
|
"grad_norm": 0.5331040024757385, |
|
"learning_rate": 0.00015950413223140498, |
|
"loss": 0.4662, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.11295034079844206, |
|
"grad_norm": 0.4501107335090637, |
|
"learning_rate": 0.00015867768595041322, |
|
"loss": 0.4593, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.11489776046738072, |
|
"grad_norm": 0.5544815063476562, |
|
"learning_rate": 0.0001578512396694215, |
|
"loss": 0.6111, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.11684518013631938, |
|
"grad_norm": 0.5640352964401245, |
|
"learning_rate": 0.00015702479338842976, |
|
"loss": 0.5499, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.11879259980525804, |
|
"grad_norm": 0.6318275332450867, |
|
"learning_rate": 0.000156198347107438, |
|
"loss": 0.4993, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.12074001947419669, |
|
"grad_norm": 0.5284496545791626, |
|
"learning_rate": 0.00015537190082644627, |
|
"loss": 0.6777, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.12268743914313535, |
|
"grad_norm": 0.44762396812438965, |
|
"learning_rate": 0.00015454545454545454, |
|
"loss": 0.3336, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.124634858812074, |
|
"grad_norm": 0.7527077794075012, |
|
"learning_rate": 0.00015371900826446281, |
|
"loss": 0.5085, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.12658227848101267, |
|
"grad_norm": 0.5384949445724487, |
|
"learning_rate": 0.00015289256198347108, |
|
"loss": 0.4182, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.12852969814995133, |
|
"grad_norm": 0.6809192895889282, |
|
"learning_rate": 0.00015206611570247935, |
|
"loss": 0.5201, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.13047711781888996, |
|
"grad_norm": 0.6038883328437805, |
|
"learning_rate": 0.00015123966942148762, |
|
"loss": 0.4331, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.13242453748782862, |
|
"grad_norm": 0.489505797624588, |
|
"learning_rate": 0.0001504132231404959, |
|
"loss": 0.3439, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.13437195715676728, |
|
"grad_norm": 0.4495028555393219, |
|
"learning_rate": 0.00014958677685950414, |
|
"loss": 0.3788, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.13631937682570594, |
|
"grad_norm": 0.3689400255680084, |
|
"learning_rate": 0.0001487603305785124, |
|
"loss": 0.299, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.1382667964946446, |
|
"grad_norm": 0.440514475107193, |
|
"learning_rate": 0.00014793388429752067, |
|
"loss": 0.2871, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.14021421616358326, |
|
"grad_norm": 0.6603713035583496, |
|
"learning_rate": 0.00014710743801652894, |
|
"loss": 0.4321, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.14216163583252192, |
|
"grad_norm": 0.353661447763443, |
|
"learning_rate": 0.0001462809917355372, |
|
"loss": 0.2402, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.14410905550146055, |
|
"grad_norm": 0.7928922772407532, |
|
"learning_rate": 0.00014545454545454546, |
|
"loss": 0.8787, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.1460564751703992, |
|
"grad_norm": 0.502176821231842, |
|
"learning_rate": 0.00014462809917355373, |
|
"loss": 0.3636, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.14800389483933787, |
|
"grad_norm": 0.5314487814903259, |
|
"learning_rate": 0.000143801652892562, |
|
"loss": 0.409, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.14995131450827653, |
|
"grad_norm": 0.6997962594032288, |
|
"learning_rate": 0.00014297520661157024, |
|
"loss": 0.5591, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.1518987341772152, |
|
"grad_norm": 0.7340275645256042, |
|
"learning_rate": 0.0001421487603305785, |
|
"loss": 0.6306, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.15384615384615385, |
|
"grad_norm": 0.722694993019104, |
|
"learning_rate": 0.00014132231404958678, |
|
"loss": 0.5656, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.15579357351509251, |
|
"grad_norm": 0.7989778518676758, |
|
"learning_rate": 0.00014049586776859505, |
|
"loss": 0.8676, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.15774099318403115, |
|
"grad_norm": 0.514163613319397, |
|
"learning_rate": 0.00013966942148760332, |
|
"loss": 0.492, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.1596884128529698, |
|
"grad_norm": 0.78263920545578, |
|
"learning_rate": 0.0001388429752066116, |
|
"loss": 0.7545, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.16163583252190847, |
|
"grad_norm": 0.5553173422813416, |
|
"learning_rate": 0.00013801652892561986, |
|
"loss": 0.5447, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.16358325219084713, |
|
"grad_norm": 0.6660541892051697, |
|
"learning_rate": 0.00013719008264462813, |
|
"loss": 0.7586, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.1655306718597858, |
|
"grad_norm": 0.5211347341537476, |
|
"learning_rate": 0.00013636363636363637, |
|
"loss": 0.3946, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.16747809152872445, |
|
"grad_norm": 0.9146146774291992, |
|
"learning_rate": 0.00013553719008264464, |
|
"loss": 0.6544, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.1694255111976631, |
|
"grad_norm": 0.5467679500579834, |
|
"learning_rate": 0.00013471074380165288, |
|
"loss": 0.5245, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.17137293086660174, |
|
"grad_norm": 0.7478577494621277, |
|
"learning_rate": 0.00013388429752066115, |
|
"loss": 0.8132, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.1733203505355404, |
|
"grad_norm": 0.6577962636947632, |
|
"learning_rate": 0.00013305785123966942, |
|
"loss": 0.59, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.17526777020447906, |
|
"grad_norm": 0.775256335735321, |
|
"learning_rate": 0.0001322314049586777, |
|
"loss": 0.5544, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.17721518987341772, |
|
"grad_norm": 0.5346238613128662, |
|
"learning_rate": 0.00013140495867768596, |
|
"loss": 0.4337, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.17916260954235638, |
|
"grad_norm": 0.578333854675293, |
|
"learning_rate": 0.00013057851239669423, |
|
"loss": 0.4799, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.18111002921129504, |
|
"grad_norm": 0.6039044857025146, |
|
"learning_rate": 0.00012975206611570247, |
|
"loss": 0.4282, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.1830574488802337, |
|
"grad_norm": 0.6463829874992371, |
|
"learning_rate": 0.00012892561983471074, |
|
"loss": 0.6641, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.18500486854917234, |
|
"grad_norm": 0.9733583331108093, |
|
"learning_rate": 0.000128099173553719, |
|
"loss": 0.6536, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.186952288218111, |
|
"grad_norm": 0.6739547252655029, |
|
"learning_rate": 0.00012727272727272728, |
|
"loss": 0.6845, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.18889970788704966, |
|
"grad_norm": 0.7450870275497437, |
|
"learning_rate": 0.00012644628099173555, |
|
"loss": 0.7029, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.19084712755598832, |
|
"grad_norm": 0.7437999248504639, |
|
"learning_rate": 0.00012561983471074382, |
|
"loss": 0.6285, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.19279454722492698, |
|
"grad_norm": 0.6908058524131775, |
|
"learning_rate": 0.0001247933884297521, |
|
"loss": 0.6161, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.19474196689386564, |
|
"grad_norm": 0.6222776174545288, |
|
"learning_rate": 0.00012396694214876033, |
|
"loss": 0.5389, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.1966893865628043, |
|
"grad_norm": 0.7415236234664917, |
|
"learning_rate": 0.0001231404958677686, |
|
"loss": 0.7564, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.19863680623174293, |
|
"grad_norm": 0.725433349609375, |
|
"learning_rate": 0.00012231404958677685, |
|
"loss": 0.5624, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.2005842259006816, |
|
"grad_norm": 0.5979681611061096, |
|
"learning_rate": 0.00012148760330578513, |
|
"loss": 0.4344, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.20253164556962025, |
|
"grad_norm": 0.6501683592796326, |
|
"learning_rate": 0.0001206611570247934, |
|
"loss": 0.5137, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.2044790652385589, |
|
"grad_norm": 0.4350631535053253, |
|
"learning_rate": 0.00011983471074380165, |
|
"loss": 0.4055, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.20642648490749757, |
|
"grad_norm": 0.7034055590629578, |
|
"learning_rate": 0.00011900826446280992, |
|
"loss": 0.6413, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.20837390457643623, |
|
"grad_norm": 0.606842041015625, |
|
"learning_rate": 0.0001181818181818182, |
|
"loss": 0.6433, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.2103213242453749, |
|
"grad_norm": 0.6555774807929993, |
|
"learning_rate": 0.00011735537190082646, |
|
"loss": 0.8194, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.21226874391431352, |
|
"grad_norm": 0.6111577749252319, |
|
"learning_rate": 0.0001165289256198347, |
|
"loss": 0.5042, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.21421616358325218, |
|
"grad_norm": 0.6553054451942444, |
|
"learning_rate": 0.00011570247933884298, |
|
"loss": 0.6989, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.21616358325219084, |
|
"grad_norm": 0.4501146972179413, |
|
"learning_rate": 0.00011487603305785125, |
|
"loss": 0.3658, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.2181110029211295, |
|
"grad_norm": 0.44687238335609436, |
|
"learning_rate": 0.0001140495867768595, |
|
"loss": 0.3615, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.22005842259006816, |
|
"grad_norm": 0.7412980198860168, |
|
"learning_rate": 0.00011322314049586777, |
|
"loss": 0.8426, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.22200584225900682, |
|
"grad_norm": 0.4677373170852661, |
|
"learning_rate": 0.00011239669421487604, |
|
"loss": 0.3759, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.22395326192794549, |
|
"grad_norm": 0.5156281590461731, |
|
"learning_rate": 0.00011157024793388431, |
|
"loss": 0.4241, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.22590068159688412, |
|
"grad_norm": 0.655049741268158, |
|
"learning_rate": 0.00011074380165289258, |
|
"loss": 0.4153, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.22784810126582278, |
|
"grad_norm": 0.4297865927219391, |
|
"learning_rate": 0.00010991735537190082, |
|
"loss": 0.3184, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.22979552093476144, |
|
"grad_norm": 0.6195393204689026, |
|
"learning_rate": 0.00010909090909090909, |
|
"loss": 0.7035, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.2317429406037001, |
|
"grad_norm": 0.6508321762084961, |
|
"learning_rate": 0.00010826446280991735, |
|
"loss": 0.6435, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.23369036027263876, |
|
"grad_norm": 0.5999849438667297, |
|
"learning_rate": 0.00010743801652892562, |
|
"loss": 0.5618, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.23563777994157742, |
|
"grad_norm": 0.4445487856864929, |
|
"learning_rate": 0.00010661157024793389, |
|
"loss": 0.3008, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.23758519961051608, |
|
"grad_norm": 0.46502161026000977, |
|
"learning_rate": 0.00010578512396694216, |
|
"loss": 0.4989, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.2395326192794547, |
|
"grad_norm": 0.5087387561798096, |
|
"learning_rate": 0.00010495867768595043, |
|
"loss": 0.3852, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.24148003894839337, |
|
"grad_norm": 0.4807125926017761, |
|
"learning_rate": 0.0001041322314049587, |
|
"loss": 0.3579, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.24342745861733203, |
|
"grad_norm": 0.5348750948905945, |
|
"learning_rate": 0.00010330578512396694, |
|
"loss": 0.4919, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.2453748782862707, |
|
"grad_norm": 0.8607493042945862, |
|
"learning_rate": 0.00010247933884297521, |
|
"loss": 0.8142, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.24732229795520935, |
|
"grad_norm": 0.4636419415473938, |
|
"learning_rate": 0.00010165289256198347, |
|
"loss": 0.407, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.249269717624148, |
|
"grad_norm": 0.49384385347366333, |
|
"learning_rate": 0.00010082644628099174, |
|
"loss": 0.4263, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.25121713729308665, |
|
"grad_norm": 0.5093722343444824, |
|
"learning_rate": 0.0001, |
|
"loss": 0.3402, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.25316455696202533, |
|
"grad_norm": 0.6678729057312012, |
|
"learning_rate": 9.917355371900827e-05, |
|
"loss": 0.5644, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.25511197663096397, |
|
"grad_norm": 0.7193764448165894, |
|
"learning_rate": 9.834710743801654e-05, |
|
"loss": 0.6274, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.25705939629990265, |
|
"grad_norm": 0.3708731234073639, |
|
"learning_rate": 9.75206611570248e-05, |
|
"loss": 0.269, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.2590068159688413, |
|
"grad_norm": 0.6703848242759705, |
|
"learning_rate": 9.669421487603306e-05, |
|
"loss": 0.5984, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.2609542356377799, |
|
"grad_norm": 0.5617924928665161, |
|
"learning_rate": 9.586776859504133e-05, |
|
"loss": 0.5166, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.2629016553067186, |
|
"grad_norm": 0.6645628213882446, |
|
"learning_rate": 9.50413223140496e-05, |
|
"loss": 0.7676, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.26484907497565724, |
|
"grad_norm": 0.6777731776237488, |
|
"learning_rate": 9.421487603305785e-05, |
|
"loss": 0.4477, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.2667964946445959, |
|
"grad_norm": 1.0802561044692993, |
|
"learning_rate": 9.338842975206612e-05, |
|
"loss": 0.7692, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.26874391431353456, |
|
"grad_norm": 0.5928775072097778, |
|
"learning_rate": 9.256198347107439e-05, |
|
"loss": 0.5721, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.27069133398247325, |
|
"grad_norm": 0.73649001121521, |
|
"learning_rate": 9.173553719008266e-05, |
|
"loss": 0.6324, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.2726387536514119, |
|
"grad_norm": 0.6584858298301697, |
|
"learning_rate": 9.090909090909092e-05, |
|
"loss": 0.5763, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.2745861733203505, |
|
"grad_norm": 0.6551913619041443, |
|
"learning_rate": 9.008264462809917e-05, |
|
"loss": 0.6005, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.2765335929892892, |
|
"grad_norm": 0.5698264837265015, |
|
"learning_rate": 8.925619834710744e-05, |
|
"loss": 0.4164, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.27848101265822783, |
|
"grad_norm": 0.5793282985687256, |
|
"learning_rate": 8.842975206611571e-05, |
|
"loss": 0.4952, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.2804284323271665, |
|
"grad_norm": 0.52625972032547, |
|
"learning_rate": 8.760330578512397e-05, |
|
"loss": 0.4534, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.28237585199610515, |
|
"grad_norm": 0.6003084778785706, |
|
"learning_rate": 8.677685950413224e-05, |
|
"loss": 0.5206, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.28432327166504384, |
|
"grad_norm": 0.5583734512329102, |
|
"learning_rate": 8.595041322314051e-05, |
|
"loss": 0.5323, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.2862706913339825, |
|
"grad_norm": 0.600612461566925, |
|
"learning_rate": 8.512396694214876e-05, |
|
"loss": 0.5463, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.2882181110029211, |
|
"grad_norm": 0.8786905407905579, |
|
"learning_rate": 8.429752066115702e-05, |
|
"loss": 0.48, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.2901655306718598, |
|
"grad_norm": 0.6033667922019958, |
|
"learning_rate": 8.347107438016529e-05, |
|
"loss": 0.4997, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.2921129503407984, |
|
"grad_norm": 0.7072091102600098, |
|
"learning_rate": 8.264462809917356e-05, |
|
"loss": 0.5164, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.2940603700097371, |
|
"grad_norm": 0.5921626687049866, |
|
"learning_rate": 8.181818181818183e-05, |
|
"loss": 0.6378, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.29600778967867575, |
|
"grad_norm": 0.5669450759887695, |
|
"learning_rate": 8.099173553719009e-05, |
|
"loss": 0.501, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.29795520934761444, |
|
"grad_norm": 0.5064852237701416, |
|
"learning_rate": 8.016528925619836e-05, |
|
"loss": 0.3843, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.29990262901655307, |
|
"grad_norm": 0.5336524248123169, |
|
"learning_rate": 7.933884297520661e-05, |
|
"loss": 0.3716, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.3018500486854917, |
|
"grad_norm": 0.44651252031326294, |
|
"learning_rate": 7.851239669421488e-05, |
|
"loss": 0.3773, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.3037974683544304, |
|
"grad_norm": 0.38342341780662537, |
|
"learning_rate": 7.768595041322314e-05, |
|
"loss": 0.2936, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.305744888023369, |
|
"grad_norm": 0.4770042598247528, |
|
"learning_rate": 7.685950413223141e-05, |
|
"loss": 0.4883, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.3076923076923077, |
|
"grad_norm": 0.6555917263031006, |
|
"learning_rate": 7.603305785123968e-05, |
|
"loss": 0.606, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.30963972736124634, |
|
"grad_norm": 0.4627241790294647, |
|
"learning_rate": 7.520661157024795e-05, |
|
"loss": 0.3273, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.31158714703018503, |
|
"grad_norm": 0.3865076005458832, |
|
"learning_rate": 7.43801652892562e-05, |
|
"loss": 0.2885, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.31353456669912366, |
|
"grad_norm": 0.37422531843185425, |
|
"learning_rate": 7.355371900826447e-05, |
|
"loss": 0.3094, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.3154819863680623, |
|
"grad_norm": 0.4276966452598572, |
|
"learning_rate": 7.272727272727273e-05, |
|
"loss": 0.3022, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.317429406037001, |
|
"grad_norm": 0.31993845105171204, |
|
"learning_rate": 7.1900826446281e-05, |
|
"loss": 0.2376, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.3193768257059396, |
|
"grad_norm": 0.3512668013572693, |
|
"learning_rate": 7.107438016528925e-05, |
|
"loss": 0.2865, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.3213242453748783, |
|
"grad_norm": 1.2840955257415771, |
|
"learning_rate": 7.024793388429752e-05, |
|
"loss": 0.9994, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.32327166504381694, |
|
"grad_norm": 0.7181655168533325, |
|
"learning_rate": 6.94214876033058e-05, |
|
"loss": 1.3493, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.3252190847127556, |
|
"grad_norm": 0.5071792602539062, |
|
"learning_rate": 6.859504132231406e-05, |
|
"loss": 0.4438, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.32716650438169426, |
|
"grad_norm": 0.7322396636009216, |
|
"learning_rate": 6.776859504132232e-05, |
|
"loss": 0.6287, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.3291139240506329, |
|
"grad_norm": 0.26013123989105225, |
|
"learning_rate": 6.694214876033058e-05, |
|
"loss": 0.2015, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.3310613437195716, |
|
"grad_norm": 0.41022831201553345, |
|
"learning_rate": 6.611570247933885e-05, |
|
"loss": 0.2814, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.3330087633885102, |
|
"grad_norm": 0.4295414388179779, |
|
"learning_rate": 6.528925619834711e-05, |
|
"loss": 0.3253, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.3349561830574489, |
|
"grad_norm": 0.5953055024147034, |
|
"learning_rate": 6.446280991735537e-05, |
|
"loss": 0.5285, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.33690360272638753, |
|
"grad_norm": 0.585370659828186, |
|
"learning_rate": 6.363636363636364e-05, |
|
"loss": 0.5672, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.3388510223953262, |
|
"grad_norm": 0.469965398311615, |
|
"learning_rate": 6.280991735537191e-05, |
|
"loss": 0.3597, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.34079844206426485, |
|
"grad_norm": 0.5699323415756226, |
|
"learning_rate": 6.198347107438017e-05, |
|
"loss": 0.4781, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.3427458617332035, |
|
"grad_norm": 0.6034254431724548, |
|
"learning_rate": 6.115702479338842e-05, |
|
"loss": 0.468, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.34469328140214217, |
|
"grad_norm": 0.646748423576355, |
|
"learning_rate": 6.03305785123967e-05, |
|
"loss": 0.4904, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.3466407010710808, |
|
"grad_norm": 0.4674849212169647, |
|
"learning_rate": 5.950413223140496e-05, |
|
"loss": 0.4888, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.3485881207400195, |
|
"grad_norm": 0.5764815807342529, |
|
"learning_rate": 5.867768595041323e-05, |
|
"loss": 0.5268, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.3505355404089581, |
|
"grad_norm": 0.5898470878601074, |
|
"learning_rate": 5.785123966942149e-05, |
|
"loss": 0.4667, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.3524829600778968, |
|
"grad_norm": 0.38661783933639526, |
|
"learning_rate": 5.702479338842975e-05, |
|
"loss": 0.3538, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.35443037974683544, |
|
"grad_norm": 0.4027338922023773, |
|
"learning_rate": 5.619834710743802e-05, |
|
"loss": 0.3705, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.3563777994157741, |
|
"grad_norm": 0.3949303925037384, |
|
"learning_rate": 5.537190082644629e-05, |
|
"loss": 0.3463, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.35832521908471276, |
|
"grad_norm": 0.5721768736839294, |
|
"learning_rate": 5.4545454545454546e-05, |
|
"loss": 0.4461, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.3602726387536514, |
|
"grad_norm": 0.5568830966949463, |
|
"learning_rate": 5.371900826446281e-05, |
|
"loss": 0.3703, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.3622200584225901, |
|
"grad_norm": 0.4944753646850586, |
|
"learning_rate": 5.289256198347108e-05, |
|
"loss": 0.4531, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.3641674780915287, |
|
"grad_norm": 0.5557789206504822, |
|
"learning_rate": 5.206611570247935e-05, |
|
"loss": 0.5426, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.3661148977604674, |
|
"grad_norm": 0.6477212309837341, |
|
"learning_rate": 5.1239669421487605e-05, |
|
"loss": 0.5483, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.36806231742940604, |
|
"grad_norm": 0.7089900374412537, |
|
"learning_rate": 5.041322314049587e-05, |
|
"loss": 0.6709, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.37000973709834467, |
|
"grad_norm": 0.5330088138580322, |
|
"learning_rate": 4.958677685950414e-05, |
|
"loss": 0.7063, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.37195715676728336, |
|
"grad_norm": 0.5296370387077332, |
|
"learning_rate": 4.87603305785124e-05, |
|
"loss": 0.5056, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.373904576436222, |
|
"grad_norm": 0.8069382309913635, |
|
"learning_rate": 4.793388429752066e-05, |
|
"loss": 0.7653, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.3758519961051607, |
|
"grad_norm": 0.43892863392829895, |
|
"learning_rate": 4.7107438016528926e-05, |
|
"loss": 0.4423, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.3777994157740993, |
|
"grad_norm": 0.7146729826927185, |
|
"learning_rate": 4.6280991735537196e-05, |
|
"loss": 0.7297, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.379746835443038, |
|
"grad_norm": 0.41250815987586975, |
|
"learning_rate": 4.545454545454546e-05, |
|
"loss": 0.404, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.38169425511197663, |
|
"grad_norm": 0.4819222390651703, |
|
"learning_rate": 4.462809917355372e-05, |
|
"loss": 0.4106, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.38364167478091526, |
|
"grad_norm": 0.41315364837646484, |
|
"learning_rate": 4.3801652892561984e-05, |
|
"loss": 0.3445, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.38558909444985395, |
|
"grad_norm": 0.4681267738342285, |
|
"learning_rate": 4.2975206611570254e-05, |
|
"loss": 0.3611, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.3875365141187926, |
|
"grad_norm": 0.48512405157089233, |
|
"learning_rate": 4.214876033057851e-05, |
|
"loss": 0.3842, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.3894839337877313, |
|
"grad_norm": 0.49099263548851013, |
|
"learning_rate": 4.132231404958678e-05, |
|
"loss": 0.3824, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.3914313534566699, |
|
"grad_norm": 0.3325710892677307, |
|
"learning_rate": 4.049586776859504e-05, |
|
"loss": 0.2954, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.3933787731256086, |
|
"grad_norm": 0.30682113766670227, |
|
"learning_rate": 3.9669421487603306e-05, |
|
"loss": 0.2274, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.3953261927945472, |
|
"grad_norm": 0.3700462579727173, |
|
"learning_rate": 3.884297520661157e-05, |
|
"loss": 0.2983, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.39727361246348586, |
|
"grad_norm": 0.4219472110271454, |
|
"learning_rate": 3.801652892561984e-05, |
|
"loss": 0.3604, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.39922103213242455, |
|
"grad_norm": 0.2869436740875244, |
|
"learning_rate": 3.71900826446281e-05, |
|
"loss": 0.2366, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.4011684518013632, |
|
"grad_norm": 0.5798669457435608, |
|
"learning_rate": 3.6363636363636364e-05, |
|
"loss": 0.6483, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.40311587147030187, |
|
"grad_norm": 0.5161038637161255, |
|
"learning_rate": 3.553719008264463e-05, |
|
"loss": 0.557, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.4050632911392405, |
|
"grad_norm": 0.43804600834846497, |
|
"learning_rate": 3.47107438016529e-05, |
|
"loss": 0.2817, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.4070107108081792, |
|
"grad_norm": 0.2725582420825958, |
|
"learning_rate": 3.388429752066116e-05, |
|
"loss": 0.2317, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.4089581304771178, |
|
"grad_norm": 0.4551626443862915, |
|
"learning_rate": 3.305785123966942e-05, |
|
"loss": 0.3523, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.41090555014605645, |
|
"grad_norm": 0.4810575842857361, |
|
"learning_rate": 3.2231404958677685e-05, |
|
"loss": 0.4526, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.41285296981499514, |
|
"grad_norm": 0.6466739773750305, |
|
"learning_rate": 3.1404958677685955e-05, |
|
"loss": 0.8117, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.4148003894839338, |
|
"grad_norm": 0.5906726717948914, |
|
"learning_rate": 3.057851239669421e-05, |
|
"loss": 0.6446, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.41674780915287246, |
|
"grad_norm": 0.5657168626785278, |
|
"learning_rate": 2.975206611570248e-05, |
|
"loss": 0.4798, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.4186952288218111, |
|
"grad_norm": 0.45190712809562683, |
|
"learning_rate": 2.8925619834710744e-05, |
|
"loss": 0.4177, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.4206426484907498, |
|
"grad_norm": 0.36461302638053894, |
|
"learning_rate": 2.809917355371901e-05, |
|
"loss": 0.3125, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.4225900681596884, |
|
"grad_norm": 0.5353040099143982, |
|
"learning_rate": 2.7272727272727273e-05, |
|
"loss": 0.6517, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.42453748782862705, |
|
"grad_norm": 0.7112963795661926, |
|
"learning_rate": 2.644628099173554e-05, |
|
"loss": 0.6613, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.42648490749756574, |
|
"grad_norm": 0.4022853374481201, |
|
"learning_rate": 2.5619834710743802e-05, |
|
"loss": 0.3589, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.42843232716650437, |
|
"grad_norm": 0.4931185841560364, |
|
"learning_rate": 2.479338842975207e-05, |
|
"loss": 0.5696, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.43037974683544306, |
|
"grad_norm": 0.28436464071273804, |
|
"learning_rate": 2.396694214876033e-05, |
|
"loss": 0.2196, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.4323271665043817, |
|
"grad_norm": 0.4250311553478241, |
|
"learning_rate": 2.3140495867768598e-05, |
|
"loss": 0.3458, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.4342745861733204, |
|
"grad_norm": 0.23114222288131714, |
|
"learning_rate": 2.231404958677686e-05, |
|
"loss": 0.2124, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.436222005842259, |
|
"grad_norm": 0.46653562784194946, |
|
"learning_rate": 2.1487603305785127e-05, |
|
"loss": 0.3793, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.43816942551119764, |
|
"grad_norm": 0.7057761549949646, |
|
"learning_rate": 2.066115702479339e-05, |
|
"loss": 0.7023, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.44011684518013633, |
|
"grad_norm": 0.4195654094219208, |
|
"learning_rate": 1.9834710743801653e-05, |
|
"loss": 0.4148, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.44206426484907496, |
|
"grad_norm": 0.4446503221988678, |
|
"learning_rate": 1.900826446280992e-05, |
|
"loss": 0.4615, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.44401168451801365, |
|
"grad_norm": 0.28336820006370544, |
|
"learning_rate": 1.8181818181818182e-05, |
|
"loss": 0.1496, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.4459591041869523, |
|
"grad_norm": 0.5226246118545532, |
|
"learning_rate": 1.735537190082645e-05, |
|
"loss": 0.5196, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.44790652385589097, |
|
"grad_norm": 0.5695878267288208, |
|
"learning_rate": 1.652892561983471e-05, |
|
"loss": 0.5308, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.4498539435248296, |
|
"grad_norm": 0.5476521849632263, |
|
"learning_rate": 1.5702479338842978e-05, |
|
"loss": 0.4681, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.45180136319376824, |
|
"grad_norm": 0.5449410676956177, |
|
"learning_rate": 1.487603305785124e-05, |
|
"loss": 0.5434, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.4537487828627069, |
|
"grad_norm": 0.47294700145721436, |
|
"learning_rate": 1.4049586776859505e-05, |
|
"loss": 0.435, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.45569620253164556, |
|
"grad_norm": 0.6122575998306274, |
|
"learning_rate": 1.322314049586777e-05, |
|
"loss": 0.5793, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.45764362220058424, |
|
"grad_norm": 0.7652103900909424, |
|
"learning_rate": 1.2396694214876034e-05, |
|
"loss": 0.8786, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.4595910418695229, |
|
"grad_norm": 0.5390903949737549, |
|
"learning_rate": 1.1570247933884299e-05, |
|
"loss": 0.4212, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.46153846153846156, |
|
"grad_norm": 0.6218725442886353, |
|
"learning_rate": 1.0743801652892564e-05, |
|
"loss": 0.5875, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.4634858812074002, |
|
"grad_norm": 0.3942383825778961, |
|
"learning_rate": 9.917355371900826e-06, |
|
"loss": 0.3841, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.46543330087633883, |
|
"grad_norm": 0.34732937812805176, |
|
"learning_rate": 9.090909090909091e-06, |
|
"loss": 0.2677, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.4673807205452775, |
|
"grad_norm": 0.5432242751121521, |
|
"learning_rate": 8.264462809917356e-06, |
|
"loss": 0.5924, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.46932814021421615, |
|
"grad_norm": 0.6005178689956665, |
|
"learning_rate": 7.43801652892562e-06, |
|
"loss": 0.5235, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.47127555988315484, |
|
"grad_norm": 0.6554625630378723, |
|
"learning_rate": 6.611570247933885e-06, |
|
"loss": 0.6429, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.47322297955209347, |
|
"grad_norm": 0.3775584399700165, |
|
"learning_rate": 5.7851239669421495e-06, |
|
"loss": 0.3247, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.47517039922103216, |
|
"grad_norm": 0.5196499824523926, |
|
"learning_rate": 4.958677685950413e-06, |
|
"loss": 0.3589, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.4771178188899708, |
|
"grad_norm": 0.462444931268692, |
|
"learning_rate": 4.132231404958678e-06, |
|
"loss": 0.4217, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.4790652385589094, |
|
"grad_norm": 0.3943294286727905, |
|
"learning_rate": 3.3057851239669424e-06, |
|
"loss": 0.2937, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.4810126582278481, |
|
"grad_norm": 0.3740336298942566, |
|
"learning_rate": 2.4793388429752066e-06, |
|
"loss": 0.3033, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.48296007789678674, |
|
"grad_norm": 0.3156736493110657, |
|
"learning_rate": 1.6528925619834712e-06, |
|
"loss": 0.2708, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.48490749756572543, |
|
"grad_norm": 0.6795382499694824, |
|
"learning_rate": 8.264462809917356e-07, |
|
"loss": 0.5882, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.48685491723466406, |
|
"grad_norm": 0.45327121019363403, |
|
"learning_rate": 0.0, |
|
"loss": 0.3451, |
|
"step": 250 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 250, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 250, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1165175740907520.0, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|