|
{ |
|
"best_metric": 0.4351051665913261, |
|
"best_model_checkpoint": "/m/triton/scratch/elec/puhe/p/palp3/MUCS/indicwav2vec_outputs/pd_warmup_500/s300_shuff100/checkpoint-1000", |
|
"epoch": 1.6, |
|
"eval_steps": 1000, |
|
"global_step": 1000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0016, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0, |
|
"loss": 65.7309, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0032, |
|
"grad_norm": 19.431121826171875, |
|
"learning_rate": 1.2e-06, |
|
"loss": 40.497, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.0048, |
|
"grad_norm": 20.21518898010254, |
|
"learning_rate": 2.4e-06, |
|
"loss": 36.2325, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0064, |
|
"grad_norm": 11.604636192321777, |
|
"learning_rate": 3.6e-06, |
|
"loss": 27.9292, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.008, |
|
"grad_norm": NaN, |
|
"learning_rate": 3.6e-06, |
|
"loss": 27.842, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.0096, |
|
"grad_norm": 13.326777458190918, |
|
"learning_rate": 4.8e-06, |
|
"loss": 28.3298, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.0112, |
|
"grad_norm": 12.479119300842285, |
|
"learning_rate": 5.999999999999999e-06, |
|
"loss": 28.7461, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.0128, |
|
"grad_norm": 11.469672203063965, |
|
"learning_rate": 7.2e-06, |
|
"loss": 26.4075, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.0144, |
|
"grad_norm": 10.416142463684082, |
|
"learning_rate": 8.4e-06, |
|
"loss": 23.0338, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.016, |
|
"grad_norm": 10.866898536682129, |
|
"learning_rate": 9.6e-06, |
|
"loss": 25.844, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0176, |
|
"grad_norm": 11.97667121887207, |
|
"learning_rate": 1.0799999999999998e-05, |
|
"loss": 24.9471, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.0192, |
|
"grad_norm": 10.51934814453125, |
|
"learning_rate": 1.1999999999999999e-05, |
|
"loss": 24.6028, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.0208, |
|
"grad_norm": 10.643861770629883, |
|
"learning_rate": 1.3199999999999997e-05, |
|
"loss": 23.1144, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.0224, |
|
"grad_norm": 11.642038345336914, |
|
"learning_rate": 1.44e-05, |
|
"loss": 23.0041, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.024, |
|
"grad_norm": 10.239847183227539, |
|
"learning_rate": 1.5599999999999996e-05, |
|
"loss": 22.1151, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.0256, |
|
"grad_norm": 10.31641960144043, |
|
"learning_rate": 1.68e-05, |
|
"loss": 20.955, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.0272, |
|
"grad_norm": 10.14737606048584, |
|
"learning_rate": 1.7999999999999997e-05, |
|
"loss": 20.0026, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.0288, |
|
"grad_norm": 10.14389419555664, |
|
"learning_rate": 1.92e-05, |
|
"loss": 20.053, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.0304, |
|
"grad_norm": 13.049298286437988, |
|
"learning_rate": 2.04e-05, |
|
"loss": 23.2289, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.032, |
|
"grad_norm": 10.918747901916504, |
|
"learning_rate": 2.1599999999999996e-05, |
|
"loss": 20.6362, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0336, |
|
"grad_norm": 10.782907485961914, |
|
"learning_rate": 2.28e-05, |
|
"loss": 19.6928, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.0352, |
|
"grad_norm": 13.15986156463623, |
|
"learning_rate": 2.3999999999999997e-05, |
|
"loss": 22.1249, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.0368, |
|
"grad_norm": 11.684220314025879, |
|
"learning_rate": 2.52e-05, |
|
"loss": 20.4745, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.0384, |
|
"grad_norm": 11.818272590637207, |
|
"learning_rate": 2.6399999999999995e-05, |
|
"loss": 20.3504, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 12.81021499633789, |
|
"learning_rate": 2.7599999999999997e-05, |
|
"loss": 21.0093, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.0416, |
|
"grad_norm": 14.015459060668945, |
|
"learning_rate": 2.88e-05, |
|
"loss": 22.1517, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.0432, |
|
"grad_norm": 11.829425811767578, |
|
"learning_rate": 2.9999999999999997e-05, |
|
"loss": 18.2946, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.0448, |
|
"grad_norm": 15.79493236541748, |
|
"learning_rate": 3.119999999999999e-05, |
|
"loss": 22.9019, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.0464, |
|
"grad_norm": 14.349289894104004, |
|
"learning_rate": 3.2399999999999995e-05, |
|
"loss": 20.221, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.048, |
|
"grad_norm": 14.111164093017578, |
|
"learning_rate": 3.36e-05, |
|
"loss": 18.9742, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.0496, |
|
"grad_norm": 15.603620529174805, |
|
"learning_rate": 3.48e-05, |
|
"loss": 20.2483, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.0512, |
|
"grad_norm": 16.44049072265625, |
|
"learning_rate": 3.5999999999999994e-05, |
|
"loss": 19.9449, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.0528, |
|
"grad_norm": 18.68276596069336, |
|
"learning_rate": 3.7199999999999996e-05, |
|
"loss": 21.1125, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.0544, |
|
"grad_norm": 14.831990242004395, |
|
"learning_rate": 3.84e-05, |
|
"loss": 17.0528, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.056, |
|
"grad_norm": 16.635009765625, |
|
"learning_rate": 3.96e-05, |
|
"loss": 18.3324, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.0576, |
|
"grad_norm": 18.090103149414062, |
|
"learning_rate": 4.08e-05, |
|
"loss": 18.495, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.0592, |
|
"grad_norm": 19.209562301635742, |
|
"learning_rate": 4.2e-05, |
|
"loss": 18.7353, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.0608, |
|
"grad_norm": 19.63134765625, |
|
"learning_rate": 4.319999999999999e-05, |
|
"loss": 17.8893, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.0624, |
|
"grad_norm": 20.713550567626953, |
|
"learning_rate": 4.4399999999999995e-05, |
|
"loss": 17.9679, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.064, |
|
"grad_norm": 25.246625900268555, |
|
"learning_rate": 4.56e-05, |
|
"loss": 19.6261, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.0656, |
|
"grad_norm": 28.108665466308594, |
|
"learning_rate": 4.68e-05, |
|
"loss": 20.795, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.0672, |
|
"grad_norm": 26.678646087646484, |
|
"learning_rate": 4.7999999999999994e-05, |
|
"loss": 19.3122, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.0688, |
|
"grad_norm": 28.03876304626465, |
|
"learning_rate": 4.9199999999999997e-05, |
|
"loss": 19.0949, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.0704, |
|
"grad_norm": 28.40298080444336, |
|
"learning_rate": 5.04e-05, |
|
"loss": 18.7253, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.072, |
|
"grad_norm": 30.833574295043945, |
|
"learning_rate": 5.1599999999999994e-05, |
|
"loss": 18.9767, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.0736, |
|
"grad_norm": 25.621904373168945, |
|
"learning_rate": 5.279999999999999e-05, |
|
"loss": 16.0377, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.0752, |
|
"grad_norm": 26.582237243652344, |
|
"learning_rate": 5.399999999999999e-05, |
|
"loss": 15.6131, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.0768, |
|
"grad_norm": 34.89401626586914, |
|
"learning_rate": 5.519999999999999e-05, |
|
"loss": 17.8755, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.0784, |
|
"grad_norm": 32.71629333496094, |
|
"learning_rate": 5.6399999999999995e-05, |
|
"loss": 16.0825, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 27.743824005126953, |
|
"learning_rate": 5.76e-05, |
|
"loss": 13.6992, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.0816, |
|
"grad_norm": Infinity, |
|
"learning_rate": 5.76e-05, |
|
"loss": 30.6838, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.0832, |
|
"grad_norm": 85.50408935546875, |
|
"learning_rate": 5.88e-05, |
|
"loss": 26.4379, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.0848, |
|
"grad_norm": Infinity, |
|
"learning_rate": 5.88e-05, |
|
"loss": 27.0726, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.0864, |
|
"grad_norm": 50.16768264770508, |
|
"learning_rate": 5.9999999999999995e-05, |
|
"loss": 18.3707, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.088, |
|
"grad_norm": 131.34494018554688, |
|
"learning_rate": 6.12e-05, |
|
"loss": 23.1862, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.0896, |
|
"grad_norm": 44.524417877197266, |
|
"learning_rate": 6.239999999999999e-05, |
|
"loss": 14.7586, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.0912, |
|
"grad_norm": 71.3255615234375, |
|
"learning_rate": 6.359999999999999e-05, |
|
"loss": 19.4668, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.0928, |
|
"grad_norm": 43.701072692871094, |
|
"learning_rate": 6.479999999999999e-05, |
|
"loss": 13.1064, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.0944, |
|
"grad_norm": 45.97714614868164, |
|
"learning_rate": 6.599999999999999e-05, |
|
"loss": 12.3225, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.096, |
|
"grad_norm": 49.00947952270508, |
|
"learning_rate": 6.72e-05, |
|
"loss": 12.3515, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.0976, |
|
"grad_norm": 54.81338119506836, |
|
"learning_rate": 6.84e-05, |
|
"loss": 12.8657, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.0992, |
|
"grad_norm": 44.2459831237793, |
|
"learning_rate": 6.96e-05, |
|
"loss": 10.2856, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.1008, |
|
"grad_norm": 73.28941345214844, |
|
"learning_rate": 7.079999999999999e-05, |
|
"loss": 9.7418, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.1024, |
|
"grad_norm": 40.017478942871094, |
|
"learning_rate": 7.199999999999999e-05, |
|
"loss": 8.9116, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.104, |
|
"grad_norm": 35.54833221435547, |
|
"learning_rate": 7.319999999999999e-05, |
|
"loss": 7.9677, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.1056, |
|
"grad_norm": 42.94904708862305, |
|
"learning_rate": 7.439999999999999e-05, |
|
"loss": 8.5134, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.1072, |
|
"grad_norm": 38.947715759277344, |
|
"learning_rate": 7.56e-05, |
|
"loss": 7.7674, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.1088, |
|
"grad_norm": 35.40559768676758, |
|
"learning_rate": 7.68e-05, |
|
"loss": 7.2294, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.1104, |
|
"grad_norm": 31.678964614868164, |
|
"learning_rate": 7.8e-05, |
|
"loss": 6.7146, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.112, |
|
"grad_norm": 31.80811309814453, |
|
"learning_rate": 7.92e-05, |
|
"loss": 6.5397, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.1136, |
|
"grad_norm": 25.740388870239258, |
|
"learning_rate": 8.04e-05, |
|
"loss": 6.0079, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.1152, |
|
"grad_norm": 23.663726806640625, |
|
"learning_rate": 8.16e-05, |
|
"loss": 5.7781, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.1168, |
|
"grad_norm": 20.862640380859375, |
|
"learning_rate": 8.28e-05, |
|
"loss": 5.5184, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.1184, |
|
"grad_norm": 17.784168243408203, |
|
"learning_rate": 8.4e-05, |
|
"loss": 5.3262, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 16.173917770385742, |
|
"learning_rate": 8.519999999999998e-05, |
|
"loss": 5.2462, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.1216, |
|
"grad_norm": 11.109197616577148, |
|
"learning_rate": 8.639999999999999e-05, |
|
"loss": 4.989, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.1232, |
|
"grad_norm": 9.369626998901367, |
|
"learning_rate": 8.759999999999999e-05, |
|
"loss": 4.9122, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.1248, |
|
"grad_norm": 9.535905838012695, |
|
"learning_rate": 8.879999999999999e-05, |
|
"loss": 4.9956, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.1264, |
|
"grad_norm": 3.752645969390869, |
|
"learning_rate": 8.999999999999999e-05, |
|
"loss": 4.7227, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.128, |
|
"grad_norm": 4.03261137008667, |
|
"learning_rate": 9.12e-05, |
|
"loss": 4.7404, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.1296, |
|
"grad_norm": 3.3785626888275146, |
|
"learning_rate": 9.24e-05, |
|
"loss": 4.6154, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.1312, |
|
"grad_norm": 3.4677908420562744, |
|
"learning_rate": 9.36e-05, |
|
"loss": 4.6707, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.1328, |
|
"grad_norm": 4.760728359222412, |
|
"learning_rate": 9.479999999999999e-05, |
|
"loss": 4.5489, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.1344, |
|
"grad_norm": 2.995011329650879, |
|
"learning_rate": 9.599999999999999e-05, |
|
"loss": 4.5604, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.136, |
|
"grad_norm": 3.4023890495300293, |
|
"learning_rate": 9.719999999999999e-05, |
|
"loss": 4.4197, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.1376, |
|
"grad_norm": 3.294135570526123, |
|
"learning_rate": 9.839999999999999e-05, |
|
"loss": 4.3887, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.1392, |
|
"grad_norm": 2.753955841064453, |
|
"learning_rate": 9.96e-05, |
|
"loss": 4.3751, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.1408, |
|
"grad_norm": 3.3620405197143555, |
|
"learning_rate": 0.0001008, |
|
"loss": 4.2838, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.1424, |
|
"grad_norm": 2.385225534439087, |
|
"learning_rate": 0.000102, |
|
"loss": 4.183, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.144, |
|
"grad_norm": 2.7207183837890625, |
|
"learning_rate": 0.00010319999999999999, |
|
"loss": 4.1382, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.1456, |
|
"grad_norm": 2.0092227458953857, |
|
"learning_rate": 0.00010439999999999999, |
|
"loss": 4.1519, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.1472, |
|
"grad_norm": 1.8385318517684937, |
|
"learning_rate": 0.00010559999999999998, |
|
"loss": 4.136, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.1488, |
|
"grad_norm": 1.3982303142547607, |
|
"learning_rate": 0.00010679999999999998, |
|
"loss": 4.084, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.1504, |
|
"grad_norm": 1.4304462671279907, |
|
"learning_rate": 0.00010799999999999998, |
|
"loss": 4.0993, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.152, |
|
"grad_norm": 5.04845666885376, |
|
"learning_rate": 0.00010919999999999998, |
|
"loss": 4.1147, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.1536, |
|
"grad_norm": 1.1188616752624512, |
|
"learning_rate": 0.00011039999999999999, |
|
"loss": 4.0138, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.1552, |
|
"grad_norm": 2.9660868644714355, |
|
"learning_rate": 0.00011159999999999999, |
|
"loss": 4.0886, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.1568, |
|
"grad_norm": 2.909925699234009, |
|
"learning_rate": 0.00011279999999999999, |
|
"loss": 4.0743, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.1584, |
|
"grad_norm": 2.233374834060669, |
|
"learning_rate": 0.00011399999999999999, |
|
"loss": 4.0643, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 1.9833197593688965, |
|
"learning_rate": 0.0001152, |
|
"loss": 4.1569, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.1616, |
|
"grad_norm": 31.450782775878906, |
|
"learning_rate": 0.0001164, |
|
"loss": 4.756, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.1632, |
|
"grad_norm": 11.571601867675781, |
|
"learning_rate": 0.0001176, |
|
"loss": 4.0323, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.1648, |
|
"grad_norm": 9.144487380981445, |
|
"learning_rate": 0.0001188, |
|
"loss": 3.9735, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.1664, |
|
"grad_norm": 16.850412368774414, |
|
"learning_rate": 0.00011999999999999999, |
|
"loss": 4.218, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.168, |
|
"grad_norm": 2.83750319480896, |
|
"learning_rate": 0.00012119999999999999, |
|
"loss": 3.9086, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.1696, |
|
"grad_norm": 1.1910499334335327, |
|
"learning_rate": 0.0001224, |
|
"loss": 3.8875, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.1712, |
|
"grad_norm": 1.7703046798706055, |
|
"learning_rate": 0.0001236, |
|
"loss": 3.9357, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.1728, |
|
"grad_norm": 4.045347690582275, |
|
"learning_rate": 0.00012479999999999997, |
|
"loss": 3.8306, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.1744, |
|
"grad_norm": 3.414834499359131, |
|
"learning_rate": 0.00012599999999999997, |
|
"loss": 3.8273, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.176, |
|
"grad_norm": 5.279017925262451, |
|
"learning_rate": 0.00012719999999999997, |
|
"loss": 3.8331, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.1776, |
|
"grad_norm": 2.1371779441833496, |
|
"learning_rate": 0.00012839999999999998, |
|
"loss": 3.8236, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.1792, |
|
"grad_norm": 0.6024636030197144, |
|
"learning_rate": 0.00012959999999999998, |
|
"loss": 3.7802, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.1808, |
|
"grad_norm": 3.284954786300659, |
|
"learning_rate": 0.00013079999999999998, |
|
"loss": 3.837, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.1824, |
|
"grad_norm": 3.2697064876556396, |
|
"learning_rate": 0.00013199999999999998, |
|
"loss": 3.8099, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.184, |
|
"grad_norm": 2.267110824584961, |
|
"learning_rate": 0.00013319999999999999, |
|
"loss": 3.7916, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.1856, |
|
"grad_norm": 0.9686329364776611, |
|
"learning_rate": 0.0001344, |
|
"loss": 3.7612, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.1872, |
|
"grad_norm": 1.1181795597076416, |
|
"learning_rate": 0.0001356, |
|
"loss": 3.817, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.1888, |
|
"grad_norm": 4.291112899780273, |
|
"learning_rate": 0.0001368, |
|
"loss": 3.75, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.1904, |
|
"grad_norm": 1.1991240978240967, |
|
"learning_rate": 0.000138, |
|
"loss": 3.8755, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.192, |
|
"grad_norm": 1.800611138343811, |
|
"learning_rate": 0.0001392, |
|
"loss": 3.7636, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.1936, |
|
"grad_norm": 0.6300956606864929, |
|
"learning_rate": 0.0001404, |
|
"loss": 3.8781, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.1952, |
|
"grad_norm": 1.9492263793945312, |
|
"learning_rate": 0.00014159999999999997, |
|
"loss": 3.758, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.1968, |
|
"grad_norm": 1.5942705869674683, |
|
"learning_rate": 0.00014279999999999997, |
|
"loss": 3.7344, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.1984, |
|
"grad_norm": 1.0828086137771606, |
|
"learning_rate": 0.00014399999999999998, |
|
"loss": 3.7535, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 1.54654860496521, |
|
"learning_rate": 0.00014519999999999998, |
|
"loss": 3.7764, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.2016, |
|
"grad_norm": 3.488430976867676, |
|
"learning_rate": 0.00014639999999999998, |
|
"loss": 3.8125, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.2032, |
|
"grad_norm": 1.4746885299682617, |
|
"learning_rate": 0.00014759999999999998, |
|
"loss": 3.8042, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.2048, |
|
"grad_norm": 2.114961862564087, |
|
"learning_rate": 0.00014879999999999998, |
|
"loss": 3.7654, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.2064, |
|
"grad_norm": 1.086930751800537, |
|
"learning_rate": 0.00015, |
|
"loss": 3.797, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.208, |
|
"grad_norm": 0.5767809152603149, |
|
"learning_rate": 0.0001512, |
|
"loss": 3.7507, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.2096, |
|
"grad_norm": 2.0122199058532715, |
|
"learning_rate": 0.0001524, |
|
"loss": 3.7508, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.2112, |
|
"grad_norm": 3.8814480304718018, |
|
"learning_rate": 0.0001536, |
|
"loss": 3.8003, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.2128, |
|
"grad_norm": 1.92637038230896, |
|
"learning_rate": 0.0001548, |
|
"loss": 3.8009, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.2144, |
|
"grad_norm": 0.7897951006889343, |
|
"learning_rate": 0.000156, |
|
"loss": 3.7989, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.216, |
|
"grad_norm": 4.2987751960754395, |
|
"learning_rate": 0.0001572, |
|
"loss": 3.7605, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.2176, |
|
"grad_norm": 1.7444305419921875, |
|
"learning_rate": 0.0001584, |
|
"loss": 3.8734, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.2192, |
|
"grad_norm": 3.6664788722991943, |
|
"learning_rate": 0.0001596, |
|
"loss": 3.7828, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.2208, |
|
"grad_norm": 1.8154056072235107, |
|
"learning_rate": 0.0001608, |
|
"loss": 3.9124, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.2224, |
|
"grad_norm": 0.8211868405342102, |
|
"learning_rate": 0.000162, |
|
"loss": 3.7861, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.224, |
|
"grad_norm": 1.5533865690231323, |
|
"learning_rate": 0.0001632, |
|
"loss": 3.787, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.2256, |
|
"grad_norm": 2.957888603210449, |
|
"learning_rate": 0.0001644, |
|
"loss": 3.7569, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.2272, |
|
"grad_norm": 3.3956074714660645, |
|
"learning_rate": 0.0001656, |
|
"loss": 3.854, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.2288, |
|
"grad_norm": 4.3899993896484375, |
|
"learning_rate": 0.0001668, |
|
"loss": 3.8498, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.2304, |
|
"grad_norm": 2.408184051513672, |
|
"learning_rate": 0.000168, |
|
"loss": 3.7483, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.232, |
|
"grad_norm": 2.7876155376434326, |
|
"learning_rate": 0.00016919999999999997, |
|
"loss": 3.7817, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.2336, |
|
"grad_norm": 2.2329955101013184, |
|
"learning_rate": 0.00017039999999999997, |
|
"loss": 3.7676, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.2352, |
|
"grad_norm": 1.5696635246276855, |
|
"learning_rate": 0.00017159999999999997, |
|
"loss": 3.811, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.2368, |
|
"grad_norm": 2.209275007247925, |
|
"learning_rate": 0.00017279999999999997, |
|
"loss": 3.8107, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.2384, |
|
"grad_norm": 1.1010030508041382, |
|
"learning_rate": 0.00017399999999999997, |
|
"loss": 3.8716, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 3.2085022926330566, |
|
"learning_rate": 0.00017519999999999998, |
|
"loss": 3.9153, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.2416, |
|
"grad_norm": 27.279512405395508, |
|
"learning_rate": 0.00017639999999999998, |
|
"loss": 4.6372, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.2432, |
|
"grad_norm": 6.734384536743164, |
|
"learning_rate": 0.00017759999999999998, |
|
"loss": 3.8794, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.2448, |
|
"grad_norm": 8.395252227783203, |
|
"learning_rate": 0.00017879999999999998, |
|
"loss": 3.9047, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.2464, |
|
"grad_norm": 2.386127471923828, |
|
"learning_rate": 0.00017999999999999998, |
|
"loss": 3.8455, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.248, |
|
"grad_norm": 5.339539051055908, |
|
"learning_rate": 0.00018119999999999999, |
|
"loss": 4.004, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.2496, |
|
"grad_norm": 6.078152179718018, |
|
"learning_rate": 0.0001824, |
|
"loss": 3.8299, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.2512, |
|
"grad_norm": 6.789322853088379, |
|
"learning_rate": 0.0001836, |
|
"loss": 3.863, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.2528, |
|
"grad_norm": 6.238964557647705, |
|
"learning_rate": 0.0001848, |
|
"loss": 3.8967, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.2544, |
|
"grad_norm": 4.713714122772217, |
|
"learning_rate": 0.000186, |
|
"loss": 3.8696, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.256, |
|
"grad_norm": 2.388154983520508, |
|
"learning_rate": 0.0001872, |
|
"loss": 3.752, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.2576, |
|
"grad_norm": 3.3313629627227783, |
|
"learning_rate": 0.00018839999999999997, |
|
"loss": 3.7579, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.2592, |
|
"grad_norm": 5.692492485046387, |
|
"learning_rate": 0.00018959999999999997, |
|
"loss": 3.7794, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.2608, |
|
"grad_norm": 5.304704189300537, |
|
"learning_rate": 0.00019079999999999998, |
|
"loss": 3.8021, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.2624, |
|
"grad_norm": 5.1804118156433105, |
|
"learning_rate": 0.00019199999999999998, |
|
"loss": 3.7633, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.264, |
|
"grad_norm": 3.4341394901275635, |
|
"learning_rate": 0.00019319999999999998, |
|
"loss": 3.802, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.2656, |
|
"grad_norm": 0.7816358804702759, |
|
"learning_rate": 0.00019439999999999998, |
|
"loss": 3.7327, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.2672, |
|
"grad_norm": 3.8968913555145264, |
|
"learning_rate": 0.00019559999999999998, |
|
"loss": 3.7701, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.2688, |
|
"grad_norm": 6.041684150695801, |
|
"learning_rate": 0.00019679999999999999, |
|
"loss": 3.7651, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.2704, |
|
"grad_norm": 1.5194474458694458, |
|
"learning_rate": 0.000198, |
|
"loss": 3.9153, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.272, |
|
"grad_norm": 5.066249847412109, |
|
"learning_rate": 0.0001992, |
|
"loss": 3.7799, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.2736, |
|
"grad_norm": 2.518998146057129, |
|
"learning_rate": 0.0002004, |
|
"loss": 3.7326, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.2752, |
|
"grad_norm": 0.759394645690918, |
|
"learning_rate": 0.0002016, |
|
"loss": 3.7889, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.2768, |
|
"grad_norm": 3.221085786819458, |
|
"learning_rate": 0.0002028, |
|
"loss": 3.709, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.2784, |
|
"grad_norm": 5.799932956695557, |
|
"learning_rate": 0.000204, |
|
"loss": 3.8104, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 4.494676113128662, |
|
"learning_rate": 0.0002052, |
|
"loss": 3.7602, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.2816, |
|
"grad_norm": 1.4825067520141602, |
|
"learning_rate": 0.00020639999999999998, |
|
"loss": 3.6961, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.2832, |
|
"grad_norm": 1.2989227771759033, |
|
"learning_rate": 0.00020759999999999998, |
|
"loss": 3.7053, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.2848, |
|
"grad_norm": 2.4774816036224365, |
|
"learning_rate": 0.00020879999999999998, |
|
"loss": 3.7072, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.2864, |
|
"grad_norm": 0.7821967005729675, |
|
"learning_rate": 0.00020999999999999998, |
|
"loss": 3.7205, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.288, |
|
"grad_norm": 2.0641937255859375, |
|
"learning_rate": 0.00021119999999999996, |
|
"loss": 3.7218, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.2896, |
|
"grad_norm": 0.7465824484825134, |
|
"learning_rate": 0.00021239999999999996, |
|
"loss": 3.7159, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.2912, |
|
"grad_norm": 1.9227252006530762, |
|
"learning_rate": 0.00021359999999999996, |
|
"loss": 3.7154, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.2928, |
|
"grad_norm": 2.588310480117798, |
|
"learning_rate": 0.00021479999999999996, |
|
"loss": 3.7532, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.2944, |
|
"grad_norm": 0.8672159314155579, |
|
"learning_rate": 0.00021599999999999996, |
|
"loss": 3.7199, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.296, |
|
"grad_norm": 2.015648365020752, |
|
"learning_rate": 0.00021719999999999997, |
|
"loss": 3.7661, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.2976, |
|
"grad_norm": 1.280931830406189, |
|
"learning_rate": 0.00021839999999999997, |
|
"loss": 3.7427, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.2992, |
|
"grad_norm": 1.2730371952056885, |
|
"learning_rate": 0.00021959999999999997, |
|
"loss": 3.8019, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.3008, |
|
"grad_norm": 0.9145507216453552, |
|
"learning_rate": 0.00022079999999999997, |
|
"loss": 3.7265, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.3024, |
|
"grad_norm": 1.7408198118209839, |
|
"learning_rate": 0.00022199999999999998, |
|
"loss": 3.6761, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.304, |
|
"grad_norm": 4.375244140625, |
|
"learning_rate": 0.00022319999999999998, |
|
"loss": 3.7797, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.3056, |
|
"grad_norm": 2.859325647354126, |
|
"learning_rate": 0.00022439999999999998, |
|
"loss": 3.7509, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.3072, |
|
"grad_norm": 2.0287702083587646, |
|
"learning_rate": 0.00022559999999999998, |
|
"loss": 3.7396, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.3088, |
|
"grad_norm": 0.7529137134552002, |
|
"learning_rate": 0.00022679999999999998, |
|
"loss": 3.7297, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.3104, |
|
"grad_norm": 0.8018739223480225, |
|
"learning_rate": 0.00022799999999999999, |
|
"loss": 3.758, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.312, |
|
"grad_norm": 1.8517733812332153, |
|
"learning_rate": 0.0002292, |
|
"loss": 3.7313, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.3136, |
|
"grad_norm": 1.8108242750167847, |
|
"learning_rate": 0.0002304, |
|
"loss": 3.726, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.3152, |
|
"grad_norm": 1.7063705921173096, |
|
"learning_rate": 0.0002316, |
|
"loss": 3.6929, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.3168, |
|
"grad_norm": 1.4768927097320557, |
|
"learning_rate": 0.0002328, |
|
"loss": 3.7741, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.3184, |
|
"grad_norm": 2.0207772254943848, |
|
"learning_rate": 0.000234, |
|
"loss": 3.8361, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.000234, |
|
"loss": 3.9766, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.3216, |
|
"grad_norm": 28.53923797607422, |
|
"learning_rate": 0.0002352, |
|
"loss": 4.9473, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.3232, |
|
"grad_norm": 19.729700088500977, |
|
"learning_rate": 0.0002364, |
|
"loss": 3.8604, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.3248, |
|
"grad_norm": 15.63243293762207, |
|
"learning_rate": 0.0002376, |
|
"loss": 4.1119, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.3264, |
|
"grad_norm": 4.515649795532227, |
|
"learning_rate": 0.0002388, |
|
"loss": 3.7548, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.328, |
|
"grad_norm": 3.835597038269043, |
|
"learning_rate": 0.00023999999999999998, |
|
"loss": 3.7632, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.3296, |
|
"grad_norm": 3.379566192626953, |
|
"learning_rate": 0.00024119999999999998, |
|
"loss": 3.9366, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.3312, |
|
"grad_norm": 1.9478070735931396, |
|
"learning_rate": 0.00024239999999999998, |
|
"loss": 3.7263, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.3328, |
|
"grad_norm": 2.06083083152771, |
|
"learning_rate": 0.00024359999999999999, |
|
"loss": 3.6968, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.3344, |
|
"grad_norm": 3.923818826675415, |
|
"learning_rate": 0.0002448, |
|
"loss": 3.7776, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.336, |
|
"grad_norm": 3.4094479084014893, |
|
"learning_rate": 0.00024599999999999996, |
|
"loss": 3.7088, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.3376, |
|
"grad_norm": 4.099003314971924, |
|
"learning_rate": 0.0002472, |
|
"loss": 3.7462, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.3392, |
|
"grad_norm": 0.9599695205688477, |
|
"learning_rate": 0.00024839999999999997, |
|
"loss": 3.7051, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.3408, |
|
"grad_norm": 3.5445761680603027, |
|
"learning_rate": 0.00024959999999999994, |
|
"loss": 3.6189, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.3424, |
|
"grad_norm": 3.2550857067108154, |
|
"learning_rate": 0.00025079999999999997, |
|
"loss": 3.5834, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.344, |
|
"grad_norm": 1.9503517150878906, |
|
"learning_rate": 0.00025199999999999995, |
|
"loss": 3.5801, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.3456, |
|
"grad_norm": 4.3856520652771, |
|
"learning_rate": 0.0002532, |
|
"loss": 3.7169, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.3472, |
|
"grad_norm": 1.5111596584320068, |
|
"learning_rate": 0.00025439999999999995, |
|
"loss": 3.5968, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.3488, |
|
"grad_norm": 1.61884605884552, |
|
"learning_rate": 0.0002556, |
|
"loss": 3.5628, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.3504, |
|
"grad_norm": 0.7278621792793274, |
|
"learning_rate": 0.00025679999999999995, |
|
"loss": 3.6218, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.352, |
|
"grad_norm": 0.9155241847038269, |
|
"learning_rate": 0.000258, |
|
"loss": 3.4842, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.3536, |
|
"grad_norm": 1.681706428527832, |
|
"learning_rate": 0.00025919999999999996, |
|
"loss": 3.4532, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.3552, |
|
"grad_norm": 2.0195343494415283, |
|
"learning_rate": 0.0002604, |
|
"loss": 3.529, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.3568, |
|
"grad_norm": 2.215996026992798, |
|
"learning_rate": 0.00026159999999999996, |
|
"loss": 3.5834, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.3584, |
|
"grad_norm": 3.6210215091705322, |
|
"learning_rate": 0.0002628, |
|
"loss": 3.5076, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 3.3881304264068604, |
|
"learning_rate": 0.00026399999999999997, |
|
"loss": 3.5052, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.3616, |
|
"grad_norm": 1.2712512016296387, |
|
"learning_rate": 0.0002652, |
|
"loss": 3.4167, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.3632, |
|
"grad_norm": 2.784555435180664, |
|
"learning_rate": 0.00026639999999999997, |
|
"loss": 3.4773, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.3648, |
|
"grad_norm": 3.9173593521118164, |
|
"learning_rate": 0.0002676, |
|
"loss": 3.4792, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.3664, |
|
"grad_norm": 5.97532844543457, |
|
"learning_rate": 0.0002688, |
|
"loss": 3.581, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.368, |
|
"grad_norm": 1.1036829948425293, |
|
"learning_rate": 0.00027, |
|
"loss": 3.4026, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.3696, |
|
"grad_norm": 2.0570623874664307, |
|
"learning_rate": 0.0002712, |
|
"loss": 3.3855, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.3712, |
|
"grad_norm": 2.33469295501709, |
|
"learning_rate": 0.0002724, |
|
"loss": 3.3493, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.3728, |
|
"grad_norm": 1.7420787811279297, |
|
"learning_rate": 0.0002736, |
|
"loss": 3.4412, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.3744, |
|
"grad_norm": 1.597578525543213, |
|
"learning_rate": 0.0002748, |
|
"loss": 3.4772, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.376, |
|
"grad_norm": 1.469741702079773, |
|
"learning_rate": 0.000276, |
|
"loss": 3.2345, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.3776, |
|
"grad_norm": 1.8439031839370728, |
|
"learning_rate": 0.0002772, |
|
"loss": 3.2449, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.3792, |
|
"grad_norm": 1.8763359785079956, |
|
"learning_rate": 0.0002784, |
|
"loss": 3.0865, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.3808, |
|
"grad_norm": 1.3372673988342285, |
|
"learning_rate": 0.00027959999999999997, |
|
"loss": 3.2648, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.3824, |
|
"grad_norm": 1.0363004207611084, |
|
"learning_rate": 0.0002808, |
|
"loss": 3.1095, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.384, |
|
"grad_norm": 1.2411675453186035, |
|
"learning_rate": 0.00028199999999999997, |
|
"loss": 3.1516, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.3856, |
|
"grad_norm": 1.1692326068878174, |
|
"learning_rate": 0.00028319999999999994, |
|
"loss": 3.1162, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.3872, |
|
"grad_norm": 1.514488697052002, |
|
"learning_rate": 0.0002844, |
|
"loss": 3.1885, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.3888, |
|
"grad_norm": 1.0548737049102783, |
|
"learning_rate": 0.00028559999999999995, |
|
"loss": 2.9995, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.3904, |
|
"grad_norm": 1.435293436050415, |
|
"learning_rate": 0.0002868, |
|
"loss": 3.1228, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.392, |
|
"grad_norm": 4.880733966827393, |
|
"learning_rate": 0.00028799999999999995, |
|
"loss": 3.523, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.3936, |
|
"grad_norm": 2.134737730026245, |
|
"learning_rate": 0.0002892, |
|
"loss": 2.9433, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.3952, |
|
"grad_norm": 1.4167667627334595, |
|
"learning_rate": 0.00029039999999999996, |
|
"loss": 3.0941, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.3968, |
|
"grad_norm": 1.7885382175445557, |
|
"learning_rate": 0.0002916, |
|
"loss": 3.076, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.3984, |
|
"grad_norm": 2.0048017501831055, |
|
"learning_rate": 0.00029279999999999996, |
|
"loss": 3.1318, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 2.751702070236206, |
|
"learning_rate": 0.000294, |
|
"loss": 3.437, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.4016, |
|
"grad_norm": 60.15131759643555, |
|
"learning_rate": 0.00029519999999999997, |
|
"loss": 6.7094, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.4032, |
|
"grad_norm": 11.904341697692871, |
|
"learning_rate": 0.0002964, |
|
"loss": 3.3368, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.4048, |
|
"grad_norm": 5.927310943603516, |
|
"learning_rate": 0.00029759999999999997, |
|
"loss": 3.084, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.4064, |
|
"grad_norm": 1.259347915649414, |
|
"learning_rate": 0.0002988, |
|
"loss": 2.8328, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.408, |
|
"grad_norm": 2.5598220825195312, |
|
"learning_rate": 0.0003, |
|
"loss": 2.9085, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.4096, |
|
"grad_norm": 1.866328239440918, |
|
"learning_rate": 0.00030119999999999995, |
|
"loss": 2.8637, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.4112, |
|
"grad_norm": 1.2088593244552612, |
|
"learning_rate": 0.0003024, |
|
"loss": 2.6899, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.4128, |
|
"grad_norm": 4.605785369873047, |
|
"learning_rate": 0.00030359999999999995, |
|
"loss": 2.9239, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.4144, |
|
"grad_norm": 1.7429507970809937, |
|
"learning_rate": 0.0003048, |
|
"loss": 2.5301, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.416, |
|
"grad_norm": 2.9041364192962646, |
|
"learning_rate": 0.00030599999999999996, |
|
"loss": 2.4356, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.4176, |
|
"grad_norm": 2.5931711196899414, |
|
"learning_rate": 0.0003072, |
|
"loss": 2.523, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.4192, |
|
"grad_norm": 1.1678705215454102, |
|
"learning_rate": 0.00030839999999999996, |
|
"loss": 2.2468, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.4208, |
|
"grad_norm": 1.711000680923462, |
|
"learning_rate": 0.0003096, |
|
"loss": 2.5393, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.4224, |
|
"grad_norm": 1.8182960748672485, |
|
"learning_rate": 0.00031079999999999997, |
|
"loss": 2.4138, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.424, |
|
"grad_norm": 1.3307945728302002, |
|
"learning_rate": 0.000312, |
|
"loss": 2.2623, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.4256, |
|
"grad_norm": 1.9444172382354736, |
|
"learning_rate": 0.00031319999999999997, |
|
"loss": 2.1209, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.4272, |
|
"grad_norm": 2.1593079566955566, |
|
"learning_rate": 0.0003144, |
|
"loss": 1.9432, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.4288, |
|
"grad_norm": 2.1220879554748535, |
|
"learning_rate": 0.0003156, |
|
"loss": 2.1439, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.4304, |
|
"grad_norm": 0.8869176506996155, |
|
"learning_rate": 0.0003168, |
|
"loss": 2.1583, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.432, |
|
"grad_norm": 1.2422492504119873, |
|
"learning_rate": 0.000318, |
|
"loss": 1.8846, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.4336, |
|
"grad_norm": 0.8885542154312134, |
|
"learning_rate": 0.0003192, |
|
"loss": 1.9293, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.4352, |
|
"grad_norm": 1.3430697917938232, |
|
"learning_rate": 0.0003204, |
|
"loss": 2.1379, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.4368, |
|
"grad_norm": 2.2559776306152344, |
|
"learning_rate": 0.0003216, |
|
"loss": 2.311, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.4384, |
|
"grad_norm": 1.0061554908752441, |
|
"learning_rate": 0.0003228, |
|
"loss": 1.8522, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.9519514441490173, |
|
"learning_rate": 0.000324, |
|
"loss": 1.9418, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.4416, |
|
"grad_norm": 1.034746527671814, |
|
"learning_rate": 0.0003252, |
|
"loss": 2.1713, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.4432, |
|
"grad_norm": 1.4670822620391846, |
|
"learning_rate": 0.0003264, |
|
"loss": 2.2635, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.4448, |
|
"grad_norm": 1.1949667930603027, |
|
"learning_rate": 0.0003276, |
|
"loss": 2.447, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.4464, |
|
"grad_norm": 2.0934438705444336, |
|
"learning_rate": 0.0003288, |
|
"loss": 1.9578, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.448, |
|
"grad_norm": 1.7934505939483643, |
|
"learning_rate": 0.00033, |
|
"loss": 2.3822, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.4496, |
|
"grad_norm": 1.0247673988342285, |
|
"learning_rate": 0.0003312, |
|
"loss": 1.8567, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.4512, |
|
"grad_norm": 2.136781692504883, |
|
"learning_rate": 0.0003324, |
|
"loss": 1.964, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.4528, |
|
"grad_norm": 0.8177749514579773, |
|
"learning_rate": 0.0003336, |
|
"loss": 1.8132, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.4544, |
|
"grad_norm": 1.0221589803695679, |
|
"learning_rate": 0.0003348, |
|
"loss": 1.8668, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.456, |
|
"grad_norm": 1.517808198928833, |
|
"learning_rate": 0.000336, |
|
"loss": 2.1496, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.4576, |
|
"grad_norm": 1.5944926738739014, |
|
"learning_rate": 0.0003372, |
|
"loss": 2.3953, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.4592, |
|
"grad_norm": 2.405046224594116, |
|
"learning_rate": 0.00033839999999999993, |
|
"loss": 1.9663, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.4608, |
|
"grad_norm": 1.6332200765609741, |
|
"learning_rate": 0.00033959999999999996, |
|
"loss": 1.7194, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.4624, |
|
"grad_norm": 1.463877558708191, |
|
"learning_rate": 0.00034079999999999994, |
|
"loss": 1.9922, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.464, |
|
"grad_norm": 1.3162888288497925, |
|
"learning_rate": 0.00034199999999999996, |
|
"loss": 1.9758, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.4656, |
|
"grad_norm": 1.7767980098724365, |
|
"learning_rate": 0.00034319999999999994, |
|
"loss": 2.2791, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.4672, |
|
"grad_norm": 1.7777577638626099, |
|
"learning_rate": 0.00034439999999999997, |
|
"loss": 1.5986, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.4688, |
|
"grad_norm": 1.4968204498291016, |
|
"learning_rate": 0.00034559999999999994, |
|
"loss": 1.7837, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.4704, |
|
"grad_norm": 1.639837384223938, |
|
"learning_rate": 0.0003467999999999999, |
|
"loss": 2.1542, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.472, |
|
"grad_norm": 2.0573577880859375, |
|
"learning_rate": 0.00034799999999999995, |
|
"loss": 2.1574, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.4736, |
|
"grad_norm": 2.6885788440704346, |
|
"learning_rate": 0.0003491999999999999, |
|
"loss": 2.0754, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.4752, |
|
"grad_norm": 2.3689215183258057, |
|
"learning_rate": 0.00035039999999999995, |
|
"loss": 2.3191, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.4768, |
|
"grad_norm": 1.4669629335403442, |
|
"learning_rate": 0.0003515999999999999, |
|
"loss": 2.1903, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.4784, |
|
"grad_norm": 6.631944179534912, |
|
"learning_rate": 0.00035279999999999996, |
|
"loss": 2.4098, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 3.869600296020508, |
|
"learning_rate": 0.00035399999999999993, |
|
"loss": 2.8155, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.4816, |
|
"grad_norm": 18.670400619506836, |
|
"learning_rate": 0.00035519999999999996, |
|
"loss": 4.0821, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.4832, |
|
"grad_norm": 6.5373640060424805, |
|
"learning_rate": 0.00035639999999999994, |
|
"loss": 2.6903, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.4848, |
|
"grad_norm": 1.8103080987930298, |
|
"learning_rate": 0.00035759999999999996, |
|
"loss": 1.715, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.4864, |
|
"grad_norm": 2.0340678691864014, |
|
"learning_rate": 0.00035879999999999994, |
|
"loss": 2.2299, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.488, |
|
"grad_norm": 2.8201074600219727, |
|
"learning_rate": 0.00035999999999999997, |
|
"loss": 2.0269, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.4896, |
|
"grad_norm": 3.9219393730163574, |
|
"learning_rate": 0.00036119999999999994, |
|
"loss": 2.0192, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.4912, |
|
"grad_norm": 2.1272470951080322, |
|
"learning_rate": 0.00036239999999999997, |
|
"loss": 2.2065, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.4928, |
|
"grad_norm": 1.2031842470169067, |
|
"learning_rate": 0.00036359999999999995, |
|
"loss": 1.6669, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.4944, |
|
"grad_norm": 2.6753766536712646, |
|
"learning_rate": 0.0003648, |
|
"loss": 1.9611, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.496, |
|
"grad_norm": 3.980130672454834, |
|
"learning_rate": 0.00036599999999999995, |
|
"loss": 1.8961, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.4976, |
|
"grad_norm": 3.4683375358581543, |
|
"learning_rate": 0.0003672, |
|
"loss": 1.6241, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.4992, |
|
"grad_norm": 3.478597402572632, |
|
"learning_rate": 0.00036839999999999996, |
|
"loss": 1.8472, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.5008, |
|
"grad_norm": 1.360845923423767, |
|
"learning_rate": 0.0003696, |
|
"loss": 1.7252, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.5024, |
|
"grad_norm": 0.9226222634315491, |
|
"learning_rate": 0.00037079999999999996, |
|
"loss": 1.4792, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.504, |
|
"grad_norm": 1.2864586114883423, |
|
"learning_rate": 0.000372, |
|
"loss": 1.5637, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.5056, |
|
"grad_norm": 1.395561695098877, |
|
"learning_rate": 0.00037319999999999996, |
|
"loss": 1.41, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.5072, |
|
"grad_norm": 1.9199680089950562, |
|
"learning_rate": 0.0003744, |
|
"loss": 1.8221, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.5088, |
|
"grad_norm": 1.2985516786575317, |
|
"learning_rate": 0.00037559999999999997, |
|
"loss": 1.9545, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.5104, |
|
"grad_norm": 0.861045777797699, |
|
"learning_rate": 0.00037679999999999994, |
|
"loss": 1.2, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.512, |
|
"grad_norm": 1.4724054336547852, |
|
"learning_rate": 0.00037799999999999997, |
|
"loss": 1.9655, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.5136, |
|
"grad_norm": 1.121520757675171, |
|
"learning_rate": 0.00037919999999999995, |
|
"loss": 1.6427, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.5152, |
|
"grad_norm": 0.8945459127426147, |
|
"learning_rate": 0.0003804, |
|
"loss": 1.8193, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.5168, |
|
"grad_norm": 1.510231614112854, |
|
"learning_rate": 0.00038159999999999995, |
|
"loss": 1.7606, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.5184, |
|
"grad_norm": 1.2624521255493164, |
|
"learning_rate": 0.0003828, |
|
"loss": 1.8403, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 1.1738777160644531, |
|
"learning_rate": 0.00038399999999999996, |
|
"loss": 1.5676, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.5216, |
|
"grad_norm": 1.4238841533660889, |
|
"learning_rate": 0.0003852, |
|
"loss": 1.3816, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.5232, |
|
"grad_norm": 0.9597876667976379, |
|
"learning_rate": 0.00038639999999999996, |
|
"loss": 1.4629, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.5248, |
|
"grad_norm": 1.1177942752838135, |
|
"learning_rate": 0.0003876, |
|
"loss": 1.4452, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.5264, |
|
"grad_norm": 1.0831985473632812, |
|
"learning_rate": 0.00038879999999999996, |
|
"loss": 1.6668, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.528, |
|
"grad_norm": 1.1985876560211182, |
|
"learning_rate": 0.00039, |
|
"loss": 1.5963, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.5296, |
|
"grad_norm": 1.4813580513000488, |
|
"learning_rate": 0.00039119999999999997, |
|
"loss": 1.6768, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.5312, |
|
"grad_norm": 1.4565857648849487, |
|
"learning_rate": 0.0003924, |
|
"loss": 1.6633, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.5328, |
|
"grad_norm": 1.1487796306610107, |
|
"learning_rate": 0.00039359999999999997, |
|
"loss": 1.4567, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.5344, |
|
"grad_norm": 1.2718472480773926, |
|
"learning_rate": 0.0003948, |
|
"loss": 1.8923, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.536, |
|
"grad_norm": 1.393419623374939, |
|
"learning_rate": 0.000396, |
|
"loss": 1.1306, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.5376, |
|
"grad_norm": 2.660311222076416, |
|
"learning_rate": 0.0003972, |
|
"loss": 1.6447, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.5392, |
|
"grad_norm": 1.1137669086456299, |
|
"learning_rate": 0.0003984, |
|
"loss": 1.6307, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.5408, |
|
"grad_norm": 1.1761451959609985, |
|
"learning_rate": 0.0003996, |
|
"loss": 1.6008, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.5424, |
|
"grad_norm": 1.350234866142273, |
|
"learning_rate": 0.0004008, |
|
"loss": 1.523, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.544, |
|
"grad_norm": 1.166507601737976, |
|
"learning_rate": 0.000402, |
|
"loss": 1.2526, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.5456, |
|
"grad_norm": 1.3881250619888306, |
|
"learning_rate": 0.0004032, |
|
"loss": 1.7568, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.5472, |
|
"grad_norm": 3.9268593788146973, |
|
"learning_rate": 0.0004044, |
|
"loss": 2.0863, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.5488, |
|
"grad_norm": 1.7067468166351318, |
|
"learning_rate": 0.0004056, |
|
"loss": 1.5831, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.5504, |
|
"grad_norm": 1.4713155031204224, |
|
"learning_rate": 0.00040679999999999997, |
|
"loss": 1.7429, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.552, |
|
"grad_norm": 1.212177038192749, |
|
"learning_rate": 0.000408, |
|
"loss": 1.6922, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.5536, |
|
"grad_norm": 1.6186903715133667, |
|
"learning_rate": 0.00040919999999999997, |
|
"loss": 1.818, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.5552, |
|
"grad_norm": 2.359452247619629, |
|
"learning_rate": 0.0004104, |
|
"loss": 2.4012, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.5568, |
|
"grad_norm": 2.3700032234191895, |
|
"learning_rate": 0.0004116, |
|
"loss": 2.296, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.5584, |
|
"grad_norm": 1.8284653425216675, |
|
"learning_rate": 0.00041279999999999995, |
|
"loss": 2.0039, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 2.491885185241699, |
|
"learning_rate": 0.0004139999999999999, |
|
"loss": 2.089, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.5616, |
|
"grad_norm": 2.3013062477111816, |
|
"learning_rate": 0.00041519999999999995, |
|
"loss": 1.6711, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.5632, |
|
"grad_norm": 1.9905917644500732, |
|
"learning_rate": 0.00041639999999999993, |
|
"loss": 1.6065, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.5648, |
|
"grad_norm": 1.480592966079712, |
|
"learning_rate": 0.00041759999999999996, |
|
"loss": 2.2154, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.5664, |
|
"grad_norm": 2.0091054439544678, |
|
"learning_rate": 0.00041879999999999993, |
|
"loss": 1.9469, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.568, |
|
"grad_norm": 4.029331207275391, |
|
"learning_rate": 0.00041999999999999996, |
|
"loss": 2.159, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.5696, |
|
"grad_norm": 3.835162401199341, |
|
"learning_rate": 0.00042119999999999994, |
|
"loss": 1.9669, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.5712, |
|
"grad_norm": 3.283651828765869, |
|
"learning_rate": 0.0004223999999999999, |
|
"loss": 1.9169, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.5728, |
|
"grad_norm": 2.7341041564941406, |
|
"learning_rate": 0.00042359999999999994, |
|
"loss": 1.9569, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.5744, |
|
"grad_norm": 3.72660231590271, |
|
"learning_rate": 0.0004247999999999999, |
|
"loss": 1.9345, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.576, |
|
"grad_norm": 1.2118239402770996, |
|
"learning_rate": 0.00042599999999999995, |
|
"loss": 1.6209, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.5776, |
|
"grad_norm": 2.773261547088623, |
|
"learning_rate": 0.0004271999999999999, |
|
"loss": 1.7308, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.5792, |
|
"grad_norm": 1.5158611536026, |
|
"learning_rate": 0.00042839999999999995, |
|
"loss": 1.5566, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.5808, |
|
"grad_norm": 1.7444158792495728, |
|
"learning_rate": 0.0004295999999999999, |
|
"loss": 1.517, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.5824, |
|
"grad_norm": 2.0275840759277344, |
|
"learning_rate": 0.00043079999999999995, |
|
"loss": 1.3961, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.584, |
|
"grad_norm": 1.0864077806472778, |
|
"learning_rate": 0.00043199999999999993, |
|
"loss": 1.3833, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.5856, |
|
"grad_norm": 1.5802364349365234, |
|
"learning_rate": 0.00043319999999999996, |
|
"loss": 1.2144, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.5872, |
|
"grad_norm": 1.1365376710891724, |
|
"learning_rate": 0.00043439999999999993, |
|
"loss": 1.4435, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.5888, |
|
"grad_norm": 2.001816511154175, |
|
"learning_rate": 0.00043559999999999996, |
|
"loss": 1.8125, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.5904, |
|
"grad_norm": 3.9782650470733643, |
|
"learning_rate": 0.00043679999999999994, |
|
"loss": 2.3917, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.592, |
|
"grad_norm": 2.559185743331909, |
|
"learning_rate": 0.00043799999999999997, |
|
"loss": 1.4065, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.5936, |
|
"grad_norm": 1.531436562538147, |
|
"learning_rate": 0.00043919999999999994, |
|
"loss": 1.3286, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.5952, |
|
"grad_norm": 2.1857967376708984, |
|
"learning_rate": 0.00044039999999999997, |
|
"loss": 1.3389, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.5968, |
|
"grad_norm": 1.2437169551849365, |
|
"learning_rate": 0.00044159999999999995, |
|
"loss": 1.305, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.5984, |
|
"grad_norm": 1.8238170146942139, |
|
"learning_rate": 0.0004428, |
|
"loss": 1.6173, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.9433510303497314, |
|
"learning_rate": 0.00044399999999999995, |
|
"loss": 1.2818, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.6016, |
|
"grad_norm": 1.7307677268981934, |
|
"learning_rate": 0.0004452, |
|
"loss": 1.3214, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.6032, |
|
"grad_norm": 1.2876639366149902, |
|
"learning_rate": 0.00044639999999999995, |
|
"loss": 1.6527, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.6048, |
|
"grad_norm": 1.0600067377090454, |
|
"learning_rate": 0.0004476, |
|
"loss": 1.4502, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.6064, |
|
"grad_norm": 1.1501784324645996, |
|
"learning_rate": 0.00044879999999999996, |
|
"loss": 1.2737, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.608, |
|
"grad_norm": 0.959336519241333, |
|
"learning_rate": 0.00045, |
|
"loss": 1.7977, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.6096, |
|
"grad_norm": 1.1226460933685303, |
|
"learning_rate": 0.00045119999999999996, |
|
"loss": 1.4758, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.6112, |
|
"grad_norm": 0.9224548935890198, |
|
"learning_rate": 0.00045239999999999994, |
|
"loss": 1.5619, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.6128, |
|
"grad_norm": 0.814278244972229, |
|
"learning_rate": 0.00045359999999999997, |
|
"loss": 1.3943, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.6144, |
|
"grad_norm": 0.9084352850914001, |
|
"learning_rate": 0.00045479999999999994, |
|
"loss": 1.5962, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.616, |
|
"grad_norm": 1.4289610385894775, |
|
"learning_rate": 0.00045599999999999997, |
|
"loss": 1.3404, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.6176, |
|
"grad_norm": 1.2523407936096191, |
|
"learning_rate": 0.00045719999999999995, |
|
"loss": 1.4539, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.6192, |
|
"grad_norm": 1.1009465456008911, |
|
"learning_rate": 0.0004584, |
|
"loss": 1.2913, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.6208, |
|
"grad_norm": 1.369174838066101, |
|
"learning_rate": 0.00045959999999999995, |
|
"loss": 1.7786, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.6224, |
|
"grad_norm": 1.6469858884811401, |
|
"learning_rate": 0.0004608, |
|
"loss": 1.5913, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.624, |
|
"grad_norm": 1.0242924690246582, |
|
"learning_rate": 0.00046199999999999995, |
|
"loss": 1.4396, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.6256, |
|
"grad_norm": 1.0944545269012451, |
|
"learning_rate": 0.0004632, |
|
"loss": 1.3629, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.6272, |
|
"grad_norm": 1.2820014953613281, |
|
"learning_rate": 0.00046439999999999996, |
|
"loss": 1.4899, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.6288, |
|
"grad_norm": 1.1084744930267334, |
|
"learning_rate": 0.0004656, |
|
"loss": 1.3476, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.6304, |
|
"grad_norm": 1.2067919969558716, |
|
"learning_rate": 0.00046679999999999996, |
|
"loss": 1.6783, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.632, |
|
"grad_norm": 1.3350406885147095, |
|
"learning_rate": 0.000468, |
|
"loss": 1.8544, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.6336, |
|
"grad_norm": 2.817788600921631, |
|
"learning_rate": 0.00046919999999999997, |
|
"loss": 1.6983, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.6352, |
|
"grad_norm": 1.577996850013733, |
|
"learning_rate": 0.0004704, |
|
"loss": 1.9165, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.6368, |
|
"grad_norm": 1.992092251777649, |
|
"learning_rate": 0.00047159999999999997, |
|
"loss": 2.2309, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.6384, |
|
"grad_norm": 1.5959856510162354, |
|
"learning_rate": 0.0004728, |
|
"loss": 1.977, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 2.959681510925293, |
|
"learning_rate": 0.000474, |
|
"loss": 1.9346, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.6416, |
|
"grad_norm": 3.0177066326141357, |
|
"learning_rate": 0.0004752, |
|
"loss": 1.9946, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.6432, |
|
"grad_norm": 3.2861762046813965, |
|
"learning_rate": 0.0004764, |
|
"loss": 1.861, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.6448, |
|
"grad_norm": 33.552921295166016, |
|
"learning_rate": 0.0004776, |
|
"loss": 7.6541, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.6464, |
|
"grad_norm": 1.975786566734314, |
|
"learning_rate": 0.0004788, |
|
"loss": 1.7211, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.648, |
|
"grad_norm": 3.4860012531280518, |
|
"learning_rate": 0.00047999999999999996, |
|
"loss": 1.8209, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.6496, |
|
"grad_norm": 3.5102968215942383, |
|
"learning_rate": 0.0004812, |
|
"loss": 1.7063, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.6512, |
|
"grad_norm": 4.879241943359375, |
|
"learning_rate": 0.00048239999999999996, |
|
"loss": 1.4615, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.6528, |
|
"grad_norm": 2.7851948738098145, |
|
"learning_rate": 0.0004836, |
|
"loss": 1.6937, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.6544, |
|
"grad_norm": 0.9268562197685242, |
|
"learning_rate": 0.00048479999999999997, |
|
"loss": 1.6795, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.656, |
|
"grad_norm": 1.127549409866333, |
|
"learning_rate": 0.000486, |
|
"loss": 1.433, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.6576, |
|
"grad_norm": 2.1968774795532227, |
|
"learning_rate": 0.00048719999999999997, |
|
"loss": 1.501, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.6592, |
|
"grad_norm": 3.553457260131836, |
|
"learning_rate": 0.0004883999999999999, |
|
"loss": 1.4432, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.6608, |
|
"grad_norm": 1.8429263830184937, |
|
"learning_rate": 0.0004896, |
|
"loss": 1.3622, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.6624, |
|
"grad_norm": 0.8631522059440613, |
|
"learning_rate": 0.0004907999999999999, |
|
"loss": 1.157, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.664, |
|
"grad_norm": 0.8570797443389893, |
|
"learning_rate": 0.0004919999999999999, |
|
"loss": 1.03, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.6656, |
|
"grad_norm": 1.2855173349380493, |
|
"learning_rate": 0.0004932, |
|
"loss": 1.5622, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.6672, |
|
"grad_norm": 1.406083106994629, |
|
"learning_rate": 0.0004944, |
|
"loss": 1.0664, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.6688, |
|
"grad_norm": 1.5346801280975342, |
|
"learning_rate": 0.0004955999999999999, |
|
"loss": 1.5327, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.6704, |
|
"grad_norm": 1.3836517333984375, |
|
"learning_rate": 0.0004967999999999999, |
|
"loss": 1.6197, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.672, |
|
"grad_norm": 1.7024117708206177, |
|
"learning_rate": 0.000498, |
|
"loss": 1.2376, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.6736, |
|
"grad_norm": 1.5172358751296997, |
|
"learning_rate": 0.0004991999999999999, |
|
"loss": 1.5432, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.6752, |
|
"grad_norm": 1.142734408378601, |
|
"learning_rate": 0.0005003999999999999, |
|
"loss": 1.2484, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.6768, |
|
"grad_norm": 0.8977586030960083, |
|
"learning_rate": 0.0005015999999999999, |
|
"loss": 1.3032, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.6784, |
|
"grad_norm": 1.1880444288253784, |
|
"learning_rate": 0.0005028, |
|
"loss": 1.4404, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 1.214245080947876, |
|
"learning_rate": 0.0005039999999999999, |
|
"loss": 1.2976, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.6816, |
|
"grad_norm": 0.9443445801734924, |
|
"learning_rate": 0.0005051999999999999, |
|
"loss": 1.418, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.6832, |
|
"grad_norm": 1.4814517498016357, |
|
"learning_rate": 0.0005064, |
|
"loss": 1.3793, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.6848, |
|
"grad_norm": 1.3838948011398315, |
|
"learning_rate": 0.0005076, |
|
"loss": 1.1823, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.6864, |
|
"grad_norm": 0.9880338311195374, |
|
"learning_rate": 0.0005087999999999999, |
|
"loss": 1.108, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.688, |
|
"grad_norm": 1.0871669054031372, |
|
"learning_rate": 0.0005099999999999999, |
|
"loss": 1.3633, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.6896, |
|
"grad_norm": 1.2696417570114136, |
|
"learning_rate": 0.0005112, |
|
"loss": 1.3324, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 0.6912, |
|
"grad_norm": 1.4589694738388062, |
|
"learning_rate": 0.0005124, |
|
"loss": 0.9689, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.6928, |
|
"grad_norm": 0.948417603969574, |
|
"learning_rate": 0.0005135999999999999, |
|
"loss": 1.2153, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 0.6944, |
|
"grad_norm": 1.4363794326782227, |
|
"learning_rate": 0.0005147999999999999, |
|
"loss": 1.7676, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.696, |
|
"grad_norm": 0.8274084329605103, |
|
"learning_rate": 0.000516, |
|
"loss": 1.247, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.6976, |
|
"grad_norm": 1.2022773027420044, |
|
"learning_rate": 0.0005172, |
|
"loss": 1.9134, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.6992, |
|
"grad_norm": 1.0256644487380981, |
|
"learning_rate": 0.0005183999999999999, |
|
"loss": 1.5249, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.7008, |
|
"grad_norm": 1.2487961053848267, |
|
"learning_rate": 0.0005195999999999999, |
|
"loss": 1.4572, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.7024, |
|
"grad_norm": 1.103989601135254, |
|
"learning_rate": 0.0005208, |
|
"loss": 1.2583, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 0.704, |
|
"grad_norm": 1.1899611949920654, |
|
"learning_rate": 0.000522, |
|
"loss": 1.249, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.7056, |
|
"grad_norm": 1.1985859870910645, |
|
"learning_rate": 0.0005231999999999999, |
|
"loss": 1.3384, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.7072, |
|
"grad_norm": 1.278523325920105, |
|
"learning_rate": 0.0005244, |
|
"loss": 1.6037, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.7088, |
|
"grad_norm": 2.5164453983306885, |
|
"learning_rate": 0.0005256, |
|
"loss": 1.7037, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 0.7104, |
|
"grad_norm": 1.4125896692276, |
|
"learning_rate": 0.0005267999999999999, |
|
"loss": 1.3628, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.712, |
|
"grad_norm": 1.4269789457321167, |
|
"learning_rate": 0.0005279999999999999, |
|
"loss": 1.8712, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.7136, |
|
"grad_norm": 1.321560263633728, |
|
"learning_rate": 0.0005292, |
|
"loss": 1.4942, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.7152, |
|
"grad_norm": 1.5753790140151978, |
|
"learning_rate": 0.0005304, |
|
"loss": 1.9359, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.7168, |
|
"grad_norm": 2.2859463691711426, |
|
"learning_rate": 0.0005315999999999999, |
|
"loss": 1.745, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.7184, |
|
"grad_norm": 2.0483174324035645, |
|
"learning_rate": 0.0005327999999999999, |
|
"loss": 1.8343, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 3.6337218284606934, |
|
"learning_rate": 0.000534, |
|
"loss": 2.2459, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.7216, |
|
"grad_norm": 13.555660247802734, |
|
"learning_rate": 0.0005352, |
|
"loss": 3.3534, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 0.7232, |
|
"grad_norm": 6.6364850997924805, |
|
"learning_rate": 0.0005363999999999999, |
|
"loss": 2.7515, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 0.7248, |
|
"grad_norm": 3.817852258682251, |
|
"learning_rate": 0.0005376, |
|
"loss": 2.012, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 0.7264, |
|
"grad_norm": 1.707593560218811, |
|
"learning_rate": 0.0005388, |
|
"loss": 1.9942, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 0.728, |
|
"grad_norm": 2.77917218208313, |
|
"learning_rate": 0.00054, |
|
"loss": 1.6863, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.7296, |
|
"grad_norm": 2.7656164169311523, |
|
"learning_rate": 0.0005411999999999999, |
|
"loss": 1.4779, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 0.7312, |
|
"grad_norm": 1.720285177230835, |
|
"learning_rate": 0.0005424, |
|
"loss": 1.8024, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 0.7328, |
|
"grad_norm": 3.847505807876587, |
|
"learning_rate": 0.0005436, |
|
"loss": 1.6034, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 0.7344, |
|
"grad_norm": 2.7850637435913086, |
|
"learning_rate": 0.0005448, |
|
"loss": 1.3834, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 0.736, |
|
"grad_norm": 1.2482507228851318, |
|
"learning_rate": 0.0005459999999999999, |
|
"loss": 1.6495, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.7376, |
|
"grad_norm": 1.2714293003082275, |
|
"learning_rate": 0.0005472, |
|
"loss": 1.3752, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 0.7392, |
|
"grad_norm": 1.8939746618270874, |
|
"learning_rate": 0.0005484, |
|
"loss": 1.4186, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 0.7408, |
|
"grad_norm": 2.5669922828674316, |
|
"learning_rate": 0.0005496, |
|
"loss": 1.2681, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 0.7424, |
|
"grad_norm": 1.9449177980422974, |
|
"learning_rate": 0.0005507999999999999, |
|
"loss": 1.3308, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 0.744, |
|
"grad_norm": 0.872009813785553, |
|
"learning_rate": 0.000552, |
|
"loss": 1.2645, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.7456, |
|
"grad_norm": 0.8539568185806274, |
|
"learning_rate": 0.0005532, |
|
"loss": 1.0258, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 0.7472, |
|
"grad_norm": 1.0018901824951172, |
|
"learning_rate": 0.0005544, |
|
"loss": 1.38, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 0.7488, |
|
"grad_norm": 0.9926770329475403, |
|
"learning_rate": 0.0005556, |
|
"loss": 1.3214, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 0.7504, |
|
"grad_norm": 1.80833101272583, |
|
"learning_rate": 0.0005568, |
|
"loss": 1.3013, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 0.752, |
|
"grad_norm": 1.4460201263427734, |
|
"learning_rate": 0.000558, |
|
"loss": 1.3652, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.7536, |
|
"grad_norm": 1.0537903308868408, |
|
"learning_rate": 0.0005591999999999999, |
|
"loss": 1.1968, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 0.7552, |
|
"grad_norm": 1.0886517763137817, |
|
"learning_rate": 0.0005604, |
|
"loss": 1.2747, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 0.7568, |
|
"grad_norm": 1.2867629528045654, |
|
"learning_rate": 0.0005616, |
|
"loss": 1.2078, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 0.7584, |
|
"grad_norm": 1.7365913391113281, |
|
"learning_rate": 0.0005627999999999999, |
|
"loss": 1.0753, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 1.7022826671600342, |
|
"learning_rate": 0.0005639999999999999, |
|
"loss": 1.2664, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.7616, |
|
"grad_norm": 1.3423399925231934, |
|
"learning_rate": 0.0005652, |
|
"loss": 1.3553, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 0.7632, |
|
"grad_norm": 1.3367669582366943, |
|
"learning_rate": 0.0005663999999999999, |
|
"loss": 1.4162, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 0.7648, |
|
"grad_norm": 0.9925369024276733, |
|
"learning_rate": 0.0005675999999999999, |
|
"loss": 1.2366, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 0.7664, |
|
"grad_norm": 1.741668939590454, |
|
"learning_rate": 0.0005688, |
|
"loss": 1.2624, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.768, |
|
"grad_norm": 1.771984577178955, |
|
"learning_rate": 0.00057, |
|
"loss": 1.0515, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.7696, |
|
"grad_norm": 1.5952296257019043, |
|
"learning_rate": 0.0005711999999999999, |
|
"loss": 1.2872, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 0.7712, |
|
"grad_norm": 1.2628203630447388, |
|
"learning_rate": 0.0005723999999999999, |
|
"loss": 1.0806, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 0.7728, |
|
"grad_norm": 1.4792985916137695, |
|
"learning_rate": 0.0005736, |
|
"loss": 1.962, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 0.7744, |
|
"grad_norm": 2.177412509918213, |
|
"learning_rate": 0.0005747999999999999, |
|
"loss": 1.4523, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 0.776, |
|
"grad_norm": 1.865964412689209, |
|
"learning_rate": 0.0005759999999999999, |
|
"loss": 1.401, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.7776, |
|
"grad_norm": 1.4428671598434448, |
|
"learning_rate": 0.0005771999999999999, |
|
"loss": 1.3322, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 0.7792, |
|
"grad_norm": 0.9237609505653381, |
|
"learning_rate": 0.0005784, |
|
"loss": 1.4704, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 0.7808, |
|
"grad_norm": 2.0012426376342773, |
|
"learning_rate": 0.0005795999999999999, |
|
"loss": 1.6074, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 0.7824, |
|
"grad_norm": 1.1296131610870361, |
|
"learning_rate": 0.0005807999999999999, |
|
"loss": 1.4989, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 0.784, |
|
"grad_norm": 1.091180682182312, |
|
"learning_rate": 0.0005819999999999999, |
|
"loss": 1.4049, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.7856, |
|
"grad_norm": 1.2416259050369263, |
|
"learning_rate": 0.0005832, |
|
"loss": 1.8097, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 0.7872, |
|
"grad_norm": 0.9992501139640808, |
|
"learning_rate": 0.0005843999999999999, |
|
"loss": 1.5731, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 0.7888, |
|
"grad_norm": 1.1882905960083008, |
|
"learning_rate": 0.0005855999999999999, |
|
"loss": 1.2349, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 0.7904, |
|
"grad_norm": 1.3324135541915894, |
|
"learning_rate": 0.0005868, |
|
"loss": 1.749, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 0.792, |
|
"grad_norm": 1.3113425970077515, |
|
"learning_rate": 0.000588, |
|
"loss": 1.3155, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.7936, |
|
"grad_norm": 1.333341121673584, |
|
"learning_rate": 0.0005891999999999999, |
|
"loss": 1.0152, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 0.7952, |
|
"grad_norm": 1.886502742767334, |
|
"learning_rate": 0.0005903999999999999, |
|
"loss": 1.5642, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 0.7968, |
|
"grad_norm": 2.012117385864258, |
|
"learning_rate": 0.0005916, |
|
"loss": 1.6069, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 0.7984, |
|
"grad_norm": 2.344853401184082, |
|
"learning_rate": 0.0005928, |
|
"loss": 2.0689, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 2.7430222034454346, |
|
"learning_rate": 0.0005939999999999999, |
|
"loss": 2.81, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.8016, |
|
"grad_norm": 2.373655319213867, |
|
"learning_rate": 0.0005951999999999999, |
|
"loss": 1.8218, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 0.8032, |
|
"grad_norm": 1.2365477085113525, |
|
"learning_rate": 0.0005964, |
|
"loss": 1.343, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 0.8048, |
|
"grad_norm": 2.100356101989746, |
|
"learning_rate": 0.0005976, |
|
"loss": 2.0204, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 0.8064, |
|
"grad_norm": 1.6222838163375854, |
|
"learning_rate": 0.0005987999999999999, |
|
"loss": 1.8366, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 0.808, |
|
"grad_norm": 3.295870542526245, |
|
"learning_rate": 0.0006, |
|
"loss": 1.6686, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.8096, |
|
"grad_norm": 3.5636391639709473, |
|
"learning_rate": 0.0005987999999999999, |
|
"loss": 1.958, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 0.8112, |
|
"grad_norm": 4.600498199462891, |
|
"learning_rate": 0.0005976, |
|
"loss": 1.4633, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 0.8128, |
|
"grad_norm": 3.999089241027832, |
|
"learning_rate": 0.0005964, |
|
"loss": 1.591, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 0.8144, |
|
"grad_norm": 2.074601173400879, |
|
"learning_rate": 0.0005951999999999999, |
|
"loss": 1.4589, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 0.816, |
|
"grad_norm": 1.2597025632858276, |
|
"learning_rate": 0.0005939999999999999, |
|
"loss": 1.517, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.8176, |
|
"grad_norm": 1.491461157798767, |
|
"learning_rate": 0.0005928, |
|
"loss": 1.1305, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 0.8192, |
|
"grad_norm": 2.2012178897857666, |
|
"learning_rate": 0.0005916, |
|
"loss": 1.2055, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 0.8208, |
|
"grad_norm": 2.303264617919922, |
|
"learning_rate": 0.0005903999999999999, |
|
"loss": 1.4843, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 0.8224, |
|
"grad_norm": 1.3678765296936035, |
|
"learning_rate": 0.0005891999999999999, |
|
"loss": 1.1635, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 0.824, |
|
"grad_norm": 1.7093764543533325, |
|
"learning_rate": 0.000588, |
|
"loss": 1.1574, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.8256, |
|
"grad_norm": 1.2002806663513184, |
|
"learning_rate": 0.0005868, |
|
"loss": 1.1366, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 0.8272, |
|
"grad_norm": 1.1055371761322021, |
|
"learning_rate": 0.0005855999999999999, |
|
"loss": 1.1778, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 0.8288, |
|
"grad_norm": 0.9321176409721375, |
|
"learning_rate": 0.0005843999999999999, |
|
"loss": 0.8233, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 0.8304, |
|
"grad_norm": 1.3442676067352295, |
|
"learning_rate": 0.0005832, |
|
"loss": 1.2463, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 0.832, |
|
"grad_norm": 0.9121391177177429, |
|
"learning_rate": 0.0005819999999999999, |
|
"loss": 1.3646, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.8336, |
|
"grad_norm": 1.1537225246429443, |
|
"learning_rate": 0.0005807999999999999, |
|
"loss": 1.4225, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 0.8352, |
|
"grad_norm": 1.0641944408416748, |
|
"learning_rate": 0.0005795999999999999, |
|
"loss": 1.0161, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 0.8368, |
|
"grad_norm": 0.7178429961204529, |
|
"learning_rate": 0.0005784, |
|
"loss": 1.1851, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 0.8384, |
|
"grad_norm": 1.4418388605117798, |
|
"learning_rate": 0.0005771999999999999, |
|
"loss": 1.8398, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 1.382843255996704, |
|
"learning_rate": 0.0005759999999999999, |
|
"loss": 1.4898, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.8416, |
|
"grad_norm": 1.0729074478149414, |
|
"learning_rate": 0.0005747999999999999, |
|
"loss": 1.2898, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 0.8432, |
|
"grad_norm": 0.9983257055282593, |
|
"learning_rate": 0.0005736, |
|
"loss": 1.1974, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 0.8448, |
|
"grad_norm": 1.1875462532043457, |
|
"learning_rate": 0.0005723999999999999, |
|
"loss": 1.1169, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 0.8464, |
|
"grad_norm": 1.318334698677063, |
|
"learning_rate": 0.0005711999999999999, |
|
"loss": 1.1376, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 0.848, |
|
"grad_norm": 1.537840485572815, |
|
"learning_rate": 0.00057, |
|
"loss": 1.2072, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.8496, |
|
"grad_norm": 1.4589056968688965, |
|
"learning_rate": 0.0005688, |
|
"loss": 1.373, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 0.8512, |
|
"grad_norm": 1.019971251487732, |
|
"learning_rate": 0.0005675999999999999, |
|
"loss": 1.0613, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 0.8528, |
|
"grad_norm": 1.0122156143188477, |
|
"learning_rate": 0.0005663999999999999, |
|
"loss": 0.9874, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 0.8544, |
|
"grad_norm": 1.1434595584869385, |
|
"learning_rate": 0.0005652, |
|
"loss": 1.2304, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 0.856, |
|
"grad_norm": 1.2336163520812988, |
|
"learning_rate": 0.0005639999999999999, |
|
"loss": 1.6633, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.8576, |
|
"grad_norm": 1.306872844696045, |
|
"learning_rate": 0.0005627999999999999, |
|
"loss": 1.7359, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 0.8592, |
|
"grad_norm": 1.2644526958465576, |
|
"learning_rate": 0.0005616, |
|
"loss": 1.2328, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 0.8608, |
|
"grad_norm": 1.262831449508667, |
|
"learning_rate": 0.0005604, |
|
"loss": 1.0457, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 0.8624, |
|
"grad_norm": 0.9390996098518372, |
|
"learning_rate": 0.0005591999999999999, |
|
"loss": 1.2251, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 0.864, |
|
"grad_norm": 1.208268165588379, |
|
"learning_rate": 0.000558, |
|
"loss": 1.3572, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.8656, |
|
"grad_norm": 1.0636659860610962, |
|
"learning_rate": 0.0005568, |
|
"loss": 1.1701, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 0.8672, |
|
"grad_norm": 0.9668271541595459, |
|
"learning_rate": 0.0005556, |
|
"loss": 1.5845, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 0.8688, |
|
"grad_norm": 0.9891708493232727, |
|
"learning_rate": 0.0005544, |
|
"loss": 1.3511, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 0.8704, |
|
"grad_norm": 1.7175298929214478, |
|
"learning_rate": 0.0005532, |
|
"loss": 1.2839, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 0.872, |
|
"grad_norm": 0.9767908453941345, |
|
"learning_rate": 0.000552, |
|
"loss": 1.5042, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.8736, |
|
"grad_norm": 1.3343541622161865, |
|
"learning_rate": 0.0005507999999999999, |
|
"loss": 1.6681, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 0.8752, |
|
"grad_norm": 1.3749518394470215, |
|
"learning_rate": 0.0005496, |
|
"loss": 1.9624, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 0.8768, |
|
"grad_norm": 2.514359712600708, |
|
"learning_rate": 0.0005484, |
|
"loss": 1.895, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 0.8784, |
|
"grad_norm": 1.514849066734314, |
|
"learning_rate": 0.0005472, |
|
"loss": 1.7603, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 1.4065580368041992, |
|
"learning_rate": 0.0005459999999999999, |
|
"loss": 1.9368, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.8816, |
|
"grad_norm": 4.109616756439209, |
|
"learning_rate": 0.0005448, |
|
"loss": 1.9927, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 0.8832, |
|
"grad_norm": 5.660764694213867, |
|
"learning_rate": 0.0005436, |
|
"loss": 2.4962, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 0.8848, |
|
"grad_norm": 2.027193307876587, |
|
"learning_rate": 0.0005424, |
|
"loss": 1.6865, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 0.8864, |
|
"grad_norm": 4.918600559234619, |
|
"learning_rate": 0.0005411999999999999, |
|
"loss": 2.1457, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 0.888, |
|
"grad_norm": 2.522416114807129, |
|
"learning_rate": 0.00054, |
|
"loss": 1.5093, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.8896, |
|
"grad_norm": 1.574242353439331, |
|
"learning_rate": 0.0005388, |
|
"loss": 1.1618, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 0.8912, |
|
"grad_norm": 2.081839084625244, |
|
"learning_rate": 0.0005376, |
|
"loss": 1.2351, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.8928, |
|
"grad_norm": 2.582669734954834, |
|
"learning_rate": 0.0005363999999999999, |
|
"loss": 1.6881, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 0.8944, |
|
"grad_norm": 2.058865785598755, |
|
"learning_rate": 0.0005352, |
|
"loss": 1.3369, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 0.896, |
|
"grad_norm": 5.098209381103516, |
|
"learning_rate": 0.000534, |
|
"loss": 1.7725, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.8976, |
|
"grad_norm": 4.526047229766846, |
|
"learning_rate": 0.0005327999999999999, |
|
"loss": 1.9443, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 0.8992, |
|
"grad_norm": 3.9119491577148438, |
|
"learning_rate": 0.0005315999999999999, |
|
"loss": 1.5399, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 0.9008, |
|
"grad_norm": 0.8436188101768494, |
|
"learning_rate": 0.0005304, |
|
"loss": 1.1211, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 0.9024, |
|
"grad_norm": 0.8627029061317444, |
|
"learning_rate": 0.0005292, |
|
"loss": 1.2774, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 0.904, |
|
"grad_norm": 1.1870328187942505, |
|
"learning_rate": 0.0005279999999999999, |
|
"loss": 1.2342, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.9056, |
|
"grad_norm": 1.0261473655700684, |
|
"learning_rate": 0.0005267999999999999, |
|
"loss": 1.0405, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 0.9072, |
|
"grad_norm": 0.6334408521652222, |
|
"learning_rate": 0.0005256, |
|
"loss": 1.0547, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 0.9088, |
|
"grad_norm": 0.7928243279457092, |
|
"learning_rate": 0.0005244, |
|
"loss": 0.9138, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 0.9104, |
|
"grad_norm": 0.9043545126914978, |
|
"learning_rate": 0.0005231999999999999, |
|
"loss": 1.0988, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 0.912, |
|
"grad_norm": 0.9109718203544617, |
|
"learning_rate": 0.000522, |
|
"loss": 1.0209, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.9136, |
|
"grad_norm": 0.8105588555335999, |
|
"learning_rate": 0.0005208, |
|
"loss": 1.2233, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 0.9152, |
|
"grad_norm": 2.0293259620666504, |
|
"learning_rate": 0.0005195999999999999, |
|
"loss": 1.4025, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 0.9168, |
|
"grad_norm": 1.1730594635009766, |
|
"learning_rate": 0.0005183999999999999, |
|
"loss": 1.481, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 0.9184, |
|
"grad_norm": 0.8027293086051941, |
|
"learning_rate": 0.0005172, |
|
"loss": 1.1333, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.8858001828193665, |
|
"learning_rate": 0.000516, |
|
"loss": 1.4056, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.9216, |
|
"grad_norm": 1.3030261993408203, |
|
"learning_rate": 0.0005147999999999999, |
|
"loss": 1.212, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 0.9232, |
|
"grad_norm": 0.9441176652908325, |
|
"learning_rate": 0.0005135999999999999, |
|
"loss": 1.6681, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 0.9248, |
|
"grad_norm": 0.9312158226966858, |
|
"learning_rate": 0.0005124, |
|
"loss": 1.2352, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 0.9264, |
|
"grad_norm": 1.1646393537521362, |
|
"learning_rate": 0.0005112, |
|
"loss": 1.1658, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 0.928, |
|
"grad_norm": 1.2840147018432617, |
|
"learning_rate": 0.0005099999999999999, |
|
"loss": 1.0198, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.9296, |
|
"grad_norm": 1.2597519159317017, |
|
"learning_rate": 0.0005087999999999999, |
|
"loss": 1.0831, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 0.9312, |
|
"grad_norm": 0.7873828411102295, |
|
"learning_rate": 0.0005076, |
|
"loss": 1.0068, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 0.9328, |
|
"grad_norm": 1.5674644708633423, |
|
"learning_rate": 0.0005064, |
|
"loss": 1.3783, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 0.9344, |
|
"grad_norm": 0.9917027354240417, |
|
"learning_rate": 0.0005051999999999999, |
|
"loss": 0.9963, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 0.936, |
|
"grad_norm": 0.9538294672966003, |
|
"learning_rate": 0.0005039999999999999, |
|
"loss": 0.9023, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.9376, |
|
"grad_norm": 1.1332108974456787, |
|
"learning_rate": 0.0005028, |
|
"loss": 1.0699, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 0.9392, |
|
"grad_norm": 1.0121673345565796, |
|
"learning_rate": 0.0005015999999999999, |
|
"loss": 1.0387, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 0.9408, |
|
"grad_norm": 1.13718581199646, |
|
"learning_rate": 0.0005003999999999999, |
|
"loss": 1.1858, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 0.9424, |
|
"grad_norm": 1.2693606615066528, |
|
"learning_rate": 0.0004991999999999999, |
|
"loss": 1.0405, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 0.944, |
|
"grad_norm": 1.4965566396713257, |
|
"learning_rate": 0.000498, |
|
"loss": 1.326, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.9456, |
|
"grad_norm": 1.2708925008773804, |
|
"learning_rate": 0.0004967999999999999, |
|
"loss": 1.2337, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 0.9472, |
|
"grad_norm": 2.260007858276367, |
|
"learning_rate": 0.0004955999999999999, |
|
"loss": 1.5223, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 0.9488, |
|
"grad_norm": 1.3588523864746094, |
|
"learning_rate": 0.0004944, |
|
"loss": 1.4486, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 0.9504, |
|
"grad_norm": 3.3943569660186768, |
|
"learning_rate": 0.0004932, |
|
"loss": 1.5663, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 0.952, |
|
"grad_norm": 1.5709065198898315, |
|
"learning_rate": 0.0004919999999999999, |
|
"loss": 1.3337, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.9536, |
|
"grad_norm": 1.6011497974395752, |
|
"learning_rate": 0.0004907999999999999, |
|
"loss": 1.4023, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 0.9552, |
|
"grad_norm": 1.2253276109695435, |
|
"learning_rate": 0.0004896, |
|
"loss": 1.0934, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 0.9568, |
|
"grad_norm": 9.454032897949219, |
|
"learning_rate": 0.0004883999999999999, |
|
"loss": 2.9201, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 0.9584, |
|
"grad_norm": 1.7915419340133667, |
|
"learning_rate": 0.00048719999999999997, |
|
"loss": 1.5261, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00048719999999999997, |
|
"loss": 1.3199, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.9616, |
|
"grad_norm": 18.144990921020508, |
|
"learning_rate": 0.000486, |
|
"loss": 4.1466, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 0.9632, |
|
"grad_norm": 1.2664769887924194, |
|
"learning_rate": 0.00048479999999999997, |
|
"loss": 1.6238, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 0.9648, |
|
"grad_norm": 2.453564405441284, |
|
"learning_rate": 0.0004836, |
|
"loss": 1.5469, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 0.9664, |
|
"grad_norm": 2.737936496734619, |
|
"learning_rate": 0.00048239999999999996, |
|
"loss": 1.3178, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 0.968, |
|
"grad_norm": 2.828806161880493, |
|
"learning_rate": 0.0004812, |
|
"loss": 1.4034, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.9696, |
|
"grad_norm": 1.4219012260437012, |
|
"learning_rate": 0.00047999999999999996, |
|
"loss": 1.0615, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 0.9712, |
|
"grad_norm": 2.025907039642334, |
|
"learning_rate": 0.0004788, |
|
"loss": 1.19, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 0.9728, |
|
"grad_norm": 0.9138876795768738, |
|
"learning_rate": 0.0004776, |
|
"loss": 1.1638, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 0.9744, |
|
"grad_norm": 0.7301196455955505, |
|
"learning_rate": 0.0004764, |
|
"loss": 1.2678, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 0.976, |
|
"grad_norm": 1.6543656587600708, |
|
"learning_rate": 0.0004752, |
|
"loss": 1.1815, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.9776, |
|
"grad_norm": 1.8122645616531372, |
|
"learning_rate": 0.000474, |
|
"loss": 1.0917, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 0.9792, |
|
"grad_norm": 1.609754204750061, |
|
"learning_rate": 0.0004728, |
|
"loss": 1.1377, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 0.9808, |
|
"grad_norm": 1.2491132020950317, |
|
"learning_rate": 0.00047159999999999997, |
|
"loss": 0.9748, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 0.9824, |
|
"grad_norm": 1.6153700351715088, |
|
"learning_rate": 0.0004704, |
|
"loss": 1.1825, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 0.984, |
|
"grad_norm": 2.4447743892669678, |
|
"learning_rate": 0.00046919999999999997, |
|
"loss": 1.1472, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.9856, |
|
"grad_norm": 0.93949294090271, |
|
"learning_rate": 0.000468, |
|
"loss": 0.9327, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 0.9872, |
|
"grad_norm": 1.0353221893310547, |
|
"learning_rate": 0.00046679999999999996, |
|
"loss": 1.3792, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 0.9888, |
|
"grad_norm": 1.5396970510482788, |
|
"learning_rate": 0.0004656, |
|
"loss": 1.4076, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 0.9904, |
|
"grad_norm": 2.1144979000091553, |
|
"learning_rate": 0.00046439999999999996, |
|
"loss": 1.2042, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 0.992, |
|
"grad_norm": 1.649453043937683, |
|
"learning_rate": 0.0004632, |
|
"loss": 1.033, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.9936, |
|
"grad_norm": 1.555721640586853, |
|
"learning_rate": 0.00046199999999999995, |
|
"loss": 1.4478, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 0.9952, |
|
"grad_norm": 2.0463335514068604, |
|
"learning_rate": 0.0004608, |
|
"loss": 1.427, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 0.9968, |
|
"grad_norm": 1.6396963596343994, |
|
"learning_rate": 0.00045959999999999995, |
|
"loss": 1.4215, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 0.9984, |
|
"grad_norm": 1.0603671073913574, |
|
"learning_rate": 0.0004584, |
|
"loss": 1.2777, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 2.150242328643799, |
|
"learning_rate": 0.00045719999999999995, |
|
"loss": 1.5985, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 1.0016, |
|
"grad_norm": 13.092059135437012, |
|
"learning_rate": 0.00045599999999999997, |
|
"loss": 3.3297, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 1.0032, |
|
"grad_norm": 5.357777118682861, |
|
"learning_rate": 0.00045479999999999994, |
|
"loss": 1.8506, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 1.0048, |
|
"grad_norm": 3.857879400253296, |
|
"learning_rate": 0.00045359999999999997, |
|
"loss": 1.4082, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 1.0064, |
|
"grad_norm": 5.025846004486084, |
|
"learning_rate": 0.00045239999999999994, |
|
"loss": 2.445, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 1.008, |
|
"grad_norm": 3.71512508392334, |
|
"learning_rate": 0.00045119999999999996, |
|
"loss": 2.0449, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 1.0096, |
|
"grad_norm": 4.47951078414917, |
|
"learning_rate": 0.00045, |
|
"loss": 1.9927, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 1.0112, |
|
"grad_norm": 2.713778257369995, |
|
"learning_rate": 0.00044879999999999996, |
|
"loss": 1.4958, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 1.0128, |
|
"grad_norm": 2.6333401203155518, |
|
"learning_rate": 0.0004476, |
|
"loss": 1.3756, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 1.0144, |
|
"grad_norm": 3.1156418323516846, |
|
"learning_rate": 0.00044639999999999995, |
|
"loss": 1.6268, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 1.016, |
|
"grad_norm": 2.78902268409729, |
|
"learning_rate": 0.0004452, |
|
"loss": 1.8211, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 1.0176, |
|
"grad_norm": 3.0045857429504395, |
|
"learning_rate": 0.00044399999999999995, |
|
"loss": 1.1757, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 1.0192, |
|
"grad_norm": 1.4704291820526123, |
|
"learning_rate": 0.0004428, |
|
"loss": 1.2811, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 1.0208, |
|
"grad_norm": 1.3432084321975708, |
|
"learning_rate": 0.00044159999999999995, |
|
"loss": 1.0086, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 1.0224, |
|
"grad_norm": 0.842569887638092, |
|
"learning_rate": 0.00044039999999999997, |
|
"loss": 1.3255, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 1.024, |
|
"grad_norm": 0.8691660761833191, |
|
"learning_rate": 0.00043919999999999994, |
|
"loss": 1.1076, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 1.0256, |
|
"grad_norm": 1.083778738975525, |
|
"learning_rate": 0.00043799999999999997, |
|
"loss": 1.0929, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 1.0272, |
|
"grad_norm": 1.1408025026321411, |
|
"learning_rate": 0.00043679999999999994, |
|
"loss": 1.2497, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 1.0288, |
|
"grad_norm": 0.8224440217018127, |
|
"learning_rate": 0.00043559999999999996, |
|
"loss": 1.2853, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 1.0304, |
|
"grad_norm": 0.7420323491096497, |
|
"learning_rate": 0.00043439999999999993, |
|
"loss": 1.187, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 1.032, |
|
"grad_norm": 0.7818359732627869, |
|
"learning_rate": 0.00043319999999999996, |
|
"loss": 1.0613, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 1.0336, |
|
"grad_norm": 1.2085120677947998, |
|
"learning_rate": 0.00043199999999999993, |
|
"loss": 1.0593, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 1.0352, |
|
"grad_norm": 0.6779820322990417, |
|
"learning_rate": 0.00043079999999999995, |
|
"loss": 0.9878, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 1.0368, |
|
"grad_norm": 1.1257340908050537, |
|
"learning_rate": 0.0004295999999999999, |
|
"loss": 1.0076, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 1.0384, |
|
"grad_norm": 0.6911525726318359, |
|
"learning_rate": 0.00042839999999999995, |
|
"loss": 1.1204, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 1.0192064046859741, |
|
"learning_rate": 0.0004271999999999999, |
|
"loss": 1.217, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.0416, |
|
"grad_norm": 0.8528116345405579, |
|
"learning_rate": 0.00042599999999999995, |
|
"loss": 1.1124, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 1.0432, |
|
"grad_norm": 0.8429757356643677, |
|
"learning_rate": 0.0004247999999999999, |
|
"loss": 1.0387, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 1.0448, |
|
"grad_norm": 1.0031050443649292, |
|
"learning_rate": 0.00042359999999999994, |
|
"loss": 1.1567, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 1.0464, |
|
"grad_norm": 1.1668227910995483, |
|
"learning_rate": 0.0004223999999999999, |
|
"loss": 1.3264, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 1.048, |
|
"grad_norm": 0.8505134582519531, |
|
"learning_rate": 0.00042119999999999994, |
|
"loss": 0.8013, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 1.0496, |
|
"grad_norm": 0.7736939191818237, |
|
"learning_rate": 0.00041999999999999996, |
|
"loss": 1.0249, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 1.0512, |
|
"grad_norm": 1.082900047302246, |
|
"learning_rate": 0.00041879999999999993, |
|
"loss": 1.0062, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 1.0528, |
|
"grad_norm": 1.9835671186447144, |
|
"learning_rate": 0.00041759999999999996, |
|
"loss": 1.4081, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 1.0544, |
|
"grad_norm": 0.9174219965934753, |
|
"learning_rate": 0.00041639999999999993, |
|
"loss": 0.8602, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 1.056, |
|
"grad_norm": 0.8988387584686279, |
|
"learning_rate": 0.00041519999999999995, |
|
"loss": 1.0813, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 1.0576, |
|
"grad_norm": 0.9161027669906616, |
|
"learning_rate": 0.0004139999999999999, |
|
"loss": 0.9137, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 1.0592, |
|
"grad_norm": 1.0347057580947876, |
|
"learning_rate": 0.00041279999999999995, |
|
"loss": 0.9526, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 1.0608, |
|
"grad_norm": 0.9697722792625427, |
|
"learning_rate": 0.0004116, |
|
"loss": 1.2357, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 1.0624, |
|
"grad_norm": 1.1263432502746582, |
|
"learning_rate": 0.0004104, |
|
"loss": 1.2324, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 1.064, |
|
"grad_norm": 1.4703420400619507, |
|
"learning_rate": 0.00040919999999999997, |
|
"loss": 1.2418, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 1.0656, |
|
"grad_norm": 1.2557834386825562, |
|
"learning_rate": 0.000408, |
|
"loss": 1.5181, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 1.0672, |
|
"grad_norm": 1.1496392488479614, |
|
"learning_rate": 0.00040679999999999997, |
|
"loss": 1.5552, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 1.0688, |
|
"grad_norm": 1.0062506198883057, |
|
"learning_rate": 0.0004056, |
|
"loss": 1.3242, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 1.0704, |
|
"grad_norm": 0.8978244066238403, |
|
"learning_rate": 0.0004044, |
|
"loss": 1.3264, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 1.072, |
|
"grad_norm": 1.2901413440704346, |
|
"learning_rate": 0.0004032, |
|
"loss": 1.3077, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 1.0735999999999999, |
|
"grad_norm": 1.202578067779541, |
|
"learning_rate": 0.000402, |
|
"loss": 1.1621, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 1.0752, |
|
"grad_norm": 1.4110441207885742, |
|
"learning_rate": 0.0004008, |
|
"loss": 1.5065, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 1.0768, |
|
"grad_norm": 1.0681021213531494, |
|
"learning_rate": 0.0003996, |
|
"loss": 1.5376, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 1.0784, |
|
"grad_norm": 1.355599045753479, |
|
"learning_rate": 0.0003984, |
|
"loss": 1.7659, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 1.245712161064148, |
|
"learning_rate": 0.0003972, |
|
"loss": 1.6335, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 1.0816, |
|
"grad_norm": 10.161030769348145, |
|
"learning_rate": 0.000396, |
|
"loss": 2.7562, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 1.0832, |
|
"grad_norm": 5.946995735168457, |
|
"learning_rate": 0.0003948, |
|
"loss": 2.2905, |
|
"step": 677 |
|
}, |
|
{ |
|
"epoch": 1.0848, |
|
"grad_norm": 3.535452365875244, |
|
"learning_rate": 0.00039359999999999997, |
|
"loss": 1.8719, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 1.0864, |
|
"grad_norm": 1.3516403436660767, |
|
"learning_rate": 0.0003924, |
|
"loss": 1.3978, |
|
"step": 679 |
|
}, |
|
{ |
|
"epoch": 1.088, |
|
"grad_norm": 1.6472234725952148, |
|
"learning_rate": 0.00039119999999999997, |
|
"loss": 1.2273, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 1.0896, |
|
"grad_norm": 3.0599803924560547, |
|
"learning_rate": 0.00039, |
|
"loss": 1.4096, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 1.0912, |
|
"grad_norm": 3.078174352645874, |
|
"learning_rate": 0.00038879999999999996, |
|
"loss": 1.2859, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 1.0928, |
|
"grad_norm": 3.5751090049743652, |
|
"learning_rate": 0.0003876, |
|
"loss": 1.4634, |
|
"step": 683 |
|
}, |
|
{ |
|
"epoch": 1.0944, |
|
"grad_norm": 2.058401584625244, |
|
"learning_rate": 0.00038639999999999996, |
|
"loss": 1.6292, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 1.096, |
|
"grad_norm": 1.8631670475006104, |
|
"learning_rate": 0.0003852, |
|
"loss": 1.1733, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 1.0976, |
|
"grad_norm": 1.9622036218643188, |
|
"learning_rate": 0.00038399999999999996, |
|
"loss": 1.0265, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 1.0992, |
|
"grad_norm": 1.0497528314590454, |
|
"learning_rate": 0.0003828, |
|
"loss": 0.9263, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 1.1008, |
|
"grad_norm": 1.0042531490325928, |
|
"learning_rate": 0.00038159999999999995, |
|
"loss": 1.0379, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 1.1024, |
|
"grad_norm": 0.7379323244094849, |
|
"learning_rate": 0.0003804, |
|
"loss": 0.8909, |
|
"step": 689 |
|
}, |
|
{ |
|
"epoch": 1.104, |
|
"grad_norm": 0.9978313446044922, |
|
"learning_rate": 0.00037919999999999995, |
|
"loss": 0.9649, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 1.1056, |
|
"grad_norm": 1.4265302419662476, |
|
"learning_rate": 0.00037799999999999997, |
|
"loss": 0.8303, |
|
"step": 691 |
|
}, |
|
{ |
|
"epoch": 1.1072, |
|
"grad_norm": 1.609402060508728, |
|
"learning_rate": 0.00037679999999999994, |
|
"loss": 1.0412, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 1.1088, |
|
"grad_norm": 1.7801131010055542, |
|
"learning_rate": 0.00037559999999999997, |
|
"loss": 1.2652, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 1.1104, |
|
"grad_norm": 0.9285919666290283, |
|
"learning_rate": 0.0003744, |
|
"loss": 1.2096, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 1.112, |
|
"grad_norm": 0.9512993693351746, |
|
"learning_rate": 0.00037319999999999996, |
|
"loss": 1.1913, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 1.1136, |
|
"grad_norm": 0.9945847392082214, |
|
"learning_rate": 0.000372, |
|
"loss": 1.1681, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 1.1152, |
|
"grad_norm": 0.8993807435035706, |
|
"learning_rate": 0.00037079999999999996, |
|
"loss": 1.2373, |
|
"step": 697 |
|
}, |
|
{ |
|
"epoch": 1.1168, |
|
"grad_norm": 1.0664983987808228, |
|
"learning_rate": 0.0003696, |
|
"loss": 1.0773, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 1.1184, |
|
"grad_norm": 1.7917791604995728, |
|
"learning_rate": 0.00036839999999999996, |
|
"loss": 1.4875, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 0.8973643183708191, |
|
"learning_rate": 0.0003672, |
|
"loss": 1.1206, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.1216, |
|
"grad_norm": 1.2886801958084106, |
|
"learning_rate": 0.00036599999999999995, |
|
"loss": 0.9865, |
|
"step": 701 |
|
}, |
|
{ |
|
"epoch": 1.1232, |
|
"grad_norm": 1.4595153331756592, |
|
"learning_rate": 0.0003648, |
|
"loss": 1.388, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 1.1248, |
|
"grad_norm": 1.1467390060424805, |
|
"learning_rate": 0.00036359999999999995, |
|
"loss": 0.9407, |
|
"step": 703 |
|
}, |
|
{ |
|
"epoch": 1.1264, |
|
"grad_norm": 1.0743904113769531, |
|
"learning_rate": 0.00036239999999999997, |
|
"loss": 1.0897, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 1.1280000000000001, |
|
"grad_norm": 0.9008836150169373, |
|
"learning_rate": 0.00036119999999999994, |
|
"loss": 0.9555, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 1.1296, |
|
"grad_norm": 1.016830325126648, |
|
"learning_rate": 0.00035999999999999997, |
|
"loss": 1.0143, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 1.1312, |
|
"grad_norm": 1.8725807666778564, |
|
"learning_rate": 0.00035879999999999994, |
|
"loss": 1.4681, |
|
"step": 707 |
|
}, |
|
{ |
|
"epoch": 1.1328, |
|
"grad_norm": 1.0106738805770874, |
|
"learning_rate": 0.00035759999999999996, |
|
"loss": 1.0107, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 1.1344, |
|
"grad_norm": 1.0680439472198486, |
|
"learning_rate": 0.00035639999999999994, |
|
"loss": 0.8917, |
|
"step": 709 |
|
}, |
|
{ |
|
"epoch": 1.1360000000000001, |
|
"grad_norm": 0.9104785323143005, |
|
"learning_rate": 0.00035519999999999996, |
|
"loss": 1.1333, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 1.1376, |
|
"grad_norm": 1.0638065338134766, |
|
"learning_rate": 0.00035399999999999993, |
|
"loss": 0.9465, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 1.1392, |
|
"grad_norm": 0.7474643588066101, |
|
"learning_rate": 0.00035279999999999996, |
|
"loss": 0.7617, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 1.1408, |
|
"grad_norm": 0.8218055963516235, |
|
"learning_rate": 0.0003515999999999999, |
|
"loss": 1.0069, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 1.1424, |
|
"grad_norm": 1.1213569641113281, |
|
"learning_rate": 0.00035039999999999995, |
|
"loss": 1.0228, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 1.144, |
|
"grad_norm": 1.0540097951889038, |
|
"learning_rate": 0.0003491999999999999, |
|
"loss": 1.2662, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 1.1456, |
|
"grad_norm": 1.019623875617981, |
|
"learning_rate": 0.00034799999999999995, |
|
"loss": 1.2725, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 1.1472, |
|
"grad_norm": 0.9017633199691772, |
|
"learning_rate": 0.0003467999999999999, |
|
"loss": 0.8269, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 1.1488, |
|
"grad_norm": 3.7017982006073, |
|
"learning_rate": 0.00034559999999999994, |
|
"loss": 2.0837, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 1.1504, |
|
"grad_norm": 1.0694856643676758, |
|
"learning_rate": 0.00034439999999999997, |
|
"loss": 1.0984, |
|
"step": 719 |
|
}, |
|
{ |
|
"epoch": 1.152, |
|
"grad_norm": 1.0227575302124023, |
|
"learning_rate": 0.00034319999999999994, |
|
"loss": 1.1509, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 1.1536, |
|
"grad_norm": 0.8948163390159607, |
|
"learning_rate": 0.00034199999999999996, |
|
"loss": 0.8165, |
|
"step": 721 |
|
}, |
|
{ |
|
"epoch": 1.1552, |
|
"grad_norm": 1.575053334236145, |
|
"learning_rate": 0.00034079999999999994, |
|
"loss": 1.4505, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 1.1568, |
|
"grad_norm": 1.6160234212875366, |
|
"learning_rate": 0.00033959999999999996, |
|
"loss": 1.5606, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 1.1584, |
|
"grad_norm": 1.469820499420166, |
|
"learning_rate": 0.00033839999999999993, |
|
"loss": 1.3345, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 2.6582064628601074, |
|
"learning_rate": 0.0003372, |
|
"loss": 1.9885, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 1.1616, |
|
"grad_norm": 8.827315330505371, |
|
"learning_rate": 0.000336, |
|
"loss": 2.7075, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 1.1632, |
|
"grad_norm": 5.201417922973633, |
|
"learning_rate": 0.0003348, |
|
"loss": 2.0887, |
|
"step": 727 |
|
}, |
|
{ |
|
"epoch": 1.1648, |
|
"grad_norm": 2.5593836307525635, |
|
"learning_rate": 0.0003336, |
|
"loss": 1.3964, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 1.1663999999999999, |
|
"grad_norm": 2.476527452468872, |
|
"learning_rate": 0.0003324, |
|
"loss": 1.6703, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 1.168, |
|
"grad_norm": 1.3854165077209473, |
|
"learning_rate": 0.0003312, |
|
"loss": 1.4116, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 1.1696, |
|
"grad_norm": 1.7695822715759277, |
|
"learning_rate": 0.00033, |
|
"loss": 1.7663, |
|
"step": 731 |
|
}, |
|
{ |
|
"epoch": 1.1712, |
|
"grad_norm": 1.7809518575668335, |
|
"learning_rate": 0.0003288, |
|
"loss": 2.2482, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 1.1728, |
|
"grad_norm": 1.5759507417678833, |
|
"learning_rate": 0.0003276, |
|
"loss": 1.0389, |
|
"step": 733 |
|
}, |
|
{ |
|
"epoch": 1.1743999999999999, |
|
"grad_norm": 3.968517780303955, |
|
"learning_rate": 0.0003264, |
|
"loss": 1.1723, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 1.176, |
|
"grad_norm": 3.551710367202759, |
|
"learning_rate": 0.0003252, |
|
"loss": 1.1122, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 1.1776, |
|
"grad_norm": 3.917438507080078, |
|
"learning_rate": 0.000324, |
|
"loss": 1.3001, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 1.1792, |
|
"grad_norm": 2.960092306137085, |
|
"learning_rate": 0.0003228, |
|
"loss": 1.3047, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 1.1808, |
|
"grad_norm": 2.588700532913208, |
|
"learning_rate": 0.0003216, |
|
"loss": 1.0372, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 1.1824, |
|
"grad_norm": 2.2377519607543945, |
|
"learning_rate": 0.0003204, |
|
"loss": 1.2321, |
|
"step": 739 |
|
}, |
|
{ |
|
"epoch": 1.184, |
|
"grad_norm": 1.3439960479736328, |
|
"learning_rate": 0.0003192, |
|
"loss": 1.2945, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 1.1856, |
|
"grad_norm": 0.8878504037857056, |
|
"learning_rate": 0.000318, |
|
"loss": 1.085, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 1.1872, |
|
"grad_norm": 0.8282541632652283, |
|
"learning_rate": 0.0003168, |
|
"loss": 1.1087, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 1.1888, |
|
"grad_norm": 0.8017875552177429, |
|
"learning_rate": 0.0003156, |
|
"loss": 0.9878, |
|
"step": 743 |
|
}, |
|
{ |
|
"epoch": 1.1904, |
|
"grad_norm": 1.347516655921936, |
|
"learning_rate": 0.0003144, |
|
"loss": 1.0613, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 1.192, |
|
"grad_norm": 1.9655871391296387, |
|
"learning_rate": 0.00031319999999999997, |
|
"loss": 1.0696, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 1.1936, |
|
"grad_norm": 1.588913083076477, |
|
"learning_rate": 0.000312, |
|
"loss": 1.2005, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 1.1952, |
|
"grad_norm": 1.29644775390625, |
|
"learning_rate": 0.00031079999999999997, |
|
"loss": 0.904, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 1.1968, |
|
"grad_norm": 2.8275885581970215, |
|
"learning_rate": 0.0003096, |
|
"loss": 1.5441, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 1.1984, |
|
"grad_norm": 1.0542739629745483, |
|
"learning_rate": 0.00030839999999999996, |
|
"loss": 0.9999, |
|
"step": 749 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 0.7280116677284241, |
|
"learning_rate": 0.0003072, |
|
"loss": 0.8393, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 1.2016, |
|
"grad_norm": 0.7934659123420715, |
|
"learning_rate": 0.00030599999999999996, |
|
"loss": 1.1939, |
|
"step": 751 |
|
}, |
|
{ |
|
"epoch": 1.2032, |
|
"grad_norm": 0.8406733274459839, |
|
"learning_rate": 0.0003048, |
|
"loss": 0.9562, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 1.2048, |
|
"grad_norm": 1.0488728284835815, |
|
"learning_rate": 0.00030359999999999995, |
|
"loss": 1.2521, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 1.2064, |
|
"grad_norm": 0.8879828453063965, |
|
"learning_rate": 0.0003024, |
|
"loss": 0.7971, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 1.208, |
|
"grad_norm": 1.3260725736618042, |
|
"learning_rate": 0.00030119999999999995, |
|
"loss": 1.2188, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 1.2096, |
|
"grad_norm": 1.1654318571090698, |
|
"learning_rate": 0.0003, |
|
"loss": 0.9387, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 1.2112, |
|
"grad_norm": 1.351473093032837, |
|
"learning_rate": 0.0002988, |
|
"loss": 1.0464, |
|
"step": 757 |
|
}, |
|
{ |
|
"epoch": 1.2128, |
|
"grad_norm": 0.9511071443557739, |
|
"learning_rate": 0.00029759999999999997, |
|
"loss": 0.9436, |
|
"step": 758 |
|
}, |
|
{ |
|
"epoch": 1.2144, |
|
"grad_norm": 1.179603934288025, |
|
"learning_rate": 0.0002964, |
|
"loss": 1.1701, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 1.216, |
|
"grad_norm": 0.8080942034721375, |
|
"learning_rate": 0.00029519999999999997, |
|
"loss": 1.3787, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 1.2176, |
|
"grad_norm": 0.7412335872650146, |
|
"learning_rate": 0.000294, |
|
"loss": 1.2793, |
|
"step": 761 |
|
}, |
|
{ |
|
"epoch": 1.2192, |
|
"grad_norm": 0.9035298824310303, |
|
"learning_rate": 0.00029279999999999996, |
|
"loss": 1.117, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 1.2208, |
|
"grad_norm": 1.026508092880249, |
|
"learning_rate": 0.0002916, |
|
"loss": 1.0756, |
|
"step": 763 |
|
}, |
|
{ |
|
"epoch": 1.2224, |
|
"grad_norm": 1.2814981937408447, |
|
"learning_rate": 0.00029039999999999996, |
|
"loss": 1.0611, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 1.224, |
|
"grad_norm": 1.475760579109192, |
|
"learning_rate": 0.0002892, |
|
"loss": 1.1253, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 1.2256, |
|
"grad_norm": 1.2571303844451904, |
|
"learning_rate": 0.00028799999999999995, |
|
"loss": 1.585, |
|
"step": 766 |
|
}, |
|
{ |
|
"epoch": 1.2272, |
|
"grad_norm": 1.2124806642532349, |
|
"learning_rate": 0.0002868, |
|
"loss": 1.3838, |
|
"step": 767 |
|
}, |
|
{ |
|
"epoch": 1.2288000000000001, |
|
"grad_norm": 2.581066131591797, |
|
"learning_rate": 0.00028559999999999995, |
|
"loss": 1.7297, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 1.2304, |
|
"grad_norm": 1.0715489387512207, |
|
"learning_rate": 0.0002844, |
|
"loss": 1.0821, |
|
"step": 769 |
|
}, |
|
{ |
|
"epoch": 1.232, |
|
"grad_norm": 1.242422342300415, |
|
"learning_rate": 0.00028319999999999994, |
|
"loss": 1.5755, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 1.2336, |
|
"grad_norm": 1.0124776363372803, |
|
"learning_rate": 0.00028199999999999997, |
|
"loss": 1.4446, |
|
"step": 771 |
|
}, |
|
{ |
|
"epoch": 1.2352, |
|
"grad_norm": 1.3531243801116943, |
|
"learning_rate": 0.0002808, |
|
"loss": 1.1064, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 1.2368000000000001, |
|
"grad_norm": 2.615983724594116, |
|
"learning_rate": 0.00027959999999999997, |
|
"loss": 1.9305, |
|
"step": 773 |
|
}, |
|
{ |
|
"epoch": 1.2384, |
|
"grad_norm": 1.1576447486877441, |
|
"learning_rate": 0.0002784, |
|
"loss": 1.2717, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 1.79608154296875, |
|
"learning_rate": 0.0002772, |
|
"loss": 1.9539, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 1.2416, |
|
"grad_norm": 14.315869331359863, |
|
"learning_rate": 0.000276, |
|
"loss": 3.4927, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 1.2432, |
|
"grad_norm": 4.236233234405518, |
|
"learning_rate": 0.0002748, |
|
"loss": 1.8021, |
|
"step": 777 |
|
}, |
|
{ |
|
"epoch": 1.2448, |
|
"grad_norm": 1.4815818071365356, |
|
"learning_rate": 0.0002736, |
|
"loss": 1.5814, |
|
"step": 778 |
|
}, |
|
{ |
|
"epoch": 1.2464, |
|
"grad_norm": 3.584576368331909, |
|
"learning_rate": 0.0002724, |
|
"loss": 2.2016, |
|
"step": 779 |
|
}, |
|
{ |
|
"epoch": 1.248, |
|
"grad_norm": 5.444706916809082, |
|
"learning_rate": 0.0002712, |
|
"loss": 2.1992, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 1.2496, |
|
"grad_norm": 1.9935747385025024, |
|
"learning_rate": 0.00027, |
|
"loss": 1.603, |
|
"step": 781 |
|
}, |
|
{ |
|
"epoch": 1.2511999999999999, |
|
"grad_norm": 1.7511041164398193, |
|
"learning_rate": 0.0002688, |
|
"loss": 1.2715, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 1.2528000000000001, |
|
"grad_norm": 3.652242422103882, |
|
"learning_rate": 0.0002676, |
|
"loss": 1.2038, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 1.2544, |
|
"grad_norm": 4.156230449676514, |
|
"learning_rate": 0.00026639999999999997, |
|
"loss": 1.1785, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 1.256, |
|
"grad_norm": 3.3570737838745117, |
|
"learning_rate": 0.0002652, |
|
"loss": 1.2539, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 1.2576, |
|
"grad_norm": 2.836935043334961, |
|
"learning_rate": 0.00026399999999999997, |
|
"loss": 1.1734, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 1.2591999999999999, |
|
"grad_norm": 2.9423153400421143, |
|
"learning_rate": 0.0002628, |
|
"loss": 1.2193, |
|
"step": 787 |
|
}, |
|
{ |
|
"epoch": 1.2608, |
|
"grad_norm": 2.4489452838897705, |
|
"learning_rate": 0.00026159999999999996, |
|
"loss": 1.291, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 1.2624, |
|
"grad_norm": 1.2637799978256226, |
|
"learning_rate": 0.0002604, |
|
"loss": 0.8955, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 1.264, |
|
"grad_norm": 0.8530015349388123, |
|
"learning_rate": 0.00025919999999999996, |
|
"loss": 1.0469, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 1.2656, |
|
"grad_norm": 0.9832066297531128, |
|
"learning_rate": 0.000258, |
|
"loss": 0.8632, |
|
"step": 791 |
|
}, |
|
{ |
|
"epoch": 1.2671999999999999, |
|
"grad_norm": 0.5436220765113831, |
|
"learning_rate": 0.00025679999999999995, |
|
"loss": 0.9851, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 1.2688, |
|
"grad_norm": 1.093661904335022, |
|
"learning_rate": 0.0002556, |
|
"loss": 1.4339, |
|
"step": 793 |
|
}, |
|
{ |
|
"epoch": 1.2704, |
|
"grad_norm": 0.7037041187286377, |
|
"learning_rate": 0.00025439999999999995, |
|
"loss": 1.1739, |
|
"step": 794 |
|
}, |
|
{ |
|
"epoch": 1.272, |
|
"grad_norm": 0.6901881098747253, |
|
"learning_rate": 0.0002532, |
|
"loss": 0.9546, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 1.2736, |
|
"grad_norm": 1.5446933507919312, |
|
"learning_rate": 0.00025199999999999995, |
|
"loss": 1.336, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 1.2752, |
|
"grad_norm": 1.7216750383377075, |
|
"learning_rate": 0.00025079999999999997, |
|
"loss": 0.8747, |
|
"step": 797 |
|
}, |
|
{ |
|
"epoch": 1.2768, |
|
"grad_norm": 1.2335959672927856, |
|
"learning_rate": 0.00024959999999999994, |
|
"loss": 0.969, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 1.2784, |
|
"grad_norm": 1.483788251876831, |
|
"learning_rate": 0.00024839999999999997, |
|
"loss": 0.8357, |
|
"step": 799 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 2.874842882156372, |
|
"learning_rate": 0.0002472, |
|
"loss": 1.2983, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.2816, |
|
"grad_norm": 0.7263085246086121, |
|
"learning_rate": 0.00024599999999999996, |
|
"loss": 1.2608, |
|
"step": 801 |
|
}, |
|
{ |
|
"epoch": 1.2832, |
|
"grad_norm": 1.0308623313903809, |
|
"learning_rate": 0.0002448, |
|
"loss": 1.117, |
|
"step": 802 |
|
}, |
|
{ |
|
"epoch": 1.2848, |
|
"grad_norm": 0.9009158611297607, |
|
"learning_rate": 0.00024359999999999999, |
|
"loss": 1.0626, |
|
"step": 803 |
|
}, |
|
{ |
|
"epoch": 1.2864, |
|
"grad_norm": 1.3145204782485962, |
|
"learning_rate": 0.00024239999999999998, |
|
"loss": 0.9173, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 1.288, |
|
"grad_norm": 1.034488320350647, |
|
"learning_rate": 0.00024119999999999998, |
|
"loss": 1.1901, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 1.2896, |
|
"grad_norm": 0.8789196610450745, |
|
"learning_rate": 0.00023999999999999998, |
|
"loss": 0.6968, |
|
"step": 806 |
|
}, |
|
{ |
|
"epoch": 1.2912, |
|
"grad_norm": 0.9626047015190125, |
|
"learning_rate": 0.0002388, |
|
"loss": 1.0333, |
|
"step": 807 |
|
}, |
|
{ |
|
"epoch": 1.2928, |
|
"grad_norm": 0.7256068587303162, |
|
"learning_rate": 0.0002376, |
|
"loss": 0.8452, |
|
"step": 808 |
|
}, |
|
{ |
|
"epoch": 1.2944, |
|
"grad_norm": 0.8389018774032593, |
|
"learning_rate": 0.0002364, |
|
"loss": 1.2869, |
|
"step": 809 |
|
}, |
|
{ |
|
"epoch": 1.296, |
|
"grad_norm": 0.7672526836395264, |
|
"learning_rate": 0.0002352, |
|
"loss": 1.1781, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 1.2976, |
|
"grad_norm": 0.9246567487716675, |
|
"learning_rate": 0.000234, |
|
"loss": 1.2095, |
|
"step": 811 |
|
}, |
|
{ |
|
"epoch": 1.2992, |
|
"grad_norm": 0.7383604049682617, |
|
"learning_rate": 0.0002328, |
|
"loss": 1.2276, |
|
"step": 812 |
|
}, |
|
{ |
|
"epoch": 1.3008, |
|
"grad_norm": 1.0119376182556152, |
|
"learning_rate": 0.0002316, |
|
"loss": 1.0281, |
|
"step": 813 |
|
}, |
|
{ |
|
"epoch": 1.3024, |
|
"grad_norm": 1.0825129747390747, |
|
"learning_rate": 0.0002304, |
|
"loss": 1.3149, |
|
"step": 814 |
|
}, |
|
{ |
|
"epoch": 1.304, |
|
"grad_norm": 0.8355119824409485, |
|
"learning_rate": 0.0002292, |
|
"loss": 1.2918, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 1.3056, |
|
"grad_norm": 0.6260655522346497, |
|
"learning_rate": 0.00022799999999999999, |
|
"loss": 0.936, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 1.3072, |
|
"grad_norm": 1.1685816049575806, |
|
"learning_rate": 0.00022679999999999998, |
|
"loss": 1.2178, |
|
"step": 817 |
|
}, |
|
{ |
|
"epoch": 1.3088, |
|
"grad_norm": 1.188957691192627, |
|
"learning_rate": 0.00022559999999999998, |
|
"loss": 1.1981, |
|
"step": 818 |
|
}, |
|
{ |
|
"epoch": 1.3104, |
|
"grad_norm": 1.8030787706375122, |
|
"learning_rate": 0.00022439999999999998, |
|
"loss": 1.247, |
|
"step": 819 |
|
}, |
|
{ |
|
"epoch": 1.312, |
|
"grad_norm": 1.1914401054382324, |
|
"learning_rate": 0.00022319999999999998, |
|
"loss": 1.0994, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 1.3136, |
|
"grad_norm": 1.3656498193740845, |
|
"learning_rate": 0.00022199999999999998, |
|
"loss": 1.3633, |
|
"step": 821 |
|
}, |
|
{ |
|
"epoch": 1.3152, |
|
"grad_norm": 1.4209223985671997, |
|
"learning_rate": 0.00022079999999999997, |
|
"loss": 1.2058, |
|
"step": 822 |
|
}, |
|
{ |
|
"epoch": 1.3168, |
|
"grad_norm": 1.3564426898956299, |
|
"learning_rate": 0.00021959999999999997, |
|
"loss": 1.6574, |
|
"step": 823 |
|
}, |
|
{ |
|
"epoch": 1.3184, |
|
"grad_norm": 2.096992015838623, |
|
"learning_rate": 0.00021839999999999997, |
|
"loss": 1.9215, |
|
"step": 824 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 1.984554409980774, |
|
"learning_rate": 0.00021719999999999997, |
|
"loss": 1.8319, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 1.3216, |
|
"grad_norm": 24.8841495513916, |
|
"learning_rate": 0.00021599999999999996, |
|
"loss": 4.94, |
|
"step": 826 |
|
}, |
|
{ |
|
"epoch": 1.3232, |
|
"grad_norm": 3.756709575653076, |
|
"learning_rate": 0.00021479999999999996, |
|
"loss": 1.9233, |
|
"step": 827 |
|
}, |
|
{ |
|
"epoch": 1.3248, |
|
"grad_norm": 1.6268775463104248, |
|
"learning_rate": 0.00021359999999999996, |
|
"loss": 1.2712, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 1.3264, |
|
"grad_norm": 3.466218948364258, |
|
"learning_rate": 0.00021239999999999996, |
|
"loss": 1.7201, |
|
"step": 829 |
|
}, |
|
{ |
|
"epoch": 1.328, |
|
"grad_norm": 1.621490716934204, |
|
"learning_rate": 0.00021119999999999996, |
|
"loss": 1.3709, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 1.3296000000000001, |
|
"grad_norm": 1.1984294652938843, |
|
"learning_rate": 0.00020999999999999998, |
|
"loss": 1.4106, |
|
"step": 831 |
|
}, |
|
{ |
|
"epoch": 1.3312, |
|
"grad_norm": 2.042137384414673, |
|
"learning_rate": 0.00020879999999999998, |
|
"loss": 1.1315, |
|
"step": 832 |
|
}, |
|
{ |
|
"epoch": 1.3328, |
|
"grad_norm": 1.9946519136428833, |
|
"learning_rate": 0.00020759999999999998, |
|
"loss": 1.0351, |
|
"step": 833 |
|
}, |
|
{ |
|
"epoch": 1.3344, |
|
"grad_norm": 2.1006603240966797, |
|
"learning_rate": 0.00020639999999999998, |
|
"loss": 1.2888, |
|
"step": 834 |
|
}, |
|
{ |
|
"epoch": 1.336, |
|
"grad_norm": 1.6302366256713867, |
|
"learning_rate": 0.0002052, |
|
"loss": 1.224, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 1.3376000000000001, |
|
"grad_norm": 2.526308059692383, |
|
"learning_rate": 0.000204, |
|
"loss": 1.0775, |
|
"step": 836 |
|
}, |
|
{ |
|
"epoch": 1.3392, |
|
"grad_norm": 2.1584436893463135, |
|
"learning_rate": 0.0002028, |
|
"loss": 1.2341, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 1.3408, |
|
"grad_norm": 1.6390632390975952, |
|
"learning_rate": 0.0002016, |
|
"loss": 1.0818, |
|
"step": 838 |
|
}, |
|
{ |
|
"epoch": 1.3424, |
|
"grad_norm": 1.6563626527786255, |
|
"learning_rate": 0.0002004, |
|
"loss": 0.9795, |
|
"step": 839 |
|
}, |
|
{ |
|
"epoch": 1.3439999999999999, |
|
"grad_norm": 0.9004166126251221, |
|
"learning_rate": 0.0001992, |
|
"loss": 1.0898, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 1.3456000000000001, |
|
"grad_norm": 0.8950443863868713, |
|
"learning_rate": 0.000198, |
|
"loss": 0.9329, |
|
"step": 841 |
|
}, |
|
{ |
|
"epoch": 1.3472, |
|
"grad_norm": 0.8782767057418823, |
|
"learning_rate": 0.00019679999999999999, |
|
"loss": 0.9925, |
|
"step": 842 |
|
}, |
|
{ |
|
"epoch": 1.3488, |
|
"grad_norm": 0.851058304309845, |
|
"learning_rate": 0.00019559999999999998, |
|
"loss": 1.0908, |
|
"step": 843 |
|
}, |
|
{ |
|
"epoch": 1.3504, |
|
"grad_norm": 5.08306360244751, |
|
"learning_rate": 0.00019439999999999998, |
|
"loss": 1.6817, |
|
"step": 844 |
|
}, |
|
{ |
|
"epoch": 1.3519999999999999, |
|
"grad_norm": 1.4701097011566162, |
|
"learning_rate": 0.00019319999999999998, |
|
"loss": 0.9395, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 1.3536000000000001, |
|
"grad_norm": 0.7278363704681396, |
|
"learning_rate": 0.00019199999999999998, |
|
"loss": 1.0193, |
|
"step": 846 |
|
}, |
|
{ |
|
"epoch": 1.3552, |
|
"grad_norm": 0.9397822618484497, |
|
"learning_rate": 0.00019079999999999998, |
|
"loss": 0.7025, |
|
"step": 847 |
|
}, |
|
{ |
|
"epoch": 1.3568, |
|
"grad_norm": 1.3745059967041016, |
|
"learning_rate": 0.00018959999999999997, |
|
"loss": 0.9564, |
|
"step": 848 |
|
}, |
|
{ |
|
"epoch": 1.3584, |
|
"grad_norm": 1.193011999130249, |
|
"learning_rate": 0.00018839999999999997, |
|
"loss": 1.1551, |
|
"step": 849 |
|
}, |
|
{ |
|
"epoch": 1.3599999999999999, |
|
"grad_norm": 0.889456570148468, |
|
"learning_rate": 0.0001872, |
|
"loss": 0.9999, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 1.3616, |
|
"grad_norm": 1.0669219493865967, |
|
"learning_rate": 0.000186, |
|
"loss": 0.8591, |
|
"step": 851 |
|
}, |
|
{ |
|
"epoch": 1.3632, |
|
"grad_norm": 0.9589920043945312, |
|
"learning_rate": 0.0001848, |
|
"loss": 0.9345, |
|
"step": 852 |
|
}, |
|
{ |
|
"epoch": 1.3648, |
|
"grad_norm": 0.7081511616706848, |
|
"learning_rate": 0.0001836, |
|
"loss": 0.7226, |
|
"step": 853 |
|
}, |
|
{ |
|
"epoch": 1.3664, |
|
"grad_norm": 0.7487833499908447, |
|
"learning_rate": 0.0001824, |
|
"loss": 1.0986, |
|
"step": 854 |
|
}, |
|
{ |
|
"epoch": 1.3679999999999999, |
|
"grad_norm": 0.9651376605033875, |
|
"learning_rate": 0.00018119999999999999, |
|
"loss": 0.8951, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 1.3696, |
|
"grad_norm": 0.9545527100563049, |
|
"learning_rate": 0.00017999999999999998, |
|
"loss": 1.0129, |
|
"step": 856 |
|
}, |
|
{ |
|
"epoch": 1.3712, |
|
"grad_norm": 1.016617774963379, |
|
"learning_rate": 0.00017879999999999998, |
|
"loss": 1.0022, |
|
"step": 857 |
|
}, |
|
{ |
|
"epoch": 1.3728, |
|
"grad_norm": 1.4556984901428223, |
|
"learning_rate": 0.00017759999999999998, |
|
"loss": 1.317, |
|
"step": 858 |
|
}, |
|
{ |
|
"epoch": 1.3744, |
|
"grad_norm": 0.790810763835907, |
|
"learning_rate": 0.00017639999999999998, |
|
"loss": 1.2533, |
|
"step": 859 |
|
}, |
|
{ |
|
"epoch": 1.376, |
|
"grad_norm": 0.9381358027458191, |
|
"learning_rate": 0.00017519999999999998, |
|
"loss": 0.9896, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 1.3776, |
|
"grad_norm": 0.8054640889167786, |
|
"learning_rate": 0.00017399999999999997, |
|
"loss": 0.9845, |
|
"step": 861 |
|
}, |
|
{ |
|
"epoch": 1.3792, |
|
"grad_norm": 0.8973929286003113, |
|
"learning_rate": 0.00017279999999999997, |
|
"loss": 1.4839, |
|
"step": 862 |
|
}, |
|
{ |
|
"epoch": 1.3808, |
|
"grad_norm": 1.5383504629135132, |
|
"learning_rate": 0.00017159999999999997, |
|
"loss": 1.0332, |
|
"step": 863 |
|
}, |
|
{ |
|
"epoch": 1.3824, |
|
"grad_norm": 0.842438817024231, |
|
"learning_rate": 0.00017039999999999997, |
|
"loss": 1.3733, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 1.384, |
|
"grad_norm": 1.3562208414077759, |
|
"learning_rate": 0.00016919999999999997, |
|
"loss": 1.3423, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 1.3856, |
|
"grad_norm": 0.8287788033485413, |
|
"learning_rate": 0.000168, |
|
"loss": 1.0022, |
|
"step": 866 |
|
}, |
|
{ |
|
"epoch": 1.3872, |
|
"grad_norm": 1.4219720363616943, |
|
"learning_rate": 0.0001668, |
|
"loss": 1.2173, |
|
"step": 867 |
|
}, |
|
{ |
|
"epoch": 1.3888, |
|
"grad_norm": 1.0144832134246826, |
|
"learning_rate": 0.0001656, |
|
"loss": 1.3067, |
|
"step": 868 |
|
}, |
|
{ |
|
"epoch": 1.3904, |
|
"grad_norm": 1.1086362600326538, |
|
"learning_rate": 0.0001644, |
|
"loss": 1.4451, |
|
"step": 869 |
|
}, |
|
{ |
|
"epoch": 1.392, |
|
"grad_norm": 1.1874333620071411, |
|
"learning_rate": 0.0001632, |
|
"loss": 1.4588, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 1.3936, |
|
"grad_norm": 1.2281416654586792, |
|
"learning_rate": 0.000162, |
|
"loss": 1.564, |
|
"step": 871 |
|
}, |
|
{ |
|
"epoch": 1.3952, |
|
"grad_norm": 1.024274468421936, |
|
"learning_rate": 0.0001608, |
|
"loss": 1.3052, |
|
"step": 872 |
|
}, |
|
{ |
|
"epoch": 1.3968, |
|
"grad_norm": 1.0952630043029785, |
|
"learning_rate": 0.0001596, |
|
"loss": 1.4589, |
|
"step": 873 |
|
}, |
|
{ |
|
"epoch": 1.3984, |
|
"grad_norm": 1.3578640222549438, |
|
"learning_rate": 0.0001584, |
|
"loss": 1.8693, |
|
"step": 874 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 1.3908575773239136, |
|
"learning_rate": 0.0001572, |
|
"loss": 2.0153, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 1.4016, |
|
"grad_norm": 3.5373663902282715, |
|
"learning_rate": 0.000156, |
|
"loss": 1.6021, |
|
"step": 876 |
|
}, |
|
{ |
|
"epoch": 1.4032, |
|
"grad_norm": 4.33758020401001, |
|
"learning_rate": 0.0001548, |
|
"loss": 1.6701, |
|
"step": 877 |
|
}, |
|
{ |
|
"epoch": 1.4048, |
|
"grad_norm": 0.7620241641998291, |
|
"learning_rate": 0.0001536, |
|
"loss": 1.1137, |
|
"step": 878 |
|
}, |
|
{ |
|
"epoch": 1.4064, |
|
"grad_norm": 3.931720733642578, |
|
"learning_rate": 0.0001524, |
|
"loss": 1.7832, |
|
"step": 879 |
|
}, |
|
{ |
|
"epoch": 1.408, |
|
"grad_norm": 4.786645412445068, |
|
"learning_rate": 0.0001512, |
|
"loss": 1.9583, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 1.4096, |
|
"grad_norm": 1.036521553993225, |
|
"learning_rate": 0.00015, |
|
"loss": 1.0819, |
|
"step": 881 |
|
}, |
|
{ |
|
"epoch": 1.4112, |
|
"grad_norm": 4.162290096282959, |
|
"learning_rate": 0.00014879999999999998, |
|
"loss": 2.0566, |
|
"step": 882 |
|
}, |
|
{ |
|
"epoch": 1.4128, |
|
"grad_norm": 2.359104633331299, |
|
"learning_rate": 0.00014759999999999998, |
|
"loss": 1.6004, |
|
"step": 883 |
|
}, |
|
{ |
|
"epoch": 1.4144, |
|
"grad_norm": 1.0020617246627808, |
|
"learning_rate": 0.00014639999999999998, |
|
"loss": 1.3709, |
|
"step": 884 |
|
}, |
|
{ |
|
"epoch": 1.416, |
|
"grad_norm": 1.8424478769302368, |
|
"learning_rate": 0.00014519999999999998, |
|
"loss": 1.2331, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 1.4176, |
|
"grad_norm": 1.1963813304901123, |
|
"learning_rate": 0.00014399999999999998, |
|
"loss": 1.0713, |
|
"step": 886 |
|
}, |
|
{ |
|
"epoch": 1.4192, |
|
"grad_norm": 1.5633505582809448, |
|
"learning_rate": 0.00014279999999999997, |
|
"loss": 1.0691, |
|
"step": 887 |
|
}, |
|
{ |
|
"epoch": 1.4208, |
|
"grad_norm": 1.1638294458389282, |
|
"learning_rate": 0.00014159999999999997, |
|
"loss": 0.9451, |
|
"step": 888 |
|
}, |
|
{ |
|
"epoch": 1.4224, |
|
"grad_norm": 1.7012614011764526, |
|
"learning_rate": 0.0001404, |
|
"loss": 0.8284, |
|
"step": 889 |
|
}, |
|
{ |
|
"epoch": 1.424, |
|
"grad_norm": 1.2200738191604614, |
|
"learning_rate": 0.0001392, |
|
"loss": 0.8185, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 1.4256, |
|
"grad_norm": 1.1766223907470703, |
|
"learning_rate": 0.000138, |
|
"loss": 0.9226, |
|
"step": 891 |
|
}, |
|
{ |
|
"epoch": 1.4272, |
|
"grad_norm": 0.9217773675918579, |
|
"learning_rate": 0.0001368, |
|
"loss": 0.8685, |
|
"step": 892 |
|
}, |
|
{ |
|
"epoch": 1.4288, |
|
"grad_norm": 0.8851170539855957, |
|
"learning_rate": 0.0001356, |
|
"loss": 0.7968, |
|
"step": 893 |
|
}, |
|
{ |
|
"epoch": 1.4304000000000001, |
|
"grad_norm": 0.8986585140228271, |
|
"learning_rate": 0.0001344, |
|
"loss": 0.9926, |
|
"step": 894 |
|
}, |
|
{ |
|
"epoch": 1.432, |
|
"grad_norm": 0.760133683681488, |
|
"learning_rate": 0.00013319999999999999, |
|
"loss": 0.761, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 1.4336, |
|
"grad_norm": 1.3854573965072632, |
|
"learning_rate": 0.00013199999999999998, |
|
"loss": 1.3453, |
|
"step": 896 |
|
}, |
|
{ |
|
"epoch": 1.4352, |
|
"grad_norm": 2.538731098175049, |
|
"learning_rate": 0.00013079999999999998, |
|
"loss": 1.405, |
|
"step": 897 |
|
}, |
|
{ |
|
"epoch": 1.4368, |
|
"grad_norm": 0.9111962914466858, |
|
"learning_rate": 0.00012959999999999998, |
|
"loss": 1.0013, |
|
"step": 898 |
|
}, |
|
{ |
|
"epoch": 1.4384000000000001, |
|
"grad_norm": 0.9045060873031616, |
|
"learning_rate": 0.00012839999999999998, |
|
"loss": 1.0312, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.8330265879631042, |
|
"learning_rate": 0.00012719999999999997, |
|
"loss": 1.1871, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.4416, |
|
"grad_norm": 0.8071021437644958, |
|
"learning_rate": 0.00012599999999999997, |
|
"loss": 1.4302, |
|
"step": 901 |
|
}, |
|
{ |
|
"epoch": 1.4432, |
|
"grad_norm": 1.5638083219528198, |
|
"learning_rate": 0.00012479999999999997, |
|
"loss": 1.2347, |
|
"step": 902 |
|
}, |
|
{ |
|
"epoch": 1.4447999999999999, |
|
"grad_norm": 1.674561619758606, |
|
"learning_rate": 0.0001236, |
|
"loss": 1.197, |
|
"step": 903 |
|
}, |
|
{ |
|
"epoch": 1.4464000000000001, |
|
"grad_norm": 1.1858874559402466, |
|
"learning_rate": 0.0001224, |
|
"loss": 0.8886, |
|
"step": 904 |
|
}, |
|
{ |
|
"epoch": 1.448, |
|
"grad_norm": 0.8216437101364136, |
|
"learning_rate": 0.00012119999999999999, |
|
"loss": 0.8166, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 1.4496, |
|
"grad_norm": 0.7633309364318848, |
|
"learning_rate": 0.00011999999999999999, |
|
"loss": 1.0883, |
|
"step": 906 |
|
}, |
|
{ |
|
"epoch": 1.4512, |
|
"grad_norm": 0.7902513742446899, |
|
"learning_rate": 0.0001188, |
|
"loss": 1.1526, |
|
"step": 907 |
|
}, |
|
{ |
|
"epoch": 1.4527999999999999, |
|
"grad_norm": 0.6996155381202698, |
|
"learning_rate": 0.0001176, |
|
"loss": 0.8232, |
|
"step": 908 |
|
}, |
|
{ |
|
"epoch": 1.4544000000000001, |
|
"grad_norm": 1.0348126888275146, |
|
"learning_rate": 0.0001164, |
|
"loss": 1.5505, |
|
"step": 909 |
|
}, |
|
{ |
|
"epoch": 1.456, |
|
"grad_norm": 0.9336599707603455, |
|
"learning_rate": 0.0001152, |
|
"loss": 1.1751, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 1.4576, |
|
"grad_norm": 0.8460116386413574, |
|
"learning_rate": 0.00011399999999999999, |
|
"loss": 1.1737, |
|
"step": 911 |
|
}, |
|
{ |
|
"epoch": 1.4592, |
|
"grad_norm": 0.8113738894462585, |
|
"learning_rate": 0.00011279999999999999, |
|
"loss": 1.0484, |
|
"step": 912 |
|
}, |
|
{ |
|
"epoch": 1.4607999999999999, |
|
"grad_norm": 0.8300177454948425, |
|
"learning_rate": 0.00011159999999999999, |
|
"loss": 0.7955, |
|
"step": 913 |
|
}, |
|
{ |
|
"epoch": 1.4624, |
|
"grad_norm": 1.2062090635299683, |
|
"learning_rate": 0.00011039999999999999, |
|
"loss": 1.3775, |
|
"step": 914 |
|
}, |
|
{ |
|
"epoch": 1.464, |
|
"grad_norm": 1.4255359172821045, |
|
"learning_rate": 0.00010919999999999998, |
|
"loss": 1.3335, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 1.4656, |
|
"grad_norm": 1.0286868810653687, |
|
"learning_rate": 0.00010799999999999998, |
|
"loss": 1.1146, |
|
"step": 916 |
|
}, |
|
{ |
|
"epoch": 1.4672, |
|
"grad_norm": 1.0137372016906738, |
|
"learning_rate": 0.00010679999999999998, |
|
"loss": 1.0296, |
|
"step": 917 |
|
}, |
|
{ |
|
"epoch": 1.4687999999999999, |
|
"grad_norm": 0.9231035113334656, |
|
"learning_rate": 0.00010559999999999998, |
|
"loss": 1.0707, |
|
"step": 918 |
|
}, |
|
{ |
|
"epoch": 1.4704, |
|
"grad_norm": 0.7874758839607239, |
|
"learning_rate": 0.00010439999999999999, |
|
"loss": 0.8588, |
|
"step": 919 |
|
}, |
|
{ |
|
"epoch": 1.472, |
|
"grad_norm": 0.9133099913597107, |
|
"learning_rate": 0.00010319999999999999, |
|
"loss": 1.3255, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 1.4736, |
|
"grad_norm": 1.030543565750122, |
|
"learning_rate": 0.000102, |
|
"loss": 1.5101, |
|
"step": 921 |
|
}, |
|
{ |
|
"epoch": 1.4752, |
|
"grad_norm": 0.9055063724517822, |
|
"learning_rate": 0.0001008, |
|
"loss": 0.9935, |
|
"step": 922 |
|
}, |
|
{ |
|
"epoch": 1.4768, |
|
"grad_norm": 1.4548567533493042, |
|
"learning_rate": 9.96e-05, |
|
"loss": 1.5182, |
|
"step": 923 |
|
}, |
|
{ |
|
"epoch": 1.4784, |
|
"grad_norm": 1.2355753183364868, |
|
"learning_rate": 9.839999999999999e-05, |
|
"loss": 1.5844, |
|
"step": 924 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 2.3780314922332764, |
|
"learning_rate": 9.719999999999999e-05, |
|
"loss": 1.9339, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 1.4816, |
|
"grad_norm": 1.8571350574493408, |
|
"learning_rate": 9.599999999999999e-05, |
|
"loss": 1.1195, |
|
"step": 926 |
|
}, |
|
{ |
|
"epoch": 1.4832, |
|
"grad_norm": 3.069366455078125, |
|
"learning_rate": 9.479999999999999e-05, |
|
"loss": 1.4374, |
|
"step": 927 |
|
}, |
|
{ |
|
"epoch": 1.4848, |
|
"grad_norm": 1.4385106563568115, |
|
"learning_rate": 9.36e-05, |
|
"loss": 1.0552, |
|
"step": 928 |
|
}, |
|
{ |
|
"epoch": 1.4864, |
|
"grad_norm": 2.345099449157715, |
|
"learning_rate": 9.24e-05, |
|
"loss": 1.7314, |
|
"step": 929 |
|
}, |
|
{ |
|
"epoch": 1.488, |
|
"grad_norm": 0.7494111657142639, |
|
"learning_rate": 9.12e-05, |
|
"loss": 0.8883, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 1.4896, |
|
"grad_norm": 0.8890445232391357, |
|
"learning_rate": 8.999999999999999e-05, |
|
"loss": 0.9971, |
|
"step": 931 |
|
}, |
|
{ |
|
"epoch": 1.4912, |
|
"grad_norm": 0.8567062616348267, |
|
"learning_rate": 8.879999999999999e-05, |
|
"loss": 0.9844, |
|
"step": 932 |
|
}, |
|
{ |
|
"epoch": 1.4928, |
|
"grad_norm": 2.1418657302856445, |
|
"learning_rate": 8.759999999999999e-05, |
|
"loss": 1.2059, |
|
"step": 933 |
|
}, |
|
{ |
|
"epoch": 1.4944, |
|
"grad_norm": 0.9829572439193726, |
|
"learning_rate": 8.639999999999999e-05, |
|
"loss": 1.2171, |
|
"step": 934 |
|
}, |
|
{ |
|
"epoch": 1.496, |
|
"grad_norm": 0.8156057000160217, |
|
"learning_rate": 8.519999999999998e-05, |
|
"loss": 0.9593, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 1.4976, |
|
"grad_norm": 1.4363347291946411, |
|
"learning_rate": 8.4e-05, |
|
"loss": 0.9601, |
|
"step": 936 |
|
}, |
|
{ |
|
"epoch": 1.4992, |
|
"grad_norm": 1.0625683069229126, |
|
"learning_rate": 8.28e-05, |
|
"loss": 1.0812, |
|
"step": 937 |
|
}, |
|
{ |
|
"epoch": 1.5008, |
|
"grad_norm": 0.9699941873550415, |
|
"learning_rate": 8.16e-05, |
|
"loss": 0.9473, |
|
"step": 938 |
|
}, |
|
{ |
|
"epoch": 1.5024, |
|
"grad_norm": 0.772868812084198, |
|
"learning_rate": 8.04e-05, |
|
"loss": 0.7998, |
|
"step": 939 |
|
}, |
|
{ |
|
"epoch": 1.504, |
|
"grad_norm": 0.9699268341064453, |
|
"learning_rate": 7.92e-05, |
|
"loss": 0.9162, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 1.5056, |
|
"grad_norm": 0.9714852571487427, |
|
"learning_rate": 7.8e-05, |
|
"loss": 0.7546, |
|
"step": 941 |
|
}, |
|
{ |
|
"epoch": 1.5072, |
|
"grad_norm": 0.7131572365760803, |
|
"learning_rate": 7.68e-05, |
|
"loss": 0.9188, |
|
"step": 942 |
|
}, |
|
{ |
|
"epoch": 1.5088, |
|
"grad_norm": 0.855786919593811, |
|
"learning_rate": 7.56e-05, |
|
"loss": 1.0208, |
|
"step": 943 |
|
}, |
|
{ |
|
"epoch": 1.5104, |
|
"grad_norm": 0.8761354684829712, |
|
"learning_rate": 7.439999999999999e-05, |
|
"loss": 0.9603, |
|
"step": 944 |
|
}, |
|
{ |
|
"epoch": 1.512, |
|
"grad_norm": 0.8056983351707458, |
|
"learning_rate": 7.319999999999999e-05, |
|
"loss": 1.0588, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 1.5135999999999998, |
|
"grad_norm": 1.0384901762008667, |
|
"learning_rate": 7.199999999999999e-05, |
|
"loss": 0.8915, |
|
"step": 946 |
|
}, |
|
{ |
|
"epoch": 1.5152, |
|
"grad_norm": 0.6739591956138611, |
|
"learning_rate": 7.079999999999999e-05, |
|
"loss": 0.7157, |
|
"step": 947 |
|
}, |
|
{ |
|
"epoch": 1.5168, |
|
"grad_norm": 0.7567741274833679, |
|
"learning_rate": 6.96e-05, |
|
"loss": 0.733, |
|
"step": 948 |
|
}, |
|
{ |
|
"epoch": 1.5184, |
|
"grad_norm": 1.2941476106643677, |
|
"learning_rate": 6.84e-05, |
|
"loss": 1.2629, |
|
"step": 949 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 0.7849064469337463, |
|
"learning_rate": 6.72e-05, |
|
"loss": 1.0687, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 1.5215999999999998, |
|
"grad_norm": 0.7167961001396179, |
|
"learning_rate": 6.599999999999999e-05, |
|
"loss": 0.9125, |
|
"step": 951 |
|
}, |
|
{ |
|
"epoch": 1.5232, |
|
"grad_norm": 1.4623581171035767, |
|
"learning_rate": 6.479999999999999e-05, |
|
"loss": 1.0419, |
|
"step": 952 |
|
}, |
|
{ |
|
"epoch": 1.5248, |
|
"grad_norm": 1.0260752439498901, |
|
"learning_rate": 6.359999999999999e-05, |
|
"loss": 0.759, |
|
"step": 953 |
|
}, |
|
{ |
|
"epoch": 1.5264, |
|
"grad_norm": 0.6854276657104492, |
|
"learning_rate": 6.239999999999999e-05, |
|
"loss": 1.1125, |
|
"step": 954 |
|
}, |
|
{ |
|
"epoch": 1.528, |
|
"grad_norm": 0.776077151298523, |
|
"learning_rate": 6.12e-05, |
|
"loss": 0.7962, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 1.5295999999999998, |
|
"grad_norm": 0.7898908257484436, |
|
"learning_rate": 5.9999999999999995e-05, |
|
"loss": 1.0421, |
|
"step": 956 |
|
}, |
|
{ |
|
"epoch": 1.5312000000000001, |
|
"grad_norm": 0.7632326483726501, |
|
"learning_rate": 5.88e-05, |
|
"loss": 0.8491, |
|
"step": 957 |
|
}, |
|
{ |
|
"epoch": 1.5328, |
|
"grad_norm": 0.7451068162918091, |
|
"learning_rate": 5.76e-05, |
|
"loss": 0.9336, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 1.5344, |
|
"grad_norm": 0.7700297832489014, |
|
"learning_rate": 5.6399999999999995e-05, |
|
"loss": 0.9, |
|
"step": 959 |
|
}, |
|
{ |
|
"epoch": 1.536, |
|
"grad_norm": 0.879389762878418, |
|
"learning_rate": 5.519999999999999e-05, |
|
"loss": 1.1537, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 1.5375999999999999, |
|
"grad_norm": 0.9828069806098938, |
|
"learning_rate": 5.399999999999999e-05, |
|
"loss": 0.9058, |
|
"step": 961 |
|
}, |
|
{ |
|
"epoch": 1.5392000000000001, |
|
"grad_norm": 0.9066524505615234, |
|
"learning_rate": 5.279999999999999e-05, |
|
"loss": 1.1331, |
|
"step": 962 |
|
}, |
|
{ |
|
"epoch": 1.5408, |
|
"grad_norm": 0.7387616038322449, |
|
"learning_rate": 5.1599999999999994e-05, |
|
"loss": 0.9451, |
|
"step": 963 |
|
}, |
|
{ |
|
"epoch": 1.5424, |
|
"grad_norm": 1.3451749086380005, |
|
"learning_rate": 5.04e-05, |
|
"loss": 1.1639, |
|
"step": 964 |
|
}, |
|
{ |
|
"epoch": 1.544, |
|
"grad_norm": 1.0739102363586426, |
|
"learning_rate": 4.9199999999999997e-05, |
|
"loss": 1.4125, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 1.5455999999999999, |
|
"grad_norm": 0.9140891432762146, |
|
"learning_rate": 4.7999999999999994e-05, |
|
"loss": 1.1883, |
|
"step": 966 |
|
}, |
|
{ |
|
"epoch": 1.5472000000000001, |
|
"grad_norm": 1.1968399286270142, |
|
"learning_rate": 4.68e-05, |
|
"loss": 1.1692, |
|
"step": 967 |
|
}, |
|
{ |
|
"epoch": 1.5488, |
|
"grad_norm": 0.9506883025169373, |
|
"learning_rate": 4.56e-05, |
|
"loss": 0.9939, |
|
"step": 968 |
|
}, |
|
{ |
|
"epoch": 1.5504, |
|
"grad_norm": 1.3204574584960938, |
|
"learning_rate": 4.4399999999999995e-05, |
|
"loss": 0.9103, |
|
"step": 969 |
|
}, |
|
{ |
|
"epoch": 1.552, |
|
"grad_norm": 1.3693453073501587, |
|
"learning_rate": 4.319999999999999e-05, |
|
"loss": 1.279, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 1.5535999999999999, |
|
"grad_norm": 1.286435842514038, |
|
"learning_rate": 4.2e-05, |
|
"loss": 1.2637, |
|
"step": 971 |
|
}, |
|
{ |
|
"epoch": 1.5552000000000001, |
|
"grad_norm": 1.6549150943756104, |
|
"learning_rate": 4.08e-05, |
|
"loss": 1.5989, |
|
"step": 972 |
|
}, |
|
{ |
|
"epoch": 1.5568, |
|
"grad_norm": 1.1446410417556763, |
|
"learning_rate": 3.96e-05, |
|
"loss": 1.5587, |
|
"step": 973 |
|
}, |
|
{ |
|
"epoch": 1.5584, |
|
"grad_norm": 1.1329030990600586, |
|
"learning_rate": 3.84e-05, |
|
"loss": 1.5915, |
|
"step": 974 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 6.757206916809082, |
|
"learning_rate": 3.7199999999999996e-05, |
|
"loss": 3.078, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 1.5615999999999999, |
|
"grad_norm": 3.0250446796417236, |
|
"learning_rate": 3.5999999999999994e-05, |
|
"loss": 1.5766, |
|
"step": 976 |
|
}, |
|
{ |
|
"epoch": 1.5632000000000001, |
|
"grad_norm": 1.796055793762207, |
|
"learning_rate": 3.48e-05, |
|
"loss": 1.4036, |
|
"step": 977 |
|
}, |
|
{ |
|
"epoch": 1.5648, |
|
"grad_norm": 2.8835206031799316, |
|
"learning_rate": 3.36e-05, |
|
"loss": 1.4255, |
|
"step": 978 |
|
}, |
|
{ |
|
"epoch": 1.5664, |
|
"grad_norm": 3.2792022228240967, |
|
"learning_rate": 3.2399999999999995e-05, |
|
"loss": 1.5776, |
|
"step": 979 |
|
}, |
|
{ |
|
"epoch": 1.568, |
|
"grad_norm": 0.9129714369773865, |
|
"learning_rate": 3.119999999999999e-05, |
|
"loss": 1.1509, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 1.5695999999999999, |
|
"grad_norm": 1.325767993927002, |
|
"learning_rate": 2.9999999999999997e-05, |
|
"loss": 1.3454, |
|
"step": 981 |
|
}, |
|
{ |
|
"epoch": 1.5712000000000002, |
|
"grad_norm": 0.8600112795829773, |
|
"learning_rate": 2.88e-05, |
|
"loss": 0.8939, |
|
"step": 982 |
|
}, |
|
{ |
|
"epoch": 1.5728, |
|
"grad_norm": 0.6780167818069458, |
|
"learning_rate": 2.7599999999999997e-05, |
|
"loss": 0.7801, |
|
"step": 983 |
|
}, |
|
{ |
|
"epoch": 1.5744, |
|
"grad_norm": 1.1986228227615356, |
|
"learning_rate": 2.6399999999999995e-05, |
|
"loss": 1.0117, |
|
"step": 984 |
|
}, |
|
{ |
|
"epoch": 1.576, |
|
"grad_norm": 1.3320322036743164, |
|
"learning_rate": 2.52e-05, |
|
"loss": 1.0323, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 1.5776, |
|
"grad_norm": 0.6832807064056396, |
|
"learning_rate": 2.3999999999999997e-05, |
|
"loss": 1.1191, |
|
"step": 986 |
|
}, |
|
{ |
|
"epoch": 1.5792000000000002, |
|
"grad_norm": 3.410607099533081, |
|
"learning_rate": 2.28e-05, |
|
"loss": 1.7286, |
|
"step": 987 |
|
}, |
|
{ |
|
"epoch": 1.5808, |
|
"grad_norm": 0.7802388668060303, |
|
"learning_rate": 2.1599999999999996e-05, |
|
"loss": 0.7245, |
|
"step": 988 |
|
}, |
|
{ |
|
"epoch": 1.5824, |
|
"grad_norm": 0.8889452815055847, |
|
"learning_rate": 2.04e-05, |
|
"loss": 0.8384, |
|
"step": 989 |
|
}, |
|
{ |
|
"epoch": 1.584, |
|
"grad_norm": 0.7941523194313049, |
|
"learning_rate": 1.92e-05, |
|
"loss": 1.0049, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 1.5856, |
|
"grad_norm": 0.8905624747276306, |
|
"learning_rate": 1.7999999999999997e-05, |
|
"loss": 0.7337, |
|
"step": 991 |
|
}, |
|
{ |
|
"epoch": 1.5872000000000002, |
|
"grad_norm": 1.1984586715698242, |
|
"learning_rate": 1.68e-05, |
|
"loss": 0.8216, |
|
"step": 992 |
|
}, |
|
{ |
|
"epoch": 1.5888, |
|
"grad_norm": 0.8467238545417786, |
|
"learning_rate": 1.5599999999999996e-05, |
|
"loss": 0.807, |
|
"step": 993 |
|
}, |
|
{ |
|
"epoch": 1.5904, |
|
"grad_norm": 1.230843424797058, |
|
"learning_rate": 1.44e-05, |
|
"loss": 1.1562, |
|
"step": 994 |
|
}, |
|
{ |
|
"epoch": 1.592, |
|
"grad_norm": 1.0017879009246826, |
|
"learning_rate": 1.3199999999999997e-05, |
|
"loss": 0.971, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 1.5936, |
|
"grad_norm": 1.124283790588379, |
|
"learning_rate": 1.1999999999999999e-05, |
|
"loss": 0.8834, |
|
"step": 996 |
|
}, |
|
{ |
|
"epoch": 1.5952, |
|
"grad_norm": 0.9626112580299377, |
|
"learning_rate": 1.0799999999999998e-05, |
|
"loss": 1.0106, |
|
"step": 997 |
|
}, |
|
{ |
|
"epoch": 1.5968, |
|
"grad_norm": 1.2808488607406616, |
|
"learning_rate": 9.6e-06, |
|
"loss": 0.9121, |
|
"step": 998 |
|
}, |
|
{ |
|
"epoch": 1.5984, |
|
"grad_norm": 0.904681921005249, |
|
"learning_rate": 8.4e-06, |
|
"loss": 0.8199, |
|
"step": 999 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 0.8806717395782471, |
|
"learning_rate": 7.2e-06, |
|
"loss": 0.972, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"eval_cer": 0.3090127193119299, |
|
"eval_loss": 1.500978946685791, |
|
"eval_runtime": 162.132, |
|
"eval_samples_per_second": 19.342, |
|
"eval_steps_per_second": 1.209, |
|
"eval_wer": 0.4351051665913261, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"step": 1000, |
|
"total_flos": 6.212261523683712e+18, |
|
"train_loss": 3.1903547011613846, |
|
"train_runtime": 2112.9078, |
|
"train_samples_per_second": 15.145, |
|
"train_steps_per_second": 0.473 |
|
} |
|
], |
|
"logging_steps": 1.0, |
|
"max_steps": 1000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 1000, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 6.212261523683712e+18, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|