|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.4803849000740192, |
|
"eval_steps": 507, |
|
"global_step": 1521, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 92.56075816528, |
|
"learning_rate": 5.000000000000001e-07, |
|
"loss": 1.6849, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"eval_loss": 1.7294352054595947, |
|
"eval_runtime": 1884.169, |
|
"eval_samples_per_second": 0.701, |
|
"eval_steps_per_second": 0.078, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 62.47252473867764, |
|
"learning_rate": 1.0000000000000002e-06, |
|
"loss": 1.4808, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 86.16163524072095, |
|
"learning_rate": 1.5e-06, |
|
"loss": 1.622, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 60.0955549964575, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 1.5835, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 29.984043062746373, |
|
"learning_rate": 2.5e-06, |
|
"loss": 1.5242, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 301.1618364383654, |
|
"learning_rate": 3e-06, |
|
"loss": 1.2748, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 109.29021464248359, |
|
"learning_rate": 3.5e-06, |
|
"loss": 1.1112, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 10.444328634170152, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 0.9269, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 6.523925938807769, |
|
"learning_rate": 4.5e-06, |
|
"loss": 0.8644, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 17.340636604619814, |
|
"learning_rate": 5e-06, |
|
"loss": 0.8207, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 44.61126816135743, |
|
"learning_rate": 4.9999969645113385e-06, |
|
"loss": 1.1849, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 4.3252416457498315, |
|
"learning_rate": 4.999987858052726e-06, |
|
"loss": 0.885, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 2.9355463597411195, |
|
"learning_rate": 4.999972680646276e-06, |
|
"loss": 0.8329, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 2.0811257726357524, |
|
"learning_rate": 4.999951432328846e-06, |
|
"loss": 0.7869, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 1.9698959240524336, |
|
"learning_rate": 4.999924113152034e-06, |
|
"loss": 0.7948, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.6266575755898287, |
|
"learning_rate": 4.999890723182183e-06, |
|
"loss": 0.8005, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.419625838626686, |
|
"learning_rate": 4.999851262500375e-06, |
|
"loss": 0.7565, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.4352187159776242, |
|
"learning_rate": 4.999805731202437e-06, |
|
"loss": 0.7602, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.2355497382357647, |
|
"learning_rate": 4.999754129398938e-06, |
|
"loss": 0.7564, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.2700121185974393, |
|
"learning_rate": 4.999696457215186e-06, |
|
"loss": 0.8125, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.1702941851369728, |
|
"learning_rate": 4.9996327147912315e-06, |
|
"loss": 0.7382, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.1798174730304447, |
|
"learning_rate": 4.999562902281866e-06, |
|
"loss": 0.7742, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.1364491950081737, |
|
"learning_rate": 4.999487019856623e-06, |
|
"loss": 0.7282, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.1131547945883817, |
|
"learning_rate": 4.999405067699773e-06, |
|
"loss": 0.7475, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.1506149455536359, |
|
"learning_rate": 4.9993170460103295e-06, |
|
"loss": 0.7072, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.2889319229905287, |
|
"learning_rate": 4.999222955002041e-06, |
|
"loss": 0.8064, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.059562137107165, |
|
"learning_rate": 4.999122794903399e-06, |
|
"loss": 0.7147, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.149120672076531, |
|
"learning_rate": 4.999016565957633e-06, |
|
"loss": 0.7792, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.1596428841302342, |
|
"learning_rate": 4.998904268422705e-06, |
|
"loss": 0.7586, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.980994033140509, |
|
"learning_rate": 4.998785902571319e-06, |
|
"loss": 0.7721, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.9969455208276253, |
|
"learning_rate": 4.998661468690914e-06, |
|
"loss": 0.7482, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.2088730805217034, |
|
"learning_rate": 4.998530967083664e-06, |
|
"loss": 0.7687, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.1772131571685314, |
|
"learning_rate": 4.998394398066477e-06, |
|
"loss": 0.7336, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.0095424896414942, |
|
"learning_rate": 4.998251761970997e-06, |
|
"loss": 0.7454, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.058399487048206, |
|
"learning_rate": 4.998103059143599e-06, |
|
"loss": 0.7424, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.0553895642417541, |
|
"learning_rate": 4.997948289945392e-06, |
|
"loss": 0.7034, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.9737916869473869, |
|
"learning_rate": 4.997787454752217e-06, |
|
"loss": 0.7225, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.0497906961287207, |
|
"learning_rate": 4.997620553954645e-06, |
|
"loss": 0.7545, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.0351771937970027, |
|
"learning_rate": 4.9974475879579745e-06, |
|
"loss": 0.6999, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.025178950396183, |
|
"learning_rate": 4.9972685571822355e-06, |
|
"loss": 0.6836, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.8990499683406185, |
|
"learning_rate": 4.997083462062185e-06, |
|
"loss": 0.749, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.1934033398652506, |
|
"learning_rate": 4.996892303047306e-06, |
|
"loss": 0.7118, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.0093747627659915, |
|
"learning_rate": 4.9966950806018075e-06, |
|
"loss": 0.6931, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.1259320175035152, |
|
"learning_rate": 4.996491795204623e-06, |
|
"loss": 0.7011, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.9340587879454492, |
|
"learning_rate": 4.996282447349408e-06, |
|
"loss": 0.7031, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.093497345220761, |
|
"learning_rate": 4.996067037544542e-06, |
|
"loss": 0.6984, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.9253425670099831, |
|
"learning_rate": 4.9958455663131235e-06, |
|
"loss": 0.7211, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.1321089016179777, |
|
"learning_rate": 4.99561803419297e-06, |
|
"loss": 0.6809, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.9225335387284348, |
|
"learning_rate": 4.995384441736622e-06, |
|
"loss": 0.6926, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.03537386787448, |
|
"learning_rate": 4.9951447895113305e-06, |
|
"loss": 0.6868, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.9586247752600078, |
|
"learning_rate": 4.994899078099064e-06, |
|
"loss": 0.6973, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.164879075983742, |
|
"learning_rate": 4.994647308096509e-06, |
|
"loss": 0.6976, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.007151099232745, |
|
"learning_rate": 4.994389480115059e-06, |
|
"loss": 0.6728, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.194002279613801, |
|
"learning_rate": 4.9941255947808226e-06, |
|
"loss": 0.7386, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.2903234642197774, |
|
"learning_rate": 4.993855652734616e-06, |
|
"loss": 0.6884, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.9696246122793456, |
|
"learning_rate": 4.993579654631963e-06, |
|
"loss": 0.7175, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 1.1155328763962573, |
|
"learning_rate": 4.993297601143095e-06, |
|
"loss": 0.7146, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.9599001918058567, |
|
"learning_rate": 4.993009492952951e-06, |
|
"loss": 0.7054, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.9180819748743216, |
|
"learning_rate": 4.992715330761167e-06, |
|
"loss": 0.6984, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.9088984413752035, |
|
"learning_rate": 4.992415115282085e-06, |
|
"loss": 0.7001, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 1.1141716959997088, |
|
"learning_rate": 4.992108847244746e-06, |
|
"loss": 0.6756, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.8636787369748702, |
|
"learning_rate": 4.991796527392888e-06, |
|
"loss": 0.7548, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 1.0780460703120494, |
|
"learning_rate": 4.991478156484945e-06, |
|
"loss": 0.7039, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.9164733455688905, |
|
"learning_rate": 4.9911537352940485e-06, |
|
"loss": 0.7385, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 1.0268015729747888, |
|
"learning_rate": 4.990823264608017e-06, |
|
"loss": 0.7283, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.9924581532223246, |
|
"learning_rate": 4.990486745229364e-06, |
|
"loss": 0.7315, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.953526608505078, |
|
"learning_rate": 4.99014417797529e-06, |
|
"loss": 0.7194, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.9721089885130799, |
|
"learning_rate": 4.9897955636776824e-06, |
|
"loss": 0.6786, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 1.095580212055755, |
|
"learning_rate": 4.989440903183112e-06, |
|
"loss": 0.6547, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 1.0176000454229972, |
|
"learning_rate": 4.989080197352834e-06, |
|
"loss": 0.7024, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.9764964284876755, |
|
"learning_rate": 4.988713447062784e-06, |
|
"loss": 0.6897, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.9720697960414958, |
|
"learning_rate": 4.9883406532035735e-06, |
|
"loss": 0.7368, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.8924650786548242, |
|
"learning_rate": 4.987961816680493e-06, |
|
"loss": 0.6986, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 1.0583230570989346, |
|
"learning_rate": 4.987576938413504e-06, |
|
"loss": 0.6488, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.8532905196520026, |
|
"learning_rate": 4.9871860193372424e-06, |
|
"loss": 0.6545, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 1.110205997163892, |
|
"learning_rate": 4.9867890604010125e-06, |
|
"loss": 0.7296, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.8585453329590255, |
|
"learning_rate": 4.986386062568787e-06, |
|
"loss": 0.6478, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 1.072721980278167, |
|
"learning_rate": 4.9859770268192e-06, |
|
"loss": 0.683, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.9114010594463168, |
|
"learning_rate": 4.985561954145552e-06, |
|
"loss": 0.7278, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.9407855861796833, |
|
"learning_rate": 4.985140845555799e-06, |
|
"loss": 0.6767, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.9716353726299576, |
|
"learning_rate": 4.984713702072561e-06, |
|
"loss": 0.7058, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.9289919706551977, |
|
"learning_rate": 4.984280524733107e-06, |
|
"loss": 0.7465, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.9719373864834964, |
|
"learning_rate": 4.9838413145893615e-06, |
|
"loss": 0.7173, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.8656006720690229, |
|
"learning_rate": 4.983396072707898e-06, |
|
"loss": 0.7138, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.9405397793417023, |
|
"learning_rate": 4.982944800169939e-06, |
|
"loss": 0.7172, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.8968931033898299, |
|
"learning_rate": 4.982487498071349e-06, |
|
"loss": 0.6959, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.969922689772122, |
|
"learning_rate": 4.982024167522638e-06, |
|
"loss": 0.706, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.8691971891272707, |
|
"learning_rate": 4.981554809648952e-06, |
|
"loss": 0.7156, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.9182420056058627, |
|
"learning_rate": 4.981079425590078e-06, |
|
"loss": 0.6694, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.9014669660783321, |
|
"learning_rate": 4.980598016500431e-06, |
|
"loss": 0.6842, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.8598791805359901, |
|
"learning_rate": 4.980110583549062e-06, |
|
"loss": 0.6707, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.8569795805490359, |
|
"learning_rate": 4.97961712791965e-06, |
|
"loss": 0.6637, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.8939370143328529, |
|
"learning_rate": 4.979117650810495e-06, |
|
"loss": 0.6927, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.9152075194521869, |
|
"learning_rate": 4.978612153434527e-06, |
|
"loss": 0.6803, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.9298563182624136, |
|
"learning_rate": 4.978100637019287e-06, |
|
"loss": 0.6776, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.9132188893536833, |
|
"learning_rate": 4.97758310280694e-06, |
|
"loss": 0.7032, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.9574161232781394, |
|
"learning_rate": 4.97705955205426e-06, |
|
"loss": 0.6731, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.8327779171567138, |
|
"learning_rate": 4.976529986032632e-06, |
|
"loss": 0.6754, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 1.1175664554073028, |
|
"learning_rate": 4.975994406028051e-06, |
|
"loss": 0.6774, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.9122529617157105, |
|
"learning_rate": 4.9754528133411144e-06, |
|
"loss": 0.6905, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.9495918841010983, |
|
"learning_rate": 4.97490520928702e-06, |
|
"loss": 0.6683, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.9199809745368361, |
|
"learning_rate": 4.9743515951955655e-06, |
|
"loss": 0.6512, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.9371967382154123, |
|
"learning_rate": 4.973791972411142e-06, |
|
"loss": 0.7004, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 1.0249455893660548, |
|
"learning_rate": 4.973226342292732e-06, |
|
"loss": 0.6902, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 1.007634999853075, |
|
"learning_rate": 4.972654706213906e-06, |
|
"loss": 0.712, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.9151412270511807, |
|
"learning_rate": 4.9720770655628216e-06, |
|
"loss": 0.689, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.9263042922044208, |
|
"learning_rate": 4.971493421742215e-06, |
|
"loss": 0.7188, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.9875225370080567, |
|
"learning_rate": 4.970903776169403e-06, |
|
"loss": 0.7313, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.9329951904410932, |
|
"learning_rate": 4.970308130276273e-06, |
|
"loss": 0.6473, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.952122809501081, |
|
"learning_rate": 4.969706485509287e-06, |
|
"loss": 0.6866, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.8675487151632736, |
|
"learning_rate": 4.969098843329475e-06, |
|
"loss": 0.6621, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 1.0124220925484857, |
|
"learning_rate": 4.968485205212429e-06, |
|
"loss": 0.696, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.8692279553117837, |
|
"learning_rate": 4.967865572648302e-06, |
|
"loss": 0.6503, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.8684082104461566, |
|
"learning_rate": 4.967239947141803e-06, |
|
"loss": 0.717, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.9363131782505817, |
|
"learning_rate": 4.966608330212198e-06, |
|
"loss": 0.67, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.9483280670290893, |
|
"learning_rate": 4.965970723393297e-06, |
|
"loss": 0.6645, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.9645550818323707, |
|
"learning_rate": 4.965327128233462e-06, |
|
"loss": 0.6402, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.9020494921501199, |
|
"learning_rate": 4.96467754629559e-06, |
|
"loss": 0.6848, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.9273157548895928, |
|
"learning_rate": 4.964021979157122e-06, |
|
"loss": 0.6888, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 1.0083484650433756, |
|
"learning_rate": 4.963360428410031e-06, |
|
"loss": 0.7223, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.9556311862868435, |
|
"learning_rate": 4.96269289566082e-06, |
|
"loss": 0.6924, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.8565613256126575, |
|
"learning_rate": 4.962019382530521e-06, |
|
"loss": 0.6551, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 1.0082880819703033, |
|
"learning_rate": 4.961339890654685e-06, |
|
"loss": 0.6208, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 1.0159617953712363, |
|
"learning_rate": 4.9606544216833865e-06, |
|
"loss": 0.7223, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.9941249295005459, |
|
"learning_rate": 4.95996297728121e-06, |
|
"loss": 0.6853, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.980987656479997, |
|
"learning_rate": 4.959265559127253e-06, |
|
"loss": 0.6531, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.8906826655489122, |
|
"learning_rate": 4.958562168915121e-06, |
|
"loss": 0.6822, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.9447717033103807, |
|
"learning_rate": 4.95785280835292e-06, |
|
"loss": 0.6735, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.9322226016156214, |
|
"learning_rate": 4.957137479163253e-06, |
|
"loss": 0.6853, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.9016538169637961, |
|
"learning_rate": 4.9564161830832214e-06, |
|
"loss": 0.6546, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.9425807696655617, |
|
"learning_rate": 4.955688921864411e-06, |
|
"loss": 0.6874, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.939091235820591, |
|
"learning_rate": 4.954955697272899e-06, |
|
"loss": 0.7002, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.8787416221103219, |
|
"learning_rate": 4.954216511089242e-06, |
|
"loss": 0.7502, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.9742122954687594, |
|
"learning_rate": 4.95347136510847e-06, |
|
"loss": 0.6657, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.8801095585358768, |
|
"learning_rate": 4.95272026114009e-06, |
|
"loss": 0.6555, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.9481198118358727, |
|
"learning_rate": 4.9519632010080765e-06, |
|
"loss": 0.6737, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.9764499477724559, |
|
"learning_rate": 4.951200186550868e-06, |
|
"loss": 0.6616, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.9677532996854814, |
|
"learning_rate": 4.9504312196213596e-06, |
|
"loss": 0.6604, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.9262840329326826, |
|
"learning_rate": 4.949656302086907e-06, |
|
"loss": 0.6634, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.9913169871400999, |
|
"learning_rate": 4.94887543582931e-06, |
|
"loss": 0.6834, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 1.037107833577886, |
|
"learning_rate": 4.948088622744819e-06, |
|
"loss": 0.7047, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 1.0270598791002519, |
|
"learning_rate": 4.947295864744121e-06, |
|
"loss": 0.6637, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.9780188865838979, |
|
"learning_rate": 4.9464971637523465e-06, |
|
"loss": 0.7196, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.9564318039750372, |
|
"learning_rate": 4.94569252170905e-06, |
|
"loss": 0.661, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 1.070366815703729, |
|
"learning_rate": 4.944881940568219e-06, |
|
"loss": 0.7288, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.951311961745425, |
|
"learning_rate": 4.944065422298262e-06, |
|
"loss": 0.6846, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.9924454965931206, |
|
"learning_rate": 4.943242968882002e-06, |
|
"loss": 0.6382, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.899515518807678, |
|
"learning_rate": 4.942414582316679e-06, |
|
"loss": 0.6949, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.9872873524132176, |
|
"learning_rate": 4.94158026461394e-06, |
|
"loss": 0.6765, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.8805014666801008, |
|
"learning_rate": 4.9407400177998335e-06, |
|
"loss": 0.6288, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.9546823658223654, |
|
"learning_rate": 4.939893843914808e-06, |
|
"loss": 0.6749, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.9963934911162335, |
|
"learning_rate": 4.939041745013703e-06, |
|
"loss": 0.6433, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.9708720605788677, |
|
"learning_rate": 4.93818372316575e-06, |
|
"loss": 0.6774, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.9473055224950518, |
|
"learning_rate": 4.937319780454559e-06, |
|
"loss": 0.685, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.9819312296607249, |
|
"learning_rate": 4.936449918978123e-06, |
|
"loss": 0.6611, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.9603793426190534, |
|
"learning_rate": 4.9355741408488045e-06, |
|
"loss": 0.6939, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 1.123192378969838, |
|
"learning_rate": 4.9346924481933345e-06, |
|
"loss": 0.6596, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 1.049989264664389, |
|
"learning_rate": 4.933804843152808e-06, |
|
"loss": 0.6664, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.9287929732517263, |
|
"learning_rate": 4.932911327882679e-06, |
|
"loss": 0.6401, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.913529924198974, |
|
"learning_rate": 4.932011904552749e-06, |
|
"loss": 0.6525, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.8908499663624638, |
|
"learning_rate": 4.931106575347171e-06, |
|
"loss": 0.6781, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.8871778748426131, |
|
"learning_rate": 4.930195342464437e-06, |
|
"loss": 0.6816, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.8819674351946986, |
|
"learning_rate": 4.929278208117378e-06, |
|
"loss": 0.6153, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.9892043785009771, |
|
"learning_rate": 4.928355174533153e-06, |
|
"loss": 0.6716, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.9215453869699415, |
|
"learning_rate": 4.927426243953252e-06, |
|
"loss": 0.6784, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.9652774517156478, |
|
"learning_rate": 4.9264914186334775e-06, |
|
"loss": 0.7372, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.8716382127920056, |
|
"learning_rate": 4.925550700843953e-06, |
|
"loss": 0.6594, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 1.02563864851558, |
|
"learning_rate": 4.924604092869109e-06, |
|
"loss": 0.6678, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.980206197810093, |
|
"learning_rate": 4.923651597007679e-06, |
|
"loss": 0.7016, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.9257361907115577, |
|
"learning_rate": 4.922693215572695e-06, |
|
"loss": 0.6537, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.9407972324104933, |
|
"learning_rate": 4.9217289508914836e-06, |
|
"loss": 0.6906, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.953792733030727, |
|
"learning_rate": 4.920758805305654e-06, |
|
"loss": 0.6874, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.9679520423391571, |
|
"learning_rate": 4.919782781171101e-06, |
|
"loss": 0.7057, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.8643570019367653, |
|
"learning_rate": 4.918800880857991e-06, |
|
"loss": 0.6224, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.9696150692028724, |
|
"learning_rate": 4.917813106750763e-06, |
|
"loss": 0.6623, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.9068337402094646, |
|
"learning_rate": 4.916819461248119e-06, |
|
"loss": 0.6756, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.9032967004326324, |
|
"learning_rate": 4.915819946763017e-06, |
|
"loss": 0.6624, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.9042931017160107, |
|
"learning_rate": 4.914814565722671e-06, |
|
"loss": 0.6711, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.8611902344265154, |
|
"learning_rate": 4.913803320568538e-06, |
|
"loss": 0.6652, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 1.1208348114642743, |
|
"learning_rate": 4.912786213756316e-06, |
|
"loss": 0.7054, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.8514149510371133, |
|
"learning_rate": 4.91176324775594e-06, |
|
"loss": 0.6777, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 1.0707648844359314, |
|
"learning_rate": 4.91073442505157e-06, |
|
"loss": 0.6793, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 1.007444952147854, |
|
"learning_rate": 4.9096997481415885e-06, |
|
"loss": 0.6532, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.8430194319590454, |
|
"learning_rate": 4.908659219538598e-06, |
|
"loss": 0.6532, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.9867088678023405, |
|
"learning_rate": 4.907612841769407e-06, |
|
"loss": 0.7107, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.9369756600022385, |
|
"learning_rate": 4.90656061737503e-06, |
|
"loss": 0.7306, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.8672758137612632, |
|
"learning_rate": 4.905502548910681e-06, |
|
"loss": 0.6412, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.8706341698734509, |
|
"learning_rate": 4.904438638945761e-06, |
|
"loss": 0.7022, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.9114920575560276, |
|
"learning_rate": 4.903368890063861e-06, |
|
"loss": 0.6328, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 1.0338522059220963, |
|
"learning_rate": 4.9022933048627496e-06, |
|
"loss": 0.7328, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.8878198226644961, |
|
"learning_rate": 4.901211885954367e-06, |
|
"loss": 0.6687, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.9423243153918482, |
|
"learning_rate": 4.900124635964823e-06, |
|
"loss": 0.6604, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.9129637670735964, |
|
"learning_rate": 4.899031557534383e-06, |
|
"loss": 0.6731, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.9896776529578333, |
|
"learning_rate": 4.8979326533174696e-06, |
|
"loss": 0.6608, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.8896434182833951, |
|
"learning_rate": 4.896827925982654e-06, |
|
"loss": 0.7074, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.8194553474739297, |
|
"learning_rate": 4.895717378212644e-06, |
|
"loss": 0.6335, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.9346190741848343, |
|
"learning_rate": 4.894601012704284e-06, |
|
"loss": 0.7068, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.874199819168889, |
|
"learning_rate": 4.893478832168545e-06, |
|
"loss": 0.6641, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.9159502482028847, |
|
"learning_rate": 4.8923508393305224e-06, |
|
"loss": 0.6731, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.8558744309352562, |
|
"learning_rate": 4.891217036929422e-06, |
|
"loss": 0.6083, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.936605889796404, |
|
"learning_rate": 4.89007742771856e-06, |
|
"loss": 0.6863, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.8664253450652424, |
|
"learning_rate": 4.8889320144653525e-06, |
|
"loss": 0.6693, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.859861919600542, |
|
"learning_rate": 4.887780799951311e-06, |
|
"loss": 0.63, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.8390583343121363, |
|
"learning_rate": 4.8866237869720334e-06, |
|
"loss": 0.7058, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.9197736229674498, |
|
"learning_rate": 4.885460978337201e-06, |
|
"loss": 0.6782, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.9686243956595285, |
|
"learning_rate": 4.884292376870567e-06, |
|
"loss": 0.6619, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.8454295473139757, |
|
"learning_rate": 4.883117985409954e-06, |
|
"loss": 0.6551, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.9658575404117379, |
|
"learning_rate": 4.881937806807241e-06, |
|
"loss": 0.647, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.8533364242044601, |
|
"learning_rate": 4.880751843928365e-06, |
|
"loss": 0.6563, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.9045931704682527, |
|
"learning_rate": 4.879560099653306e-06, |
|
"loss": 0.6667, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.8723782063352707, |
|
"learning_rate": 4.8783625768760865e-06, |
|
"loss": 0.6425, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.8891929748147736, |
|
"learning_rate": 4.877159278504759e-06, |
|
"loss": 0.638, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.8842932376273528, |
|
"learning_rate": 4.875950207461403e-06, |
|
"loss": 0.6437, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.8634008607871101, |
|
"learning_rate": 4.8747353666821155e-06, |
|
"loss": 0.6229, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.8749107029341481, |
|
"learning_rate": 4.873514759117004e-06, |
|
"loss": 0.6785, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.8531375678504259, |
|
"learning_rate": 4.872288387730182e-06, |
|
"loss": 0.637, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.8582175373389599, |
|
"learning_rate": 4.871056255499758e-06, |
|
"loss": 0.6444, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 1.0656353331347395, |
|
"learning_rate": 4.86981836541783e-06, |
|
"loss": 0.6756, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.8937004109826505, |
|
"learning_rate": 4.8685747204904796e-06, |
|
"loss": 0.652, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.9024529450351501, |
|
"learning_rate": 4.867325323737765e-06, |
|
"loss": 0.6606, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.847448497225188, |
|
"learning_rate": 4.866070178193707e-06, |
|
"loss": 0.6445, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.9790412024382392, |
|
"learning_rate": 4.86480928690629e-06, |
|
"loss": 0.6377, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.9166672541084665, |
|
"learning_rate": 4.863542652937453e-06, |
|
"loss": 0.6568, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.9053859567083279, |
|
"learning_rate": 4.862270279363076e-06, |
|
"loss": 0.6729, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.8963082010943075, |
|
"learning_rate": 4.860992169272981e-06, |
|
"loss": 0.7187, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.8625928467278408, |
|
"learning_rate": 4.859708325770919e-06, |
|
"loss": 0.6569, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.9572711715249694, |
|
"learning_rate": 4.858418751974564e-06, |
|
"loss": 0.6858, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.9116206947499983, |
|
"learning_rate": 4.857123451015503e-06, |
|
"loss": 0.6341, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.9469293485174248, |
|
"learning_rate": 4.855822426039236e-06, |
|
"loss": 0.6475, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.9320786521604542, |
|
"learning_rate": 4.854515680205159e-06, |
|
"loss": 0.7021, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 1.038184288949622, |
|
"learning_rate": 4.853203216686562e-06, |
|
"loss": 0.5999, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.904022305585284, |
|
"learning_rate": 4.851885038670618e-06, |
|
"loss": 0.6555, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 1.323477712215313, |
|
"learning_rate": 4.8505611493583815e-06, |
|
"loss": 0.6297, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 1.2176660817908431, |
|
"learning_rate": 4.849231551964771e-06, |
|
"loss": 0.6771, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.8930100103847493, |
|
"learning_rate": 4.84789624971857e-06, |
|
"loss": 0.6296, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 1.0104643222544036, |
|
"learning_rate": 4.846555245862413e-06, |
|
"loss": 0.6329, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 1.0006315975644342, |
|
"learning_rate": 4.845208543652783e-06, |
|
"loss": 0.6448, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.9719470767684136, |
|
"learning_rate": 4.843856146359999e-06, |
|
"loss": 0.6538, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 1.025734064898448, |
|
"learning_rate": 4.842498057268209e-06, |
|
"loss": 0.6335, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.8825687655634383, |
|
"learning_rate": 4.841134279675386e-06, |
|
"loss": 0.6728, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.9411597091872842, |
|
"learning_rate": 4.839764816893315e-06, |
|
"loss": 0.6756, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.8838664041310971, |
|
"learning_rate": 4.838389672247585e-06, |
|
"loss": 0.6593, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.8745260448073756, |
|
"learning_rate": 4.837008849077588e-06, |
|
"loss": 0.6397, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.8847557689287797, |
|
"learning_rate": 4.835622350736499e-06, |
|
"loss": 0.659, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.875366668905105, |
|
"learning_rate": 4.8342301805912815e-06, |
|
"loss": 0.627, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.8537326681265562, |
|
"learning_rate": 4.832832342022666e-06, |
|
"loss": 0.6477, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.8593654316451845, |
|
"learning_rate": 4.831428838425153e-06, |
|
"loss": 0.6698, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.9220753709893383, |
|
"learning_rate": 4.830019673206997e-06, |
|
"loss": 0.6477, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.8570602383594389, |
|
"learning_rate": 4.828604849790201e-06, |
|
"loss": 0.6815, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.9015787695209869, |
|
"learning_rate": 4.827184371610511e-06, |
|
"loss": 0.6293, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.900429307913389, |
|
"learning_rate": 4.8257582421174025e-06, |
|
"loss": 0.6799, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.8833032292011391, |
|
"learning_rate": 4.824326464774076e-06, |
|
"loss": 0.6665, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.9806435575510756, |
|
"learning_rate": 4.822889043057446e-06, |
|
"loss": 0.6828, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.8273111947815865, |
|
"learning_rate": 4.821445980458134e-06, |
|
"loss": 0.6265, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.9867943428090185, |
|
"learning_rate": 4.8199972804804615e-06, |
|
"loss": 0.6856, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.941670204697157, |
|
"learning_rate": 4.8185429466424375e-06, |
|
"loss": 0.6241, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.950190491805473, |
|
"learning_rate": 4.817082982475753e-06, |
|
"loss": 0.6733, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.9059424297489737, |
|
"learning_rate": 4.815617391525772e-06, |
|
"loss": 0.6717, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.833767481386628, |
|
"learning_rate": 4.814146177351523e-06, |
|
"loss": 0.6167, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.9120074119227388, |
|
"learning_rate": 4.812669343525688e-06, |
|
"loss": 0.6383, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 1.0846687157648525, |
|
"learning_rate": 4.811186893634597e-06, |
|
"loss": 0.6542, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.9273661121621178, |
|
"learning_rate": 4.809698831278217e-06, |
|
"loss": 0.6658, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.8614935020984111, |
|
"learning_rate": 4.808205160070147e-06, |
|
"loss": 0.6632, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.8640864376511351, |
|
"learning_rate": 4.806705883637604e-06, |
|
"loss": 0.6714, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.8588959960471779, |
|
"learning_rate": 4.8052010056214184e-06, |
|
"loss": 0.6365, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.9735734771646465, |
|
"learning_rate": 4.80369052967602e-06, |
|
"loss": 0.6939, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.9404781469926247, |
|
"learning_rate": 4.802174459469435e-06, |
|
"loss": 0.6291, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.9663046015849157, |
|
"learning_rate": 4.800652798683277e-06, |
|
"loss": 0.6516, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.8807991294468791, |
|
"learning_rate": 4.799125551012731e-06, |
|
"loss": 0.6364, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.9527629477737198, |
|
"learning_rate": 4.7975927201665515e-06, |
|
"loss": 0.6298, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.8525983935565186, |
|
"learning_rate": 4.796054309867053e-06, |
|
"loss": 0.6371, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.9405625850498484, |
|
"learning_rate": 4.794510323850096e-06, |
|
"loss": 0.6264, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.9478829826012205, |
|
"learning_rate": 4.79296076586508e-06, |
|
"loss": 0.6328, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.9599057130243777, |
|
"learning_rate": 4.791405639674941e-06, |
|
"loss": 0.6151, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.959409086045542, |
|
"learning_rate": 4.789844949056131e-06, |
|
"loss": 0.6369, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.8492616257179605, |
|
"learning_rate": 4.788278697798619e-06, |
|
"loss": 0.6554, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.9806877124101798, |
|
"learning_rate": 4.7867068897058725e-06, |
|
"loss": 0.6579, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.9000450903810971, |
|
"learning_rate": 4.785129528594858e-06, |
|
"loss": 0.6235, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.9010530779092166, |
|
"learning_rate": 4.783546618296025e-06, |
|
"loss": 0.658, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.9784681165218017, |
|
"learning_rate": 4.781958162653298e-06, |
|
"loss": 0.6676, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.9576730133216942, |
|
"learning_rate": 4.780364165524067e-06, |
|
"loss": 0.6228, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 1.0020359961640437, |
|
"learning_rate": 4.778764630779184e-06, |
|
"loss": 0.647, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.8688473783118098, |
|
"learning_rate": 4.77715956230294e-06, |
|
"loss": 0.652, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.9278357926734184, |
|
"learning_rate": 4.775548963993072e-06, |
|
"loss": 0.6291, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.9690649138149913, |
|
"learning_rate": 4.77393283976074e-06, |
|
"loss": 0.6509, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.9407962797966478, |
|
"learning_rate": 4.7723111935305275e-06, |
|
"loss": 0.6197, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.9768285903734765, |
|
"learning_rate": 4.770684029240425e-06, |
|
"loss": 0.6518, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 1.0077965365925552, |
|
"learning_rate": 4.769051350841822e-06, |
|
"loss": 0.6195, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.9474683243935313, |
|
"learning_rate": 4.767413162299501e-06, |
|
"loss": 0.6504, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.8585900243531269, |
|
"learning_rate": 4.765769467591626e-06, |
|
"loss": 0.6539, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.8904877736571806, |
|
"learning_rate": 4.764120270709727e-06, |
|
"loss": 0.6321, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.8628282782617729, |
|
"learning_rate": 4.7624655756587e-06, |
|
"loss": 0.648, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 2.669416127497495, |
|
"learning_rate": 4.760805386456793e-06, |
|
"loss": 0.6594, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.8981381784754892, |
|
"learning_rate": 4.759139707135592e-06, |
|
"loss": 0.6769, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.7879080017640494, |
|
"learning_rate": 4.757468541740019e-06, |
|
"loss": 0.6281, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.8871289305729931, |
|
"learning_rate": 4.755791894328317e-06, |
|
"loss": 0.6321, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.8299591322141577, |
|
"learning_rate": 4.75410976897204e-06, |
|
"loss": 0.6505, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.8909590264299827, |
|
"learning_rate": 4.752422169756048e-06, |
|
"loss": 0.6811, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.9254976675249351, |
|
"learning_rate": 4.75072910077849e-06, |
|
"loss": 0.6545, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.8588476509107275, |
|
"learning_rate": 4.7490305661508006e-06, |
|
"loss": 0.667, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.8412766854913993, |
|
"learning_rate": 4.747326569997684e-06, |
|
"loss": 0.6003, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.8760795055942986, |
|
"learning_rate": 4.74561711645711e-06, |
|
"loss": 0.6364, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.83956409107638, |
|
"learning_rate": 4.7439022096803024e-06, |
|
"loss": 0.6575, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.8380872276529072, |
|
"learning_rate": 4.742181853831721e-06, |
|
"loss": 0.6424, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.8125631213297304, |
|
"learning_rate": 4.740456053089065e-06, |
|
"loss": 0.6663, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.8823004853380154, |
|
"learning_rate": 4.7387248116432524e-06, |
|
"loss": 0.6301, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.8640770572532905, |
|
"learning_rate": 4.736988133698416e-06, |
|
"loss": 0.6406, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.8823776031811036, |
|
"learning_rate": 4.735246023471885e-06, |
|
"loss": 0.6115, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.993013360555103, |
|
"learning_rate": 4.733498485194188e-06, |
|
"loss": 0.6643, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.8904288915714781, |
|
"learning_rate": 4.731745523109029e-06, |
|
"loss": 0.6555, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.8926370506160105, |
|
"learning_rate": 4.729987141473286e-06, |
|
"loss": 0.6601, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.953062196989694, |
|
"learning_rate": 4.728223344556996e-06, |
|
"loss": 0.6809, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.8761001219987443, |
|
"learning_rate": 4.7264541366433495e-06, |
|
"loss": 0.6833, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.9566539767394971, |
|
"learning_rate": 4.724679522028672e-06, |
|
"loss": 0.6385, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.9891726403966534, |
|
"learning_rate": 4.722899505022424e-06, |
|
"loss": 0.6999, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.8037801938204335, |
|
"learning_rate": 4.721114089947181e-06, |
|
"loss": 0.5879, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.9130617389052504, |
|
"learning_rate": 4.71932328113863e-06, |
|
"loss": 0.6373, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.8259173345318234, |
|
"learning_rate": 4.717527082945555e-06, |
|
"loss": 0.6019, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.8371433538017989, |
|
"learning_rate": 4.715725499729826e-06, |
|
"loss": 0.6329, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.9057753463438688, |
|
"learning_rate": 4.713918535866392e-06, |
|
"loss": 0.6281, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.9029311795134953, |
|
"learning_rate": 4.712106195743269e-06, |
|
"loss": 0.6324, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.8858712104718339, |
|
"learning_rate": 4.710288483761524e-06, |
|
"loss": 0.6032, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.8645912795937802, |
|
"learning_rate": 4.708465404335277e-06, |
|
"loss": 0.6165, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.8006257669108777, |
|
"learning_rate": 4.706636961891673e-06, |
|
"loss": 0.6829, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.8617001997741787, |
|
"learning_rate": 4.704803160870888e-06, |
|
"loss": 0.6338, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.8966975029556187, |
|
"learning_rate": 4.702964005726106e-06, |
|
"loss": 0.6447, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 1.0966911622953233, |
|
"learning_rate": 4.701119500923516e-06, |
|
"loss": 0.6652, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.882038752989712, |
|
"learning_rate": 4.699269650942296e-06, |
|
"loss": 0.6022, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.9811354762670036, |
|
"learning_rate": 4.697414460274605e-06, |
|
"loss": 0.6672, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.8694666503441387, |
|
"learning_rate": 4.6955539334255714e-06, |
|
"loss": 0.6327, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.8671894922067316, |
|
"learning_rate": 4.693688074913282e-06, |
|
"loss": 0.6306, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.8737657642067588, |
|
"learning_rate": 4.69181688926877e-06, |
|
"loss": 0.6059, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.9300407192703628, |
|
"learning_rate": 4.689940381036005e-06, |
|
"loss": 0.6382, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.836435132546773, |
|
"learning_rate": 4.6880585547718845e-06, |
|
"loss": 0.6493, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.858710237692544, |
|
"learning_rate": 4.686171415046217e-06, |
|
"loss": 0.6392, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.9070208021531481, |
|
"learning_rate": 4.684278966441716e-06, |
|
"loss": 0.6074, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.8430640165966782, |
|
"learning_rate": 4.682381213553986e-06, |
|
"loss": 0.6386, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.9181203953874308, |
|
"learning_rate": 4.680478160991514e-06, |
|
"loss": 0.6246, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.8224529815938104, |
|
"learning_rate": 4.678569813375654e-06, |
|
"loss": 0.6367, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.8840051978175106, |
|
"learning_rate": 4.676656175340621e-06, |
|
"loss": 0.6802, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.851500472699274, |
|
"learning_rate": 4.674737251533476e-06, |
|
"loss": 0.6415, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.9237880941549678, |
|
"learning_rate": 4.672813046614116e-06, |
|
"loss": 0.6324, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.835807900767003, |
|
"learning_rate": 4.670883565255264e-06, |
|
"loss": 0.6196, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.8636603022537266, |
|
"learning_rate": 4.668948812142454e-06, |
|
"loss": 0.6054, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.8467987581796347, |
|
"learning_rate": 4.6670087919740224e-06, |
|
"loss": 0.6364, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.7976168655689396, |
|
"learning_rate": 4.665063509461098e-06, |
|
"loss": 0.6407, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.8787272073060952, |
|
"learning_rate": 4.663112969327584e-06, |
|
"loss": 0.6275, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.9257599968777892, |
|
"learning_rate": 4.661157176310159e-06, |
|
"loss": 0.6561, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.8310789526292252, |
|
"learning_rate": 4.659196135158251e-06, |
|
"loss": 0.6718, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.909951412942146, |
|
"learning_rate": 4.657229850634033e-06, |
|
"loss": 0.6861, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.8697647332483404, |
|
"learning_rate": 4.6552583275124145e-06, |
|
"loss": 0.6131, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.9004276489872411, |
|
"learning_rate": 4.653281570581023e-06, |
|
"loss": 0.6187, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.8571109763430029, |
|
"learning_rate": 4.651299584640198e-06, |
|
"loss": 0.6096, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.8734397889718317, |
|
"learning_rate": 4.6493123745029765e-06, |
|
"loss": 0.6915, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.8669799668424895, |
|
"learning_rate": 4.64731994499508e-06, |
|
"loss": 0.6401, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.8542352737214168, |
|
"learning_rate": 4.645322300954908e-06, |
|
"loss": 0.6559, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.870542275037437, |
|
"learning_rate": 4.643319447233521e-06, |
|
"loss": 0.6826, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.8583311322550404, |
|
"learning_rate": 4.641311388694629e-06, |
|
"loss": 0.6258, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.8737060383296704, |
|
"learning_rate": 4.639298130214585e-06, |
|
"loss": 0.6339, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.8912967573704136, |
|
"learning_rate": 4.637279676682367e-06, |
|
"loss": 0.6469, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.8715196658497849, |
|
"learning_rate": 4.635256032999569e-06, |
|
"loss": 0.6646, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.8662783788798808, |
|
"learning_rate": 4.633227204080389e-06, |
|
"loss": 0.6338, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.8927387239503449, |
|
"learning_rate": 4.631193194851617e-06, |
|
"loss": 0.6251, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.9498780462874753, |
|
"learning_rate": 4.629154010252624e-06, |
|
"loss": 0.615, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.956056661851013, |
|
"learning_rate": 4.627109655235345e-06, |
|
"loss": 0.6302, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.9067649800180307, |
|
"learning_rate": 4.625060134764273e-06, |
|
"loss": 0.6604, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 1.027192573699629, |
|
"learning_rate": 4.623005453816447e-06, |
|
"loss": 0.5859, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.9203828984363205, |
|
"learning_rate": 4.620945617381435e-06, |
|
"loss": 0.6202, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.951292364398945, |
|
"learning_rate": 4.618880630461324e-06, |
|
"loss": 0.65, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.9118476668711323, |
|
"learning_rate": 4.6168104980707105e-06, |
|
"loss": 0.6487, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.878750842802124, |
|
"learning_rate": 4.614735225236685e-06, |
|
"loss": 0.615, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.9060696025734105, |
|
"learning_rate": 4.612654816998821e-06, |
|
"loss": 0.6248, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.9609216437070113, |
|
"learning_rate": 4.610569278409164e-06, |
|
"loss": 0.6778, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.898695374253286, |
|
"learning_rate": 4.608478614532215e-06, |
|
"loss": 0.6336, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.9341703247271076, |
|
"learning_rate": 4.606382830444925e-06, |
|
"loss": 0.7034, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.9888477139866373, |
|
"learning_rate": 4.604281931236675e-06, |
|
"loss": 0.6101, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.8782761928393398, |
|
"learning_rate": 4.602175922009272e-06, |
|
"loss": 0.5853, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.9482746932782271, |
|
"learning_rate": 4.600064807876929e-06, |
|
"loss": 0.6652, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.8564815677643716, |
|
"learning_rate": 4.597948593966256e-06, |
|
"loss": 0.6782, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.9240455443174742, |
|
"learning_rate": 4.595827285416248e-06, |
|
"loss": 0.6275, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.8929439029031102, |
|
"learning_rate": 4.59370088737827e-06, |
|
"loss": 0.6398, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.9148342501219822, |
|
"learning_rate": 4.59156940501605e-06, |
|
"loss": 0.6359, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.946705923671764, |
|
"learning_rate": 4.589432843505659e-06, |
|
"loss": 0.6128, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.9114594357972271, |
|
"learning_rate": 4.587291208035504e-06, |
|
"loss": 0.6329, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.8637275857494376, |
|
"learning_rate": 4.585144503806312e-06, |
|
"loss": 0.6444, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.9422654099393534, |
|
"learning_rate": 4.5829927360311224e-06, |
|
"loss": 0.619, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.8652896213966386, |
|
"learning_rate": 4.5808359099352675e-06, |
|
"loss": 0.5979, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 1.088012826550748, |
|
"learning_rate": 4.578674030756364e-06, |
|
"loss": 0.6621, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.8470033759027679, |
|
"learning_rate": 4.576507103744299e-06, |
|
"loss": 0.6089, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.8949204387411863, |
|
"learning_rate": 4.574335134161219e-06, |
|
"loss": 0.6306, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.9395950836852421, |
|
"learning_rate": 4.572158127281516e-06, |
|
"loss": 0.6559, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.8035578632803461, |
|
"learning_rate": 4.569976088391813e-06, |
|
"loss": 0.6262, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.8860765726891832, |
|
"learning_rate": 4.567789022790953e-06, |
|
"loss": 0.6469, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.9318594734495474, |
|
"learning_rate": 4.565596935789987e-06, |
|
"loss": 0.6153, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.8747678520637625, |
|
"learning_rate": 4.5633998327121595e-06, |
|
"loss": 0.6666, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.8349863455158287, |
|
"learning_rate": 4.561197718892896e-06, |
|
"loss": 0.6017, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.8835445705707418, |
|
"learning_rate": 4.558990599679787e-06, |
|
"loss": 0.6418, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.8929109099342171, |
|
"learning_rate": 4.556778480432584e-06, |
|
"loss": 0.6326, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.8364797596101992, |
|
"learning_rate": 4.554561366523176e-06, |
|
"loss": 0.6782, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.9260600125305487, |
|
"learning_rate": 4.552339263335581e-06, |
|
"loss": 0.595, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.8599195336450693, |
|
"learning_rate": 4.550112176265937e-06, |
|
"loss": 0.5968, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.9826419950054966, |
|
"learning_rate": 4.54788011072248e-06, |
|
"loss": 0.6456, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.9203378835364787, |
|
"learning_rate": 4.5456430721255384e-06, |
|
"loss": 0.7066, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.8939684740473081, |
|
"learning_rate": 4.5434010659075165e-06, |
|
"loss": 0.647, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.8917246609121712, |
|
"learning_rate": 4.541154097512881e-06, |
|
"loss": 0.6568, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.9075876860734314, |
|
"learning_rate": 4.538902172398151e-06, |
|
"loss": 0.6798, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.8110519815726133, |
|
"learning_rate": 4.53664529603188e-06, |
|
"loss": 0.5882, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.8925518187452417, |
|
"learning_rate": 4.534383473894646e-06, |
|
"loss": 0.6181, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.9114329206202244, |
|
"learning_rate": 4.532116711479039e-06, |
|
"loss": 0.6026, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.846641811272952, |
|
"learning_rate": 4.529845014289642e-06, |
|
"loss": 0.5825, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.8740877163769825, |
|
"learning_rate": 4.527568387843025e-06, |
|
"loss": 0.6989, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.9447837582033838, |
|
"learning_rate": 4.525286837667726e-06, |
|
"loss": 0.6582, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.8674554766425348, |
|
"learning_rate": 4.523000369304244e-06, |
|
"loss": 0.6217, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.8500815565912371, |
|
"learning_rate": 4.520708988305014e-06, |
|
"loss": 0.5941, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 1.0164281137951803, |
|
"learning_rate": 4.518412700234407e-06, |
|
"loss": 0.6469, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.8420383857000564, |
|
"learning_rate": 4.516111510668707e-06, |
|
"loss": 0.6237, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.9412530438844999, |
|
"learning_rate": 4.513805425196103e-06, |
|
"loss": 0.6386, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.9175391505475148, |
|
"learning_rate": 4.511494449416671e-06, |
|
"loss": 0.6372, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.8965680547901378, |
|
"learning_rate": 4.509178588942365e-06, |
|
"loss": 0.6398, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 1.161349181913427, |
|
"learning_rate": 4.506857849396998e-06, |
|
"loss": 0.7087, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.9562290751417603, |
|
"learning_rate": 4.504532236416234e-06, |
|
"loss": 0.6969, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.9026029147011317, |
|
"learning_rate": 4.502201755647571e-06, |
|
"loss": 0.6637, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.9230372085626913, |
|
"learning_rate": 4.499866412750324e-06, |
|
"loss": 0.618, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.9413375641434438, |
|
"learning_rate": 4.4975262133956235e-06, |
|
"loss": 0.6054, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.9455820405339215, |
|
"learning_rate": 4.495181163266384e-06, |
|
"loss": 0.6606, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.9648271805379507, |
|
"learning_rate": 4.492831268057307e-06, |
|
"loss": 0.6676, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.8351069990624824, |
|
"learning_rate": 4.490476533474854e-06, |
|
"loss": 0.6479, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.9101496071500632, |
|
"learning_rate": 4.488116965237244e-06, |
|
"loss": 0.6413, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.8280285227695013, |
|
"learning_rate": 4.485752569074429e-06, |
|
"loss": 0.6223, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.8992598915107023, |
|
"learning_rate": 4.4833833507280884e-06, |
|
"loss": 0.6742, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.8923843348896123, |
|
"learning_rate": 4.48100931595161e-06, |
|
"loss": 0.5947, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.8664845993947492, |
|
"learning_rate": 4.478630470510078e-06, |
|
"loss": 0.6369, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.8726591582806938, |
|
"learning_rate": 4.476246820180259e-06, |
|
"loss": 0.6554, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.8410498499068318, |
|
"learning_rate": 4.473858370750589e-06, |
|
"loss": 0.6293, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.8755653995170062, |
|
"learning_rate": 4.4714651280211555e-06, |
|
"loss": 0.634, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.9524417549442475, |
|
"learning_rate": 4.469067097803689e-06, |
|
"loss": 0.6435, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.7960663615043987, |
|
"learning_rate": 4.466664285921543e-06, |
|
"loss": 0.6331, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.8888628596240454, |
|
"learning_rate": 4.464256698209685e-06, |
|
"loss": 0.6552, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.9118498141803711, |
|
"learning_rate": 4.461844340514678e-06, |
|
"loss": 0.5826, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.854812295494578, |
|
"learning_rate": 4.459427218694671e-06, |
|
"loss": 0.6156, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.8894677182707545, |
|
"learning_rate": 4.457005338619379e-06, |
|
"loss": 0.6316, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.92320969218264, |
|
"learning_rate": 4.454578706170075e-06, |
|
"loss": 0.6139, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.9227943336374008, |
|
"learning_rate": 4.452147327239571e-06, |
|
"loss": 0.6377, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.8967111083779129, |
|
"learning_rate": 4.4497112077322045e-06, |
|
"loss": 0.6004, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.8705464350911942, |
|
"learning_rate": 4.447270353563828e-06, |
|
"loss": 0.7149, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.9047682842604887, |
|
"learning_rate": 4.444824770661788e-06, |
|
"loss": 0.6214, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.8332571186935797, |
|
"learning_rate": 4.442374464964916e-06, |
|
"loss": 0.61, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.9191644952904162, |
|
"learning_rate": 4.439919442423513e-06, |
|
"loss": 0.6974, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.8351147721256271, |
|
"learning_rate": 4.437459708999332e-06, |
|
"loss": 0.6633, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.892336463583514, |
|
"learning_rate": 4.434995270665569e-06, |
|
"loss": 0.6406, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8650896389402535, |
|
"learning_rate": 4.432526133406843e-06, |
|
"loss": 0.6345, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8153513627930788, |
|
"learning_rate": 4.430052303219185e-06, |
|
"loss": 0.6188, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8565464920766526, |
|
"learning_rate": 4.42757378611002e-06, |
|
"loss": 0.631, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8399364160683044, |
|
"learning_rate": 4.425090588098158e-06, |
|
"loss": 0.6577, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8912303770048037, |
|
"learning_rate": 4.422602715213774e-06, |
|
"loss": 0.6249, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8665388252830907, |
|
"learning_rate": 4.4201101734983965e-06, |
|
"loss": 0.6273, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8464141917671866, |
|
"learning_rate": 4.4176129690048905e-06, |
|
"loss": 0.5851, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8811314655875526, |
|
"learning_rate": 4.415111107797445e-06, |
|
"loss": 0.6214, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8083079113454086, |
|
"learning_rate": 4.412604595951558e-06, |
|
"loss": 0.6344, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8720341388764445, |
|
"learning_rate": 4.410093439554019e-06, |
|
"loss": 0.632, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8502476104259754, |
|
"learning_rate": 4.407577644702897e-06, |
|
"loss": 0.6184, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.8813994548710394, |
|
"learning_rate": 4.405057217507527e-06, |
|
"loss": 0.637, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.917114660031616, |
|
"learning_rate": 4.40253216408849e-06, |
|
"loss": 0.7224, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.8092450819851889, |
|
"learning_rate": 4.400002490577604e-06, |
|
"loss": 0.6247, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.9134141894112806, |
|
"learning_rate": 4.397468203117905e-06, |
|
"loss": 0.6813, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.8877869537323849, |
|
"learning_rate": 4.394929307863633e-06, |
|
"loss": 0.6205, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.866374769727955, |
|
"learning_rate": 4.392385810980218e-06, |
|
"loss": 0.6463, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 1.071041707971838, |
|
"learning_rate": 4.3898377186442665e-06, |
|
"loss": 0.6571, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.8411856812284965, |
|
"learning_rate": 4.38728503704354e-06, |
|
"loss": 0.6411, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.851584061852996, |
|
"learning_rate": 4.38472777237695e-06, |
|
"loss": 0.6966, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.8291807888706711, |
|
"learning_rate": 4.382165930854534e-06, |
|
"loss": 0.6025, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.8079509788666187, |
|
"learning_rate": 4.379599518697444e-06, |
|
"loss": 0.6461, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.8116984175253492, |
|
"learning_rate": 4.377028542137933e-06, |
|
"loss": 0.6172, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.8194670399241157, |
|
"learning_rate": 4.374453007419336e-06, |
|
"loss": 0.6244, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.8544501253810839, |
|
"learning_rate": 4.371872920796059e-06, |
|
"loss": 0.6085, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.8597333545854512, |
|
"learning_rate": 4.369288288533561e-06, |
|
"loss": 0.6757, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.912646868045953, |
|
"learning_rate": 4.366699116908339e-06, |
|
"loss": 0.6487, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.8476447032290744, |
|
"learning_rate": 4.364105412207914e-06, |
|
"loss": 0.5805, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.9241438674933707, |
|
"learning_rate": 4.3615071807308165e-06, |
|
"loss": 0.6758, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.8122777487524169, |
|
"learning_rate": 4.358904428786565e-06, |
|
"loss": 0.6153, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.9018439615133529, |
|
"learning_rate": 4.356297162695662e-06, |
|
"loss": 0.6453, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.9627678054923827, |
|
"learning_rate": 4.353685388789567e-06, |
|
"loss": 0.6282, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.8406101285686195, |
|
"learning_rate": 4.351069113410689e-06, |
|
"loss": 0.6078, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.835904176007674, |
|
"learning_rate": 4.348448342912365e-06, |
|
"loss": 0.6436, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.8568187537534087, |
|
"learning_rate": 4.345823083658855e-06, |
|
"loss": 0.6677, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.8347696865421858, |
|
"learning_rate": 4.34319334202531e-06, |
|
"loss": 0.6171, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.8348209799902971, |
|
"learning_rate": 4.340559124397774e-06, |
|
"loss": 0.6239, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.8852176634720593, |
|
"learning_rate": 4.3379204371731555e-06, |
|
"loss": 0.6573, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.8637190549045177, |
|
"learning_rate": 4.335277286759218e-06, |
|
"loss": 0.6394, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.8678111098963714, |
|
"learning_rate": 4.332629679574566e-06, |
|
"loss": 0.6011, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.9549679439132955, |
|
"learning_rate": 4.3299776220486235e-06, |
|
"loss": 0.6205, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.834971592773752, |
|
"learning_rate": 4.3273211206216235e-06, |
|
"loss": 0.6121, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.8863673003507534, |
|
"learning_rate": 4.324660181744589e-06, |
|
"loss": 0.6118, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.8984092199591801, |
|
"learning_rate": 4.321994811879321e-06, |
|
"loss": 0.6897, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.8880703411725283, |
|
"learning_rate": 4.319325017498379e-06, |
|
"loss": 0.5674, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.8213603864968535, |
|
"learning_rate": 4.316650805085068e-06, |
|
"loss": 0.5959, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.8521624706722543, |
|
"learning_rate": 4.31397218113342e-06, |
|
"loss": 0.6439, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.8407539759191036, |
|
"learning_rate": 4.311289152148182e-06, |
|
"loss": 0.6495, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.8832549384112272, |
|
"learning_rate": 4.308601724644797e-06, |
|
"loss": 0.6291, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.9007160602566869, |
|
"learning_rate": 4.305909905149389e-06, |
|
"loss": 0.6219, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.8798855352674151, |
|
"learning_rate": 4.303213700198749e-06, |
|
"loss": 0.6365, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.841611556568396, |
|
"learning_rate": 4.300513116340317e-06, |
|
"loss": 0.6464, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.9363258404112693, |
|
"learning_rate": 4.297808160132165e-06, |
|
"loss": 0.6181, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.8390395946351076, |
|
"learning_rate": 4.295098838142985e-06, |
|
"loss": 0.6186, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.8925228288644416, |
|
"learning_rate": 4.292385156952069e-06, |
|
"loss": 0.5934, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.8557639299914508, |
|
"learning_rate": 4.289667123149296e-06, |
|
"loss": 0.6458, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.8078616638133473, |
|
"learning_rate": 4.2869447433351165e-06, |
|
"loss": 0.6045, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"eval_loss": 0.6127398610115051, |
|
"eval_runtime": 1883.8371, |
|
"eval_samples_per_second": 0.701, |
|
"eval_steps_per_second": 0.078, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.8362298388304847, |
|
"learning_rate": 4.284218024120531e-06, |
|
"loss": 0.706, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.9282283575285319, |
|
"learning_rate": 4.2814869721270805e-06, |
|
"loss": 0.7029, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.8047669518916841, |
|
"learning_rate": 4.278751593986826e-06, |
|
"loss": 0.6209, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.8454914208142845, |
|
"learning_rate": 4.276011896342336e-06, |
|
"loss": 0.621, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.8078879153619455, |
|
"learning_rate": 4.273267885846667e-06, |
|
"loss": 0.6198, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.8462035097672782, |
|
"learning_rate": 4.270519569163348e-06, |
|
"loss": 0.6244, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.816672394583137, |
|
"learning_rate": 4.267766952966369e-06, |
|
"loss": 0.5808, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.8937772002097515, |
|
"learning_rate": 4.2650100439401565e-06, |
|
"loss": 0.6484, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.8250884650740417, |
|
"learning_rate": 4.262248848779563e-06, |
|
"loss": 0.6666, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.8610024828211696, |
|
"learning_rate": 4.259483374189851e-06, |
|
"loss": 0.5824, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.9330039890470406, |
|
"learning_rate": 4.256713626886674e-06, |
|
"loss": 0.6528, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 1.0641883529059932, |
|
"learning_rate": 4.253939613596058e-06, |
|
"loss": 0.6483, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.8107458510268721, |
|
"learning_rate": 4.251161341054396e-06, |
|
"loss": 0.5966, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.8834018306419973, |
|
"learning_rate": 4.248378816008418e-06, |
|
"loss": 0.6007, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.8655653950439697, |
|
"learning_rate": 4.245592045215182e-06, |
|
"loss": 0.6102, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.8412296785814055, |
|
"learning_rate": 4.242801035442059e-06, |
|
"loss": 0.6269, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.8492279489616671, |
|
"learning_rate": 4.240005793466709e-06, |
|
"loss": 0.6272, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.8524964006471626, |
|
"learning_rate": 4.237206326077073e-06, |
|
"loss": 0.6157, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.9309187782618586, |
|
"learning_rate": 4.234402640071355e-06, |
|
"loss": 0.6323, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.8767215287912865, |
|
"learning_rate": 4.231594742257997e-06, |
|
"loss": 0.6546, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.8857382937100764, |
|
"learning_rate": 4.228782639455674e-06, |
|
"loss": 0.5967, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.8878805691610596, |
|
"learning_rate": 4.225966338493272e-06, |
|
"loss": 0.6291, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.8533831355843948, |
|
"learning_rate": 4.223145846209868e-06, |
|
"loss": 0.6365, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.8780516405279318, |
|
"learning_rate": 4.220321169454723e-06, |
|
"loss": 0.621, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.8555312461179767, |
|
"learning_rate": 4.217492315087255e-06, |
|
"loss": 0.647, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.8609895152612035, |
|
"learning_rate": 4.214659289977028e-06, |
|
"loss": 0.5775, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.8296488468656131, |
|
"learning_rate": 4.211822101003734e-06, |
|
"loss": 0.6499, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.821402081187483, |
|
"learning_rate": 4.2089807550571786e-06, |
|
"loss": 0.5918, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.918003540760866, |
|
"learning_rate": 4.20613525903726e-06, |
|
"loss": 0.6449, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.8811180879144805, |
|
"learning_rate": 4.203285619853954e-06, |
|
"loss": 0.6343, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.8032697641623496, |
|
"learning_rate": 4.200431844427299e-06, |
|
"loss": 0.5849, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.9224510847524761, |
|
"learning_rate": 4.197573939687378e-06, |
|
"loss": 0.6671, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.8597523417116901, |
|
"learning_rate": 4.1947119125743e-06, |
|
"loss": 0.6001, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.858547476629971, |
|
"learning_rate": 4.191845770038186e-06, |
|
"loss": 0.6614, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.8316297295683677, |
|
"learning_rate": 4.188975519039151e-06, |
|
"loss": 0.7046, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.8750165143725187, |
|
"learning_rate": 4.1861011665472864e-06, |
|
"loss": 0.5857, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.8504473498402192, |
|
"learning_rate": 4.183222719542643e-06, |
|
"loss": 0.6212, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.8861097216987638, |
|
"learning_rate": 4.180340185015216e-06, |
|
"loss": 0.6298, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.8764506901623238, |
|
"learning_rate": 4.177453569964925e-06, |
|
"loss": 0.6299, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.8954878186353425, |
|
"learning_rate": 4.174562881401602e-06, |
|
"loss": 0.6435, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.8004194533630808, |
|
"learning_rate": 4.171668126344968e-06, |
|
"loss": 0.6462, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.8154879145019431, |
|
"learning_rate": 4.168769311824619e-06, |
|
"loss": 0.6249, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.8753691348791529, |
|
"learning_rate": 4.1658664448800105e-06, |
|
"loss": 0.6409, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.9129753096435906, |
|
"learning_rate": 4.162959532560438e-06, |
|
"loss": 0.6042, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.8135110815096118, |
|
"learning_rate": 4.160048581925022e-06, |
|
"loss": 0.6103, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.853647599461546, |
|
"learning_rate": 4.1571336000426865e-06, |
|
"loss": 0.6703, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.8443698357876949, |
|
"learning_rate": 4.154214593992149e-06, |
|
"loss": 0.5785, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.8279089687921122, |
|
"learning_rate": 4.151291570861897e-06, |
|
"loss": 0.5805, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.7909744989541445, |
|
"learning_rate": 4.1483645377501726e-06, |
|
"loss": 0.6439, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.8201980138747228, |
|
"learning_rate": 4.145433501764958e-06, |
|
"loss": 0.6641, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.8926818528101442, |
|
"learning_rate": 4.142498470023952e-06, |
|
"loss": 0.6016, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.8717064096758335, |
|
"learning_rate": 4.139559449654561e-06, |
|
"loss": 0.6323, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.850010692441458, |
|
"learning_rate": 4.136616447793874e-06, |
|
"loss": 0.5979, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.8329484168696811, |
|
"learning_rate": 4.133669471588651e-06, |
|
"loss": 0.6214, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.8319091624927326, |
|
"learning_rate": 4.130718528195303e-06, |
|
"loss": 0.652, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.9140031940509377, |
|
"learning_rate": 4.127763624779873e-06, |
|
"loss": 0.6062, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.904103009449078, |
|
"learning_rate": 4.124804768518022e-06, |
|
"loss": 0.6201, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.8410022890114313, |
|
"learning_rate": 4.121841966595009e-06, |
|
"loss": 0.6086, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.8940417054615136, |
|
"learning_rate": 4.118875226205677e-06, |
|
"loss": 0.6522, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.9155247713200788, |
|
"learning_rate": 4.11590455455443e-06, |
|
"loss": 0.5841, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.8745417589672394, |
|
"learning_rate": 4.11292995885522e-06, |
|
"loss": 0.6258, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.8552912069311052, |
|
"learning_rate": 4.1099514463315286e-06, |
|
"loss": 0.6056, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.8366769779121708, |
|
"learning_rate": 4.106969024216348e-06, |
|
"loss": 0.6652, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.8135609584995043, |
|
"learning_rate": 4.103982699752167e-06, |
|
"loss": 0.6228, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.8377970374149817, |
|
"learning_rate": 4.1009924801909475e-06, |
|
"loss": 0.6313, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.8318819640389654, |
|
"learning_rate": 4.097998372794111e-06, |
|
"loss": 0.6533, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.8758708325816799, |
|
"learning_rate": 4.095000384832522e-06, |
|
"loss": 0.6517, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.8190389257290644, |
|
"learning_rate": 4.091998523586466e-06, |
|
"loss": 0.6052, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.8446608189730106, |
|
"learning_rate": 4.088992796345637e-06, |
|
"loss": 0.6314, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.8652594802837115, |
|
"learning_rate": 4.085983210409114e-06, |
|
"loss": 0.6195, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.8351116333122931, |
|
"learning_rate": 4.0829697730853505e-06, |
|
"loss": 0.624, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.8704459285009329, |
|
"learning_rate": 4.0799524916921475e-06, |
|
"loss": 0.6504, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.903077675003753, |
|
"learning_rate": 4.076931373556646e-06, |
|
"loss": 0.6218, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 1.2268852729835975, |
|
"learning_rate": 4.073906426015301e-06, |
|
"loss": 0.6487, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.8429954352584018, |
|
"learning_rate": 4.0708776564138685e-06, |
|
"loss": 0.6383, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.8555120915062305, |
|
"learning_rate": 4.067845072107384e-06, |
|
"loss": 0.6088, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.8550901642574732, |
|
"learning_rate": 4.064808680460149e-06, |
|
"loss": 0.6196, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.8040403373015266, |
|
"learning_rate": 4.061768488845707e-06, |
|
"loss": 0.6224, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.8175245764774571, |
|
"learning_rate": 4.058724504646834e-06, |
|
"loss": 0.5801, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.8718375181238908, |
|
"learning_rate": 4.055676735255513e-06, |
|
"loss": 0.6243, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.7799454280424266, |
|
"learning_rate": 4.0526251880729205e-06, |
|
"loss": 0.6152, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.8395548950336932, |
|
"learning_rate": 4.049569870509404e-06, |
|
"loss": 0.6391, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.8893708670958256, |
|
"learning_rate": 4.046510789984471e-06, |
|
"loss": 0.6104, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.7958226662626613, |
|
"learning_rate": 4.043447953926763e-06, |
|
"loss": 0.6467, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.8439920849678747, |
|
"learning_rate": 4.040381369774045e-06, |
|
"loss": 0.5917, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.8561339618371849, |
|
"learning_rate": 4.03731104497318e-06, |
|
"loss": 0.6868, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.880827080603074, |
|
"learning_rate": 4.034236986980119e-06, |
|
"loss": 0.6217, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.9125142796540433, |
|
"learning_rate": 4.031159203259876e-06, |
|
"loss": 0.6241, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.8470398941235018, |
|
"learning_rate": 4.028077701286512e-06, |
|
"loss": 0.689, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.8695577595082639, |
|
"learning_rate": 4.02499248854312e-06, |
|
"loss": 0.6194, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.8400799986259637, |
|
"learning_rate": 4.021903572521802e-06, |
|
"loss": 0.6123, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.8364906173856129, |
|
"learning_rate": 4.018810960723654e-06, |
|
"loss": 0.6403, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.8426020752299488, |
|
"learning_rate": 4.015714660658745e-06, |
|
"loss": 0.6261, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.8639159393487216, |
|
"learning_rate": 4.012614679846103e-06, |
|
"loss": 0.6469, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.8426962439957932, |
|
"learning_rate": 4.009511025813694e-06, |
|
"loss": 0.6292, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.8311029210618062, |
|
"learning_rate": 4.0064037060984015e-06, |
|
"loss": 0.6388, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.8530506717592211, |
|
"learning_rate": 4.003292728246015e-06, |
|
"loss": 0.6716, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.8440602463846489, |
|
"learning_rate": 4.000178099811203e-06, |
|
"loss": 0.5825, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.8527447447182788, |
|
"learning_rate": 3.997059828357501e-06, |
|
"loss": 0.6405, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.8141210608713626, |
|
"learning_rate": 3.993937921457292e-06, |
|
"loss": 0.5794, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.8167088274295972, |
|
"learning_rate": 3.990812386691786e-06, |
|
"loss": 0.5785, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.8416848345796162, |
|
"learning_rate": 3.987683231651003e-06, |
|
"loss": 0.6082, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.841934252054496, |
|
"learning_rate": 3.984550463933754e-06, |
|
"loss": 0.6462, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.8803588200069083, |
|
"learning_rate": 3.981414091147626e-06, |
|
"loss": 0.5668, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.8446219400321544, |
|
"learning_rate": 3.978274120908957e-06, |
|
"loss": 0.6334, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.837550005498191, |
|
"learning_rate": 3.975130560842821e-06, |
|
"loss": 0.6179, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 1.004865035425333, |
|
"learning_rate": 3.971983418583012e-06, |
|
"loss": 0.6747, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.8557125045552572, |
|
"learning_rate": 3.968832701772022e-06, |
|
"loss": 0.6069, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.8832288543033557, |
|
"learning_rate": 3.965678418061023e-06, |
|
"loss": 0.6512, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 1.050854968266542, |
|
"learning_rate": 3.962520575109849e-06, |
|
"loss": 0.6139, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.8367647706094705, |
|
"learning_rate": 3.9593591805869755e-06, |
|
"loss": 0.6423, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.9902929127011927, |
|
"learning_rate": 3.956194242169506e-06, |
|
"loss": 0.5839, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.8538710862131312, |
|
"learning_rate": 3.953025767543148e-06, |
|
"loss": 0.6355, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.8406210033884886, |
|
"learning_rate": 3.949853764402196e-06, |
|
"loss": 0.6368, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.9035073071147527, |
|
"learning_rate": 3.946678240449515e-06, |
|
"loss": 0.6007, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.8519818744840768, |
|
"learning_rate": 3.943499203396517e-06, |
|
"loss": 0.6058, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.8848827749849268, |
|
"learning_rate": 3.940316660963147e-06, |
|
"loss": 0.6592, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.8514691094807268, |
|
"learning_rate": 3.937130620877863e-06, |
|
"loss": 0.6305, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.8673549125405013, |
|
"learning_rate": 3.933941090877615e-06, |
|
"loss": 0.6525, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.8170507594506436, |
|
"learning_rate": 3.93074807870783e-06, |
|
"loss": 0.6054, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.9160705523256952, |
|
"learning_rate": 3.927551592122389e-06, |
|
"loss": 0.646, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.8218544345647915, |
|
"learning_rate": 3.92435163888361e-06, |
|
"loss": 0.627, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.9097780389616746, |
|
"learning_rate": 3.921148226762231e-06, |
|
"loss": 0.6122, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.7955169076765822, |
|
"learning_rate": 3.9179413635373895e-06, |
|
"loss": 0.6221, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.9196174957020604, |
|
"learning_rate": 3.914731056996604e-06, |
|
"loss": 0.6248, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.8564166348680387, |
|
"learning_rate": 3.911517314935752e-06, |
|
"loss": 0.6629, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.861955774140363, |
|
"learning_rate": 3.908300145159055e-06, |
|
"loss": 0.5845, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.8237943505322816, |
|
"learning_rate": 3.905079555479061e-06, |
|
"loss": 0.5752, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.8254490831345543, |
|
"learning_rate": 3.90185555371662e-06, |
|
"loss": 0.6124, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.7892820407938396, |
|
"learning_rate": 3.898628147700869e-06, |
|
"loss": 0.6145, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.8608765813825437, |
|
"learning_rate": 3.895397345269211e-06, |
|
"loss": 0.6205, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.8395603666500568, |
|
"learning_rate": 3.892163154267295e-06, |
|
"loss": 0.6649, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.9131396543054635, |
|
"learning_rate": 3.888925582549006e-06, |
|
"loss": 0.6121, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.9090584462851223, |
|
"learning_rate": 3.88568463797643e-06, |
|
"loss": 0.5809, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.7938541089543004, |
|
"learning_rate": 3.882440328419849e-06, |
|
"loss": 0.6079, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.8068071932206416, |
|
"learning_rate": 3.879192661757715e-06, |
|
"loss": 0.5728, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.8578705877241041, |
|
"learning_rate": 3.875941645876631e-06, |
|
"loss": 0.6426, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.9087218932248154, |
|
"learning_rate": 3.872687288671335e-06, |
|
"loss": 0.5947, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.8484390918118917, |
|
"learning_rate": 3.869429598044679e-06, |
|
"loss": 0.5985, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.9202346447435071, |
|
"learning_rate": 3.866168581907609e-06, |
|
"loss": 0.6258, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.9340901392950748, |
|
"learning_rate": 3.8629042481791475e-06, |
|
"loss": 0.6289, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.8635034011239118, |
|
"learning_rate": 3.859636604786372e-06, |
|
"loss": 0.643, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.8415497288944668, |
|
"learning_rate": 3.856365659664399e-06, |
|
"loss": 0.6341, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.8687282559878431, |
|
"learning_rate": 3.853091420756362e-06, |
|
"loss": 0.6481, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 1.0072749621790154, |
|
"learning_rate": 3.849813896013392e-06, |
|
"loss": 0.6715, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.8246193635953545, |
|
"learning_rate": 3.846533093394601e-06, |
|
"loss": 0.5799, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.8558794445823575, |
|
"learning_rate": 3.8432490208670605e-06, |
|
"loss": 0.6019, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.8349778480098637, |
|
"learning_rate": 3.839961686405782e-06, |
|
"loss": 0.6254, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.8199486664572769, |
|
"learning_rate": 3.836671097993698e-06, |
|
"loss": 0.6129, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.9459661992839397, |
|
"learning_rate": 3.833377263621646e-06, |
|
"loss": 0.6327, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.8250918846236105, |
|
"learning_rate": 3.830080191288342e-06, |
|
"loss": 0.6636, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.9390093353108422, |
|
"learning_rate": 3.826779889000366e-06, |
|
"loss": 0.6384, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.7978120610077674, |
|
"learning_rate": 3.823476364772143e-06, |
|
"loss": 0.5865, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.9061114135678785, |
|
"learning_rate": 3.82016962662592e-06, |
|
"loss": 0.6342, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.8072077947758938, |
|
"learning_rate": 3.816859682591752e-06, |
|
"loss": 0.5847, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.8123266086238162, |
|
"learning_rate": 3.8135465407074756e-06, |
|
"loss": 0.5951, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.8400833347987473, |
|
"learning_rate": 3.810230209018694e-06, |
|
"loss": 0.6343, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.8774221118702056, |
|
"learning_rate": 3.8069106955787593e-06, |
|
"loss": 0.6027, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.8469674582077675, |
|
"learning_rate": 3.8035880084487454e-06, |
|
"loss": 0.6427, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.8713388963235409, |
|
"learning_rate": 3.8002621556974367e-06, |
|
"loss": 0.6498, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.8494210390587402, |
|
"learning_rate": 3.796933145401304e-06, |
|
"loss": 0.7154, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.8273270122789752, |
|
"learning_rate": 3.7936009856444854e-06, |
|
"loss": 0.5999, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.8382150794392165, |
|
"learning_rate": 3.790265684518767e-06, |
|
"loss": 0.626, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.788293287889455, |
|
"learning_rate": 3.7869272501235644e-06, |
|
"loss": 0.595, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.8272921725994223, |
|
"learning_rate": 3.7835856905659015e-06, |
|
"loss": 0.6007, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.8622000588979104, |
|
"learning_rate": 3.7802410139603908e-06, |
|
"loss": 0.599, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.8358507677494942, |
|
"learning_rate": 3.7768932284292147e-06, |
|
"loss": 0.641, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.8625794861605586, |
|
"learning_rate": 3.773542342102105e-06, |
|
"loss": 0.5919, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.8416399086204703, |
|
"learning_rate": 3.770188363116324e-06, |
|
"loss": 0.5554, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.8418267359188863, |
|
"learning_rate": 3.766831299616644e-06, |
|
"loss": 0.6304, |
|
"step": 677 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.8762169067425597, |
|
"learning_rate": 3.7634711597553274e-06, |
|
"loss": 0.6476, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.9404479760658981, |
|
"learning_rate": 3.7601079516921076e-06, |
|
"loss": 0.6499, |
|
"step": 679 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.8947001559317359, |
|
"learning_rate": 3.7567416835941674e-06, |
|
"loss": 0.69, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.7966747108540816, |
|
"learning_rate": 3.7533723636361236e-06, |
|
"loss": 0.6372, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.8544936589074825, |
|
"learning_rate": 3.7500000000000005e-06, |
|
"loss": 0.5779, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.8178872102657411, |
|
"learning_rate": 3.746624600875216e-06, |
|
"loss": 0.5636, |
|
"step": 683 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.8332061128370889, |
|
"learning_rate": 3.7432461744585595e-06, |
|
"loss": 0.5995, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.863474799322471, |
|
"learning_rate": 3.7398647289541703e-06, |
|
"loss": 0.629, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.865438085828472, |
|
"learning_rate": 3.736480272573519e-06, |
|
"loss": 0.638, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.8433955975404765, |
|
"learning_rate": 3.7330928135353904e-06, |
|
"loss": 0.6567, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.9145902886360865, |
|
"learning_rate": 3.7297023600658586e-06, |
|
"loss": 0.6144, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.8556239801373773, |
|
"learning_rate": 3.7263089203982698e-06, |
|
"loss": 0.6016, |
|
"step": 689 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.8208625440531927, |
|
"learning_rate": 3.722912502773224e-06, |
|
"loss": 0.644, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.9949073359631555, |
|
"learning_rate": 3.719513115438548e-06, |
|
"loss": 0.6436, |
|
"step": 691 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.8308047358059517, |
|
"learning_rate": 3.716110766649285e-06, |
|
"loss": 0.5918, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.8488700445625043, |
|
"learning_rate": 3.712705464667667e-06, |
|
"loss": 0.5992, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.8496884126493682, |
|
"learning_rate": 3.7092972177630998e-06, |
|
"loss": 0.652, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.8883418066922215, |
|
"learning_rate": 3.7058860342121385e-06, |
|
"loss": 0.5873, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.8400599741065571, |
|
"learning_rate": 3.7024719222984696e-06, |
|
"loss": 0.6079, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.8438768318930868, |
|
"learning_rate": 3.699054890312892e-06, |
|
"loss": 0.6272, |
|
"step": 697 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.8881813441097738, |
|
"learning_rate": 3.695634946553296e-06, |
|
"loss": 0.602, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.8390601326013466, |
|
"learning_rate": 3.6922120993246406e-06, |
|
"loss": 0.6486, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.8182123415184759, |
|
"learning_rate": 3.6887863569389388e-06, |
|
"loss": 0.6081, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.825114071415925, |
|
"learning_rate": 3.6853577277152295e-06, |
|
"loss": 0.6142, |
|
"step": 701 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.8525114312935935, |
|
"learning_rate": 3.681926219979568e-06, |
|
"loss": 0.6433, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.9146597445291013, |
|
"learning_rate": 3.6784918420649952e-06, |
|
"loss": 0.6623, |
|
"step": 703 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.8462781824353365, |
|
"learning_rate": 3.675054602311522e-06, |
|
"loss": 0.66, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.9014059561209856, |
|
"learning_rate": 3.6716145090661115e-06, |
|
"loss": 0.5729, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.9206429639224136, |
|
"learning_rate": 3.6681715706826555e-06, |
|
"loss": 0.6226, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.8518509022390937, |
|
"learning_rate": 3.6647257955219535e-06, |
|
"loss": 0.6359, |
|
"step": 707 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.9151576998752515, |
|
"learning_rate": 3.6612771919516944e-06, |
|
"loss": 0.6264, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.8746527924270058, |
|
"learning_rate": 3.6578257683464363e-06, |
|
"loss": 0.6372, |
|
"step": 709 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.8963857822510269, |
|
"learning_rate": 3.654371533087586e-06, |
|
"loss": 0.6381, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.893619853609527, |
|
"learning_rate": 3.6509144945633747e-06, |
|
"loss": 0.6118, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.8612944499918704, |
|
"learning_rate": 3.6474546611688446e-06, |
|
"loss": 0.6569, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.9267879655807548, |
|
"learning_rate": 3.643992041305824e-06, |
|
"loss": 0.6296, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.8324239177328866, |
|
"learning_rate": 3.640526643382908e-06, |
|
"loss": 0.596, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.8303516050409866, |
|
"learning_rate": 3.6370584758154366e-06, |
|
"loss": 0.627, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.8257066905266303, |
|
"learning_rate": 3.6335875470254763e-06, |
|
"loss": 0.6204, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.8617688384824457, |
|
"learning_rate": 3.6301138654418e-06, |
|
"loss": 0.5991, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.8331747116361415, |
|
"learning_rate": 3.626637439499864e-06, |
|
"loss": 0.6316, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.9111998000553265, |
|
"learning_rate": 3.623158277641789e-06, |
|
"loss": 0.5996, |
|
"step": 719 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.9347960312555601, |
|
"learning_rate": 3.6196763883163407e-06, |
|
"loss": 0.6208, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.8228735877621802, |
|
"learning_rate": 3.6161917799789076e-06, |
|
"loss": 0.6654, |
|
"step": 721 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.7692626652167038, |
|
"learning_rate": 3.612704461091481e-06, |
|
"loss": 0.6309, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.8558647649798884, |
|
"learning_rate": 3.6092144401226347e-06, |
|
"loss": 0.5716, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.8195467347471203, |
|
"learning_rate": 3.6057217255475034e-06, |
|
"loss": 0.5914, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.8194512413081951, |
|
"learning_rate": 3.6022263258477636e-06, |
|
"loss": 0.6197, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.8704175696506994, |
|
"learning_rate": 3.598728249511613e-06, |
|
"loss": 0.6095, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.8493811573838203, |
|
"learning_rate": 3.595227505033747e-06, |
|
"loss": 0.6232, |
|
"step": 727 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.790925203596828, |
|
"learning_rate": 3.5917241009153424e-06, |
|
"loss": 0.5643, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.8269684214873226, |
|
"learning_rate": 3.588218045664035e-06, |
|
"loss": 0.5822, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.8194321737982796, |
|
"learning_rate": 3.5847093477938955e-06, |
|
"loss": 0.6121, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 1.7164330849560612, |
|
"learning_rate": 3.5811980158254156e-06, |
|
"loss": 0.6089, |
|
"step": 731 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.8748550271603309, |
|
"learning_rate": 3.5776840582854815e-06, |
|
"loss": 0.5805, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.8636172206842424, |
|
"learning_rate": 3.5741674837073563e-06, |
|
"loss": 0.6128, |
|
"step": 733 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.8282087905244047, |
|
"learning_rate": 3.5706483006306567e-06, |
|
"loss": 0.6459, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.8848259311647253, |
|
"learning_rate": 3.5671265176013363e-06, |
|
"loss": 0.6277, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.8119988447523083, |
|
"learning_rate": 3.5636021431716604e-06, |
|
"loss": 0.623, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.8342845309664513, |
|
"learning_rate": 3.5600751859001873e-06, |
|
"loss": 0.6282, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.8410918679240155, |
|
"learning_rate": 3.556545654351749e-06, |
|
"loss": 0.591, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.8796336800622948, |
|
"learning_rate": 3.553013557097428e-06, |
|
"loss": 0.6341, |
|
"step": 739 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.7917875470526526, |
|
"learning_rate": 3.549478902714536e-06, |
|
"loss": 0.604, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.8337369916578986, |
|
"learning_rate": 3.545941699786596e-06, |
|
"loss": 0.6125, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.8490891695105536, |
|
"learning_rate": 3.542401956903321e-06, |
|
"loss": 0.5988, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.8452367560916697, |
|
"learning_rate": 3.5388596826605885e-06, |
|
"loss": 0.5787, |
|
"step": 743 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.8688484430410042, |
|
"learning_rate": 3.5353148856604265e-06, |
|
"loss": 0.6813, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.8286016233816287, |
|
"learning_rate": 3.531767574510987e-06, |
|
"loss": 0.6342, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.8515417361722135, |
|
"learning_rate": 3.5282177578265295e-06, |
|
"loss": 0.6163, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.788811099590474, |
|
"learning_rate": 3.5246654442273952e-06, |
|
"loss": 0.5822, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.799045340559827, |
|
"learning_rate": 3.521110642339991e-06, |
|
"loss": 0.6142, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.8437467452590093, |
|
"learning_rate": 3.5175533607967656e-06, |
|
"loss": 0.5822, |
|
"step": 749 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.8563735262905348, |
|
"learning_rate": 3.513993608236188e-06, |
|
"loss": 0.6669, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.8355492451040473, |
|
"learning_rate": 3.51043139330273e-06, |
|
"loss": 0.6051, |
|
"step": 751 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.9036193008645055, |
|
"learning_rate": 3.5068667246468437e-06, |
|
"loss": 0.5772, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.9220169686470918, |
|
"learning_rate": 3.503299610924935e-06, |
|
"loss": 0.5966, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.8451566800102057, |
|
"learning_rate": 3.499730060799352e-06, |
|
"loss": 0.6292, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.9102528483160476, |
|
"learning_rate": 3.496158082938359e-06, |
|
"loss": 0.5723, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 1.0868296175811145, |
|
"learning_rate": 3.492583686016113e-06, |
|
"loss": 0.6884, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.8535175516558623, |
|
"learning_rate": 3.4890068787126475e-06, |
|
"loss": 0.5513, |
|
"step": 757 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.8854803755423458, |
|
"learning_rate": 3.485427669713849e-06, |
|
"loss": 0.5858, |
|
"step": 758 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.8274288286764514, |
|
"learning_rate": 3.481846067711436e-06, |
|
"loss": 0.5906, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.8383612336715415, |
|
"learning_rate": 3.4782620814029376e-06, |
|
"loss": 0.6489, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.8219239495153828, |
|
"learning_rate": 3.474675719491675e-06, |
|
"loss": 0.6328, |
|
"step": 761 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.87449995235705, |
|
"learning_rate": 3.471086990686737e-06, |
|
"loss": 0.5782, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.8486500828100978, |
|
"learning_rate": 3.4674959037029593e-06, |
|
"loss": 0.6338, |
|
"step": 763 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.8464185221334688, |
|
"learning_rate": 3.4639024672609045e-06, |
|
"loss": 0.616, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.8795568484465794, |
|
"learning_rate": 3.4603066900868425e-06, |
|
"loss": 0.62, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.7898963360806371, |
|
"learning_rate": 3.4567085809127247e-06, |
|
"loss": 0.6165, |
|
"step": 766 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.834736333485077, |
|
"learning_rate": 3.4531081484761676e-06, |
|
"loss": 0.6291, |
|
"step": 767 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.8198807461377275, |
|
"learning_rate": 3.4495054015204282e-06, |
|
"loss": 0.6341, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.8089496064861047, |
|
"learning_rate": 3.4459003487943842e-06, |
|
"loss": 0.6323, |
|
"step": 769 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.9015873348579728, |
|
"learning_rate": 3.4422929990525133e-06, |
|
"loss": 0.5976, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.8101825375577076, |
|
"learning_rate": 3.438683361054872e-06, |
|
"loss": 0.6088, |
|
"step": 771 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.8464641333673643, |
|
"learning_rate": 3.4350714435670706e-06, |
|
"loss": 0.6348, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.876759997482892, |
|
"learning_rate": 3.4314572553602577e-06, |
|
"loss": 0.6229, |
|
"step": 773 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.7958084255478849, |
|
"learning_rate": 3.427840805211095e-06, |
|
"loss": 0.6231, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.8287287958310313, |
|
"learning_rate": 3.424222101901738e-06, |
|
"loss": 0.5963, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.8142217551737962, |
|
"learning_rate": 3.4206011542198115e-06, |
|
"loss": 0.6336, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.805828999419117, |
|
"learning_rate": 3.416977970958393e-06, |
|
"loss": 0.5936, |
|
"step": 777 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.8224599427095683, |
|
"learning_rate": 3.4133525609159883e-06, |
|
"loss": 0.6746, |
|
"step": 778 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.8304652958460922, |
|
"learning_rate": 3.409724932896509e-06, |
|
"loss": 0.6736, |
|
"step": 779 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.7912513015896195, |
|
"learning_rate": 3.406095095709254e-06, |
|
"loss": 0.6085, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.8246365476659197, |
|
"learning_rate": 3.4024630581688895e-06, |
|
"loss": 0.5968, |
|
"step": 781 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.8288445554081133, |
|
"learning_rate": 3.398828829095419e-06, |
|
"loss": 0.6241, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.8548262155784426, |
|
"learning_rate": 3.395192417314174e-06, |
|
"loss": 0.6391, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.7713543131187703, |
|
"learning_rate": 3.391553831655783e-06, |
|
"loss": 0.5857, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.8085787898759782, |
|
"learning_rate": 3.3879130809561543e-06, |
|
"loss": 0.5905, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.8110232754655301, |
|
"learning_rate": 3.384270174056454e-06, |
|
"loss": 0.5678, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.8263659965643595, |
|
"learning_rate": 3.3806251198030843e-06, |
|
"loss": 0.6386, |
|
"step": 787 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.9161733903902445, |
|
"learning_rate": 3.3769779270476614e-06, |
|
"loss": 0.6271, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.8281838229897066, |
|
"learning_rate": 3.3733286046469955e-06, |
|
"loss": 0.5625, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.8275110443615774, |
|
"learning_rate": 3.369677161463068e-06, |
|
"loss": 0.606, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.8291505968037055, |
|
"learning_rate": 3.3660236063630116e-06, |
|
"loss": 0.6052, |
|
"step": 791 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.8513874699652647, |
|
"learning_rate": 3.3623679482190834e-06, |
|
"loss": 0.6159, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.861694134467597, |
|
"learning_rate": 3.358710195908653e-06, |
|
"loss": 0.6178, |
|
"step": 793 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.8353375015177325, |
|
"learning_rate": 3.3550503583141726e-06, |
|
"loss": 0.563, |
|
"step": 794 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.8544792831170283, |
|
"learning_rate": 3.351388444323157e-06, |
|
"loss": 0.6131, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.8454973440583377, |
|
"learning_rate": 3.3477244628281667e-06, |
|
"loss": 0.5999, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.7948897675289834, |
|
"learning_rate": 3.3440584227267814e-06, |
|
"loss": 0.6587, |
|
"step": 797 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.8207217899114025, |
|
"learning_rate": 3.3403903329215777e-06, |
|
"loss": 0.5899, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.9250944862790339, |
|
"learning_rate": 3.3367202023201128e-06, |
|
"loss": 0.6491, |
|
"step": 799 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.7983017822178072, |
|
"learning_rate": 3.3330480398348988e-06, |
|
"loss": 0.5705, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.8699135194548229, |
|
"learning_rate": 3.3293738543833807e-06, |
|
"loss": 0.6335, |
|
"step": 801 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.84153902102691, |
|
"learning_rate": 3.3256976548879183e-06, |
|
"loss": 0.6058, |
|
"step": 802 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.8542237620487997, |
|
"learning_rate": 3.3220194502757602e-06, |
|
"loss": 0.6194, |
|
"step": 803 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.7924943868095786, |
|
"learning_rate": 3.3183392494790264e-06, |
|
"loss": 0.6274, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.8336114140592508, |
|
"learning_rate": 3.3146570614346814e-06, |
|
"loss": 0.5773, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.8639260233810916, |
|
"learning_rate": 3.3109728950845184e-06, |
|
"loss": 0.5685, |
|
"step": 806 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.8158312287141508, |
|
"learning_rate": 3.3072867593751335e-06, |
|
"loss": 0.5698, |
|
"step": 807 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.8751459349914905, |
|
"learning_rate": 3.303598663257904e-06, |
|
"loss": 0.6269, |
|
"step": 808 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.8053527944592563, |
|
"learning_rate": 3.2999086156889715e-06, |
|
"loss": 0.6604, |
|
"step": 809 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.8610616380496376, |
|
"learning_rate": 3.2962166256292116e-06, |
|
"loss": 0.6122, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.8426268663118117, |
|
"learning_rate": 3.292522702044221e-06, |
|
"loss": 0.6023, |
|
"step": 811 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.8283059132461468, |
|
"learning_rate": 3.288826853904289e-06, |
|
"loss": 0.6368, |
|
"step": 812 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.9098952954025415, |
|
"learning_rate": 3.285129090184381e-06, |
|
"loss": 0.5802, |
|
"step": 813 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.881481454188651, |
|
"learning_rate": 3.281429419864112e-06, |
|
"loss": 0.5848, |
|
"step": 814 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.8281945418633309, |
|
"learning_rate": 3.277727851927727e-06, |
|
"loss": 0.6097, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.882553902183949, |
|
"learning_rate": 3.2740243953640827e-06, |
|
"loss": 0.6596, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.8657763942244997, |
|
"learning_rate": 3.2703190591666174e-06, |
|
"loss": 0.6209, |
|
"step": 817 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.8358399157957145, |
|
"learning_rate": 3.2666118523333363e-06, |
|
"loss": 0.5879, |
|
"step": 818 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.8429824602810644, |
|
"learning_rate": 3.2629027838667863e-06, |
|
"loss": 0.6147, |
|
"step": 819 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.895622236703056, |
|
"learning_rate": 3.259191862774037e-06, |
|
"loss": 0.6342, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.869081721554779, |
|
"learning_rate": 3.2554790980666545e-06, |
|
"loss": 0.6288, |
|
"step": 821 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.8548842199692267, |
|
"learning_rate": 3.2517644987606827e-06, |
|
"loss": 0.6071, |
|
"step": 822 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.9265871658506936, |
|
"learning_rate": 3.2480480738766222e-06, |
|
"loss": 0.6125, |
|
"step": 823 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.8372005330678954, |
|
"learning_rate": 3.244329832439404e-06, |
|
"loss": 0.6184, |
|
"step": 824 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.7956763361747755, |
|
"learning_rate": 3.2406097834783724e-06, |
|
"loss": 0.6217, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.8181556551626763, |
|
"learning_rate": 3.236887936027261e-06, |
|
"loss": 0.5953, |
|
"step": 826 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.8857603345703071, |
|
"learning_rate": 3.2331642991241695e-06, |
|
"loss": 0.6547, |
|
"step": 827 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.8699553256972498, |
|
"learning_rate": 3.2294388818115447e-06, |
|
"loss": 0.6462, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.8260959526938845, |
|
"learning_rate": 3.225711693136156e-06, |
|
"loss": 0.6375, |
|
"step": 829 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.8585673049197299, |
|
"learning_rate": 3.221982742149075e-06, |
|
"loss": 0.5986, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.8401813928908182, |
|
"learning_rate": 3.2182520379056516e-06, |
|
"loss": 0.6087, |
|
"step": 831 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.8659784868378224, |
|
"learning_rate": 3.2145195894654947e-06, |
|
"loss": 0.6199, |
|
"step": 832 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.9223475931615208, |
|
"learning_rate": 3.210785405892448e-06, |
|
"loss": 0.5873, |
|
"step": 833 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.9447111876979355, |
|
"learning_rate": 3.207049496254569e-06, |
|
"loss": 0.5778, |
|
"step": 834 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.9766649870230963, |
|
"learning_rate": 3.203311869624107e-06, |
|
"loss": 0.6555, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.8477023018068901, |
|
"learning_rate": 3.199572535077481e-06, |
|
"loss": 0.5809, |
|
"step": 836 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.8206501838600238, |
|
"learning_rate": 3.195831501695256e-06, |
|
"loss": 0.6056, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.870594387289095, |
|
"learning_rate": 3.1920887785621233e-06, |
|
"loss": 0.6143, |
|
"step": 838 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.8149272829160001, |
|
"learning_rate": 3.1883443747668806e-06, |
|
"loss": 0.6117, |
|
"step": 839 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.8529746046688006, |
|
"learning_rate": 3.1845982994024006e-06, |
|
"loss": 0.6374, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.8412258694783452, |
|
"learning_rate": 3.180850561565621e-06, |
|
"loss": 0.6003, |
|
"step": 841 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.8444286299374963, |
|
"learning_rate": 3.1771011703575134e-06, |
|
"loss": 0.6059, |
|
"step": 842 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.8312527111815757, |
|
"learning_rate": 3.173350134883066e-06, |
|
"loss": 0.6153, |
|
"step": 843 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.8372205351537885, |
|
"learning_rate": 3.169597464251258e-06, |
|
"loss": 0.5939, |
|
"step": 844 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.807605821367219, |
|
"learning_rate": 3.1658431675750424e-06, |
|
"loss": 0.6059, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.7860979892611595, |
|
"learning_rate": 3.162087253971318e-06, |
|
"loss": 0.6629, |
|
"step": 846 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.940414176770529, |
|
"learning_rate": 3.1583297325609117e-06, |
|
"loss": 0.6028, |
|
"step": 847 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.8151977099439447, |
|
"learning_rate": 3.1545706124685553e-06, |
|
"loss": 0.6662, |
|
"step": 848 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.8471672058322735, |
|
"learning_rate": 3.1508099028228613e-06, |
|
"loss": 0.5856, |
|
"step": 849 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.8663713675701746, |
|
"learning_rate": 3.147047612756302e-06, |
|
"loss": 0.5481, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.7769840915544656, |
|
"learning_rate": 3.1432837514051893e-06, |
|
"loss": 0.5613, |
|
"step": 851 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.8592230044240439, |
|
"learning_rate": 3.139518327909651e-06, |
|
"loss": 0.6063, |
|
"step": 852 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.8597869095388183, |
|
"learning_rate": 3.1357513514136044e-06, |
|
"loss": 0.5747, |
|
"step": 853 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.8372404454667257, |
|
"learning_rate": 3.1319828310647437e-06, |
|
"loss": 0.6134, |
|
"step": 854 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.9298489053411231, |
|
"learning_rate": 3.1282127760145094e-06, |
|
"loss": 0.6078, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.9416390217217425, |
|
"learning_rate": 3.1244411954180677e-06, |
|
"loss": 0.5963, |
|
"step": 856 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.802868802927103, |
|
"learning_rate": 3.120668098434291e-06, |
|
"loss": 0.6058, |
|
"step": 857 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.9220598631922068, |
|
"learning_rate": 3.116893494225734e-06, |
|
"loss": 0.5966, |
|
"step": 858 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.9215503847795026, |
|
"learning_rate": 3.113117391958612e-06, |
|
"loss": 0.6112, |
|
"step": 859 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.7987198854171116, |
|
"learning_rate": 3.109339800802777e-06, |
|
"loss": 0.6388, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.8460741120385099, |
|
"learning_rate": 3.1055607299316966e-06, |
|
"loss": 0.5802, |
|
"step": 861 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.9262573376433733, |
|
"learning_rate": 3.1017801885224332e-06, |
|
"loss": 0.6559, |
|
"step": 862 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.8649803631458961, |
|
"learning_rate": 3.097998185755618e-06, |
|
"loss": 0.5749, |
|
"step": 863 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.8129582206860377, |
|
"learning_rate": 3.094214730815433e-06, |
|
"loss": 0.623, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.8894384638551944, |
|
"learning_rate": 3.0904298328895865e-06, |
|
"loss": 0.6156, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.9197272755662508, |
|
"learning_rate": 3.0866435011692884e-06, |
|
"loss": 0.6072, |
|
"step": 866 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.833467743832381, |
|
"learning_rate": 3.082855744849234e-06, |
|
"loss": 0.5942, |
|
"step": 867 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.8447839217824836, |
|
"learning_rate": 3.0790665731275764e-06, |
|
"loss": 0.5904, |
|
"step": 868 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.8347608532015025, |
|
"learning_rate": 3.0752759952059047e-06, |
|
"loss": 0.6212, |
|
"step": 869 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.7674030693122568, |
|
"learning_rate": 3.0714840202892243e-06, |
|
"loss": 0.6607, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.9057260645967865, |
|
"learning_rate": 3.0676906575859335e-06, |
|
"loss": 0.6377, |
|
"step": 871 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.7996224752255925, |
|
"learning_rate": 3.063895916307799e-06, |
|
"loss": 0.6332, |
|
"step": 872 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.8266387730228465, |
|
"learning_rate": 3.0600998056699367e-06, |
|
"loss": 0.5976, |
|
"step": 873 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.7758894383405645, |
|
"learning_rate": 3.056302334890786e-06, |
|
"loss": 0.5835, |
|
"step": 874 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.8070378254400489, |
|
"learning_rate": 3.0525035131920926e-06, |
|
"loss": 0.5945, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.8667704019187145, |
|
"learning_rate": 3.0487033497988794e-06, |
|
"loss": 0.605, |
|
"step": 876 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.8087365050450245, |
|
"learning_rate": 3.0449018539394274e-06, |
|
"loss": 0.6171, |
|
"step": 877 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.8106936486403415, |
|
"learning_rate": 3.0410990348452572e-06, |
|
"loss": 0.5878, |
|
"step": 878 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.8343865081867896, |
|
"learning_rate": 3.037294901751099e-06, |
|
"loss": 0.6159, |
|
"step": 879 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.9111920750589219, |
|
"learning_rate": 3.0334894638948753e-06, |
|
"loss": 0.6618, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.7905509417329776, |
|
"learning_rate": 3.0296827305176767e-06, |
|
"loss": 0.6421, |
|
"step": 881 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.8528082694556166, |
|
"learning_rate": 3.02587471086374e-06, |
|
"loss": 0.5966, |
|
"step": 882 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.8683617424120555, |
|
"learning_rate": 3.022065414180425e-06, |
|
"loss": 0.614, |
|
"step": 883 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.810261949769031, |
|
"learning_rate": 3.0182548497181946e-06, |
|
"loss": 0.6054, |
|
"step": 884 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.8458357510533312, |
|
"learning_rate": 3.0144430267305874e-06, |
|
"loss": 0.6012, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.8221353641749821, |
|
"learning_rate": 3.0106299544742013e-06, |
|
"loss": 0.6554, |
|
"step": 886 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.8230859903173897, |
|
"learning_rate": 3.006815642208665e-06, |
|
"loss": 0.6155, |
|
"step": 887 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.8897307384369784, |
|
"learning_rate": 3.0030000991966213e-06, |
|
"loss": 0.5947, |
|
"step": 888 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.8616094455160354, |
|
"learning_rate": 2.999183334703699e-06, |
|
"loss": 0.6392, |
|
"step": 889 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.8045566406491605, |
|
"learning_rate": 2.9953653579984945e-06, |
|
"loss": 0.613, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.8491995106395055, |
|
"learning_rate": 2.991546178352548e-06, |
|
"loss": 0.6732, |
|
"step": 891 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.8568125335081748, |
|
"learning_rate": 2.9877258050403214e-06, |
|
"loss": 0.6375, |
|
"step": 892 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.8403200791165655, |
|
"learning_rate": 2.9839042473391734e-06, |
|
"loss": 0.6355, |
|
"step": 893 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.8160289798230667, |
|
"learning_rate": 2.980081514529341e-06, |
|
"loss": 0.5816, |
|
"step": 894 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.8325072686110231, |
|
"learning_rate": 2.9762576158939127e-06, |
|
"loss": 0.6107, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.8532640950528609, |
|
"learning_rate": 2.972432560718811e-06, |
|
"loss": 0.6047, |
|
"step": 896 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.825222634002753, |
|
"learning_rate": 2.9686063582927636e-06, |
|
"loss": 0.6503, |
|
"step": 897 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.8374408245675915, |
|
"learning_rate": 2.964779017907287e-06, |
|
"loss": 0.5852, |
|
"step": 898 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.8009118608947914, |
|
"learning_rate": 2.9609505488566585e-06, |
|
"loss": 0.5691, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.8167294558750782, |
|
"learning_rate": 2.9571209604378993e-06, |
|
"loss": 0.6175, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.8232727662689424, |
|
"learning_rate": 2.9532902619507465e-06, |
|
"loss": 0.5732, |
|
"step": 901 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.846978796971522, |
|
"learning_rate": 2.9494584626976318e-06, |
|
"loss": 0.6105, |
|
"step": 902 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.8164610234787385, |
|
"learning_rate": 2.9456255719836648e-06, |
|
"loss": 0.566, |
|
"step": 903 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.8582764336563975, |
|
"learning_rate": 2.941791599116601e-06, |
|
"loss": 0.6275, |
|
"step": 904 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.777249993314137, |
|
"learning_rate": 2.9379565534068243e-06, |
|
"loss": 0.5805, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.8500028002405607, |
|
"learning_rate": 2.9341204441673267e-06, |
|
"loss": 0.5914, |
|
"step": 906 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.8175240407487714, |
|
"learning_rate": 2.93028328071368e-06, |
|
"loss": 0.586, |
|
"step": 907 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.8046556687399473, |
|
"learning_rate": 2.926445072364017e-06, |
|
"loss": 0.6353, |
|
"step": 908 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.8172777770063785, |
|
"learning_rate": 2.9226058284390093e-06, |
|
"loss": 0.6274, |
|
"step": 909 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.9129786089314338, |
|
"learning_rate": 2.9187655582618413e-06, |
|
"loss": 0.6398, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.8123191016829922, |
|
"learning_rate": 2.91492427115819e-06, |
|
"loss": 0.5638, |
|
"step": 911 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.7801688483442424, |
|
"learning_rate": 2.911081976456202e-06, |
|
"loss": 0.5872, |
|
"step": 912 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.8096931952713965, |
|
"learning_rate": 2.9072386834864723e-06, |
|
"loss": 0.5545, |
|
"step": 913 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.8346885149840537, |
|
"learning_rate": 2.903394401582017e-06, |
|
"loss": 0.593, |
|
"step": 914 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.8708319246348792, |
|
"learning_rate": 2.899549140078256e-06, |
|
"loss": 0.6032, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.8116588188201339, |
|
"learning_rate": 2.895702908312987e-06, |
|
"loss": 0.6215, |
|
"step": 916 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.8068958158119758, |
|
"learning_rate": 2.8918557156263662e-06, |
|
"loss": 0.5682, |
|
"step": 917 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.7858662353988665, |
|
"learning_rate": 2.888007571360879e-06, |
|
"loss": 0.5923, |
|
"step": 918 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.8771050496241385, |
|
"learning_rate": 2.8841584848613254e-06, |
|
"loss": 0.6514, |
|
"step": 919 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.8296510990241097, |
|
"learning_rate": 2.880308465474792e-06, |
|
"loss": 0.606, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.7889136835873297, |
|
"learning_rate": 2.876457522550631e-06, |
|
"loss": 0.5862, |
|
"step": 921 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 2.925493035659946, |
|
"learning_rate": 2.872605665440436e-06, |
|
"loss": 0.6542, |
|
"step": 922 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.8507683639836144, |
|
"learning_rate": 2.8687529034980244e-06, |
|
"loss": 0.5491, |
|
"step": 923 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.9174925682822145, |
|
"learning_rate": 2.8648992460794054e-06, |
|
"loss": 0.6571, |
|
"step": 924 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.7942861243562647, |
|
"learning_rate": 2.8610447025427685e-06, |
|
"loss": 0.6239, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.8075965123150352, |
|
"learning_rate": 2.8571892822484502e-06, |
|
"loss": 0.6194, |
|
"step": 926 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.7953063343629782, |
|
"learning_rate": 2.8533329945589192e-06, |
|
"loss": 0.5954, |
|
"step": 927 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.8417169635844408, |
|
"learning_rate": 2.849475848838749e-06, |
|
"loss": 0.5967, |
|
"step": 928 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.8434785540022463, |
|
"learning_rate": 2.8456178544545976e-06, |
|
"loss": 0.5901, |
|
"step": 929 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.8173771863580619, |
|
"learning_rate": 2.841759020775184e-06, |
|
"loss": 0.6201, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.8148818341326955, |
|
"learning_rate": 2.8378993571712638e-06, |
|
"loss": 0.5838, |
|
"step": 931 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.8076114270975754, |
|
"learning_rate": 2.8340388730156097e-06, |
|
"loss": 0.5921, |
|
"step": 932 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.829687636666006, |
|
"learning_rate": 2.8301775776829875e-06, |
|
"loss": 0.6039, |
|
"step": 933 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.9197178593956111, |
|
"learning_rate": 2.82631548055013e-06, |
|
"loss": 0.5839, |
|
"step": 934 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.8011771461166681, |
|
"learning_rate": 2.822452590995719e-06, |
|
"loss": 0.625, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.8071429478414894, |
|
"learning_rate": 2.8185889184003614e-06, |
|
"loss": 0.6258, |
|
"step": 936 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.8473353081081582, |
|
"learning_rate": 2.814724472146564e-06, |
|
"loss": 0.6264, |
|
"step": 937 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.8351071263143917, |
|
"learning_rate": 2.8108592616187135e-06, |
|
"loss": 0.5849, |
|
"step": 938 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.8433252071598873, |
|
"learning_rate": 2.806993296203052e-06, |
|
"loss": 0.5811, |
|
"step": 939 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 1.6097167209325385, |
|
"learning_rate": 2.8031265852876537e-06, |
|
"loss": 0.5958, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.868697036480233, |
|
"learning_rate": 2.7992591382624064e-06, |
|
"loss": 0.6281, |
|
"step": 941 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.851428078011761, |
|
"learning_rate": 2.7953909645189824e-06, |
|
"loss": 0.6174, |
|
"step": 942 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.8278638035201089, |
|
"learning_rate": 2.791522073450819e-06, |
|
"loss": 0.6232, |
|
"step": 943 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.795856052219787, |
|
"learning_rate": 2.7876524744530967e-06, |
|
"loss": 0.6053, |
|
"step": 944 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.8545427598167615, |
|
"learning_rate": 2.7837821769227154e-06, |
|
"loss": 0.606, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.7872291169661336, |
|
"learning_rate": 2.7799111902582697e-06, |
|
"loss": 0.5937, |
|
"step": 946 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.8222342924067905, |
|
"learning_rate": 2.7760395238600286e-06, |
|
"loss": 0.6429, |
|
"step": 947 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.8324916626961192, |
|
"learning_rate": 2.7721671871299115e-06, |
|
"loss": 0.594, |
|
"step": 948 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.7949219615171061, |
|
"learning_rate": 2.7682941894714664e-06, |
|
"loss": 0.6344, |
|
"step": 949 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.7915081056376099, |
|
"learning_rate": 2.764420540289845e-06, |
|
"loss": 0.647, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.8221089204021264, |
|
"learning_rate": 2.7605462489917817e-06, |
|
"loss": 0.613, |
|
"step": 951 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.7556912305467864, |
|
"learning_rate": 2.7566713249855715e-06, |
|
"loss": 0.5927, |
|
"step": 952 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.8516869877340751, |
|
"learning_rate": 2.7527957776810432e-06, |
|
"loss": 0.5737, |
|
"step": 953 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.8954878925348272, |
|
"learning_rate": 2.748919616489542e-06, |
|
"loss": 0.5881, |
|
"step": 954 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.7814783166367757, |
|
"learning_rate": 2.7450428508239024e-06, |
|
"loss": 0.5817, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.7761831452626766, |
|
"learning_rate": 2.7411654900984257e-06, |
|
"loss": 0.6423, |
|
"step": 956 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.8610326997473378, |
|
"learning_rate": 2.7372875437288603e-06, |
|
"loss": 0.5698, |
|
"step": 957 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.8037534451700079, |
|
"learning_rate": 2.733409021132377e-06, |
|
"loss": 0.5618, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.8267194190693397, |
|
"learning_rate": 2.729529931727544e-06, |
|
"loss": 0.6146, |
|
"step": 959 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.8595187209919314, |
|
"learning_rate": 2.7256502849343062e-06, |
|
"loss": 0.6139, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.9151590718279631, |
|
"learning_rate": 2.7217700901739637e-06, |
|
"loss": 0.5933, |
|
"step": 961 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.7743662392639246, |
|
"learning_rate": 2.717889356869146e-06, |
|
"loss": 0.6067, |
|
"step": 962 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.7692375464347264, |
|
"learning_rate": 2.7140080944437903e-06, |
|
"loss": 0.6007, |
|
"step": 963 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.8243938846614656, |
|
"learning_rate": 2.710126312323119e-06, |
|
"loss": 0.6439, |
|
"step": 964 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.869984720476869, |
|
"learning_rate": 2.706244019933618e-06, |
|
"loss": 0.6077, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.8579806687629068, |
|
"learning_rate": 2.7023612267030085e-06, |
|
"loss": 0.6735, |
|
"step": 966 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.7651338842188523, |
|
"learning_rate": 2.6984779420602324e-06, |
|
"loss": 0.5633, |
|
"step": 967 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.8845999584695565, |
|
"learning_rate": 2.694594175435422e-06, |
|
"loss": 0.6655, |
|
"step": 968 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.9552402309050972, |
|
"learning_rate": 2.6907099362598815e-06, |
|
"loss": 0.6015, |
|
"step": 969 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.7881059310601948, |
|
"learning_rate": 2.686825233966061e-06, |
|
"loss": 0.6393, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.8145877809003402, |
|
"learning_rate": 2.6829400779875377e-06, |
|
"loss": 0.5862, |
|
"step": 971 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.8749887484577397, |
|
"learning_rate": 2.679054477758988e-06, |
|
"loss": 0.6127, |
|
"step": 972 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.8610636451400849, |
|
"learning_rate": 2.6751684427161684e-06, |
|
"loss": 0.6101, |
|
"step": 973 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.8990043854731499, |
|
"learning_rate": 2.6712819822958917e-06, |
|
"loss": 0.6828, |
|
"step": 974 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.8412551914209038, |
|
"learning_rate": 2.6673951059360036e-06, |
|
"loss": 0.613, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.9159241308527234, |
|
"learning_rate": 2.663507823075358e-06, |
|
"loss": 0.5765, |
|
"step": 976 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.8840845410112719, |
|
"learning_rate": 2.6596201431537977e-06, |
|
"loss": 0.6241, |
|
"step": 977 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.853193970562842, |
|
"learning_rate": 2.6557320756121306e-06, |
|
"loss": 0.552, |
|
"step": 978 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.8275645940427406, |
|
"learning_rate": 2.651843629892103e-06, |
|
"loss": 0.568, |
|
"step": 979 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.7816643401147234, |
|
"learning_rate": 2.647954815436382e-06, |
|
"loss": 0.6002, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.8122193509157052, |
|
"learning_rate": 2.6440656416885302e-06, |
|
"loss": 0.6705, |
|
"step": 981 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.7916641408063855, |
|
"learning_rate": 2.6401761180929798e-06, |
|
"loss": 0.5727, |
|
"step": 982 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.823136144697698, |
|
"learning_rate": 2.6362862540950163e-06, |
|
"loss": 0.6495, |
|
"step": 983 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.9635376499725725, |
|
"learning_rate": 2.6323960591407487e-06, |
|
"loss": 0.6187, |
|
"step": 984 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.8237142858677388, |
|
"learning_rate": 2.6285055426770935e-06, |
|
"loss": 0.5672, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.841511477918458, |
|
"learning_rate": 2.624614714151743e-06, |
|
"loss": 0.5995, |
|
"step": 986 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.865394366604303, |
|
"learning_rate": 2.6207235830131517e-06, |
|
"loss": 0.5992, |
|
"step": 987 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.8810526841971326, |
|
"learning_rate": 2.616832158710506e-06, |
|
"loss": 0.65, |
|
"step": 988 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.83728728954254, |
|
"learning_rate": 2.612940450693706e-06, |
|
"loss": 0.5785, |
|
"step": 989 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.8094909685342696, |
|
"learning_rate": 2.6090484684133406e-06, |
|
"loss": 0.6586, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.8383394818111478, |
|
"learning_rate": 2.6051562213206633e-06, |
|
"loss": 0.6471, |
|
"step": 991 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.8008496703545273, |
|
"learning_rate": 2.6012637188675717e-06, |
|
"loss": 0.6041, |
|
"step": 992 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.8631369907215043, |
|
"learning_rate": 2.5973709705065836e-06, |
|
"loss": 0.6147, |
|
"step": 993 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.8266779122262543, |
|
"learning_rate": 2.593477985690815e-06, |
|
"loss": 0.6196, |
|
"step": 994 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.7893764214118881, |
|
"learning_rate": 2.589584773873953e-06, |
|
"loss": 0.6189, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.8482167729005808, |
|
"learning_rate": 2.5856913445102383e-06, |
|
"loss": 0.6131, |
|
"step": 996 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.8436526224259587, |
|
"learning_rate": 2.5817977070544408e-06, |
|
"loss": 0.6285, |
|
"step": 997 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.8898122115831852, |
|
"learning_rate": 2.577903870961833e-06, |
|
"loss": 0.5756, |
|
"step": 998 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.8022952501948407, |
|
"learning_rate": 2.5740098456881718e-06, |
|
"loss": 0.5794, |
|
"step": 999 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.7719948709570122, |
|
"learning_rate": 2.5701156406896726e-06, |
|
"loss": 0.5584, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.8361192518803613, |
|
"learning_rate": 2.5662212654229884e-06, |
|
"loss": 0.5866, |
|
"step": 1001 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.8128856577929028, |
|
"learning_rate": 2.5623267293451827e-06, |
|
"loss": 0.5976, |
|
"step": 1002 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.8061981083828511, |
|
"learning_rate": 2.5584320419137127e-06, |
|
"loss": 0.5893, |
|
"step": 1003 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.8541963254842264, |
|
"learning_rate": 2.554537212586403e-06, |
|
"loss": 0.5794, |
|
"step": 1004 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.8341460804008688, |
|
"learning_rate": 2.5506422508214207e-06, |
|
"loss": 0.5897, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.8003569253426912, |
|
"learning_rate": 2.5467471660772557e-06, |
|
"loss": 0.62, |
|
"step": 1006 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.8162229148335823, |
|
"learning_rate": 2.5428519678126974e-06, |
|
"loss": 0.5881, |
|
"step": 1007 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.8407184867945466, |
|
"learning_rate": 2.538956665486808e-06, |
|
"loss": 0.6183, |
|
"step": 1008 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.815433687807614, |
|
"learning_rate": 2.535061268558906e-06, |
|
"loss": 0.6176, |
|
"step": 1009 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.7869662406073055, |
|
"learning_rate": 2.531165786488538e-06, |
|
"loss": 0.5672, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.8189601120399821, |
|
"learning_rate": 2.527270228735456e-06, |
|
"loss": 0.5557, |
|
"step": 1011 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.8194269588514274, |
|
"learning_rate": 2.5233746047595984e-06, |
|
"loss": 0.5744, |
|
"step": 1012 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.933322233446511, |
|
"learning_rate": 2.5194789240210623e-06, |
|
"loss": 0.5976, |
|
"step": 1013 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.941214440145205, |
|
"learning_rate": 2.5155831959800843e-06, |
|
"loss": 0.5986, |
|
"step": 1014 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.5868439078330994, |
|
"eval_runtime": 1883.2861, |
|
"eval_samples_per_second": 0.701, |
|
"eval_steps_per_second": 0.078, |
|
"step": 1014 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.8184065254427949, |
|
"learning_rate": 2.511687430097014e-06, |
|
"loss": 0.5699, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.7766493329966468, |
|
"learning_rate": 2.5077916358322948e-06, |
|
"loss": 0.5714, |
|
"step": 1016 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 2.8105346789430676, |
|
"learning_rate": 2.503895822646438e-06, |
|
"loss": 0.6318, |
|
"step": 1017 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.8941499596953436, |
|
"learning_rate": 2.5e-06, |
|
"loss": 0.5792, |
|
"step": 1018 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.8358223804872863, |
|
"learning_rate": 2.4961041773535625e-06, |
|
"loss": 0.5513, |
|
"step": 1019 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.8368099930696999, |
|
"learning_rate": 2.492208364167706e-06, |
|
"loss": 0.6077, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.7869895833465728, |
|
"learning_rate": 2.488312569902987e-06, |
|
"loss": 0.5731, |
|
"step": 1021 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.8558739115192562, |
|
"learning_rate": 2.484416804019916e-06, |
|
"loss": 0.6019, |
|
"step": 1022 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.8428793143348562, |
|
"learning_rate": 2.480521075978938e-06, |
|
"loss": 0.6974, |
|
"step": 1023 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.8337532547752103, |
|
"learning_rate": 2.4766253952404024e-06, |
|
"loss": 0.6301, |
|
"step": 1024 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.8224455861498798, |
|
"learning_rate": 2.4727297712645446e-06, |
|
"loss": 0.5924, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.8478732523363851, |
|
"learning_rate": 2.4688342135114625e-06, |
|
"loss": 0.6113, |
|
"step": 1026 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.8517697941831913, |
|
"learning_rate": 2.464938731441095e-06, |
|
"loss": 0.6138, |
|
"step": 1027 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.8490356238096585, |
|
"learning_rate": 2.461043334513193e-06, |
|
"loss": 0.5697, |
|
"step": 1028 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.8095862996841214, |
|
"learning_rate": 2.457148032187304e-06, |
|
"loss": 0.5818, |
|
"step": 1029 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.8949465514679057, |
|
"learning_rate": 2.453252833922745e-06, |
|
"loss": 0.6591, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.8836025523334905, |
|
"learning_rate": 2.44935774917858e-06, |
|
"loss": 0.5924, |
|
"step": 1031 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.8054372100465728, |
|
"learning_rate": 2.4454627874135976e-06, |
|
"loss": 0.6437, |
|
"step": 1032 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.8416983532954821, |
|
"learning_rate": 2.441567958086288e-06, |
|
"loss": 0.6281, |
|
"step": 1033 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.8757745197214527, |
|
"learning_rate": 2.4376732706548185e-06, |
|
"loss": 0.625, |
|
"step": 1034 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 1.2326242381833146, |
|
"learning_rate": 2.4337787345770133e-06, |
|
"loss": 0.583, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 1.1535345183481485, |
|
"learning_rate": 2.429884359310328e-06, |
|
"loss": 0.5372, |
|
"step": 1036 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.9975727902534105, |
|
"learning_rate": 2.425990154311829e-06, |
|
"loss": 0.5382, |
|
"step": 1037 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 1.0719972982620805, |
|
"learning_rate": 2.4220961290381676e-06, |
|
"loss": 0.5734, |
|
"step": 1038 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 1.1215516398780767, |
|
"learning_rate": 2.41820229294556e-06, |
|
"loss": 0.5519, |
|
"step": 1039 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 1.0021989076040982, |
|
"learning_rate": 2.4143086554897626e-06, |
|
"loss": 0.5286, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 1.0179021321813106, |
|
"learning_rate": 2.4104152261260485e-06, |
|
"loss": 0.532, |
|
"step": 1041 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 1.0203705451501146, |
|
"learning_rate": 2.4065220143091863e-06, |
|
"loss": 0.5245, |
|
"step": 1042 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 1.0303824771293035, |
|
"learning_rate": 2.4026290294934177e-06, |
|
"loss": 0.4896, |
|
"step": 1043 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.9195191354878299, |
|
"learning_rate": 2.3987362811324296e-06, |
|
"loss": 0.5332, |
|
"step": 1044 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.9225671787625556, |
|
"learning_rate": 2.394843778679338e-06, |
|
"loss": 0.5867, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 1.00883437367367, |
|
"learning_rate": 2.3909515315866606e-06, |
|
"loss": 0.4902, |
|
"step": 1046 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.9819564894542347, |
|
"learning_rate": 2.3870595493062953e-06, |
|
"loss": 0.5548, |
|
"step": 1047 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.8996578093054595, |
|
"learning_rate": 2.3831678412894947e-06, |
|
"loss": 0.5143, |
|
"step": 1048 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.8841287268967455, |
|
"learning_rate": 2.379276416986849e-06, |
|
"loss": 0.5577, |
|
"step": 1049 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.9154027270624303, |
|
"learning_rate": 2.375385285848257e-06, |
|
"loss": 0.5219, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.8503941542526697, |
|
"learning_rate": 2.371494457322907e-06, |
|
"loss": 0.5076, |
|
"step": 1051 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.8631593873373877, |
|
"learning_rate": 2.367603940859251e-06, |
|
"loss": 0.5437, |
|
"step": 1052 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.8357328748238451, |
|
"learning_rate": 2.363713745904984e-06, |
|
"loss": 0.5229, |
|
"step": 1053 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.8775120825084444, |
|
"learning_rate": 2.3598238819070206e-06, |
|
"loss": 0.5367, |
|
"step": 1054 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.8609100919426709, |
|
"learning_rate": 2.3559343583114706e-06, |
|
"loss": 0.523, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 1.0193673849353349, |
|
"learning_rate": 2.352045184563618e-06, |
|
"loss": 0.5475, |
|
"step": 1056 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.8529786933163018, |
|
"learning_rate": 2.348156370107897e-06, |
|
"loss": 0.5565, |
|
"step": 1057 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.8417460689300821, |
|
"learning_rate": 2.3442679243878698e-06, |
|
"loss": 0.5207, |
|
"step": 1058 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.8108039618515722, |
|
"learning_rate": 2.3403798568462023e-06, |
|
"loss": 0.5359, |
|
"step": 1059 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 0.8716560315889638, |
|
"learning_rate": 2.3364921769246423e-06, |
|
"loss": 0.5143, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 0.8639220899847455, |
|
"learning_rate": 2.332604894063997e-06, |
|
"loss": 0.5175, |
|
"step": 1061 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 0.8512693059616497, |
|
"learning_rate": 2.3287180177041087e-06, |
|
"loss": 0.5519, |
|
"step": 1062 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 0.9325436792044128, |
|
"learning_rate": 2.3248315572838316e-06, |
|
"loss": 0.5362, |
|
"step": 1063 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 0.8560972703486215, |
|
"learning_rate": 2.3209455222410125e-06, |
|
"loss": 0.5012, |
|
"step": 1064 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 0.8789242584496589, |
|
"learning_rate": 2.3170599220124636e-06, |
|
"loss": 0.5097, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 0.897194591755581, |
|
"learning_rate": 2.3131747660339396e-06, |
|
"loss": 0.5278, |
|
"step": 1066 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 0.8574035467764353, |
|
"learning_rate": 2.3092900637401193e-06, |
|
"loss": 0.5137, |
|
"step": 1067 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 0.8291833001739086, |
|
"learning_rate": 2.3054058245645787e-06, |
|
"loss": 0.4965, |
|
"step": 1068 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 0.8821188097966408, |
|
"learning_rate": 2.301522057939768e-06, |
|
"loss": 0.5873, |
|
"step": 1069 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.8766883160011891, |
|
"learning_rate": 2.297638773296992e-06, |
|
"loss": 0.4979, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.8589072722935909, |
|
"learning_rate": 2.2937559800663828e-06, |
|
"loss": 0.5203, |
|
"step": 1071 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.8389777642684669, |
|
"learning_rate": 2.2898736876768816e-06, |
|
"loss": 0.476, |
|
"step": 1072 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.9599021584638144, |
|
"learning_rate": 2.2859919055562105e-06, |
|
"loss": 0.4946, |
|
"step": 1073 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.881803755504705, |
|
"learning_rate": 2.2821106431308546e-06, |
|
"loss": 0.4973, |
|
"step": 1074 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.8829215158890488, |
|
"learning_rate": 2.278229909826037e-06, |
|
"loss": 0.5237, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.9262781203072397, |
|
"learning_rate": 2.274349715065694e-06, |
|
"loss": 0.5156, |
|
"step": 1076 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.8654049538846458, |
|
"learning_rate": 2.2704700682724566e-06, |
|
"loss": 0.4867, |
|
"step": 1077 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.8403716826368274, |
|
"learning_rate": 2.2665909788676236e-06, |
|
"loss": 0.5088, |
|
"step": 1078 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.8617657693206535, |
|
"learning_rate": 2.26271245627114e-06, |
|
"loss": 0.4882, |
|
"step": 1079 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 0.8273414350307483, |
|
"learning_rate": 2.258834509901575e-06, |
|
"loss": 0.5354, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 0.8280137361700848, |
|
"learning_rate": 2.2549571491760985e-06, |
|
"loss": 0.5704, |
|
"step": 1081 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 0.8361140070473022, |
|
"learning_rate": 2.251080383510459e-06, |
|
"loss": 0.5245, |
|
"step": 1082 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 0.8655386471036016, |
|
"learning_rate": 2.247204222318957e-06, |
|
"loss": 0.587, |
|
"step": 1083 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 0.9843887188585405, |
|
"learning_rate": 2.2433286750144293e-06, |
|
"loss": 0.5266, |
|
"step": 1084 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 0.8831089197841557, |
|
"learning_rate": 2.239453751008219e-06, |
|
"loss": 0.5406, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 0.8311371050772496, |
|
"learning_rate": 2.235579459710156e-06, |
|
"loss": 0.5388, |
|
"step": 1086 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 0.90750142486625, |
|
"learning_rate": 2.231705810528534e-06, |
|
"loss": 0.5039, |
|
"step": 1087 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 0.8767416694124908, |
|
"learning_rate": 2.2278328128700893e-06, |
|
"loss": 0.5046, |
|
"step": 1088 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 0.8728415037121665, |
|
"learning_rate": 2.2239604761399718e-06, |
|
"loss": 0.4735, |
|
"step": 1089 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 0.8770872360734615, |
|
"learning_rate": 2.2200888097417308e-06, |
|
"loss": 0.5053, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 0.8320867717236903, |
|
"learning_rate": 2.216217823077285e-06, |
|
"loss": 0.5004, |
|
"step": 1091 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 0.8721327690081925, |
|
"learning_rate": 2.2123475255469037e-06, |
|
"loss": 0.4854, |
|
"step": 1092 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 0.8869046238947459, |
|
"learning_rate": 2.2084779265491817e-06, |
|
"loss": 0.5033, |
|
"step": 1093 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 0.835536916011144, |
|
"learning_rate": 2.204609035481018e-06, |
|
"loss": 0.5067, |
|
"step": 1094 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 0.8380276943680528, |
|
"learning_rate": 2.2007408617375944e-06, |
|
"loss": 0.5093, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 0.8932865363338237, |
|
"learning_rate": 2.1968734147123467e-06, |
|
"loss": 0.5084, |
|
"step": 1096 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 0.8600518274007936, |
|
"learning_rate": 2.193006703796949e-06, |
|
"loss": 0.5917, |
|
"step": 1097 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 0.836626172768628, |
|
"learning_rate": 2.1891407383812878e-06, |
|
"loss": 0.5629, |
|
"step": 1098 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 0.8260250095624573, |
|
"learning_rate": 2.1852755278534373e-06, |
|
"loss": 0.4975, |
|
"step": 1099 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 0.7884121566415102, |
|
"learning_rate": 2.18141108159964e-06, |
|
"loss": 0.518, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 0.8223086356273321, |
|
"learning_rate": 2.1775474090042824e-06, |
|
"loss": 0.5487, |
|
"step": 1101 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 0.8115486306598928, |
|
"learning_rate": 2.173684519449872e-06, |
|
"loss": 0.501, |
|
"step": 1102 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 0.9335530764840646, |
|
"learning_rate": 2.1698224223170137e-06, |
|
"loss": 0.5315, |
|
"step": 1103 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 0.8643469465750571, |
|
"learning_rate": 2.1659611269843907e-06, |
|
"loss": 0.5308, |
|
"step": 1104 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 0.8449777143283284, |
|
"learning_rate": 2.162100642828737e-06, |
|
"loss": 0.5167, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 0.9015414602562898, |
|
"learning_rate": 2.158240979224817e-06, |
|
"loss": 0.5348, |
|
"step": 1106 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 0.8325170189781876, |
|
"learning_rate": 2.1543821455454032e-06, |
|
"loss": 0.53, |
|
"step": 1107 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 0.8539221844749437, |
|
"learning_rate": 2.1505241511612522e-06, |
|
"loss": 0.5015, |
|
"step": 1108 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 0.873096579016518, |
|
"learning_rate": 2.146667005441082e-06, |
|
"loss": 0.5332, |
|
"step": 1109 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 0.841238373167198, |
|
"learning_rate": 2.1428107177515506e-06, |
|
"loss": 0.4869, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 0.8375524609776578, |
|
"learning_rate": 2.138955297457233e-06, |
|
"loss": 0.5044, |
|
"step": 1111 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 0.9097304276305382, |
|
"learning_rate": 2.1351007539205954e-06, |
|
"loss": 0.5649, |
|
"step": 1112 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 0.8306588930318208, |
|
"learning_rate": 2.1312470965019765e-06, |
|
"loss": 0.5202, |
|
"step": 1113 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 0.8594460674957857, |
|
"learning_rate": 2.1273943345595637e-06, |
|
"loss": 0.5991, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 0.8475877410188423, |
|
"learning_rate": 2.1235424774493695e-06, |
|
"loss": 0.5033, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 0.9233261317257966, |
|
"learning_rate": 2.1196915345252085e-06, |
|
"loss": 0.5227, |
|
"step": 1116 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 0.914929787489972, |
|
"learning_rate": 2.1158415151386746e-06, |
|
"loss": 0.5294, |
|
"step": 1117 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 0.882493446028043, |
|
"learning_rate": 2.111992428639121e-06, |
|
"loss": 0.5561, |
|
"step": 1118 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 0.915426790816166, |
|
"learning_rate": 2.108144284373634e-06, |
|
"loss": 0.5047, |
|
"step": 1119 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 0.8250367018979915, |
|
"learning_rate": 2.104297091687013e-06, |
|
"loss": 0.5211, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 0.9742813220296467, |
|
"learning_rate": 2.1004508599217444e-06, |
|
"loss": 0.5057, |
|
"step": 1121 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 0.8125299066591511, |
|
"learning_rate": 2.0966055984179834e-06, |
|
"loss": 0.5092, |
|
"step": 1122 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 0.8468524958201039, |
|
"learning_rate": 2.0927613165135285e-06, |
|
"loss": 0.533, |
|
"step": 1123 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 0.8387395179689563, |
|
"learning_rate": 2.088918023543798e-06, |
|
"loss": 0.5478, |
|
"step": 1124 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 0.8252815008223313, |
|
"learning_rate": 2.0850757288418103e-06, |
|
"loss": 0.4967, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 0.8652347115335814, |
|
"learning_rate": 2.0812344417381595e-06, |
|
"loss": 0.55, |
|
"step": 1126 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 0.8597046573753564, |
|
"learning_rate": 2.077394171560991e-06, |
|
"loss": 0.5135, |
|
"step": 1127 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 0.8329136911134295, |
|
"learning_rate": 2.073554927635983e-06, |
|
"loss": 0.5471, |
|
"step": 1128 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 0.868905523386283, |
|
"learning_rate": 2.0697167192863205e-06, |
|
"loss": 0.5365, |
|
"step": 1129 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 0.8327359527602765, |
|
"learning_rate": 2.0658795558326745e-06, |
|
"loss": 0.539, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.8620052059128163, |
|
"learning_rate": 2.0620434465931765e-06, |
|
"loss": 0.5234, |
|
"step": 1131 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.8773898625310315, |
|
"learning_rate": 2.0582084008834003e-06, |
|
"loss": 0.5039, |
|
"step": 1132 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.8027974104327941, |
|
"learning_rate": 2.0543744280163365e-06, |
|
"loss": 0.5066, |
|
"step": 1133 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.8434888762216062, |
|
"learning_rate": 2.0505415373023686e-06, |
|
"loss": 0.5226, |
|
"step": 1134 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.8448822472541239, |
|
"learning_rate": 2.0467097380492547e-06, |
|
"loss": 0.5518, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.8367870393130636, |
|
"learning_rate": 2.0428790395621015e-06, |
|
"loss": 0.4899, |
|
"step": 1136 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.82710622957581, |
|
"learning_rate": 2.039049451143342e-06, |
|
"loss": 0.5517, |
|
"step": 1137 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.8866030136538112, |
|
"learning_rate": 2.035220982092714e-06, |
|
"loss": 0.4669, |
|
"step": 1138 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.8769901729762373, |
|
"learning_rate": 2.0313936417072368e-06, |
|
"loss": 0.556, |
|
"step": 1139 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.8857705371994514, |
|
"learning_rate": 2.02756743928119e-06, |
|
"loss": 0.5566, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 0.8730050406027847, |
|
"learning_rate": 2.0237423841060877e-06, |
|
"loss": 0.5235, |
|
"step": 1141 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 0.8405396726023919, |
|
"learning_rate": 2.0199184854706597e-06, |
|
"loss": 0.5198, |
|
"step": 1142 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 0.858763068434379, |
|
"learning_rate": 2.016095752660828e-06, |
|
"loss": 0.5224, |
|
"step": 1143 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 0.920500499730532, |
|
"learning_rate": 2.01227419495968e-06, |
|
"loss": 0.5122, |
|
"step": 1144 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 0.8770479418764661, |
|
"learning_rate": 2.0084538216474525e-06, |
|
"loss": 0.5316, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 0.8601190347275052, |
|
"learning_rate": 2.004634642001507e-06, |
|
"loss": 0.5577, |
|
"step": 1146 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 0.876531830412254, |
|
"learning_rate": 2.000816665296302e-06, |
|
"loss": 0.5613, |
|
"step": 1147 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 0.870565552665694, |
|
"learning_rate": 1.9969999008033796e-06, |
|
"loss": 0.4973, |
|
"step": 1148 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 0.8403707362329739, |
|
"learning_rate": 1.9931843577913356e-06, |
|
"loss": 0.545, |
|
"step": 1149 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 0.8208475370045443, |
|
"learning_rate": 1.9893700455257996e-06, |
|
"loss": 0.5037, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 0.859174085734411, |
|
"learning_rate": 1.985556973269413e-06, |
|
"loss": 0.4886, |
|
"step": 1151 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 0.8904536919872101, |
|
"learning_rate": 1.9817451502818063e-06, |
|
"loss": 0.5288, |
|
"step": 1152 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 0.8382646221041071, |
|
"learning_rate": 1.9779345858195757e-06, |
|
"loss": 0.5334, |
|
"step": 1153 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 0.8477438049549787, |
|
"learning_rate": 1.974125289136261e-06, |
|
"loss": 0.491, |
|
"step": 1154 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 0.8678972728643963, |
|
"learning_rate": 1.970317269482324e-06, |
|
"loss": 0.5094, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 0.8250099417025093, |
|
"learning_rate": 1.9665105361051255e-06, |
|
"loss": 0.5077, |
|
"step": 1156 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 0.819944697867349, |
|
"learning_rate": 1.9627050982489014e-06, |
|
"loss": 0.5141, |
|
"step": 1157 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 0.8218366216426203, |
|
"learning_rate": 1.958900965154743e-06, |
|
"loss": 0.5182, |
|
"step": 1158 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 0.8122072565245823, |
|
"learning_rate": 1.9550981460605734e-06, |
|
"loss": 0.5154, |
|
"step": 1159 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 0.8123124739985441, |
|
"learning_rate": 1.9512966502011222e-06, |
|
"loss": 0.5327, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.8387045926191694, |
|
"learning_rate": 1.9474964868079082e-06, |
|
"loss": 0.5527, |
|
"step": 1161 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.8535839422791967, |
|
"learning_rate": 1.9436976651092143e-06, |
|
"loss": 0.4916, |
|
"step": 1162 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.9006072458090121, |
|
"learning_rate": 1.9399001943300645e-06, |
|
"loss": 0.5529, |
|
"step": 1163 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.8354717139216293, |
|
"learning_rate": 1.936104083692202e-06, |
|
"loss": 0.534, |
|
"step": 1164 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.8491139421838702, |
|
"learning_rate": 1.9323093424140673e-06, |
|
"loss": 0.5744, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.8395416950156503, |
|
"learning_rate": 1.9285159797107766e-06, |
|
"loss": 0.5425, |
|
"step": 1166 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.8247770821972967, |
|
"learning_rate": 1.9247240047940966e-06, |
|
"loss": 0.5533, |
|
"step": 1167 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.8228136864560915, |
|
"learning_rate": 1.920933426872425e-06, |
|
"loss": 0.5182, |
|
"step": 1168 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 1.1160073812032258, |
|
"learning_rate": 1.917144255150767e-06, |
|
"loss": 0.4845, |
|
"step": 1169 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.8487318241192247, |
|
"learning_rate": 1.913356498830713e-06, |
|
"loss": 0.5039, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.813560302312254, |
|
"learning_rate": 1.909570167110415e-06, |
|
"loss": 0.4765, |
|
"step": 1171 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 0.8851975504941262, |
|
"learning_rate": 1.9057852691845677e-06, |
|
"loss": 0.5466, |
|
"step": 1172 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 0.8372722129182711, |
|
"learning_rate": 1.9020018142443834e-06, |
|
"loss": 0.5694, |
|
"step": 1173 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 0.8192604770151811, |
|
"learning_rate": 1.8982198114775683e-06, |
|
"loss": 0.4879, |
|
"step": 1174 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 0.914589074878136, |
|
"learning_rate": 1.8944392700683043e-06, |
|
"loss": 0.4966, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 0.8767357123088995, |
|
"learning_rate": 1.8906601991972233e-06, |
|
"loss": 0.5017, |
|
"step": 1176 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 0.8882490824865216, |
|
"learning_rate": 1.8868826080413876e-06, |
|
"loss": 0.5551, |
|
"step": 1177 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 0.8476779519428936, |
|
"learning_rate": 1.8831065057742658e-06, |
|
"loss": 0.4783, |
|
"step": 1178 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 0.8512763436343278, |
|
"learning_rate": 1.8793319015657091e-06, |
|
"loss": 0.5213, |
|
"step": 1179 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 0.8607307116274757, |
|
"learning_rate": 1.8755588045819325e-06, |
|
"loss": 0.4992, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 0.8317228987977554, |
|
"learning_rate": 1.8717872239854912e-06, |
|
"loss": 0.5402, |
|
"step": 1181 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 0.8130324795780219, |
|
"learning_rate": 1.868017168935256e-06, |
|
"loss": 0.4824, |
|
"step": 1182 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 0.8365085778490724, |
|
"learning_rate": 1.8642486485863954e-06, |
|
"loss": 0.4984, |
|
"step": 1183 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 0.8483038315719312, |
|
"learning_rate": 1.86048167209035e-06, |
|
"loss": 0.5359, |
|
"step": 1184 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 0.8834538360237251, |
|
"learning_rate": 1.856716248594811e-06, |
|
"loss": 0.5172, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 0.8485142661493278, |
|
"learning_rate": 1.852952387243698e-06, |
|
"loss": 0.5387, |
|
"step": 1186 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 0.8548372665093342, |
|
"learning_rate": 1.8491900971771387e-06, |
|
"loss": 0.5053, |
|
"step": 1187 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 0.8266022680630544, |
|
"learning_rate": 1.845429387531445e-06, |
|
"loss": 0.4635, |
|
"step": 1188 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 1.4463463544398012, |
|
"learning_rate": 1.841670267439088e-06, |
|
"loss": 0.5502, |
|
"step": 1189 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 1.066516598019322, |
|
"learning_rate": 1.8379127460286819e-06, |
|
"loss": 0.5575, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 0.8740864704012543, |
|
"learning_rate": 1.8341568324249584e-06, |
|
"loss": 0.5216, |
|
"step": 1191 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.871996718634586, |
|
"learning_rate": 1.8304025357487427e-06, |
|
"loss": 0.4912, |
|
"step": 1192 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.850465659469331, |
|
"learning_rate": 1.8266498651169352e-06, |
|
"loss": 0.5158, |
|
"step": 1193 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.8760990178982471, |
|
"learning_rate": 1.8228988296424877e-06, |
|
"loss": 0.5018, |
|
"step": 1194 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.8918830024236992, |
|
"learning_rate": 1.81914943843438e-06, |
|
"loss": 0.519, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.8935633421686892, |
|
"learning_rate": 1.8154017005976e-06, |
|
"loss": 0.5218, |
|
"step": 1196 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.8765577491519292, |
|
"learning_rate": 1.8116556252331202e-06, |
|
"loss": 0.5261, |
|
"step": 1197 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 1.0527012351304872, |
|
"learning_rate": 1.8079112214378769e-06, |
|
"loss": 0.5157, |
|
"step": 1198 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.9044830696403585, |
|
"learning_rate": 1.804168498304745e-06, |
|
"loss": 0.5406, |
|
"step": 1199 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.9098746231032456, |
|
"learning_rate": 1.8004274649225201e-06, |
|
"loss": 0.4842, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.8101267276957087, |
|
"learning_rate": 1.7966881303758938e-06, |
|
"loss": 0.5022, |
|
"step": 1201 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 0.8601108497410651, |
|
"learning_rate": 1.7929505037454315e-06, |
|
"loss": 0.539, |
|
"step": 1202 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 0.9301076032455242, |
|
"learning_rate": 1.7892145941075526e-06, |
|
"loss": 0.499, |
|
"step": 1203 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 0.8382841406985904, |
|
"learning_rate": 1.7854804105345064e-06, |
|
"loss": 0.4949, |
|
"step": 1204 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 0.7832696813211808, |
|
"learning_rate": 1.781747962094349e-06, |
|
"loss": 0.5268, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 1.5062442738319717, |
|
"learning_rate": 1.778017257850926e-06, |
|
"loss": 0.5479, |
|
"step": 1206 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 0.8783161930303991, |
|
"learning_rate": 1.7742883068638447e-06, |
|
"loss": 0.5626, |
|
"step": 1207 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 0.8421190790352715, |
|
"learning_rate": 1.7705611181884558e-06, |
|
"loss": 0.4973, |
|
"step": 1208 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 0.8586808607306615, |
|
"learning_rate": 1.7668357008758309e-06, |
|
"loss": 0.5486, |
|
"step": 1209 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 0.8786355969325144, |
|
"learning_rate": 1.7631120639727396e-06, |
|
"loss": 0.549, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 0.8405432278222555, |
|
"learning_rate": 1.7593902165216284e-06, |
|
"loss": 0.5359, |
|
"step": 1211 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 0.8264936691418094, |
|
"learning_rate": 1.7556701675605968e-06, |
|
"loss": 0.5371, |
|
"step": 1212 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 0.9617454619106705, |
|
"learning_rate": 1.7519519261233786e-06, |
|
"loss": 0.5113, |
|
"step": 1213 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 0.8526826972774398, |
|
"learning_rate": 1.7482355012393177e-06, |
|
"loss": 0.5378, |
|
"step": 1214 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 0.8728494810565016, |
|
"learning_rate": 1.7445209019333461e-06, |
|
"loss": 0.5407, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 0.857666749325946, |
|
"learning_rate": 1.7408081372259633e-06, |
|
"loss": 0.5249, |
|
"step": 1216 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 0.8141921993616225, |
|
"learning_rate": 1.7370972161332145e-06, |
|
"loss": 0.5557, |
|
"step": 1217 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 0.8202544755723353, |
|
"learning_rate": 1.7333881476666648e-06, |
|
"loss": 0.5309, |
|
"step": 1218 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 0.8405460573500194, |
|
"learning_rate": 1.729680940833383e-06, |
|
"loss": 0.5613, |
|
"step": 1219 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 0.8405391706338762, |
|
"learning_rate": 1.7259756046359177e-06, |
|
"loss": 0.5584, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 0.8070114183042183, |
|
"learning_rate": 1.722272148072273e-06, |
|
"loss": 0.4852, |
|
"step": 1221 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 1.0385662932606796, |
|
"learning_rate": 1.7185705801358892e-06, |
|
"loss": 0.5621, |
|
"step": 1222 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 0.828652122809735, |
|
"learning_rate": 1.7148709098156202e-06, |
|
"loss": 0.5196, |
|
"step": 1223 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 0.8849290847244835, |
|
"learning_rate": 1.711173146095712e-06, |
|
"loss": 0.5295, |
|
"step": 1224 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 0.8370522639101268, |
|
"learning_rate": 1.7074772979557802e-06, |
|
"loss": 0.5236, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 0.848931593624677, |
|
"learning_rate": 1.7037833743707892e-06, |
|
"loss": 0.5112, |
|
"step": 1226 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 0.8235990692349079, |
|
"learning_rate": 1.70009138431103e-06, |
|
"loss": 0.5512, |
|
"step": 1227 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 0.8625547584596783, |
|
"learning_rate": 1.6964013367420967e-06, |
|
"loss": 0.5659, |
|
"step": 1228 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 0.9109010489606646, |
|
"learning_rate": 1.6927132406248677e-06, |
|
"loss": 0.5395, |
|
"step": 1229 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 0.8938813845309608, |
|
"learning_rate": 1.6890271049154826e-06, |
|
"loss": 0.5082, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 0.9755938984650637, |
|
"learning_rate": 1.6853429385653196e-06, |
|
"loss": 0.442, |
|
"step": 1231 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 0.8864534981870436, |
|
"learning_rate": 1.681660750520975e-06, |
|
"loss": 0.5798, |
|
"step": 1232 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 0.825845565798459, |
|
"learning_rate": 1.6779805497242408e-06, |
|
"loss": 0.5486, |
|
"step": 1233 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 0.8145101154028018, |
|
"learning_rate": 1.6743023451120831e-06, |
|
"loss": 0.5443, |
|
"step": 1234 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 0.8320797796170017, |
|
"learning_rate": 1.6706261456166205e-06, |
|
"loss": 0.5519, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 0.8679853610693097, |
|
"learning_rate": 1.6669519601651025e-06, |
|
"loss": 0.536, |
|
"step": 1236 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 0.8110400454384976, |
|
"learning_rate": 1.6632797976798887e-06, |
|
"loss": 0.5495, |
|
"step": 1237 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 0.8752279680550583, |
|
"learning_rate": 1.6596096670784238e-06, |
|
"loss": 0.5846, |
|
"step": 1238 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 0.8412302712094935, |
|
"learning_rate": 1.6559415772732196e-06, |
|
"loss": 0.5659, |
|
"step": 1239 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 0.8718052511648309, |
|
"learning_rate": 1.6522755371718333e-06, |
|
"loss": 0.5236, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 0.9039324471509512, |
|
"learning_rate": 1.648611555676843e-06, |
|
"loss": 0.4747, |
|
"step": 1241 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 0.8454310553906655, |
|
"learning_rate": 1.6449496416858285e-06, |
|
"loss": 0.5182, |
|
"step": 1242 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 0.8701924234537203, |
|
"learning_rate": 1.6412898040913471e-06, |
|
"loss": 0.5497, |
|
"step": 1243 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 0.9223507906717088, |
|
"learning_rate": 1.6376320517809168e-06, |
|
"loss": 0.5635, |
|
"step": 1244 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 0.8437046940111125, |
|
"learning_rate": 1.633976393636989e-06, |
|
"loss": 0.5864, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 0.879160486631292, |
|
"learning_rate": 1.6303228385369318e-06, |
|
"loss": 0.5609, |
|
"step": 1246 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 0.8357804205086922, |
|
"learning_rate": 1.6266713953530045e-06, |
|
"loss": 0.5142, |
|
"step": 1247 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 0.8437698095618436, |
|
"learning_rate": 1.6230220729523388e-06, |
|
"loss": 0.5106, |
|
"step": 1248 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 0.8381766600073225, |
|
"learning_rate": 1.6193748801969164e-06, |
|
"loss": 0.5059, |
|
"step": 1249 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 0.8089802551550114, |
|
"learning_rate": 1.6157298259435466e-06, |
|
"loss": 0.5444, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 0.8237115403177346, |
|
"learning_rate": 1.6120869190438459e-06, |
|
"loss": 0.5366, |
|
"step": 1251 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 0.823513876570041, |
|
"learning_rate": 1.6084461683442176e-06, |
|
"loss": 0.4898, |
|
"step": 1252 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 0.8648410791159006, |
|
"learning_rate": 1.6048075826858263e-06, |
|
"loss": 0.4691, |
|
"step": 1253 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 0.8527907981659381, |
|
"learning_rate": 1.6011711709045813e-06, |
|
"loss": 0.5748, |
|
"step": 1254 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 0.8715966168588507, |
|
"learning_rate": 1.5975369418311113e-06, |
|
"loss": 0.5731, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 0.8652434216006221, |
|
"learning_rate": 1.5939049042907463e-06, |
|
"loss": 0.4938, |
|
"step": 1256 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 0.867539622172284, |
|
"learning_rate": 1.5902750671034917e-06, |
|
"loss": 0.568, |
|
"step": 1257 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 0.868058810993107, |
|
"learning_rate": 1.5866474390840126e-06, |
|
"loss": 0.4484, |
|
"step": 1258 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 0.9002735620775817, |
|
"learning_rate": 1.5830220290416076e-06, |
|
"loss": 0.4973, |
|
"step": 1259 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 0.8218130712360244, |
|
"learning_rate": 1.579398845780189e-06, |
|
"loss": 0.5075, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 0.8586002794730077, |
|
"learning_rate": 1.5757778980982627e-06, |
|
"loss": 0.5493, |
|
"step": 1261 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 0.8409823695015988, |
|
"learning_rate": 1.5721591947889053e-06, |
|
"loss": 0.5497, |
|
"step": 1262 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 0.8714768875682657, |
|
"learning_rate": 1.5685427446397427e-06, |
|
"loss": 0.5181, |
|
"step": 1263 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 0.8488522102521581, |
|
"learning_rate": 1.5649285564329296e-06, |
|
"loss": 0.5135, |
|
"step": 1264 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 0.829116202704699, |
|
"learning_rate": 1.5613166389451285e-06, |
|
"loss": 0.5438, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 0.8449102078102267, |
|
"learning_rate": 1.5577070009474872e-06, |
|
"loss": 0.5013, |
|
"step": 1266 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 0.8289934445211119, |
|
"learning_rate": 1.5540996512056164e-06, |
|
"loss": 0.4822, |
|
"step": 1267 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 0.8322114648194353, |
|
"learning_rate": 1.5504945984795728e-06, |
|
"loss": 0.537, |
|
"step": 1268 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 0.9184160820530575, |
|
"learning_rate": 1.5468918515238335e-06, |
|
"loss": 0.5226, |
|
"step": 1269 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 0.8510238103173541, |
|
"learning_rate": 1.5432914190872757e-06, |
|
"loss": 0.5205, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 0.8008762117447774, |
|
"learning_rate": 1.539693309913158e-06, |
|
"loss": 0.5399, |
|
"step": 1271 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 0.826055280829751, |
|
"learning_rate": 1.536097532739096e-06, |
|
"loss": 0.489, |
|
"step": 1272 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.8562622049996536, |
|
"learning_rate": 1.5325040962970417e-06, |
|
"loss": 0.5016, |
|
"step": 1273 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.844101830193045, |
|
"learning_rate": 1.5289130093132634e-06, |
|
"loss": 0.5558, |
|
"step": 1274 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.8227563017575598, |
|
"learning_rate": 1.5253242805083255e-06, |
|
"loss": 0.5476, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.8425935565033811, |
|
"learning_rate": 1.521737918597063e-06, |
|
"loss": 0.5131, |
|
"step": 1276 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.826212794209584, |
|
"learning_rate": 1.5181539322885652e-06, |
|
"loss": 0.4938, |
|
"step": 1277 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.8629173681301252, |
|
"learning_rate": 1.514572330286152e-06, |
|
"loss": 0.5486, |
|
"step": 1278 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.8277794706177418, |
|
"learning_rate": 1.5109931212873535e-06, |
|
"loss": 0.5377, |
|
"step": 1279 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.8308448570400866, |
|
"learning_rate": 1.5074163139838876e-06, |
|
"loss": 0.5088, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.81071051812824, |
|
"learning_rate": 1.5038419170616414e-06, |
|
"loss": 0.5223, |
|
"step": 1281 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.8252737345921397, |
|
"learning_rate": 1.500269939200648e-06, |
|
"loss": 0.5008, |
|
"step": 1282 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.8487607485310287, |
|
"learning_rate": 1.4967003890750656e-06, |
|
"loss": 0.5308, |
|
"step": 1283 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.8792962398751665, |
|
"learning_rate": 1.4931332753531575e-06, |
|
"loss": 0.487, |
|
"step": 1284 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.8212234225764866, |
|
"learning_rate": 1.4895686066972703e-06, |
|
"loss": 0.5355, |
|
"step": 1285 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.8994603967645555, |
|
"learning_rate": 1.4860063917638128e-06, |
|
"loss": 0.5481, |
|
"step": 1286 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.8614194025677114, |
|
"learning_rate": 1.4824466392032357e-06, |
|
"loss": 0.5097, |
|
"step": 1287 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.8182116413003467, |
|
"learning_rate": 1.47888935766001e-06, |
|
"loss": 0.4924, |
|
"step": 1288 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.8770055245291976, |
|
"learning_rate": 1.4753345557726062e-06, |
|
"loss": 0.5111, |
|
"step": 1289 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.8394898011564801, |
|
"learning_rate": 1.4717822421734717e-06, |
|
"loss": 0.5277, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.8058559316639713, |
|
"learning_rate": 1.4682324254890135e-06, |
|
"loss": 0.523, |
|
"step": 1291 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.7860388327736769, |
|
"learning_rate": 1.4646851143395746e-06, |
|
"loss": 0.4844, |
|
"step": 1292 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 0.8242523348097235, |
|
"learning_rate": 1.4611403173394126e-06, |
|
"loss": 0.5732, |
|
"step": 1293 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 0.8353721700700213, |
|
"learning_rate": 1.4575980430966808e-06, |
|
"loss": 0.4996, |
|
"step": 1294 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 0.8206241441867873, |
|
"learning_rate": 1.4540583002134046e-06, |
|
"loss": 0.521, |
|
"step": 1295 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 0.875531119521352, |
|
"learning_rate": 1.4505210972854655e-06, |
|
"loss": 0.5482, |
|
"step": 1296 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 0.868631540546499, |
|
"learning_rate": 1.446986442902574e-06, |
|
"loss": 0.5639, |
|
"step": 1297 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 0.8629511645382901, |
|
"learning_rate": 1.443454345648252e-06, |
|
"loss": 0.5073, |
|
"step": 1298 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 0.8668932056309039, |
|
"learning_rate": 1.4399248140998137e-06, |
|
"loss": 0.5021, |
|
"step": 1299 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 0.8406523008882236, |
|
"learning_rate": 1.4363978568283412e-06, |
|
"loss": 0.5292, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 0.8323082828363668, |
|
"learning_rate": 1.4328734823986646e-06, |
|
"loss": 0.5356, |
|
"step": 1301 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 0.8645802843341719, |
|
"learning_rate": 1.4293516993693428e-06, |
|
"loss": 0.4955, |
|
"step": 1302 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 0.8645413919216907, |
|
"learning_rate": 1.4258325162926441e-06, |
|
"loss": 0.5722, |
|
"step": 1303 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 0.8603360249105225, |
|
"learning_rate": 1.422315941714518e-06, |
|
"loss": 0.5156, |
|
"step": 1304 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 0.8732380772710749, |
|
"learning_rate": 1.4188019841745842e-06, |
|
"loss": 0.5251, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 0.8714793855351259, |
|
"learning_rate": 1.415290652206105e-06, |
|
"loss": 0.5038, |
|
"step": 1306 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 0.8986369199774957, |
|
"learning_rate": 1.4117819543359654e-06, |
|
"loss": 0.5489, |
|
"step": 1307 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 0.8433980899589901, |
|
"learning_rate": 1.4082758990846578e-06, |
|
"loss": 0.5858, |
|
"step": 1308 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 0.8586816956803768, |
|
"learning_rate": 1.404772494966254e-06, |
|
"loss": 0.5097, |
|
"step": 1309 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 0.9354727861293575, |
|
"learning_rate": 1.4012717504883874e-06, |
|
"loss": 0.4896, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 0.8427571754719572, |
|
"learning_rate": 1.3977736741522366e-06, |
|
"loss": 0.492, |
|
"step": 1311 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 0.8402672905014476, |
|
"learning_rate": 1.3942782744524974e-06, |
|
"loss": 0.5384, |
|
"step": 1312 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.8641085714229687, |
|
"learning_rate": 1.3907855598773653e-06, |
|
"loss": 0.5261, |
|
"step": 1313 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.816971428882062, |
|
"learning_rate": 1.387295538908519e-06, |
|
"loss": 0.5279, |
|
"step": 1314 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.898509528920645, |
|
"learning_rate": 1.3838082200210932e-06, |
|
"loss": 0.5414, |
|
"step": 1315 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.8669152452662072, |
|
"learning_rate": 1.3803236116836593e-06, |
|
"loss": 0.5266, |
|
"step": 1316 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.8403833902639722, |
|
"learning_rate": 1.376841722358212e-06, |
|
"loss": 0.5235, |
|
"step": 1317 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.8349523991129191, |
|
"learning_rate": 1.3733625605001365e-06, |
|
"loss": 0.5136, |
|
"step": 1318 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.8440010555764079, |
|
"learning_rate": 1.369886134558201e-06, |
|
"loss": 0.4955, |
|
"step": 1319 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.7988252461471034, |
|
"learning_rate": 1.3664124529745246e-06, |
|
"loss": 0.5381, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.877094769693697, |
|
"learning_rate": 1.362941524184564e-06, |
|
"loss": 0.5286, |
|
"step": 1321 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.8414959026786313, |
|
"learning_rate": 1.3594733566170925e-06, |
|
"loss": 0.4852, |
|
"step": 1322 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.9058731971619174, |
|
"learning_rate": 1.3560079586941766e-06, |
|
"loss": 0.5355, |
|
"step": 1323 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 0.8686256016386007, |
|
"learning_rate": 1.3525453388311554e-06, |
|
"loss": 0.5522, |
|
"step": 1324 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 0.8196416502539858, |
|
"learning_rate": 1.3490855054366264e-06, |
|
"loss": 0.5253, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 0.8320155071145593, |
|
"learning_rate": 1.3456284669124159e-06, |
|
"loss": 0.5154, |
|
"step": 1326 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 0.8887063336416359, |
|
"learning_rate": 1.342174231653564e-06, |
|
"loss": 0.5719, |
|
"step": 1327 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 0.8331842496882823, |
|
"learning_rate": 1.3387228080483062e-06, |
|
"loss": 0.5759, |
|
"step": 1328 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 0.8993547546754129, |
|
"learning_rate": 1.3352742044780476e-06, |
|
"loss": 0.5117, |
|
"step": 1329 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 0.8021294174670146, |
|
"learning_rate": 1.331828429317345e-06, |
|
"loss": 0.4924, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 0.8900622286069916, |
|
"learning_rate": 1.328385490933889e-06, |
|
"loss": 0.5195, |
|
"step": 1331 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 0.8870834553083442, |
|
"learning_rate": 1.324945397688479e-06, |
|
"loss": 0.5323, |
|
"step": 1332 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 0.8478044883847651, |
|
"learning_rate": 1.3215081579350058e-06, |
|
"loss": 0.4996, |
|
"step": 1333 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.8778574019603872, |
|
"learning_rate": 1.3180737800204329e-06, |
|
"loss": 0.5149, |
|
"step": 1334 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.8290157714031899, |
|
"learning_rate": 1.3146422722847713e-06, |
|
"loss": 0.513, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.894326770956352, |
|
"learning_rate": 1.3112136430610623e-06, |
|
"loss": 0.5152, |
|
"step": 1336 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.8685419016543828, |
|
"learning_rate": 1.3077879006753598e-06, |
|
"loss": 0.5237, |
|
"step": 1337 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.8125668883178335, |
|
"learning_rate": 1.3043650534467054e-06, |
|
"loss": 0.4873, |
|
"step": 1338 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.834975050326325, |
|
"learning_rate": 1.3009451096871084e-06, |
|
"loss": 0.5691, |
|
"step": 1339 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.8755759557839874, |
|
"learning_rate": 1.2975280777015315e-06, |
|
"loss": 0.5196, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.8301532732296394, |
|
"learning_rate": 1.2941139657878622e-06, |
|
"loss": 0.5081, |
|
"step": 1341 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.8555569411472547, |
|
"learning_rate": 1.2907027822369006e-06, |
|
"loss": 0.521, |
|
"step": 1342 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.8171279451549595, |
|
"learning_rate": 1.2872945353323335e-06, |
|
"loss": 0.5011, |
|
"step": 1343 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 0.844852389770996, |
|
"learning_rate": 1.2838892333507156e-06, |
|
"loss": 0.5047, |
|
"step": 1344 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 0.8393860493297807, |
|
"learning_rate": 1.2804868845614527e-06, |
|
"loss": 0.5287, |
|
"step": 1345 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 0.8452976842547626, |
|
"learning_rate": 1.2770874972267777e-06, |
|
"loss": 0.5767, |
|
"step": 1346 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 0.9456544440817343, |
|
"learning_rate": 1.2736910796017302e-06, |
|
"loss": 0.5577, |
|
"step": 1347 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 0.860914389892166, |
|
"learning_rate": 1.2702976399341422e-06, |
|
"loss": 0.4828, |
|
"step": 1348 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 0.8359148347078158, |
|
"learning_rate": 1.266907186464611e-06, |
|
"loss": 0.5229, |
|
"step": 1349 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 0.853947398083973, |
|
"learning_rate": 1.2635197274264814e-06, |
|
"loss": 0.512, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 0.8505728667108701, |
|
"learning_rate": 1.2601352710458314e-06, |
|
"loss": 0.5042, |
|
"step": 1351 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 0.8253375510200085, |
|
"learning_rate": 1.256753825541442e-06, |
|
"loss": 0.5126, |
|
"step": 1352 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 0.8616990286305032, |
|
"learning_rate": 1.2533753991247844e-06, |
|
"loss": 0.5608, |
|
"step": 1353 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.8378597147853472, |
|
"learning_rate": 1.2500000000000007e-06, |
|
"loss": 0.5187, |
|
"step": 1354 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.9086208709461716, |
|
"learning_rate": 1.2466276363638779e-06, |
|
"loss": 0.5061, |
|
"step": 1355 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.8385951040137415, |
|
"learning_rate": 1.2432583164058332e-06, |
|
"loss": 0.5084, |
|
"step": 1356 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.8514512681064714, |
|
"learning_rate": 1.2398920483078939e-06, |
|
"loss": 0.5416, |
|
"step": 1357 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.858376398756373, |
|
"learning_rate": 1.236528840244674e-06, |
|
"loss": 0.5461, |
|
"step": 1358 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.851819908353371, |
|
"learning_rate": 1.2331687003833568e-06, |
|
"loss": 0.5274, |
|
"step": 1359 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.8445208446367949, |
|
"learning_rate": 1.2298116368836772e-06, |
|
"loss": 0.5389, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.8329173435518172, |
|
"learning_rate": 1.2264576578978956e-06, |
|
"loss": 0.5165, |
|
"step": 1361 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.8594454164946037, |
|
"learning_rate": 1.2231067715707868e-06, |
|
"loss": 0.5141, |
|
"step": 1362 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.8661621348694677, |
|
"learning_rate": 1.219758986039611e-06, |
|
"loss": 0.5218, |
|
"step": 1363 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 0.8454693078199651, |
|
"learning_rate": 1.2164143094340993e-06, |
|
"loss": 0.5527, |
|
"step": 1364 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 0.8321073962447422, |
|
"learning_rate": 1.2130727498764356e-06, |
|
"loss": 0.5145, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 0.8296351178962421, |
|
"learning_rate": 1.2097343154812332e-06, |
|
"loss": 0.5133, |
|
"step": 1366 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 0.8251182765902709, |
|
"learning_rate": 1.2063990143555154e-06, |
|
"loss": 0.5415, |
|
"step": 1367 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 0.8856513775190644, |
|
"learning_rate": 1.203066854598696e-06, |
|
"loss": 0.5115, |
|
"step": 1368 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 0.8870324349365732, |
|
"learning_rate": 1.1997378443025633e-06, |
|
"loss": 0.5582, |
|
"step": 1369 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 0.811227552350083, |
|
"learning_rate": 1.196411991551255e-06, |
|
"loss": 0.5474, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 0.9187732616387055, |
|
"learning_rate": 1.1930893044212411e-06, |
|
"loss": 0.5533, |
|
"step": 1371 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 1.0617325433409386, |
|
"learning_rate": 1.1897697909813058e-06, |
|
"loss": 0.5388, |
|
"step": 1372 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 0.8471073077031555, |
|
"learning_rate": 1.1864534592925254e-06, |
|
"loss": 0.5961, |
|
"step": 1373 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 0.9528791078610926, |
|
"learning_rate": 1.183140317408248e-06, |
|
"loss": 0.532, |
|
"step": 1374 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 0.8614098986436766, |
|
"learning_rate": 1.1798303733740801e-06, |
|
"loss": 0.5106, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 0.852366435933281, |
|
"learning_rate": 1.1765236352278573e-06, |
|
"loss": 0.5071, |
|
"step": 1376 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 0.8731758573160572, |
|
"learning_rate": 1.1732201109996343e-06, |
|
"loss": 0.5606, |
|
"step": 1377 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 0.9275550484350488, |
|
"learning_rate": 1.169919808711659e-06, |
|
"loss": 0.5238, |
|
"step": 1378 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 0.7780678419258684, |
|
"learning_rate": 1.1666227363783538e-06, |
|
"loss": 0.4793, |
|
"step": 1379 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 0.8584699266737411, |
|
"learning_rate": 1.1633289020063019e-06, |
|
"loss": 0.5423, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 0.8682777436540573, |
|
"learning_rate": 1.160038313594219e-06, |
|
"loss": 0.5799, |
|
"step": 1381 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 0.8575287908111978, |
|
"learning_rate": 1.1567509791329402e-06, |
|
"loss": 0.529, |
|
"step": 1382 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 0.9265206565081486, |
|
"learning_rate": 1.1534669066054e-06, |
|
"loss": 0.5503, |
|
"step": 1383 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 0.8295819727309957, |
|
"learning_rate": 1.1501861039866095e-06, |
|
"loss": 0.4875, |
|
"step": 1384 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 0.8593146867018754, |
|
"learning_rate": 1.1469085792436387e-06, |
|
"loss": 0.5509, |
|
"step": 1385 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 0.8943872555576587, |
|
"learning_rate": 1.1436343403356019e-06, |
|
"loss": 0.5332, |
|
"step": 1386 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 0.8429618370537719, |
|
"learning_rate": 1.140363395213629e-06, |
|
"loss": 0.5318, |
|
"step": 1387 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 0.8476452956026092, |
|
"learning_rate": 1.137095751820853e-06, |
|
"loss": 0.5178, |
|
"step": 1388 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 0.8419589103929042, |
|
"learning_rate": 1.1338314180923917e-06, |
|
"loss": 0.4904, |
|
"step": 1389 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 0.8643668678324875, |
|
"learning_rate": 1.130570401955322e-06, |
|
"loss": 0.5265, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 0.8253871370418006, |
|
"learning_rate": 1.1273127113286656e-06, |
|
"loss": 0.5147, |
|
"step": 1391 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 0.8605565052324231, |
|
"learning_rate": 1.1240583541233702e-06, |
|
"loss": 0.4983, |
|
"step": 1392 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 0.809370363413381, |
|
"learning_rate": 1.1208073382422866e-06, |
|
"loss": 0.5307, |
|
"step": 1393 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 0.8132567924919997, |
|
"learning_rate": 1.1175596715801515e-06, |
|
"loss": 0.5131, |
|
"step": 1394 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 0.7892671050985064, |
|
"learning_rate": 1.1143153620235705e-06, |
|
"loss": 0.4723, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 0.8104584079322232, |
|
"learning_rate": 1.1110744174509952e-06, |
|
"loss": 0.5551, |
|
"step": 1396 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 0.8060049696366666, |
|
"learning_rate": 1.1078368457327046e-06, |
|
"loss": 0.5028, |
|
"step": 1397 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 0.8460495542218409, |
|
"learning_rate": 1.1046026547307907e-06, |
|
"loss": 0.4973, |
|
"step": 1398 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 0.8536239473618448, |
|
"learning_rate": 1.1013718522991315e-06, |
|
"loss": 0.5046, |
|
"step": 1399 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 0.9247482810981781, |
|
"learning_rate": 1.0981444462833802e-06, |
|
"loss": 0.4894, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 0.8622494879666638, |
|
"learning_rate": 1.0949204445209394e-06, |
|
"loss": 0.5327, |
|
"step": 1401 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 0.8257850136186105, |
|
"learning_rate": 1.0916998548409449e-06, |
|
"loss": 0.4969, |
|
"step": 1402 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 0.8836478758114623, |
|
"learning_rate": 1.0884826850642492e-06, |
|
"loss": 0.6262, |
|
"step": 1403 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 0.8205614037500113, |
|
"learning_rate": 1.0852689430033971e-06, |
|
"loss": 0.5491, |
|
"step": 1404 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 0.8571764912452442, |
|
"learning_rate": 1.0820586364626103e-06, |
|
"loss": 0.5243, |
|
"step": 1405 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 0.8413673785079766, |
|
"learning_rate": 1.0788517732377696e-06, |
|
"loss": 0.5038, |
|
"step": 1406 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 0.8770926597666683, |
|
"learning_rate": 1.0756483611163915e-06, |
|
"loss": 0.5344, |
|
"step": 1407 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 0.8422754850515988, |
|
"learning_rate": 1.0724484078776121e-06, |
|
"loss": 0.5303, |
|
"step": 1408 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 0.8403033998613175, |
|
"learning_rate": 1.0692519212921707e-06, |
|
"loss": 0.5188, |
|
"step": 1409 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 0.8385664461898923, |
|
"learning_rate": 1.0660589091223854e-06, |
|
"loss": 0.5444, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 0.8943616740406539, |
|
"learning_rate": 1.0628693791221373e-06, |
|
"loss": 0.508, |
|
"step": 1411 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 0.8544571616386037, |
|
"learning_rate": 1.0596833390368536e-06, |
|
"loss": 0.5298, |
|
"step": 1412 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 0.8486718285936482, |
|
"learning_rate": 1.0565007966034844e-06, |
|
"loss": 0.507, |
|
"step": 1413 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 0.8971813572743373, |
|
"learning_rate": 1.0533217595504859e-06, |
|
"loss": 0.475, |
|
"step": 1414 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 0.8752701639286048, |
|
"learning_rate": 1.0501462355978047e-06, |
|
"loss": 0.5151, |
|
"step": 1415 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 0.8555844838654592, |
|
"learning_rate": 1.0469742324568536e-06, |
|
"loss": 0.4924, |
|
"step": 1416 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 0.8243277858133128, |
|
"learning_rate": 1.043805757830495e-06, |
|
"loss": 0.5035, |
|
"step": 1417 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 0.8033661658811092, |
|
"learning_rate": 1.040640819413026e-06, |
|
"loss": 0.5021, |
|
"step": 1418 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 0.8232809177991901, |
|
"learning_rate": 1.0374794248901526e-06, |
|
"loss": 0.5381, |
|
"step": 1419 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 0.8668238731300797, |
|
"learning_rate": 1.0343215819389782e-06, |
|
"loss": 0.4718, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 0.8871599864379339, |
|
"learning_rate": 1.031167298227979e-06, |
|
"loss": 0.5344, |
|
"step": 1421 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 0.8344980205686212, |
|
"learning_rate": 1.0280165814169884e-06, |
|
"loss": 0.5528, |
|
"step": 1422 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 0.8720667208065984, |
|
"learning_rate": 1.02486943915718e-06, |
|
"loss": 0.5077, |
|
"step": 1423 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 0.8227881125973793, |
|
"learning_rate": 1.0217258790910447e-06, |
|
"loss": 0.534, |
|
"step": 1424 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 0.8654325720102245, |
|
"learning_rate": 1.0185859088523748e-06, |
|
"loss": 0.5244, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 0.8877041052770681, |
|
"learning_rate": 1.0154495360662464e-06, |
|
"loss": 0.5493, |
|
"step": 1426 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 1.0407163692565926, |
|
"learning_rate": 1.0123167683489986e-06, |
|
"loss": 0.4669, |
|
"step": 1427 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 0.8652377608101248, |
|
"learning_rate": 1.009187613308215e-06, |
|
"loss": 0.5635, |
|
"step": 1428 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 0.8291447446237054, |
|
"learning_rate": 1.0060620785427083e-06, |
|
"loss": 0.4751, |
|
"step": 1429 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 0.8784145833569265, |
|
"learning_rate": 1.0029401716424994e-06, |
|
"loss": 0.5167, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 0.8547561883276696, |
|
"learning_rate": 9.99821900188798e-07, |
|
"loss": 0.558, |
|
"step": 1431 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 0.8269118338521745, |
|
"learning_rate": 9.967072717539852e-07, |
|
"loss": 0.4938, |
|
"step": 1432 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 0.8286690177143268, |
|
"learning_rate": 9.935962939015983e-07, |
|
"loss": 0.5467, |
|
"step": 1433 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 0.8603505088361836, |
|
"learning_rate": 9.90488974186306e-07, |
|
"loss": 0.5451, |
|
"step": 1434 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.8771141261263814, |
|
"learning_rate": 9.873853201538972e-07, |
|
"loss": 0.5511, |
|
"step": 1435 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.8120198874711716, |
|
"learning_rate": 9.842853393412557e-07, |
|
"loss": 0.5337, |
|
"step": 1436 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.8569042192567161, |
|
"learning_rate": 9.811890392763468e-07, |
|
"loss": 0.5171, |
|
"step": 1437 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.8100510149829135, |
|
"learning_rate": 9.780964274781984e-07, |
|
"loss": 0.4836, |
|
"step": 1438 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.8778142778025754, |
|
"learning_rate": 9.750075114568805e-07, |
|
"loss": 0.4857, |
|
"step": 1439 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.8415118605491905, |
|
"learning_rate": 9.719222987134878e-07, |
|
"loss": 0.5077, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.8545136377771673, |
|
"learning_rate": 9.688407967401248e-07, |
|
"loss": 0.5581, |
|
"step": 1441 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.8314147470505274, |
|
"learning_rate": 9.657630130198819e-07, |
|
"loss": 0.5927, |
|
"step": 1442 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.8434714833853477, |
|
"learning_rate": 9.626889550268204e-07, |
|
"loss": 0.4988, |
|
"step": 1443 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.8407604245838901, |
|
"learning_rate": 9.596186302259563e-07, |
|
"loss": 0.5065, |
|
"step": 1444 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 0.8431911766178656, |
|
"learning_rate": 9.56552046073238e-07, |
|
"loss": 0.5128, |
|
"step": 1445 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 0.8665214143114242, |
|
"learning_rate": 9.534892100155296e-07, |
|
"loss": 0.5105, |
|
"step": 1446 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 0.8245774786573771, |
|
"learning_rate": 9.504301294905966e-07, |
|
"loss": 0.5175, |
|
"step": 1447 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 0.8498733195274237, |
|
"learning_rate": 9.473748119270807e-07, |
|
"loss": 0.5111, |
|
"step": 1448 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 0.8649542163477292, |
|
"learning_rate": 9.443232647444869e-07, |
|
"loss": 0.5531, |
|
"step": 1449 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 0.8203647257128365, |
|
"learning_rate": 9.412754953531664e-07, |
|
"loss": 0.4999, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 0.8907771180053089, |
|
"learning_rate": 9.382315111542939e-07, |
|
"loss": 0.5346, |
|
"step": 1451 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 0.8827015600189333, |
|
"learning_rate": 9.351913195398523e-07, |
|
"loss": 0.5276, |
|
"step": 1452 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 0.8206255951064396, |
|
"learning_rate": 9.32154927892617e-07, |
|
"loss": 0.4765, |
|
"step": 1453 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 0.8314786262814058, |
|
"learning_rate": 9.291223435861318e-07, |
|
"loss": 0.506, |
|
"step": 1454 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 0.9540637457834469, |
|
"learning_rate": 9.260935739846993e-07, |
|
"loss": 0.5246, |
|
"step": 1455 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 0.910183703936372, |
|
"learning_rate": 9.230686264433547e-07, |
|
"loss": 0.514, |
|
"step": 1456 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 0.8388131083458139, |
|
"learning_rate": 9.200475083078528e-07, |
|
"loss": 0.5195, |
|
"step": 1457 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 0.9048019571114666, |
|
"learning_rate": 9.170302269146509e-07, |
|
"loss": 0.5151, |
|
"step": 1458 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 0.8568997007043364, |
|
"learning_rate": 9.140167895908867e-07, |
|
"loss": 0.5045, |
|
"step": 1459 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 0.8016795001871563, |
|
"learning_rate": 9.110072036543635e-07, |
|
"loss": 0.4999, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 0.8403773782643171, |
|
"learning_rate": 9.080014764135345e-07, |
|
"loss": 0.4793, |
|
"step": 1461 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 0.8198349391449373, |
|
"learning_rate": 9.04999615167479e-07, |
|
"loss": 0.5186, |
|
"step": 1462 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 0.8102637591378546, |
|
"learning_rate": 9.02001627205889e-07, |
|
"loss": 0.5143, |
|
"step": 1463 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 0.841717937055826, |
|
"learning_rate": 8.990075198090531e-07, |
|
"loss": 0.4987, |
|
"step": 1464 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 0.983003680660331, |
|
"learning_rate": 8.960173002478336e-07, |
|
"loss": 0.5241, |
|
"step": 1465 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 0.8831136377325097, |
|
"learning_rate": 8.930309757836517e-07, |
|
"loss": 0.5507, |
|
"step": 1466 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 0.859848781870307, |
|
"learning_rate": 8.900485536684725e-07, |
|
"loss": 0.5265, |
|
"step": 1467 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 0.7951663558174535, |
|
"learning_rate": 8.870700411447817e-07, |
|
"loss": 0.5044, |
|
"step": 1468 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 0.8179578542889253, |
|
"learning_rate": 8.840954454455713e-07, |
|
"loss": 0.4785, |
|
"step": 1469 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 0.9213326990807826, |
|
"learning_rate": 8.811247737943243e-07, |
|
"loss": 0.5342, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 0.9283254621009271, |
|
"learning_rate": 8.781580334049919e-07, |
|
"loss": 0.5181, |
|
"step": 1471 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 0.83886555440469, |
|
"learning_rate": 8.751952314819786e-07, |
|
"loss": 0.5146, |
|
"step": 1472 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 0.8515206673929563, |
|
"learning_rate": 8.722363752201277e-07, |
|
"loss": 0.4973, |
|
"step": 1473 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 0.8350109546756519, |
|
"learning_rate": 8.692814718046979e-07, |
|
"loss": 0.4989, |
|
"step": 1474 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 0.8586124134805211, |
|
"learning_rate": 8.663305284113491e-07, |
|
"loss": 0.518, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.8467810373632807, |
|
"learning_rate": 8.633835522061268e-07, |
|
"loss": 0.5421, |
|
"step": 1476 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.8278264544831425, |
|
"learning_rate": 8.604405503454399e-07, |
|
"loss": 0.5455, |
|
"step": 1477 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.8577344792736433, |
|
"learning_rate": 8.575015299760492e-07, |
|
"loss": 0.5441, |
|
"step": 1478 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.8264037326728428, |
|
"learning_rate": 8.545664982350438e-07, |
|
"loss": 0.5094, |
|
"step": 1479 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.9149752535547083, |
|
"learning_rate": 8.516354622498279e-07, |
|
"loss": 0.5404, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.9869351503652237, |
|
"learning_rate": 8.48708429138104e-07, |
|
"loss": 0.604, |
|
"step": 1481 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.9089542327118894, |
|
"learning_rate": 8.457854060078521e-07, |
|
"loss": 0.536, |
|
"step": 1482 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.9092749602931682, |
|
"learning_rate": 8.428663999573142e-07, |
|
"loss": 0.5203, |
|
"step": 1483 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.9055831460836471, |
|
"learning_rate": 8.399514180749796e-07, |
|
"loss": 0.5279, |
|
"step": 1484 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.8096286326174174, |
|
"learning_rate": 8.370404674395632e-07, |
|
"loss": 0.516, |
|
"step": 1485 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 0.8676917398179171, |
|
"learning_rate": 8.341335551199903e-07, |
|
"loss": 0.5169, |
|
"step": 1486 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 0.8134259776638351, |
|
"learning_rate": 8.31230688175382e-07, |
|
"loss": 0.5191, |
|
"step": 1487 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 0.8730503437885434, |
|
"learning_rate": 8.283318736550333e-07, |
|
"loss": 0.4963, |
|
"step": 1488 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 0.886556042277387, |
|
"learning_rate": 8.254371185983981e-07, |
|
"loss": 0.5149, |
|
"step": 1489 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 0.7871526041338943, |
|
"learning_rate": 8.225464300350752e-07, |
|
"loss": 0.5248, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 0.8422784707849628, |
|
"learning_rate": 8.196598149847845e-07, |
|
"loss": 0.524, |
|
"step": 1491 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 0.812381669542707, |
|
"learning_rate": 8.16777280457357e-07, |
|
"loss": 0.5066, |
|
"step": 1492 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 0.8271280669942666, |
|
"learning_rate": 8.138988334527143e-07, |
|
"loss": 0.5497, |
|
"step": 1493 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 0.8558092006065072, |
|
"learning_rate": 8.110244809608494e-07, |
|
"loss": 0.5186, |
|
"step": 1494 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 0.8690544832253072, |
|
"learning_rate": 8.081542299618139e-07, |
|
"loss": 0.5535, |
|
"step": 1495 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 0.8161077141719014, |
|
"learning_rate": 8.052880874257005e-07, |
|
"loss": 0.5123, |
|
"step": 1496 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 0.8550803144571056, |
|
"learning_rate": 8.024260603126229e-07, |
|
"loss": 0.5533, |
|
"step": 1497 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 0.8164787852606792, |
|
"learning_rate": 7.995681555727011e-07, |
|
"loss": 0.5557, |
|
"step": 1498 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 0.8298798471160497, |
|
"learning_rate": 7.967143801460467e-07, |
|
"loss": 0.5136, |
|
"step": 1499 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 0.9003048698578533, |
|
"learning_rate": 7.938647409627412e-07, |
|
"loss": 0.5162, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 0.8011312821601281, |
|
"learning_rate": 7.910192449428216e-07, |
|
"loss": 0.5319, |
|
"step": 1501 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 0.8224781558901803, |
|
"learning_rate": 7.881778989962663e-07, |
|
"loss": 0.517, |
|
"step": 1502 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 0.8642229333640068, |
|
"learning_rate": 7.853407100229732e-07, |
|
"loss": 0.5474, |
|
"step": 1503 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 0.8944127773161233, |
|
"learning_rate": 7.825076849127458e-07, |
|
"loss": 0.5109, |
|
"step": 1504 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 0.8649465092530685, |
|
"learning_rate": 7.796788305452777e-07, |
|
"loss": 0.5325, |
|
"step": 1505 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 0.8417242917461267, |
|
"learning_rate": 7.768541537901325e-07, |
|
"loss": 0.4949, |
|
"step": 1506 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 0.8376649820399588, |
|
"learning_rate": 7.740336615067293e-07, |
|
"loss": 0.5107, |
|
"step": 1507 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 0.8001119104415296, |
|
"learning_rate": 7.712173605443269e-07, |
|
"loss": 0.4866, |
|
"step": 1508 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 0.8139978607710936, |
|
"learning_rate": 7.684052577420043e-07, |
|
"loss": 0.51, |
|
"step": 1509 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 0.8446594795021735, |
|
"learning_rate": 7.655973599286459e-07, |
|
"loss": 0.5401, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 0.8643631004269995, |
|
"learning_rate": 7.627936739229272e-07, |
|
"loss": 0.4849, |
|
"step": 1511 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 0.8850121686014331, |
|
"learning_rate": 7.599942065332916e-07, |
|
"loss": 0.5277, |
|
"step": 1512 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 0.8166347055921038, |
|
"learning_rate": 7.57198964557942e-07, |
|
"loss": 0.5308, |
|
"step": 1513 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 1.1188794899577097, |
|
"learning_rate": 7.544079547848183e-07, |
|
"loss": 0.5531, |
|
"step": 1514 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 0.8944795237318417, |
|
"learning_rate": 7.516211839915821e-07, |
|
"loss": 0.5206, |
|
"step": 1515 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 0.8394799079994575, |
|
"learning_rate": 7.488386589456043e-07, |
|
"loss": 0.5489, |
|
"step": 1516 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 0.8314016040692069, |
|
"learning_rate": 7.46060386403942e-07, |
|
"loss": 0.5076, |
|
"step": 1517 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 0.811099300082134, |
|
"learning_rate": 7.432863731133271e-07, |
|
"loss": 0.5455, |
|
"step": 1518 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 0.8584417095096406, |
|
"learning_rate": 7.405166258101495e-07, |
|
"loss": 0.505, |
|
"step": 1519 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 0.8558832162720049, |
|
"learning_rate": 7.377511512204375e-07, |
|
"loss": 0.5656, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 0.9736385454138048, |
|
"learning_rate": 7.349899560598439e-07, |
|
"loss": 0.5136, |
|
"step": 1521 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"eval_loss": 0.5786285400390625, |
|
"eval_runtime": 1882.6522, |
|
"eval_samples_per_second": 0.702, |
|
"eval_steps_per_second": 0.078, |
|
"step": 1521 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 2026, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 507, |
|
"total_flos": 716431750594560.0, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|