|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 500, |
|
"global_step": 520, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 6.25e-06, |
|
"loss": 2.1094, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.25e-05, |
|
"loss": 2.1719, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.8750000000000002e-05, |
|
"loss": 1.8203, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.5e-05, |
|
"loss": 1.1953, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 3.125e-05, |
|
"loss": 1.0938, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 3.7500000000000003e-05, |
|
"loss": 1.125, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.375e-05, |
|
"loss": 1.0625, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 5e-05, |
|
"loss": 1.0859, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 5.6250000000000005e-05, |
|
"loss": 1.0156, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 6.25e-05, |
|
"loss": 1.0, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 6.875e-05, |
|
"loss": 0.9844, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 7.500000000000001e-05, |
|
"loss": 0.9805, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 8.125000000000001e-05, |
|
"loss": 0.9727, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 8.75e-05, |
|
"loss": 0.9648, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.375e-05, |
|
"loss": 0.9258, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0001, |
|
"loss": 0.9531, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.999902864657691e-05, |
|
"loss": 0.9453, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.999611462404875e-05, |
|
"loss": 0.957, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.999125804563732e-05, |
|
"loss": 0.9375, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.998445910004082e-05, |
|
"loss": 0.9102, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.997571805142639e-05, |
|
"loss": 0.9375, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.996503523941994e-05, |
|
"loss": 0.918, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.99524110790929e-05, |
|
"loss": 0.8906, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.993784606094612e-05, |
|
"loss": 0.9102, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.992134075089084e-05, |
|
"loss": 0.8984, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.99028957902266e-05, |
|
"loss": 0.8945, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.988251189561645e-05, |
|
"loss": 0.9219, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.986018985905901e-05, |
|
"loss": 0.8789, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.983593054785776e-05, |
|
"loss": 0.8984, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.980973490458728e-05, |
|
"loss": 0.8945, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.978160394705668e-05, |
|
"loss": 0.8789, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.975153876827008e-05, |
|
"loss": 0.9062, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.971954053638399e-05, |
|
"loss": 0.875, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.968561049466214e-05, |
|
"loss": 0.8594, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.964974996142698e-05, |
|
"loss": 0.8672, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.961196033000861e-05, |
|
"loss": 0.8711, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.957224306869053e-05, |
|
"loss": 0.8828, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.953059972065265e-05, |
|
"loss": 0.8594, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.948703190391131e-05, |
|
"loss": 0.8672, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.944154131125642e-05, |
|
"loss": 0.8789, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.939412971018574e-05, |
|
"loss": 0.875, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.934479894283606e-05, |
|
"loss": 0.8672, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.92935509259118e-05, |
|
"loss": 0.8594, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.924038765061042e-05, |
|
"loss": 0.8711, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.918531118254507e-05, |
|
"loss": 0.8672, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.912832366166442e-05, |
|
"loss": 0.8516, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.906942730216939e-05, |
|
"loss": 0.8477, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.900862439242719e-05, |
|
"loss": 0.8555, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.894591729488242e-05, |
|
"loss": 0.875, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.888130844596524e-05, |
|
"loss": 0.8438, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.881480035599667e-05, |
|
"loss": 0.8516, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.874639560909117e-05, |
|
"loss": 0.8789, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.867609686305617e-05, |
|
"loss": 0.8711, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.860390684928873e-05, |
|
"loss": 0.8594, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.852982837266955e-05, |
|
"loss": 0.8555, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.84538643114539e-05, |
|
"loss": 0.8633, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.837601761715983e-05, |
|
"loss": 0.8438, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.829629131445342e-05, |
|
"loss": 0.8438, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.82146885010314e-05, |
|
"loss": 0.8555, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.81312123475006e-05, |
|
"loss": 0.8477, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.804586609725499e-05, |
|
"loss": 0.875, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.79586530663494e-05, |
|
"loss": 0.8555, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.78695766433709e-05, |
|
"loss": 0.8438, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.777864028930705e-05, |
|
"loss": 0.8594, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.768584753741134e-05, |
|
"loss": 0.8203, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.759120199306613e-05, |
|
"loss": 0.8281, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.74947073336423e-05, |
|
"loss": 0.8438, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.73963673083566e-05, |
|
"loss": 0.8594, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.72961857381258e-05, |
|
"loss": 0.8555, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.719416651541839e-05, |
|
"loss": 0.8242, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.709031360410318e-05, |
|
"loss": 0.8555, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.698463103929542e-05, |
|
"loss": 0.8359, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.687712292719997e-05, |
|
"loss": 0.8438, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.67677934449517e-05, |
|
"loss": 0.8594, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.665664684045333e-05, |
|
"loss": 0.8516, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.654368743221022e-05, |
|
"loss": 0.8242, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.642891960916268e-05, |
|
"loss": 0.8281, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.631234783051544e-05, |
|
"loss": 0.832, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.619397662556435e-05, |
|
"loss": 0.8398, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.607381059352038e-05, |
|
"loss": 0.8359, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.595185440333103e-05, |
|
"loss": 0.8086, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.582811279349882e-05, |
|
"loss": 0.8281, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.570259057189717e-05, |
|
"loss": 0.8125, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.557529261558367e-05, |
|
"loss": 0.8359, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.544622387061055e-05, |
|
"loss": 0.8398, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.53153893518325e-05, |
|
"loss": 0.8516, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.518279414271183e-05, |
|
"loss": 0.8555, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.504844339512095e-05, |
|
"loss": 0.8438, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.491234232914221e-05, |
|
"loss": 0.8203, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.477449623286505e-05, |
|
"loss": 0.8242, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.463491046218058e-05, |
|
"loss": 0.8164, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.449359044057345e-05, |
|
"loss": 0.8164, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.435054165891109e-05, |
|
"loss": 0.8164, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.420576967523049e-05, |
|
"loss": 0.8359, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.405928011452211e-05, |
|
"loss": 0.8398, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.391107866851143e-05, |
|
"loss": 0.8164, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.376117109543769e-05, |
|
"loss": 0.8516, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.360956321983028e-05, |
|
"loss": 0.8281, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.345626093228233e-05, |
|
"loss": 0.8164, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.330127018922194e-05, |
|
"loss": 0.8555, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.314459701268065e-05, |
|
"loss": 0.8125, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.298624749005951e-05, |
|
"loss": 0.8242, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.282622777389258e-05, |
|
"loss": 0.8281, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.266454408160779e-05, |
|
"loss": 0.832, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.250120269528546e-05, |
|
"loss": 0.8242, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.233620996141421e-05, |
|
"loss": 0.8438, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.21695722906443e-05, |
|
"loss": 0.8281, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.200129615753859e-05, |
|
"loss": 0.8203, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.183138810032099e-05, |
|
"loss": 0.8398, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.165985472062246e-05, |
|
"loss": 0.8359, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.148670268322438e-05, |
|
"loss": 0.8281, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.131193871579975e-05, |
|
"loss": 0.8125, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.113556960865167e-05, |
|
"loss": 0.8359, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.09576022144496e-05, |
|
"loss": 0.8281, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.077804344796302e-05, |
|
"loss": 0.8047, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.059690028579284e-05, |
|
"loss": 0.8125, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.041417976610027e-05, |
|
"loss": 0.8203, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.022988898833342e-05, |
|
"loss": 0.832, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.004403511295141e-05, |
|
"loss": 0.8242, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.985662536114613e-05, |
|
"loss": 0.8203, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.966766701456177e-05, |
|
"loss": 0.8164, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.947716741501177e-05, |
|
"loss": 0.8242, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.928513396419368e-05, |
|
"loss": 0.8008, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.90915741234015e-05, |
|
"loss": 0.8125, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.889649541323574e-05, |
|
"loss": 0.832, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.869990541331138e-05, |
|
"loss": 0.8047, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.850181176196315e-05, |
|
"loss": 0.8086, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.83022221559489e-05, |
|
"loss": 0.8164, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.810114435015054e-05, |
|
"loss": 0.8047, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.789858615727265e-05, |
|
"loss": 0.8047, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.7694555447539e-05, |
|
"loss": 0.8359, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.748906014838672e-05, |
|
"loss": 0.832, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.728210824415827e-05, |
|
"loss": 0.8086, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.707370777579133e-05, |
|
"loss": 0.8008, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.68638668405062e-05, |
|
"loss": 0.8164, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.665259359149132e-05, |
|
"loss": 0.8125, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.643989623758643e-05, |
|
"loss": 0.8281, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.622578304296364e-05, |
|
"loss": 0.8125, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.601026232680634e-05, |
|
"loss": 0.8047, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.579334246298593e-05, |
|
"loss": 0.8008, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.557503187973651e-05, |
|
"loss": 0.8164, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.535533905932738e-05, |
|
"loss": 0.8164, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.513427253773346e-05, |
|
"loss": 0.8047, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.491184090430364e-05, |
|
"loss": 0.8008, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.468805280142709e-05, |
|
"loss": 0.7969, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.446291692419736e-05, |
|
"loss": 0.8086, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.423644202007467e-05, |
|
"loss": 0.7969, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.400863688854597e-05, |
|
"loss": 0.8281, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.377951038078302e-05, |
|
"loss": 0.8203, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.354907139929851e-05, |
|
"loss": 0.8164, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.33173288976002e-05, |
|
"loss": 0.8203, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.308429187984297e-05, |
|
"loss": 0.8125, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.284996940047903e-05, |
|
"loss": 0.8242, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.261437056390606e-05, |
|
"loss": 0.8125, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.237750452411353e-05, |
|
"loss": 0.8164, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.213938048432697e-05, |
|
"loss": 0.8164, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.190000769665044e-05, |
|
"loss": 0.8203, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.1659395461707e-05, |
|
"loss": 0.7969, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 8.141755312827736e-05, |
|
"loss": 0.8125, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 8.117449009293668e-05, |
|
"loss": 0.8086, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 8.093021579968941e-05, |
|
"loss": 0.8047, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 8.068473973960238e-05, |
|
"loss": 0.8047, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 8.043807145043604e-05, |
|
"loss": 0.793, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 8.019022051627388e-05, |
|
"loss": 0.8359, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.994119656715002e-05, |
|
"loss": 0.7891, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.969100927867507e-05, |
|
"loss": 0.8047, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.943966837166023e-05, |
|
"loss": 0.8008, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.91871836117395e-05, |
|
"loss": 0.8125, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.89335648089903e-05, |
|
"loss": 0.793, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.86788218175523e-05, |
|
"loss": 0.8125, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.842296453524463e-05, |
|
"loss": 0.8047, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.81660029031811e-05, |
|
"loss": 0.7969, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.79079469053842e-05, |
|
"loss": 0.7969, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.764880656839696e-05, |
|
"loss": 0.8203, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.738859196089358e-05, |
|
"loss": 0.8008, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.712731319328798e-05, |
|
"loss": 0.8203, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.68649804173412e-05, |
|
"loss": 0.7891, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.660160382576683e-05, |
|
"loss": 0.8008, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.633719365183504e-05, |
|
"loss": 0.8125, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.60717601689749e-05, |
|
"loss": 0.793, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.580531369037533e-05, |
|
"loss": 0.7969, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.553786456858429e-05, |
|
"loss": 0.7969, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.526942319510655e-05, |
|
"loss": 0.8008, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.500000000000001e-05, |
|
"loss": 0.8008, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.472960545147038e-05, |
|
"loss": 0.8008, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.445825005546448e-05, |
|
"loss": 0.8125, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.4185944355262e-05, |
|
"loss": 0.8203, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.391269893106592e-05, |
|
"loss": 0.8164, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.363852439959135e-05, |
|
"loss": 0.8203, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.33634314136531e-05, |
|
"loss": 0.8164, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.308743066175172e-05, |
|
"loss": 0.7969, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.281053286765815e-05, |
|
"loss": 0.7852, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.253274878999727e-05, |
|
"loss": 0.8164, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.225408922182961e-05, |
|
"loss": 0.8203, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 7.197456499023225e-05, |
|
"loss": 0.7891, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 7.169418695587791e-05, |
|
"loss": 0.8125, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 7.141296601261314e-05, |
|
"loss": 0.8008, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 7.113091308703498e-05, |
|
"loss": 0.7891, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 7.084803913806641e-05, |
|
"loss": 0.8086, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 7.056435515653059e-05, |
|
"loss": 0.8125, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 7.027987216472377e-05, |
|
"loss": 0.8125, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.999460121598704e-05, |
|
"loss": 0.7852, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.970855339427698e-05, |
|
"loss": 0.7891, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.942173981373474e-05, |
|
"loss": 0.8125, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.91341716182545e-05, |
|
"loss": 0.8125, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.884585998105026e-05, |
|
"loss": 0.8047, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.855681610422189e-05, |
|
"loss": 0.8008, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.826705121831976e-05, |
|
"loss": 0.8047, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.797657658190839e-05, |
|
"loss": 0.8164, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.768540348112907e-05, |
|
"loss": 0.8047, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.739354322926136e-05, |
|
"loss": 0.8164, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.710100716628344e-05, |
|
"loss": 0.7969, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.680780665843155e-05, |
|
"loss": 0.7969, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.651395309775837e-05, |
|
"loss": 0.793, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.621945790169036e-05, |
|
"loss": 0.8047, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.592433251258423e-05, |
|
"loss": 0.7969, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.562858839728223e-05, |
|
"loss": 0.7969, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.533223704666672e-05, |
|
"loss": 0.7773, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.503528997521366e-05, |
|
"loss": 0.8086, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.473775872054521e-05, |
|
"loss": 0.8086, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.44396548429815e-05, |
|
"loss": 0.793, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.414098992509138e-05, |
|
"loss": 0.7852, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.384177557124247e-05, |
|
"loss": 0.7969, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.354202340715026e-05, |
|
"loss": 0.8164, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.324174507942637e-05, |
|
"loss": 0.7695, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.294095225512603e-05, |
|
"loss": 0.7852, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 6.263965662129487e-05, |
|
"loss": 0.7891, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 6.233786988451468e-05, |
|
"loss": 0.8008, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 6.203560377044866e-05, |
|
"loss": 0.793, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 6.173287002338577e-05, |
|
"loss": 0.793, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 6.142968040578449e-05, |
|
"loss": 0.7969, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 6.112604669781572e-05, |
|
"loss": 0.8047, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 6.0821980696905146e-05, |
|
"loss": 0.7812, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 6.0517494217274794e-05, |
|
"loss": 0.8047, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 6.021259908948402e-05, |
|
"loss": 0.7734, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.9907307159969884e-05, |
|
"loss": 0.7734, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.960163029058682e-05, |
|
"loss": 0.793, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.9295580358145744e-05, |
|
"loss": 0.7812, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.898916925395264e-05, |
|
"loss": 0.7969, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.868240888334653e-05, |
|
"loss": 0.7891, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.837531116523682e-05, |
|
"loss": 0.7695, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.806788803164034e-05, |
|
"loss": 0.7812, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.7760151427217576e-05, |
|
"loss": 0.8008, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.745211330880872e-05, |
|
"loss": 0.7969, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.714378564496901e-05, |
|
"loss": 0.7773, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.683518041550368e-05, |
|
"loss": 0.7812, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.6526309611002594e-05, |
|
"loss": 0.7773, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.621718523237427e-05, |
|
"loss": 0.7969, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.590781929037965e-05, |
|
"loss": 0.7852, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.559822380516539e-05, |
|
"loss": 0.8008, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.5288410805796895e-05, |
|
"loss": 0.7852, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.497839232979084e-05, |
|
"loss": 0.8008, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.466818042264753e-05, |
|
"loss": 0.7969, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.435778713738292e-05, |
|
"loss": 0.793, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.404722453406017e-05, |
|
"loss": 0.7812, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.373650467932122e-05, |
|
"loss": 0.7773, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.3425639645917834e-05, |
|
"loss": 0.793, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.311464151224261e-05, |
|
"loss": 0.7734, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.2803522361859594e-05, |
|
"loss": 0.7773, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.249229428303486e-05, |
|
"loss": 0.7773, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.218096936826681e-05, |
|
"loss": 0.7812, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.18695597138163e-05, |
|
"loss": 0.8047, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 5.155807741923666e-05, |
|
"loss": 0.7852, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 5.124653458690365e-05, |
|
"loss": 0.7617, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 5.0934943321545115e-05, |
|
"loss": 0.7852, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 5.062331572977076e-05, |
|
"loss": 0.7852, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 5.031166391960168e-05, |
|
"loss": 0.7812, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 5e-05, |
|
"loss": 0.8008, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.968833608039832e-05, |
|
"loss": 0.7852, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.9376684270229254e-05, |
|
"loss": 0.7773, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.9065056678454904e-05, |
|
"loss": 0.7812, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.875346541309637e-05, |
|
"loss": 0.793, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.844192258076336e-05, |
|
"loss": 0.7812, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.813044028618373e-05, |
|
"loss": 0.8008, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.781903063173321e-05, |
|
"loss": 0.793, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.750770571696514e-05, |
|
"loss": 0.7852, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.7196477638140404e-05, |
|
"loss": 0.7617, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.68853584877574e-05, |
|
"loss": 0.7734, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.657436035408217e-05, |
|
"loss": 0.7734, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.626349532067879e-05, |
|
"loss": 0.7773, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.595277546593984e-05, |
|
"loss": 0.7695, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.564221286261709e-05, |
|
"loss": 0.7891, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.5331819577352474e-05, |
|
"loss": 0.7734, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.502160767020918e-05, |
|
"loss": 0.7734, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.471158919420312e-05, |
|
"loss": 0.793, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.4401776194834613e-05, |
|
"loss": 0.7812, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.409218070962036e-05, |
|
"loss": 0.7891, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.378281476762576e-05, |
|
"loss": 0.7773, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.347369038899744e-05, |
|
"loss": 0.7734, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.316481958449634e-05, |
|
"loss": 0.7812, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.285621435503101e-05, |
|
"loss": 0.7734, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.254788669119127e-05, |
|
"loss": 0.7812, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.223984857278242e-05, |
|
"loss": 0.7812, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.1932111968359664e-05, |
|
"loss": 0.7773, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.162468883476319e-05, |
|
"loss": 0.7891, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.131759111665349e-05, |
|
"loss": 0.7773, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.101083074604737e-05, |
|
"loss": 0.7773, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.0704419641854274e-05, |
|
"loss": 0.8008, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.03983697094132e-05, |
|
"loss": 0.7656, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.0092692840030134e-05, |
|
"loss": 0.7773, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.978740091051599e-05, |
|
"loss": 0.7695, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.9482505782725224e-05, |
|
"loss": 0.7578, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.917801930309486e-05, |
|
"loss": 0.7812, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.887395330218429e-05, |
|
"loss": 0.7773, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.857031959421553e-05, |
|
"loss": 0.7773, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.8267129976614254e-05, |
|
"loss": 0.7734, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.7964396229551364e-05, |
|
"loss": 0.7617, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.7662130115485314e-05, |
|
"loss": 0.7656, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.7360343378705124e-05, |
|
"loss": 0.7852, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.705904774487396e-05, |
|
"loss": 0.7812, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.675825492057364e-05, |
|
"loss": 0.7773, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.6457976592849754e-05, |
|
"loss": 0.7695, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.6158224428757535e-05, |
|
"loss": 0.7578, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.585901007490863e-05, |
|
"loss": 0.793, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.556034515701852e-05, |
|
"loss": 0.7656, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.5262241279454785e-05, |
|
"loss": 0.7617, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.4964710024786354e-05, |
|
"loss": 0.7812, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.4667762953333295e-05, |
|
"loss": 0.7539, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.4371411602717784e-05, |
|
"loss": 0.7812, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.4075667487415785e-05, |
|
"loss": 0.7656, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.3780542098309654e-05, |
|
"loss": 0.7578, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.3486046902241664e-05, |
|
"loss": 0.7539, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.319219334156847e-05, |
|
"loss": 0.7656, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.289899283371657e-05, |
|
"loss": 0.7695, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.2606456770738636e-05, |
|
"loss": 0.7773, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.231459651887093e-05, |
|
"loss": 0.793, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.2023423418091626e-05, |
|
"loss": 0.7656, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.173294878168025e-05, |
|
"loss": 0.7617, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.1443183895778105e-05, |
|
"loss": 0.7773, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.115414001894974e-05, |
|
"loss": 0.7578, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.086582838174551e-05, |
|
"loss": 0.7656, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.0578260186265265e-05, |
|
"loss": 0.7539, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.029144660572304e-05, |
|
"loss": 0.7656, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.000539878401296e-05, |
|
"loss": 0.7734, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.9720127835276256e-05, |
|
"loss": 0.7617, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.9435644843469436e-05, |
|
"loss": 0.7734, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.9151960861933614e-05, |
|
"loss": 0.7539, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.886908691296504e-05, |
|
"loss": 0.7852, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.858703398738686e-05, |
|
"loss": 0.7852, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.8305813044122097e-05, |
|
"loss": 0.7656, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.8025435009767747e-05, |
|
"loss": 0.7656, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.774591077817038e-05, |
|
"loss": 0.7656, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.746725121000273e-05, |
|
"loss": 0.7656, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.718946713234185e-05, |
|
"loss": 0.7617, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.6912569338248315e-05, |
|
"loss": 0.7656, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.66365685863469e-05, |
|
"loss": 0.7578, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.636147560040866e-05, |
|
"loss": 0.7695, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.6087301068934106e-05, |
|
"loss": 0.7617, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.581405564473801e-05, |
|
"loss": 0.7695, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.5541749944535554e-05, |
|
"loss": 0.7617, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.527039454852963e-05, |
|
"loss": 0.7812, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.500000000000001e-05, |
|
"loss": 0.7578, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.473057680489348e-05, |
|
"loss": 0.7617, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.4462135431415733e-05, |
|
"loss": 0.7656, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.4194686309624663e-05, |
|
"loss": 0.75, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.39282398310251e-05, |
|
"loss": 0.7617, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.366280634816496e-05, |
|
"loss": 0.7656, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.3398396174233178e-05, |
|
"loss": 0.7695, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.3135019582658802e-05, |
|
"loss": 0.7422, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.2872686806712035e-05, |
|
"loss": 0.7695, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.261140803910644e-05, |
|
"loss": 0.7695, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.235119343160303e-05, |
|
"loss": 0.7539, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.2092053094615813e-05, |
|
"loss": 0.7695, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.1833997096818898e-05, |
|
"loss": 0.7773, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.157703546475539e-05, |
|
"loss": 0.7461, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.132117818244771e-05, |
|
"loss": 0.75, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.1066435191009715e-05, |
|
"loss": 0.7422, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.0812816388260518e-05, |
|
"loss": 0.7617, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.056033162833977e-05, |
|
"loss": 0.7695, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.0308990721324927e-05, |
|
"loss": 0.7617, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.0058803432849987e-05, |
|
"loss": 0.7656, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.980977948372612e-05, |
|
"loss": 0.7461, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.9561928549563968e-05, |
|
"loss": 0.7461, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.931526026039764e-05, |
|
"loss": 0.7461, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.906978420031059e-05, |
|
"loss": 0.7422, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.8825509907063327e-05, |
|
"loss": 0.75, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.8582446871722636e-05, |
|
"loss": 0.7344, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.8340604538293015e-05, |
|
"loss": 0.7461, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.8099992303349577e-05, |
|
"loss": 0.7422, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.7860619515673033e-05, |
|
"loss": 0.7617, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.7622495475886487e-05, |
|
"loss": 0.7539, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.738562943609396e-05, |
|
"loss": 0.7539, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.7150030599520984e-05, |
|
"loss": 0.7617, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.691570812015704e-05, |
|
"loss": 0.7422, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.6682671102399805e-05, |
|
"loss": 0.75, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.6450928600701504e-05, |
|
"loss": 0.75, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.622048961921699e-05, |
|
"loss": 0.7617, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.599136311145402e-05, |
|
"loss": 0.7695, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.5763557979925324e-05, |
|
"loss": 0.7461, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.553708307580265e-05, |
|
"loss": 0.7617, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.531194719857292e-05, |
|
"loss": 0.793, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.5088159095696363e-05, |
|
"loss": 0.7539, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.4865727462266543e-05, |
|
"loss": 0.7734, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.4644660940672627e-05, |
|
"loss": 0.7734, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.4424968120263504e-05, |
|
"loss": 0.7461, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.4206657537014079e-05, |
|
"loss": 0.7578, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.398973767319368e-05, |
|
"loss": 0.7578, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.3774216957036367e-05, |
|
"loss": 0.7539, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.3560103762413584e-05, |
|
"loss": 0.7539, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.3347406408508695e-05, |
|
"loss": 0.7539, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.3136133159493802e-05, |
|
"loss": 0.7617, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.2926292224208664e-05, |
|
"loss": 0.7422, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.2717891755841722e-05, |
|
"loss": 0.7734, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.2510939851613285e-05, |
|
"loss": 0.7734, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.230544455246101e-05, |
|
"loss": 0.75, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.2101413842727345e-05, |
|
"loss": 0.7461, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.1898855649849461e-05, |
|
"loss": 0.7539, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.1697777844051105e-05, |
|
"loss": 0.7305, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.1498188238036861e-05, |
|
"loss": 0.7422, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.130009458668863e-05, |
|
"loss": 0.75, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.1103504586764263e-05, |
|
"loss": 0.7539, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.090842587659851e-05, |
|
"loss": 0.7617, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.0714866035806326e-05, |
|
"loss": 0.7422, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.0522832584988234e-05, |
|
"loss": 0.7617, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.0332332985438248e-05, |
|
"loss": 0.7656, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.0143374638853891e-05, |
|
"loss": 0.7656, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.955964887048607e-06, |
|
"loss": 0.7656, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.770111011666583e-06, |
|
"loss": 0.7617, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.58582023389974e-06, |
|
"loss": 0.7461, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.403099714207175e-06, |
|
"loss": 0.7539, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.221956552036992e-06, |
|
"loss": 0.7422, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.042397785550405e-06, |
|
"loss": 0.7461, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.864430391348332e-06, |
|
"loss": 0.7578, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 8.688061284200266e-06, |
|
"loss": 0.7383, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 8.513297316775625e-06, |
|
"loss": 0.7539, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 8.34014527937756e-06, |
|
"loss": 0.75, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 8.168611899679013e-06, |
|
"loss": 0.7383, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 7.998703842461431e-06, |
|
"loss": 0.7695, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 7.830427709355725e-06, |
|
"loss": 0.7383, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.663790038585793e-06, |
|
"loss": 0.7578, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.498797304714544e-06, |
|
"loss": 0.7539, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.33545591839222e-06, |
|
"loss": 0.7227, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.173772226107434e-06, |
|
"loss": 0.75, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.013752509940485e-06, |
|
"loss": 0.7617, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.855402987319348e-06, |
|
"loss": 0.7344, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.698729810778065e-06, |
|
"loss": 0.7539, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.54373906771768e-06, |
|
"loss": 0.7656, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.390436780169734e-06, |
|
"loss": 0.7344, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.238828904562316e-06, |
|
"loss": 0.7305, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 6.088921331488568e-06, |
|
"loss": 0.7617, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.94071988547788e-06, |
|
"loss": 0.7422, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.794230324769517e-06, |
|
"loss": 0.7383, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.649458341088915e-06, |
|
"loss": 0.7617, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.506409559426573e-06, |
|
"loss": 0.7461, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 5.365089537819434e-06, |
|
"loss": 0.75, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 5.2255037671349535e-06, |
|
"loss": 0.7266, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 5.087657670857798e-06, |
|
"loss": 0.7578, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.951556604879048e-06, |
|
"loss": 0.75, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.8172058572881765e-06, |
|
"loss": 0.7461, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.684610648167503e-06, |
|
"loss": 0.7578, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.5537761293894535e-06, |
|
"loss": 0.7578, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.424707384416344e-06, |
|
"loss": 0.7383, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.29740942810285e-06, |
|
"loss": 0.7383, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.1718872065011904e-06, |
|
"loss": 0.7422, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 4.048145596668967e-06, |
|
"loss": 0.7305, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.9261894064796135e-06, |
|
"loss": 0.7617, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.8060233744356633e-06, |
|
"loss": 0.7695, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.687652169484568e-06, |
|
"loss": 0.7383, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.5710803908373224e-06, |
|
"loss": 0.75, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.4563125677897932e-06, |
|
"loss": 0.7578, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 3.343353159546675e-06, |
|
"loss": 0.7383, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 3.2322065550483007e-06, |
|
"loss": 0.7461, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 3.1228770728000455e-06, |
|
"loss": 0.7461, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 3.0153689607045845e-06, |
|
"loss": 0.7539, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.9096863958968268e-06, |
|
"loss": 0.7422, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.8058334845816213e-06, |
|
"loss": 0.7461, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.7038142618741992e-06, |
|
"loss": 0.7617, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.603632691643415e-06, |
|
"loss": 0.7422, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.5052926663577e-06, |
|
"loss": 0.7656, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.408798006933882e-06, |
|
"loss": 0.7461, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.314152462588659e-06, |
|
"loss": 0.7422, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.221359710692961e-06, |
|
"loss": 0.75, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.1304233566290964e-06, |
|
"loss": 0.7578, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.041346933650612e-06, |
|
"loss": 0.7539, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.9541339027450256e-06, |
|
"loss": 0.75, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.8687876524993987e-06, |
|
"loss": 0.7656, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.785311498968617e-06, |
|
"loss": 0.7578, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.70370868554659e-06, |
|
"loss": 0.7305, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.6239823828401945e-06, |
|
"loss": 0.7266, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.5461356885461075e-06, |
|
"loss": 0.7617, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.4701716273304521e-06, |
|
"loss": 0.7383, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.3960931507112752e-06, |
|
"loss": 0.7617, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.3239031369438326e-06, |
|
"loss": 0.75, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.2536043909088191e-06, |
|
"loss": 0.7539, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.1851996440033319e-06, |
|
"loss": 0.7578, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.118691554034773e-06, |
|
"loss": 0.75, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.0540827051175818e-06, |
|
"loss": 0.75, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 9.913756075728087e-07, |
|
"loss": 0.7695, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 9.305726978306173e-07, |
|
"loss": 0.75, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 8.716763383355864e-07, |
|
"loss": 0.7656, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 8.146888174549339e-07, |
|
"loss": 0.7773, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 7.596123493895991e-07, |
|
"loss": 0.7695, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 7.064490740882057e-07, |
|
"loss": 0.7461, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 6.552010571639456e-07, |
|
"loss": 0.7539, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 6.058702898142643e-07, |
|
"loss": 0.7656, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 5.584586887435739e-07, |
|
"loss": 0.7539, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 5.129680960887007e-07, |
|
"loss": 0.7461, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 4.6940027934735954e-07, |
|
"loss": 0.7656, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 4.277569313094809e-07, |
|
"loss": 0.7656, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 3.8803966999139684e-07, |
|
"loss": 0.7734, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 3.50250038573019e-07, |
|
"loss": 0.7461, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 3.143895053378698e-07, |
|
"loss": 0.7539, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.8045946361601183e-07, |
|
"loss": 0.7383, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.4846123172992954e-07, |
|
"loss": 0.7422, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.1839605294330933e-07, |
|
"loss": 0.7578, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.9026509541272275e-07, |
|
"loss": 0.7422, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.640694521422459e-07, |
|
"loss": 0.7422, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.3981014094099353e-07, |
|
"loss": 0.7461, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.1748810438355628e-07, |
|
"loss": 0.75, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 9.710420977340762e-08, |
|
"loss": 0.7461, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 7.865924910916977e-08, |
|
"loss": 0.7539, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 6.215393905388278e-08, |
|
"loss": 0.7578, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 4.7588920907110094e-08, |
|
"loss": 0.7422, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 3.496476058006959e-08, |
|
"loss": 0.7422, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.4281948573617874e-08, |
|
"loss": 0.7578, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.5540899959187727e-08, |
|
"loss": 0.7539, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 8.741954362678772e-09, |
|
"loss": 0.7656, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 3.885375951256931e-09, |
|
"loss": 0.75, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.713534230904041e-10, |
|
"loss": 0.75, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 0.0, |
|
"loss": 0.7578, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 520, |
|
"total_flos": 2.976872756622131e+16, |
|
"train_loss": 0.17339993990384617, |
|
"train_runtime": 2924.9681, |
|
"train_samples_per_second": 728.716, |
|
"train_steps_per_second": 0.178 |
|
} |
|
], |
|
"logging_steps": 1.0, |
|
"max_steps": 520, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 200, |
|
"total_flos": 2.976872756622131e+16, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|