diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,65667 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.999680034129693, + "eval_steps": 500, + "global_step": 9375, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003199658703071672, + "grad_norm": 7.29244327545166, + "learning_rate": 1.0660980810234543e-08, + "loss": 1.0554, + "step": 1 + }, + { + "epoch": 0.0006399317406143345, + "grad_norm": 6.493714332580566, + "learning_rate": 2.1321961620469085e-08, + "loss": 1.0764, + "step": 2 + }, + { + "epoch": 0.0009598976109215017, + "grad_norm": 6.872972011566162, + "learning_rate": 3.1982942430703625e-08, + "loss": 1.053, + "step": 3 + }, + { + "epoch": 0.001279863481228669, + "grad_norm": 7.674505233764648, + "learning_rate": 4.264392324093817e-08, + "loss": 1.1434, + "step": 4 + }, + { + "epoch": 0.0015998293515358361, + "grad_norm": 7.034660816192627, + "learning_rate": 5.330490405117271e-08, + "loss": 1.0999, + "step": 5 + }, + { + "epoch": 0.0019197952218430034, + "grad_norm": 6.8419647216796875, + "learning_rate": 6.396588486140725e-08, + "loss": 1.0741, + "step": 6 + }, + { + "epoch": 0.0022397610921501706, + "grad_norm": 7.366310119628906, + "learning_rate": 7.462686567164179e-08, + "loss": 1.1859, + "step": 7 + }, + { + "epoch": 0.002559726962457338, + "grad_norm": 7.265549659729004, + "learning_rate": 8.528784648187634e-08, + "loss": 1.0836, + "step": 8 + }, + { + "epoch": 0.002879692832764505, + "grad_norm": 7.0671586990356445, + "learning_rate": 9.59488272921109e-08, + "loss": 1.0631, + "step": 9 + }, + { + "epoch": 0.0031996587030716723, + "grad_norm": 6.615237712860107, + "learning_rate": 1.0660980810234542e-07, + "loss": 0.9951, + "step": 10 + }, + { + "epoch": 0.0035196245733788395, + "grad_norm": 7.323988914489746, + "learning_rate": 1.1727078891257997e-07, + "loss": 1.1467, + "step": 11 + }, + { + "epoch": 0.0038395904436860067, + "grad_norm": 7.808572292327881, + "learning_rate": 1.279317697228145e-07, + "loss": 1.1524, + "step": 12 + }, + { + "epoch": 0.004159556313993174, + "grad_norm": 7.407500267028809, + "learning_rate": 1.3859275053304905e-07, + "loss": 1.0871, + "step": 13 + }, + { + "epoch": 0.004479522184300341, + "grad_norm": 7.308926105499268, + "learning_rate": 1.4925373134328358e-07, + "loss": 1.1039, + "step": 14 + }, + { + "epoch": 0.004799488054607509, + "grad_norm": 7.3174004554748535, + "learning_rate": 1.5991471215351813e-07, + "loss": 1.145, + "step": 15 + }, + { + "epoch": 0.005119453924914676, + "grad_norm": 7.225617408752441, + "learning_rate": 1.7057569296375268e-07, + "loss": 1.1349, + "step": 16 + }, + { + "epoch": 0.005439419795221843, + "grad_norm": 6.9253668785095215, + "learning_rate": 1.812366737739872e-07, + "loss": 1.0338, + "step": 17 + }, + { + "epoch": 0.00575938566552901, + "grad_norm": 7.221156597137451, + "learning_rate": 1.918976545842218e-07, + "loss": 1.1135, + "step": 18 + }, + { + "epoch": 0.006079351535836178, + "grad_norm": 6.97838020324707, + "learning_rate": 2.0255863539445632e-07, + "loss": 1.054, + "step": 19 + }, + { + "epoch": 0.0063993174061433445, + "grad_norm": 6.8710455894470215, + "learning_rate": 2.1321961620469084e-07, + "loss": 1.0712, + "step": 20 + }, + { + "epoch": 0.006719283276450512, + "grad_norm": 7.243351936340332, + "learning_rate": 2.2388059701492537e-07, + "loss": 1.0826, + "step": 21 + }, + { + "epoch": 0.007039249146757679, + "grad_norm": 6.599764347076416, + "learning_rate": 2.3454157782515995e-07, + "loss": 1.1181, + "step": 22 + }, + { + "epoch": 0.007359215017064847, + "grad_norm": 6.794438362121582, + "learning_rate": 2.4520255863539447e-07, + "loss": 1.0853, + "step": 23 + }, + { + "epoch": 0.007679180887372013, + "grad_norm": 6.50425386428833, + "learning_rate": 2.55863539445629e-07, + "loss": 1.0862, + "step": 24 + }, + { + "epoch": 0.00799914675767918, + "grad_norm": 6.425209045410156, + "learning_rate": 2.665245202558635e-07, + "loss": 1.0684, + "step": 25 + }, + { + "epoch": 0.008319112627986349, + "grad_norm": 6.9282708168029785, + "learning_rate": 2.771855010660981e-07, + "loss": 1.1113, + "step": 26 + }, + { + "epoch": 0.008639078498293516, + "grad_norm": 6.989166259765625, + "learning_rate": 2.8784648187633263e-07, + "loss": 1.0686, + "step": 27 + }, + { + "epoch": 0.008959044368600682, + "grad_norm": 7.104398727416992, + "learning_rate": 2.9850746268656716e-07, + "loss": 1.101, + "step": 28 + }, + { + "epoch": 0.009279010238907849, + "grad_norm": 6.411932468414307, + "learning_rate": 3.0916844349680174e-07, + "loss": 1.076, + "step": 29 + }, + { + "epoch": 0.009598976109215018, + "grad_norm": 5.924315452575684, + "learning_rate": 3.1982942430703626e-07, + "loss": 1.0632, + "step": 30 + }, + { + "epoch": 0.009918941979522184, + "grad_norm": 6.04094123840332, + "learning_rate": 3.3049040511727084e-07, + "loss": 1.0843, + "step": 31 + }, + { + "epoch": 0.010238907849829351, + "grad_norm": 5.795870304107666, + "learning_rate": 3.4115138592750537e-07, + "loss": 1.0012, + "step": 32 + }, + { + "epoch": 0.010558873720136518, + "grad_norm": 5.819528579711914, + "learning_rate": 3.518123667377399e-07, + "loss": 1.0604, + "step": 33 + }, + { + "epoch": 0.010878839590443687, + "grad_norm": 5.332779884338379, + "learning_rate": 3.624733475479744e-07, + "loss": 1.0586, + "step": 34 + }, + { + "epoch": 0.011198805460750853, + "grad_norm": 5.464365005493164, + "learning_rate": 3.7313432835820895e-07, + "loss": 1.1297, + "step": 35 + }, + { + "epoch": 0.01151877133105802, + "grad_norm": 5.652714729309082, + "learning_rate": 3.837953091684436e-07, + "loss": 1.0346, + "step": 36 + }, + { + "epoch": 0.011838737201365187, + "grad_norm": 5.79964542388916, + "learning_rate": 3.944562899786781e-07, + "loss": 1.0312, + "step": 37 + }, + { + "epoch": 0.012158703071672355, + "grad_norm": 5.450277805328369, + "learning_rate": 4.0511727078891263e-07, + "loss": 1.0336, + "step": 38 + }, + { + "epoch": 0.012478668941979522, + "grad_norm": 4.844827175140381, + "learning_rate": 4.1577825159914716e-07, + "loss": 0.9825, + "step": 39 + }, + { + "epoch": 0.012798634812286689, + "grad_norm": 3.600152015686035, + "learning_rate": 4.264392324093817e-07, + "loss": 0.9513, + "step": 40 + }, + { + "epoch": 0.013118600682593856, + "grad_norm": 3.4587929248809814, + "learning_rate": 4.371002132196162e-07, + "loss": 0.9135, + "step": 41 + }, + { + "epoch": 0.013438566552901024, + "grad_norm": 2.950002431869507, + "learning_rate": 4.4776119402985074e-07, + "loss": 0.9524, + "step": 42 + }, + { + "epoch": 0.013758532423208191, + "grad_norm": 3.295619249343872, + "learning_rate": 4.5842217484008537e-07, + "loss": 0.9572, + "step": 43 + }, + { + "epoch": 0.014078498293515358, + "grad_norm": 2.7801365852355957, + "learning_rate": 4.690831556503199e-07, + "loss": 0.9389, + "step": 44 + }, + { + "epoch": 0.014398464163822526, + "grad_norm": 2.845637321472168, + "learning_rate": 4.797441364605544e-07, + "loss": 0.9185, + "step": 45 + }, + { + "epoch": 0.014718430034129693, + "grad_norm": 2.6487975120544434, + "learning_rate": 4.904051172707889e-07, + "loss": 0.9893, + "step": 46 + }, + { + "epoch": 0.01503839590443686, + "grad_norm": 2.6472787857055664, + "learning_rate": 5.010660980810235e-07, + "loss": 0.9619, + "step": 47 + }, + { + "epoch": 0.015358361774744027, + "grad_norm": 2.818723440170288, + "learning_rate": 5.11727078891258e-07, + "loss": 0.9178, + "step": 48 + }, + { + "epoch": 0.015678327645051195, + "grad_norm": 2.410865306854248, + "learning_rate": 5.223880597014925e-07, + "loss": 0.9459, + "step": 49 + }, + { + "epoch": 0.01599829351535836, + "grad_norm": 2.5202646255493164, + "learning_rate": 5.33049040511727e-07, + "loss": 0.9107, + "step": 50 + }, + { + "epoch": 0.01631825938566553, + "grad_norm": 2.329850196838379, + "learning_rate": 5.437100213219617e-07, + "loss": 0.9128, + "step": 51 + }, + { + "epoch": 0.016638225255972697, + "grad_norm": 2.3990280628204346, + "learning_rate": 5.543710021321962e-07, + "loss": 0.9408, + "step": 52 + }, + { + "epoch": 0.016958191126279862, + "grad_norm": 2.338702440261841, + "learning_rate": 5.650319829424307e-07, + "loss": 0.9071, + "step": 53 + }, + { + "epoch": 0.01727815699658703, + "grad_norm": 2.126091957092285, + "learning_rate": 5.756929637526653e-07, + "loss": 0.9262, + "step": 54 + }, + { + "epoch": 0.0175981228668942, + "grad_norm": 2.0692224502563477, + "learning_rate": 5.863539445628998e-07, + "loss": 0.9058, + "step": 55 + }, + { + "epoch": 0.017918088737201365, + "grad_norm": 2.1477673053741455, + "learning_rate": 5.970149253731343e-07, + "loss": 0.8858, + "step": 56 + }, + { + "epoch": 0.018238054607508533, + "grad_norm": 2.15913462638855, + "learning_rate": 6.076759061833689e-07, + "loss": 0.8942, + "step": 57 + }, + { + "epoch": 0.018558020477815698, + "grad_norm": 2.243981122970581, + "learning_rate": 6.183368869936035e-07, + "loss": 0.8919, + "step": 58 + }, + { + "epoch": 0.018877986348122867, + "grad_norm": 2.0859246253967285, + "learning_rate": 6.28997867803838e-07, + "loss": 0.9206, + "step": 59 + }, + { + "epoch": 0.019197952218430035, + "grad_norm": 2.022251844406128, + "learning_rate": 6.396588486140725e-07, + "loss": 0.8063, + "step": 60 + }, + { + "epoch": 0.0195179180887372, + "grad_norm": 2.079070806503296, + "learning_rate": 6.50319829424307e-07, + "loss": 0.8856, + "step": 61 + }, + { + "epoch": 0.01983788395904437, + "grad_norm": 1.8525385856628418, + "learning_rate": 6.609808102345417e-07, + "loss": 0.8495, + "step": 62 + }, + { + "epoch": 0.020157849829351537, + "grad_norm": 1.8756672143936157, + "learning_rate": 6.716417910447762e-07, + "loss": 0.8234, + "step": 63 + }, + { + "epoch": 0.020477815699658702, + "grad_norm": 1.7776811122894287, + "learning_rate": 6.823027718550107e-07, + "loss": 0.899, + "step": 64 + }, + { + "epoch": 0.02079778156996587, + "grad_norm": 1.680071234703064, + "learning_rate": 6.929637526652453e-07, + "loss": 0.9274, + "step": 65 + }, + { + "epoch": 0.021117747440273036, + "grad_norm": 1.493758201599121, + "learning_rate": 7.036247334754798e-07, + "loss": 0.8671, + "step": 66 + }, + { + "epoch": 0.021437713310580205, + "grad_norm": 1.4272754192352295, + "learning_rate": 7.142857142857143e-07, + "loss": 0.8085, + "step": 67 + }, + { + "epoch": 0.021757679180887373, + "grad_norm": 1.404809594154358, + "learning_rate": 7.249466950959488e-07, + "loss": 0.8463, + "step": 68 + }, + { + "epoch": 0.022077645051194538, + "grad_norm": 1.2687548398971558, + "learning_rate": 7.356076759061834e-07, + "loss": 0.8656, + "step": 69 + }, + { + "epoch": 0.022397610921501707, + "grad_norm": 1.2395495176315308, + "learning_rate": 7.462686567164179e-07, + "loss": 0.8554, + "step": 70 + }, + { + "epoch": 0.022717576791808875, + "grad_norm": 1.2378782033920288, + "learning_rate": 7.569296375266526e-07, + "loss": 0.848, + "step": 71 + }, + { + "epoch": 0.02303754266211604, + "grad_norm": 1.4011038541793823, + "learning_rate": 7.675906183368872e-07, + "loss": 0.826, + "step": 72 + }, + { + "epoch": 0.02335750853242321, + "grad_norm": 1.261948585510254, + "learning_rate": 7.782515991471217e-07, + "loss": 0.842, + "step": 73 + }, + { + "epoch": 0.023677474402730374, + "grad_norm": 1.0836261510849, + "learning_rate": 7.889125799573562e-07, + "loss": 0.8043, + "step": 74 + }, + { + "epoch": 0.023997440273037542, + "grad_norm": 1.2844727039337158, + "learning_rate": 7.995735607675907e-07, + "loss": 0.8443, + "step": 75 + }, + { + "epoch": 0.02431740614334471, + "grad_norm": 1.0323970317840576, + "learning_rate": 8.102345415778253e-07, + "loss": 0.7497, + "step": 76 + }, + { + "epoch": 0.024637372013651876, + "grad_norm": 1.0269299745559692, + "learning_rate": 8.208955223880598e-07, + "loss": 0.7616, + "step": 77 + }, + { + "epoch": 0.024957337883959044, + "grad_norm": 1.0544461011886597, + "learning_rate": 8.315565031982943e-07, + "loss": 0.7697, + "step": 78 + }, + { + "epoch": 0.025277303754266213, + "grad_norm": 1.0795966386795044, + "learning_rate": 8.422174840085288e-07, + "loss": 0.768, + "step": 79 + }, + { + "epoch": 0.025597269624573378, + "grad_norm": 0.9690151214599609, + "learning_rate": 8.528784648187634e-07, + "loss": 0.7908, + "step": 80 + }, + { + "epoch": 0.025917235494880547, + "grad_norm": 1.0053969621658325, + "learning_rate": 8.635394456289979e-07, + "loss": 0.8469, + "step": 81 + }, + { + "epoch": 0.02623720136518771, + "grad_norm": 0.928873598575592, + "learning_rate": 8.742004264392324e-07, + "loss": 0.76, + "step": 82 + }, + { + "epoch": 0.02655716723549488, + "grad_norm": 0.9888681769371033, + "learning_rate": 8.848614072494669e-07, + "loss": 0.7964, + "step": 83 + }, + { + "epoch": 0.02687713310580205, + "grad_norm": 0.8391404747962952, + "learning_rate": 8.955223880597015e-07, + "loss": 0.6921, + "step": 84 + }, + { + "epoch": 0.027197098976109214, + "grad_norm": 0.9005293846130371, + "learning_rate": 9.06183368869936e-07, + "loss": 0.7879, + "step": 85 + }, + { + "epoch": 0.027517064846416382, + "grad_norm": 0.8991631269454956, + "learning_rate": 9.168443496801707e-07, + "loss": 0.7402, + "step": 86 + }, + { + "epoch": 0.02783703071672355, + "grad_norm": 0.8436546921730042, + "learning_rate": 9.275053304904053e-07, + "loss": 0.7411, + "step": 87 + }, + { + "epoch": 0.028156996587030716, + "grad_norm": 0.8756288290023804, + "learning_rate": 9.381663113006398e-07, + "loss": 0.7711, + "step": 88 + }, + { + "epoch": 0.028476962457337884, + "grad_norm": 0.7887997031211853, + "learning_rate": 9.488272921108743e-07, + "loss": 0.7588, + "step": 89 + }, + { + "epoch": 0.028796928327645053, + "grad_norm": 0.8141859769821167, + "learning_rate": 9.594882729211088e-07, + "loss": 0.7743, + "step": 90 + }, + { + "epoch": 0.029116894197952218, + "grad_norm": 0.8312386274337769, + "learning_rate": 9.701492537313434e-07, + "loss": 0.7418, + "step": 91 + }, + { + "epoch": 0.029436860068259386, + "grad_norm": 0.8288974165916443, + "learning_rate": 9.808102345415779e-07, + "loss": 0.7381, + "step": 92 + }, + { + "epoch": 0.02975682593856655, + "grad_norm": 0.8249732255935669, + "learning_rate": 9.914712153518124e-07, + "loss": 0.7541, + "step": 93 + }, + { + "epoch": 0.03007679180887372, + "grad_norm": 0.7712000608444214, + "learning_rate": 1.002132196162047e-06, + "loss": 0.7684, + "step": 94 + }, + { + "epoch": 0.03039675767918089, + "grad_norm": 0.7715640664100647, + "learning_rate": 1.0127931769722815e-06, + "loss": 0.8098, + "step": 95 + }, + { + "epoch": 0.030716723549488054, + "grad_norm": 0.7638954520225525, + "learning_rate": 1.023454157782516e-06, + "loss": 0.7735, + "step": 96 + }, + { + "epoch": 0.031036689419795222, + "grad_norm": 0.7430364489555359, + "learning_rate": 1.0341151385927505e-06, + "loss": 0.745, + "step": 97 + }, + { + "epoch": 0.03135665529010239, + "grad_norm": 0.7673352956771851, + "learning_rate": 1.044776119402985e-06, + "loss": 0.756, + "step": 98 + }, + { + "epoch": 0.031676621160409556, + "grad_norm": 0.8949801921844482, + "learning_rate": 1.0554371002132196e-06, + "loss": 0.7666, + "step": 99 + }, + { + "epoch": 0.03199658703071672, + "grad_norm": 0.8077754974365234, + "learning_rate": 1.066098081023454e-06, + "loss": 0.7326, + "step": 100 + }, + { + "epoch": 0.03231655290102389, + "grad_norm": 0.7525066137313843, + "learning_rate": 1.0767590618336886e-06, + "loss": 0.7114, + "step": 101 + }, + { + "epoch": 0.03263651877133106, + "grad_norm": 0.7346336841583252, + "learning_rate": 1.0874200426439234e-06, + "loss": 0.7588, + "step": 102 + }, + { + "epoch": 0.03295648464163822, + "grad_norm": 0.8265507221221924, + "learning_rate": 1.0980810234541579e-06, + "loss": 0.7152, + "step": 103 + }, + { + "epoch": 0.033276450511945395, + "grad_norm": 0.7094560861587524, + "learning_rate": 1.1087420042643924e-06, + "loss": 0.715, + "step": 104 + }, + { + "epoch": 0.03359641638225256, + "grad_norm": 0.7683343887329102, + "learning_rate": 1.119402985074627e-06, + "loss": 0.7334, + "step": 105 + }, + { + "epoch": 0.033916382252559725, + "grad_norm": 0.7544600963592529, + "learning_rate": 1.1300639658848615e-06, + "loss": 0.7015, + "step": 106 + }, + { + "epoch": 0.0342363481228669, + "grad_norm": 0.7648451924324036, + "learning_rate": 1.140724946695096e-06, + "loss": 0.7146, + "step": 107 + }, + { + "epoch": 0.03455631399317406, + "grad_norm": 0.7159047722816467, + "learning_rate": 1.1513859275053305e-06, + "loss": 0.6871, + "step": 108 + }, + { + "epoch": 0.03487627986348123, + "grad_norm": 0.6753017902374268, + "learning_rate": 1.162046908315565e-06, + "loss": 0.6809, + "step": 109 + }, + { + "epoch": 0.0351962457337884, + "grad_norm": 0.6882673501968384, + "learning_rate": 1.1727078891257996e-06, + "loss": 0.7309, + "step": 110 + }, + { + "epoch": 0.035516211604095564, + "grad_norm": 0.6664094924926758, + "learning_rate": 1.183368869936034e-06, + "loss": 0.7106, + "step": 111 + }, + { + "epoch": 0.03583617747440273, + "grad_norm": 0.7424576878547668, + "learning_rate": 1.1940298507462686e-06, + "loss": 0.7242, + "step": 112 + }, + { + "epoch": 0.036156143344709894, + "grad_norm": 0.7396127581596375, + "learning_rate": 1.2046908315565034e-06, + "loss": 0.719, + "step": 113 + }, + { + "epoch": 0.036476109215017066, + "grad_norm": 0.6853981018066406, + "learning_rate": 1.2153518123667379e-06, + "loss": 0.6913, + "step": 114 + }, + { + "epoch": 0.03679607508532423, + "grad_norm": 0.7118975520133972, + "learning_rate": 1.2260127931769724e-06, + "loss": 0.7491, + "step": 115 + }, + { + "epoch": 0.037116040955631396, + "grad_norm": 0.6694145798683167, + "learning_rate": 1.236673773987207e-06, + "loss": 0.6652, + "step": 116 + }, + { + "epoch": 0.03743600682593857, + "grad_norm": 0.6602712273597717, + "learning_rate": 1.2473347547974415e-06, + "loss": 0.7452, + "step": 117 + }, + { + "epoch": 0.03775597269624573, + "grad_norm": 0.6255103349685669, + "learning_rate": 1.257995735607676e-06, + "loss": 0.7292, + "step": 118 + }, + { + "epoch": 0.0380759385665529, + "grad_norm": 0.738187313079834, + "learning_rate": 1.2686567164179105e-06, + "loss": 0.7191, + "step": 119 + }, + { + "epoch": 0.03839590443686007, + "grad_norm": 0.6636806130409241, + "learning_rate": 1.279317697228145e-06, + "loss": 0.6932, + "step": 120 + }, + { + "epoch": 0.038715870307167236, + "grad_norm": 0.6825982332229614, + "learning_rate": 1.2899786780383796e-06, + "loss": 0.7343, + "step": 121 + }, + { + "epoch": 0.0390358361774744, + "grad_norm": 0.740442156791687, + "learning_rate": 1.300639658848614e-06, + "loss": 0.7303, + "step": 122 + }, + { + "epoch": 0.03935580204778157, + "grad_norm": 0.5918875336647034, + "learning_rate": 1.3113006396588488e-06, + "loss": 0.7187, + "step": 123 + }, + { + "epoch": 0.03967576791808874, + "grad_norm": 0.7803345918655396, + "learning_rate": 1.3219616204690834e-06, + "loss": 0.8031, + "step": 124 + }, + { + "epoch": 0.0399957337883959, + "grad_norm": 0.6779199838638306, + "learning_rate": 1.3326226012793179e-06, + "loss": 0.7096, + "step": 125 + }, + { + "epoch": 0.040315699658703075, + "grad_norm": 0.6925889253616333, + "learning_rate": 1.3432835820895524e-06, + "loss": 0.6824, + "step": 126 + }, + { + "epoch": 0.04063566552901024, + "grad_norm": 0.7514821290969849, + "learning_rate": 1.353944562899787e-06, + "loss": 0.7371, + "step": 127 + }, + { + "epoch": 0.040955631399317405, + "grad_norm": 0.6688573360443115, + "learning_rate": 1.3646055437100215e-06, + "loss": 0.7117, + "step": 128 + }, + { + "epoch": 0.04127559726962458, + "grad_norm": 0.6557217240333557, + "learning_rate": 1.375266524520256e-06, + "loss": 0.6998, + "step": 129 + }, + { + "epoch": 0.04159556313993174, + "grad_norm": 0.7434080243110657, + "learning_rate": 1.3859275053304905e-06, + "loss": 0.7321, + "step": 130 + }, + { + "epoch": 0.04191552901023891, + "grad_norm": 0.669056236743927, + "learning_rate": 1.396588486140725e-06, + "loss": 0.7098, + "step": 131 + }, + { + "epoch": 0.04223549488054607, + "grad_norm": 0.5930835604667664, + "learning_rate": 1.4072494669509596e-06, + "loss": 0.6943, + "step": 132 + }, + { + "epoch": 0.042555460750853244, + "grad_norm": 0.6236996054649353, + "learning_rate": 1.417910447761194e-06, + "loss": 0.6928, + "step": 133 + }, + { + "epoch": 0.04287542662116041, + "grad_norm": 0.6656635403633118, + "learning_rate": 1.4285714285714286e-06, + "loss": 0.6943, + "step": 134 + }, + { + "epoch": 0.043195392491467574, + "grad_norm": 0.7729142308235168, + "learning_rate": 1.4392324093816632e-06, + "loss": 0.7283, + "step": 135 + }, + { + "epoch": 0.043515358361774746, + "grad_norm": 0.6872455477714539, + "learning_rate": 1.4498933901918977e-06, + "loss": 0.706, + "step": 136 + }, + { + "epoch": 0.04383532423208191, + "grad_norm": 0.6097860336303711, + "learning_rate": 1.4605543710021322e-06, + "loss": 0.6711, + "step": 137 + }, + { + "epoch": 0.044155290102389076, + "grad_norm": 0.7215981483459473, + "learning_rate": 1.4712153518123667e-06, + "loss": 0.75, + "step": 138 + }, + { + "epoch": 0.04447525597269625, + "grad_norm": 0.6652256846427917, + "learning_rate": 1.4818763326226013e-06, + "loss": 0.7266, + "step": 139 + }, + { + "epoch": 0.04479522184300341, + "grad_norm": 0.7119171619415283, + "learning_rate": 1.4925373134328358e-06, + "loss": 0.6318, + "step": 140 + }, + { + "epoch": 0.04511518771331058, + "grad_norm": 0.7838914394378662, + "learning_rate": 1.5031982942430705e-06, + "loss": 0.6842, + "step": 141 + }, + { + "epoch": 0.04543515358361775, + "grad_norm": 0.7664074897766113, + "learning_rate": 1.5138592750533053e-06, + "loss": 0.7572, + "step": 142 + }, + { + "epoch": 0.045755119453924915, + "grad_norm": 0.6630170941352844, + "learning_rate": 1.5245202558635398e-06, + "loss": 0.6776, + "step": 143 + }, + { + "epoch": 0.04607508532423208, + "grad_norm": 0.6675658822059631, + "learning_rate": 1.5351812366737743e-06, + "loss": 0.7035, + "step": 144 + }, + { + "epoch": 0.04639505119453925, + "grad_norm": 0.6449645161628723, + "learning_rate": 1.5458422174840088e-06, + "loss": 0.7138, + "step": 145 + }, + { + "epoch": 0.04671501706484642, + "grad_norm": 0.6699844002723694, + "learning_rate": 1.5565031982942434e-06, + "loss": 0.6748, + "step": 146 + }, + { + "epoch": 0.04703498293515358, + "grad_norm": 0.6595117449760437, + "learning_rate": 1.5671641791044779e-06, + "loss": 0.6752, + "step": 147 + }, + { + "epoch": 0.04735494880546075, + "grad_norm": 0.6826941967010498, + "learning_rate": 1.5778251599147124e-06, + "loss": 0.7394, + "step": 148 + }, + { + "epoch": 0.04767491467576792, + "grad_norm": 0.7005380392074585, + "learning_rate": 1.588486140724947e-06, + "loss": 0.6968, + "step": 149 + }, + { + "epoch": 0.047994880546075085, + "grad_norm": 0.608120322227478, + "learning_rate": 1.5991471215351815e-06, + "loss": 0.6177, + "step": 150 + }, + { + "epoch": 0.04831484641638225, + "grad_norm": 0.7013968825340271, + "learning_rate": 1.609808102345416e-06, + "loss": 0.6854, + "step": 151 + }, + { + "epoch": 0.04863481228668942, + "grad_norm": 0.6775135397911072, + "learning_rate": 1.6204690831556505e-06, + "loss": 0.7177, + "step": 152 + }, + { + "epoch": 0.04895477815699659, + "grad_norm": 0.7195072174072266, + "learning_rate": 1.631130063965885e-06, + "loss": 0.6894, + "step": 153 + }, + { + "epoch": 0.04927474402730375, + "grad_norm": 0.7061877846717834, + "learning_rate": 1.6417910447761196e-06, + "loss": 0.7465, + "step": 154 + }, + { + "epoch": 0.049594709897610924, + "grad_norm": 0.6255117058753967, + "learning_rate": 1.652452025586354e-06, + "loss": 0.6758, + "step": 155 + }, + { + "epoch": 0.04991467576791809, + "grad_norm": 0.7507020235061646, + "learning_rate": 1.6631130063965886e-06, + "loss": 0.7578, + "step": 156 + }, + { + "epoch": 0.050234641638225254, + "grad_norm": 0.6676733493804932, + "learning_rate": 1.6737739872068232e-06, + "loss": 0.6807, + "step": 157 + }, + { + "epoch": 0.050554607508532426, + "grad_norm": 0.6160469651222229, + "learning_rate": 1.6844349680170577e-06, + "loss": 0.662, + "step": 158 + }, + { + "epoch": 0.05087457337883959, + "grad_norm": 0.6552683711051941, + "learning_rate": 1.6950959488272922e-06, + "loss": 0.6405, + "step": 159 + }, + { + "epoch": 0.051194539249146756, + "grad_norm": 0.6595558524131775, + "learning_rate": 1.7057569296375267e-06, + "loss": 0.6813, + "step": 160 + }, + { + "epoch": 0.05151450511945393, + "grad_norm": 0.6022697687149048, + "learning_rate": 1.7164179104477613e-06, + "loss": 0.6718, + "step": 161 + }, + { + "epoch": 0.05183447098976109, + "grad_norm": 0.631741464138031, + "learning_rate": 1.7270788912579958e-06, + "loss": 0.6571, + "step": 162 + }, + { + "epoch": 0.05215443686006826, + "grad_norm": 0.712632417678833, + "learning_rate": 1.7377398720682303e-06, + "loss": 0.6367, + "step": 163 + }, + { + "epoch": 0.05247440273037542, + "grad_norm": 0.691933274269104, + "learning_rate": 1.7484008528784648e-06, + "loss": 0.6726, + "step": 164 + }, + { + "epoch": 0.052794368600682595, + "grad_norm": 0.648144543170929, + "learning_rate": 1.7590618336886994e-06, + "loss": 0.7117, + "step": 165 + }, + { + "epoch": 0.05311433447098976, + "grad_norm": 0.7375291585922241, + "learning_rate": 1.7697228144989339e-06, + "loss": 0.7251, + "step": 166 + }, + { + "epoch": 0.053434300341296925, + "grad_norm": 0.6844941973686218, + "learning_rate": 1.7803837953091684e-06, + "loss": 0.6893, + "step": 167 + }, + { + "epoch": 0.0537542662116041, + "grad_norm": 0.7131636142730713, + "learning_rate": 1.791044776119403e-06, + "loss": 0.7042, + "step": 168 + }, + { + "epoch": 0.05407423208191126, + "grad_norm": 0.7594255208969116, + "learning_rate": 1.8017057569296375e-06, + "loss": 0.7, + "step": 169 + }, + { + "epoch": 0.05439419795221843, + "grad_norm": 0.7435464859008789, + "learning_rate": 1.812366737739872e-06, + "loss": 0.697, + "step": 170 + }, + { + "epoch": 0.0547141638225256, + "grad_norm": 0.7214502692222595, + "learning_rate": 1.8230277185501067e-06, + "loss": 0.7183, + "step": 171 + }, + { + "epoch": 0.055034129692832764, + "grad_norm": 0.6937075257301331, + "learning_rate": 1.8336886993603415e-06, + "loss": 0.6498, + "step": 172 + }, + { + "epoch": 0.05535409556313993, + "grad_norm": 0.6126171350479126, + "learning_rate": 1.844349680170576e-06, + "loss": 0.6458, + "step": 173 + }, + { + "epoch": 0.0556740614334471, + "grad_norm": 0.697283148765564, + "learning_rate": 1.8550106609808105e-06, + "loss": 0.6651, + "step": 174 + }, + { + "epoch": 0.05599402730375427, + "grad_norm": 0.6571756601333618, + "learning_rate": 1.865671641791045e-06, + "loss": 0.6113, + "step": 175 + }, + { + "epoch": 0.05631399317406143, + "grad_norm": 0.658406674861908, + "learning_rate": 1.8763326226012796e-06, + "loss": 0.6706, + "step": 176 + }, + { + "epoch": 0.056633959044368604, + "grad_norm": 0.7434971332550049, + "learning_rate": 1.886993603411514e-06, + "loss": 0.6733, + "step": 177 + }, + { + "epoch": 0.05695392491467577, + "grad_norm": 0.6445094347000122, + "learning_rate": 1.8976545842217486e-06, + "loss": 0.6544, + "step": 178 + }, + { + "epoch": 0.057273890784982934, + "grad_norm": 0.6629904508590698, + "learning_rate": 1.908315565031983e-06, + "loss": 0.7276, + "step": 179 + }, + { + "epoch": 0.057593856655290106, + "grad_norm": 0.699326753616333, + "learning_rate": 1.9189765458422177e-06, + "loss": 0.6755, + "step": 180 + }, + { + "epoch": 0.05791382252559727, + "grad_norm": 0.8005274534225464, + "learning_rate": 1.929637526652452e-06, + "loss": 0.6551, + "step": 181 + }, + { + "epoch": 0.058233788395904436, + "grad_norm": 0.7194087505340576, + "learning_rate": 1.9402985074626867e-06, + "loss": 0.6459, + "step": 182 + }, + { + "epoch": 0.0585537542662116, + "grad_norm": 0.7291178703308105, + "learning_rate": 1.9509594882729213e-06, + "loss": 0.7131, + "step": 183 + }, + { + "epoch": 0.05887372013651877, + "grad_norm": 0.6386391520500183, + "learning_rate": 1.9616204690831558e-06, + "loss": 0.6826, + "step": 184 + }, + { + "epoch": 0.05919368600682594, + "grad_norm": 0.6472675800323486, + "learning_rate": 1.9722814498933903e-06, + "loss": 0.6299, + "step": 185 + }, + { + "epoch": 0.0595136518771331, + "grad_norm": 0.6684521436691284, + "learning_rate": 1.982942430703625e-06, + "loss": 0.6535, + "step": 186 + }, + { + "epoch": 0.059833617747440275, + "grad_norm": 0.6821104884147644, + "learning_rate": 1.9936034115138594e-06, + "loss": 0.6948, + "step": 187 + }, + { + "epoch": 0.06015358361774744, + "grad_norm": 0.6314494013786316, + "learning_rate": 2.004264392324094e-06, + "loss": 0.664, + "step": 188 + }, + { + "epoch": 0.060473549488054605, + "grad_norm": 0.7831369638442993, + "learning_rate": 2.0149253731343284e-06, + "loss": 0.7075, + "step": 189 + }, + { + "epoch": 0.06079351535836178, + "grad_norm": 0.6174335479736328, + "learning_rate": 2.025586353944563e-06, + "loss": 0.6666, + "step": 190 + }, + { + "epoch": 0.06111348122866894, + "grad_norm": 0.6839204430580139, + "learning_rate": 2.0362473347547975e-06, + "loss": 0.6791, + "step": 191 + }, + { + "epoch": 0.06143344709897611, + "grad_norm": 0.6346204280853271, + "learning_rate": 2.046908315565032e-06, + "loss": 0.6579, + "step": 192 + }, + { + "epoch": 0.06175341296928328, + "grad_norm": 0.7199333906173706, + "learning_rate": 2.0575692963752665e-06, + "loss": 0.6369, + "step": 193 + }, + { + "epoch": 0.062073378839590444, + "grad_norm": 0.6710346937179565, + "learning_rate": 2.068230277185501e-06, + "loss": 0.6483, + "step": 194 + }, + { + "epoch": 0.06239334470989761, + "grad_norm": 0.7106640338897705, + "learning_rate": 2.0788912579957356e-06, + "loss": 0.662, + "step": 195 + }, + { + "epoch": 0.06271331058020478, + "grad_norm": 0.8064489364624023, + "learning_rate": 2.08955223880597e-06, + "loss": 0.6897, + "step": 196 + }, + { + "epoch": 0.06303327645051195, + "grad_norm": 0.7295928597450256, + "learning_rate": 2.1002132196162046e-06, + "loss": 0.6644, + "step": 197 + }, + { + "epoch": 0.06335324232081911, + "grad_norm": 0.7068210244178772, + "learning_rate": 2.110874200426439e-06, + "loss": 0.6854, + "step": 198 + }, + { + "epoch": 0.06367320819112628, + "grad_norm": 0.7519869208335876, + "learning_rate": 2.1215351812366737e-06, + "loss": 0.6667, + "step": 199 + }, + { + "epoch": 0.06399317406143344, + "grad_norm": 0.7251495122909546, + "learning_rate": 2.132196162046908e-06, + "loss": 0.6764, + "step": 200 + }, + { + "epoch": 0.06431313993174062, + "grad_norm": 0.5924597382545471, + "learning_rate": 2.1428571428571427e-06, + "loss": 0.6447, + "step": 201 + }, + { + "epoch": 0.06463310580204779, + "grad_norm": 0.7425607442855835, + "learning_rate": 2.1535181236673773e-06, + "loss": 0.6867, + "step": 202 + }, + { + "epoch": 0.06495307167235495, + "grad_norm": 0.662427544593811, + "learning_rate": 2.1641791044776118e-06, + "loss": 0.6814, + "step": 203 + }, + { + "epoch": 0.06527303754266212, + "grad_norm": 0.7188402414321899, + "learning_rate": 2.1748400852878467e-06, + "loss": 0.7082, + "step": 204 + }, + { + "epoch": 0.06559300341296928, + "grad_norm": 0.6829525828361511, + "learning_rate": 2.1855010660980813e-06, + "loss": 0.6382, + "step": 205 + }, + { + "epoch": 0.06591296928327645, + "grad_norm": 0.713671088218689, + "learning_rate": 2.1961620469083158e-06, + "loss": 0.6556, + "step": 206 + }, + { + "epoch": 0.06623293515358361, + "grad_norm": 0.7159529328346252, + "learning_rate": 2.2068230277185503e-06, + "loss": 0.6594, + "step": 207 + }, + { + "epoch": 0.06655290102389079, + "grad_norm": 0.856481671333313, + "learning_rate": 2.217484008528785e-06, + "loss": 0.6609, + "step": 208 + }, + { + "epoch": 0.06687286689419795, + "grad_norm": 0.6801454424858093, + "learning_rate": 2.2281449893390194e-06, + "loss": 0.619, + "step": 209 + }, + { + "epoch": 0.06719283276450512, + "grad_norm": 0.7858098149299622, + "learning_rate": 2.238805970149254e-06, + "loss": 0.6505, + "step": 210 + }, + { + "epoch": 0.06751279863481228, + "grad_norm": 0.6824889183044434, + "learning_rate": 2.2494669509594884e-06, + "loss": 0.6682, + "step": 211 + }, + { + "epoch": 0.06783276450511945, + "grad_norm": 0.691604733467102, + "learning_rate": 2.260127931769723e-06, + "loss": 0.7058, + "step": 212 + }, + { + "epoch": 0.06815273037542662, + "grad_norm": 0.732110321521759, + "learning_rate": 2.2707889125799575e-06, + "loss": 0.6431, + "step": 213 + }, + { + "epoch": 0.0684726962457338, + "grad_norm": 0.6957806348800659, + "learning_rate": 2.281449893390192e-06, + "loss": 0.6827, + "step": 214 + }, + { + "epoch": 0.06879266211604096, + "grad_norm": 0.7264412641525269, + "learning_rate": 2.2921108742004265e-06, + "loss": 0.6411, + "step": 215 + }, + { + "epoch": 0.06911262798634812, + "grad_norm": 0.6580216884613037, + "learning_rate": 2.302771855010661e-06, + "loss": 0.6748, + "step": 216 + }, + { + "epoch": 0.06943259385665529, + "grad_norm": 0.7710157036781311, + "learning_rate": 2.3134328358208956e-06, + "loss": 0.6895, + "step": 217 + }, + { + "epoch": 0.06975255972696245, + "grad_norm": 0.769323468208313, + "learning_rate": 2.32409381663113e-06, + "loss": 0.6593, + "step": 218 + }, + { + "epoch": 0.07007252559726962, + "grad_norm": 0.6831722259521484, + "learning_rate": 2.3347547974413646e-06, + "loss": 0.6512, + "step": 219 + }, + { + "epoch": 0.0703924914675768, + "grad_norm": 0.7421032786369324, + "learning_rate": 2.345415778251599e-06, + "loss": 0.6667, + "step": 220 + }, + { + "epoch": 0.07071245733788396, + "grad_norm": 0.8579223155975342, + "learning_rate": 2.3560767590618337e-06, + "loss": 0.6856, + "step": 221 + }, + { + "epoch": 0.07103242320819113, + "grad_norm": 0.7443462610244751, + "learning_rate": 2.366737739872068e-06, + "loss": 0.7142, + "step": 222 + }, + { + "epoch": 0.0713523890784983, + "grad_norm": 0.6016785502433777, + "learning_rate": 2.3773987206823027e-06, + "loss": 0.6414, + "step": 223 + }, + { + "epoch": 0.07167235494880546, + "grad_norm": 0.6656649708747864, + "learning_rate": 2.3880597014925373e-06, + "loss": 0.6648, + "step": 224 + }, + { + "epoch": 0.07199232081911262, + "grad_norm": 0.6443098187446594, + "learning_rate": 2.398720682302772e-06, + "loss": 0.637, + "step": 225 + }, + { + "epoch": 0.07231228668941979, + "grad_norm": 0.6778454780578613, + "learning_rate": 2.4093816631130067e-06, + "loss": 0.6501, + "step": 226 + }, + { + "epoch": 0.07263225255972697, + "grad_norm": 0.6492361426353455, + "learning_rate": 2.4200426439232413e-06, + "loss": 0.7013, + "step": 227 + }, + { + "epoch": 0.07295221843003413, + "grad_norm": 0.8034964203834534, + "learning_rate": 2.4307036247334758e-06, + "loss": 0.7234, + "step": 228 + }, + { + "epoch": 0.0732721843003413, + "grad_norm": 0.6665212512016296, + "learning_rate": 2.4413646055437103e-06, + "loss": 0.6744, + "step": 229 + }, + { + "epoch": 0.07359215017064846, + "grad_norm": 0.6973935961723328, + "learning_rate": 2.452025586353945e-06, + "loss": 0.6579, + "step": 230 + }, + { + "epoch": 0.07391211604095563, + "grad_norm": 0.698112964630127, + "learning_rate": 2.4626865671641794e-06, + "loss": 0.6221, + "step": 231 + }, + { + "epoch": 0.07423208191126279, + "grad_norm": 0.8335263133049011, + "learning_rate": 2.473347547974414e-06, + "loss": 0.6552, + "step": 232 + }, + { + "epoch": 0.07455204778156997, + "grad_norm": 0.6335951685905457, + "learning_rate": 2.4840085287846484e-06, + "loss": 0.6366, + "step": 233 + }, + { + "epoch": 0.07487201365187714, + "grad_norm": 0.7226898074150085, + "learning_rate": 2.494669509594883e-06, + "loss": 0.6684, + "step": 234 + }, + { + "epoch": 0.0751919795221843, + "grad_norm": 0.7174168825149536, + "learning_rate": 2.5053304904051175e-06, + "loss": 0.6656, + "step": 235 + }, + { + "epoch": 0.07551194539249147, + "grad_norm": 0.7624191045761108, + "learning_rate": 2.515991471215352e-06, + "loss": 0.6451, + "step": 236 + }, + { + "epoch": 0.07583191126279863, + "grad_norm": 0.7072004675865173, + "learning_rate": 2.5266524520255865e-06, + "loss": 0.6608, + "step": 237 + }, + { + "epoch": 0.0761518771331058, + "grad_norm": 0.6365002989768982, + "learning_rate": 2.537313432835821e-06, + "loss": 0.6563, + "step": 238 + }, + { + "epoch": 0.07647184300341298, + "grad_norm": 0.7351357340812683, + "learning_rate": 2.5479744136460556e-06, + "loss": 0.6749, + "step": 239 + }, + { + "epoch": 0.07679180887372014, + "grad_norm": 0.6643660068511963, + "learning_rate": 2.55863539445629e-06, + "loss": 0.621, + "step": 240 + }, + { + "epoch": 0.0771117747440273, + "grad_norm": 0.767565131187439, + "learning_rate": 2.5692963752665246e-06, + "loss": 0.6579, + "step": 241 + }, + { + "epoch": 0.07743174061433447, + "grad_norm": 0.6990678906440735, + "learning_rate": 2.579957356076759e-06, + "loss": 0.6516, + "step": 242 + }, + { + "epoch": 0.07775170648464164, + "grad_norm": 0.6894639134407043, + "learning_rate": 2.5906183368869937e-06, + "loss": 0.6645, + "step": 243 + }, + { + "epoch": 0.0780716723549488, + "grad_norm": 0.7961663007736206, + "learning_rate": 2.601279317697228e-06, + "loss": 0.6722, + "step": 244 + }, + { + "epoch": 0.07839163822525597, + "grad_norm": 0.6465956568717957, + "learning_rate": 2.6119402985074627e-06, + "loss": 0.6401, + "step": 245 + }, + { + "epoch": 0.07871160409556315, + "grad_norm": 0.6665163040161133, + "learning_rate": 2.6226012793176977e-06, + "loss": 0.6665, + "step": 246 + }, + { + "epoch": 0.07903156996587031, + "grad_norm": 0.6720198392868042, + "learning_rate": 2.6332622601279318e-06, + "loss": 0.6444, + "step": 247 + }, + { + "epoch": 0.07935153583617748, + "grad_norm": 0.6681743860244751, + "learning_rate": 2.6439232409381667e-06, + "loss": 0.6919, + "step": 248 + }, + { + "epoch": 0.07967150170648464, + "grad_norm": 0.7217524647712708, + "learning_rate": 2.654584221748401e-06, + "loss": 0.6661, + "step": 249 + }, + { + "epoch": 0.0799914675767918, + "grad_norm": 0.6475330591201782, + "learning_rate": 2.6652452025586358e-06, + "loss": 0.6383, + "step": 250 + }, + { + "epoch": 0.08031143344709897, + "grad_norm": 0.6881884336471558, + "learning_rate": 2.6759061833688703e-06, + "loss": 0.6602, + "step": 251 + }, + { + "epoch": 0.08063139931740615, + "grad_norm": 0.6619604825973511, + "learning_rate": 2.686567164179105e-06, + "loss": 0.6399, + "step": 252 + }, + { + "epoch": 0.08095136518771331, + "grad_norm": 0.6265876889228821, + "learning_rate": 2.6972281449893394e-06, + "loss": 0.5992, + "step": 253 + }, + { + "epoch": 0.08127133105802048, + "grad_norm": 0.675489604473114, + "learning_rate": 2.707889125799574e-06, + "loss": 0.6298, + "step": 254 + }, + { + "epoch": 0.08159129692832764, + "grad_norm": 0.6868592500686646, + "learning_rate": 2.7185501066098084e-06, + "loss": 0.6043, + "step": 255 + }, + { + "epoch": 0.08191126279863481, + "grad_norm": 0.6755011677742004, + "learning_rate": 2.729211087420043e-06, + "loss": 0.6325, + "step": 256 + }, + { + "epoch": 0.08223122866894197, + "grad_norm": 0.680072009563446, + "learning_rate": 2.7398720682302775e-06, + "loss": 0.6401, + "step": 257 + }, + { + "epoch": 0.08255119453924915, + "grad_norm": 0.6612681746482849, + "learning_rate": 2.750533049040512e-06, + "loss": 0.6355, + "step": 258 + }, + { + "epoch": 0.08287116040955632, + "grad_norm": 0.7801066637039185, + "learning_rate": 2.7611940298507465e-06, + "loss": 0.6424, + "step": 259 + }, + { + "epoch": 0.08319112627986348, + "grad_norm": 0.663550615310669, + "learning_rate": 2.771855010660981e-06, + "loss": 0.6217, + "step": 260 + }, + { + "epoch": 0.08351109215017065, + "grad_norm": 0.641058623790741, + "learning_rate": 2.7825159914712156e-06, + "loss": 0.6314, + "step": 261 + }, + { + "epoch": 0.08383105802047781, + "grad_norm": 0.8522260189056396, + "learning_rate": 2.79317697228145e-06, + "loss": 0.6258, + "step": 262 + }, + { + "epoch": 0.08415102389078498, + "grad_norm": 0.6692214012145996, + "learning_rate": 2.8038379530916846e-06, + "loss": 0.6255, + "step": 263 + }, + { + "epoch": 0.08447098976109214, + "grad_norm": 0.6521351933479309, + "learning_rate": 2.814498933901919e-06, + "loss": 0.6366, + "step": 264 + }, + { + "epoch": 0.08479095563139932, + "grad_norm": 0.6812143921852112, + "learning_rate": 2.825159914712154e-06, + "loss": 0.6331, + "step": 265 + }, + { + "epoch": 0.08511092150170649, + "grad_norm": 0.6537390351295471, + "learning_rate": 2.835820895522388e-06, + "loss": 0.667, + "step": 266 + }, + { + "epoch": 0.08543088737201365, + "grad_norm": 0.6082245111465454, + "learning_rate": 2.846481876332623e-06, + "loss": 0.6321, + "step": 267 + }, + { + "epoch": 0.08575085324232082, + "grad_norm": 0.6908940672874451, + "learning_rate": 2.8571428571428573e-06, + "loss": 0.6164, + "step": 268 + }, + { + "epoch": 0.08607081911262798, + "grad_norm": 0.7633948922157288, + "learning_rate": 2.867803837953092e-06, + "loss": 0.6714, + "step": 269 + }, + { + "epoch": 0.08639078498293515, + "grad_norm": 0.6619597673416138, + "learning_rate": 2.8784648187633263e-06, + "loss": 0.6453, + "step": 270 + }, + { + "epoch": 0.08671075085324233, + "grad_norm": 0.7117516398429871, + "learning_rate": 2.8891257995735613e-06, + "loss": 0.6731, + "step": 271 + }, + { + "epoch": 0.08703071672354949, + "grad_norm": 0.7704693078994751, + "learning_rate": 2.8997867803837954e-06, + "loss": 0.646, + "step": 272 + }, + { + "epoch": 0.08735068259385666, + "grad_norm": 0.6669108867645264, + "learning_rate": 2.9104477611940303e-06, + "loss": 0.6283, + "step": 273 + }, + { + "epoch": 0.08767064846416382, + "grad_norm": 0.7196122407913208, + "learning_rate": 2.9211087420042644e-06, + "loss": 0.6208, + "step": 274 + }, + { + "epoch": 0.08799061433447099, + "grad_norm": 0.7552131414413452, + "learning_rate": 2.9317697228144994e-06, + "loss": 0.6619, + "step": 275 + }, + { + "epoch": 0.08831058020477815, + "grad_norm": 0.6823918223381042, + "learning_rate": 2.9424307036247335e-06, + "loss": 0.6141, + "step": 276 + }, + { + "epoch": 0.08863054607508532, + "grad_norm": 0.7164825797080994, + "learning_rate": 2.9530916844349684e-06, + "loss": 0.6555, + "step": 277 + }, + { + "epoch": 0.0889505119453925, + "grad_norm": 0.6870175004005432, + "learning_rate": 2.9637526652452025e-06, + "loss": 0.6142, + "step": 278 + }, + { + "epoch": 0.08927047781569966, + "grad_norm": 0.7144311666488647, + "learning_rate": 2.9744136460554375e-06, + "loss": 0.6431, + "step": 279 + }, + { + "epoch": 0.08959044368600683, + "grad_norm": 0.6681182384490967, + "learning_rate": 2.9850746268656716e-06, + "loss": 0.682, + "step": 280 + }, + { + "epoch": 0.08991040955631399, + "grad_norm": 0.6359437704086304, + "learning_rate": 2.9957356076759065e-06, + "loss": 0.6527, + "step": 281 + }, + { + "epoch": 0.09023037542662116, + "grad_norm": 0.718871533870697, + "learning_rate": 3.006396588486141e-06, + "loss": 0.6613, + "step": 282 + }, + { + "epoch": 0.09055034129692832, + "grad_norm": 0.7304879426956177, + "learning_rate": 3.0170575692963756e-06, + "loss": 0.6771, + "step": 283 + }, + { + "epoch": 0.0908703071672355, + "grad_norm": 0.6561257839202881, + "learning_rate": 3.0277185501066105e-06, + "loss": 0.5928, + "step": 284 + }, + { + "epoch": 0.09119027303754267, + "grad_norm": 0.7376087307929993, + "learning_rate": 3.0383795309168446e-06, + "loss": 0.6923, + "step": 285 + }, + { + "epoch": 0.09151023890784983, + "grad_norm": 0.6895893216133118, + "learning_rate": 3.0490405117270796e-06, + "loss": 0.6311, + "step": 286 + }, + { + "epoch": 0.091830204778157, + "grad_norm": 0.7078286409378052, + "learning_rate": 3.0597014925373137e-06, + "loss": 0.645, + "step": 287 + }, + { + "epoch": 0.09215017064846416, + "grad_norm": 0.8378876447677612, + "learning_rate": 3.0703624733475486e-06, + "loss": 0.6613, + "step": 288 + }, + { + "epoch": 0.09247013651877133, + "grad_norm": 0.7053190469741821, + "learning_rate": 3.0810234541577827e-06, + "loss": 0.5924, + "step": 289 + }, + { + "epoch": 0.0927901023890785, + "grad_norm": 0.6344333291053772, + "learning_rate": 3.0916844349680177e-06, + "loss": 0.6003, + "step": 290 + }, + { + "epoch": 0.09311006825938567, + "grad_norm": 0.6472283601760864, + "learning_rate": 3.1023454157782518e-06, + "loss": 0.6304, + "step": 291 + }, + { + "epoch": 0.09343003412969283, + "grad_norm": 0.6516737341880798, + "learning_rate": 3.1130063965884867e-06, + "loss": 0.6816, + "step": 292 + }, + { + "epoch": 0.09375, + "grad_norm": 0.730661153793335, + "learning_rate": 3.123667377398721e-06, + "loss": 0.624, + "step": 293 + }, + { + "epoch": 0.09406996587030717, + "grad_norm": 0.7991647720336914, + "learning_rate": 3.1343283582089558e-06, + "loss": 0.637, + "step": 294 + }, + { + "epoch": 0.09438993174061433, + "grad_norm": 0.6912937760353088, + "learning_rate": 3.14498933901919e-06, + "loss": 0.6645, + "step": 295 + }, + { + "epoch": 0.0947098976109215, + "grad_norm": 0.7326075434684753, + "learning_rate": 3.155650319829425e-06, + "loss": 0.6634, + "step": 296 + }, + { + "epoch": 0.09502986348122867, + "grad_norm": 0.736692488193512, + "learning_rate": 3.166311300639659e-06, + "loss": 0.6611, + "step": 297 + }, + { + "epoch": 0.09534982935153584, + "grad_norm": 0.6914559602737427, + "learning_rate": 3.176972281449894e-06, + "loss": 0.6449, + "step": 298 + }, + { + "epoch": 0.095669795221843, + "grad_norm": 0.7045454382896423, + "learning_rate": 3.187633262260128e-06, + "loss": 0.6794, + "step": 299 + }, + { + "epoch": 0.09598976109215017, + "grad_norm": 0.7817652821540833, + "learning_rate": 3.198294243070363e-06, + "loss": 0.6839, + "step": 300 + }, + { + "epoch": 0.09630972696245733, + "grad_norm": 0.7341583967208862, + "learning_rate": 3.208955223880597e-06, + "loss": 0.6486, + "step": 301 + }, + { + "epoch": 0.0966296928327645, + "grad_norm": 0.6466298699378967, + "learning_rate": 3.219616204690832e-06, + "loss": 0.6218, + "step": 302 + }, + { + "epoch": 0.09694965870307168, + "grad_norm": 0.7359744906425476, + "learning_rate": 3.230277185501066e-06, + "loss": 0.6357, + "step": 303 + }, + { + "epoch": 0.09726962457337884, + "grad_norm": 0.7613527774810791, + "learning_rate": 3.240938166311301e-06, + "loss": 0.662, + "step": 304 + }, + { + "epoch": 0.09758959044368601, + "grad_norm": 0.7327111959457397, + "learning_rate": 3.251599147121535e-06, + "loss": 0.6443, + "step": 305 + }, + { + "epoch": 0.09790955631399317, + "grad_norm": 0.7447808384895325, + "learning_rate": 3.26226012793177e-06, + "loss": 0.6432, + "step": 306 + }, + { + "epoch": 0.09822952218430034, + "grad_norm": 0.705234706401825, + "learning_rate": 3.272921108742004e-06, + "loss": 0.6817, + "step": 307 + }, + { + "epoch": 0.0985494880546075, + "grad_norm": 0.743067741394043, + "learning_rate": 3.283582089552239e-06, + "loss": 0.6824, + "step": 308 + }, + { + "epoch": 0.09886945392491468, + "grad_norm": 0.6528238654136658, + "learning_rate": 3.2942430703624733e-06, + "loss": 0.6466, + "step": 309 + }, + { + "epoch": 0.09918941979522185, + "grad_norm": 0.6550853252410889, + "learning_rate": 3.304904051172708e-06, + "loss": 0.5895, + "step": 310 + }, + { + "epoch": 0.09950938566552901, + "grad_norm": 0.6330170631408691, + "learning_rate": 3.3155650319829423e-06, + "loss": 0.6457, + "step": 311 + }, + { + "epoch": 0.09982935153583618, + "grad_norm": 0.7128297686576843, + "learning_rate": 3.3262260127931773e-06, + "loss": 0.6464, + "step": 312 + }, + { + "epoch": 0.10014931740614334, + "grad_norm": 0.636508047580719, + "learning_rate": 3.336886993603412e-06, + "loss": 0.6088, + "step": 313 + }, + { + "epoch": 0.10046928327645051, + "grad_norm": 0.8207855224609375, + "learning_rate": 3.3475479744136463e-06, + "loss": 0.7245, + "step": 314 + }, + { + "epoch": 0.10078924914675767, + "grad_norm": 0.6568029522895813, + "learning_rate": 3.3582089552238813e-06, + "loss": 0.6385, + "step": 315 + }, + { + "epoch": 0.10110921501706485, + "grad_norm": 0.7073377966880798, + "learning_rate": 3.3688699360341154e-06, + "loss": 0.6677, + "step": 316 + }, + { + "epoch": 0.10142918088737202, + "grad_norm": 0.6592832207679749, + "learning_rate": 3.3795309168443503e-06, + "loss": 0.6273, + "step": 317 + }, + { + "epoch": 0.10174914675767918, + "grad_norm": 0.6786140203475952, + "learning_rate": 3.3901918976545844e-06, + "loss": 0.6669, + "step": 318 + }, + { + "epoch": 0.10206911262798635, + "grad_norm": 0.6719084978103638, + "learning_rate": 3.4008528784648194e-06, + "loss": 0.6335, + "step": 319 + }, + { + "epoch": 0.10238907849829351, + "grad_norm": 0.6851851940155029, + "learning_rate": 3.4115138592750535e-06, + "loss": 0.6011, + "step": 320 + }, + { + "epoch": 0.10270904436860068, + "grad_norm": 0.6099117994308472, + "learning_rate": 3.4221748400852884e-06, + "loss": 0.6332, + "step": 321 + }, + { + "epoch": 0.10302901023890786, + "grad_norm": 0.6866176724433899, + "learning_rate": 3.4328358208955225e-06, + "loss": 0.6017, + "step": 322 + }, + { + "epoch": 0.10334897610921502, + "grad_norm": 0.7074094414710999, + "learning_rate": 3.4434968017057575e-06, + "loss": 0.6177, + "step": 323 + }, + { + "epoch": 0.10366894197952219, + "grad_norm": 0.6538453102111816, + "learning_rate": 3.4541577825159916e-06, + "loss": 0.644, + "step": 324 + }, + { + "epoch": 0.10398890784982935, + "grad_norm": 0.6807860732078552, + "learning_rate": 3.4648187633262265e-06, + "loss": 0.6325, + "step": 325 + }, + { + "epoch": 0.10430887372013652, + "grad_norm": 0.8274717330932617, + "learning_rate": 3.4754797441364606e-06, + "loss": 0.6364, + "step": 326 + }, + { + "epoch": 0.10462883959044368, + "grad_norm": 0.7771089673042297, + "learning_rate": 3.4861407249466956e-06, + "loss": 0.6476, + "step": 327 + }, + { + "epoch": 0.10494880546075085, + "grad_norm": 0.7471724152565002, + "learning_rate": 3.4968017057569297e-06, + "loss": 0.6656, + "step": 328 + }, + { + "epoch": 0.10526877133105803, + "grad_norm": 0.7084169387817383, + "learning_rate": 3.5074626865671646e-06, + "loss": 0.6722, + "step": 329 + }, + { + "epoch": 0.10558873720136519, + "grad_norm": 0.700748085975647, + "learning_rate": 3.5181236673773987e-06, + "loss": 0.622, + "step": 330 + }, + { + "epoch": 0.10590870307167236, + "grad_norm": 0.8253344893455505, + "learning_rate": 3.5287846481876337e-06, + "loss": 0.6405, + "step": 331 + }, + { + "epoch": 0.10622866894197952, + "grad_norm": 0.7408985495567322, + "learning_rate": 3.5394456289978678e-06, + "loss": 0.6814, + "step": 332 + }, + { + "epoch": 0.10654863481228669, + "grad_norm": 0.9821094274520874, + "learning_rate": 3.5501066098081027e-06, + "loss": 0.6398, + "step": 333 + }, + { + "epoch": 0.10686860068259385, + "grad_norm": 0.6583428978919983, + "learning_rate": 3.560767590618337e-06, + "loss": 0.6206, + "step": 334 + }, + { + "epoch": 0.10718856655290103, + "grad_norm": 0.5934454798698425, + "learning_rate": 3.5714285714285718e-06, + "loss": 0.6149, + "step": 335 + }, + { + "epoch": 0.1075085324232082, + "grad_norm": 0.6851022839546204, + "learning_rate": 3.582089552238806e-06, + "loss": 0.6245, + "step": 336 + }, + { + "epoch": 0.10782849829351536, + "grad_norm": 0.6943828463554382, + "learning_rate": 3.592750533049041e-06, + "loss": 0.6542, + "step": 337 + }, + { + "epoch": 0.10814846416382252, + "grad_norm": 0.6568377017974854, + "learning_rate": 3.603411513859275e-06, + "loss": 0.6493, + "step": 338 + }, + { + "epoch": 0.10846843003412969, + "grad_norm": 0.6676375865936279, + "learning_rate": 3.61407249466951e-06, + "loss": 0.6573, + "step": 339 + }, + { + "epoch": 0.10878839590443685, + "grad_norm": 0.6609268188476562, + "learning_rate": 3.624733475479744e-06, + "loss": 0.6402, + "step": 340 + }, + { + "epoch": 0.10910836177474403, + "grad_norm": 0.6636278033256531, + "learning_rate": 3.635394456289979e-06, + "loss": 0.6256, + "step": 341 + }, + { + "epoch": 0.1094283276450512, + "grad_norm": 0.6692773699760437, + "learning_rate": 3.6460554371002135e-06, + "loss": 0.6314, + "step": 342 + }, + { + "epoch": 0.10974829351535836, + "grad_norm": 0.7020422220230103, + "learning_rate": 3.656716417910448e-06, + "loss": 0.6009, + "step": 343 + }, + { + "epoch": 0.11006825938566553, + "grad_norm": 0.5868972539901733, + "learning_rate": 3.667377398720683e-06, + "loss": 0.627, + "step": 344 + }, + { + "epoch": 0.1103882252559727, + "grad_norm": 0.6617975831031799, + "learning_rate": 3.678038379530917e-06, + "loss": 0.6157, + "step": 345 + }, + { + "epoch": 0.11070819112627986, + "grad_norm": 0.6820520758628845, + "learning_rate": 3.688699360341152e-06, + "loss": 0.6337, + "step": 346 + }, + { + "epoch": 0.11102815699658702, + "grad_norm": 0.7438987493515015, + "learning_rate": 3.699360341151386e-06, + "loss": 0.6758, + "step": 347 + }, + { + "epoch": 0.1113481228668942, + "grad_norm": 0.7671152353286743, + "learning_rate": 3.710021321961621e-06, + "loss": 0.6857, + "step": 348 + }, + { + "epoch": 0.11166808873720137, + "grad_norm": 0.7241283655166626, + "learning_rate": 3.720682302771855e-06, + "loss": 0.6179, + "step": 349 + }, + { + "epoch": 0.11198805460750853, + "grad_norm": 0.6965861916542053, + "learning_rate": 3.73134328358209e-06, + "loss": 0.6618, + "step": 350 + }, + { + "epoch": 0.1123080204778157, + "grad_norm": 0.7182785868644714, + "learning_rate": 3.742004264392324e-06, + "loss": 0.6689, + "step": 351 + }, + { + "epoch": 0.11262798634812286, + "grad_norm": 0.7257066369056702, + "learning_rate": 3.752665245202559e-06, + "loss": 0.6041, + "step": 352 + }, + { + "epoch": 0.11294795221843003, + "grad_norm": 0.6591728925704956, + "learning_rate": 3.7633262260127933e-06, + "loss": 0.6401, + "step": 353 + }, + { + "epoch": 0.11326791808873721, + "grad_norm": 0.6794602870941162, + "learning_rate": 3.773987206823028e-06, + "loss": 0.6376, + "step": 354 + }, + { + "epoch": 0.11358788395904437, + "grad_norm": 0.7658209204673767, + "learning_rate": 3.7846481876332623e-06, + "loss": 0.634, + "step": 355 + }, + { + "epoch": 0.11390784982935154, + "grad_norm": 0.6292558312416077, + "learning_rate": 3.7953091684434973e-06, + "loss": 0.6086, + "step": 356 + }, + { + "epoch": 0.1142278156996587, + "grad_norm": 0.6457000374794006, + "learning_rate": 3.8059701492537314e-06, + "loss": 0.605, + "step": 357 + }, + { + "epoch": 0.11454778156996587, + "grad_norm": 0.7265068292617798, + "learning_rate": 3.816631130063966e-06, + "loss": 0.6344, + "step": 358 + }, + { + "epoch": 0.11486774744027303, + "grad_norm": 0.6964762806892395, + "learning_rate": 3.827292110874201e-06, + "loss": 0.6327, + "step": 359 + }, + { + "epoch": 0.11518771331058021, + "grad_norm": 0.717154860496521, + "learning_rate": 3.837953091684435e-06, + "loss": 0.6271, + "step": 360 + }, + { + "epoch": 0.11550767918088738, + "grad_norm": 0.687904953956604, + "learning_rate": 3.84861407249467e-06, + "loss": 0.6407, + "step": 361 + }, + { + "epoch": 0.11582764505119454, + "grad_norm": 0.6947673559188843, + "learning_rate": 3.859275053304904e-06, + "loss": 0.614, + "step": 362 + }, + { + "epoch": 0.1161476109215017, + "grad_norm": 0.7783513069152832, + "learning_rate": 3.869936034115139e-06, + "loss": 0.622, + "step": 363 + }, + { + "epoch": 0.11646757679180887, + "grad_norm": 0.7462539076805115, + "learning_rate": 3.8805970149253735e-06, + "loss": 0.6867, + "step": 364 + }, + { + "epoch": 0.11678754266211604, + "grad_norm": 0.6062483787536621, + "learning_rate": 3.891257995735608e-06, + "loss": 0.6055, + "step": 365 + }, + { + "epoch": 0.1171075085324232, + "grad_norm": 0.7094764113426208, + "learning_rate": 3.9019189765458425e-06, + "loss": 0.63, + "step": 366 + }, + { + "epoch": 0.11742747440273038, + "grad_norm": 0.7285301089286804, + "learning_rate": 3.912579957356077e-06, + "loss": 0.6302, + "step": 367 + }, + { + "epoch": 0.11774744027303755, + "grad_norm": 0.6298773884773254, + "learning_rate": 3.9232409381663116e-06, + "loss": 0.6001, + "step": 368 + }, + { + "epoch": 0.11806740614334471, + "grad_norm": 0.7302676439285278, + "learning_rate": 3.933901918976546e-06, + "loss": 0.6395, + "step": 369 + }, + { + "epoch": 0.11838737201365188, + "grad_norm": 0.6924223303794861, + "learning_rate": 3.944562899786781e-06, + "loss": 0.5949, + "step": 370 + }, + { + "epoch": 0.11870733788395904, + "grad_norm": 0.6996256709098816, + "learning_rate": 3.955223880597015e-06, + "loss": 0.6594, + "step": 371 + }, + { + "epoch": 0.1190273037542662, + "grad_norm": 0.7116991281509399, + "learning_rate": 3.96588486140725e-06, + "loss": 0.6228, + "step": 372 + }, + { + "epoch": 0.11934726962457338, + "grad_norm": 0.6261963844299316, + "learning_rate": 3.976545842217484e-06, + "loss": 0.606, + "step": 373 + }, + { + "epoch": 0.11966723549488055, + "grad_norm": 0.7136417627334595, + "learning_rate": 3.987206823027719e-06, + "loss": 0.5899, + "step": 374 + }, + { + "epoch": 0.11998720136518772, + "grad_norm": 0.6391706466674805, + "learning_rate": 3.997867803837953e-06, + "loss": 0.6355, + "step": 375 + }, + { + "epoch": 0.12030716723549488, + "grad_norm": 0.6646369099617004, + "learning_rate": 4.008528784648188e-06, + "loss": 0.6122, + "step": 376 + }, + { + "epoch": 0.12062713310580205, + "grad_norm": 0.6632401347160339, + "learning_rate": 4.019189765458423e-06, + "loss": 0.6471, + "step": 377 + }, + { + "epoch": 0.12094709897610921, + "grad_norm": 0.6242071390151978, + "learning_rate": 4.029850746268657e-06, + "loss": 0.5711, + "step": 378 + }, + { + "epoch": 0.12126706484641639, + "grad_norm": 0.641707181930542, + "learning_rate": 4.040511727078892e-06, + "loss": 0.6318, + "step": 379 + }, + { + "epoch": 0.12158703071672355, + "grad_norm": 0.7148804068565369, + "learning_rate": 4.051172707889126e-06, + "loss": 0.6299, + "step": 380 + }, + { + "epoch": 0.12190699658703072, + "grad_norm": 0.7760037183761597, + "learning_rate": 4.061833688699361e-06, + "loss": 0.6155, + "step": 381 + }, + { + "epoch": 0.12222696245733788, + "grad_norm": 0.679989755153656, + "learning_rate": 4.072494669509595e-06, + "loss": 0.6409, + "step": 382 + }, + { + "epoch": 0.12254692832764505, + "grad_norm": 0.6069170236587524, + "learning_rate": 4.08315565031983e-06, + "loss": 0.607, + "step": 383 + }, + { + "epoch": 0.12286689419795221, + "grad_norm": 0.6047013401985168, + "learning_rate": 4.093816631130064e-06, + "loss": 0.6175, + "step": 384 + }, + { + "epoch": 0.12318686006825938, + "grad_norm": 0.6415553689002991, + "learning_rate": 4.104477611940299e-06, + "loss": 0.6168, + "step": 385 + }, + { + "epoch": 0.12350682593856656, + "grad_norm": 0.7326337695121765, + "learning_rate": 4.115138592750533e-06, + "loss": 0.6247, + "step": 386 + }, + { + "epoch": 0.12382679180887372, + "grad_norm": 0.7146400809288025, + "learning_rate": 4.125799573560768e-06, + "loss": 0.62, + "step": 387 + }, + { + "epoch": 0.12414675767918089, + "grad_norm": 0.6480631828308105, + "learning_rate": 4.136460554371002e-06, + "loss": 0.6511, + "step": 388 + }, + { + "epoch": 0.12446672354948805, + "grad_norm": 0.6476372480392456, + "learning_rate": 4.1471215351812375e-06, + "loss": 0.6027, + "step": 389 + }, + { + "epoch": 0.12478668941979522, + "grad_norm": 0.644311785697937, + "learning_rate": 4.157782515991471e-06, + "loss": 0.6371, + "step": 390 + }, + { + "epoch": 0.1251066552901024, + "grad_norm": 0.6093184351921082, + "learning_rate": 4.1684434968017065e-06, + "loss": 0.638, + "step": 391 + }, + { + "epoch": 0.12542662116040956, + "grad_norm": 0.6396116018295288, + "learning_rate": 4.17910447761194e-06, + "loss": 0.6051, + "step": 392 + }, + { + "epoch": 0.12574658703071673, + "grad_norm": 0.6135868430137634, + "learning_rate": 4.1897654584221756e-06, + "loss": 0.6662, + "step": 393 + }, + { + "epoch": 0.1260665529010239, + "grad_norm": 0.661790132522583, + "learning_rate": 4.200426439232409e-06, + "loss": 0.623, + "step": 394 + }, + { + "epoch": 0.12638651877133106, + "grad_norm": 0.687401533126831, + "learning_rate": 4.211087420042645e-06, + "loss": 0.6448, + "step": 395 + }, + { + "epoch": 0.12670648464163822, + "grad_norm": 0.6437802910804749, + "learning_rate": 4.221748400852878e-06, + "loss": 0.6103, + "step": 396 + }, + { + "epoch": 0.1270264505119454, + "grad_norm": 0.7061891555786133, + "learning_rate": 4.232409381663114e-06, + "loss": 0.634, + "step": 397 + }, + { + "epoch": 0.12734641638225255, + "grad_norm": 0.6025312542915344, + "learning_rate": 4.243070362473347e-06, + "loss": 0.6258, + "step": 398 + }, + { + "epoch": 0.12766638225255972, + "grad_norm": 0.9356108903884888, + "learning_rate": 4.253731343283583e-06, + "loss": 0.5903, + "step": 399 + }, + { + "epoch": 0.12798634812286688, + "grad_norm": 0.6520940661430359, + "learning_rate": 4.264392324093816e-06, + "loss": 0.5926, + "step": 400 + }, + { + "epoch": 0.12830631399317405, + "grad_norm": 0.6317100524902344, + "learning_rate": 4.275053304904052e-06, + "loss": 0.5891, + "step": 401 + }, + { + "epoch": 0.12862627986348124, + "grad_norm": 0.6645535826683044, + "learning_rate": 4.2857142857142855e-06, + "loss": 0.572, + "step": 402 + }, + { + "epoch": 0.1289462457337884, + "grad_norm": 0.6676063537597656, + "learning_rate": 4.296375266524521e-06, + "loss": 0.6334, + "step": 403 + }, + { + "epoch": 0.12926621160409557, + "grad_norm": 0.6608290672302246, + "learning_rate": 4.3070362473347545e-06, + "loss": 0.6003, + "step": 404 + }, + { + "epoch": 0.12958617747440274, + "grad_norm": 0.7465037107467651, + "learning_rate": 4.31769722814499e-06, + "loss": 0.6297, + "step": 405 + }, + { + "epoch": 0.1299061433447099, + "grad_norm": 0.7195454239845276, + "learning_rate": 4.3283582089552236e-06, + "loss": 0.6276, + "step": 406 + }, + { + "epoch": 0.13022610921501707, + "grad_norm": 0.7333062887191772, + "learning_rate": 4.339019189765459e-06, + "loss": 0.5858, + "step": 407 + }, + { + "epoch": 0.13054607508532423, + "grad_norm": 0.7343773245811462, + "learning_rate": 4.3496801705756935e-06, + "loss": 0.6304, + "step": 408 + }, + { + "epoch": 0.1308660409556314, + "grad_norm": 0.7145223021507263, + "learning_rate": 4.360341151385928e-06, + "loss": 0.676, + "step": 409 + }, + { + "epoch": 0.13118600682593856, + "grad_norm": 0.6226415634155273, + "learning_rate": 4.3710021321961625e-06, + "loss": 0.6018, + "step": 410 + }, + { + "epoch": 0.13150597269624573, + "grad_norm": 0.7167772650718689, + "learning_rate": 4.381663113006397e-06, + "loss": 0.6159, + "step": 411 + }, + { + "epoch": 0.1318259385665529, + "grad_norm": 0.6907845139503479, + "learning_rate": 4.3923240938166316e-06, + "loss": 0.6303, + "step": 412 + }, + { + "epoch": 0.13214590443686006, + "grad_norm": 0.729068398475647, + "learning_rate": 4.402985074626866e-06, + "loss": 0.603, + "step": 413 + }, + { + "epoch": 0.13246587030716722, + "grad_norm": 0.6759557127952576, + "learning_rate": 4.413646055437101e-06, + "loss": 0.6493, + "step": 414 + }, + { + "epoch": 0.13278583617747441, + "grad_norm": 0.6795688271522522, + "learning_rate": 4.424307036247335e-06, + "loss": 0.631, + "step": 415 + }, + { + "epoch": 0.13310580204778158, + "grad_norm": 0.6279520988464355, + "learning_rate": 4.43496801705757e-06, + "loss": 0.596, + "step": 416 + }, + { + "epoch": 0.13342576791808874, + "grad_norm": 0.6398487091064453, + "learning_rate": 4.445628997867804e-06, + "loss": 0.6141, + "step": 417 + }, + { + "epoch": 0.1337457337883959, + "grad_norm": 0.6193283200263977, + "learning_rate": 4.456289978678039e-06, + "loss": 0.6011, + "step": 418 + }, + { + "epoch": 0.13406569965870307, + "grad_norm": 0.6516242027282715, + "learning_rate": 4.466950959488273e-06, + "loss": 0.6447, + "step": 419 + }, + { + "epoch": 0.13438566552901024, + "grad_norm": 0.6823792457580566, + "learning_rate": 4.477611940298508e-06, + "loss": 0.6377, + "step": 420 + }, + { + "epoch": 0.1347056313993174, + "grad_norm": 0.6426964402198792, + "learning_rate": 4.488272921108742e-06, + "loss": 0.5983, + "step": 421 + }, + { + "epoch": 0.13502559726962457, + "grad_norm": 0.6207419633865356, + "learning_rate": 4.498933901918977e-06, + "loss": 0.632, + "step": 422 + }, + { + "epoch": 0.13534556313993173, + "grad_norm": 0.6856409907341003, + "learning_rate": 4.509594882729211e-06, + "loss": 0.6397, + "step": 423 + }, + { + "epoch": 0.1356655290102389, + "grad_norm": 0.666517972946167, + "learning_rate": 4.520255863539446e-06, + "loss": 0.6226, + "step": 424 + }, + { + "epoch": 0.13598549488054607, + "grad_norm": 0.7025275230407715, + "learning_rate": 4.53091684434968e-06, + "loss": 0.6276, + "step": 425 + }, + { + "epoch": 0.13630546075085323, + "grad_norm": 0.7011261582374573, + "learning_rate": 4.541577825159915e-06, + "loss": 0.6618, + "step": 426 + }, + { + "epoch": 0.13662542662116042, + "grad_norm": 0.6024221777915955, + "learning_rate": 4.5522388059701495e-06, + "loss": 0.574, + "step": 427 + }, + { + "epoch": 0.1369453924914676, + "grad_norm": 0.6258330345153809, + "learning_rate": 4.562899786780384e-06, + "loss": 0.6353, + "step": 428 + }, + { + "epoch": 0.13726535836177475, + "grad_norm": 0.633340060710907, + "learning_rate": 4.5735607675906185e-06, + "loss": 0.5696, + "step": 429 + }, + { + "epoch": 0.13758532423208192, + "grad_norm": 0.6408478617668152, + "learning_rate": 4.584221748400853e-06, + "loss": 0.6102, + "step": 430 + }, + { + "epoch": 0.13790529010238908, + "grad_norm": 0.7019578814506531, + "learning_rate": 4.5948827292110876e-06, + "loss": 0.651, + "step": 431 + }, + { + "epoch": 0.13822525597269625, + "grad_norm": 0.6981320381164551, + "learning_rate": 4.605543710021322e-06, + "loss": 0.6474, + "step": 432 + }, + { + "epoch": 0.1385452218430034, + "grad_norm": 0.6756162643432617, + "learning_rate": 4.616204690831557e-06, + "loss": 0.5541, + "step": 433 + }, + { + "epoch": 0.13886518771331058, + "grad_norm": 0.7130482792854309, + "learning_rate": 4.626865671641791e-06, + "loss": 0.5995, + "step": 434 + }, + { + "epoch": 0.13918515358361774, + "grad_norm": 0.6483554244041443, + "learning_rate": 4.637526652452026e-06, + "loss": 0.5764, + "step": 435 + }, + { + "epoch": 0.1395051194539249, + "grad_norm": 0.7095510959625244, + "learning_rate": 4.64818763326226e-06, + "loss": 0.6112, + "step": 436 + }, + { + "epoch": 0.13982508532423207, + "grad_norm": 0.8624890446662903, + "learning_rate": 4.658848614072495e-06, + "loss": 0.6447, + "step": 437 + }, + { + "epoch": 0.14014505119453924, + "grad_norm": 0.7431426644325256, + "learning_rate": 4.669509594882729e-06, + "loss": 0.6282, + "step": 438 + }, + { + "epoch": 0.1404650170648464, + "grad_norm": 0.6454997062683105, + "learning_rate": 4.680170575692965e-06, + "loss": 0.6507, + "step": 439 + }, + { + "epoch": 0.1407849829351536, + "grad_norm": 0.7619973421096802, + "learning_rate": 4.690831556503198e-06, + "loss": 0.6045, + "step": 440 + }, + { + "epoch": 0.14110494880546076, + "grad_norm": 0.678459644317627, + "learning_rate": 4.701492537313434e-06, + "loss": 0.6377, + "step": 441 + }, + { + "epoch": 0.14142491467576793, + "grad_norm": 0.6140632033348083, + "learning_rate": 4.712153518123667e-06, + "loss": 0.6237, + "step": 442 + }, + { + "epoch": 0.1417448805460751, + "grad_norm": 0.6932960748672485, + "learning_rate": 4.722814498933903e-06, + "loss": 0.6598, + "step": 443 + }, + { + "epoch": 0.14206484641638226, + "grad_norm": 0.8086076974868774, + "learning_rate": 4.733475479744136e-06, + "loss": 0.6339, + "step": 444 + }, + { + "epoch": 0.14238481228668942, + "grad_norm": 0.7320109009742737, + "learning_rate": 4.744136460554372e-06, + "loss": 0.5965, + "step": 445 + }, + { + "epoch": 0.1427047781569966, + "grad_norm": 0.6901565790176392, + "learning_rate": 4.7547974413646055e-06, + "loss": 0.6033, + "step": 446 + }, + { + "epoch": 0.14302474402730375, + "grad_norm": 0.7285332679748535, + "learning_rate": 4.765458422174841e-06, + "loss": 0.6248, + "step": 447 + }, + { + "epoch": 0.14334470989761092, + "grad_norm": 0.6956148147583008, + "learning_rate": 4.7761194029850745e-06, + "loss": 0.6356, + "step": 448 + }, + { + "epoch": 0.14366467576791808, + "grad_norm": 0.673285186290741, + "learning_rate": 4.78678038379531e-06, + "loss": 0.6328, + "step": 449 + }, + { + "epoch": 0.14398464163822525, + "grad_norm": 0.7397879958152771, + "learning_rate": 4.797441364605544e-06, + "loss": 0.6247, + "step": 450 + }, + { + "epoch": 0.1443046075085324, + "grad_norm": 0.825232982635498, + "learning_rate": 4.808102345415779e-06, + "loss": 0.6125, + "step": 451 + }, + { + "epoch": 0.14462457337883958, + "grad_norm": 0.727387011051178, + "learning_rate": 4.8187633262260135e-06, + "loss": 0.6363, + "step": 452 + }, + { + "epoch": 0.14494453924914677, + "grad_norm": 0.6681481599807739, + "learning_rate": 4.829424307036248e-06, + "loss": 0.6234, + "step": 453 + }, + { + "epoch": 0.14526450511945393, + "grad_norm": 0.7631570100784302, + "learning_rate": 4.8400852878464825e-06, + "loss": 0.6238, + "step": 454 + }, + { + "epoch": 0.1455844709897611, + "grad_norm": 0.6824873089790344, + "learning_rate": 4.850746268656717e-06, + "loss": 0.6299, + "step": 455 + }, + { + "epoch": 0.14590443686006827, + "grad_norm": 0.6673473119735718, + "learning_rate": 4.8614072494669516e-06, + "loss": 0.6139, + "step": 456 + }, + { + "epoch": 0.14622440273037543, + "grad_norm": 0.6729328036308289, + "learning_rate": 4.872068230277186e-06, + "loss": 0.5745, + "step": 457 + }, + { + "epoch": 0.1465443686006826, + "grad_norm": 0.7668410539627075, + "learning_rate": 4.882729211087421e-06, + "loss": 0.6114, + "step": 458 + }, + { + "epoch": 0.14686433447098976, + "grad_norm": 0.657248318195343, + "learning_rate": 4.893390191897655e-06, + "loss": 0.639, + "step": 459 + }, + { + "epoch": 0.14718430034129693, + "grad_norm": 0.6458208560943604, + "learning_rate": 4.90405117270789e-06, + "loss": 0.6091, + "step": 460 + }, + { + "epoch": 0.1475042662116041, + "grad_norm": 0.6650400161743164, + "learning_rate": 4.914712153518124e-06, + "loss": 0.5947, + "step": 461 + }, + { + "epoch": 0.14782423208191126, + "grad_norm": 0.6707201600074768, + "learning_rate": 4.925373134328359e-06, + "loss": 0.592, + "step": 462 + }, + { + "epoch": 0.14814419795221842, + "grad_norm": 0.7198519706726074, + "learning_rate": 4.936034115138593e-06, + "loss": 0.5947, + "step": 463 + }, + { + "epoch": 0.14846416382252559, + "grad_norm": 0.7003646492958069, + "learning_rate": 4.946695095948828e-06, + "loss": 0.6417, + "step": 464 + }, + { + "epoch": 0.14878412969283278, + "grad_norm": 0.6954401731491089, + "learning_rate": 4.957356076759062e-06, + "loss": 0.6477, + "step": 465 + }, + { + "epoch": 0.14910409556313994, + "grad_norm": 0.7278227806091309, + "learning_rate": 4.968017057569297e-06, + "loss": 0.651, + "step": 466 + }, + { + "epoch": 0.1494240614334471, + "grad_norm": 0.8354657888412476, + "learning_rate": 4.978678038379531e-06, + "loss": 0.5986, + "step": 467 + }, + { + "epoch": 0.14974402730375427, + "grad_norm": 0.7701458930969238, + "learning_rate": 4.989339019189766e-06, + "loss": 0.6188, + "step": 468 + }, + { + "epoch": 0.15006399317406144, + "grad_norm": 0.6838735938072205, + "learning_rate": 5e-06, + "loss": 0.6067, + "step": 469 + }, + { + "epoch": 0.1503839590443686, + "grad_norm": 0.6737359762191772, + "learning_rate": 5.010660980810235e-06, + "loss": 0.632, + "step": 470 + }, + { + "epoch": 0.15070392491467577, + "grad_norm": 0.7790951132774353, + "learning_rate": 5.02132196162047e-06, + "loss": 0.5928, + "step": 471 + }, + { + "epoch": 0.15102389078498293, + "grad_norm": 0.7367972135543823, + "learning_rate": 5.031982942430704e-06, + "loss": 0.6217, + "step": 472 + }, + { + "epoch": 0.1513438566552901, + "grad_norm": 0.7044655084609985, + "learning_rate": 5.0426439232409385e-06, + "loss": 0.6131, + "step": 473 + }, + { + "epoch": 0.15166382252559726, + "grad_norm": 0.6171327233314514, + "learning_rate": 5.053304904051173e-06, + "loss": 0.6232, + "step": 474 + }, + { + "epoch": 0.15198378839590443, + "grad_norm": 0.7287469506263733, + "learning_rate": 5.063965884861408e-06, + "loss": 0.5999, + "step": 475 + }, + { + "epoch": 0.1523037542662116, + "grad_norm": 0.7412554621696472, + "learning_rate": 5.074626865671642e-06, + "loss": 0.6647, + "step": 476 + }, + { + "epoch": 0.15262372013651876, + "grad_norm": 0.7282645106315613, + "learning_rate": 5.085287846481877e-06, + "loss": 0.6364, + "step": 477 + }, + { + "epoch": 0.15294368600682595, + "grad_norm": 0.7131372094154358, + "learning_rate": 5.095948827292111e-06, + "loss": 0.6549, + "step": 478 + }, + { + "epoch": 0.15326365187713312, + "grad_norm": 0.7518019676208496, + "learning_rate": 5.1066098081023465e-06, + "loss": 0.6298, + "step": 479 + }, + { + "epoch": 0.15358361774744028, + "grad_norm": 0.6324292421340942, + "learning_rate": 5.11727078891258e-06, + "loss": 0.6214, + "step": 480 + }, + { + "epoch": 0.15390358361774745, + "grad_norm": 0.692251980304718, + "learning_rate": 5.127931769722815e-06, + "loss": 0.5769, + "step": 481 + }, + { + "epoch": 0.1542235494880546, + "grad_norm": 0.6692104339599609, + "learning_rate": 5.138592750533049e-06, + "loss": 0.5986, + "step": 482 + }, + { + "epoch": 0.15454351535836178, + "grad_norm": 0.6897379159927368, + "learning_rate": 5.149253731343285e-06, + "loss": 0.6281, + "step": 483 + }, + { + "epoch": 0.15486348122866894, + "grad_norm": 0.6690213680267334, + "learning_rate": 5.159914712153518e-06, + "loss": 0.574, + "step": 484 + }, + { + "epoch": 0.1551834470989761, + "grad_norm": 0.6694300770759583, + "learning_rate": 5.170575692963753e-06, + "loss": 0.6249, + "step": 485 + }, + { + "epoch": 0.15550341296928327, + "grad_norm": 0.6471107006072998, + "learning_rate": 5.181236673773987e-06, + "loss": 0.6231, + "step": 486 + }, + { + "epoch": 0.15582337883959044, + "grad_norm": 0.618867039680481, + "learning_rate": 5.191897654584223e-06, + "loss": 0.6061, + "step": 487 + }, + { + "epoch": 0.1561433447098976, + "grad_norm": 0.6762124300003052, + "learning_rate": 5.202558635394456e-06, + "loss": 0.6264, + "step": 488 + }, + { + "epoch": 0.15646331058020477, + "grad_norm": 0.6503360867500305, + "learning_rate": 5.213219616204691e-06, + "loss": 0.6345, + "step": 489 + }, + { + "epoch": 0.15678327645051193, + "grad_norm": 0.7969846725463867, + "learning_rate": 5.2238805970149255e-06, + "loss": 0.6233, + "step": 490 + }, + { + "epoch": 0.15710324232081913, + "grad_norm": 0.7003441452980042, + "learning_rate": 5.234541577825161e-06, + "loss": 0.634, + "step": 491 + }, + { + "epoch": 0.1574232081911263, + "grad_norm": 0.6595831513404846, + "learning_rate": 5.245202558635395e-06, + "loss": 0.6253, + "step": 492 + }, + { + "epoch": 0.15774317406143346, + "grad_norm": 0.7349994778633118, + "learning_rate": 5.255863539445629e-06, + "loss": 0.6214, + "step": 493 + }, + { + "epoch": 0.15806313993174062, + "grad_norm": 0.7195835113525391, + "learning_rate": 5.2665245202558636e-06, + "loss": 0.5856, + "step": 494 + }, + { + "epoch": 0.15838310580204779, + "grad_norm": 0.6510567665100098, + "learning_rate": 5.277185501066099e-06, + "loss": 0.6374, + "step": 495 + }, + { + "epoch": 0.15870307167235495, + "grad_norm": 0.7259697318077087, + "learning_rate": 5.2878464818763335e-06, + "loss": 0.6446, + "step": 496 + }, + { + "epoch": 0.15902303754266212, + "grad_norm": 0.6870812177658081, + "learning_rate": 5.298507462686567e-06, + "loss": 0.5873, + "step": 497 + }, + { + "epoch": 0.15934300341296928, + "grad_norm": 0.8159638047218323, + "learning_rate": 5.309168443496802e-06, + "loss": 0.6401, + "step": 498 + }, + { + "epoch": 0.15966296928327645, + "grad_norm": 0.6727353930473328, + "learning_rate": 5.319829424307037e-06, + "loss": 0.5834, + "step": 499 + }, + { + "epoch": 0.1599829351535836, + "grad_norm": 0.6303347945213318, + "learning_rate": 5.3304904051172716e-06, + "loss": 0.5554, + "step": 500 + }, + { + "epoch": 0.16030290102389078, + "grad_norm": 0.6501518487930298, + "learning_rate": 5.341151385927505e-06, + "loss": 0.6153, + "step": 501 + }, + { + "epoch": 0.16062286689419794, + "grad_norm": 0.7419226169586182, + "learning_rate": 5.351812366737741e-06, + "loss": 0.6119, + "step": 502 + }, + { + "epoch": 0.1609428327645051, + "grad_norm": 0.6859214901924133, + "learning_rate": 5.362473347547975e-06, + "loss": 0.6041, + "step": 503 + }, + { + "epoch": 0.1612627986348123, + "grad_norm": 0.68569016456604, + "learning_rate": 5.37313432835821e-06, + "loss": 0.6317, + "step": 504 + }, + { + "epoch": 0.16158276450511946, + "grad_norm": 0.6246910095214844, + "learning_rate": 5.383795309168443e-06, + "loss": 0.6128, + "step": 505 + }, + { + "epoch": 0.16190273037542663, + "grad_norm": 0.7596530914306641, + "learning_rate": 5.394456289978679e-06, + "loss": 0.6383, + "step": 506 + }, + { + "epoch": 0.1622226962457338, + "grad_norm": 0.6581939458847046, + "learning_rate": 5.405117270788913e-06, + "loss": 0.5816, + "step": 507 + }, + { + "epoch": 0.16254266211604096, + "grad_norm": 0.6166406273841858, + "learning_rate": 5.415778251599148e-06, + "loss": 0.6065, + "step": 508 + }, + { + "epoch": 0.16286262798634812, + "grad_norm": 0.7130981683731079, + "learning_rate": 5.4264392324093815e-06, + "loss": 0.6438, + "step": 509 + }, + { + "epoch": 0.1631825938566553, + "grad_norm": 0.7229422926902771, + "learning_rate": 5.437100213219617e-06, + "loss": 0.6152, + "step": 510 + }, + { + "epoch": 0.16350255972696245, + "grad_norm": 0.6925520896911621, + "learning_rate": 5.447761194029851e-06, + "loss": 0.6183, + "step": 511 + }, + { + "epoch": 0.16382252559726962, + "grad_norm": 0.7318209409713745, + "learning_rate": 5.458422174840086e-06, + "loss": 0.6151, + "step": 512 + }, + { + "epoch": 0.16414249146757678, + "grad_norm": 0.7651650905609131, + "learning_rate": 5.4690831556503196e-06, + "loss": 0.6298, + "step": 513 + }, + { + "epoch": 0.16446245733788395, + "grad_norm": 0.7928692698478699, + "learning_rate": 5.479744136460555e-06, + "loss": 0.6212, + "step": 514 + }, + { + "epoch": 0.16478242320819111, + "grad_norm": 0.7141736745834351, + "learning_rate": 5.4904051172707895e-06, + "loss": 0.6023, + "step": 515 + }, + { + "epoch": 0.1651023890784983, + "grad_norm": 0.6715232133865356, + "learning_rate": 5.501066098081024e-06, + "loss": 0.6173, + "step": 516 + }, + { + "epoch": 0.16542235494880547, + "grad_norm": 0.6879045963287354, + "learning_rate": 5.511727078891258e-06, + "loss": 0.6054, + "step": 517 + }, + { + "epoch": 0.16574232081911264, + "grad_norm": 0.7811073660850525, + "learning_rate": 5.522388059701493e-06, + "loss": 0.5491, + "step": 518 + }, + { + "epoch": 0.1660622866894198, + "grad_norm": 0.6594018340110779, + "learning_rate": 5.5330490405117276e-06, + "loss": 0.5957, + "step": 519 + }, + { + "epoch": 0.16638225255972697, + "grad_norm": 0.6076728701591492, + "learning_rate": 5.543710021321962e-06, + "loss": 0.5769, + "step": 520 + }, + { + "epoch": 0.16670221843003413, + "grad_norm": 0.6521137952804565, + "learning_rate": 5.554371002132196e-06, + "loss": 0.6136, + "step": 521 + }, + { + "epoch": 0.1670221843003413, + "grad_norm": 0.6845177412033081, + "learning_rate": 5.565031982942431e-06, + "loss": 0.5994, + "step": 522 + }, + { + "epoch": 0.16734215017064846, + "grad_norm": 0.690648078918457, + "learning_rate": 5.575692963752666e-06, + "loss": 0.6106, + "step": 523 + }, + { + "epoch": 0.16766211604095563, + "grad_norm": 0.6770439743995667, + "learning_rate": 5.5863539445629e-06, + "loss": 0.5759, + "step": 524 + }, + { + "epoch": 0.1679820819112628, + "grad_norm": 0.7323155403137207, + "learning_rate": 5.597014925373134e-06, + "loss": 0.6161, + "step": 525 + }, + { + "epoch": 0.16830204778156996, + "grad_norm": 0.6350544691085815, + "learning_rate": 5.607675906183369e-06, + "loss": 0.5853, + "step": 526 + }, + { + "epoch": 0.16862201365187712, + "grad_norm": 0.6570029854774475, + "learning_rate": 5.618336886993604e-06, + "loss": 0.6346, + "step": 527 + }, + { + "epoch": 0.1689419795221843, + "grad_norm": 0.7343910932540894, + "learning_rate": 5.628997867803838e-06, + "loss": 0.6016, + "step": 528 + }, + { + "epoch": 0.16926194539249148, + "grad_norm": 0.6532744765281677, + "learning_rate": 5.639658848614073e-06, + "loss": 0.6323, + "step": 529 + }, + { + "epoch": 0.16958191126279865, + "grad_norm": 0.617245078086853, + "learning_rate": 5.650319829424308e-06, + "loss": 0.6216, + "step": 530 + }, + { + "epoch": 0.1699018771331058, + "grad_norm": 0.6686919927597046, + "learning_rate": 5.660980810234542e-06, + "loss": 0.5816, + "step": 531 + }, + { + "epoch": 0.17022184300341298, + "grad_norm": 0.6508616805076599, + "learning_rate": 5.671641791044776e-06, + "loss": 0.6232, + "step": 532 + }, + { + "epoch": 0.17054180887372014, + "grad_norm": 0.6479763984680176, + "learning_rate": 5.682302771855012e-06, + "loss": 0.6242, + "step": 533 + }, + { + "epoch": 0.1708617747440273, + "grad_norm": 0.6898424029350281, + "learning_rate": 5.692963752665246e-06, + "loss": 0.6153, + "step": 534 + }, + { + "epoch": 0.17118174061433447, + "grad_norm": 0.631033718585968, + "learning_rate": 5.70362473347548e-06, + "loss": 0.5956, + "step": 535 + }, + { + "epoch": 0.17150170648464164, + "grad_norm": 0.7228789329528809, + "learning_rate": 5.7142857142857145e-06, + "loss": 0.6204, + "step": 536 + }, + { + "epoch": 0.1718216723549488, + "grad_norm": 0.7352203726768494, + "learning_rate": 5.72494669509595e-06, + "loss": 0.6203, + "step": 537 + }, + { + "epoch": 0.17214163822525597, + "grad_norm": 0.6712698936462402, + "learning_rate": 5.735607675906184e-06, + "loss": 0.5926, + "step": 538 + }, + { + "epoch": 0.17246160409556313, + "grad_norm": 0.7073250412940979, + "learning_rate": 5.746268656716418e-06, + "loss": 0.616, + "step": 539 + }, + { + "epoch": 0.1727815699658703, + "grad_norm": 0.6370400786399841, + "learning_rate": 5.756929637526653e-06, + "loss": 0.576, + "step": 540 + }, + { + "epoch": 0.17310153583617746, + "grad_norm": 0.8138288259506226, + "learning_rate": 5.767590618336888e-06, + "loss": 0.5674, + "step": 541 + }, + { + "epoch": 0.17342150170648465, + "grad_norm": 0.6876887083053589, + "learning_rate": 5.7782515991471225e-06, + "loss": 0.6106, + "step": 542 + }, + { + "epoch": 0.17374146757679182, + "grad_norm": 0.6476050615310669, + "learning_rate": 5.788912579957356e-06, + "loss": 0.6193, + "step": 543 + }, + { + "epoch": 0.17406143344709898, + "grad_norm": 0.6442128419876099, + "learning_rate": 5.799573560767591e-06, + "loss": 0.5954, + "step": 544 + }, + { + "epoch": 0.17438139931740615, + "grad_norm": 0.7466898560523987, + "learning_rate": 5.810234541577826e-06, + "loss": 0.6207, + "step": 545 + }, + { + "epoch": 0.17470136518771331, + "grad_norm": 0.8017808794975281, + "learning_rate": 5.820895522388061e-06, + "loss": 0.5983, + "step": 546 + }, + { + "epoch": 0.17502133105802048, + "grad_norm": 0.6698926091194153, + "learning_rate": 5.831556503198294e-06, + "loss": 0.6028, + "step": 547 + }, + { + "epoch": 0.17534129692832764, + "grad_norm": 0.6686583757400513, + "learning_rate": 5.842217484008529e-06, + "loss": 0.6043, + "step": 548 + }, + { + "epoch": 0.1756612627986348, + "grad_norm": 0.7330739498138428, + "learning_rate": 5.852878464818764e-06, + "loss": 0.5881, + "step": 549 + }, + { + "epoch": 0.17598122866894197, + "grad_norm": 0.7499303221702576, + "learning_rate": 5.863539445628999e-06, + "loss": 0.5953, + "step": 550 + }, + { + "epoch": 0.17630119453924914, + "grad_norm": 0.6294394135475159, + "learning_rate": 5.874200426439232e-06, + "loss": 0.591, + "step": 551 + }, + { + "epoch": 0.1766211604095563, + "grad_norm": 0.6672900319099426, + "learning_rate": 5.884861407249467e-06, + "loss": 0.5864, + "step": 552 + }, + { + "epoch": 0.17694112627986347, + "grad_norm": 0.7012150883674622, + "learning_rate": 5.895522388059702e-06, + "loss": 0.5784, + "step": 553 + }, + { + "epoch": 0.17726109215017063, + "grad_norm": 0.6924875378608704, + "learning_rate": 5.906183368869937e-06, + "loss": 0.6435, + "step": 554 + }, + { + "epoch": 0.17758105802047783, + "grad_norm": 0.8030469417572021, + "learning_rate": 5.9168443496801705e-06, + "loss": 0.6521, + "step": 555 + }, + { + "epoch": 0.177901023890785, + "grad_norm": 0.6786168813705444, + "learning_rate": 5.927505330490405e-06, + "loss": 0.5929, + "step": 556 + }, + { + "epoch": 0.17822098976109216, + "grad_norm": 0.6793124079704285, + "learning_rate": 5.93816631130064e-06, + "loss": 0.6022, + "step": 557 + }, + { + "epoch": 0.17854095563139932, + "grad_norm": 0.7449572086334229, + "learning_rate": 5.948827292110875e-06, + "loss": 0.6223, + "step": 558 + }, + { + "epoch": 0.1788609215017065, + "grad_norm": 0.6794044971466064, + "learning_rate": 5.959488272921109e-06, + "loss": 0.555, + "step": 559 + }, + { + "epoch": 0.17918088737201365, + "grad_norm": 0.6000784635543823, + "learning_rate": 5.970149253731343e-06, + "loss": 0.5957, + "step": 560 + }, + { + "epoch": 0.17950085324232082, + "grad_norm": 0.7043874859809875, + "learning_rate": 5.9808102345415785e-06, + "loss": 0.5595, + "step": 561 + }, + { + "epoch": 0.17982081911262798, + "grad_norm": 0.7883280515670776, + "learning_rate": 5.991471215351813e-06, + "loss": 0.5963, + "step": 562 + }, + { + "epoch": 0.18014078498293515, + "grad_norm": 0.6450311541557312, + "learning_rate": 6.002132196162047e-06, + "loss": 0.6307, + "step": 563 + }, + { + "epoch": 0.1804607508532423, + "grad_norm": 0.6373298764228821, + "learning_rate": 6.012793176972282e-06, + "loss": 0.5746, + "step": 564 + }, + { + "epoch": 0.18078071672354948, + "grad_norm": 0.8034327030181885, + "learning_rate": 6.023454157782517e-06, + "loss": 0.6252, + "step": 565 + }, + { + "epoch": 0.18110068259385664, + "grad_norm": 0.6212670207023621, + "learning_rate": 6.034115138592751e-06, + "loss": 0.627, + "step": 566 + }, + { + "epoch": 0.18142064846416384, + "grad_norm": 0.7424978017807007, + "learning_rate": 6.044776119402986e-06, + "loss": 0.6091, + "step": 567 + }, + { + "epoch": 0.181740614334471, + "grad_norm": 0.6831108331680298, + "learning_rate": 6.055437100213221e-06, + "loss": 0.5944, + "step": 568 + }, + { + "epoch": 0.18206058020477817, + "grad_norm": 0.6674385666847229, + "learning_rate": 6.066098081023455e-06, + "loss": 0.6182, + "step": 569 + }, + { + "epoch": 0.18238054607508533, + "grad_norm": 0.6555317044258118, + "learning_rate": 6.076759061833689e-06, + "loss": 0.614, + "step": 570 + }, + { + "epoch": 0.1827005119453925, + "grad_norm": 0.6911100745201111, + "learning_rate": 6.087420042643924e-06, + "loss": 0.6122, + "step": 571 + }, + { + "epoch": 0.18302047781569966, + "grad_norm": 0.6371263861656189, + "learning_rate": 6.098081023454159e-06, + "loss": 0.6368, + "step": 572 + }, + { + "epoch": 0.18334044368600683, + "grad_norm": 0.6716217398643494, + "learning_rate": 6.108742004264393e-06, + "loss": 0.5839, + "step": 573 + }, + { + "epoch": 0.183660409556314, + "grad_norm": 0.7504799962043762, + "learning_rate": 6.119402985074627e-06, + "loss": 0.6295, + "step": 574 + }, + { + "epoch": 0.18398037542662116, + "grad_norm": 0.678804337978363, + "learning_rate": 6.130063965884862e-06, + "loss": 0.6007, + "step": 575 + }, + { + "epoch": 0.18430034129692832, + "grad_norm": 0.7213124632835388, + "learning_rate": 6.140724946695097e-06, + "loss": 0.6269, + "step": 576 + }, + { + "epoch": 0.1846203071672355, + "grad_norm": 0.6804861426353455, + "learning_rate": 6.151385927505331e-06, + "loss": 0.5593, + "step": 577 + }, + { + "epoch": 0.18494027303754265, + "grad_norm": 0.744810163974762, + "learning_rate": 6.1620469083155655e-06, + "loss": 0.6072, + "step": 578 + }, + { + "epoch": 0.18526023890784982, + "grad_norm": 0.6842663288116455, + "learning_rate": 6.1727078891258e-06, + "loss": 0.6033, + "step": 579 + }, + { + "epoch": 0.185580204778157, + "grad_norm": 0.7466291189193726, + "learning_rate": 6.183368869936035e-06, + "loss": 0.6218, + "step": 580 + }, + { + "epoch": 0.18590017064846417, + "grad_norm": 0.7174020409584045, + "learning_rate": 6.194029850746269e-06, + "loss": 0.6095, + "step": 581 + }, + { + "epoch": 0.18622013651877134, + "grad_norm": 0.7337679862976074, + "learning_rate": 6.2046908315565036e-06, + "loss": 0.6252, + "step": 582 + }, + { + "epoch": 0.1865401023890785, + "grad_norm": 0.6675230264663696, + "learning_rate": 6.215351812366738e-06, + "loss": 0.6061, + "step": 583 + }, + { + "epoch": 0.18686006825938567, + "grad_norm": 0.7611439228057861, + "learning_rate": 6.2260127931769735e-06, + "loss": 0.6308, + "step": 584 + }, + { + "epoch": 0.18718003412969283, + "grad_norm": 0.7200325131416321, + "learning_rate": 6.236673773987207e-06, + "loss": 0.6027, + "step": 585 + }, + { + "epoch": 0.1875, + "grad_norm": 0.6008200645446777, + "learning_rate": 6.247334754797442e-06, + "loss": 0.5991, + "step": 586 + }, + { + "epoch": 0.18781996587030717, + "grad_norm": 0.7074070572853088, + "learning_rate": 6.257995735607676e-06, + "loss": 0.5954, + "step": 587 + }, + { + "epoch": 0.18813993174061433, + "grad_norm": 0.6179446578025818, + "learning_rate": 6.2686567164179116e-06, + "loss": 0.5912, + "step": 588 + }, + { + "epoch": 0.1884598976109215, + "grad_norm": 0.6689445376396179, + "learning_rate": 6.279317697228145e-06, + "loss": 0.6092, + "step": 589 + }, + { + "epoch": 0.18877986348122866, + "grad_norm": 0.688685417175293, + "learning_rate": 6.28997867803838e-06, + "loss": 0.6181, + "step": 590 + }, + { + "epoch": 0.18909982935153583, + "grad_norm": 0.8675934672355652, + "learning_rate": 6.300639658848614e-06, + "loss": 0.6463, + "step": 591 + }, + { + "epoch": 0.189419795221843, + "grad_norm": 0.714806854724884, + "learning_rate": 6.31130063965885e-06, + "loss": 0.6366, + "step": 592 + }, + { + "epoch": 0.18973976109215018, + "grad_norm": 0.7190007567405701, + "learning_rate": 6.321961620469083e-06, + "loss": 0.6041, + "step": 593 + }, + { + "epoch": 0.19005972696245735, + "grad_norm": 0.7643619179725647, + "learning_rate": 6.332622601279318e-06, + "loss": 0.6026, + "step": 594 + }, + { + "epoch": 0.1903796928327645, + "grad_norm": 0.6657800078392029, + "learning_rate": 6.343283582089553e-06, + "loss": 0.6191, + "step": 595 + }, + { + "epoch": 0.19069965870307168, + "grad_norm": 0.6771308779716492, + "learning_rate": 6.353944562899788e-06, + "loss": 0.6488, + "step": 596 + }, + { + "epoch": 0.19101962457337884, + "grad_norm": 0.7485832571983337, + "learning_rate": 6.3646055437100215e-06, + "loss": 0.6351, + "step": 597 + }, + { + "epoch": 0.191339590443686, + "grad_norm": 0.6575362682342529, + "learning_rate": 6.375266524520256e-06, + "loss": 0.5739, + "step": 598 + }, + { + "epoch": 0.19165955631399317, + "grad_norm": 0.6349213123321533, + "learning_rate": 6.385927505330491e-06, + "loss": 0.5851, + "step": 599 + }, + { + "epoch": 0.19197952218430034, + "grad_norm": 0.6870304942131042, + "learning_rate": 6.396588486140726e-06, + "loss": 0.6312, + "step": 600 + }, + { + "epoch": 0.1922994880546075, + "grad_norm": 0.5886594653129578, + "learning_rate": 6.4072494669509596e-06, + "loss": 0.6055, + "step": 601 + }, + { + "epoch": 0.19261945392491467, + "grad_norm": 0.7129337191581726, + "learning_rate": 6.417910447761194e-06, + "loss": 0.558, + "step": 602 + }, + { + "epoch": 0.19293941979522183, + "grad_norm": 0.6680113077163696, + "learning_rate": 6.4285714285714295e-06, + "loss": 0.5845, + "step": 603 + }, + { + "epoch": 0.193259385665529, + "grad_norm": 0.7426353096961975, + "learning_rate": 6.439232409381664e-06, + "loss": 0.6856, + "step": 604 + }, + { + "epoch": 0.19357935153583616, + "grad_norm": 0.7182149291038513, + "learning_rate": 6.449893390191898e-06, + "loss": 0.5754, + "step": 605 + }, + { + "epoch": 0.19389931740614336, + "grad_norm": 0.6413835883140564, + "learning_rate": 6.460554371002132e-06, + "loss": 0.62, + "step": 606 + }, + { + "epoch": 0.19421928327645052, + "grad_norm": 0.7080693244934082, + "learning_rate": 6.4712153518123676e-06, + "loss": 0.6363, + "step": 607 + }, + { + "epoch": 0.1945392491467577, + "grad_norm": 0.6945396661758423, + "learning_rate": 6.481876332622602e-06, + "loss": 0.6092, + "step": 608 + }, + { + "epoch": 0.19485921501706485, + "grad_norm": 0.7453746199607849, + "learning_rate": 6.492537313432837e-06, + "loss": 0.6184, + "step": 609 + }, + { + "epoch": 0.19517918088737202, + "grad_norm": 0.7010712623596191, + "learning_rate": 6.50319829424307e-06, + "loss": 0.6062, + "step": 610 + }, + { + "epoch": 0.19549914675767918, + "grad_norm": 0.626008927822113, + "learning_rate": 6.513859275053306e-06, + "loss": 0.6378, + "step": 611 + }, + { + "epoch": 0.19581911262798635, + "grad_norm": 0.6873940229415894, + "learning_rate": 6.52452025586354e-06, + "loss": 0.6199, + "step": 612 + }, + { + "epoch": 0.1961390784982935, + "grad_norm": 0.7339004278182983, + "learning_rate": 6.535181236673775e-06, + "loss": 0.609, + "step": 613 + }, + { + "epoch": 0.19645904436860068, + "grad_norm": 0.7159809470176697, + "learning_rate": 6.545842217484008e-06, + "loss": 0.6354, + "step": 614 + }, + { + "epoch": 0.19677901023890784, + "grad_norm": 0.7771121263504028, + "learning_rate": 6.556503198294244e-06, + "loss": 0.6504, + "step": 615 + }, + { + "epoch": 0.197098976109215, + "grad_norm": 0.7306795120239258, + "learning_rate": 6.567164179104478e-06, + "loss": 0.6061, + "step": 616 + }, + { + "epoch": 0.19741894197952217, + "grad_norm": 0.651380181312561, + "learning_rate": 6.577825159914713e-06, + "loss": 0.5689, + "step": 617 + }, + { + "epoch": 0.19773890784982937, + "grad_norm": 0.6777538061141968, + "learning_rate": 6.5884861407249465e-06, + "loss": 0.5912, + "step": 618 + }, + { + "epoch": 0.19805887372013653, + "grad_norm": 0.6833059191703796, + "learning_rate": 6.599147121535182e-06, + "loss": 0.5937, + "step": 619 + }, + { + "epoch": 0.1983788395904437, + "grad_norm": 0.6252380609512329, + "learning_rate": 6.609808102345416e-06, + "loss": 0.5579, + "step": 620 + }, + { + "epoch": 0.19869880546075086, + "grad_norm": 0.7200431823730469, + "learning_rate": 6.620469083155651e-06, + "loss": 0.579, + "step": 621 + }, + { + "epoch": 0.19901877133105803, + "grad_norm": 0.6220011115074158, + "learning_rate": 6.631130063965885e-06, + "loss": 0.5765, + "step": 622 + }, + { + "epoch": 0.1993387372013652, + "grad_norm": 0.6783512234687805, + "learning_rate": 6.64179104477612e-06, + "loss": 0.5552, + "step": 623 + }, + { + "epoch": 0.19965870307167236, + "grad_norm": 0.6065102815628052, + "learning_rate": 6.6524520255863545e-06, + "loss": 0.6104, + "step": 624 + }, + { + "epoch": 0.19997866894197952, + "grad_norm": 0.789917528629303, + "learning_rate": 6.663113006396589e-06, + "loss": 0.5819, + "step": 625 + }, + { + "epoch": 0.20029863481228669, + "grad_norm": 0.7503900527954102, + "learning_rate": 6.673773987206824e-06, + "loss": 0.6099, + "step": 626 + }, + { + "epoch": 0.20061860068259385, + "grad_norm": 0.7439480423927307, + "learning_rate": 6.684434968017058e-06, + "loss": 0.6757, + "step": 627 + }, + { + "epoch": 0.20093856655290102, + "grad_norm": 0.7218211889266968, + "learning_rate": 6.695095948827293e-06, + "loss": 0.5991, + "step": 628 + }, + { + "epoch": 0.20125853242320818, + "grad_norm": 0.7553074359893799, + "learning_rate": 6.705756929637527e-06, + "loss": 0.6263, + "step": 629 + }, + { + "epoch": 0.20157849829351535, + "grad_norm": 0.7821144461631775, + "learning_rate": 6.7164179104477625e-06, + "loss": 0.6385, + "step": 630 + }, + { + "epoch": 0.20189846416382254, + "grad_norm": 0.7380728125572205, + "learning_rate": 6.727078891257996e-06, + "loss": 0.6083, + "step": 631 + }, + { + "epoch": 0.2022184300341297, + "grad_norm": 0.6631566286087036, + "learning_rate": 6.737739872068231e-06, + "loss": 0.5956, + "step": 632 + }, + { + "epoch": 0.20253839590443687, + "grad_norm": 0.7742174863815308, + "learning_rate": 6.748400852878465e-06, + "loss": 0.5878, + "step": 633 + }, + { + "epoch": 0.20285836177474403, + "grad_norm": 0.7614405751228333, + "learning_rate": 6.759061833688701e-06, + "loss": 0.5883, + "step": 634 + }, + { + "epoch": 0.2031783276450512, + "grad_norm": 0.6650667786598206, + "learning_rate": 6.769722814498934e-06, + "loss": 0.5952, + "step": 635 + }, + { + "epoch": 0.20349829351535836, + "grad_norm": 0.710259735584259, + "learning_rate": 6.780383795309169e-06, + "loss": 0.5938, + "step": 636 + }, + { + "epoch": 0.20381825938566553, + "grad_norm": 0.6996580362319946, + "learning_rate": 6.791044776119403e-06, + "loss": 0.6285, + "step": 637 + }, + { + "epoch": 0.2041382252559727, + "grad_norm": 0.7071582078933716, + "learning_rate": 6.801705756929639e-06, + "loss": 0.5862, + "step": 638 + }, + { + "epoch": 0.20445819112627986, + "grad_norm": 0.7259476780891418, + "learning_rate": 6.812366737739872e-06, + "loss": 0.608, + "step": 639 + }, + { + "epoch": 0.20477815699658702, + "grad_norm": 0.7430290579795837, + "learning_rate": 6.823027718550107e-06, + "loss": 0.6497, + "step": 640 + }, + { + "epoch": 0.2050981228668942, + "grad_norm": 0.7619306445121765, + "learning_rate": 6.8336886993603415e-06, + "loss": 0.6543, + "step": 641 + }, + { + "epoch": 0.20541808873720135, + "grad_norm": 0.8050456643104553, + "learning_rate": 6.844349680170577e-06, + "loss": 0.6278, + "step": 642 + }, + { + "epoch": 0.20573805460750852, + "grad_norm": 0.7514400482177734, + "learning_rate": 6.8550106609808105e-06, + "loss": 0.6198, + "step": 643 + }, + { + "epoch": 0.2060580204778157, + "grad_norm": 0.7751818299293518, + "learning_rate": 6.865671641791045e-06, + "loss": 0.6206, + "step": 644 + }, + { + "epoch": 0.20637798634812288, + "grad_norm": 0.7462335824966431, + "learning_rate": 6.8763326226012796e-06, + "loss": 0.6169, + "step": 645 + }, + { + "epoch": 0.20669795221843004, + "grad_norm": 0.6527459621429443, + "learning_rate": 6.886993603411515e-06, + "loss": 0.6172, + "step": 646 + }, + { + "epoch": 0.2070179180887372, + "grad_norm": 0.7508776187896729, + "learning_rate": 6.8976545842217495e-06, + "loss": 0.5796, + "step": 647 + }, + { + "epoch": 0.20733788395904437, + "grad_norm": 0.6763598918914795, + "learning_rate": 6.908315565031983e-06, + "loss": 0.6227, + "step": 648 + }, + { + "epoch": 0.20765784982935154, + "grad_norm": 0.7581509351730347, + "learning_rate": 6.918976545842218e-06, + "loss": 0.6063, + "step": 649 + }, + { + "epoch": 0.2079778156996587, + "grad_norm": 0.7387362122535706, + "learning_rate": 6.929637526652453e-06, + "loss": 0.5696, + "step": 650 + }, + { + "epoch": 0.20829778156996587, + "grad_norm": 0.7688075304031372, + "learning_rate": 6.9402985074626876e-06, + "loss": 0.5871, + "step": 651 + }, + { + "epoch": 0.20861774744027303, + "grad_norm": 0.6743102073669434, + "learning_rate": 6.950959488272921e-06, + "loss": 0.61, + "step": 652 + }, + { + "epoch": 0.2089377133105802, + "grad_norm": 0.7545280456542969, + "learning_rate": 6.961620469083156e-06, + "loss": 0.614, + "step": 653 + }, + { + "epoch": 0.20925767918088736, + "grad_norm": 0.694983720779419, + "learning_rate": 6.972281449893391e-06, + "loss": 0.6409, + "step": 654 + }, + { + "epoch": 0.20957764505119453, + "grad_norm": 0.7406476736068726, + "learning_rate": 6.982942430703626e-06, + "loss": 0.6163, + "step": 655 + }, + { + "epoch": 0.2098976109215017, + "grad_norm": 0.6860409379005432, + "learning_rate": 6.993603411513859e-06, + "loss": 0.5756, + "step": 656 + }, + { + "epoch": 0.21021757679180889, + "grad_norm": 0.662331759929657, + "learning_rate": 7.004264392324095e-06, + "loss": 0.644, + "step": 657 + }, + { + "epoch": 0.21053754266211605, + "grad_norm": 0.7123412489891052, + "learning_rate": 7.014925373134329e-06, + "loss": 0.6, + "step": 658 + }, + { + "epoch": 0.21085750853242322, + "grad_norm": 0.7488113641738892, + "learning_rate": 7.025586353944564e-06, + "loss": 0.5972, + "step": 659 + }, + { + "epoch": 0.21117747440273038, + "grad_norm": 0.7227650880813599, + "learning_rate": 7.0362473347547975e-06, + "loss": 0.622, + "step": 660 + }, + { + "epoch": 0.21149744027303755, + "grad_norm": 0.7823781967163086, + "learning_rate": 7.046908315565033e-06, + "loss": 0.6404, + "step": 661 + }, + { + "epoch": 0.2118174061433447, + "grad_norm": 0.6650978922843933, + "learning_rate": 7.057569296375267e-06, + "loss": 0.6049, + "step": 662 + }, + { + "epoch": 0.21213737201365188, + "grad_norm": 0.6199343204498291, + "learning_rate": 7.068230277185502e-06, + "loss": 0.5808, + "step": 663 + }, + { + "epoch": 0.21245733788395904, + "grad_norm": 0.7609500885009766, + "learning_rate": 7.0788912579957356e-06, + "loss": 0.6099, + "step": 664 + }, + { + "epoch": 0.2127773037542662, + "grad_norm": 0.6560154557228088, + "learning_rate": 7.089552238805971e-06, + "loss": 0.59, + "step": 665 + }, + { + "epoch": 0.21309726962457337, + "grad_norm": 0.6989269852638245, + "learning_rate": 7.1002132196162055e-06, + "loss": 0.6321, + "step": 666 + }, + { + "epoch": 0.21341723549488054, + "grad_norm": 0.7098025679588318, + "learning_rate": 7.11087420042644e-06, + "loss": 0.5795, + "step": 667 + }, + { + "epoch": 0.2137372013651877, + "grad_norm": 0.6742767691612244, + "learning_rate": 7.121535181236674e-06, + "loss": 0.6138, + "step": 668 + }, + { + "epoch": 0.2140571672354949, + "grad_norm": 0.7612442970275879, + "learning_rate": 7.132196162046909e-06, + "loss": 0.6684, + "step": 669 + }, + { + "epoch": 0.21437713310580206, + "grad_norm": 0.7280893325805664, + "learning_rate": 7.1428571428571436e-06, + "loss": 0.6422, + "step": 670 + }, + { + "epoch": 0.21469709897610922, + "grad_norm": 0.6309712529182434, + "learning_rate": 7.153518123667378e-06, + "loss": 0.5463, + "step": 671 + }, + { + "epoch": 0.2150170648464164, + "grad_norm": 0.6505151987075806, + "learning_rate": 7.164179104477612e-06, + "loss": 0.6028, + "step": 672 + }, + { + "epoch": 0.21533703071672355, + "grad_norm": 0.6818749308586121, + "learning_rate": 7.174840085287847e-06, + "loss": 0.6038, + "step": 673 + }, + { + "epoch": 0.21565699658703072, + "grad_norm": 0.6691836714744568, + "learning_rate": 7.185501066098082e-06, + "loss": 0.6081, + "step": 674 + }, + { + "epoch": 0.21597696245733788, + "grad_norm": 0.6752950549125671, + "learning_rate": 7.196162046908316e-06, + "loss": 0.6063, + "step": 675 + }, + { + "epoch": 0.21629692832764505, + "grad_norm": 0.7235035300254822, + "learning_rate": 7.20682302771855e-06, + "loss": 0.6094, + "step": 676 + }, + { + "epoch": 0.21661689419795221, + "grad_norm": 0.6782038807868958, + "learning_rate": 7.217484008528785e-06, + "loss": 0.6016, + "step": 677 + }, + { + "epoch": 0.21693686006825938, + "grad_norm": 0.7983608841896057, + "learning_rate": 7.22814498933902e-06, + "loss": 0.6242, + "step": 678 + }, + { + "epoch": 0.21725682593856654, + "grad_norm": 0.8160451650619507, + "learning_rate": 7.238805970149254e-06, + "loss": 0.6348, + "step": 679 + }, + { + "epoch": 0.2175767918088737, + "grad_norm": 0.808131754398346, + "learning_rate": 7.249466950959488e-06, + "loss": 0.6194, + "step": 680 + }, + { + "epoch": 0.21789675767918087, + "grad_norm": 0.8144751787185669, + "learning_rate": 7.260127931769723e-06, + "loss": 0.6214, + "step": 681 + }, + { + "epoch": 0.21821672354948807, + "grad_norm": 0.8087030053138733, + "learning_rate": 7.270788912579958e-06, + "loss": 0.6095, + "step": 682 + }, + { + "epoch": 0.21853668941979523, + "grad_norm": 0.8073561191558838, + "learning_rate": 7.281449893390192e-06, + "loss": 0.6004, + "step": 683 + }, + { + "epoch": 0.2188566552901024, + "grad_norm": 0.7714781761169434, + "learning_rate": 7.292110874200427e-06, + "loss": 0.5931, + "step": 684 + }, + { + "epoch": 0.21917662116040956, + "grad_norm": 0.7731906771659851, + "learning_rate": 7.302771855010662e-06, + "loss": 0.5817, + "step": 685 + }, + { + "epoch": 0.21949658703071673, + "grad_norm": 0.8933697938919067, + "learning_rate": 7.313432835820896e-06, + "loss": 0.6729, + "step": 686 + }, + { + "epoch": 0.2198165529010239, + "grad_norm": 0.8103617429733276, + "learning_rate": 7.3240938166311305e-06, + "loss": 0.6505, + "step": 687 + }, + { + "epoch": 0.22013651877133106, + "grad_norm": 0.6836149096488953, + "learning_rate": 7.334754797441366e-06, + "loss": 0.5906, + "step": 688 + }, + { + "epoch": 0.22045648464163822, + "grad_norm": 0.8601536750793457, + "learning_rate": 7.3454157782516e-06, + "loss": 0.6226, + "step": 689 + }, + { + "epoch": 0.2207764505119454, + "grad_norm": 0.8050239086151123, + "learning_rate": 7.356076759061834e-06, + "loss": 0.582, + "step": 690 + }, + { + "epoch": 0.22109641638225255, + "grad_norm": 0.7658442258834839, + "learning_rate": 7.366737739872069e-06, + "loss": 0.5756, + "step": 691 + }, + { + "epoch": 0.22141638225255972, + "grad_norm": 0.9899559020996094, + "learning_rate": 7.377398720682304e-06, + "loss": 0.661, + "step": 692 + }, + { + "epoch": 0.22173634812286688, + "grad_norm": 0.663729190826416, + "learning_rate": 7.3880597014925385e-06, + "loss": 0.5754, + "step": 693 + }, + { + "epoch": 0.22205631399317405, + "grad_norm": 0.6831468939781189, + "learning_rate": 7.398720682302772e-06, + "loss": 0.6276, + "step": 694 + }, + { + "epoch": 0.22237627986348124, + "grad_norm": 0.7177518606185913, + "learning_rate": 7.409381663113007e-06, + "loss": 0.5654, + "step": 695 + }, + { + "epoch": 0.2226962457337884, + "grad_norm": 0.8334975838661194, + "learning_rate": 7.420042643923242e-06, + "loss": 0.5664, + "step": 696 + }, + { + "epoch": 0.22301621160409557, + "grad_norm": 0.7577218413352966, + "learning_rate": 7.430703624733477e-06, + "loss": 0.6004, + "step": 697 + }, + { + "epoch": 0.22333617747440274, + "grad_norm": 0.8322723507881165, + "learning_rate": 7.44136460554371e-06, + "loss": 0.6105, + "step": 698 + }, + { + "epoch": 0.2236561433447099, + "grad_norm": 0.8639805316925049, + "learning_rate": 7.452025586353945e-06, + "loss": 0.6352, + "step": 699 + }, + { + "epoch": 0.22397610921501707, + "grad_norm": 0.7729711532592773, + "learning_rate": 7.46268656716418e-06, + "loss": 0.6066, + "step": 700 + }, + { + "epoch": 0.22429607508532423, + "grad_norm": 0.906555712223053, + "learning_rate": 7.473347547974415e-06, + "loss": 0.6466, + "step": 701 + }, + { + "epoch": 0.2246160409556314, + "grad_norm": 0.8130903244018555, + "learning_rate": 7.484008528784648e-06, + "loss": 0.5961, + "step": 702 + }, + { + "epoch": 0.22493600682593856, + "grad_norm": 0.8586740493774414, + "learning_rate": 7.494669509594883e-06, + "loss": 0.5668, + "step": 703 + }, + { + "epoch": 0.22525597269624573, + "grad_norm": 0.952433705329895, + "learning_rate": 7.505330490405118e-06, + "loss": 0.5882, + "step": 704 + }, + { + "epoch": 0.2255759385665529, + "grad_norm": 0.7945285439491272, + "learning_rate": 7.515991471215353e-06, + "loss": 0.6012, + "step": 705 + }, + { + "epoch": 0.22589590443686006, + "grad_norm": 0.7891091108322144, + "learning_rate": 7.5266524520255865e-06, + "loss": 0.6197, + "step": 706 + }, + { + "epoch": 0.22621587030716722, + "grad_norm": 0.7546164393424988, + "learning_rate": 7.537313432835821e-06, + "loss": 0.5517, + "step": 707 + }, + { + "epoch": 0.22653583617747441, + "grad_norm": 0.7216648459434509, + "learning_rate": 7.547974413646056e-06, + "loss": 0.6037, + "step": 708 + }, + { + "epoch": 0.22685580204778158, + "grad_norm": 0.7756451368331909, + "learning_rate": 7.558635394456291e-06, + "loss": 0.6059, + "step": 709 + }, + { + "epoch": 0.22717576791808874, + "grad_norm": 0.6714305281639099, + "learning_rate": 7.569296375266525e-06, + "loss": 0.6041, + "step": 710 + }, + { + "epoch": 0.2274957337883959, + "grad_norm": 0.698634147644043, + "learning_rate": 7.579957356076759e-06, + "loss": 0.6272, + "step": 711 + }, + { + "epoch": 0.22781569965870307, + "grad_norm": 0.7570346593856812, + "learning_rate": 7.5906183368869945e-06, + "loss": 0.5946, + "step": 712 + }, + { + "epoch": 0.22813566552901024, + "grad_norm": 0.7321618795394897, + "learning_rate": 7.601279317697229e-06, + "loss": 0.5984, + "step": 713 + }, + { + "epoch": 0.2284556313993174, + "grad_norm": 0.7243728637695312, + "learning_rate": 7.611940298507463e-06, + "loss": 0.6318, + "step": 714 + }, + { + "epoch": 0.22877559726962457, + "grad_norm": 0.7606075406074524, + "learning_rate": 7.622601279317697e-06, + "loss": 0.577, + "step": 715 + }, + { + "epoch": 0.22909556313993173, + "grad_norm": 0.5844424366950989, + "learning_rate": 7.633262260127933e-06, + "loss": 0.5897, + "step": 716 + }, + { + "epoch": 0.2294155290102389, + "grad_norm": 0.6975888609886169, + "learning_rate": 7.643923240938167e-06, + "loss": 0.5822, + "step": 717 + }, + { + "epoch": 0.22973549488054607, + "grad_norm": 0.7610205411911011, + "learning_rate": 7.654584221748402e-06, + "loss": 0.6291, + "step": 718 + }, + { + "epoch": 0.23005546075085323, + "grad_norm": 0.6593003869056702, + "learning_rate": 7.665245202558636e-06, + "loss": 0.6267, + "step": 719 + }, + { + "epoch": 0.23037542662116042, + "grad_norm": 0.7518547773361206, + "learning_rate": 7.67590618336887e-06, + "loss": 0.6667, + "step": 720 + }, + { + "epoch": 0.2306953924914676, + "grad_norm": 0.6882115006446838, + "learning_rate": 7.686567164179105e-06, + "loss": 0.5997, + "step": 721 + }, + { + "epoch": 0.23101535836177475, + "grad_norm": 0.647657036781311, + "learning_rate": 7.69722814498934e-06, + "loss": 0.622, + "step": 722 + }, + { + "epoch": 0.23133532423208192, + "grad_norm": 0.6777190566062927, + "learning_rate": 7.707889125799574e-06, + "loss": 0.5917, + "step": 723 + }, + { + "epoch": 0.23165529010238908, + "grad_norm": 0.7947143316268921, + "learning_rate": 7.718550106609809e-06, + "loss": 0.5911, + "step": 724 + }, + { + "epoch": 0.23197525597269625, + "grad_norm": 0.6618469953536987, + "learning_rate": 7.729211087420043e-06, + "loss": 0.593, + "step": 725 + }, + { + "epoch": 0.2322952218430034, + "grad_norm": 0.6656261682510376, + "learning_rate": 7.739872068230278e-06, + "loss": 0.6148, + "step": 726 + }, + { + "epoch": 0.23261518771331058, + "grad_norm": 0.7834561467170715, + "learning_rate": 7.750533049040512e-06, + "loss": 0.6014, + "step": 727 + }, + { + "epoch": 0.23293515358361774, + "grad_norm": 0.7411160469055176, + "learning_rate": 7.761194029850747e-06, + "loss": 0.5903, + "step": 728 + }, + { + "epoch": 0.2332551194539249, + "grad_norm": 0.684743344783783, + "learning_rate": 7.771855010660981e-06, + "loss": 0.597, + "step": 729 + }, + { + "epoch": 0.23357508532423207, + "grad_norm": 0.6873225569725037, + "learning_rate": 7.782515991471216e-06, + "loss": 0.5827, + "step": 730 + }, + { + "epoch": 0.23389505119453924, + "grad_norm": 0.6318769454956055, + "learning_rate": 7.79317697228145e-06, + "loss": 0.5915, + "step": 731 + }, + { + "epoch": 0.2342150170648464, + "grad_norm": 0.6759801506996155, + "learning_rate": 7.803837953091685e-06, + "loss": 0.6325, + "step": 732 + }, + { + "epoch": 0.2345349829351536, + "grad_norm": 0.6570876836776733, + "learning_rate": 7.81449893390192e-06, + "loss": 0.616, + "step": 733 + }, + { + "epoch": 0.23485494880546076, + "grad_norm": 0.7084457278251648, + "learning_rate": 7.825159914712154e-06, + "loss": 0.583, + "step": 734 + }, + { + "epoch": 0.23517491467576793, + "grad_norm": 0.7025005221366882, + "learning_rate": 7.835820895522389e-06, + "loss": 0.5998, + "step": 735 + }, + { + "epoch": 0.2354948805460751, + "grad_norm": 0.7843663692474365, + "learning_rate": 7.846481876332623e-06, + "loss": 0.6449, + "step": 736 + }, + { + "epoch": 0.23581484641638226, + "grad_norm": 0.6961464285850525, + "learning_rate": 7.857142857142858e-06, + "loss": 0.6293, + "step": 737 + }, + { + "epoch": 0.23613481228668942, + "grad_norm": 0.6443120837211609, + "learning_rate": 7.867803837953092e-06, + "loss": 0.5757, + "step": 738 + }, + { + "epoch": 0.2364547781569966, + "grad_norm": 0.7188411951065063, + "learning_rate": 7.878464818763327e-06, + "loss": 0.5754, + "step": 739 + }, + { + "epoch": 0.23677474402730375, + "grad_norm": 0.7644008994102478, + "learning_rate": 7.889125799573561e-06, + "loss": 0.6046, + "step": 740 + }, + { + "epoch": 0.23709470989761092, + "grad_norm": 0.7195684313774109, + "learning_rate": 7.899786780383796e-06, + "loss": 0.5944, + "step": 741 + }, + { + "epoch": 0.23741467576791808, + "grad_norm": 0.7057017087936401, + "learning_rate": 7.91044776119403e-06, + "loss": 0.5895, + "step": 742 + }, + { + "epoch": 0.23773464163822525, + "grad_norm": 0.7110957503318787, + "learning_rate": 7.921108742004265e-06, + "loss": 0.6243, + "step": 743 + }, + { + "epoch": 0.2380546075085324, + "grad_norm": 0.6075982451438904, + "learning_rate": 7.9317697228145e-06, + "loss": 0.5831, + "step": 744 + }, + { + "epoch": 0.23837457337883958, + "grad_norm": 0.6880046129226685, + "learning_rate": 7.942430703624734e-06, + "loss": 0.56, + "step": 745 + }, + { + "epoch": 0.23869453924914677, + "grad_norm": 0.7850193381309509, + "learning_rate": 7.953091684434968e-06, + "loss": 0.6023, + "step": 746 + }, + { + "epoch": 0.23901450511945393, + "grad_norm": 0.8188655376434326, + "learning_rate": 7.963752665245203e-06, + "loss": 0.6267, + "step": 747 + }, + { + "epoch": 0.2393344709897611, + "grad_norm": 0.7466493248939514, + "learning_rate": 7.974413646055437e-06, + "loss": 0.6166, + "step": 748 + }, + { + "epoch": 0.23965443686006827, + "grad_norm": 0.6109229922294617, + "learning_rate": 7.985074626865672e-06, + "loss": 0.5878, + "step": 749 + }, + { + "epoch": 0.23997440273037543, + "grad_norm": 0.7116419672966003, + "learning_rate": 7.995735607675907e-06, + "loss": 0.5625, + "step": 750 + }, + { + "epoch": 0.2402943686006826, + "grad_norm": 0.7238773107528687, + "learning_rate": 8.006396588486141e-06, + "loss": 0.6143, + "step": 751 + }, + { + "epoch": 0.24061433447098976, + "grad_norm": 0.611035943031311, + "learning_rate": 8.017057569296376e-06, + "loss": 0.6131, + "step": 752 + }, + { + "epoch": 0.24093430034129693, + "grad_norm": 0.6717967391014099, + "learning_rate": 8.02771855010661e-06, + "loss": 0.5653, + "step": 753 + }, + { + "epoch": 0.2412542662116041, + "grad_norm": 0.6449785828590393, + "learning_rate": 8.038379530916846e-06, + "loss": 0.5731, + "step": 754 + }, + { + "epoch": 0.24157423208191126, + "grad_norm": 0.5837206840515137, + "learning_rate": 8.049040511727079e-06, + "loss": 0.5718, + "step": 755 + }, + { + "epoch": 0.24189419795221842, + "grad_norm": 0.6526334285736084, + "learning_rate": 8.059701492537314e-06, + "loss": 0.6007, + "step": 756 + }, + { + "epoch": 0.24221416382252559, + "grad_norm": 0.6099703907966614, + "learning_rate": 8.070362473347548e-06, + "loss": 0.5914, + "step": 757 + }, + { + "epoch": 0.24253412969283278, + "grad_norm": 0.6639705896377563, + "learning_rate": 8.081023454157784e-06, + "loss": 0.601, + "step": 758 + }, + { + "epoch": 0.24285409556313994, + "grad_norm": 0.66612309217453, + "learning_rate": 8.091684434968017e-06, + "loss": 0.5493, + "step": 759 + }, + { + "epoch": 0.2431740614334471, + "grad_norm": 0.6732496023178101, + "learning_rate": 8.102345415778252e-06, + "loss": 0.5995, + "step": 760 + }, + { + "epoch": 0.24349402730375427, + "grad_norm": 0.6970329284667969, + "learning_rate": 8.113006396588486e-06, + "loss": 0.5815, + "step": 761 + }, + { + "epoch": 0.24381399317406144, + "grad_norm": 0.7264329791069031, + "learning_rate": 8.123667377398723e-06, + "loss": 0.6032, + "step": 762 + }, + { + "epoch": 0.2441339590443686, + "grad_norm": 0.6926742792129517, + "learning_rate": 8.134328358208955e-06, + "loss": 0.571, + "step": 763 + }, + { + "epoch": 0.24445392491467577, + "grad_norm": 0.7230720520019531, + "learning_rate": 8.14498933901919e-06, + "loss": 0.6081, + "step": 764 + }, + { + "epoch": 0.24477389078498293, + "grad_norm": 0.7081652879714966, + "learning_rate": 8.155650319829424e-06, + "loss": 0.5771, + "step": 765 + }, + { + "epoch": 0.2450938566552901, + "grad_norm": 0.6964278817176819, + "learning_rate": 8.16631130063966e-06, + "loss": 0.6184, + "step": 766 + }, + { + "epoch": 0.24541382252559726, + "grad_norm": 0.6587214469909668, + "learning_rate": 8.176972281449893e-06, + "loss": 0.5783, + "step": 767 + }, + { + "epoch": 0.24573378839590443, + "grad_norm": 0.7075352668762207, + "learning_rate": 8.187633262260128e-06, + "loss": 0.5982, + "step": 768 + }, + { + "epoch": 0.2460537542662116, + "grad_norm": 0.6418300271034241, + "learning_rate": 8.198294243070363e-06, + "loss": 0.5858, + "step": 769 + }, + { + "epoch": 0.24637372013651876, + "grad_norm": 0.5982336401939392, + "learning_rate": 8.208955223880599e-06, + "loss": 0.5749, + "step": 770 + }, + { + "epoch": 0.24669368600682595, + "grad_norm": 0.696170449256897, + "learning_rate": 8.219616204690832e-06, + "loss": 0.5657, + "step": 771 + }, + { + "epoch": 0.24701365187713312, + "grad_norm": 0.6634019017219543, + "learning_rate": 8.230277185501066e-06, + "loss": 0.5397, + "step": 772 + }, + { + "epoch": 0.24733361774744028, + "grad_norm": 0.6658785939216614, + "learning_rate": 8.2409381663113e-06, + "loss": 0.5805, + "step": 773 + }, + { + "epoch": 0.24765358361774745, + "grad_norm": 0.6904900670051575, + "learning_rate": 8.251599147121537e-06, + "loss": 0.5917, + "step": 774 + }, + { + "epoch": 0.2479735494880546, + "grad_norm": 0.6993851065635681, + "learning_rate": 8.26226012793177e-06, + "loss": 0.5809, + "step": 775 + }, + { + "epoch": 0.24829351535836178, + "grad_norm": 0.7204084396362305, + "learning_rate": 8.272921108742004e-06, + "loss": 0.5708, + "step": 776 + }, + { + "epoch": 0.24861348122866894, + "grad_norm": 0.6606250405311584, + "learning_rate": 8.283582089552239e-06, + "loss": 0.5826, + "step": 777 + }, + { + "epoch": 0.2489334470989761, + "grad_norm": 0.5864621996879578, + "learning_rate": 8.294243070362475e-06, + "loss": 0.5777, + "step": 778 + }, + { + "epoch": 0.24925341296928327, + "grad_norm": 0.6498123407363892, + "learning_rate": 8.304904051172708e-06, + "loss": 0.6237, + "step": 779 + }, + { + "epoch": 0.24957337883959044, + "grad_norm": 0.6379949450492859, + "learning_rate": 8.315565031982942e-06, + "loss": 0.5634, + "step": 780 + }, + { + "epoch": 0.2498933447098976, + "grad_norm": 0.6023986339569092, + "learning_rate": 8.326226012793177e-06, + "loss": 0.5861, + "step": 781 + }, + { + "epoch": 0.2502133105802048, + "grad_norm": 0.7081618309020996, + "learning_rate": 8.336886993603413e-06, + "loss": 0.5642, + "step": 782 + }, + { + "epoch": 0.25053327645051193, + "grad_norm": 0.5907355546951294, + "learning_rate": 8.347547974413648e-06, + "loss": 0.5535, + "step": 783 + }, + { + "epoch": 0.2508532423208191, + "grad_norm": 0.7312473058700562, + "learning_rate": 8.35820895522388e-06, + "loss": 0.5675, + "step": 784 + }, + { + "epoch": 0.25117320819112626, + "grad_norm": 0.6056268215179443, + "learning_rate": 8.368869936034117e-06, + "loss": 0.5623, + "step": 785 + }, + { + "epoch": 0.25149317406143346, + "grad_norm": 0.5965462923049927, + "learning_rate": 8.379530916844351e-06, + "loss": 0.6102, + "step": 786 + }, + { + "epoch": 0.2518131399317406, + "grad_norm": 0.6213732957839966, + "learning_rate": 8.390191897654586e-06, + "loss": 0.575, + "step": 787 + }, + { + "epoch": 0.2521331058020478, + "grad_norm": 0.6415621042251587, + "learning_rate": 8.400852878464819e-06, + "loss": 0.595, + "step": 788 + }, + { + "epoch": 0.2524530716723549, + "grad_norm": 0.6275827884674072, + "learning_rate": 8.411513859275055e-06, + "loss": 0.5753, + "step": 789 + }, + { + "epoch": 0.2527730375426621, + "grad_norm": 0.7514165639877319, + "learning_rate": 8.42217484008529e-06, + "loss": 0.6222, + "step": 790 + }, + { + "epoch": 0.2530930034129693, + "grad_norm": 0.6514053344726562, + "learning_rate": 8.432835820895524e-06, + "loss": 0.5706, + "step": 791 + }, + { + "epoch": 0.25341296928327645, + "grad_norm": 0.7108455300331116, + "learning_rate": 8.443496801705757e-06, + "loss": 0.6316, + "step": 792 + }, + { + "epoch": 0.25373293515358364, + "grad_norm": 0.6999474763870239, + "learning_rate": 8.454157782515993e-06, + "loss": 0.5585, + "step": 793 + }, + { + "epoch": 0.2540529010238908, + "grad_norm": 0.6626673936843872, + "learning_rate": 8.464818763326227e-06, + "loss": 0.5968, + "step": 794 + }, + { + "epoch": 0.25437286689419797, + "grad_norm": 0.6229158043861389, + "learning_rate": 8.475479744136462e-06, + "loss": 0.6297, + "step": 795 + }, + { + "epoch": 0.2546928327645051, + "grad_norm": 0.6660380959510803, + "learning_rate": 8.486140724946695e-06, + "loss": 0.6002, + "step": 796 + }, + { + "epoch": 0.2550127986348123, + "grad_norm": 0.6904699206352234, + "learning_rate": 8.496801705756931e-06, + "loss": 0.605, + "step": 797 + }, + { + "epoch": 0.25533276450511944, + "grad_norm": 0.7006338834762573, + "learning_rate": 8.507462686567165e-06, + "loss": 0.5965, + "step": 798 + }, + { + "epoch": 0.25565273037542663, + "grad_norm": 0.6655489206314087, + "learning_rate": 8.5181236673774e-06, + "loss": 0.6186, + "step": 799 + }, + { + "epoch": 0.25597269624573377, + "grad_norm": 0.7310318946838379, + "learning_rate": 8.528784648187633e-06, + "loss": 0.6423, + "step": 800 + }, + { + "epoch": 0.25629266211604096, + "grad_norm": 0.737553060054779, + "learning_rate": 8.539445628997869e-06, + "loss": 0.6044, + "step": 801 + }, + { + "epoch": 0.2566126279863481, + "grad_norm": 0.7621052265167236, + "learning_rate": 8.550106609808104e-06, + "loss": 0.6067, + "step": 802 + }, + { + "epoch": 0.2569325938566553, + "grad_norm": 0.7435135245323181, + "learning_rate": 8.560767590618338e-06, + "loss": 0.5814, + "step": 803 + }, + { + "epoch": 0.2572525597269625, + "grad_norm": 0.7449132204055786, + "learning_rate": 8.571428571428571e-06, + "loss": 0.6206, + "step": 804 + }, + { + "epoch": 0.2575725255972696, + "grad_norm": 0.6484292149543762, + "learning_rate": 8.582089552238807e-06, + "loss": 0.6, + "step": 805 + }, + { + "epoch": 0.2578924914675768, + "grad_norm": 0.6599040031433105, + "learning_rate": 8.592750533049042e-06, + "loss": 0.5526, + "step": 806 + }, + { + "epoch": 0.25821245733788395, + "grad_norm": 0.7006849646568298, + "learning_rate": 8.603411513859276e-06, + "loss": 0.6108, + "step": 807 + }, + { + "epoch": 0.25853242320819114, + "grad_norm": 0.6509926319122314, + "learning_rate": 8.614072494669509e-06, + "loss": 0.5835, + "step": 808 + }, + { + "epoch": 0.2588523890784983, + "grad_norm": 0.6741862297058105, + "learning_rate": 8.624733475479745e-06, + "loss": 0.5516, + "step": 809 + }, + { + "epoch": 0.2591723549488055, + "grad_norm": 0.6708396673202515, + "learning_rate": 8.63539445628998e-06, + "loss": 0.5815, + "step": 810 + }, + { + "epoch": 0.2594923208191126, + "grad_norm": 0.7102819085121155, + "learning_rate": 8.646055437100214e-06, + "loss": 0.6043, + "step": 811 + }, + { + "epoch": 0.2598122866894198, + "grad_norm": 0.6952850222587585, + "learning_rate": 8.656716417910447e-06, + "loss": 0.5943, + "step": 812 + }, + { + "epoch": 0.26013225255972694, + "grad_norm": 0.7286235094070435, + "learning_rate": 8.667377398720683e-06, + "loss": 0.5735, + "step": 813 + }, + { + "epoch": 0.26045221843003413, + "grad_norm": 0.6984187960624695, + "learning_rate": 8.678038379530918e-06, + "loss": 0.5924, + "step": 814 + }, + { + "epoch": 0.26077218430034127, + "grad_norm": 0.6510578393936157, + "learning_rate": 8.688699360341152e-06, + "loss": 0.5999, + "step": 815 + }, + { + "epoch": 0.26109215017064846, + "grad_norm": 0.6672604084014893, + "learning_rate": 8.699360341151387e-06, + "loss": 0.6006, + "step": 816 + }, + { + "epoch": 0.26141211604095566, + "grad_norm": 0.6719098687171936, + "learning_rate": 8.710021321961621e-06, + "loss": 0.5706, + "step": 817 + }, + { + "epoch": 0.2617320819112628, + "grad_norm": 0.6405746340751648, + "learning_rate": 8.720682302771856e-06, + "loss": 0.5847, + "step": 818 + }, + { + "epoch": 0.26205204778157, + "grad_norm": 0.6537840962409973, + "learning_rate": 8.73134328358209e-06, + "loss": 0.5908, + "step": 819 + }, + { + "epoch": 0.2623720136518771, + "grad_norm": 0.685456395149231, + "learning_rate": 8.742004264392325e-06, + "loss": 0.5854, + "step": 820 + }, + { + "epoch": 0.2626919795221843, + "grad_norm": 0.7683976888656616, + "learning_rate": 8.75266524520256e-06, + "loss": 0.6454, + "step": 821 + }, + { + "epoch": 0.26301194539249145, + "grad_norm": 0.6867132782936096, + "learning_rate": 8.763326226012794e-06, + "loss": 0.562, + "step": 822 + }, + { + "epoch": 0.26333191126279865, + "grad_norm": 0.6225679516792297, + "learning_rate": 8.773987206823029e-06, + "loss": 0.6013, + "step": 823 + }, + { + "epoch": 0.2636518771331058, + "grad_norm": 0.7322930097579956, + "learning_rate": 8.784648187633263e-06, + "loss": 0.5666, + "step": 824 + }, + { + "epoch": 0.263971843003413, + "grad_norm": 0.6259191632270813, + "learning_rate": 8.795309168443498e-06, + "loss": 0.5527, + "step": 825 + }, + { + "epoch": 0.2642918088737201, + "grad_norm": 0.7102184295654297, + "learning_rate": 8.805970149253732e-06, + "loss": 0.6152, + "step": 826 + }, + { + "epoch": 0.2646117747440273, + "grad_norm": 0.7099477648735046, + "learning_rate": 8.816631130063967e-06, + "loss": 0.6105, + "step": 827 + }, + { + "epoch": 0.26493174061433444, + "grad_norm": 0.7012109160423279, + "learning_rate": 8.827292110874201e-06, + "loss": 0.6027, + "step": 828 + }, + { + "epoch": 0.26525170648464164, + "grad_norm": 0.765407145023346, + "learning_rate": 8.837953091684436e-06, + "loss": 0.5808, + "step": 829 + }, + { + "epoch": 0.26557167235494883, + "grad_norm": 0.6566329598426819, + "learning_rate": 8.84861407249467e-06, + "loss": 0.5603, + "step": 830 + }, + { + "epoch": 0.26589163822525597, + "grad_norm": 0.7441750764846802, + "learning_rate": 8.859275053304905e-06, + "loss": 0.5913, + "step": 831 + }, + { + "epoch": 0.26621160409556316, + "grad_norm": 0.6909970045089722, + "learning_rate": 8.86993603411514e-06, + "loss": 0.602, + "step": 832 + }, + { + "epoch": 0.2665315699658703, + "grad_norm": 0.6908165812492371, + "learning_rate": 8.880597014925374e-06, + "loss": 0.6129, + "step": 833 + }, + { + "epoch": 0.2668515358361775, + "grad_norm": 0.73782879114151, + "learning_rate": 8.891257995735608e-06, + "loss": 0.568, + "step": 834 + }, + { + "epoch": 0.2671715017064846, + "grad_norm": 0.6002166867256165, + "learning_rate": 8.901918976545843e-06, + "loss": 0.5844, + "step": 835 + }, + { + "epoch": 0.2674914675767918, + "grad_norm": 0.5845295190811157, + "learning_rate": 8.912579957356077e-06, + "loss": 0.594, + "step": 836 + }, + { + "epoch": 0.26781143344709896, + "grad_norm": 0.783158004283905, + "learning_rate": 8.923240938166312e-06, + "loss": 0.633, + "step": 837 + }, + { + "epoch": 0.26813139931740615, + "grad_norm": 0.6837108135223389, + "learning_rate": 8.933901918976547e-06, + "loss": 0.5938, + "step": 838 + }, + { + "epoch": 0.2684513651877133, + "grad_norm": 0.6747543811798096, + "learning_rate": 8.944562899786781e-06, + "loss": 0.5842, + "step": 839 + }, + { + "epoch": 0.2687713310580205, + "grad_norm": 0.7062118053436279, + "learning_rate": 8.955223880597016e-06, + "loss": 0.5453, + "step": 840 + }, + { + "epoch": 0.2690912969283277, + "grad_norm": 0.7308565974235535, + "learning_rate": 8.96588486140725e-06, + "loss": 0.6113, + "step": 841 + }, + { + "epoch": 0.2694112627986348, + "grad_norm": 0.6708758473396301, + "learning_rate": 8.976545842217485e-06, + "loss": 0.6004, + "step": 842 + }, + { + "epoch": 0.269731228668942, + "grad_norm": 0.7115057706832886, + "learning_rate": 8.987206823027719e-06, + "loss": 0.5802, + "step": 843 + }, + { + "epoch": 0.27005119453924914, + "grad_norm": 0.6746699810028076, + "learning_rate": 8.997867803837954e-06, + "loss": 0.6277, + "step": 844 + }, + { + "epoch": 0.27037116040955633, + "grad_norm": 0.7693876028060913, + "learning_rate": 9.008528784648188e-06, + "loss": 0.6117, + "step": 845 + }, + { + "epoch": 0.27069112627986347, + "grad_norm": 0.6683522462844849, + "learning_rate": 9.019189765458423e-06, + "loss": 0.5984, + "step": 846 + }, + { + "epoch": 0.27101109215017066, + "grad_norm": 0.6747117638587952, + "learning_rate": 9.029850746268657e-06, + "loss": 0.6316, + "step": 847 + }, + { + "epoch": 0.2713310580204778, + "grad_norm": 0.6540120244026184, + "learning_rate": 9.040511727078892e-06, + "loss": 0.5743, + "step": 848 + }, + { + "epoch": 0.271651023890785, + "grad_norm": 0.6651832461357117, + "learning_rate": 9.051172707889126e-06, + "loss": 0.615, + "step": 849 + }, + { + "epoch": 0.27197098976109213, + "grad_norm": 0.5970577001571655, + "learning_rate": 9.06183368869936e-06, + "loss": 0.574, + "step": 850 + }, + { + "epoch": 0.2722909556313993, + "grad_norm": 0.6609485745429993, + "learning_rate": 9.072494669509595e-06, + "loss": 0.5525, + "step": 851 + }, + { + "epoch": 0.27261092150170646, + "grad_norm": 0.6449378132820129, + "learning_rate": 9.08315565031983e-06, + "loss": 0.6593, + "step": 852 + }, + { + "epoch": 0.27293088737201365, + "grad_norm": 0.7078551054000854, + "learning_rate": 9.093816631130064e-06, + "loss": 0.6115, + "step": 853 + }, + { + "epoch": 0.27325085324232085, + "grad_norm": 0.6257254481315613, + "learning_rate": 9.104477611940299e-06, + "loss": 0.568, + "step": 854 + }, + { + "epoch": 0.273570819112628, + "grad_norm": 0.70318603515625, + "learning_rate": 9.115138592750533e-06, + "loss": 0.588, + "step": 855 + }, + { + "epoch": 0.2738907849829352, + "grad_norm": 0.6586565971374512, + "learning_rate": 9.125799573560768e-06, + "loss": 0.5872, + "step": 856 + }, + { + "epoch": 0.2742107508532423, + "grad_norm": 0.7630588412284851, + "learning_rate": 9.136460554371003e-06, + "loss": 0.6025, + "step": 857 + }, + { + "epoch": 0.2745307167235495, + "grad_norm": 0.6974326968193054, + "learning_rate": 9.147121535181237e-06, + "loss": 0.6098, + "step": 858 + }, + { + "epoch": 0.27485068259385664, + "grad_norm": 0.7954372763633728, + "learning_rate": 9.157782515991472e-06, + "loss": 0.6076, + "step": 859 + }, + { + "epoch": 0.27517064846416384, + "grad_norm": 0.717241644859314, + "learning_rate": 9.168443496801706e-06, + "loss": 0.6285, + "step": 860 + }, + { + "epoch": 0.275490614334471, + "grad_norm": 0.7189328074455261, + "learning_rate": 9.17910447761194e-06, + "loss": 0.605, + "step": 861 + }, + { + "epoch": 0.27581058020477817, + "grad_norm": 0.8277390003204346, + "learning_rate": 9.189765458422175e-06, + "loss": 0.589, + "step": 862 + }, + { + "epoch": 0.2761305460750853, + "grad_norm": 0.6832088232040405, + "learning_rate": 9.200426439232411e-06, + "loss": 0.6062, + "step": 863 + }, + { + "epoch": 0.2764505119453925, + "grad_norm": 0.653217077255249, + "learning_rate": 9.211087420042644e-06, + "loss": 0.5561, + "step": 864 + }, + { + "epoch": 0.27677047781569963, + "grad_norm": 0.759894073009491, + "learning_rate": 9.221748400852879e-06, + "loss": 0.6274, + "step": 865 + }, + { + "epoch": 0.2770904436860068, + "grad_norm": 0.7022303938865662, + "learning_rate": 9.232409381663113e-06, + "loss": 0.5755, + "step": 866 + }, + { + "epoch": 0.277410409556314, + "grad_norm": 0.7041699886322021, + "learning_rate": 9.24307036247335e-06, + "loss": 0.5774, + "step": 867 + }, + { + "epoch": 0.27773037542662116, + "grad_norm": 0.6180068850517273, + "learning_rate": 9.253731343283582e-06, + "loss": 0.5866, + "step": 868 + }, + { + "epoch": 0.27805034129692835, + "grad_norm": 0.7739641666412354, + "learning_rate": 9.264392324093817e-06, + "loss": 0.6054, + "step": 869 + }, + { + "epoch": 0.2783703071672355, + "grad_norm": 0.7242615818977356, + "learning_rate": 9.275053304904051e-06, + "loss": 0.5805, + "step": 870 + }, + { + "epoch": 0.2786902730375427, + "grad_norm": 0.8257843255996704, + "learning_rate": 9.285714285714288e-06, + "loss": 0.5771, + "step": 871 + }, + { + "epoch": 0.2790102389078498, + "grad_norm": 0.8289785385131836, + "learning_rate": 9.29637526652452e-06, + "loss": 0.6556, + "step": 872 + }, + { + "epoch": 0.279330204778157, + "grad_norm": 0.6636307835578918, + "learning_rate": 9.307036247334755e-06, + "loss": 0.6051, + "step": 873 + }, + { + "epoch": 0.27965017064846415, + "grad_norm": 0.7275445461273193, + "learning_rate": 9.31769722814499e-06, + "loss": 0.5676, + "step": 874 + }, + { + "epoch": 0.27997013651877134, + "grad_norm": 0.6746795773506165, + "learning_rate": 9.328358208955226e-06, + "loss": 0.5767, + "step": 875 + }, + { + "epoch": 0.2802901023890785, + "grad_norm": 0.6168851256370544, + "learning_rate": 9.339019189765458e-06, + "loss": 0.6061, + "step": 876 + }, + { + "epoch": 0.28061006825938567, + "grad_norm": 0.7217643857002258, + "learning_rate": 9.349680170575693e-06, + "loss": 0.6003, + "step": 877 + }, + { + "epoch": 0.2809300341296928, + "grad_norm": 0.6689902544021606, + "learning_rate": 9.36034115138593e-06, + "loss": 0.6247, + "step": 878 + }, + { + "epoch": 0.28125, + "grad_norm": 0.6647174954414368, + "learning_rate": 9.371002132196164e-06, + "loss": 0.603, + "step": 879 + }, + { + "epoch": 0.2815699658703072, + "grad_norm": 0.662013828754425, + "learning_rate": 9.381663113006397e-06, + "loss": 0.6026, + "step": 880 + }, + { + "epoch": 0.28188993174061433, + "grad_norm": 0.6372009515762329, + "learning_rate": 9.392324093816631e-06, + "loss": 0.5461, + "step": 881 + }, + { + "epoch": 0.2822098976109215, + "grad_norm": 0.7451372146606445, + "learning_rate": 9.402985074626867e-06, + "loss": 0.6046, + "step": 882 + }, + { + "epoch": 0.28252986348122866, + "grad_norm": 0.6218172907829285, + "learning_rate": 9.413646055437102e-06, + "loss": 0.5474, + "step": 883 + }, + { + "epoch": 0.28284982935153585, + "grad_norm": 0.5694857835769653, + "learning_rate": 9.424307036247335e-06, + "loss": 0.5917, + "step": 884 + }, + { + "epoch": 0.283169795221843, + "grad_norm": 0.5968525409698486, + "learning_rate": 9.43496801705757e-06, + "loss": 0.5733, + "step": 885 + }, + { + "epoch": 0.2834897610921502, + "grad_norm": 0.7009020447731018, + "learning_rate": 9.445628997867805e-06, + "loss": 0.6102, + "step": 886 + }, + { + "epoch": 0.2838097269624573, + "grad_norm": 0.654617965221405, + "learning_rate": 9.45628997867804e-06, + "loss": 0.6114, + "step": 887 + }, + { + "epoch": 0.2841296928327645, + "grad_norm": 0.6574299931526184, + "learning_rate": 9.466950959488273e-06, + "loss": 0.6115, + "step": 888 + }, + { + "epoch": 0.28444965870307165, + "grad_norm": 0.7500622868537903, + "learning_rate": 9.477611940298507e-06, + "loss": 0.6424, + "step": 889 + }, + { + "epoch": 0.28476962457337884, + "grad_norm": 0.6697449684143066, + "learning_rate": 9.488272921108744e-06, + "loss": 0.5415, + "step": 890 + }, + { + "epoch": 0.285089590443686, + "grad_norm": 0.6367616653442383, + "learning_rate": 9.498933901918978e-06, + "loss": 0.5739, + "step": 891 + }, + { + "epoch": 0.2854095563139932, + "grad_norm": 0.6737374663352966, + "learning_rate": 9.509594882729211e-06, + "loss": 0.5941, + "step": 892 + }, + { + "epoch": 0.28572952218430037, + "grad_norm": 0.6352681517601013, + "learning_rate": 9.520255863539445e-06, + "loss": 0.588, + "step": 893 + }, + { + "epoch": 0.2860494880546075, + "grad_norm": 0.6935127377510071, + "learning_rate": 9.530916844349682e-06, + "loss": 0.5845, + "step": 894 + }, + { + "epoch": 0.2863694539249147, + "grad_norm": 0.6597230434417725, + "learning_rate": 9.541577825159916e-06, + "loss": 0.6159, + "step": 895 + }, + { + "epoch": 0.28668941979522183, + "grad_norm": 0.7148585319519043, + "learning_rate": 9.552238805970149e-06, + "loss": 0.5976, + "step": 896 + }, + { + "epoch": 0.287009385665529, + "grad_norm": 0.7132214307785034, + "learning_rate": 9.562899786780384e-06, + "loss": 0.6219, + "step": 897 + }, + { + "epoch": 0.28732935153583616, + "grad_norm": 0.73183673620224, + "learning_rate": 9.57356076759062e-06, + "loss": 0.5652, + "step": 898 + }, + { + "epoch": 0.28764931740614336, + "grad_norm": 0.7999463677406311, + "learning_rate": 9.584221748400854e-06, + "loss": 0.558, + "step": 899 + }, + { + "epoch": 0.2879692832764505, + "grad_norm": 0.8034552335739136, + "learning_rate": 9.594882729211089e-06, + "loss": 0.6331, + "step": 900 + }, + { + "epoch": 0.2882892491467577, + "grad_norm": 0.660112738609314, + "learning_rate": 9.605543710021322e-06, + "loss": 0.5843, + "step": 901 + }, + { + "epoch": 0.2886092150170648, + "grad_norm": 0.8363476395606995, + "learning_rate": 9.616204690831558e-06, + "loss": 0.5984, + "step": 902 + }, + { + "epoch": 0.288929180887372, + "grad_norm": 0.7044869065284729, + "learning_rate": 9.626865671641792e-06, + "loss": 0.584, + "step": 903 + }, + { + "epoch": 0.28924914675767915, + "grad_norm": 0.6434544920921326, + "learning_rate": 9.637526652452027e-06, + "loss": 0.5776, + "step": 904 + }, + { + "epoch": 0.28956911262798635, + "grad_norm": 0.7166516780853271, + "learning_rate": 9.64818763326226e-06, + "loss": 0.5585, + "step": 905 + }, + { + "epoch": 0.28988907849829354, + "grad_norm": 0.6601057052612305, + "learning_rate": 9.658848614072496e-06, + "loss": 0.5893, + "step": 906 + }, + { + "epoch": 0.2902090443686007, + "grad_norm": 0.6507968902587891, + "learning_rate": 9.66950959488273e-06, + "loss": 0.633, + "step": 907 + }, + { + "epoch": 0.29052901023890787, + "grad_norm": 0.6383233070373535, + "learning_rate": 9.680170575692965e-06, + "loss": 0.5559, + "step": 908 + }, + { + "epoch": 0.290848976109215, + "grad_norm": 0.6499894857406616, + "learning_rate": 9.6908315565032e-06, + "loss": 0.5959, + "step": 909 + }, + { + "epoch": 0.2911689419795222, + "grad_norm": 0.62410569190979, + "learning_rate": 9.701492537313434e-06, + "loss": 0.5943, + "step": 910 + }, + { + "epoch": 0.29148890784982934, + "grad_norm": 0.7339252829551697, + "learning_rate": 9.712153518123669e-06, + "loss": 0.5782, + "step": 911 + }, + { + "epoch": 0.29180887372013653, + "grad_norm": 0.6893714666366577, + "learning_rate": 9.722814498933903e-06, + "loss": 0.6149, + "step": 912 + }, + { + "epoch": 0.29212883959044367, + "grad_norm": 0.6200568079948425, + "learning_rate": 9.733475479744138e-06, + "loss": 0.5349, + "step": 913 + }, + { + "epoch": 0.29244880546075086, + "grad_norm": 0.6621658205986023, + "learning_rate": 9.744136460554372e-06, + "loss": 0.5936, + "step": 914 + }, + { + "epoch": 0.292768771331058, + "grad_norm": 0.6932737827301025, + "learning_rate": 9.754797441364607e-06, + "loss": 0.6119, + "step": 915 + }, + { + "epoch": 0.2930887372013652, + "grad_norm": 0.6547825932502747, + "learning_rate": 9.765458422174841e-06, + "loss": 0.557, + "step": 916 + }, + { + "epoch": 0.2934087030716723, + "grad_norm": 0.7168145179748535, + "learning_rate": 9.776119402985076e-06, + "loss": 0.591, + "step": 917 + }, + { + "epoch": 0.2937286689419795, + "grad_norm": 0.7716021537780762, + "learning_rate": 9.78678038379531e-06, + "loss": 0.5465, + "step": 918 + }, + { + "epoch": 0.2940486348122867, + "grad_norm": 0.6409813165664673, + "learning_rate": 9.797441364605545e-06, + "loss": 0.6201, + "step": 919 + }, + { + "epoch": 0.29436860068259385, + "grad_norm": 0.724183201789856, + "learning_rate": 9.80810234541578e-06, + "loss": 0.5796, + "step": 920 + }, + { + "epoch": 0.29468856655290104, + "grad_norm": 0.7610300183296204, + "learning_rate": 9.818763326226014e-06, + "loss": 0.5688, + "step": 921 + }, + { + "epoch": 0.2950085324232082, + "grad_norm": 0.6418836116790771, + "learning_rate": 9.829424307036248e-06, + "loss": 0.6027, + "step": 922 + }, + { + "epoch": 0.2953284982935154, + "grad_norm": 0.678600549697876, + "learning_rate": 9.840085287846483e-06, + "loss": 0.5505, + "step": 923 + }, + { + "epoch": 0.2956484641638225, + "grad_norm": 0.7092781066894531, + "learning_rate": 9.850746268656717e-06, + "loss": 0.5945, + "step": 924 + }, + { + "epoch": 0.2959684300341297, + "grad_norm": 0.6756407022476196, + "learning_rate": 9.861407249466952e-06, + "loss": 0.5891, + "step": 925 + }, + { + "epoch": 0.29628839590443684, + "grad_norm": 0.6679978966712952, + "learning_rate": 9.872068230277187e-06, + "loss": 0.5535, + "step": 926 + }, + { + "epoch": 0.29660836177474403, + "grad_norm": 0.6598780751228333, + "learning_rate": 9.882729211087421e-06, + "loss": 0.612, + "step": 927 + }, + { + "epoch": 0.29692832764505117, + "grad_norm": 0.6880948543548584, + "learning_rate": 9.893390191897656e-06, + "loss": 0.6034, + "step": 928 + }, + { + "epoch": 0.29724829351535836, + "grad_norm": 0.7000877857208252, + "learning_rate": 9.90405117270789e-06, + "loss": 0.5231, + "step": 929 + }, + { + "epoch": 0.29756825938566556, + "grad_norm": 0.7053589820861816, + "learning_rate": 9.914712153518125e-06, + "loss": 0.5982, + "step": 930 + }, + { + "epoch": 0.2978882252559727, + "grad_norm": 0.7295544743537903, + "learning_rate": 9.925373134328359e-06, + "loss": 0.5584, + "step": 931 + }, + { + "epoch": 0.2982081911262799, + "grad_norm": 0.657467782497406, + "learning_rate": 9.936034115138594e-06, + "loss": 0.568, + "step": 932 + }, + { + "epoch": 0.298528156996587, + "grad_norm": 0.716752290725708, + "learning_rate": 9.946695095948828e-06, + "loss": 0.5728, + "step": 933 + }, + { + "epoch": 0.2988481228668942, + "grad_norm": 0.6869081854820251, + "learning_rate": 9.957356076759063e-06, + "loss": 0.5801, + "step": 934 + }, + { + "epoch": 0.29916808873720135, + "grad_norm": 0.6525644063949585, + "learning_rate": 9.968017057569297e-06, + "loss": 0.5332, + "step": 935 + }, + { + "epoch": 0.29948805460750855, + "grad_norm": 0.7127978205680847, + "learning_rate": 9.978678038379532e-06, + "loss": 0.6113, + "step": 936 + }, + { + "epoch": 0.2998080204778157, + "grad_norm": 0.7229118943214417, + "learning_rate": 9.989339019189766e-06, + "loss": 0.6064, + "step": 937 + }, + { + "epoch": 0.3001279863481229, + "grad_norm": 0.8196436762809753, + "learning_rate": 1e-05, + "loss": 0.5811, + "step": 938 + }, + { + "epoch": 0.30044795221843, + "grad_norm": 0.8263665437698364, + "learning_rate": 9.999999653371993e-06, + "loss": 0.6201, + "step": 939 + }, + { + "epoch": 0.3007679180887372, + "grad_norm": 0.84651118516922, + "learning_rate": 9.999998613488017e-06, + "loss": 0.594, + "step": 940 + }, + { + "epoch": 0.30108788395904434, + "grad_norm": 0.7862319350242615, + "learning_rate": 9.999996880348222e-06, + "loss": 0.6073, + "step": 941 + }, + { + "epoch": 0.30140784982935154, + "grad_norm": 0.7739344239234924, + "learning_rate": 9.999994453952842e-06, + "loss": 0.6434, + "step": 942 + }, + { + "epoch": 0.30172781569965873, + "grad_norm": 0.6697371006011963, + "learning_rate": 9.999991334302215e-06, + "loss": 0.5778, + "step": 943 + }, + { + "epoch": 0.30204778156996587, + "grad_norm": 0.7090111970901489, + "learning_rate": 9.999987521396774e-06, + "loss": 0.5715, + "step": 944 + }, + { + "epoch": 0.30236774744027306, + "grad_norm": 0.7254164218902588, + "learning_rate": 9.99998301523705e-06, + "loss": 0.5908, + "step": 945 + }, + { + "epoch": 0.3026877133105802, + "grad_norm": 0.74472975730896, + "learning_rate": 9.999977815823665e-06, + "loss": 0.6066, + "step": 946 + }, + { + "epoch": 0.3030076791808874, + "grad_norm": 0.7350637316703796, + "learning_rate": 9.99997192315734e-06, + "loss": 0.5776, + "step": 947 + }, + { + "epoch": 0.3033276450511945, + "grad_norm": 0.6912806034088135, + "learning_rate": 9.999965337238892e-06, + "loss": 0.5896, + "step": 948 + }, + { + "epoch": 0.3036476109215017, + "grad_norm": 0.7248418927192688, + "learning_rate": 9.999958058069237e-06, + "loss": 0.6134, + "step": 949 + }, + { + "epoch": 0.30396757679180886, + "grad_norm": 0.7095966339111328, + "learning_rate": 9.999950085649382e-06, + "loss": 0.5876, + "step": 950 + }, + { + "epoch": 0.30428754266211605, + "grad_norm": 0.7419503331184387, + "learning_rate": 9.999941419980432e-06, + "loss": 0.5915, + "step": 951 + }, + { + "epoch": 0.3046075085324232, + "grad_norm": 0.6916894912719727, + "learning_rate": 9.99993206106359e-06, + "loss": 0.5961, + "step": 952 + }, + { + "epoch": 0.3049274744027304, + "grad_norm": 0.7504989504814148, + "learning_rate": 9.99992200890015e-06, + "loss": 0.6087, + "step": 953 + }, + { + "epoch": 0.3052474402730375, + "grad_norm": 0.624904453754425, + "learning_rate": 9.99991126349151e-06, + "loss": 0.6079, + "step": 954 + }, + { + "epoch": 0.3055674061433447, + "grad_norm": 0.671826958656311, + "learning_rate": 9.999899824839159e-06, + "loss": 0.5845, + "step": 955 + }, + { + "epoch": 0.3058873720136519, + "grad_norm": 0.7129992246627808, + "learning_rate": 9.999887692944681e-06, + "loss": 0.5992, + "step": 956 + }, + { + "epoch": 0.30620733788395904, + "grad_norm": 0.6746633052825928, + "learning_rate": 9.99987486780976e-06, + "loss": 0.5832, + "step": 957 + }, + { + "epoch": 0.30652730375426623, + "grad_norm": 0.6484116911888123, + "learning_rate": 9.999861349436174e-06, + "loss": 0.5421, + "step": 958 + }, + { + "epoch": 0.30684726962457337, + "grad_norm": 0.7942398190498352, + "learning_rate": 9.999847137825795e-06, + "loss": 0.5889, + "step": 959 + }, + { + "epoch": 0.30716723549488056, + "grad_norm": 0.6215122938156128, + "learning_rate": 9.999832232980598e-06, + "loss": 0.5349, + "step": 960 + }, + { + "epoch": 0.3074872013651877, + "grad_norm": 0.6754631400108337, + "learning_rate": 9.999816634902645e-06, + "loss": 0.5662, + "step": 961 + }, + { + "epoch": 0.3078071672354949, + "grad_norm": 0.6522238850593567, + "learning_rate": 9.999800343594101e-06, + "loss": 0.6217, + "step": 962 + }, + { + "epoch": 0.30812713310580203, + "grad_norm": 0.6306779384613037, + "learning_rate": 9.999783359057226e-06, + "loss": 0.5367, + "step": 963 + }, + { + "epoch": 0.3084470989761092, + "grad_norm": 0.6738232374191284, + "learning_rate": 9.999765681294371e-06, + "loss": 0.6207, + "step": 964 + }, + { + "epoch": 0.30876706484641636, + "grad_norm": 0.774307906627655, + "learning_rate": 9.999747310307993e-06, + "loss": 0.6226, + "step": 965 + }, + { + "epoch": 0.30908703071672355, + "grad_norm": 0.6412689685821533, + "learning_rate": 9.999728246100632e-06, + "loss": 0.5872, + "step": 966 + }, + { + "epoch": 0.3094069965870307, + "grad_norm": 0.6333892941474915, + "learning_rate": 9.999708488674936e-06, + "loss": 0.5863, + "step": 967 + }, + { + "epoch": 0.3097269624573379, + "grad_norm": 0.6409503817558289, + "learning_rate": 9.999688038033643e-06, + "loss": 0.5536, + "step": 968 + }, + { + "epoch": 0.3100469283276451, + "grad_norm": 0.6600871086120605, + "learning_rate": 9.99966689417959e-06, + "loss": 0.5638, + "step": 969 + }, + { + "epoch": 0.3103668941979522, + "grad_norm": 0.6778677701950073, + "learning_rate": 9.999645057115706e-06, + "loss": 0.5806, + "step": 970 + }, + { + "epoch": 0.3106868600682594, + "grad_norm": 0.6836385726928711, + "learning_rate": 9.99962252684502e-06, + "loss": 0.5788, + "step": 971 + }, + { + "epoch": 0.31100682593856654, + "grad_norm": 0.6408164501190186, + "learning_rate": 9.999599303370658e-06, + "loss": 0.5407, + "step": 972 + }, + { + "epoch": 0.31132679180887374, + "grad_norm": 0.7385551929473877, + "learning_rate": 9.999575386695834e-06, + "loss": 0.5931, + "step": 973 + }, + { + "epoch": 0.3116467576791809, + "grad_norm": 0.7113818526268005, + "learning_rate": 9.999550776823872e-06, + "loss": 0.5816, + "step": 974 + }, + { + "epoch": 0.31196672354948807, + "grad_norm": 0.6312804818153381, + "learning_rate": 9.999525473758178e-06, + "loss": 0.5758, + "step": 975 + }, + { + "epoch": 0.3122866894197952, + "grad_norm": 0.6690606474876404, + "learning_rate": 9.999499477502263e-06, + "loss": 0.594, + "step": 976 + }, + { + "epoch": 0.3126066552901024, + "grad_norm": 0.7593414187431335, + "learning_rate": 9.999472788059732e-06, + "loss": 0.6296, + "step": 977 + }, + { + "epoch": 0.31292662116040953, + "grad_norm": 0.6702037453651428, + "learning_rate": 9.999445405434283e-06, + "loss": 0.5851, + "step": 978 + }, + { + "epoch": 0.3132465870307167, + "grad_norm": 0.6028419137001038, + "learning_rate": 9.999417329629715e-06, + "loss": 0.5407, + "step": 979 + }, + { + "epoch": 0.31356655290102387, + "grad_norm": 0.7517277002334595, + "learning_rate": 9.999388560649919e-06, + "loss": 0.6294, + "step": 980 + }, + { + "epoch": 0.31388651877133106, + "grad_norm": 0.6471433639526367, + "learning_rate": 9.999359098498884e-06, + "loss": 0.5681, + "step": 981 + }, + { + "epoch": 0.31420648464163825, + "grad_norm": 0.6184797286987305, + "learning_rate": 9.999328943180696e-06, + "loss": 0.5607, + "step": 982 + }, + { + "epoch": 0.3145264505119454, + "grad_norm": 0.652275562286377, + "learning_rate": 9.999298094699538e-06, + "loss": 0.57, + "step": 983 + }, + { + "epoch": 0.3148464163822526, + "grad_norm": 0.6612683534622192, + "learning_rate": 9.999266553059683e-06, + "loss": 0.5986, + "step": 984 + }, + { + "epoch": 0.3151663822525597, + "grad_norm": 0.6565192937850952, + "learning_rate": 9.999234318265507e-06, + "loss": 0.5933, + "step": 985 + }, + { + "epoch": 0.3154863481228669, + "grad_norm": 0.6525236964225769, + "learning_rate": 9.999201390321476e-06, + "loss": 0.574, + "step": 986 + }, + { + "epoch": 0.31580631399317405, + "grad_norm": 0.6861425638198853, + "learning_rate": 9.999167769232162e-06, + "loss": 0.6212, + "step": 987 + }, + { + "epoch": 0.31612627986348124, + "grad_norm": 0.5889517068862915, + "learning_rate": 9.999133455002219e-06, + "loss": 0.5596, + "step": 988 + }, + { + "epoch": 0.3164462457337884, + "grad_norm": 0.5472261905670166, + "learning_rate": 9.999098447636411e-06, + "loss": 0.5732, + "step": 989 + }, + { + "epoch": 0.31676621160409557, + "grad_norm": 0.6761401891708374, + "learning_rate": 9.999062747139587e-06, + "loss": 0.5878, + "step": 990 + }, + { + "epoch": 0.3170861774744027, + "grad_norm": 0.7099294662475586, + "learning_rate": 9.999026353516702e-06, + "loss": 0.6142, + "step": 991 + }, + { + "epoch": 0.3174061433447099, + "grad_norm": 0.7079955339431763, + "learning_rate": 9.998989266772797e-06, + "loss": 0.5645, + "step": 992 + }, + { + "epoch": 0.31772610921501704, + "grad_norm": 0.6157618761062622, + "learning_rate": 9.998951486913015e-06, + "loss": 0.5805, + "step": 993 + }, + { + "epoch": 0.31804607508532423, + "grad_norm": 0.6414501667022705, + "learning_rate": 9.998913013942597e-06, + "loss": 0.5719, + "step": 994 + }, + { + "epoch": 0.3183660409556314, + "grad_norm": 0.7088600993156433, + "learning_rate": 9.998873847866876e-06, + "loss": 0.5769, + "step": 995 + }, + { + "epoch": 0.31868600682593856, + "grad_norm": 0.6993066668510437, + "learning_rate": 9.998833988691281e-06, + "loss": 0.5656, + "step": 996 + }, + { + "epoch": 0.31900597269624575, + "grad_norm": 0.700746476650238, + "learning_rate": 9.998793436421342e-06, + "loss": 0.6116, + "step": 997 + }, + { + "epoch": 0.3193259385665529, + "grad_norm": 0.7170496582984924, + "learning_rate": 9.998752191062678e-06, + "loss": 0.6043, + "step": 998 + }, + { + "epoch": 0.3196459044368601, + "grad_norm": 0.7763385772705078, + "learning_rate": 9.998710252621009e-06, + "loss": 0.5861, + "step": 999 + }, + { + "epoch": 0.3199658703071672, + "grad_norm": 0.6285117268562317, + "learning_rate": 9.99866762110215e-06, + "loss": 0.5729, + "step": 1000 + }, + { + "epoch": 0.3202858361774744, + "grad_norm": 0.7778967618942261, + "learning_rate": 9.998624296512011e-06, + "loss": 0.6236, + "step": 1001 + }, + { + "epoch": 0.32060580204778155, + "grad_norm": 0.6404998302459717, + "learning_rate": 9.9985802788566e-06, + "loss": 0.5518, + "step": 1002 + }, + { + "epoch": 0.32092576791808874, + "grad_norm": 0.7212223410606384, + "learning_rate": 9.99853556814202e-06, + "loss": 0.5896, + "step": 1003 + }, + { + "epoch": 0.3212457337883959, + "grad_norm": 0.634592592716217, + "learning_rate": 9.998490164374472e-06, + "loss": 0.6042, + "step": 1004 + }, + { + "epoch": 0.3215656996587031, + "grad_norm": 0.676922082901001, + "learning_rate": 9.998444067560247e-06, + "loss": 0.6182, + "step": 1005 + }, + { + "epoch": 0.3218856655290102, + "grad_norm": 0.6989039182662964, + "learning_rate": 9.998397277705741e-06, + "loss": 0.5595, + "step": 1006 + }, + { + "epoch": 0.3222056313993174, + "grad_norm": 0.628720223903656, + "learning_rate": 9.998349794817436e-06, + "loss": 0.5595, + "step": 1007 + }, + { + "epoch": 0.3225255972696246, + "grad_norm": 0.7136312127113342, + "learning_rate": 9.998301618901923e-06, + "loss": 0.594, + "step": 1008 + }, + { + "epoch": 0.32284556313993173, + "grad_norm": 0.5778558254241943, + "learning_rate": 9.998252749965873e-06, + "loss": 0.5656, + "step": 1009 + }, + { + "epoch": 0.3231655290102389, + "grad_norm": 0.713370144367218, + "learning_rate": 9.99820318801607e-06, + "loss": 0.6236, + "step": 1010 + }, + { + "epoch": 0.32348549488054607, + "grad_norm": 0.6834118366241455, + "learning_rate": 9.99815293305938e-06, + "loss": 0.629, + "step": 1011 + }, + { + "epoch": 0.32380546075085326, + "grad_norm": 0.6257147789001465, + "learning_rate": 9.998101985102773e-06, + "loss": 0.5491, + "step": 1012 + }, + { + "epoch": 0.3241254266211604, + "grad_norm": 0.7445933818817139, + "learning_rate": 9.998050344153313e-06, + "loss": 0.6007, + "step": 1013 + }, + { + "epoch": 0.3244453924914676, + "grad_norm": 0.7208285927772522, + "learning_rate": 9.99799801021816e-06, + "loss": 0.5681, + "step": 1014 + }, + { + "epoch": 0.3247653583617747, + "grad_norm": 0.6081100106239319, + "learning_rate": 9.99794498330457e-06, + "loss": 0.5463, + "step": 1015 + }, + { + "epoch": 0.3250853242320819, + "grad_norm": 0.6651808619499207, + "learning_rate": 9.997891263419896e-06, + "loss": 0.581, + "step": 1016 + }, + { + "epoch": 0.32540529010238906, + "grad_norm": 0.5771706700325012, + "learning_rate": 9.997836850571584e-06, + "loss": 0.5767, + "step": 1017 + }, + { + "epoch": 0.32572525597269625, + "grad_norm": 0.7077763080596924, + "learning_rate": 9.997781744767182e-06, + "loss": 0.5861, + "step": 1018 + }, + { + "epoch": 0.3260452218430034, + "grad_norm": 0.6261406540870667, + "learning_rate": 9.997725946014328e-06, + "loss": 0.5665, + "step": 1019 + }, + { + "epoch": 0.3263651877133106, + "grad_norm": 0.6279311776161194, + "learning_rate": 9.997669454320759e-06, + "loss": 0.5788, + "step": 1020 + }, + { + "epoch": 0.32668515358361777, + "grad_norm": 0.7319290041923523, + "learning_rate": 9.997612269694308e-06, + "loss": 0.5647, + "step": 1021 + }, + { + "epoch": 0.3270051194539249, + "grad_norm": 0.5971276760101318, + "learning_rate": 9.997554392142903e-06, + "loss": 0.5847, + "step": 1022 + }, + { + "epoch": 0.3273250853242321, + "grad_norm": 0.6742677092552185, + "learning_rate": 9.997495821674572e-06, + "loss": 0.6257, + "step": 1023 + }, + { + "epoch": 0.32764505119453924, + "grad_norm": 0.7395401000976562, + "learning_rate": 9.99743655829743e-06, + "loss": 0.6391, + "step": 1024 + }, + { + "epoch": 0.32796501706484643, + "grad_norm": 0.6834313273429871, + "learning_rate": 9.997376602019698e-06, + "loss": 0.5521, + "step": 1025 + }, + { + "epoch": 0.32828498293515357, + "grad_norm": 0.7318760752677917, + "learning_rate": 9.99731595284969e-06, + "loss": 0.589, + "step": 1026 + }, + { + "epoch": 0.32860494880546076, + "grad_norm": 0.6454777717590332, + "learning_rate": 9.99725461079581e-06, + "loss": 0.6045, + "step": 1027 + }, + { + "epoch": 0.3289249146757679, + "grad_norm": 0.6907580494880676, + "learning_rate": 9.997192575866569e-06, + "loss": 0.5911, + "step": 1028 + }, + { + "epoch": 0.3292448805460751, + "grad_norm": 0.7321333885192871, + "learning_rate": 9.997129848070563e-06, + "loss": 0.5866, + "step": 1029 + }, + { + "epoch": 0.32956484641638223, + "grad_norm": 0.6222965717315674, + "learning_rate": 9.997066427416494e-06, + "loss": 0.5784, + "step": 1030 + }, + { + "epoch": 0.3298848122866894, + "grad_norm": 0.7147840261459351, + "learning_rate": 9.997002313913153e-06, + "loss": 0.5809, + "step": 1031 + }, + { + "epoch": 0.3302047781569966, + "grad_norm": 0.7093286514282227, + "learning_rate": 9.996937507569429e-06, + "loss": 0.5888, + "step": 1032 + }, + { + "epoch": 0.33052474402730375, + "grad_norm": 0.6205602884292603, + "learning_rate": 9.996872008394307e-06, + "loss": 0.5634, + "step": 1033 + }, + { + "epoch": 0.33084470989761094, + "grad_norm": 0.6194776892662048, + "learning_rate": 9.996805816396873e-06, + "loss": 0.5492, + "step": 1034 + }, + { + "epoch": 0.3311646757679181, + "grad_norm": 0.6360750198364258, + "learning_rate": 9.996738931586298e-06, + "loss": 0.6012, + "step": 1035 + }, + { + "epoch": 0.3314846416382253, + "grad_norm": 0.646390438079834, + "learning_rate": 9.996671353971861e-06, + "loss": 0.5748, + "step": 1036 + }, + { + "epoch": 0.3318046075085324, + "grad_norm": 0.6864686608314514, + "learning_rate": 9.996603083562928e-06, + "loss": 0.5952, + "step": 1037 + }, + { + "epoch": 0.3321245733788396, + "grad_norm": 0.6476268172264099, + "learning_rate": 9.996534120368968e-06, + "loss": 0.5659, + "step": 1038 + }, + { + "epoch": 0.33244453924914674, + "grad_norm": 0.651563823223114, + "learning_rate": 9.99646446439954e-06, + "loss": 0.6185, + "step": 1039 + }, + { + "epoch": 0.33276450511945393, + "grad_norm": 0.6136869192123413, + "learning_rate": 9.996394115664302e-06, + "loss": 0.5927, + "step": 1040 + }, + { + "epoch": 0.3330844709897611, + "grad_norm": 0.6799325346946716, + "learning_rate": 9.996323074173012e-06, + "loss": 0.6027, + "step": 1041 + }, + { + "epoch": 0.33340443686006827, + "grad_norm": 0.6832480430603027, + "learning_rate": 9.996251339935517e-06, + "loss": 0.6224, + "step": 1042 + }, + { + "epoch": 0.3337244027303754, + "grad_norm": 0.6809660196304321, + "learning_rate": 9.996178912961761e-06, + "loss": 0.5949, + "step": 1043 + }, + { + "epoch": 0.3340443686006826, + "grad_norm": 0.654651403427124, + "learning_rate": 9.996105793261787e-06, + "loss": 0.55, + "step": 1044 + }, + { + "epoch": 0.3343643344709898, + "grad_norm": 0.6393281817436218, + "learning_rate": 9.996031980845737e-06, + "loss": 0.5628, + "step": 1045 + }, + { + "epoch": 0.3346843003412969, + "grad_norm": 0.691877007484436, + "learning_rate": 9.995957475723844e-06, + "loss": 0.6134, + "step": 1046 + }, + { + "epoch": 0.3350042662116041, + "grad_norm": 0.6541376113891602, + "learning_rate": 9.995882277906432e-06, + "loss": 0.5532, + "step": 1047 + }, + { + "epoch": 0.33532423208191126, + "grad_norm": 0.6597897410392761, + "learning_rate": 9.995806387403935e-06, + "loss": 0.5665, + "step": 1048 + }, + { + "epoch": 0.33564419795221845, + "grad_norm": 0.6826539635658264, + "learning_rate": 9.995729804226873e-06, + "loss": 0.5538, + "step": 1049 + }, + { + "epoch": 0.3359641638225256, + "grad_norm": 0.6731455326080322, + "learning_rate": 9.995652528385861e-06, + "loss": 0.5992, + "step": 1050 + }, + { + "epoch": 0.3362841296928328, + "grad_norm": 0.7085751295089722, + "learning_rate": 9.995574559891618e-06, + "loss": 0.5933, + "step": 1051 + }, + { + "epoch": 0.3366040955631399, + "grad_norm": 0.7039170861244202, + "learning_rate": 9.995495898754952e-06, + "loss": 0.6041, + "step": 1052 + }, + { + "epoch": 0.3369240614334471, + "grad_norm": 0.6699726581573486, + "learning_rate": 9.99541654498677e-06, + "loss": 0.6477, + "step": 1053 + }, + { + "epoch": 0.33724402730375425, + "grad_norm": 0.7025903463363647, + "learning_rate": 9.995336498598075e-06, + "loss": 0.599, + "step": 1054 + }, + { + "epoch": 0.33756399317406144, + "grad_norm": 0.6295349597930908, + "learning_rate": 9.995255759599963e-06, + "loss": 0.6067, + "step": 1055 + }, + { + "epoch": 0.3378839590443686, + "grad_norm": 0.7312242388725281, + "learning_rate": 9.99517432800363e-06, + "loss": 0.6038, + "step": 1056 + }, + { + "epoch": 0.33820392491467577, + "grad_norm": 0.6863213181495667, + "learning_rate": 9.99509220382037e-06, + "loss": 0.5819, + "step": 1057 + }, + { + "epoch": 0.33852389078498296, + "grad_norm": 0.5715763568878174, + "learning_rate": 9.995009387061564e-06, + "loss": 0.5727, + "step": 1058 + }, + { + "epoch": 0.3388438566552901, + "grad_norm": 0.6914387345314026, + "learning_rate": 9.9949258777387e-06, + "loss": 0.6037, + "step": 1059 + }, + { + "epoch": 0.3391638225255973, + "grad_norm": 0.6119387149810791, + "learning_rate": 9.994841675863352e-06, + "loss": 0.5823, + "step": 1060 + }, + { + "epoch": 0.33948378839590443, + "grad_norm": 0.6641659736633301, + "learning_rate": 9.994756781447197e-06, + "loss": 0.5566, + "step": 1061 + }, + { + "epoch": 0.3398037542662116, + "grad_norm": 0.736589252948761, + "learning_rate": 9.994671194502006e-06, + "loss": 0.6084, + "step": 1062 + }, + { + "epoch": 0.34012372013651876, + "grad_norm": 0.6191753149032593, + "learning_rate": 9.994584915039644e-06, + "loss": 0.6031, + "step": 1063 + }, + { + "epoch": 0.34044368600682595, + "grad_norm": 0.6466872692108154, + "learning_rate": 9.994497943072076e-06, + "loss": 0.544, + "step": 1064 + }, + { + "epoch": 0.3407636518771331, + "grad_norm": 0.7583022117614746, + "learning_rate": 9.99441027861136e-06, + "loss": 0.5698, + "step": 1065 + }, + { + "epoch": 0.3410836177474403, + "grad_norm": 0.6253617405891418, + "learning_rate": 9.99432192166965e-06, + "loss": 0.5644, + "step": 1066 + }, + { + "epoch": 0.3414035836177474, + "grad_norm": 0.6997573375701904, + "learning_rate": 9.994232872259197e-06, + "loss": 0.6113, + "step": 1067 + }, + { + "epoch": 0.3417235494880546, + "grad_norm": 0.8160573840141296, + "learning_rate": 9.99414313039235e-06, + "loss": 0.5807, + "step": 1068 + }, + { + "epoch": 0.34204351535836175, + "grad_norm": 0.7675189971923828, + "learning_rate": 9.994052696081546e-06, + "loss": 0.5939, + "step": 1069 + }, + { + "epoch": 0.34236348122866894, + "grad_norm": 0.6997158527374268, + "learning_rate": 9.99396156933933e-06, + "loss": 0.5717, + "step": 1070 + }, + { + "epoch": 0.34268344709897613, + "grad_norm": 0.7480242252349854, + "learning_rate": 9.993869750178334e-06, + "loss": 0.5421, + "step": 1071 + }, + { + "epoch": 0.3430034129692833, + "grad_norm": 0.637569010257721, + "learning_rate": 9.993777238611292e-06, + "loss": 0.5676, + "step": 1072 + }, + { + "epoch": 0.34332337883959047, + "grad_norm": 0.6248916983604431, + "learning_rate": 9.993684034651024e-06, + "loss": 0.5755, + "step": 1073 + }, + { + "epoch": 0.3436433447098976, + "grad_norm": 0.7089166045188904, + "learning_rate": 9.993590138310458e-06, + "loss": 0.5866, + "step": 1074 + }, + { + "epoch": 0.3439633105802048, + "grad_norm": 0.7098347544670105, + "learning_rate": 9.993495549602615e-06, + "loss": 0.5986, + "step": 1075 + }, + { + "epoch": 0.34428327645051193, + "grad_norm": 0.7500614523887634, + "learning_rate": 9.993400268540603e-06, + "loss": 0.5966, + "step": 1076 + }, + { + "epoch": 0.3446032423208191, + "grad_norm": 0.841600775718689, + "learning_rate": 9.993304295137638e-06, + "loss": 0.5983, + "step": 1077 + }, + { + "epoch": 0.34492320819112626, + "grad_norm": 0.6450603008270264, + "learning_rate": 9.993207629407026e-06, + "loss": 0.5679, + "step": 1078 + }, + { + "epoch": 0.34524317406143346, + "grad_norm": 0.717484176158905, + "learning_rate": 9.993110271362168e-06, + "loss": 0.5646, + "step": 1079 + }, + { + "epoch": 0.3455631399317406, + "grad_norm": 0.63707035779953, + "learning_rate": 9.993012221016565e-06, + "loss": 0.5142, + "step": 1080 + }, + { + "epoch": 0.3458831058020478, + "grad_norm": 0.6633713841438293, + "learning_rate": 9.99291347838381e-06, + "loss": 0.5782, + "step": 1081 + }, + { + "epoch": 0.3462030716723549, + "grad_norm": 0.6746697425842285, + "learning_rate": 9.992814043477595e-06, + "loss": 0.5736, + "step": 1082 + }, + { + "epoch": 0.3465230375426621, + "grad_norm": 0.6532283425331116, + "learning_rate": 9.992713916311707e-06, + "loss": 0.5737, + "step": 1083 + }, + { + "epoch": 0.3468430034129693, + "grad_norm": 0.6815126538276672, + "learning_rate": 9.992613096900027e-06, + "loss": 0.5999, + "step": 1084 + }, + { + "epoch": 0.34716296928327645, + "grad_norm": 0.7754561901092529, + "learning_rate": 9.992511585256537e-06, + "loss": 0.5579, + "step": 1085 + }, + { + "epoch": 0.34748293515358364, + "grad_norm": 0.7235525250434875, + "learning_rate": 9.992409381395306e-06, + "loss": 0.578, + "step": 1086 + }, + { + "epoch": 0.3478029010238908, + "grad_norm": 0.7848513126373291, + "learning_rate": 9.992306485330511e-06, + "loss": 0.586, + "step": 1087 + }, + { + "epoch": 0.34812286689419797, + "grad_norm": 0.6942896246910095, + "learning_rate": 9.992202897076416e-06, + "loss": 0.5753, + "step": 1088 + }, + { + "epoch": 0.3484428327645051, + "grad_norm": 0.7098321318626404, + "learning_rate": 9.992098616647384e-06, + "loss": 0.5201, + "step": 1089 + }, + { + "epoch": 0.3487627986348123, + "grad_norm": 0.7250136137008667, + "learning_rate": 9.99199364405787e-06, + "loss": 0.5745, + "step": 1090 + }, + { + "epoch": 0.34908276450511944, + "grad_norm": 0.6978463530540466, + "learning_rate": 9.991887979322434e-06, + "loss": 0.5581, + "step": 1091 + }, + { + "epoch": 0.34940273037542663, + "grad_norm": 0.6456747055053711, + "learning_rate": 9.991781622455723e-06, + "loss": 0.5557, + "step": 1092 + }, + { + "epoch": 0.34972269624573377, + "grad_norm": 0.6560348272323608, + "learning_rate": 9.991674573472488e-06, + "loss": 0.5489, + "step": 1093 + }, + { + "epoch": 0.35004266211604096, + "grad_norm": 0.6581658720970154, + "learning_rate": 9.991566832387564e-06, + "loss": 0.5708, + "step": 1094 + }, + { + "epoch": 0.3503626279863481, + "grad_norm": 0.6679834127426147, + "learning_rate": 9.991458399215897e-06, + "loss": 0.6162, + "step": 1095 + }, + { + "epoch": 0.3506825938566553, + "grad_norm": 0.644548773765564, + "learning_rate": 9.991349273972517e-06, + "loss": 0.5614, + "step": 1096 + }, + { + "epoch": 0.3510025597269625, + "grad_norm": 0.645207941532135, + "learning_rate": 9.991239456672555e-06, + "loss": 0.595, + "step": 1097 + }, + { + "epoch": 0.3513225255972696, + "grad_norm": 0.6927055716514587, + "learning_rate": 9.991128947331236e-06, + "loss": 0.6157, + "step": 1098 + }, + { + "epoch": 0.3516424914675768, + "grad_norm": 0.66288822889328, + "learning_rate": 9.991017745963886e-06, + "loss": 0.5706, + "step": 1099 + }, + { + "epoch": 0.35196245733788395, + "grad_norm": 0.5749547481536865, + "learning_rate": 9.990905852585922e-06, + "loss": 0.5642, + "step": 1100 + }, + { + "epoch": 0.35228242320819114, + "grad_norm": 0.726407527923584, + "learning_rate": 9.990793267212857e-06, + "loss": 0.5749, + "step": 1101 + }, + { + "epoch": 0.3526023890784983, + "grad_norm": 0.6601400971412659, + "learning_rate": 9.9906799898603e-06, + "loss": 0.6701, + "step": 1102 + }, + { + "epoch": 0.3529223549488055, + "grad_norm": 0.614708662033081, + "learning_rate": 9.99056602054396e-06, + "loss": 0.5712, + "step": 1103 + }, + { + "epoch": 0.3532423208191126, + "grad_norm": 0.6331604719161987, + "learning_rate": 9.990451359279636e-06, + "loss": 0.5656, + "step": 1104 + }, + { + "epoch": 0.3535622866894198, + "grad_norm": 0.6421919465065002, + "learning_rate": 9.990336006083228e-06, + "loss": 0.5533, + "step": 1105 + }, + { + "epoch": 0.35388225255972694, + "grad_norm": 0.6681429147720337, + "learning_rate": 9.99021996097073e-06, + "loss": 0.5741, + "step": 1106 + }, + { + "epoch": 0.35420221843003413, + "grad_norm": 0.7112416625022888, + "learning_rate": 9.99010322395823e-06, + "loss": 0.5495, + "step": 1107 + }, + { + "epoch": 0.35452218430034127, + "grad_norm": 0.705014169216156, + "learning_rate": 9.989985795061914e-06, + "loss": 0.59, + "step": 1108 + }, + { + "epoch": 0.35484215017064846, + "grad_norm": 0.7376593351364136, + "learning_rate": 9.989867674298066e-06, + "loss": 0.5733, + "step": 1109 + }, + { + "epoch": 0.35516211604095566, + "grad_norm": 0.6666154861450195, + "learning_rate": 9.989748861683061e-06, + "loss": 0.5523, + "step": 1110 + }, + { + "epoch": 0.3554820819112628, + "grad_norm": 0.7035896182060242, + "learning_rate": 9.989629357233375e-06, + "loss": 0.5774, + "step": 1111 + }, + { + "epoch": 0.35580204778157, + "grad_norm": 0.6515471935272217, + "learning_rate": 9.989509160965575e-06, + "loss": 0.5701, + "step": 1112 + }, + { + "epoch": 0.3561220136518771, + "grad_norm": 0.6917015314102173, + "learning_rate": 9.989388272896328e-06, + "loss": 0.5746, + "step": 1113 + }, + { + "epoch": 0.3564419795221843, + "grad_norm": 0.6861441731452942, + "learning_rate": 9.989266693042394e-06, + "loss": 0.605, + "step": 1114 + }, + { + "epoch": 0.35676194539249145, + "grad_norm": 0.5988866686820984, + "learning_rate": 9.98914442142063e-06, + "loss": 0.5874, + "step": 1115 + }, + { + "epoch": 0.35708191126279865, + "grad_norm": 0.644522488117218, + "learning_rate": 9.989021458047992e-06, + "loss": 0.5827, + "step": 1116 + }, + { + "epoch": 0.3574018771331058, + "grad_norm": 0.7073920369148254, + "learning_rate": 9.988897802941523e-06, + "loss": 0.5459, + "step": 1117 + }, + { + "epoch": 0.357721843003413, + "grad_norm": 0.5924735069274902, + "learning_rate": 9.988773456118376e-06, + "loss": 0.5737, + "step": 1118 + }, + { + "epoch": 0.3580418088737201, + "grad_norm": 0.8429334163665771, + "learning_rate": 9.988648417595785e-06, + "loss": 0.626, + "step": 1119 + }, + { + "epoch": 0.3583617747440273, + "grad_norm": 0.7303470969200134, + "learning_rate": 9.988522687391092e-06, + "loss": 0.5688, + "step": 1120 + }, + { + "epoch": 0.35868174061433444, + "grad_norm": 0.6042421460151672, + "learning_rate": 9.988396265521725e-06, + "loss": 0.5737, + "step": 1121 + }, + { + "epoch": 0.35900170648464164, + "grad_norm": 0.7154580950737, + "learning_rate": 9.988269152005215e-06, + "loss": 0.6298, + "step": 1122 + }, + { + "epoch": 0.35932167235494883, + "grad_norm": 0.6817564368247986, + "learning_rate": 9.988141346859186e-06, + "loss": 0.6032, + "step": 1123 + }, + { + "epoch": 0.35964163822525597, + "grad_norm": 0.7040557861328125, + "learning_rate": 9.988012850101359e-06, + "loss": 0.592, + "step": 1124 + }, + { + "epoch": 0.35996160409556316, + "grad_norm": 0.6956594586372375, + "learning_rate": 9.987883661749548e-06, + "loss": 0.5889, + "step": 1125 + }, + { + "epoch": 0.3602815699658703, + "grad_norm": 0.6449658870697021, + "learning_rate": 9.987753781821667e-06, + "loss": 0.6034, + "step": 1126 + }, + { + "epoch": 0.3606015358361775, + "grad_norm": 0.7003522515296936, + "learning_rate": 9.987623210335725e-06, + "loss": 0.5691, + "step": 1127 + }, + { + "epoch": 0.3609215017064846, + "grad_norm": 0.6019575595855713, + "learning_rate": 9.987491947309825e-06, + "loss": 0.526, + "step": 1128 + }, + { + "epoch": 0.3612414675767918, + "grad_norm": 0.7500554323196411, + "learning_rate": 9.987359992762165e-06, + "loss": 0.622, + "step": 1129 + }, + { + "epoch": 0.36156143344709896, + "grad_norm": 0.6906517148017883, + "learning_rate": 9.987227346711043e-06, + "loss": 0.5249, + "step": 1130 + }, + { + "epoch": 0.36188139931740615, + "grad_norm": 0.6303286552429199, + "learning_rate": 9.987094009174849e-06, + "loss": 0.5833, + "step": 1131 + }, + { + "epoch": 0.3622013651877133, + "grad_norm": 0.6838951706886292, + "learning_rate": 9.98695998017207e-06, + "loss": 0.556, + "step": 1132 + }, + { + "epoch": 0.3625213310580205, + "grad_norm": 0.7039484977722168, + "learning_rate": 9.986825259721292e-06, + "loss": 0.5779, + "step": 1133 + }, + { + "epoch": 0.3628412969283277, + "grad_norm": 0.6559299826622009, + "learning_rate": 9.986689847841192e-06, + "loss": 0.5974, + "step": 1134 + }, + { + "epoch": 0.3631612627986348, + "grad_norm": 0.7664186954498291, + "learning_rate": 9.986553744550546e-06, + "loss": 0.5808, + "step": 1135 + }, + { + "epoch": 0.363481228668942, + "grad_norm": 0.6697313189506531, + "learning_rate": 9.986416949868223e-06, + "loss": 0.5865, + "step": 1136 + }, + { + "epoch": 0.36380119453924914, + "grad_norm": 0.709850013256073, + "learning_rate": 9.986279463813192e-06, + "loss": 0.5888, + "step": 1137 + }, + { + "epoch": 0.36412116040955633, + "grad_norm": 0.7761910557746887, + "learning_rate": 9.986141286404514e-06, + "loss": 0.6414, + "step": 1138 + }, + { + "epoch": 0.36444112627986347, + "grad_norm": 0.6782816052436829, + "learning_rate": 9.98600241766135e-06, + "loss": 0.5639, + "step": 1139 + }, + { + "epoch": 0.36476109215017066, + "grad_norm": 0.6862764954566956, + "learning_rate": 9.985862857602951e-06, + "loss": 0.5562, + "step": 1140 + }, + { + "epoch": 0.3650810580204778, + "grad_norm": 0.7213887572288513, + "learning_rate": 9.985722606248669e-06, + "loss": 0.6093, + "step": 1141 + }, + { + "epoch": 0.365401023890785, + "grad_norm": 0.747785210609436, + "learning_rate": 9.985581663617948e-06, + "loss": 0.5812, + "step": 1142 + }, + { + "epoch": 0.36572098976109213, + "grad_norm": 0.5790359973907471, + "learning_rate": 9.985440029730335e-06, + "loss": 0.5696, + "step": 1143 + }, + { + "epoch": 0.3660409556313993, + "grad_norm": 0.653598964214325, + "learning_rate": 9.985297704605461e-06, + "loss": 0.5617, + "step": 1144 + }, + { + "epoch": 0.36636092150170646, + "grad_norm": 0.6908425092697144, + "learning_rate": 9.985154688263065e-06, + "loss": 0.5403, + "step": 1145 + }, + { + "epoch": 0.36668088737201365, + "grad_norm": 0.5721466541290283, + "learning_rate": 9.985010980722974e-06, + "loss": 0.5788, + "step": 1146 + }, + { + "epoch": 0.36700085324232085, + "grad_norm": 0.6806579232215881, + "learning_rate": 9.984866582005112e-06, + "loss": 0.5775, + "step": 1147 + }, + { + "epoch": 0.367320819112628, + "grad_norm": 0.6131672859191895, + "learning_rate": 9.9847214921295e-06, + "loss": 0.5577, + "step": 1148 + }, + { + "epoch": 0.3676407849829352, + "grad_norm": 0.6523510217666626, + "learning_rate": 9.98457571111626e-06, + "loss": 0.5776, + "step": 1149 + }, + { + "epoch": 0.3679607508532423, + "grad_norm": 0.6214600801467896, + "learning_rate": 9.9844292389856e-06, + "loss": 0.5581, + "step": 1150 + }, + { + "epoch": 0.3682807167235495, + "grad_norm": 0.6363633871078491, + "learning_rate": 9.98428207575783e-06, + "loss": 0.6062, + "step": 1151 + }, + { + "epoch": 0.36860068259385664, + "grad_norm": 0.7081869840621948, + "learning_rate": 9.984134221453352e-06, + "loss": 0.5975, + "step": 1152 + }, + { + "epoch": 0.36892064846416384, + "grad_norm": 0.6446593403816223, + "learning_rate": 9.98398567609267e-06, + "loss": 0.6211, + "step": 1153 + }, + { + "epoch": 0.369240614334471, + "grad_norm": 0.7253109216690063, + "learning_rate": 9.983836439696378e-06, + "loss": 0.6514, + "step": 1154 + }, + { + "epoch": 0.36956058020477817, + "grad_norm": 0.6849066019058228, + "learning_rate": 9.983686512285166e-06, + "loss": 0.5941, + "step": 1155 + }, + { + "epoch": 0.3698805460750853, + "grad_norm": 0.7073627710342407, + "learning_rate": 9.983535893879825e-06, + "loss": 0.627, + "step": 1156 + }, + { + "epoch": 0.3702005119453925, + "grad_norm": 0.7770823240280151, + "learning_rate": 9.983384584501236e-06, + "loss": 0.639, + "step": 1157 + }, + { + "epoch": 0.37052047781569963, + "grad_norm": 0.6505449414253235, + "learning_rate": 9.98323258417038e-06, + "loss": 0.5884, + "step": 1158 + }, + { + "epoch": 0.3708404436860068, + "grad_norm": 0.6462435722351074, + "learning_rate": 9.983079892908332e-06, + "loss": 0.5755, + "step": 1159 + }, + { + "epoch": 0.371160409556314, + "grad_norm": 0.6581071019172668, + "learning_rate": 9.982926510736261e-06, + "loss": 0.5478, + "step": 1160 + }, + { + "epoch": 0.37148037542662116, + "grad_norm": 0.6703523397445679, + "learning_rate": 9.982772437675436e-06, + "loss": 0.6015, + "step": 1161 + }, + { + "epoch": 0.37180034129692835, + "grad_norm": 0.6885740160942078, + "learning_rate": 9.982617673747218e-06, + "loss": 0.5669, + "step": 1162 + }, + { + "epoch": 0.3721203071672355, + "grad_norm": 0.7145320177078247, + "learning_rate": 9.982462218973064e-06, + "loss": 0.6137, + "step": 1163 + }, + { + "epoch": 0.3724402730375427, + "grad_norm": 0.6247287392616272, + "learning_rate": 9.98230607337453e-06, + "loss": 0.5928, + "step": 1164 + }, + { + "epoch": 0.3727602389078498, + "grad_norm": 0.6764070987701416, + "learning_rate": 9.982149236973264e-06, + "loss": 0.5466, + "step": 1165 + }, + { + "epoch": 0.373080204778157, + "grad_norm": 0.6221293807029724, + "learning_rate": 9.981991709791012e-06, + "loss": 0.5462, + "step": 1166 + }, + { + "epoch": 0.37340017064846415, + "grad_norm": 0.6133841276168823, + "learning_rate": 9.981833491849618e-06, + "loss": 0.5722, + "step": 1167 + }, + { + "epoch": 0.37372013651877134, + "grad_norm": 0.7736768126487732, + "learning_rate": 9.981674583171017e-06, + "loss": 0.6171, + "step": 1168 + }, + { + "epoch": 0.3740401023890785, + "grad_norm": 0.575372040271759, + "learning_rate": 9.98151498377724e-06, + "loss": 0.572, + "step": 1169 + }, + { + "epoch": 0.37436006825938567, + "grad_norm": 0.6014509201049805, + "learning_rate": 9.98135469369042e-06, + "loss": 0.5583, + "step": 1170 + }, + { + "epoch": 0.3746800341296928, + "grad_norm": 0.6997368335723877, + "learning_rate": 9.981193712932778e-06, + "loss": 0.5789, + "step": 1171 + }, + { + "epoch": 0.375, + "grad_norm": 0.6987643837928772, + "learning_rate": 9.981032041526635e-06, + "loss": 0.5355, + "step": 1172 + }, + { + "epoch": 0.3753199658703072, + "grad_norm": 0.5802475214004517, + "learning_rate": 9.980869679494407e-06, + "loss": 0.5657, + "step": 1173 + }, + { + "epoch": 0.37563993174061433, + "grad_norm": 0.6369011402130127, + "learning_rate": 9.980706626858607e-06, + "loss": 0.5567, + "step": 1174 + }, + { + "epoch": 0.3759598976109215, + "grad_norm": 0.5437308549880981, + "learning_rate": 9.980542883641839e-06, + "loss": 0.5418, + "step": 1175 + }, + { + "epoch": 0.37627986348122866, + "grad_norm": 0.5967769622802734, + "learning_rate": 9.98037844986681e-06, + "loss": 0.5721, + "step": 1176 + }, + { + "epoch": 0.37659982935153585, + "grad_norm": 0.7096673250198364, + "learning_rate": 9.980213325556319e-06, + "loss": 0.6103, + "step": 1177 + }, + { + "epoch": 0.376919795221843, + "grad_norm": 0.5754175186157227, + "learning_rate": 9.980047510733257e-06, + "loss": 0.5682, + "step": 1178 + }, + { + "epoch": 0.3772397610921502, + "grad_norm": 0.6432886123657227, + "learning_rate": 9.979881005420615e-06, + "loss": 0.5549, + "step": 1179 + }, + { + "epoch": 0.3775597269624573, + "grad_norm": 0.6181249618530273, + "learning_rate": 9.979713809641482e-06, + "loss": 0.5719, + "step": 1180 + }, + { + "epoch": 0.3778796928327645, + "grad_norm": 0.7071648240089417, + "learning_rate": 9.97954592341904e-06, + "loss": 0.575, + "step": 1181 + }, + { + "epoch": 0.37819965870307165, + "grad_norm": 0.6493441462516785, + "learning_rate": 9.979377346776564e-06, + "loss": 0.5015, + "step": 1182 + }, + { + "epoch": 0.37851962457337884, + "grad_norm": 0.6092111468315125, + "learning_rate": 9.979208079737428e-06, + "loss": 0.5714, + "step": 1183 + }, + { + "epoch": 0.378839590443686, + "grad_norm": 0.6894551515579224, + "learning_rate": 9.979038122325103e-06, + "loss": 0.6282, + "step": 1184 + }, + { + "epoch": 0.3791595563139932, + "grad_norm": 0.6173356175422668, + "learning_rate": 9.978867474563151e-06, + "loss": 0.5896, + "step": 1185 + }, + { + "epoch": 0.37947952218430037, + "grad_norm": 0.6287574172019958, + "learning_rate": 9.978696136475235e-06, + "loss": 0.5881, + "step": 1186 + }, + { + "epoch": 0.3797994880546075, + "grad_norm": 0.6383042335510254, + "learning_rate": 9.97852410808511e-06, + "loss": 0.598, + "step": 1187 + }, + { + "epoch": 0.3801194539249147, + "grad_norm": 0.6773865818977356, + "learning_rate": 9.978351389416626e-06, + "loss": 0.6118, + "step": 1188 + }, + { + "epoch": 0.38043941979522183, + "grad_norm": 0.6683553457260132, + "learning_rate": 9.978177980493736e-06, + "loss": 0.6278, + "step": 1189 + }, + { + "epoch": 0.380759385665529, + "grad_norm": 0.6252800822257996, + "learning_rate": 9.978003881340478e-06, + "loss": 0.5507, + "step": 1190 + }, + { + "epoch": 0.38107935153583616, + "grad_norm": 0.6656942367553711, + "learning_rate": 9.977829091980996e-06, + "loss": 0.6245, + "step": 1191 + }, + { + "epoch": 0.38139931740614336, + "grad_norm": 0.6880273222923279, + "learning_rate": 9.977653612439518e-06, + "loss": 0.5634, + "step": 1192 + }, + { + "epoch": 0.3817192832764505, + "grad_norm": 0.6544172167778015, + "learning_rate": 9.97747744274038e-06, + "loss": 0.5966, + "step": 1193 + }, + { + "epoch": 0.3820392491467577, + "grad_norm": 0.6122605204582214, + "learning_rate": 9.977300582908007e-06, + "loss": 0.6154, + "step": 1194 + }, + { + "epoch": 0.3823592150170648, + "grad_norm": 0.739066481590271, + "learning_rate": 9.97712303296692e-06, + "loss": 0.5971, + "step": 1195 + }, + { + "epoch": 0.382679180887372, + "grad_norm": 0.6078752875328064, + "learning_rate": 9.976944792941736e-06, + "loss": 0.561, + "step": 1196 + }, + { + "epoch": 0.38299914675767915, + "grad_norm": 0.6688116788864136, + "learning_rate": 9.976765862857171e-06, + "loss": 0.5797, + "step": 1197 + }, + { + "epoch": 0.38331911262798635, + "grad_norm": 0.6825562715530396, + "learning_rate": 9.976586242738032e-06, + "loss": 0.5775, + "step": 1198 + }, + { + "epoch": 0.38363907849829354, + "grad_norm": 0.6867298483848572, + "learning_rate": 9.976405932609221e-06, + "loss": 0.581, + "step": 1199 + }, + { + "epoch": 0.3839590443686007, + "grad_norm": 0.7204470634460449, + "learning_rate": 9.976224932495741e-06, + "loss": 0.5947, + "step": 1200 + }, + { + "epoch": 0.38427901023890787, + "grad_norm": 0.8085878491401672, + "learning_rate": 9.97604324242269e-06, + "loss": 0.6117, + "step": 1201 + }, + { + "epoch": 0.384598976109215, + "grad_norm": 0.6830131411552429, + "learning_rate": 9.975860862415255e-06, + "loss": 0.58, + "step": 1202 + }, + { + "epoch": 0.3849189419795222, + "grad_norm": 0.6403172016143799, + "learning_rate": 9.975677792498725e-06, + "loss": 0.5923, + "step": 1203 + }, + { + "epoch": 0.38523890784982934, + "grad_norm": 0.6732902526855469, + "learning_rate": 9.975494032698485e-06, + "loss": 0.5615, + "step": 1204 + }, + { + "epoch": 0.38555887372013653, + "grad_norm": 0.726446270942688, + "learning_rate": 9.975309583040012e-06, + "loss": 0.5781, + "step": 1205 + }, + { + "epoch": 0.38587883959044367, + "grad_norm": 0.7127211093902588, + "learning_rate": 9.975124443548878e-06, + "loss": 0.573, + "step": 1206 + }, + { + "epoch": 0.38619880546075086, + "grad_norm": 0.6206179261207581, + "learning_rate": 9.974938614250755e-06, + "loss": 0.5707, + "step": 1207 + }, + { + "epoch": 0.386518771331058, + "grad_norm": 0.8432472944259644, + "learning_rate": 9.974752095171409e-06, + "loss": 0.5884, + "step": 1208 + }, + { + "epoch": 0.3868387372013652, + "grad_norm": 0.8524286150932312, + "learning_rate": 9.9745648863367e-06, + "loss": 0.6203, + "step": 1209 + }, + { + "epoch": 0.3871587030716723, + "grad_norm": 0.6269184947013855, + "learning_rate": 9.974376987772584e-06, + "loss": 0.5799, + "step": 1210 + }, + { + "epoch": 0.3874786689419795, + "grad_norm": 0.7409312129020691, + "learning_rate": 9.974188399505116e-06, + "loss": 0.5737, + "step": 1211 + }, + { + "epoch": 0.3877986348122867, + "grad_norm": 0.6829506754875183, + "learning_rate": 9.97399912156044e-06, + "loss": 0.5729, + "step": 1212 + }, + { + "epoch": 0.38811860068259385, + "grad_norm": 0.7614309191703796, + "learning_rate": 9.973809153964803e-06, + "loss": 0.559, + "step": 1213 + }, + { + "epoch": 0.38843856655290104, + "grad_norm": 0.6158562898635864, + "learning_rate": 9.973618496744544e-06, + "loss": 0.5986, + "step": 1214 + }, + { + "epoch": 0.3887585324232082, + "grad_norm": 0.5494482517242432, + "learning_rate": 9.973427149926098e-06, + "loss": 0.5347, + "step": 1215 + }, + { + "epoch": 0.3890784982935154, + "grad_norm": 0.7281996607780457, + "learning_rate": 9.973235113535992e-06, + "loss": 0.566, + "step": 1216 + }, + { + "epoch": 0.3893984641638225, + "grad_norm": 0.7141414284706116, + "learning_rate": 9.973042387600856e-06, + "loss": 0.5825, + "step": 1217 + }, + { + "epoch": 0.3897184300341297, + "grad_norm": 0.5919783115386963, + "learning_rate": 9.972848972147408e-06, + "loss": 0.5699, + "step": 1218 + }, + { + "epoch": 0.39003839590443684, + "grad_norm": 0.6098409295082092, + "learning_rate": 9.97265486720247e-06, + "loss": 0.5485, + "step": 1219 + }, + { + "epoch": 0.39035836177474403, + "grad_norm": 0.5585481524467468, + "learning_rate": 9.972460072792951e-06, + "loss": 0.5359, + "step": 1220 + }, + { + "epoch": 0.39067832764505117, + "grad_norm": 0.620502233505249, + "learning_rate": 9.972264588945863e-06, + "loss": 0.6053, + "step": 1221 + }, + { + "epoch": 0.39099829351535836, + "grad_norm": 0.6805775165557861, + "learning_rate": 9.972068415688307e-06, + "loss": 0.6028, + "step": 1222 + }, + { + "epoch": 0.39131825938566556, + "grad_norm": 0.6690160632133484, + "learning_rate": 9.971871553047484e-06, + "loss": 0.5774, + "step": 1223 + }, + { + "epoch": 0.3916382252559727, + "grad_norm": 0.7072525024414062, + "learning_rate": 9.971674001050687e-06, + "loss": 0.6017, + "step": 1224 + }, + { + "epoch": 0.3919581911262799, + "grad_norm": 0.5692431926727295, + "learning_rate": 9.971475759725308e-06, + "loss": 0.574, + "step": 1225 + }, + { + "epoch": 0.392278156996587, + "grad_norm": 0.6000913977622986, + "learning_rate": 9.971276829098837e-06, + "loss": 0.5986, + "step": 1226 + }, + { + "epoch": 0.3925981228668942, + "grad_norm": 0.7071651816368103, + "learning_rate": 9.971077209198851e-06, + "loss": 0.5838, + "step": 1227 + }, + { + "epoch": 0.39291808873720135, + "grad_norm": 0.6404964327812195, + "learning_rate": 9.970876900053031e-06, + "loss": 0.5755, + "step": 1228 + }, + { + "epoch": 0.39323805460750855, + "grad_norm": 0.5926943421363831, + "learning_rate": 9.970675901689148e-06, + "loss": 0.5661, + "step": 1229 + }, + { + "epoch": 0.3935580204778157, + "grad_norm": 0.6548840999603271, + "learning_rate": 9.970474214135072e-06, + "loss": 0.6253, + "step": 1230 + }, + { + "epoch": 0.3938779863481229, + "grad_norm": 0.5912249684333801, + "learning_rate": 9.970271837418765e-06, + "loss": 0.5632, + "step": 1231 + }, + { + "epoch": 0.39419795221843, + "grad_norm": 0.5992922782897949, + "learning_rate": 9.97006877156829e-06, + "loss": 0.535, + "step": 1232 + }, + { + "epoch": 0.3945179180887372, + "grad_norm": 0.6178451776504517, + "learning_rate": 9.9698650166118e-06, + "loss": 0.5984, + "step": 1233 + }, + { + "epoch": 0.39483788395904434, + "grad_norm": 0.6016383767127991, + "learning_rate": 9.969660572577548e-06, + "loss": 0.5452, + "step": 1234 + }, + { + "epoch": 0.39515784982935154, + "grad_norm": 0.5990430116653442, + "learning_rate": 9.969455439493877e-06, + "loss": 0.5966, + "step": 1235 + }, + { + "epoch": 0.39547781569965873, + "grad_norm": 0.6854697465896606, + "learning_rate": 9.969249617389232e-06, + "loss": 0.6192, + "step": 1236 + }, + { + "epoch": 0.39579778156996587, + "grad_norm": 0.558914065361023, + "learning_rate": 9.969043106292149e-06, + "loss": 0.589, + "step": 1237 + }, + { + "epoch": 0.39611774744027306, + "grad_norm": 0.6286885738372803, + "learning_rate": 9.96883590623126e-06, + "loss": 0.5764, + "step": 1238 + }, + { + "epoch": 0.3964377133105802, + "grad_norm": 0.6284793615341187, + "learning_rate": 9.968628017235297e-06, + "loss": 0.5925, + "step": 1239 + }, + { + "epoch": 0.3967576791808874, + "grad_norm": 0.6545090079307556, + "learning_rate": 9.96841943933308e-06, + "loss": 0.6065, + "step": 1240 + }, + { + "epoch": 0.3970776450511945, + "grad_norm": 0.5711103677749634, + "learning_rate": 9.968210172553533e-06, + "loss": 0.5231, + "step": 1241 + }, + { + "epoch": 0.3973976109215017, + "grad_norm": 0.6072283983230591, + "learning_rate": 9.968000216925668e-06, + "loss": 0.5538, + "step": 1242 + }, + { + "epoch": 0.39771757679180886, + "grad_norm": 0.6144881248474121, + "learning_rate": 9.967789572478596e-06, + "loss": 0.5628, + "step": 1243 + }, + { + "epoch": 0.39803754266211605, + "grad_norm": 0.6187753677368164, + "learning_rate": 9.967578239241523e-06, + "loss": 0.555, + "step": 1244 + }, + { + "epoch": 0.3983575085324232, + "grad_norm": 0.6248455047607422, + "learning_rate": 9.967366217243752e-06, + "loss": 0.5834, + "step": 1245 + }, + { + "epoch": 0.3986774744027304, + "grad_norm": 0.6310631632804871, + "learning_rate": 9.967153506514677e-06, + "loss": 0.5562, + "step": 1246 + }, + { + "epoch": 0.3989974402730375, + "grad_norm": 0.6546112298965454, + "learning_rate": 9.966940107083795e-06, + "loss": 0.5814, + "step": 1247 + }, + { + "epoch": 0.3993174061433447, + "grad_norm": 0.5986621975898743, + "learning_rate": 9.96672601898069e-06, + "loss": 0.5235, + "step": 1248 + }, + { + "epoch": 0.3996373720136519, + "grad_norm": 0.7620200514793396, + "learning_rate": 9.966511242235048e-06, + "loss": 0.5646, + "step": 1249 + }, + { + "epoch": 0.39995733788395904, + "grad_norm": 0.6559775471687317, + "learning_rate": 9.966295776876648e-06, + "loss": 0.5526, + "step": 1250 + }, + { + "epoch": 0.40027730375426623, + "grad_norm": 0.544110119342804, + "learning_rate": 9.966079622935363e-06, + "loss": 0.5504, + "step": 1251 + }, + { + "epoch": 0.40059726962457337, + "grad_norm": 0.6554135680198669, + "learning_rate": 9.965862780441165e-06, + "loss": 0.5796, + "step": 1252 + }, + { + "epoch": 0.40091723549488056, + "grad_norm": 0.6634418368339539, + "learning_rate": 9.965645249424118e-06, + "loss": 0.6091, + "step": 1253 + }, + { + "epoch": 0.4012372013651877, + "grad_norm": 0.57181316614151, + "learning_rate": 9.965427029914383e-06, + "loss": 0.5736, + "step": 1254 + }, + { + "epoch": 0.4015571672354949, + "grad_norm": 0.6040608286857605, + "learning_rate": 9.965208121942217e-06, + "loss": 0.5524, + "step": 1255 + }, + { + "epoch": 0.40187713310580203, + "grad_norm": 0.7191135287284851, + "learning_rate": 9.964988525537971e-06, + "loss": 0.5519, + "step": 1256 + }, + { + "epoch": 0.4021970989761092, + "grad_norm": 0.599594235420227, + "learning_rate": 9.964768240732094e-06, + "loss": 0.5674, + "step": 1257 + }, + { + "epoch": 0.40251706484641636, + "grad_norm": 0.6317551136016846, + "learning_rate": 9.964547267555127e-06, + "loss": 0.5634, + "step": 1258 + }, + { + "epoch": 0.40283703071672355, + "grad_norm": 0.6445105075836182, + "learning_rate": 9.964325606037709e-06, + "loss": 0.5879, + "step": 1259 + }, + { + "epoch": 0.4031569965870307, + "grad_norm": 0.6589673161506653, + "learning_rate": 9.964103256210573e-06, + "loss": 0.5316, + "step": 1260 + }, + { + "epoch": 0.4034769624573379, + "grad_norm": 0.6195188164710999, + "learning_rate": 9.96388021810455e-06, + "loss": 0.6039, + "step": 1261 + }, + { + "epoch": 0.4037969283276451, + "grad_norm": 0.6745509505271912, + "learning_rate": 9.963656491750562e-06, + "loss": 0.559, + "step": 1262 + }, + { + "epoch": 0.4041168941979522, + "grad_norm": 0.5846664905548096, + "learning_rate": 9.963432077179629e-06, + "loss": 0.5752, + "step": 1263 + }, + { + "epoch": 0.4044368600682594, + "grad_norm": 0.6875172853469849, + "learning_rate": 9.963206974422868e-06, + "loss": 0.5997, + "step": 1264 + }, + { + "epoch": 0.40475682593856654, + "grad_norm": 0.6228593587875366, + "learning_rate": 9.962981183511491e-06, + "loss": 0.5657, + "step": 1265 + }, + { + "epoch": 0.40507679180887374, + "grad_norm": 0.5766012072563171, + "learning_rate": 9.9627547044768e-06, + "loss": 0.5441, + "step": 1266 + }, + { + "epoch": 0.4053967576791809, + "grad_norm": 0.6164464354515076, + "learning_rate": 9.9625275373502e-06, + "loss": 0.5954, + "step": 1267 + }, + { + "epoch": 0.40571672354948807, + "grad_norm": 0.6023998856544495, + "learning_rate": 9.962299682163185e-06, + "loss": 0.5657, + "step": 1268 + }, + { + "epoch": 0.4060366894197952, + "grad_norm": 0.6143546104431152, + "learning_rate": 9.962071138947352e-06, + "loss": 0.5329, + "step": 1269 + }, + { + "epoch": 0.4063566552901024, + "grad_norm": 0.6571676731109619, + "learning_rate": 9.961841907734385e-06, + "loss": 0.5595, + "step": 1270 + }, + { + "epoch": 0.40667662116040953, + "grad_norm": 0.591964602470398, + "learning_rate": 9.961611988556067e-06, + "loss": 0.5509, + "step": 1271 + }, + { + "epoch": 0.4069965870307167, + "grad_norm": 0.6351963877677917, + "learning_rate": 9.96138138144428e-06, + "loss": 0.5731, + "step": 1272 + }, + { + "epoch": 0.40731655290102387, + "grad_norm": 0.598535418510437, + "learning_rate": 9.961150086430995e-06, + "loss": 0.5716, + "step": 1273 + }, + { + "epoch": 0.40763651877133106, + "grad_norm": 0.8259955644607544, + "learning_rate": 9.960918103548281e-06, + "loss": 0.6108, + "step": 1274 + }, + { + "epoch": 0.40795648464163825, + "grad_norm": 0.6393885016441345, + "learning_rate": 9.960685432828305e-06, + "loss": 0.5721, + "step": 1275 + }, + { + "epoch": 0.4082764505119454, + "grad_norm": 0.7137302756309509, + "learning_rate": 9.960452074303327e-06, + "loss": 0.5791, + "step": 1276 + }, + { + "epoch": 0.4085964163822526, + "grad_norm": 0.7044225335121155, + "learning_rate": 9.960218028005699e-06, + "loss": 0.5997, + "step": 1277 + }, + { + "epoch": 0.4089163822525597, + "grad_norm": 0.6670733094215393, + "learning_rate": 9.959983293967875e-06, + "loss": 0.5852, + "step": 1278 + }, + { + "epoch": 0.4092363481228669, + "grad_norm": 0.6231539845466614, + "learning_rate": 9.9597478722224e-06, + "loss": 0.5777, + "step": 1279 + }, + { + "epoch": 0.40955631399317405, + "grad_norm": 0.6410960555076599, + "learning_rate": 9.959511762801916e-06, + "loss": 0.5573, + "step": 1280 + }, + { + "epoch": 0.40987627986348124, + "grad_norm": 0.67142254114151, + "learning_rate": 9.95927496573916e-06, + "loss": 0.5713, + "step": 1281 + }, + { + "epoch": 0.4101962457337884, + "grad_norm": 0.6224493980407715, + "learning_rate": 9.959037481066964e-06, + "loss": 0.5823, + "step": 1282 + }, + { + "epoch": 0.41051621160409557, + "grad_norm": 0.7081955075263977, + "learning_rate": 9.958799308818254e-06, + "loss": 0.5603, + "step": 1283 + }, + { + "epoch": 0.4108361774744027, + "grad_norm": 0.6348280310630798, + "learning_rate": 9.958560449026055e-06, + "loss": 0.594, + "step": 1284 + }, + { + "epoch": 0.4111561433447099, + "grad_norm": 0.689378023147583, + "learning_rate": 9.958320901723487e-06, + "loss": 0.5512, + "step": 1285 + }, + { + "epoch": 0.41147610921501704, + "grad_norm": 0.7600042223930359, + "learning_rate": 9.958080666943757e-06, + "loss": 0.5554, + "step": 1286 + }, + { + "epoch": 0.41179607508532423, + "grad_norm": 0.5788906216621399, + "learning_rate": 9.95783974472018e-06, + "loss": 0.5857, + "step": 1287 + }, + { + "epoch": 0.4121160409556314, + "grad_norm": 0.6882662177085876, + "learning_rate": 9.957598135086157e-06, + "loss": 0.5523, + "step": 1288 + }, + { + "epoch": 0.41243600682593856, + "grad_norm": 0.7677709460258484, + "learning_rate": 9.957355838075188e-06, + "loss": 0.556, + "step": 1289 + }, + { + "epoch": 0.41275597269624575, + "grad_norm": 0.6081621646881104, + "learning_rate": 9.95711285372087e-06, + "loss": 0.5768, + "step": 1290 + }, + { + "epoch": 0.4130759385665529, + "grad_norm": 0.6662126779556274, + "learning_rate": 9.95686918205689e-06, + "loss": 0.6074, + "step": 1291 + }, + { + "epoch": 0.4133959044368601, + "grad_norm": 0.7383588552474976, + "learning_rate": 9.956624823117036e-06, + "loss": 0.6067, + "step": 1292 + }, + { + "epoch": 0.4137158703071672, + "grad_norm": 0.6374850869178772, + "learning_rate": 9.956379776935185e-06, + "loss": 0.5683, + "step": 1293 + }, + { + "epoch": 0.4140358361774744, + "grad_norm": 0.6523234844207764, + "learning_rate": 9.956134043545317e-06, + "loss": 0.6137, + "step": 1294 + }, + { + "epoch": 0.41435580204778155, + "grad_norm": 0.7293864488601685, + "learning_rate": 9.955887622981501e-06, + "loss": 0.599, + "step": 1295 + }, + { + "epoch": 0.41467576791808874, + "grad_norm": 0.7201833724975586, + "learning_rate": 9.955640515277905e-06, + "loss": 0.5998, + "step": 1296 + }, + { + "epoch": 0.4149957337883959, + "grad_norm": 0.6556185483932495, + "learning_rate": 9.955392720468789e-06, + "loss": 0.5898, + "step": 1297 + }, + { + "epoch": 0.4153156996587031, + "grad_norm": 0.6244340538978577, + "learning_rate": 9.95514423858851e-06, + "loss": 0.6161, + "step": 1298 + }, + { + "epoch": 0.4156356655290102, + "grad_norm": 0.7177112698554993, + "learning_rate": 9.954895069671523e-06, + "loss": 0.5785, + "step": 1299 + }, + { + "epoch": 0.4159556313993174, + "grad_norm": 0.6149005889892578, + "learning_rate": 9.954645213752372e-06, + "loss": 0.5764, + "step": 1300 + }, + { + "epoch": 0.4162755972696246, + "grad_norm": 0.6709874868392944, + "learning_rate": 9.954394670865701e-06, + "loss": 0.6015, + "step": 1301 + }, + { + "epoch": 0.41659556313993173, + "grad_norm": 0.5673186779022217, + "learning_rate": 9.95414344104625e-06, + "loss": 0.5538, + "step": 1302 + }, + { + "epoch": 0.4169155290102389, + "grad_norm": 0.5811302065849304, + "learning_rate": 9.953891524328851e-06, + "loss": 0.5742, + "step": 1303 + }, + { + "epoch": 0.41723549488054607, + "grad_norm": 0.7059139609336853, + "learning_rate": 9.953638920748432e-06, + "loss": 0.563, + "step": 1304 + }, + { + "epoch": 0.41755546075085326, + "grad_norm": 0.7833511233329773, + "learning_rate": 9.953385630340016e-06, + "loss": 0.615, + "step": 1305 + }, + { + "epoch": 0.4178754266211604, + "grad_norm": 0.7137935161590576, + "learning_rate": 9.953131653138726e-06, + "loss": 0.5711, + "step": 1306 + }, + { + "epoch": 0.4181953924914676, + "grad_norm": 0.5633505582809448, + "learning_rate": 9.952876989179772e-06, + "loss": 0.5677, + "step": 1307 + }, + { + "epoch": 0.4185153583617747, + "grad_norm": 0.5868337750434875, + "learning_rate": 9.952621638498465e-06, + "loss": 0.5379, + "step": 1308 + }, + { + "epoch": 0.4188353242320819, + "grad_norm": 0.6548082828521729, + "learning_rate": 9.95236560113021e-06, + "loss": 0.6171, + "step": 1309 + }, + { + "epoch": 0.41915529010238906, + "grad_norm": 0.6385195851325989, + "learning_rate": 9.952108877110505e-06, + "loss": 0.5671, + "step": 1310 + }, + { + "epoch": 0.41947525597269625, + "grad_norm": 0.6664000153541565, + "learning_rate": 9.951851466474949e-06, + "loss": 0.5631, + "step": 1311 + }, + { + "epoch": 0.4197952218430034, + "grad_norm": 0.6223545074462891, + "learning_rate": 9.95159336925923e-06, + "loss": 0.5661, + "step": 1312 + }, + { + "epoch": 0.4201151877133106, + "grad_norm": 0.6739969253540039, + "learning_rate": 9.95133458549913e-06, + "loss": 0.5956, + "step": 1313 + }, + { + "epoch": 0.42043515358361777, + "grad_norm": 0.6768685579299927, + "learning_rate": 9.951075115230537e-06, + "loss": 0.5835, + "step": 1314 + }, + { + "epoch": 0.4207551194539249, + "grad_norm": 0.615283727645874, + "learning_rate": 9.950814958489421e-06, + "loss": 0.5621, + "step": 1315 + }, + { + "epoch": 0.4210750853242321, + "grad_norm": 0.5856695175170898, + "learning_rate": 9.950554115311855e-06, + "loss": 0.5479, + "step": 1316 + }, + { + "epoch": 0.42139505119453924, + "grad_norm": 0.6586756110191345, + "learning_rate": 9.950292585734007e-06, + "loss": 0.5547, + "step": 1317 + }, + { + "epoch": 0.42171501706484643, + "grad_norm": 0.6436223387718201, + "learning_rate": 9.950030369792133e-06, + "loss": 0.5959, + "step": 1318 + }, + { + "epoch": 0.42203498293515357, + "grad_norm": 0.5667495727539062, + "learning_rate": 9.949767467522597e-06, + "loss": 0.5868, + "step": 1319 + }, + { + "epoch": 0.42235494880546076, + "grad_norm": 0.6080712080001831, + "learning_rate": 9.949503878961846e-06, + "loss": 0.5511, + "step": 1320 + }, + { + "epoch": 0.4226749146757679, + "grad_norm": 0.6536353826522827, + "learning_rate": 9.949239604146427e-06, + "loss": 0.5547, + "step": 1321 + }, + { + "epoch": 0.4229948805460751, + "grad_norm": 0.6362671256065369, + "learning_rate": 9.948974643112984e-06, + "loss": 0.5693, + "step": 1322 + }, + { + "epoch": 0.42331484641638223, + "grad_norm": 0.6655951738357544, + "learning_rate": 9.948708995898251e-06, + "loss": 0.6167, + "step": 1323 + }, + { + "epoch": 0.4236348122866894, + "grad_norm": 0.678941011428833, + "learning_rate": 9.948442662539064e-06, + "loss": 0.5996, + "step": 1324 + }, + { + "epoch": 0.4239547781569966, + "grad_norm": 0.6312676668167114, + "learning_rate": 9.948175643072347e-06, + "loss": 0.575, + "step": 1325 + }, + { + "epoch": 0.42427474402730375, + "grad_norm": 0.663619875907898, + "learning_rate": 9.947907937535127e-06, + "loss": 0.5598, + "step": 1326 + }, + { + "epoch": 0.42459470989761094, + "grad_norm": 0.6370262503623962, + "learning_rate": 9.947639545964516e-06, + "loss": 0.5459, + "step": 1327 + }, + { + "epoch": 0.4249146757679181, + "grad_norm": 0.6334279775619507, + "learning_rate": 9.947370468397731e-06, + "loss": 0.5924, + "step": 1328 + }, + { + "epoch": 0.4252346416382253, + "grad_norm": 0.6385119557380676, + "learning_rate": 9.947100704872079e-06, + "loss": 0.521, + "step": 1329 + }, + { + "epoch": 0.4255546075085324, + "grad_norm": 0.5783963799476624, + "learning_rate": 9.946830255424964e-06, + "loss": 0.5408, + "step": 1330 + }, + { + "epoch": 0.4258745733788396, + "grad_norm": 0.6530171036720276, + "learning_rate": 9.94655912009388e-06, + "loss": 0.567, + "step": 1331 + }, + { + "epoch": 0.42619453924914674, + "grad_norm": 0.6820838451385498, + "learning_rate": 9.946287298916426e-06, + "loss": 0.6293, + "step": 1332 + }, + { + "epoch": 0.42651450511945393, + "grad_norm": 0.6430705785751343, + "learning_rate": 9.946014791930285e-06, + "loss": 0.5808, + "step": 1333 + }, + { + "epoch": 0.4268344709897611, + "grad_norm": 0.69588702917099, + "learning_rate": 9.945741599173244e-06, + "loss": 0.5724, + "step": 1334 + }, + { + "epoch": 0.42715443686006827, + "grad_norm": 0.6550391912460327, + "learning_rate": 9.94546772068318e-06, + "loss": 0.5921, + "step": 1335 + }, + { + "epoch": 0.4274744027303754, + "grad_norm": 0.6094030737876892, + "learning_rate": 9.945193156498068e-06, + "loss": 0.5448, + "step": 1336 + }, + { + "epoch": 0.4277943686006826, + "grad_norm": 0.7154574394226074, + "learning_rate": 9.944917906655976e-06, + "loss": 0.615, + "step": 1337 + }, + { + "epoch": 0.4281143344709898, + "grad_norm": 0.5963408946990967, + "learning_rate": 9.944641971195065e-06, + "loss": 0.5823, + "step": 1338 + }, + { + "epoch": 0.4284343003412969, + "grad_norm": 0.6101605892181396, + "learning_rate": 9.944365350153598e-06, + "loss": 0.5914, + "step": 1339 + }, + { + "epoch": 0.4287542662116041, + "grad_norm": 0.655142068862915, + "learning_rate": 9.944088043569925e-06, + "loss": 0.5732, + "step": 1340 + }, + { + "epoch": 0.42907423208191126, + "grad_norm": 0.6475589275360107, + "learning_rate": 9.9438100514825e-06, + "loss": 0.5754, + "step": 1341 + }, + { + "epoch": 0.42939419795221845, + "grad_norm": 0.6456246376037598, + "learning_rate": 9.943531373929863e-06, + "loss": 0.5733, + "step": 1342 + }, + { + "epoch": 0.4297141638225256, + "grad_norm": 0.7405171990394592, + "learning_rate": 9.943252010950653e-06, + "loss": 0.5914, + "step": 1343 + }, + { + "epoch": 0.4300341296928328, + "grad_norm": 0.5655635595321655, + "learning_rate": 9.942971962583605e-06, + "loss": 0.5319, + "step": 1344 + }, + { + "epoch": 0.4303540955631399, + "grad_norm": 0.6160438656806946, + "learning_rate": 9.942691228867548e-06, + "loss": 0.5907, + "step": 1345 + }, + { + "epoch": 0.4306740614334471, + "grad_norm": 0.6776102781295776, + "learning_rate": 9.942409809841404e-06, + "loss": 0.5641, + "step": 1346 + }, + { + "epoch": 0.43099402730375425, + "grad_norm": 0.6737133860588074, + "learning_rate": 9.942127705544196e-06, + "loss": 0.5942, + "step": 1347 + }, + { + "epoch": 0.43131399317406144, + "grad_norm": 0.6324644684791565, + "learning_rate": 9.941844916015035e-06, + "loss": 0.601, + "step": 1348 + }, + { + "epoch": 0.4316339590443686, + "grad_norm": 0.620689868927002, + "learning_rate": 9.941561441293134e-06, + "loss": 0.5364, + "step": 1349 + }, + { + "epoch": 0.43195392491467577, + "grad_norm": 0.6571909189224243, + "learning_rate": 9.941277281417792e-06, + "loss": 0.6054, + "step": 1350 + }, + { + "epoch": 0.43227389078498296, + "grad_norm": 0.5578046441078186, + "learning_rate": 9.94099243642841e-06, + "loss": 0.5415, + "step": 1351 + }, + { + "epoch": 0.4325938566552901, + "grad_norm": 0.5772393345832825, + "learning_rate": 9.940706906364484e-06, + "loss": 0.5384, + "step": 1352 + }, + { + "epoch": 0.4329138225255973, + "grad_norm": 0.6861337423324585, + "learning_rate": 9.9404206912656e-06, + "loss": 0.5939, + "step": 1353 + }, + { + "epoch": 0.43323378839590443, + "grad_norm": 0.6963403820991516, + "learning_rate": 9.940133791171445e-06, + "loss": 0.5802, + "step": 1354 + }, + { + "epoch": 0.4335537542662116, + "grad_norm": 0.564021646976471, + "learning_rate": 9.939846206121795e-06, + "loss": 0.5282, + "step": 1355 + }, + { + "epoch": 0.43387372013651876, + "grad_norm": 0.6498206257820129, + "learning_rate": 9.939557936156527e-06, + "loss": 0.5259, + "step": 1356 + }, + { + "epoch": 0.43419368600682595, + "grad_norm": 0.6931615471839905, + "learning_rate": 9.939268981315609e-06, + "loss": 0.5777, + "step": 1357 + }, + { + "epoch": 0.4345136518771331, + "grad_norm": 0.6055909395217896, + "learning_rate": 9.938979341639105e-06, + "loss": 0.5769, + "step": 1358 + }, + { + "epoch": 0.4348336177474403, + "grad_norm": 0.7766366004943848, + "learning_rate": 9.938689017167173e-06, + "loss": 0.573, + "step": 1359 + }, + { + "epoch": 0.4351535836177474, + "grad_norm": 0.6288790106773376, + "learning_rate": 9.938398007940067e-06, + "loss": 0.5826, + "step": 1360 + }, + { + "epoch": 0.4354735494880546, + "grad_norm": 0.6430128216743469, + "learning_rate": 9.938106313998136e-06, + "loss": 0.5781, + "step": 1361 + }, + { + "epoch": 0.43579351535836175, + "grad_norm": 0.6517099142074585, + "learning_rate": 9.937813935381823e-06, + "loss": 0.5561, + "step": 1362 + }, + { + "epoch": 0.43611348122866894, + "grad_norm": 0.6719871163368225, + "learning_rate": 9.93752087213167e-06, + "loss": 0.57, + "step": 1363 + }, + { + "epoch": 0.43643344709897613, + "grad_norm": 0.6535536050796509, + "learning_rate": 9.937227124288306e-06, + "loss": 0.5627, + "step": 1364 + }, + { + "epoch": 0.4367534129692833, + "grad_norm": 0.689322829246521, + "learning_rate": 9.936932691892463e-06, + "loss": 0.5457, + "step": 1365 + }, + { + "epoch": 0.43707337883959047, + "grad_norm": 0.6880068182945251, + "learning_rate": 9.936637574984964e-06, + "loss": 0.5631, + "step": 1366 + }, + { + "epoch": 0.4373933447098976, + "grad_norm": 0.7001758217811584, + "learning_rate": 9.936341773606723e-06, + "loss": 0.6106, + "step": 1367 + }, + { + "epoch": 0.4377133105802048, + "grad_norm": 0.6887964010238647, + "learning_rate": 9.936045287798759e-06, + "loss": 0.5849, + "step": 1368 + }, + { + "epoch": 0.43803327645051193, + "grad_norm": 0.7089389562606812, + "learning_rate": 9.935748117602178e-06, + "loss": 0.5952, + "step": 1369 + }, + { + "epoch": 0.4383532423208191, + "grad_norm": 0.6172699928283691, + "learning_rate": 9.93545026305818e-06, + "loss": 0.5147, + "step": 1370 + }, + { + "epoch": 0.43867320819112626, + "grad_norm": 0.6297218799591064, + "learning_rate": 9.935151724208068e-06, + "loss": 0.572, + "step": 1371 + }, + { + "epoch": 0.43899317406143346, + "grad_norm": 0.7111726999282837, + "learning_rate": 9.93485250109323e-06, + "loss": 0.579, + "step": 1372 + }, + { + "epoch": 0.4393131399317406, + "grad_norm": 0.6565432548522949, + "learning_rate": 9.934552593755158e-06, + "loss": 0.615, + "step": 1373 + }, + { + "epoch": 0.4396331058020478, + "grad_norm": 0.6117081642150879, + "learning_rate": 9.934252002235433e-06, + "loss": 0.5565, + "step": 1374 + }, + { + "epoch": 0.4399530716723549, + "grad_norm": 0.6642066836357117, + "learning_rate": 9.93395072657573e-06, + "loss": 0.5598, + "step": 1375 + }, + { + "epoch": 0.4402730375426621, + "grad_norm": 0.6202808022499084, + "learning_rate": 9.933648766817823e-06, + "loss": 0.5716, + "step": 1376 + }, + { + "epoch": 0.4405930034129693, + "grad_norm": 0.5759875178337097, + "learning_rate": 9.933346123003581e-06, + "loss": 0.5462, + "step": 1377 + }, + { + "epoch": 0.44091296928327645, + "grad_norm": 0.7093290686607361, + "learning_rate": 9.933042795174964e-06, + "loss": 0.5705, + "step": 1378 + }, + { + "epoch": 0.44123293515358364, + "grad_norm": 0.6566049456596375, + "learning_rate": 9.932738783374028e-06, + "loss": 0.5653, + "step": 1379 + }, + { + "epoch": 0.4415529010238908, + "grad_norm": 0.631485641002655, + "learning_rate": 9.932434087642924e-06, + "loss": 0.5672, + "step": 1380 + }, + { + "epoch": 0.44187286689419797, + "grad_norm": 0.6641365885734558, + "learning_rate": 9.932128708023901e-06, + "loss": 0.538, + "step": 1381 + }, + { + "epoch": 0.4421928327645051, + "grad_norm": 0.574982762336731, + "learning_rate": 9.931822644559298e-06, + "loss": 0.5212, + "step": 1382 + }, + { + "epoch": 0.4425127986348123, + "grad_norm": 0.6498568058013916, + "learning_rate": 9.931515897291555e-06, + "loss": 0.5818, + "step": 1383 + }, + { + "epoch": 0.44283276450511944, + "grad_norm": 0.6279152035713196, + "learning_rate": 9.931208466263198e-06, + "loss": 0.5548, + "step": 1384 + }, + { + "epoch": 0.44315273037542663, + "grad_norm": 0.6007976531982422, + "learning_rate": 9.930900351516857e-06, + "loss": 0.5456, + "step": 1385 + }, + { + "epoch": 0.44347269624573377, + "grad_norm": 0.6235578060150146, + "learning_rate": 9.930591553095248e-06, + "loss": 0.6002, + "step": 1386 + }, + { + "epoch": 0.44379266211604096, + "grad_norm": 0.6414409279823303, + "learning_rate": 9.930282071041188e-06, + "loss": 0.5304, + "step": 1387 + }, + { + "epoch": 0.4441126279863481, + "grad_norm": 0.6062139272689819, + "learning_rate": 9.92997190539759e-06, + "loss": 0.5881, + "step": 1388 + }, + { + "epoch": 0.4444325938566553, + "grad_norm": 0.6121253967285156, + "learning_rate": 9.929661056207455e-06, + "loss": 0.6046, + "step": 1389 + }, + { + "epoch": 0.4447525597269625, + "grad_norm": 0.6918778419494629, + "learning_rate": 9.929349523513884e-06, + "loss": 0.5734, + "step": 1390 + }, + { + "epoch": 0.4450725255972696, + "grad_norm": 0.6338698267936707, + "learning_rate": 9.929037307360073e-06, + "loss": 0.5541, + "step": 1391 + }, + { + "epoch": 0.4453924914675768, + "grad_norm": 0.5875884890556335, + "learning_rate": 9.928724407789307e-06, + "loss": 0.5652, + "step": 1392 + }, + { + "epoch": 0.44571245733788395, + "grad_norm": 0.6555237770080566, + "learning_rate": 9.928410824844974e-06, + "loss": 0.5742, + "step": 1393 + }, + { + "epoch": 0.44603242320819114, + "grad_norm": 0.5992658734321594, + "learning_rate": 9.92809655857055e-06, + "loss": 0.5878, + "step": 1394 + }, + { + "epoch": 0.4463523890784983, + "grad_norm": 0.6018292307853699, + "learning_rate": 9.927781609009611e-06, + "loss": 0.5827, + "step": 1395 + }, + { + "epoch": 0.4466723549488055, + "grad_norm": 0.6248888969421387, + "learning_rate": 9.927465976205823e-06, + "loss": 0.5768, + "step": 1396 + }, + { + "epoch": 0.4469923208191126, + "grad_norm": 0.6832228899002075, + "learning_rate": 9.927149660202949e-06, + "loss": 0.5906, + "step": 1397 + }, + { + "epoch": 0.4473122866894198, + "grad_norm": 0.5579515695571899, + "learning_rate": 9.926832661044848e-06, + "loss": 0.5318, + "step": 1398 + }, + { + "epoch": 0.44763225255972694, + "grad_norm": 0.7638499736785889, + "learning_rate": 9.92651497877547e-06, + "loss": 0.557, + "step": 1399 + }, + { + "epoch": 0.44795221843003413, + "grad_norm": 0.5726019740104675, + "learning_rate": 9.926196613438866e-06, + "loss": 0.5835, + "step": 1400 + }, + { + "epoch": 0.44827218430034127, + "grad_norm": 0.669023871421814, + "learning_rate": 9.925877565079172e-06, + "loss": 0.5944, + "step": 1401 + }, + { + "epoch": 0.44859215017064846, + "grad_norm": 0.6804897785186768, + "learning_rate": 9.92555783374063e-06, + "loss": 0.5345, + "step": 1402 + }, + { + "epoch": 0.44891211604095566, + "grad_norm": 0.7297967076301575, + "learning_rate": 9.925237419467567e-06, + "loss": 0.6176, + "step": 1403 + }, + { + "epoch": 0.4492320819112628, + "grad_norm": 0.6806580424308777, + "learning_rate": 9.924916322304413e-06, + "loss": 0.5738, + "step": 1404 + }, + { + "epoch": 0.44955204778157, + "grad_norm": 0.597823977470398, + "learning_rate": 9.924594542295684e-06, + "loss": 0.5565, + "step": 1405 + }, + { + "epoch": 0.4498720136518771, + "grad_norm": 0.6245985627174377, + "learning_rate": 9.924272079485996e-06, + "loss": 0.5889, + "step": 1406 + }, + { + "epoch": 0.4501919795221843, + "grad_norm": 0.6150023937225342, + "learning_rate": 9.923948933920064e-06, + "loss": 0.5652, + "step": 1407 + }, + { + "epoch": 0.45051194539249145, + "grad_norm": 0.6009223461151123, + "learning_rate": 9.923625105642687e-06, + "loss": 0.5729, + "step": 1408 + }, + { + "epoch": 0.45083191126279865, + "grad_norm": 0.5567916631698608, + "learning_rate": 9.923300594698765e-06, + "loss": 0.5315, + "step": 1409 + }, + { + "epoch": 0.4511518771331058, + "grad_norm": 0.612503707408905, + "learning_rate": 9.922975401133292e-06, + "loss": 0.5928, + "step": 1410 + }, + { + "epoch": 0.451471843003413, + "grad_norm": 0.6755174398422241, + "learning_rate": 9.922649524991359e-06, + "loss": 0.5171, + "step": 1411 + }, + { + "epoch": 0.4517918088737201, + "grad_norm": 0.6449228525161743, + "learning_rate": 9.922322966318148e-06, + "loss": 0.6091, + "step": 1412 + }, + { + "epoch": 0.4521117747440273, + "grad_norm": 0.6059353947639465, + "learning_rate": 9.921995725158934e-06, + "loss": 0.5472, + "step": 1413 + }, + { + "epoch": 0.45243174061433444, + "grad_norm": 0.6328505277633667, + "learning_rate": 9.921667801559092e-06, + "loss": 0.5961, + "step": 1414 + }, + { + "epoch": 0.45275170648464164, + "grad_norm": 0.6243288516998291, + "learning_rate": 9.921339195564088e-06, + "loss": 0.5667, + "step": 1415 + }, + { + "epoch": 0.45307167235494883, + "grad_norm": 0.5473213791847229, + "learning_rate": 9.921009907219485e-06, + "loss": 0.5603, + "step": 1416 + }, + { + "epoch": 0.45339163822525597, + "grad_norm": 0.5441519021987915, + "learning_rate": 9.920679936570937e-06, + "loss": 0.6045, + "step": 1417 + }, + { + "epoch": 0.45371160409556316, + "grad_norm": 0.5660945773124695, + "learning_rate": 9.920349283664197e-06, + "loss": 0.5489, + "step": 1418 + }, + { + "epoch": 0.4540315699658703, + "grad_norm": 0.5919684767723083, + "learning_rate": 9.920017948545109e-06, + "loss": 0.5622, + "step": 1419 + }, + { + "epoch": 0.4543515358361775, + "grad_norm": 0.6301105618476868, + "learning_rate": 9.919685931259615e-06, + "loss": 0.5791, + "step": 1420 + }, + { + "epoch": 0.4546715017064846, + "grad_norm": 0.6944140195846558, + "learning_rate": 9.919353231853746e-06, + "loss": 0.5882, + "step": 1421 + }, + { + "epoch": 0.4549914675767918, + "grad_norm": 0.6283155083656311, + "learning_rate": 9.919019850373635e-06, + "loss": 0.5745, + "step": 1422 + }, + { + "epoch": 0.45531143344709896, + "grad_norm": 0.5469874143600464, + "learning_rate": 9.918685786865503e-06, + "loss": 0.53, + "step": 1423 + }, + { + "epoch": 0.45563139931740615, + "grad_norm": 0.5966683030128479, + "learning_rate": 9.91835104137567e-06, + "loss": 0.5867, + "step": 1424 + }, + { + "epoch": 0.4559513651877133, + "grad_norm": 0.6650198101997375, + "learning_rate": 9.918015613950548e-06, + "loss": 0.5862, + "step": 1425 + }, + { + "epoch": 0.4562713310580205, + "grad_norm": 0.6126238107681274, + "learning_rate": 9.917679504636644e-06, + "loss": 0.5622, + "step": 1426 + }, + { + "epoch": 0.4565912969283277, + "grad_norm": 0.6327012777328491, + "learning_rate": 9.91734271348056e-06, + "loss": 0.5256, + "step": 1427 + }, + { + "epoch": 0.4569112627986348, + "grad_norm": 0.608069896697998, + "learning_rate": 9.917005240528994e-06, + "loss": 0.5402, + "step": 1428 + }, + { + "epoch": 0.457231228668942, + "grad_norm": 0.6601377725601196, + "learning_rate": 9.916667085828735e-06, + "loss": 0.562, + "step": 1429 + }, + { + "epoch": 0.45755119453924914, + "grad_norm": 0.7051287889480591, + "learning_rate": 9.916328249426671e-06, + "loss": 0.5794, + "step": 1430 + }, + { + "epoch": 0.45787116040955633, + "grad_norm": 0.5601382255554199, + "learning_rate": 9.915988731369782e-06, + "loss": 0.5622, + "step": 1431 + }, + { + "epoch": 0.45819112627986347, + "grad_norm": 0.5865865349769592, + "learning_rate": 9.91564853170514e-06, + "loss": 0.5529, + "step": 1432 + }, + { + "epoch": 0.45851109215017066, + "grad_norm": 0.6290679574012756, + "learning_rate": 9.915307650479915e-06, + "loss": 0.5504, + "step": 1433 + }, + { + "epoch": 0.4588310580204778, + "grad_norm": 0.6417455077171326, + "learning_rate": 9.91496608774137e-06, + "loss": 0.5786, + "step": 1434 + }, + { + "epoch": 0.459151023890785, + "grad_norm": 0.638545572757721, + "learning_rate": 9.914623843536866e-06, + "loss": 0.5837, + "step": 1435 + }, + { + "epoch": 0.45947098976109213, + "grad_norm": 0.5751261711120605, + "learning_rate": 9.914280917913851e-06, + "loss": 0.5992, + "step": 1436 + }, + { + "epoch": 0.4597909556313993, + "grad_norm": 0.6094741225242615, + "learning_rate": 9.913937310919877e-06, + "loss": 0.5261, + "step": 1437 + }, + { + "epoch": 0.46011092150170646, + "grad_norm": 0.5998948812484741, + "learning_rate": 9.913593022602583e-06, + "loss": 0.5936, + "step": 1438 + }, + { + "epoch": 0.46043088737201365, + "grad_norm": 0.7354366779327393, + "learning_rate": 9.913248053009705e-06, + "loss": 0.6342, + "step": 1439 + }, + { + "epoch": 0.46075085324232085, + "grad_norm": 0.6196147799491882, + "learning_rate": 9.912902402189075e-06, + "loss": 0.576, + "step": 1440 + }, + { + "epoch": 0.461070819112628, + "grad_norm": 0.6172345280647278, + "learning_rate": 9.912556070188615e-06, + "loss": 0.5871, + "step": 1441 + }, + { + "epoch": 0.4613907849829352, + "grad_norm": 0.7067185044288635, + "learning_rate": 9.912209057056345e-06, + "loss": 0.5775, + "step": 1442 + }, + { + "epoch": 0.4617107508532423, + "grad_norm": 0.607168436050415, + "learning_rate": 9.911861362840382e-06, + "loss": 0.579, + "step": 1443 + }, + { + "epoch": 0.4620307167235495, + "grad_norm": 0.6715491414070129, + "learning_rate": 9.911512987588932e-06, + "loss": 0.6184, + "step": 1444 + }, + { + "epoch": 0.46235068259385664, + "grad_norm": 0.6675828695297241, + "learning_rate": 9.911163931350296e-06, + "loss": 0.5802, + "step": 1445 + }, + { + "epoch": 0.46267064846416384, + "grad_norm": 0.5708353519439697, + "learning_rate": 9.910814194172872e-06, + "loss": 0.5371, + "step": 1446 + }, + { + "epoch": 0.462990614334471, + "grad_norm": 0.599196195602417, + "learning_rate": 9.910463776105154e-06, + "loss": 0.5975, + "step": 1447 + }, + { + "epoch": 0.46331058020477817, + "grad_norm": 0.635504961013794, + "learning_rate": 9.910112677195724e-06, + "loss": 0.5694, + "step": 1448 + }, + { + "epoch": 0.4636305460750853, + "grad_norm": 0.6071615815162659, + "learning_rate": 9.909760897493267e-06, + "loss": 0.5395, + "step": 1449 + }, + { + "epoch": 0.4639505119453925, + "grad_norm": 0.6165795922279358, + "learning_rate": 9.909408437046553e-06, + "loss": 0.5666, + "step": 1450 + }, + { + "epoch": 0.46427047781569963, + "grad_norm": 0.6419926285743713, + "learning_rate": 9.909055295904453e-06, + "loss": 0.5628, + "step": 1451 + }, + { + "epoch": 0.4645904436860068, + "grad_norm": 0.6607449650764465, + "learning_rate": 9.908701474115931e-06, + "loss": 0.5526, + "step": 1452 + }, + { + "epoch": 0.464910409556314, + "grad_norm": 0.6067793369293213, + "learning_rate": 9.908346971730045e-06, + "loss": 0.5443, + "step": 1453 + }, + { + "epoch": 0.46523037542662116, + "grad_norm": 0.6271002292633057, + "learning_rate": 9.907991788795945e-06, + "loss": 0.5797, + "step": 1454 + }, + { + "epoch": 0.46555034129692835, + "grad_norm": 0.6588297486305237, + "learning_rate": 9.90763592536288e-06, + "loss": 0.5894, + "step": 1455 + }, + { + "epoch": 0.4658703071672355, + "grad_norm": 0.6160594820976257, + "learning_rate": 9.90727938148019e-06, + "loss": 0.5708, + "step": 1456 + }, + { + "epoch": 0.4661902730375427, + "grad_norm": 0.742313027381897, + "learning_rate": 9.90692215719731e-06, + "loss": 0.5786, + "step": 1457 + }, + { + "epoch": 0.4665102389078498, + "grad_norm": 0.6778494119644165, + "learning_rate": 9.906564252563769e-06, + "loss": 0.5975, + "step": 1458 + }, + { + "epoch": 0.466830204778157, + "grad_norm": 0.6413978934288025, + "learning_rate": 9.906205667629193e-06, + "loss": 0.5772, + "step": 1459 + }, + { + "epoch": 0.46715017064846415, + "grad_norm": 0.6484549641609192, + "learning_rate": 9.905846402443297e-06, + "loss": 0.5799, + "step": 1460 + }, + { + "epoch": 0.46747013651877134, + "grad_norm": 0.5865228176116943, + "learning_rate": 9.905486457055897e-06, + "loss": 0.5508, + "step": 1461 + }, + { + "epoch": 0.4677901023890785, + "grad_norm": 0.6155765652656555, + "learning_rate": 9.905125831516899e-06, + "loss": 0.5502, + "step": 1462 + }, + { + "epoch": 0.46811006825938567, + "grad_norm": 0.5499103665351868, + "learning_rate": 9.904764525876301e-06, + "loss": 0.5334, + "step": 1463 + }, + { + "epoch": 0.4684300341296928, + "grad_norm": 0.5945428609848022, + "learning_rate": 9.904402540184203e-06, + "loss": 0.6077, + "step": 1464 + }, + { + "epoch": 0.46875, + "grad_norm": 0.6026771068572998, + "learning_rate": 9.904039874490791e-06, + "loss": 0.5386, + "step": 1465 + }, + { + "epoch": 0.4690699658703072, + "grad_norm": 0.564043402671814, + "learning_rate": 9.903676528846353e-06, + "loss": 0.5591, + "step": 1466 + }, + { + "epoch": 0.46938993174061433, + "grad_norm": 0.6140351295471191, + "learning_rate": 9.903312503301263e-06, + "loss": 0.5584, + "step": 1467 + }, + { + "epoch": 0.4697098976109215, + "grad_norm": 0.564509391784668, + "learning_rate": 9.902947797905997e-06, + "loss": 0.5536, + "step": 1468 + }, + { + "epoch": 0.47002986348122866, + "grad_norm": 0.5968620777130127, + "learning_rate": 9.90258241271112e-06, + "loss": 0.6038, + "step": 1469 + }, + { + "epoch": 0.47034982935153585, + "grad_norm": 0.7169108986854553, + "learning_rate": 9.902216347767292e-06, + "loss": 0.5891, + "step": 1470 + }, + { + "epoch": 0.470669795221843, + "grad_norm": 0.6527320146560669, + "learning_rate": 9.901849603125271e-06, + "loss": 0.6206, + "step": 1471 + }, + { + "epoch": 0.4709897610921502, + "grad_norm": 0.695573091506958, + "learning_rate": 9.901482178835907e-06, + "loss": 0.5959, + "step": 1472 + }, + { + "epoch": 0.4713097269624573, + "grad_norm": 0.6598088145256042, + "learning_rate": 9.90111407495014e-06, + "loss": 0.5587, + "step": 1473 + }, + { + "epoch": 0.4716296928327645, + "grad_norm": 0.6323714852333069, + "learning_rate": 9.90074529151901e-06, + "loss": 0.5583, + "step": 1474 + }, + { + "epoch": 0.47194965870307165, + "grad_norm": 0.6153364181518555, + "learning_rate": 9.90037582859365e-06, + "loss": 0.5852, + "step": 1475 + }, + { + "epoch": 0.47226962457337884, + "grad_norm": 0.6652483344078064, + "learning_rate": 9.900005686225288e-06, + "loss": 0.587, + "step": 1476 + }, + { + "epoch": 0.472589590443686, + "grad_norm": 0.6001269817352295, + "learning_rate": 9.899634864465242e-06, + "loss": 0.5545, + "step": 1477 + }, + { + "epoch": 0.4729095563139932, + "grad_norm": 0.6979308128356934, + "learning_rate": 9.899263363364926e-06, + "loss": 0.5771, + "step": 1478 + }, + { + "epoch": 0.47322952218430037, + "grad_norm": 0.6638851165771484, + "learning_rate": 9.898891182975852e-06, + "loss": 0.602, + "step": 1479 + }, + { + "epoch": 0.4735494880546075, + "grad_norm": 0.593870997428894, + "learning_rate": 9.898518323349621e-06, + "loss": 0.5897, + "step": 1480 + }, + { + "epoch": 0.4738694539249147, + "grad_norm": 0.5697798132896423, + "learning_rate": 9.898144784537933e-06, + "loss": 0.5594, + "step": 1481 + }, + { + "epoch": 0.47418941979522183, + "grad_norm": 0.642444908618927, + "learning_rate": 9.897770566592578e-06, + "loss": 0.5916, + "step": 1482 + }, + { + "epoch": 0.474509385665529, + "grad_norm": 0.6214845776557922, + "learning_rate": 9.897395669565439e-06, + "loss": 0.5523, + "step": 1483 + }, + { + "epoch": 0.47482935153583616, + "grad_norm": 0.6055586934089661, + "learning_rate": 9.897020093508502e-06, + "loss": 0.5692, + "step": 1484 + }, + { + "epoch": 0.47514931740614336, + "grad_norm": 0.7527763843536377, + "learning_rate": 9.896643838473836e-06, + "loss": 0.5906, + "step": 1485 + }, + { + "epoch": 0.4754692832764505, + "grad_norm": 0.7165666818618774, + "learning_rate": 9.896266904513612e-06, + "loss": 0.596, + "step": 1486 + }, + { + "epoch": 0.4757892491467577, + "grad_norm": 0.5976629257202148, + "learning_rate": 9.89588929168009e-06, + "loss": 0.5478, + "step": 1487 + }, + { + "epoch": 0.4761092150170648, + "grad_norm": 0.6347126960754395, + "learning_rate": 9.89551100002563e-06, + "loss": 0.5746, + "step": 1488 + }, + { + "epoch": 0.476429180887372, + "grad_norm": 0.6473086476325989, + "learning_rate": 9.895132029602679e-06, + "loss": 0.62, + "step": 1489 + }, + { + "epoch": 0.47674914675767915, + "grad_norm": 0.664999783039093, + "learning_rate": 9.894752380463782e-06, + "loss": 0.561, + "step": 1490 + }, + { + "epoch": 0.47706911262798635, + "grad_norm": 0.5583751201629639, + "learning_rate": 9.894372052661581e-06, + "loss": 0.5326, + "step": 1491 + }, + { + "epoch": 0.47738907849829354, + "grad_norm": 0.6377359628677368, + "learning_rate": 9.893991046248806e-06, + "loss": 0.5871, + "step": 1492 + }, + { + "epoch": 0.4777090443686007, + "grad_norm": 0.6565425992012024, + "learning_rate": 9.893609361278284e-06, + "loss": 0.5779, + "step": 1493 + }, + { + "epoch": 0.47802901023890787, + "grad_norm": 0.6273458003997803, + "learning_rate": 9.893226997802936e-06, + "loss": 0.5339, + "step": 1494 + }, + { + "epoch": 0.478348976109215, + "grad_norm": 0.5752052068710327, + "learning_rate": 9.892843955875782e-06, + "loss": 0.5491, + "step": 1495 + }, + { + "epoch": 0.4786689419795222, + "grad_norm": 0.5871261954307556, + "learning_rate": 9.892460235549923e-06, + "loss": 0.6143, + "step": 1496 + }, + { + "epoch": 0.47898890784982934, + "grad_norm": 0.6665970683097839, + "learning_rate": 9.89207583687857e-06, + "loss": 0.591, + "step": 1497 + }, + { + "epoch": 0.47930887372013653, + "grad_norm": 0.6098381876945496, + "learning_rate": 9.891690759915016e-06, + "loss": 0.5687, + "step": 1498 + }, + { + "epoch": 0.47962883959044367, + "grad_norm": 0.6258179545402527, + "learning_rate": 9.891305004712652e-06, + "loss": 0.5505, + "step": 1499 + }, + { + "epoch": 0.47994880546075086, + "grad_norm": 0.6166390180587769, + "learning_rate": 9.890918571324966e-06, + "loss": 0.5425, + "step": 1500 + }, + { + "epoch": 0.480268771331058, + "grad_norm": 0.7183076739311218, + "learning_rate": 9.890531459805535e-06, + "loss": 0.5511, + "step": 1501 + }, + { + "epoch": 0.4805887372013652, + "grad_norm": 0.6725587248802185, + "learning_rate": 9.890143670208034e-06, + "loss": 0.56, + "step": 1502 + }, + { + "epoch": 0.4809087030716723, + "grad_norm": 0.6727641820907593, + "learning_rate": 9.88975520258623e-06, + "loss": 0.5736, + "step": 1503 + }, + { + "epoch": 0.4812286689419795, + "grad_norm": 0.6162008047103882, + "learning_rate": 9.889366056993985e-06, + "loss": 0.5851, + "step": 1504 + }, + { + "epoch": 0.4815486348122867, + "grad_norm": 0.6258243322372437, + "learning_rate": 9.888976233485254e-06, + "loss": 0.5272, + "step": 1505 + }, + { + "epoch": 0.48186860068259385, + "grad_norm": 0.6168959736824036, + "learning_rate": 9.888585732114086e-06, + "loss": 0.5632, + "step": 1506 + }, + { + "epoch": 0.48218856655290104, + "grad_norm": 0.6138393878936768, + "learning_rate": 9.888194552934627e-06, + "loss": 0.5812, + "step": 1507 + }, + { + "epoch": 0.4825085324232082, + "grad_norm": 0.5661769509315491, + "learning_rate": 9.887802696001112e-06, + "loss": 0.562, + "step": 1508 + }, + { + "epoch": 0.4828284982935154, + "grad_norm": 0.6120560765266418, + "learning_rate": 9.887410161367873e-06, + "loss": 0.5828, + "step": 1509 + }, + { + "epoch": 0.4831484641638225, + "grad_norm": 0.6099132895469666, + "learning_rate": 9.887016949089334e-06, + "loss": 0.5826, + "step": 1510 + }, + { + "epoch": 0.4834684300341297, + "grad_norm": 0.6493592858314514, + "learning_rate": 9.886623059220018e-06, + "loss": 0.5573, + "step": 1511 + }, + { + "epoch": 0.48378839590443684, + "grad_norm": 0.6510331034660339, + "learning_rate": 9.886228491814533e-06, + "loss": 0.6168, + "step": 1512 + }, + { + "epoch": 0.48410836177474403, + "grad_norm": 0.6180525422096252, + "learning_rate": 9.885833246927592e-06, + "loss": 0.5541, + "step": 1513 + }, + { + "epoch": 0.48442832764505117, + "grad_norm": 0.6133224964141846, + "learning_rate": 9.885437324613993e-06, + "loss": 0.5828, + "step": 1514 + }, + { + "epoch": 0.48474829351535836, + "grad_norm": 0.6018075942993164, + "learning_rate": 9.88504072492863e-06, + "loss": 0.5662, + "step": 1515 + }, + { + "epoch": 0.48506825938566556, + "grad_norm": 0.5534994006156921, + "learning_rate": 9.884643447926496e-06, + "loss": 0.5982, + "step": 1516 + }, + { + "epoch": 0.4853882252559727, + "grad_norm": 0.5935896039009094, + "learning_rate": 9.88424549366267e-06, + "loss": 0.5697, + "step": 1517 + }, + { + "epoch": 0.4857081911262799, + "grad_norm": 0.5847989320755005, + "learning_rate": 9.88384686219233e-06, + "loss": 0.5552, + "step": 1518 + }, + { + "epoch": 0.486028156996587, + "grad_norm": 0.647017240524292, + "learning_rate": 9.883447553570749e-06, + "loss": 0.5672, + "step": 1519 + }, + { + "epoch": 0.4863481228668942, + "grad_norm": 0.634661078453064, + "learning_rate": 9.88304756785329e-06, + "loss": 0.5312, + "step": 1520 + }, + { + "epoch": 0.48666808873720135, + "grad_norm": 0.5421854257583618, + "learning_rate": 9.88264690509541e-06, + "loss": 0.5452, + "step": 1521 + }, + { + "epoch": 0.48698805460750855, + "grad_norm": 0.6323463320732117, + "learning_rate": 9.882245565352662e-06, + "loss": 0.5914, + "step": 1522 + }, + { + "epoch": 0.4873080204778157, + "grad_norm": 0.6671409010887146, + "learning_rate": 9.881843548680694e-06, + "loss": 0.5507, + "step": 1523 + }, + { + "epoch": 0.4876279863481229, + "grad_norm": 0.6813701391220093, + "learning_rate": 9.881440855135245e-06, + "loss": 0.6005, + "step": 1524 + }, + { + "epoch": 0.48794795221843, + "grad_norm": 0.5512477159500122, + "learning_rate": 9.881037484772148e-06, + "loss": 0.5743, + "step": 1525 + }, + { + "epoch": 0.4882679180887372, + "grad_norm": 0.6768247485160828, + "learning_rate": 9.880633437647333e-06, + "loss": 0.5925, + "step": 1526 + }, + { + "epoch": 0.48858788395904434, + "grad_norm": 0.6516999006271362, + "learning_rate": 9.880228713816819e-06, + "loss": 0.5655, + "step": 1527 + }, + { + "epoch": 0.48890784982935154, + "grad_norm": 0.5851047039031982, + "learning_rate": 9.879823313336723e-06, + "loss": 0.5573, + "step": 1528 + }, + { + "epoch": 0.48922781569965873, + "grad_norm": 0.6633613705635071, + "learning_rate": 9.879417236263255e-06, + "loss": 0.5639, + "step": 1529 + }, + { + "epoch": 0.48954778156996587, + "grad_norm": 0.5621408224105835, + "learning_rate": 9.879010482652716e-06, + "loss": 0.5566, + "step": 1530 + }, + { + "epoch": 0.48986774744027306, + "grad_norm": 0.7079163789749146, + "learning_rate": 9.878603052561503e-06, + "loss": 0.5656, + "step": 1531 + }, + { + "epoch": 0.4901877133105802, + "grad_norm": 0.6103261709213257, + "learning_rate": 9.87819494604611e-06, + "loss": 0.5482, + "step": 1532 + }, + { + "epoch": 0.4905076791808874, + "grad_norm": 0.6498193144798279, + "learning_rate": 9.877786163163119e-06, + "loss": 0.5401, + "step": 1533 + }, + { + "epoch": 0.4908276450511945, + "grad_norm": 0.6479784250259399, + "learning_rate": 9.877376703969207e-06, + "loss": 0.5726, + "step": 1534 + }, + { + "epoch": 0.4911476109215017, + "grad_norm": 0.6524590253829956, + "learning_rate": 9.876966568521148e-06, + "loss": 0.5684, + "step": 1535 + }, + { + "epoch": 0.49146757679180886, + "grad_norm": 0.5773913264274597, + "learning_rate": 9.876555756875807e-06, + "loss": 0.5931, + "step": 1536 + }, + { + "epoch": 0.49178754266211605, + "grad_norm": 0.7117459177970886, + "learning_rate": 9.876144269090144e-06, + "loss": 0.5374, + "step": 1537 + }, + { + "epoch": 0.4921075085324232, + "grad_norm": 0.5589005351066589, + "learning_rate": 9.875732105221211e-06, + "loss": 0.5523, + "step": 1538 + }, + { + "epoch": 0.4924274744027304, + "grad_norm": 0.6223816275596619, + "learning_rate": 9.875319265326158e-06, + "loss": 0.5407, + "step": 1539 + }, + { + "epoch": 0.4927474402730375, + "grad_norm": 0.6707935333251953, + "learning_rate": 9.874905749462222e-06, + "loss": 0.5952, + "step": 1540 + }, + { + "epoch": 0.4930674061433447, + "grad_norm": 0.5833324790000916, + "learning_rate": 9.87449155768674e-06, + "loss": 0.5847, + "step": 1541 + }, + { + "epoch": 0.4933873720136519, + "grad_norm": 0.6765121817588806, + "learning_rate": 9.874076690057139e-06, + "loss": 0.5881, + "step": 1542 + }, + { + "epoch": 0.49370733788395904, + "grad_norm": 0.6047101020812988, + "learning_rate": 9.87366114663094e-06, + "loss": 0.58, + "step": 1543 + }, + { + "epoch": 0.49402730375426623, + "grad_norm": 0.6373699307441711, + "learning_rate": 9.873244927465762e-06, + "loss": 0.5745, + "step": 1544 + }, + { + "epoch": 0.49434726962457337, + "grad_norm": 0.6877090334892273, + "learning_rate": 9.87282803261931e-06, + "loss": 0.5312, + "step": 1545 + }, + { + "epoch": 0.49466723549488056, + "grad_norm": 0.6187264919281006, + "learning_rate": 9.872410462149392e-06, + "loss": 0.5746, + "step": 1546 + }, + { + "epoch": 0.4949872013651877, + "grad_norm": 0.5987083911895752, + "learning_rate": 9.8719922161139e-06, + "loss": 0.5413, + "step": 1547 + }, + { + "epoch": 0.4953071672354949, + "grad_norm": 0.662311315536499, + "learning_rate": 9.871573294570825e-06, + "loss": 0.6106, + "step": 1548 + }, + { + "epoch": 0.49562713310580203, + "grad_norm": 0.6173791289329529, + "learning_rate": 9.871153697578254e-06, + "loss": 0.586, + "step": 1549 + }, + { + "epoch": 0.4959470989761092, + "grad_norm": 0.587100625038147, + "learning_rate": 9.870733425194362e-06, + "loss": 0.5345, + "step": 1550 + }, + { + "epoch": 0.49626706484641636, + "grad_norm": 0.6461461782455444, + "learning_rate": 9.870312477477421e-06, + "loss": 0.5861, + "step": 1551 + }, + { + "epoch": 0.49658703071672355, + "grad_norm": 0.6375117897987366, + "learning_rate": 9.869890854485795e-06, + "loss": 0.6194, + "step": 1552 + }, + { + "epoch": 0.4969069965870307, + "grad_norm": 0.6479952931404114, + "learning_rate": 9.869468556277945e-06, + "loss": 0.5479, + "step": 1553 + }, + { + "epoch": 0.4972269624573379, + "grad_norm": 0.6575143337249756, + "learning_rate": 9.869045582912419e-06, + "loss": 0.5822, + "step": 1554 + }, + { + "epoch": 0.4975469283276451, + "grad_norm": 0.5700595378875732, + "learning_rate": 9.868621934447869e-06, + "loss": 0.5529, + "step": 1555 + }, + { + "epoch": 0.4978668941979522, + "grad_norm": 0.62436842918396, + "learning_rate": 9.868197610943028e-06, + "loss": 0.5779, + "step": 1556 + }, + { + "epoch": 0.4981868600682594, + "grad_norm": 0.6780308485031128, + "learning_rate": 9.867772612456731e-06, + "loss": 0.5869, + "step": 1557 + }, + { + "epoch": 0.49850682593856654, + "grad_norm": 0.6406908631324768, + "learning_rate": 9.867346939047909e-06, + "loss": 0.5717, + "step": 1558 + }, + { + "epoch": 0.49882679180887374, + "grad_norm": 0.6032212376594543, + "learning_rate": 9.866920590775575e-06, + "loss": 0.5775, + "step": 1559 + }, + { + "epoch": 0.4991467576791809, + "grad_norm": 0.5806117653846741, + "learning_rate": 9.866493567698847e-06, + "loss": 0.5663, + "step": 1560 + }, + { + "epoch": 0.49946672354948807, + "grad_norm": 0.5774424076080322, + "learning_rate": 9.866065869876932e-06, + "loss": 0.5421, + "step": 1561 + }, + { + "epoch": 0.4997866894197952, + "grad_norm": 0.5617887377738953, + "learning_rate": 9.86563749736913e-06, + "loss": 0.5653, + "step": 1562 + }, + { + "epoch": 0.5001066552901023, + "grad_norm": 0.6689897179603577, + "learning_rate": 9.865208450234834e-06, + "loss": 0.5195, + "step": 1563 + }, + { + "epoch": 0.5004266211604096, + "grad_norm": 0.6852822303771973, + "learning_rate": 9.864778728533535e-06, + "loss": 0.5569, + "step": 1564 + }, + { + "epoch": 0.5007465870307167, + "grad_norm": 0.5707159638404846, + "learning_rate": 9.864348332324811e-06, + "loss": 0.5549, + "step": 1565 + }, + { + "epoch": 0.5010665529010239, + "grad_norm": 0.7322942018508911, + "learning_rate": 9.86391726166834e-06, + "loss": 0.6051, + "step": 1566 + }, + { + "epoch": 0.5013865187713311, + "grad_norm": 0.602660596370697, + "learning_rate": 9.863485516623889e-06, + "loss": 0.5289, + "step": 1567 + }, + { + "epoch": 0.5017064846416383, + "grad_norm": 0.6694923043251038, + "learning_rate": 9.86305309725132e-06, + "loss": 0.5844, + "step": 1568 + }, + { + "epoch": 0.5020264505119454, + "grad_norm": 0.6712828874588013, + "learning_rate": 9.86262000361059e-06, + "loss": 0.6245, + "step": 1569 + }, + { + "epoch": 0.5023464163822525, + "grad_norm": 0.5861441493034363, + "learning_rate": 9.862186235761746e-06, + "loss": 0.582, + "step": 1570 + }, + { + "epoch": 0.5026663822525598, + "grad_norm": 0.5756668448448181, + "learning_rate": 9.86175179376493e-06, + "loss": 0.5677, + "step": 1571 + }, + { + "epoch": 0.5029863481228669, + "grad_norm": 0.6047782301902771, + "learning_rate": 9.861316677680379e-06, + "loss": 0.5764, + "step": 1572 + }, + { + "epoch": 0.503306313993174, + "grad_norm": 0.6280213594436646, + "learning_rate": 9.860880887568423e-06, + "loss": 0.613, + "step": 1573 + }, + { + "epoch": 0.5036262798634812, + "grad_norm": 0.645076334476471, + "learning_rate": 9.860444423489485e-06, + "loss": 0.5216, + "step": 1574 + }, + { + "epoch": 0.5039462457337884, + "grad_norm": 0.6336789727210999, + "learning_rate": 9.860007285504079e-06, + "loss": 0.5584, + "step": 1575 + }, + { + "epoch": 0.5042662116040956, + "grad_norm": 0.5808457732200623, + "learning_rate": 9.859569473672816e-06, + "loss": 0.5566, + "step": 1576 + }, + { + "epoch": 0.5045861774744027, + "grad_norm": 0.5337081551551819, + "learning_rate": 9.859130988056402e-06, + "loss": 0.5731, + "step": 1577 + }, + { + "epoch": 0.5049061433447098, + "grad_norm": 0.6150102019309998, + "learning_rate": 9.858691828715627e-06, + "loss": 0.56, + "step": 1578 + }, + { + "epoch": 0.5052261092150171, + "grad_norm": 0.5708062052726746, + "learning_rate": 9.858251995711388e-06, + "loss": 0.5633, + "step": 1579 + }, + { + "epoch": 0.5055460750853242, + "grad_norm": 0.6003345251083374, + "learning_rate": 9.857811489104663e-06, + "loss": 0.5849, + "step": 1580 + }, + { + "epoch": 0.5058660409556314, + "grad_norm": 0.5903301239013672, + "learning_rate": 9.857370308956532e-06, + "loss": 0.5836, + "step": 1581 + }, + { + "epoch": 0.5061860068259386, + "grad_norm": 0.5505856275558472, + "learning_rate": 9.856928455328163e-06, + "loss": 0.5519, + "step": 1582 + }, + { + "epoch": 0.5065059726962458, + "grad_norm": 0.6287134289741516, + "learning_rate": 9.856485928280822e-06, + "loss": 0.5945, + "step": 1583 + }, + { + "epoch": 0.5068259385665529, + "grad_norm": 0.6269447207450867, + "learning_rate": 9.856042727875864e-06, + "loss": 0.6088, + "step": 1584 + }, + { + "epoch": 0.50714590443686, + "grad_norm": 0.6402657628059387, + "learning_rate": 9.85559885417474e-06, + "loss": 0.5714, + "step": 1585 + }, + { + "epoch": 0.5074658703071673, + "grad_norm": 0.6432064771652222, + "learning_rate": 9.855154307238993e-06, + "loss": 0.5953, + "step": 1586 + }, + { + "epoch": 0.5077858361774744, + "grad_norm": 0.5881626605987549, + "learning_rate": 9.854709087130261e-06, + "loss": 0.5384, + "step": 1587 + }, + { + "epoch": 0.5081058020477816, + "grad_norm": 0.6792723536491394, + "learning_rate": 9.854263193910274e-06, + "loss": 0.6163, + "step": 1588 + }, + { + "epoch": 0.5084257679180887, + "grad_norm": 0.6320458054542542, + "learning_rate": 9.853816627640854e-06, + "loss": 0.6078, + "step": 1589 + }, + { + "epoch": 0.5087457337883959, + "grad_norm": 0.6325360536575317, + "learning_rate": 9.85336938838392e-06, + "loss": 0.6139, + "step": 1590 + }, + { + "epoch": 0.5090656996587031, + "grad_norm": 0.5258231163024902, + "learning_rate": 9.852921476201482e-06, + "loss": 0.5716, + "step": 1591 + }, + { + "epoch": 0.5093856655290102, + "grad_norm": 0.575778067111969, + "learning_rate": 9.852472891155642e-06, + "loss": 0.5962, + "step": 1592 + }, + { + "epoch": 0.5097056313993175, + "grad_norm": 0.6357591152191162, + "learning_rate": 9.852023633308599e-06, + "loss": 0.5391, + "step": 1593 + }, + { + "epoch": 0.5100255972696246, + "grad_norm": 0.6247155070304871, + "learning_rate": 9.851573702722641e-06, + "loss": 0.5548, + "step": 1594 + }, + { + "epoch": 0.5103455631399317, + "grad_norm": 0.6489351987838745, + "learning_rate": 9.851123099460153e-06, + "loss": 0.5942, + "step": 1595 + }, + { + "epoch": 0.5106655290102389, + "grad_norm": 0.5720632672309875, + "learning_rate": 9.850671823583614e-06, + "loss": 0.5263, + "step": 1596 + }, + { + "epoch": 0.5109854948805461, + "grad_norm": 0.6225847601890564, + "learning_rate": 9.850219875155586e-06, + "loss": 0.5797, + "step": 1597 + }, + { + "epoch": 0.5113054607508533, + "grad_norm": 0.669680655002594, + "learning_rate": 9.849767254238741e-06, + "loss": 0.5773, + "step": 1598 + }, + { + "epoch": 0.5116254266211604, + "grad_norm": 0.6217535138130188, + "learning_rate": 9.84931396089583e-06, + "loss": 0.53, + "step": 1599 + }, + { + "epoch": 0.5119453924914675, + "grad_norm": 0.5527267456054688, + "learning_rate": 9.848859995189705e-06, + "loss": 0.5335, + "step": 1600 + }, + { + "epoch": 0.5122653583617748, + "grad_norm": 0.640916645526886, + "learning_rate": 9.84840535718331e-06, + "loss": 0.5725, + "step": 1601 + }, + { + "epoch": 0.5125853242320819, + "grad_norm": 0.6258885264396667, + "learning_rate": 9.847950046939679e-06, + "loss": 0.5809, + "step": 1602 + }, + { + "epoch": 0.512905290102389, + "grad_norm": 0.6089369058609009, + "learning_rate": 9.84749406452194e-06, + "loss": 0.5531, + "step": 1603 + }, + { + "epoch": 0.5132252559726962, + "grad_norm": 0.6907296180725098, + "learning_rate": 9.847037409993318e-06, + "loss": 0.5515, + "step": 1604 + }, + { + "epoch": 0.5135452218430034, + "grad_norm": 0.6565805077552795, + "learning_rate": 9.846580083417128e-06, + "loss": 0.5793, + "step": 1605 + }, + { + "epoch": 0.5138651877133106, + "grad_norm": 0.5883509516716003, + "learning_rate": 9.84612208485678e-06, + "loss": 0.5424, + "step": 1606 + }, + { + "epoch": 0.5141851535836177, + "grad_norm": 0.6544880867004395, + "learning_rate": 9.845663414375772e-06, + "loss": 0.569, + "step": 1607 + }, + { + "epoch": 0.514505119453925, + "grad_norm": 0.6848540306091309, + "learning_rate": 9.845204072037705e-06, + "loss": 0.573, + "step": 1608 + }, + { + "epoch": 0.5148250853242321, + "grad_norm": 0.6440815329551697, + "learning_rate": 9.844744057906263e-06, + "loss": 0.5743, + "step": 1609 + }, + { + "epoch": 0.5151450511945392, + "grad_norm": 0.5966008901596069, + "learning_rate": 9.844283372045231e-06, + "loss": 0.5526, + "step": 1610 + }, + { + "epoch": 0.5154650170648464, + "grad_norm": 0.6927117705345154, + "learning_rate": 9.843822014518479e-06, + "loss": 0.581, + "step": 1611 + }, + { + "epoch": 0.5157849829351536, + "grad_norm": 0.6129295229911804, + "learning_rate": 9.843359985389978e-06, + "loss": 0.6324, + "step": 1612 + }, + { + "epoch": 0.5161049488054608, + "grad_norm": 0.5667604207992554, + "learning_rate": 9.84289728472379e-06, + "loss": 0.5525, + "step": 1613 + }, + { + "epoch": 0.5164249146757679, + "grad_norm": 0.6120064854621887, + "learning_rate": 9.842433912584066e-06, + "loss": 0.6035, + "step": 1614 + }, + { + "epoch": 0.516744880546075, + "grad_norm": 0.5590425133705139, + "learning_rate": 9.841969869035055e-06, + "loss": 0.5454, + "step": 1615 + }, + { + "epoch": 0.5170648464163823, + "grad_norm": 0.5539948344230652, + "learning_rate": 9.841505154141095e-06, + "loss": 0.5396, + "step": 1616 + }, + { + "epoch": 0.5173848122866894, + "grad_norm": 0.5885478258132935, + "learning_rate": 9.841039767966622e-06, + "loss": 0.5496, + "step": 1617 + }, + { + "epoch": 0.5177047781569966, + "grad_norm": 0.5543431043624878, + "learning_rate": 9.840573710576163e-06, + "loss": 0.5502, + "step": 1618 + }, + { + "epoch": 0.5180247440273038, + "grad_norm": 0.5092489123344421, + "learning_rate": 9.840106982034335e-06, + "loss": 0.5557, + "step": 1619 + }, + { + "epoch": 0.518344709897611, + "grad_norm": 0.5805311799049377, + "learning_rate": 9.83963958240585e-06, + "loss": 0.5748, + "step": 1620 + }, + { + "epoch": 0.5186646757679181, + "grad_norm": 0.5589832663536072, + "learning_rate": 9.839171511755514e-06, + "loss": 0.5251, + "step": 1621 + }, + { + "epoch": 0.5189846416382252, + "grad_norm": 0.6321656703948975, + "learning_rate": 9.838702770148227e-06, + "loss": 0.5652, + "step": 1622 + }, + { + "epoch": 0.5193046075085325, + "grad_norm": 0.5689617991447449, + "learning_rate": 9.83823335764898e-06, + "loss": 0.5823, + "step": 1623 + }, + { + "epoch": 0.5196245733788396, + "grad_norm": 0.6367678642272949, + "learning_rate": 9.837763274322858e-06, + "loss": 0.5428, + "step": 1624 + }, + { + "epoch": 0.5199445392491467, + "grad_norm": 0.6572225093841553, + "learning_rate": 9.837292520235037e-06, + "loss": 0.5924, + "step": 1625 + }, + { + "epoch": 0.5202645051194539, + "grad_norm": 0.5359474420547485, + "learning_rate": 9.836821095450789e-06, + "loss": 0.5428, + "step": 1626 + }, + { + "epoch": 0.5205844709897611, + "grad_norm": 0.6847219467163086, + "learning_rate": 9.836349000035477e-06, + "loss": 0.5699, + "step": 1627 + }, + { + "epoch": 0.5209044368600683, + "grad_norm": 0.5682468414306641, + "learning_rate": 9.835876234054558e-06, + "loss": 0.5705, + "step": 1628 + }, + { + "epoch": 0.5212244027303754, + "grad_norm": 0.6233001947402954, + "learning_rate": 9.835402797573582e-06, + "loss": 0.602, + "step": 1629 + }, + { + "epoch": 0.5215443686006825, + "grad_norm": 0.6758689880371094, + "learning_rate": 9.834928690658191e-06, + "loss": 0.5808, + "step": 1630 + }, + { + "epoch": 0.5218643344709898, + "grad_norm": 0.6670978665351868, + "learning_rate": 9.83445391337412e-06, + "loss": 0.5779, + "step": 1631 + }, + { + "epoch": 0.5221843003412969, + "grad_norm": 0.5305650234222412, + "learning_rate": 9.833978465787198e-06, + "loss": 0.5356, + "step": 1632 + }, + { + "epoch": 0.5225042662116041, + "grad_norm": 0.7348790764808655, + "learning_rate": 9.833502347963346e-06, + "loss": 0.539, + "step": 1633 + }, + { + "epoch": 0.5228242320819113, + "grad_norm": 0.7539340257644653, + "learning_rate": 9.83302555996858e-06, + "loss": 0.566, + "step": 1634 + }, + { + "epoch": 0.5231441979522184, + "grad_norm": 0.6037124395370483, + "learning_rate": 9.832548101869004e-06, + "loss": 0.5512, + "step": 1635 + }, + { + "epoch": 0.5234641638225256, + "grad_norm": 0.6238608956336975, + "learning_rate": 9.832069973730821e-06, + "loss": 0.5866, + "step": 1636 + }, + { + "epoch": 0.5237841296928327, + "grad_norm": 0.6644582152366638, + "learning_rate": 9.831591175620323e-06, + "loss": 0.5994, + "step": 1637 + }, + { + "epoch": 0.52410409556314, + "grad_norm": 0.6156083345413208, + "learning_rate": 9.831111707603896e-06, + "loss": 0.5531, + "step": 1638 + }, + { + "epoch": 0.5244240614334471, + "grad_norm": 0.6902807950973511, + "learning_rate": 9.830631569748018e-06, + "loss": 0.5558, + "step": 1639 + }, + { + "epoch": 0.5247440273037542, + "grad_norm": 0.6093671321868896, + "learning_rate": 9.83015076211926e-06, + "loss": 0.5589, + "step": 1640 + }, + { + "epoch": 0.5250639931740614, + "grad_norm": 0.6647526025772095, + "learning_rate": 9.829669284784292e-06, + "loss": 0.5418, + "step": 1641 + }, + { + "epoch": 0.5253839590443686, + "grad_norm": 0.6453307271003723, + "learning_rate": 9.829187137809865e-06, + "loss": 0.5448, + "step": 1642 + }, + { + "epoch": 0.5257039249146758, + "grad_norm": 0.5826766490936279, + "learning_rate": 9.828704321262833e-06, + "loss": 0.5196, + "step": 1643 + }, + { + "epoch": 0.5260238907849829, + "grad_norm": 0.6424474716186523, + "learning_rate": 9.828220835210136e-06, + "loss": 0.5954, + "step": 1644 + }, + { + "epoch": 0.5263438566552902, + "grad_norm": 0.5636090040206909, + "learning_rate": 9.827736679718813e-06, + "loss": 0.5072, + "step": 1645 + }, + { + "epoch": 0.5266638225255973, + "grad_norm": 0.5572488307952881, + "learning_rate": 9.827251854855992e-06, + "loss": 0.5685, + "step": 1646 + }, + { + "epoch": 0.5269837883959044, + "grad_norm": 0.6091232895851135, + "learning_rate": 9.826766360688893e-06, + "loss": 0.5586, + "step": 1647 + }, + { + "epoch": 0.5273037542662116, + "grad_norm": 0.5716792941093445, + "learning_rate": 9.82628019728483e-06, + "loss": 0.547, + "step": 1648 + }, + { + "epoch": 0.5276237201365188, + "grad_norm": 0.5463572144508362, + "learning_rate": 9.825793364711214e-06, + "loss": 0.5681, + "step": 1649 + }, + { + "epoch": 0.527943686006826, + "grad_norm": 0.6867672204971313, + "learning_rate": 9.825305863035542e-06, + "loss": 0.6118, + "step": 1650 + }, + { + "epoch": 0.5282636518771331, + "grad_norm": 0.608964204788208, + "learning_rate": 9.824817692325406e-06, + "loss": 0.5524, + "step": 1651 + }, + { + "epoch": 0.5285836177474402, + "grad_norm": 0.6046398282051086, + "learning_rate": 9.824328852648493e-06, + "loss": 0.5688, + "step": 1652 + }, + { + "epoch": 0.5289035836177475, + "grad_norm": 0.5403018593788147, + "learning_rate": 9.823839344072582e-06, + "loss": 0.5252, + "step": 1653 + }, + { + "epoch": 0.5292235494880546, + "grad_norm": 0.5631974935531616, + "learning_rate": 9.82334916666554e-06, + "loss": 0.5623, + "step": 1654 + }, + { + "epoch": 0.5295435153583617, + "grad_norm": 0.6233628392219543, + "learning_rate": 9.822858320495335e-06, + "loss": 0.5735, + "step": 1655 + }, + { + "epoch": 0.5298634812286689, + "grad_norm": 0.5567300915718079, + "learning_rate": 9.822366805630022e-06, + "loss": 0.534, + "step": 1656 + }, + { + "epoch": 0.5301834470989761, + "grad_norm": 0.6214247345924377, + "learning_rate": 9.82187462213775e-06, + "loss": 0.5767, + "step": 1657 + }, + { + "epoch": 0.5305034129692833, + "grad_norm": 0.6399384140968323, + "learning_rate": 9.82138177008676e-06, + "loss": 0.5787, + "step": 1658 + }, + { + "epoch": 0.5308233788395904, + "grad_norm": 0.584986686706543, + "learning_rate": 9.820888249545388e-06, + "loss": 0.5501, + "step": 1659 + }, + { + "epoch": 0.5311433447098977, + "grad_norm": 0.5978248715400696, + "learning_rate": 9.82039406058206e-06, + "loss": 0.5829, + "step": 1660 + }, + { + "epoch": 0.5314633105802048, + "grad_norm": 0.6715003252029419, + "learning_rate": 9.819899203265298e-06, + "loss": 0.5932, + "step": 1661 + }, + { + "epoch": 0.5317832764505119, + "grad_norm": 0.622571587562561, + "learning_rate": 9.81940367766371e-06, + "loss": 0.6027, + "step": 1662 + }, + { + "epoch": 0.5321032423208191, + "grad_norm": 0.646557092666626, + "learning_rate": 9.818907483846007e-06, + "loss": 0.5819, + "step": 1663 + }, + { + "epoch": 0.5324232081911263, + "grad_norm": 0.6905471086502075, + "learning_rate": 9.818410621880983e-06, + "loss": 0.609, + "step": 1664 + }, + { + "epoch": 0.5327431740614335, + "grad_norm": 0.6914578676223755, + "learning_rate": 9.81791309183753e-06, + "loss": 0.5885, + "step": 1665 + }, + { + "epoch": 0.5330631399317406, + "grad_norm": 0.5626682043075562, + "learning_rate": 9.81741489378463e-06, + "loss": 0.5536, + "step": 1666 + }, + { + "epoch": 0.5333831058020477, + "grad_norm": 0.663912296295166, + "learning_rate": 9.816916027791359e-06, + "loss": 0.5744, + "step": 1667 + }, + { + "epoch": 0.533703071672355, + "grad_norm": 0.6111547946929932, + "learning_rate": 9.816416493926887e-06, + "loss": 0.5534, + "step": 1668 + }, + { + "epoch": 0.5340230375426621, + "grad_norm": 0.5859062671661377, + "learning_rate": 9.815916292260474e-06, + "loss": 0.572, + "step": 1669 + }, + { + "epoch": 0.5343430034129693, + "grad_norm": 0.5649756789207458, + "learning_rate": 9.815415422861472e-06, + "loss": 0.5531, + "step": 1670 + }, + { + "epoch": 0.5346629692832765, + "grad_norm": 0.5535812973976135, + "learning_rate": 9.81491388579933e-06, + "loss": 0.5256, + "step": 1671 + }, + { + "epoch": 0.5349829351535836, + "grad_norm": 0.6786409020423889, + "learning_rate": 9.814411681143585e-06, + "loss": 0.5855, + "step": 1672 + }, + { + "epoch": 0.5353029010238908, + "grad_norm": 0.5840369462966919, + "learning_rate": 9.813908808963868e-06, + "loss": 0.5619, + "step": 1673 + }, + { + "epoch": 0.5356228668941979, + "grad_norm": 0.5695275664329529, + "learning_rate": 9.813405269329903e-06, + "loss": 0.5961, + "step": 1674 + }, + { + "epoch": 0.5359428327645052, + "grad_norm": 0.607681155204773, + "learning_rate": 9.812901062311508e-06, + "loss": 0.5416, + "step": 1675 + }, + { + "epoch": 0.5362627986348123, + "grad_norm": 0.5815897583961487, + "learning_rate": 9.81239618797859e-06, + "loss": 0.5771, + "step": 1676 + }, + { + "epoch": 0.5365827645051194, + "grad_norm": 0.544284999370575, + "learning_rate": 9.81189064640115e-06, + "loss": 0.5203, + "step": 1677 + }, + { + "epoch": 0.5369027303754266, + "grad_norm": 0.5429071187973022, + "learning_rate": 9.811384437649285e-06, + "loss": 0.5808, + "step": 1678 + }, + { + "epoch": 0.5372226962457338, + "grad_norm": 0.5531368851661682, + "learning_rate": 9.810877561793178e-06, + "loss": 0.5617, + "step": 1679 + }, + { + "epoch": 0.537542662116041, + "grad_norm": 0.7014932036399841, + "learning_rate": 9.81037001890311e-06, + "loss": 0.5843, + "step": 1680 + }, + { + "epoch": 0.5378626279863481, + "grad_norm": 0.6051618456840515, + "learning_rate": 9.809861809049453e-06, + "loss": 0.56, + "step": 1681 + }, + { + "epoch": 0.5381825938566553, + "grad_norm": 0.5951033234596252, + "learning_rate": 9.809352932302666e-06, + "loss": 0.5714, + "step": 1682 + }, + { + "epoch": 0.5385025597269625, + "grad_norm": 0.5959492325782776, + "learning_rate": 9.808843388733314e-06, + "loss": 0.553, + "step": 1683 + }, + { + "epoch": 0.5388225255972696, + "grad_norm": 0.5903566479682922, + "learning_rate": 9.808333178412038e-06, + "loss": 0.5638, + "step": 1684 + }, + { + "epoch": 0.5391424914675768, + "grad_norm": 0.5624989867210388, + "learning_rate": 9.807822301409583e-06, + "loss": 0.5699, + "step": 1685 + }, + { + "epoch": 0.539462457337884, + "grad_norm": 0.6225293278694153, + "learning_rate": 9.807310757796782e-06, + "loss": 0.5659, + "step": 1686 + }, + { + "epoch": 0.5397824232081911, + "grad_norm": 0.6644200682640076, + "learning_rate": 9.80679854764456e-06, + "loss": 0.5929, + "step": 1687 + }, + { + "epoch": 0.5401023890784983, + "grad_norm": 0.6093822121620178, + "learning_rate": 9.806285671023938e-06, + "loss": 0.5704, + "step": 1688 + }, + { + "epoch": 0.5404223549488054, + "grad_norm": 0.6620171666145325, + "learning_rate": 9.805772128006027e-06, + "loss": 0.5991, + "step": 1689 + }, + { + "epoch": 0.5407423208191127, + "grad_norm": 0.6101434230804443, + "learning_rate": 9.805257918662026e-06, + "loss": 0.5798, + "step": 1690 + }, + { + "epoch": 0.5410622866894198, + "grad_norm": 0.5370322465896606, + "learning_rate": 9.804743043063235e-06, + "loss": 0.5582, + "step": 1691 + }, + { + "epoch": 0.5413822525597269, + "grad_norm": 0.7034540176391602, + "learning_rate": 9.804227501281041e-06, + "loss": 0.6284, + "step": 1692 + }, + { + "epoch": 0.5417022184300341, + "grad_norm": 0.6034532785415649, + "learning_rate": 9.803711293386925e-06, + "loss": 0.5597, + "step": 1693 + }, + { + "epoch": 0.5420221843003413, + "grad_norm": 0.6423254013061523, + "learning_rate": 9.803194419452457e-06, + "loss": 0.5648, + "step": 1694 + }, + { + "epoch": 0.5423421501706485, + "grad_norm": 0.6289602518081665, + "learning_rate": 9.802676879549307e-06, + "loss": 0.5663, + "step": 1695 + }, + { + "epoch": 0.5426621160409556, + "grad_norm": 0.5871692895889282, + "learning_rate": 9.802158673749226e-06, + "loss": 0.5469, + "step": 1696 + }, + { + "epoch": 0.5429820819112628, + "grad_norm": 0.6468052864074707, + "learning_rate": 9.801639802124073e-06, + "loss": 0.5633, + "step": 1697 + }, + { + "epoch": 0.54330204778157, + "grad_norm": 0.6076865196228027, + "learning_rate": 9.80112026474578e-06, + "loss": 0.5436, + "step": 1698 + }, + { + "epoch": 0.5436220136518771, + "grad_norm": 0.6306731700897217, + "learning_rate": 9.80060006168639e-06, + "loss": 0.5826, + "step": 1699 + }, + { + "epoch": 0.5439419795221843, + "grad_norm": 0.5897573828697205, + "learning_rate": 9.800079193018025e-06, + "loss": 0.5587, + "step": 1700 + }, + { + "epoch": 0.5442619453924915, + "grad_norm": 0.6413190364837646, + "learning_rate": 9.799557658812903e-06, + "loss": 0.5394, + "step": 1701 + }, + { + "epoch": 0.5445819112627986, + "grad_norm": 0.5940037965774536, + "learning_rate": 9.79903545914334e-06, + "loss": 0.5601, + "step": 1702 + }, + { + "epoch": 0.5449018771331058, + "grad_norm": 0.6278110146522522, + "learning_rate": 9.798512594081737e-06, + "loss": 0.5293, + "step": 1703 + }, + { + "epoch": 0.5452218430034129, + "grad_norm": 0.7648538947105408, + "learning_rate": 9.797989063700591e-06, + "loss": 0.5833, + "step": 1704 + }, + { + "epoch": 0.5455418088737202, + "grad_norm": 0.5689882040023804, + "learning_rate": 9.797464868072489e-06, + "loss": 0.5842, + "step": 1705 + }, + { + "epoch": 0.5458617747440273, + "grad_norm": 0.5978536605834961, + "learning_rate": 9.79694000727011e-06, + "loss": 0.609, + "step": 1706 + }, + { + "epoch": 0.5461817406143344, + "grad_norm": 0.7378367185592651, + "learning_rate": 9.796414481366227e-06, + "loss": 0.5615, + "step": 1707 + }, + { + "epoch": 0.5465017064846417, + "grad_norm": 0.7020463347434998, + "learning_rate": 9.795888290433709e-06, + "loss": 0.6027, + "step": 1708 + }, + { + "epoch": 0.5468216723549488, + "grad_norm": 0.5986933708190918, + "learning_rate": 9.795361434545508e-06, + "loss": 0.5398, + "step": 1709 + }, + { + "epoch": 0.547141638225256, + "grad_norm": 0.7171380519866943, + "learning_rate": 9.794833913774678e-06, + "loss": 0.5354, + "step": 1710 + }, + { + "epoch": 0.5474616040955631, + "grad_norm": 0.647392749786377, + "learning_rate": 9.794305728194353e-06, + "loss": 0.5626, + "step": 1711 + }, + { + "epoch": 0.5477815699658704, + "grad_norm": 0.670254647731781, + "learning_rate": 9.793776877877774e-06, + "loss": 0.578, + "step": 1712 + }, + { + "epoch": 0.5481015358361775, + "grad_norm": 0.6468953490257263, + "learning_rate": 9.793247362898264e-06, + "loss": 0.5564, + "step": 1713 + }, + { + "epoch": 0.5484215017064846, + "grad_norm": 0.6417179107666016, + "learning_rate": 9.79271718332924e-06, + "loss": 0.5667, + "step": 1714 + }, + { + "epoch": 0.5487414675767918, + "grad_norm": 0.603004515171051, + "learning_rate": 9.792186339244211e-06, + "loss": 0.5482, + "step": 1715 + }, + { + "epoch": 0.549061433447099, + "grad_norm": 0.6157732009887695, + "learning_rate": 9.791654830716784e-06, + "loss": 0.5758, + "step": 1716 + }, + { + "epoch": 0.5493813993174061, + "grad_norm": 0.6211690902709961, + "learning_rate": 9.79112265782065e-06, + "loss": 0.5542, + "step": 1717 + }, + { + "epoch": 0.5497013651877133, + "grad_norm": 0.6685905456542969, + "learning_rate": 9.790589820629594e-06, + "loss": 0.5799, + "step": 1718 + }, + { + "epoch": 0.5500213310580204, + "grad_norm": 0.6300346255302429, + "learning_rate": 9.790056319217496e-06, + "loss": 0.6207, + "step": 1719 + }, + { + "epoch": 0.5503412969283277, + "grad_norm": 0.6636313796043396, + "learning_rate": 9.789522153658328e-06, + "loss": 0.6008, + "step": 1720 + }, + { + "epoch": 0.5506612627986348, + "grad_norm": 0.6509947180747986, + "learning_rate": 9.78898732402615e-06, + "loss": 0.5185, + "step": 1721 + }, + { + "epoch": 0.550981228668942, + "grad_norm": 0.553341269493103, + "learning_rate": 9.78845183039512e-06, + "loss": 0.5621, + "step": 1722 + }, + { + "epoch": 0.5513011945392492, + "grad_norm": 0.5741779804229736, + "learning_rate": 9.78791567283948e-06, + "loss": 0.5271, + "step": 1723 + }, + { + "epoch": 0.5516211604095563, + "grad_norm": 0.5845010876655579, + "learning_rate": 9.787378851433575e-06, + "loss": 0.5661, + "step": 1724 + }, + { + "epoch": 0.5519411262798635, + "grad_norm": 0.5319338440895081, + "learning_rate": 9.786841366251832e-06, + "loss": 0.5589, + "step": 1725 + }, + { + "epoch": 0.5522610921501706, + "grad_norm": 0.6268966197967529, + "learning_rate": 9.786303217368776e-06, + "loss": 0.5764, + "step": 1726 + }, + { + "epoch": 0.5525810580204779, + "grad_norm": 0.6318864226341248, + "learning_rate": 9.78576440485902e-06, + "loss": 0.6013, + "step": 1727 + }, + { + "epoch": 0.552901023890785, + "grad_norm": 0.5269508957862854, + "learning_rate": 9.78522492879727e-06, + "loss": 0.5649, + "step": 1728 + }, + { + "epoch": 0.5532209897610921, + "grad_norm": 0.5737921595573425, + "learning_rate": 9.78468478925833e-06, + "loss": 0.5322, + "step": 1729 + }, + { + "epoch": 0.5535409556313993, + "grad_norm": 0.6844146847724915, + "learning_rate": 9.784143986317084e-06, + "loss": 0.5964, + "step": 1730 + }, + { + "epoch": 0.5538609215017065, + "grad_norm": 0.5287778377532959, + "learning_rate": 9.783602520048524e-06, + "loss": 0.5648, + "step": 1731 + }, + { + "epoch": 0.5541808873720137, + "grad_norm": 0.599004328250885, + "learning_rate": 9.783060390527715e-06, + "loss": 0.5581, + "step": 1732 + }, + { + "epoch": 0.5545008532423208, + "grad_norm": 0.6111088395118713, + "learning_rate": 9.78251759782983e-06, + "loss": 0.5774, + "step": 1733 + }, + { + "epoch": 0.554820819112628, + "grad_norm": 0.6105920672416687, + "learning_rate": 9.781974142030128e-06, + "loss": 0.5592, + "step": 1734 + }, + { + "epoch": 0.5551407849829352, + "grad_norm": 0.583831250667572, + "learning_rate": 9.781430023203957e-06, + "loss": 0.5793, + "step": 1735 + }, + { + "epoch": 0.5554607508532423, + "grad_norm": 0.5499464273452759, + "learning_rate": 9.780885241426762e-06, + "loss": 0.5533, + "step": 1736 + }, + { + "epoch": 0.5557807167235495, + "grad_norm": 0.6046625971794128, + "learning_rate": 9.780339796774076e-06, + "loss": 0.5736, + "step": 1737 + }, + { + "epoch": 0.5561006825938567, + "grad_norm": 0.5838026404380798, + "learning_rate": 9.779793689321528e-06, + "loss": 0.561, + "step": 1738 + }, + { + "epoch": 0.5564206484641638, + "grad_norm": 0.5161740183830261, + "learning_rate": 9.779246919144833e-06, + "loss": 0.5488, + "step": 1739 + }, + { + "epoch": 0.556740614334471, + "grad_norm": 0.5640830993652344, + "learning_rate": 9.778699486319803e-06, + "loss": 0.5456, + "step": 1740 + }, + { + "epoch": 0.5570605802047781, + "grad_norm": 0.6633821725845337, + "learning_rate": 9.77815139092234e-06, + "loss": 0.5319, + "step": 1741 + }, + { + "epoch": 0.5573805460750854, + "grad_norm": 0.5820072889328003, + "learning_rate": 9.77760263302844e-06, + "loss": 0.5577, + "step": 1742 + }, + { + "epoch": 0.5577005119453925, + "grad_norm": 0.6018913388252258, + "learning_rate": 9.777053212714188e-06, + "loss": 0.5492, + "step": 1743 + }, + { + "epoch": 0.5580204778156996, + "grad_norm": 0.6947749853134155, + "learning_rate": 9.776503130055758e-06, + "loss": 0.5873, + "step": 1744 + }, + { + "epoch": 0.5583404436860068, + "grad_norm": 0.6588101387023926, + "learning_rate": 9.775952385129423e-06, + "loss": 0.5815, + "step": 1745 + }, + { + "epoch": 0.558660409556314, + "grad_norm": 0.5936509370803833, + "learning_rate": 9.775400978011547e-06, + "loss": 0.6015, + "step": 1746 + }, + { + "epoch": 0.5589803754266212, + "grad_norm": 0.6646673679351807, + "learning_rate": 9.774848908778577e-06, + "loss": 0.5585, + "step": 1747 + }, + { + "epoch": 0.5593003412969283, + "grad_norm": 0.5602323412895203, + "learning_rate": 9.774296177507065e-06, + "loss": 0.5442, + "step": 1748 + }, + { + "epoch": 0.5596203071672355, + "grad_norm": 0.5378860235214233, + "learning_rate": 9.773742784273643e-06, + "loss": 0.5255, + "step": 1749 + }, + { + "epoch": 0.5599402730375427, + "grad_norm": 0.6054778099060059, + "learning_rate": 9.773188729155042e-06, + "loss": 0.5897, + "step": 1750 + }, + { + "epoch": 0.5602602389078498, + "grad_norm": 0.5988175272941589, + "learning_rate": 9.77263401222808e-06, + "loss": 0.6037, + "step": 1751 + }, + { + "epoch": 0.560580204778157, + "grad_norm": 0.6248955726623535, + "learning_rate": 9.772078633569671e-06, + "loss": 0.5725, + "step": 1752 + }, + { + "epoch": 0.5609001706484642, + "grad_norm": 0.6076432466506958, + "learning_rate": 9.771522593256817e-06, + "loss": 0.563, + "step": 1753 + }, + { + "epoch": 0.5612201365187713, + "grad_norm": 0.612916111946106, + "learning_rate": 9.770965891366617e-06, + "loss": 0.596, + "step": 1754 + }, + { + "epoch": 0.5615401023890785, + "grad_norm": 0.598166286945343, + "learning_rate": 9.770408527976257e-06, + "loss": 0.5782, + "step": 1755 + }, + { + "epoch": 0.5618600682593856, + "grad_norm": 0.5937659740447998, + "learning_rate": 9.769850503163015e-06, + "loss": 0.5321, + "step": 1756 + }, + { + "epoch": 0.5621800341296929, + "grad_norm": 0.5992322564125061, + "learning_rate": 9.76929181700426e-06, + "loss": 0.5703, + "step": 1757 + }, + { + "epoch": 0.5625, + "grad_norm": 0.6490613222122192, + "learning_rate": 9.76873246957746e-06, + "loss": 0.5482, + "step": 1758 + }, + { + "epoch": 0.5628199658703071, + "grad_norm": 0.642821192741394, + "learning_rate": 9.768172460960166e-06, + "loss": 0.5629, + "step": 1759 + }, + { + "epoch": 0.5631399317406144, + "grad_norm": 0.6676160097122192, + "learning_rate": 9.767611791230023e-06, + "loss": 0.5194, + "step": 1760 + }, + { + "epoch": 0.5634598976109215, + "grad_norm": 0.6234229207038879, + "learning_rate": 9.767050460464769e-06, + "loss": 0.5573, + "step": 1761 + }, + { + "epoch": 0.5637798634812287, + "grad_norm": 0.6466017365455627, + "learning_rate": 9.766488468742235e-06, + "loss": 0.5885, + "step": 1762 + }, + { + "epoch": 0.5640998293515358, + "grad_norm": 0.559739887714386, + "learning_rate": 9.76592581614034e-06, + "loss": 0.5735, + "step": 1763 + }, + { + "epoch": 0.564419795221843, + "grad_norm": 0.6247774958610535, + "learning_rate": 9.765362502737098e-06, + "loss": 0.5771, + "step": 1764 + }, + { + "epoch": 0.5647397610921502, + "grad_norm": 0.5745558738708496, + "learning_rate": 9.764798528610612e-06, + "loss": 0.5715, + "step": 1765 + }, + { + "epoch": 0.5650597269624573, + "grad_norm": 0.5948444604873657, + "learning_rate": 9.764233893839077e-06, + "loss": 0.5802, + "step": 1766 + }, + { + "epoch": 0.5653796928327645, + "grad_norm": 0.5603572726249695, + "learning_rate": 9.763668598500782e-06, + "loss": 0.5397, + "step": 1767 + }, + { + "epoch": 0.5656996587030717, + "grad_norm": 0.5986836552619934, + "learning_rate": 9.763102642674104e-06, + "loss": 0.5411, + "step": 1768 + }, + { + "epoch": 0.5660196245733788, + "grad_norm": 0.6480838060379028, + "learning_rate": 9.762536026437515e-06, + "loss": 0.5741, + "step": 1769 + }, + { + "epoch": 0.566339590443686, + "grad_norm": 0.6260193586349487, + "learning_rate": 9.761968749869576e-06, + "loss": 0.5964, + "step": 1770 + }, + { + "epoch": 0.5666595563139932, + "grad_norm": 0.6947192549705505, + "learning_rate": 9.76140081304894e-06, + "loss": 0.5766, + "step": 1771 + }, + { + "epoch": 0.5669795221843004, + "grad_norm": 0.5749015808105469, + "learning_rate": 9.760832216054356e-06, + "loss": 0.5477, + "step": 1772 + }, + { + "epoch": 0.5672994880546075, + "grad_norm": 0.6033229231834412, + "learning_rate": 9.760262958964656e-06, + "loss": 0.5639, + "step": 1773 + }, + { + "epoch": 0.5676194539249146, + "grad_norm": 0.6450707912445068, + "learning_rate": 9.75969304185877e-06, + "loss": 0.5523, + "step": 1774 + }, + { + "epoch": 0.5679394197952219, + "grad_norm": 0.6025317907333374, + "learning_rate": 9.759122464815718e-06, + "loss": 0.5907, + "step": 1775 + }, + { + "epoch": 0.568259385665529, + "grad_norm": 0.7295170426368713, + "learning_rate": 9.75855122791461e-06, + "loss": 0.5803, + "step": 1776 + }, + { + "epoch": 0.5685793515358362, + "grad_norm": 0.613014280796051, + "learning_rate": 9.757979331234651e-06, + "loss": 0.5558, + "step": 1777 + }, + { + "epoch": 0.5688993174061433, + "grad_norm": 0.5929551720619202, + "learning_rate": 9.757406774855134e-06, + "loss": 0.5863, + "step": 1778 + }, + { + "epoch": 0.5692192832764505, + "grad_norm": 0.6426234841346741, + "learning_rate": 9.756833558855443e-06, + "loss": 0.5739, + "step": 1779 + }, + { + "epoch": 0.5695392491467577, + "grad_norm": 0.719906747341156, + "learning_rate": 9.756259683315056e-06, + "loss": 0.5883, + "step": 1780 + }, + { + "epoch": 0.5698592150170648, + "grad_norm": 0.6483733057975769, + "learning_rate": 9.755685148313544e-06, + "loss": 0.6078, + "step": 1781 + }, + { + "epoch": 0.570179180887372, + "grad_norm": 0.6298717260360718, + "learning_rate": 9.755109953930564e-06, + "loss": 0.5331, + "step": 1782 + }, + { + "epoch": 0.5704991467576792, + "grad_norm": 0.7005184888839722, + "learning_rate": 9.754534100245867e-06, + "loss": 0.5949, + "step": 1783 + }, + { + "epoch": 0.5708191126279863, + "grad_norm": 0.5551364421844482, + "learning_rate": 9.7539575873393e-06, + "loss": 0.5433, + "step": 1784 + }, + { + "epoch": 0.5711390784982935, + "grad_norm": 0.6557787656784058, + "learning_rate": 9.753380415290793e-06, + "loss": 0.5892, + "step": 1785 + }, + { + "epoch": 0.5714590443686007, + "grad_norm": 0.718512237071991, + "learning_rate": 9.752802584180372e-06, + "loss": 0.6151, + "step": 1786 + }, + { + "epoch": 0.5717790102389079, + "grad_norm": 0.5826222896575928, + "learning_rate": 9.752224094088156e-06, + "loss": 0.5744, + "step": 1787 + }, + { + "epoch": 0.572098976109215, + "grad_norm": 0.695049524307251, + "learning_rate": 9.751644945094352e-06, + "loss": 0.5994, + "step": 1788 + }, + { + "epoch": 0.5724189419795221, + "grad_norm": 0.6085717082023621, + "learning_rate": 9.751065137279261e-06, + "loss": 0.5471, + "step": 1789 + }, + { + "epoch": 0.5727389078498294, + "grad_norm": 0.5604434013366699, + "learning_rate": 9.750484670723272e-06, + "loss": 0.5563, + "step": 1790 + }, + { + "epoch": 0.5730588737201365, + "grad_norm": 0.6504515409469604, + "learning_rate": 9.749903545506868e-06, + "loss": 0.5488, + "step": 1791 + }, + { + "epoch": 0.5733788395904437, + "grad_norm": 0.7121344208717346, + "learning_rate": 9.749321761710625e-06, + "loss": 0.5697, + "step": 1792 + }, + { + "epoch": 0.5736988054607508, + "grad_norm": 0.6262866258621216, + "learning_rate": 9.748739319415203e-06, + "loss": 0.5913, + "step": 1793 + }, + { + "epoch": 0.574018771331058, + "grad_norm": 0.6223553419113159, + "learning_rate": 9.748156218701363e-06, + "loss": 0.5731, + "step": 1794 + }, + { + "epoch": 0.5743387372013652, + "grad_norm": 0.693714439868927, + "learning_rate": 9.747572459649952e-06, + "loss": 0.5876, + "step": 1795 + }, + { + "epoch": 0.5746587030716723, + "grad_norm": 0.5756598114967346, + "learning_rate": 9.746988042341907e-06, + "loss": 0.5376, + "step": 1796 + }, + { + "epoch": 0.5749786689419796, + "grad_norm": 0.6898940801620483, + "learning_rate": 9.746402966858259e-06, + "loss": 0.5538, + "step": 1797 + }, + { + "epoch": 0.5752986348122867, + "grad_norm": 0.62108314037323, + "learning_rate": 9.74581723328013e-06, + "loss": 0.5493, + "step": 1798 + }, + { + "epoch": 0.5756186006825939, + "grad_norm": 0.6687491536140442, + "learning_rate": 9.745230841688733e-06, + "loss": 0.5688, + "step": 1799 + }, + { + "epoch": 0.575938566552901, + "grad_norm": 0.6858484148979187, + "learning_rate": 9.74464379216537e-06, + "loss": 0.5609, + "step": 1800 + }, + { + "epoch": 0.5762585324232082, + "grad_norm": 0.6215513944625854, + "learning_rate": 9.744056084791439e-06, + "loss": 0.5785, + "step": 1801 + }, + { + "epoch": 0.5765784982935154, + "grad_norm": 0.6347957849502563, + "learning_rate": 9.743467719648423e-06, + "loss": 0.5399, + "step": 1802 + }, + { + "epoch": 0.5768984641638225, + "grad_norm": 0.6881147623062134, + "learning_rate": 9.742878696817901e-06, + "loss": 0.5333, + "step": 1803 + }, + { + "epoch": 0.5772184300341296, + "grad_norm": 0.7048043608665466, + "learning_rate": 9.742289016381542e-06, + "loss": 0.5787, + "step": 1804 + }, + { + "epoch": 0.5775383959044369, + "grad_norm": 0.6560038924217224, + "learning_rate": 9.741698678421108e-06, + "loss": 0.5611, + "step": 1805 + }, + { + "epoch": 0.577858361774744, + "grad_norm": 0.6992453932762146, + "learning_rate": 9.741107683018444e-06, + "loss": 0.5972, + "step": 1806 + }, + { + "epoch": 0.5781783276450512, + "grad_norm": 0.6837847232818604, + "learning_rate": 9.7405160302555e-06, + "loss": 0.563, + "step": 1807 + }, + { + "epoch": 0.5784982935153583, + "grad_norm": 0.8289478421211243, + "learning_rate": 9.739923720214305e-06, + "loss": 0.5804, + "step": 1808 + }, + { + "epoch": 0.5788182593856656, + "grad_norm": 0.666740357875824, + "learning_rate": 9.739330752976981e-06, + "loss": 0.5798, + "step": 1809 + }, + { + "epoch": 0.5791382252559727, + "grad_norm": 0.6751397252082825, + "learning_rate": 9.738737128625748e-06, + "loss": 0.5375, + "step": 1810 + }, + { + "epoch": 0.5794581911262798, + "grad_norm": 0.6747352480888367, + "learning_rate": 9.73814284724291e-06, + "loss": 0.5687, + "step": 1811 + }, + { + "epoch": 0.5797781569965871, + "grad_norm": 0.6946969628334045, + "learning_rate": 9.737547908910867e-06, + "loss": 0.5897, + "step": 1812 + }, + { + "epoch": 0.5800981228668942, + "grad_norm": 0.6373754739761353, + "learning_rate": 9.736952313712108e-06, + "loss": 0.5496, + "step": 1813 + }, + { + "epoch": 0.5804180887372014, + "grad_norm": 0.6169047355651855, + "learning_rate": 9.736356061729212e-06, + "loss": 0.575, + "step": 1814 + }, + { + "epoch": 0.5807380546075085, + "grad_norm": 0.6824852228164673, + "learning_rate": 9.735759153044849e-06, + "loss": 0.5506, + "step": 1815 + }, + { + "epoch": 0.5810580204778157, + "grad_norm": 0.5575690865516663, + "learning_rate": 9.735161587741782e-06, + "loss": 0.5717, + "step": 1816 + }, + { + "epoch": 0.5813779863481229, + "grad_norm": 0.5755314230918884, + "learning_rate": 9.734563365902865e-06, + "loss": 0.5444, + "step": 1817 + }, + { + "epoch": 0.58169795221843, + "grad_norm": 0.6323144435882568, + "learning_rate": 9.733964487611044e-06, + "loss": 0.5629, + "step": 1818 + }, + { + "epoch": 0.5820179180887372, + "grad_norm": 0.5643534064292908, + "learning_rate": 9.733364952949348e-06, + "loss": 0.5611, + "step": 1819 + }, + { + "epoch": 0.5823378839590444, + "grad_norm": 0.6021159887313843, + "learning_rate": 9.732764762000908e-06, + "loss": 0.5562, + "step": 1820 + }, + { + "epoch": 0.5826578498293515, + "grad_norm": 0.5586191415786743, + "learning_rate": 9.73216391484894e-06, + "loss": 0.5187, + "step": 1821 + }, + { + "epoch": 0.5829778156996587, + "grad_norm": 0.6042790412902832, + "learning_rate": 9.731562411576751e-06, + "loss": 0.5652, + "step": 1822 + }, + { + "epoch": 0.5832977815699659, + "grad_norm": 0.5898236036300659, + "learning_rate": 9.730960252267744e-06, + "loss": 0.5713, + "step": 1823 + }, + { + "epoch": 0.5836177474402731, + "grad_norm": 0.5529911518096924, + "learning_rate": 9.730357437005404e-06, + "loss": 0.5775, + "step": 1824 + }, + { + "epoch": 0.5839377133105802, + "grad_norm": 0.5990542769432068, + "learning_rate": 9.729753965873316e-06, + "loss": 0.6122, + "step": 1825 + }, + { + "epoch": 0.5842576791808873, + "grad_norm": 0.6193751692771912, + "learning_rate": 9.729149838955152e-06, + "loss": 0.5542, + "step": 1826 + }, + { + "epoch": 0.5845776450511946, + "grad_norm": 0.5816980004310608, + "learning_rate": 9.728545056334672e-06, + "loss": 0.5646, + "step": 1827 + }, + { + "epoch": 0.5848976109215017, + "grad_norm": 0.6045491099357605, + "learning_rate": 9.72793961809573e-06, + "loss": 0.5797, + "step": 1828 + }, + { + "epoch": 0.5852175767918089, + "grad_norm": 0.6036240458488464, + "learning_rate": 9.727333524322274e-06, + "loss": 0.5432, + "step": 1829 + }, + { + "epoch": 0.585537542662116, + "grad_norm": 0.516349732875824, + "learning_rate": 9.726726775098338e-06, + "loss": 0.5456, + "step": 1830 + }, + { + "epoch": 0.5858575085324232, + "grad_norm": 0.655970573425293, + "learning_rate": 9.726119370508048e-06, + "loss": 0.5891, + "step": 1831 + }, + { + "epoch": 0.5861774744027304, + "grad_norm": 0.5936410427093506, + "learning_rate": 9.72551131063562e-06, + "loss": 0.5418, + "step": 1832 + }, + { + "epoch": 0.5864974402730375, + "grad_norm": 0.573919951915741, + "learning_rate": 9.724902595565367e-06, + "loss": 0.5551, + "step": 1833 + }, + { + "epoch": 0.5868174061433447, + "grad_norm": 0.593414843082428, + "learning_rate": 9.724293225381681e-06, + "loss": 0.5491, + "step": 1834 + }, + { + "epoch": 0.5871373720136519, + "grad_norm": 0.6796587109565735, + "learning_rate": 9.723683200169059e-06, + "loss": 0.564, + "step": 1835 + }, + { + "epoch": 0.587457337883959, + "grad_norm": 0.7001802921295166, + "learning_rate": 9.723072520012077e-06, + "loss": 0.5585, + "step": 1836 + }, + { + "epoch": 0.5877773037542662, + "grad_norm": 0.5853090882301331, + "learning_rate": 9.722461184995408e-06, + "loss": 0.5905, + "step": 1837 + }, + { + "epoch": 0.5880972696245734, + "grad_norm": 0.6028174161911011, + "learning_rate": 9.721849195203814e-06, + "loss": 0.5431, + "step": 1838 + }, + { + "epoch": 0.5884172354948806, + "grad_norm": 0.6703921556472778, + "learning_rate": 9.721236550722152e-06, + "loss": 0.5828, + "step": 1839 + }, + { + "epoch": 0.5887372013651877, + "grad_norm": 0.5544477105140686, + "learning_rate": 9.720623251635358e-06, + "loss": 0.5307, + "step": 1840 + }, + { + "epoch": 0.5890571672354948, + "grad_norm": 0.5607783794403076, + "learning_rate": 9.720009298028473e-06, + "loss": 0.5603, + "step": 1841 + }, + { + "epoch": 0.5893771331058021, + "grad_norm": 0.5793670415878296, + "learning_rate": 9.71939468998662e-06, + "loss": 0.5483, + "step": 1842 + }, + { + "epoch": 0.5896970989761092, + "grad_norm": 0.6787139773368835, + "learning_rate": 9.718779427595017e-06, + "loss": 0.5509, + "step": 1843 + }, + { + "epoch": 0.5900170648464164, + "grad_norm": 0.613737165927887, + "learning_rate": 9.71816351093897e-06, + "loss": 0.5673, + "step": 1844 + }, + { + "epoch": 0.5903370307167235, + "grad_norm": 0.5790629982948303, + "learning_rate": 9.717546940103874e-06, + "loss": 0.5594, + "step": 1845 + }, + { + "epoch": 0.5906569965870307, + "grad_norm": 0.5192556381225586, + "learning_rate": 9.71692971517522e-06, + "loss": 0.5669, + "step": 1846 + }, + { + "epoch": 0.5909769624573379, + "grad_norm": 0.5302321314811707, + "learning_rate": 9.716311836238585e-06, + "loss": 0.5579, + "step": 1847 + }, + { + "epoch": 0.591296928327645, + "grad_norm": 0.5794715285301208, + "learning_rate": 9.715693303379643e-06, + "loss": 0.5409, + "step": 1848 + }, + { + "epoch": 0.5916168941979523, + "grad_norm": 0.6204904913902283, + "learning_rate": 9.71507411668415e-06, + "loss": 0.6097, + "step": 1849 + }, + { + "epoch": 0.5919368600682594, + "grad_norm": 0.6074978709220886, + "learning_rate": 9.714454276237958e-06, + "loss": 0.5656, + "step": 1850 + }, + { + "epoch": 0.5922568259385665, + "grad_norm": 0.6031474471092224, + "learning_rate": 9.713833782127008e-06, + "loss": 0.5845, + "step": 1851 + }, + { + "epoch": 0.5925767918088737, + "grad_norm": 0.5594843626022339, + "learning_rate": 9.713212634437337e-06, + "loss": 0.5373, + "step": 1852 + }, + { + "epoch": 0.5928967576791809, + "grad_norm": 0.5694454908370972, + "learning_rate": 9.71259083325506e-06, + "loss": 0.5959, + "step": 1853 + }, + { + "epoch": 0.5932167235494881, + "grad_norm": 0.6202699542045593, + "learning_rate": 9.711968378666396e-06, + "loss": 0.5807, + "step": 1854 + }, + { + "epoch": 0.5935366894197952, + "grad_norm": 0.5497620701789856, + "learning_rate": 9.711345270757649e-06, + "loss": 0.597, + "step": 1855 + }, + { + "epoch": 0.5938566552901023, + "grad_norm": 0.597712516784668, + "learning_rate": 9.71072150961521e-06, + "loss": 0.5641, + "step": 1856 + }, + { + "epoch": 0.5941766211604096, + "grad_norm": 0.5572855472564697, + "learning_rate": 9.71009709532557e-06, + "loss": 0.5706, + "step": 1857 + }, + { + "epoch": 0.5944965870307167, + "grad_norm": 0.614568829536438, + "learning_rate": 9.709472027975299e-06, + "loss": 0.5732, + "step": 1858 + }, + { + "epoch": 0.5948165529010239, + "grad_norm": 0.5438926815986633, + "learning_rate": 9.708846307651066e-06, + "loss": 0.5323, + "step": 1859 + }, + { + "epoch": 0.5951365187713311, + "grad_norm": 0.5313812494277954, + "learning_rate": 9.708219934439627e-06, + "loss": 0.5536, + "step": 1860 + }, + { + "epoch": 0.5954564846416383, + "grad_norm": 0.61684250831604, + "learning_rate": 9.70759290842783e-06, + "loss": 0.5556, + "step": 1861 + }, + { + "epoch": 0.5957764505119454, + "grad_norm": 0.5726536512374878, + "learning_rate": 9.706965229702614e-06, + "loss": 0.5442, + "step": 1862 + }, + { + "epoch": 0.5960964163822525, + "grad_norm": 0.6879222393035889, + "learning_rate": 9.706336898351005e-06, + "loss": 0.5752, + "step": 1863 + }, + { + "epoch": 0.5964163822525598, + "grad_norm": 0.5403534770011902, + "learning_rate": 9.705707914460122e-06, + "loss": 0.5306, + "step": 1864 + }, + { + "epoch": 0.5967363481228669, + "grad_norm": 0.5859735608100891, + "learning_rate": 9.70507827811718e-06, + "loss": 0.5654, + "step": 1865 + }, + { + "epoch": 0.597056313993174, + "grad_norm": 0.6655783653259277, + "learning_rate": 9.704447989409472e-06, + "loss": 0.5517, + "step": 1866 + }, + { + "epoch": 0.5973762798634812, + "grad_norm": 0.7582557201385498, + "learning_rate": 9.703817048424389e-06, + "loss": 0.6227, + "step": 1867 + }, + { + "epoch": 0.5976962457337884, + "grad_norm": 0.5447322726249695, + "learning_rate": 9.703185455249416e-06, + "loss": 0.5584, + "step": 1868 + }, + { + "epoch": 0.5980162116040956, + "grad_norm": 0.6286244988441467, + "learning_rate": 9.70255320997212e-06, + "loss": 0.5837, + "step": 1869 + }, + { + "epoch": 0.5983361774744027, + "grad_norm": 0.6168583631515503, + "learning_rate": 9.701920312680163e-06, + "loss": 0.5357, + "step": 1870 + }, + { + "epoch": 0.5986561433447098, + "grad_norm": 0.5949482321739197, + "learning_rate": 9.7012867634613e-06, + "loss": 0.5357, + "step": 1871 + }, + { + "epoch": 0.5989761092150171, + "grad_norm": 0.5717710256576538, + "learning_rate": 9.70065256240337e-06, + "loss": 0.5891, + "step": 1872 + }, + { + "epoch": 0.5992960750853242, + "grad_norm": 0.6137530207633972, + "learning_rate": 9.70001770959431e-06, + "loss": 0.5506, + "step": 1873 + }, + { + "epoch": 0.5996160409556314, + "grad_norm": 0.623816192150116, + "learning_rate": 9.699382205122138e-06, + "loss": 0.569, + "step": 1874 + }, + { + "epoch": 0.5999360068259386, + "grad_norm": 0.5570641160011292, + "learning_rate": 9.69874604907497e-06, + "loss": 0.5678, + "step": 1875 + }, + { + "epoch": 0.6002559726962458, + "grad_norm": 0.5365951657295227, + "learning_rate": 9.698109241541009e-06, + "loss": 0.5651, + "step": 1876 + }, + { + "epoch": 0.6005759385665529, + "grad_norm": 0.5629282593727112, + "learning_rate": 9.697471782608552e-06, + "loss": 0.5368, + "step": 1877 + }, + { + "epoch": 0.60089590443686, + "grad_norm": 0.5601006150245667, + "learning_rate": 9.69683367236598e-06, + "loss": 0.5681, + "step": 1878 + }, + { + "epoch": 0.6012158703071673, + "grad_norm": 0.5880118012428284, + "learning_rate": 9.696194910901766e-06, + "loss": 0.5722, + "step": 1879 + }, + { + "epoch": 0.6015358361774744, + "grad_norm": 0.5651369690895081, + "learning_rate": 9.695555498304481e-06, + "loss": 0.5514, + "step": 1880 + }, + { + "epoch": 0.6018558020477816, + "grad_norm": 0.614946722984314, + "learning_rate": 9.694915434662777e-06, + "loss": 0.506, + "step": 1881 + }, + { + "epoch": 0.6021757679180887, + "grad_norm": 0.5546241402626038, + "learning_rate": 9.6942747200654e-06, + "loss": 0.5185, + "step": 1882 + }, + { + "epoch": 0.6024957337883959, + "grad_norm": 0.6413306593894958, + "learning_rate": 9.693633354601187e-06, + "loss": 0.6012, + "step": 1883 + }, + { + "epoch": 0.6028156996587031, + "grad_norm": 0.5133163332939148, + "learning_rate": 9.69299133835906e-06, + "loss": 0.5371, + "step": 1884 + }, + { + "epoch": 0.6031356655290102, + "grad_norm": 0.5547313094139099, + "learning_rate": 9.692348671428038e-06, + "loss": 0.5699, + "step": 1885 + }, + { + "epoch": 0.6034556313993175, + "grad_norm": 0.5818104147911072, + "learning_rate": 9.691705353897229e-06, + "loss": 0.5507, + "step": 1886 + }, + { + "epoch": 0.6037755972696246, + "grad_norm": 0.5582051873207092, + "learning_rate": 9.69106138585583e-06, + "loss": 0.5771, + "step": 1887 + }, + { + "epoch": 0.6040955631399317, + "grad_norm": 0.66030353307724, + "learning_rate": 9.690416767393124e-06, + "loss": 0.5983, + "step": 1888 + }, + { + "epoch": 0.6044155290102389, + "grad_norm": 0.5691258907318115, + "learning_rate": 9.689771498598491e-06, + "loss": 0.5335, + "step": 1889 + }, + { + "epoch": 0.6047354948805461, + "grad_norm": 0.663033664226532, + "learning_rate": 9.689125579561398e-06, + "loss": 0.5597, + "step": 1890 + }, + { + "epoch": 0.6050554607508533, + "grad_norm": 0.6218370795249939, + "learning_rate": 9.688479010371403e-06, + "loss": 0.5757, + "step": 1891 + }, + { + "epoch": 0.6053754266211604, + "grad_norm": 0.6317211389541626, + "learning_rate": 9.687831791118151e-06, + "loss": 0.6012, + "step": 1892 + }, + { + "epoch": 0.6056953924914675, + "grad_norm": 0.5636645555496216, + "learning_rate": 9.687183921891383e-06, + "loss": 0.5671, + "step": 1893 + }, + { + "epoch": 0.6060153583617748, + "grad_norm": 0.5526756644248962, + "learning_rate": 9.686535402780926e-06, + "loss": 0.5263, + "step": 1894 + }, + { + "epoch": 0.6063353242320819, + "grad_norm": 0.5666709542274475, + "learning_rate": 9.685886233876696e-06, + "loss": 0.5286, + "step": 1895 + }, + { + "epoch": 0.606655290102389, + "grad_norm": 0.579018235206604, + "learning_rate": 9.685236415268702e-06, + "loss": 0.5574, + "step": 1896 + }, + { + "epoch": 0.6069752559726962, + "grad_norm": 0.5683245658874512, + "learning_rate": 9.684585947047045e-06, + "loss": 0.582, + "step": 1897 + }, + { + "epoch": 0.6072952218430034, + "grad_norm": 0.6050066351890564, + "learning_rate": 9.683934829301909e-06, + "loss": 0.5789, + "step": 1898 + }, + { + "epoch": 0.6076151877133106, + "grad_norm": 0.5764254927635193, + "learning_rate": 9.683283062123574e-06, + "loss": 0.572, + "step": 1899 + }, + { + "epoch": 0.6079351535836177, + "grad_norm": 0.4990369975566864, + "learning_rate": 9.682630645602409e-06, + "loss": 0.5779, + "step": 1900 + }, + { + "epoch": 0.608255119453925, + "grad_norm": 0.5741258859634399, + "learning_rate": 9.681977579828869e-06, + "loss": 0.5449, + "step": 1901 + }, + { + "epoch": 0.6085750853242321, + "grad_norm": 0.570561945438385, + "learning_rate": 9.681323864893509e-06, + "loss": 0.5505, + "step": 1902 + }, + { + "epoch": 0.6088950511945392, + "grad_norm": 0.6162533164024353, + "learning_rate": 9.68066950088696e-06, + "loss": 0.5623, + "step": 1903 + }, + { + "epoch": 0.6092150170648464, + "grad_norm": 0.6171571016311646, + "learning_rate": 9.680014487899957e-06, + "loss": 0.5874, + "step": 1904 + }, + { + "epoch": 0.6095349829351536, + "grad_norm": 0.5293908715248108, + "learning_rate": 9.679358826023312e-06, + "loss": 0.5515, + "step": 1905 + }, + { + "epoch": 0.6098549488054608, + "grad_norm": 0.5890594720840454, + "learning_rate": 9.678702515347937e-06, + "loss": 0.57, + "step": 1906 + }, + { + "epoch": 0.6101749146757679, + "grad_norm": 0.5866312384605408, + "learning_rate": 9.678045555964831e-06, + "loss": 0.575, + "step": 1907 + }, + { + "epoch": 0.610494880546075, + "grad_norm": 0.6178193092346191, + "learning_rate": 9.67738794796508e-06, + "loss": 0.5976, + "step": 1908 + }, + { + "epoch": 0.6108148464163823, + "grad_norm": 0.594685971736908, + "learning_rate": 9.676729691439864e-06, + "loss": 0.5369, + "step": 1909 + }, + { + "epoch": 0.6111348122866894, + "grad_norm": 0.6227867603302002, + "learning_rate": 9.676070786480447e-06, + "loss": 0.5444, + "step": 1910 + }, + { + "epoch": 0.6114547781569966, + "grad_norm": 0.6000764966011047, + "learning_rate": 9.675411233178193e-06, + "loss": 0.53, + "step": 1911 + }, + { + "epoch": 0.6117747440273038, + "grad_norm": 0.6180450916290283, + "learning_rate": 9.674751031624546e-06, + "loss": 0.5318, + "step": 1912 + }, + { + "epoch": 0.612094709897611, + "grad_norm": 0.572875440120697, + "learning_rate": 9.674090181911044e-06, + "loss": 0.5393, + "step": 1913 + }, + { + "epoch": 0.6124146757679181, + "grad_norm": 0.6435962915420532, + "learning_rate": 9.673428684129317e-06, + "loss": 0.5432, + "step": 1914 + }, + { + "epoch": 0.6127346416382252, + "grad_norm": 0.6471882462501526, + "learning_rate": 9.672766538371079e-06, + "loss": 0.5597, + "step": 1915 + }, + { + "epoch": 0.6130546075085325, + "grad_norm": 0.621203601360321, + "learning_rate": 9.67210374472814e-06, + "loss": 0.5464, + "step": 1916 + }, + { + "epoch": 0.6133745733788396, + "grad_norm": 0.5741633176803589, + "learning_rate": 9.671440303292395e-06, + "loss": 0.5824, + "step": 1917 + }, + { + "epoch": 0.6136945392491467, + "grad_norm": 0.559482991695404, + "learning_rate": 9.670776214155833e-06, + "loss": 0.5802, + "step": 1918 + }, + { + "epoch": 0.6140145051194539, + "grad_norm": 0.5610204935073853, + "learning_rate": 9.670111477410528e-06, + "loss": 0.5608, + "step": 1919 + }, + { + "epoch": 0.6143344709897611, + "grad_norm": 0.6152501106262207, + "learning_rate": 9.66944609314865e-06, + "loss": 0.5532, + "step": 1920 + }, + { + "epoch": 0.6146544368600683, + "grad_norm": 0.599856436252594, + "learning_rate": 9.668780061462455e-06, + "loss": 0.5923, + "step": 1921 + }, + { + "epoch": 0.6149744027303754, + "grad_norm": 0.661961555480957, + "learning_rate": 9.668113382444287e-06, + "loss": 0.5971, + "step": 1922 + }, + { + "epoch": 0.6152943686006825, + "grad_norm": 0.605215847492218, + "learning_rate": 9.66744605618658e-06, + "loss": 0.5498, + "step": 1923 + }, + { + "epoch": 0.6156143344709898, + "grad_norm": 0.5865672826766968, + "learning_rate": 9.666778082781865e-06, + "loss": 0.5347, + "step": 1924 + }, + { + "epoch": 0.6159343003412969, + "grad_norm": 0.6111371517181396, + "learning_rate": 9.666109462322753e-06, + "loss": 0.59, + "step": 1925 + }, + { + "epoch": 0.6162542662116041, + "grad_norm": 0.5758267641067505, + "learning_rate": 9.665440194901951e-06, + "loss": 0.5529, + "step": 1926 + }, + { + "epoch": 0.6165742320819113, + "grad_norm": 0.6464254260063171, + "learning_rate": 9.664770280612255e-06, + "loss": 0.5979, + "step": 1927 + }, + { + "epoch": 0.6168941979522184, + "grad_norm": 0.6565837264060974, + "learning_rate": 9.664099719546547e-06, + "loss": 0.5606, + "step": 1928 + }, + { + "epoch": 0.6172141638225256, + "grad_norm": 0.5908191800117493, + "learning_rate": 9.663428511797801e-06, + "loss": 0.5834, + "step": 1929 + }, + { + "epoch": 0.6175341296928327, + "grad_norm": 0.5503283143043518, + "learning_rate": 9.66275665745908e-06, + "loss": 0.5443, + "step": 1930 + }, + { + "epoch": 0.61785409556314, + "grad_norm": 0.6072626113891602, + "learning_rate": 9.66208415662354e-06, + "loss": 0.5551, + "step": 1931 + }, + { + "epoch": 0.6181740614334471, + "grad_norm": 0.5384520292282104, + "learning_rate": 9.661411009384426e-06, + "loss": 0.5487, + "step": 1932 + }, + { + "epoch": 0.6184940273037542, + "grad_norm": 0.5442841649055481, + "learning_rate": 9.660737215835066e-06, + "loss": 0.5345, + "step": 1933 + }, + { + "epoch": 0.6188139931740614, + "grad_norm": 0.5890656113624573, + "learning_rate": 9.660062776068883e-06, + "loss": 0.5655, + "step": 1934 + }, + { + "epoch": 0.6191339590443686, + "grad_norm": 0.5688875317573547, + "learning_rate": 9.659387690179392e-06, + "loss": 0.5339, + "step": 1935 + }, + { + "epoch": 0.6194539249146758, + "grad_norm": 0.5539190769195557, + "learning_rate": 9.658711958260189e-06, + "loss": 0.5863, + "step": 1936 + }, + { + "epoch": 0.6197738907849829, + "grad_norm": 0.6129565834999084, + "learning_rate": 9.658035580404972e-06, + "loss": 0.5742, + "step": 1937 + }, + { + "epoch": 0.6200938566552902, + "grad_norm": 0.5914084911346436, + "learning_rate": 9.657358556707516e-06, + "loss": 0.5388, + "step": 1938 + }, + { + "epoch": 0.6204138225255973, + "grad_norm": 0.5551041960716248, + "learning_rate": 9.656680887261693e-06, + "loss": 0.5513, + "step": 1939 + }, + { + "epoch": 0.6207337883959044, + "grad_norm": 0.6423928141593933, + "learning_rate": 9.656002572161463e-06, + "loss": 0.5914, + "step": 1940 + }, + { + "epoch": 0.6210537542662116, + "grad_norm": 0.585381031036377, + "learning_rate": 9.655323611500876e-06, + "loss": 0.5434, + "step": 1941 + }, + { + "epoch": 0.6213737201365188, + "grad_norm": 0.6105824708938599, + "learning_rate": 9.65464400537407e-06, + "loss": 0.5868, + "step": 1942 + }, + { + "epoch": 0.621693686006826, + "grad_norm": 0.5563498139381409, + "learning_rate": 9.65396375387527e-06, + "loss": 0.5621, + "step": 1943 + }, + { + "epoch": 0.6220136518771331, + "grad_norm": 0.606756329536438, + "learning_rate": 9.6532828570988e-06, + "loss": 0.545, + "step": 1944 + }, + { + "epoch": 0.6223336177474402, + "grad_norm": 0.4880354106426239, + "learning_rate": 9.652601315139063e-06, + "loss": 0.5487, + "step": 1945 + }, + { + "epoch": 0.6226535836177475, + "grad_norm": 0.527872622013092, + "learning_rate": 9.651919128090557e-06, + "loss": 0.5838, + "step": 1946 + }, + { + "epoch": 0.6229735494880546, + "grad_norm": 0.5351196527481079, + "learning_rate": 9.651236296047866e-06, + "loss": 0.549, + "step": 1947 + }, + { + "epoch": 0.6232935153583617, + "grad_norm": 0.6507715582847595, + "learning_rate": 9.650552819105667e-06, + "loss": 0.591, + "step": 1948 + }, + { + "epoch": 0.6236134812286689, + "grad_norm": 0.6204143762588501, + "learning_rate": 9.649868697358727e-06, + "loss": 0.5482, + "step": 1949 + }, + { + "epoch": 0.6239334470989761, + "grad_norm": 0.568878173828125, + "learning_rate": 9.649183930901895e-06, + "loss": 0.5811, + "step": 1950 + }, + { + "epoch": 0.6242534129692833, + "grad_norm": 0.6354591846466064, + "learning_rate": 9.64849851983012e-06, + "loss": 0.5746, + "step": 1951 + }, + { + "epoch": 0.6245733788395904, + "grad_norm": 0.6112775802612305, + "learning_rate": 9.647812464238434e-06, + "loss": 0.5806, + "step": 1952 + }, + { + "epoch": 0.6248933447098977, + "grad_norm": 0.5902199745178223, + "learning_rate": 9.647125764221956e-06, + "loss": 0.5939, + "step": 1953 + }, + { + "epoch": 0.6252133105802048, + "grad_norm": 0.544347882270813, + "learning_rate": 9.646438419875901e-06, + "loss": 0.585, + "step": 1954 + }, + { + "epoch": 0.6255332764505119, + "grad_norm": 0.5880687832832336, + "learning_rate": 9.645750431295568e-06, + "loss": 0.5594, + "step": 1955 + }, + { + "epoch": 0.6258532423208191, + "grad_norm": 0.711134135723114, + "learning_rate": 9.645061798576352e-06, + "loss": 0.5548, + "step": 1956 + }, + { + "epoch": 0.6261732081911263, + "grad_norm": 0.5626425743103027, + "learning_rate": 9.644372521813728e-06, + "loss": 0.5494, + "step": 1957 + }, + { + "epoch": 0.6264931740614335, + "grad_norm": 0.692373514175415, + "learning_rate": 9.643682601103265e-06, + "loss": 0.6095, + "step": 1958 + }, + { + "epoch": 0.6268131399317406, + "grad_norm": 0.5923463702201843, + "learning_rate": 9.642992036540624e-06, + "loss": 0.5349, + "step": 1959 + }, + { + "epoch": 0.6271331058020477, + "grad_norm": 0.6308784484863281, + "learning_rate": 9.64230082822155e-06, + "loss": 0.5608, + "step": 1960 + }, + { + "epoch": 0.627453071672355, + "grad_norm": 0.6047257781028748, + "learning_rate": 9.641608976241883e-06, + "loss": 0.5927, + "step": 1961 + }, + { + "epoch": 0.6277730375426621, + "grad_norm": 0.7537582516670227, + "learning_rate": 9.640916480697549e-06, + "loss": 0.5708, + "step": 1962 + }, + { + "epoch": 0.6280930034129693, + "grad_norm": 0.6602718234062195, + "learning_rate": 9.640223341684559e-06, + "loss": 0.5749, + "step": 1963 + }, + { + "epoch": 0.6284129692832765, + "grad_norm": 0.5211908221244812, + "learning_rate": 9.639529559299022e-06, + "loss": 0.5495, + "step": 1964 + }, + { + "epoch": 0.6287329351535836, + "grad_norm": 0.6491508483886719, + "learning_rate": 9.638835133637129e-06, + "loss": 0.5618, + "step": 1965 + }, + { + "epoch": 0.6290529010238908, + "grad_norm": 0.6506140828132629, + "learning_rate": 9.638140064795165e-06, + "loss": 0.5724, + "step": 1966 + }, + { + "epoch": 0.6293728668941979, + "grad_norm": 0.6765151619911194, + "learning_rate": 9.6374443528695e-06, + "loss": 0.6029, + "step": 1967 + }, + { + "epoch": 0.6296928327645052, + "grad_norm": 0.5773376822471619, + "learning_rate": 9.636747997956597e-06, + "loss": 0.5702, + "step": 1968 + }, + { + "epoch": 0.6300127986348123, + "grad_norm": 0.6763083338737488, + "learning_rate": 9.636051000153006e-06, + "loss": 0.5499, + "step": 1969 + }, + { + "epoch": 0.6303327645051194, + "grad_norm": 0.6451863050460815, + "learning_rate": 9.635353359555366e-06, + "loss": 0.5408, + "step": 1970 + }, + { + "epoch": 0.6306527303754266, + "grad_norm": 0.6031057834625244, + "learning_rate": 9.634655076260409e-06, + "loss": 0.5545, + "step": 1971 + }, + { + "epoch": 0.6309726962457338, + "grad_norm": 0.6576459407806396, + "learning_rate": 9.633956150364948e-06, + "loss": 0.5346, + "step": 1972 + }, + { + "epoch": 0.631292662116041, + "grad_norm": 0.6280539035797119, + "learning_rate": 9.63325658196589e-06, + "loss": 0.5397, + "step": 1973 + }, + { + "epoch": 0.6316126279863481, + "grad_norm": 0.5733145475387573, + "learning_rate": 9.632556371160235e-06, + "loss": 0.58, + "step": 1974 + }, + { + "epoch": 0.6319325938566553, + "grad_norm": 0.642741858959198, + "learning_rate": 9.631855518045066e-06, + "loss": 0.5692, + "step": 1975 + }, + { + "epoch": 0.6322525597269625, + "grad_norm": 0.6114027500152588, + "learning_rate": 9.631154022717559e-06, + "loss": 0.5559, + "step": 1976 + }, + { + "epoch": 0.6325725255972696, + "grad_norm": 0.525458812713623, + "learning_rate": 9.630451885274972e-06, + "loss": 0.5216, + "step": 1977 + }, + { + "epoch": 0.6328924914675768, + "grad_norm": 0.6430639624595642, + "learning_rate": 9.629749105814664e-06, + "loss": 0.5731, + "step": 1978 + }, + { + "epoch": 0.633212457337884, + "grad_norm": 0.6785262823104858, + "learning_rate": 9.629045684434069e-06, + "loss": 0.5897, + "step": 1979 + }, + { + "epoch": 0.6335324232081911, + "grad_norm": 0.5230071544647217, + "learning_rate": 9.628341621230724e-06, + "loss": 0.5355, + "step": 1980 + }, + { + "epoch": 0.6338523890784983, + "grad_norm": 0.6526293754577637, + "learning_rate": 9.627636916302244e-06, + "loss": 0.5579, + "step": 1981 + }, + { + "epoch": 0.6341723549488054, + "grad_norm": 0.5032544136047363, + "learning_rate": 9.626931569746337e-06, + "loss": 0.5699, + "step": 1982 + }, + { + "epoch": 0.6344923208191127, + "grad_norm": 0.5933758616447449, + "learning_rate": 9.626225581660802e-06, + "loss": 0.531, + "step": 1983 + }, + { + "epoch": 0.6348122866894198, + "grad_norm": 0.692962110042572, + "learning_rate": 9.625518952143526e-06, + "loss": 0.5833, + "step": 1984 + }, + { + "epoch": 0.6351322525597269, + "grad_norm": 0.6170587539672852, + "learning_rate": 9.624811681292482e-06, + "loss": 0.5408, + "step": 1985 + }, + { + "epoch": 0.6354522184300341, + "grad_norm": 0.589077889919281, + "learning_rate": 9.624103769205735e-06, + "loss": 0.5583, + "step": 1986 + }, + { + "epoch": 0.6357721843003413, + "grad_norm": 0.5993129014968872, + "learning_rate": 9.623395215981438e-06, + "loss": 0.558, + "step": 1987 + }, + { + "epoch": 0.6360921501706485, + "grad_norm": 0.5902277231216431, + "learning_rate": 9.62268602171783e-06, + "loss": 0.5258, + "step": 1988 + }, + { + "epoch": 0.6364121160409556, + "grad_norm": 0.6470785140991211, + "learning_rate": 9.621976186513246e-06, + "loss": 0.5756, + "step": 1989 + }, + { + "epoch": 0.6367320819112628, + "grad_norm": 0.6402467489242554, + "learning_rate": 9.621265710466103e-06, + "loss": 0.549, + "step": 1990 + }, + { + "epoch": 0.63705204778157, + "grad_norm": 0.6703201532363892, + "learning_rate": 9.62055459367491e-06, + "loss": 0.5687, + "step": 1991 + }, + { + "epoch": 0.6373720136518771, + "grad_norm": 0.5826023817062378, + "learning_rate": 9.619842836238262e-06, + "loss": 0.576, + "step": 1992 + }, + { + "epoch": 0.6376919795221843, + "grad_norm": 0.5884040594100952, + "learning_rate": 9.619130438254848e-06, + "loss": 0.5591, + "step": 1993 + }, + { + "epoch": 0.6380119453924915, + "grad_norm": 0.612905740737915, + "learning_rate": 9.618417399823442e-06, + "loss": 0.5536, + "step": 1994 + }, + { + "epoch": 0.6383319112627986, + "grad_norm": 0.6277275681495667, + "learning_rate": 9.617703721042908e-06, + "loss": 0.5701, + "step": 1995 + }, + { + "epoch": 0.6386518771331058, + "grad_norm": 0.5961500406265259, + "learning_rate": 9.616989402012197e-06, + "loss": 0.5349, + "step": 1996 + }, + { + "epoch": 0.6389718430034129, + "grad_norm": 0.7005389928817749, + "learning_rate": 9.616274442830351e-06, + "loss": 0.5559, + "step": 1997 + }, + { + "epoch": 0.6392918088737202, + "grad_norm": 0.5778456926345825, + "learning_rate": 9.615558843596499e-06, + "loss": 0.5772, + "step": 1998 + }, + { + "epoch": 0.6396117747440273, + "grad_norm": 0.6484526991844177, + "learning_rate": 9.614842604409863e-06, + "loss": 0.5679, + "step": 1999 + }, + { + "epoch": 0.6399317406143344, + "grad_norm": 0.6132345795631409, + "learning_rate": 9.614125725369748e-06, + "loss": 0.6166, + "step": 2000 + }, + { + "epoch": 0.6402517064846417, + "grad_norm": 0.6483284831047058, + "learning_rate": 9.61340820657555e-06, + "loss": 0.5865, + "step": 2001 + }, + { + "epoch": 0.6405716723549488, + "grad_norm": 0.6628165245056152, + "learning_rate": 9.612690048126751e-06, + "loss": 0.5806, + "step": 2002 + }, + { + "epoch": 0.640891638225256, + "grad_norm": 0.651597261428833, + "learning_rate": 9.611971250122932e-06, + "loss": 0.5479, + "step": 2003 + }, + { + "epoch": 0.6412116040955631, + "grad_norm": 0.6118643283843994, + "learning_rate": 9.611251812663748e-06, + "loss": 0.5947, + "step": 2004 + }, + { + "epoch": 0.6415315699658704, + "grad_norm": 0.6847550272941589, + "learning_rate": 9.610531735848953e-06, + "loss": 0.5963, + "step": 2005 + }, + { + "epoch": 0.6418515358361775, + "grad_norm": 0.6172077059745789, + "learning_rate": 9.609811019778388e-06, + "loss": 0.5258, + "step": 2006 + }, + { + "epoch": 0.6421715017064846, + "grad_norm": 0.7246551513671875, + "learning_rate": 9.609089664551976e-06, + "loss": 0.5895, + "step": 2007 + }, + { + "epoch": 0.6424914675767918, + "grad_norm": 0.5504890084266663, + "learning_rate": 9.60836767026974e-06, + "loss": 0.5393, + "step": 2008 + }, + { + "epoch": 0.642811433447099, + "grad_norm": 0.6666256189346313, + "learning_rate": 9.607645037031783e-06, + "loss": 0.5494, + "step": 2009 + }, + { + "epoch": 0.6431313993174061, + "grad_norm": 0.6688473224639893, + "learning_rate": 9.606921764938297e-06, + "loss": 0.5617, + "step": 2010 + }, + { + "epoch": 0.6434513651877133, + "grad_norm": 0.6392967104911804, + "learning_rate": 9.606197854089565e-06, + "loss": 0.5448, + "step": 2011 + }, + { + "epoch": 0.6437713310580204, + "grad_norm": 0.5535218715667725, + "learning_rate": 9.60547330458596e-06, + "loss": 0.5338, + "step": 2012 + }, + { + "epoch": 0.6440912969283277, + "grad_norm": 0.6675567626953125, + "learning_rate": 9.604748116527941e-06, + "loss": 0.5717, + "step": 2013 + }, + { + "epoch": 0.6444112627986348, + "grad_norm": 0.7083868384361267, + "learning_rate": 9.604022290016058e-06, + "loss": 0.5345, + "step": 2014 + }, + { + "epoch": 0.644731228668942, + "grad_norm": 0.6258317232131958, + "learning_rate": 9.603295825150945e-06, + "loss": 0.5257, + "step": 2015 + }, + { + "epoch": 0.6450511945392492, + "grad_norm": 0.6635270118713379, + "learning_rate": 9.602568722033325e-06, + "loss": 0.5718, + "step": 2016 + }, + { + "epoch": 0.6453711604095563, + "grad_norm": 0.5791119933128357, + "learning_rate": 9.601840980764016e-06, + "loss": 0.5769, + "step": 2017 + }, + { + "epoch": 0.6456911262798635, + "grad_norm": 0.5586481094360352, + "learning_rate": 9.601112601443921e-06, + "loss": 0.5455, + "step": 2018 + }, + { + "epoch": 0.6460110921501706, + "grad_norm": 0.6536003947257996, + "learning_rate": 9.600383584174026e-06, + "loss": 0.5536, + "step": 2019 + }, + { + "epoch": 0.6463310580204779, + "grad_norm": 0.5298084616661072, + "learning_rate": 9.599653929055415e-06, + "loss": 0.541, + "step": 2020 + }, + { + "epoch": 0.646651023890785, + "grad_norm": 0.5338242650032043, + "learning_rate": 9.598923636189252e-06, + "loss": 0.5341, + "step": 2021 + }, + { + "epoch": 0.6469709897610921, + "grad_norm": 0.5623722076416016, + "learning_rate": 9.598192705676793e-06, + "loss": 0.5432, + "step": 2022 + }, + { + "epoch": 0.6472909556313993, + "grad_norm": 0.6378763914108276, + "learning_rate": 9.597461137619385e-06, + "loss": 0.5895, + "step": 2023 + }, + { + "epoch": 0.6476109215017065, + "grad_norm": 0.5924339294433594, + "learning_rate": 9.59672893211846e-06, + "loss": 0.5071, + "step": 2024 + }, + { + "epoch": 0.6479308873720137, + "grad_norm": 0.5660730600357056, + "learning_rate": 9.595996089275536e-06, + "loss": 0.5536, + "step": 2025 + }, + { + "epoch": 0.6482508532423208, + "grad_norm": 0.6002275347709656, + "learning_rate": 9.595262609192227e-06, + "loss": 0.5031, + "step": 2026 + }, + { + "epoch": 0.648570819112628, + "grad_norm": 0.6197472214698792, + "learning_rate": 9.594528491970228e-06, + "loss": 0.5654, + "step": 2027 + }, + { + "epoch": 0.6488907849829352, + "grad_norm": 0.5491067171096802, + "learning_rate": 9.593793737711328e-06, + "loss": 0.5654, + "step": 2028 + }, + { + "epoch": 0.6492107508532423, + "grad_norm": 0.6247221827507019, + "learning_rate": 9.593058346517399e-06, + "loss": 0.559, + "step": 2029 + }, + { + "epoch": 0.6495307167235495, + "grad_norm": 0.636074960231781, + "learning_rate": 9.592322318490404e-06, + "loss": 0.5868, + "step": 2030 + }, + { + "epoch": 0.6498506825938567, + "grad_norm": 0.5542947053909302, + "learning_rate": 9.591585653732396e-06, + "loss": 0.5166, + "step": 2031 + }, + { + "epoch": 0.6501706484641638, + "grad_norm": 0.5761454105377197, + "learning_rate": 9.590848352345512e-06, + "loss": 0.5599, + "step": 2032 + }, + { + "epoch": 0.650490614334471, + "grad_norm": 0.6053789258003235, + "learning_rate": 9.590110414431981e-06, + "loss": 0.5886, + "step": 2033 + }, + { + "epoch": 0.6508105802047781, + "grad_norm": 0.5515697598457336, + "learning_rate": 9.58937184009412e-06, + "loss": 0.5624, + "step": 2034 + }, + { + "epoch": 0.6511305460750854, + "grad_norm": 0.6260466575622559, + "learning_rate": 9.588632629434331e-06, + "loss": 0.5709, + "step": 2035 + }, + { + "epoch": 0.6514505119453925, + "grad_norm": 0.5678718090057373, + "learning_rate": 9.587892782555109e-06, + "loss": 0.5663, + "step": 2036 + }, + { + "epoch": 0.6517704778156996, + "grad_norm": 0.6218145489692688, + "learning_rate": 9.587152299559033e-06, + "loss": 0.5648, + "step": 2037 + }, + { + "epoch": 0.6520904436860068, + "grad_norm": 0.5651326179504395, + "learning_rate": 9.586411180548771e-06, + "loss": 0.5434, + "step": 2038 + }, + { + "epoch": 0.652410409556314, + "grad_norm": 0.5573369264602661, + "learning_rate": 9.585669425627084e-06, + "loss": 0.5597, + "step": 2039 + }, + { + "epoch": 0.6527303754266212, + "grad_norm": 0.6214823126792908, + "learning_rate": 9.584927034896811e-06, + "loss": 0.5441, + "step": 2040 + }, + { + "epoch": 0.6530503412969283, + "grad_norm": 0.6007993221282959, + "learning_rate": 9.58418400846089e-06, + "loss": 0.5621, + "step": 2041 + }, + { + "epoch": 0.6533703071672355, + "grad_norm": 0.5891932249069214, + "learning_rate": 9.583440346422343e-06, + "loss": 0.5827, + "step": 2042 + }, + { + "epoch": 0.6536902730375427, + "grad_norm": 0.6093569993972778, + "learning_rate": 9.582696048884277e-06, + "loss": 0.5996, + "step": 2043 + }, + { + "epoch": 0.6540102389078498, + "grad_norm": 0.616129457950592, + "learning_rate": 9.58195111594989e-06, + "loss": 0.5865, + "step": 2044 + }, + { + "epoch": 0.654330204778157, + "grad_norm": 0.6123115420341492, + "learning_rate": 9.581205547722469e-06, + "loss": 0.5576, + "step": 2045 + }, + { + "epoch": 0.6546501706484642, + "grad_norm": 0.6001349687576294, + "learning_rate": 9.580459344305388e-06, + "loss": 0.5784, + "step": 2046 + }, + { + "epoch": 0.6549701365187713, + "grad_norm": 0.5365661382675171, + "learning_rate": 9.579712505802108e-06, + "loss": 0.5567, + "step": 2047 + }, + { + "epoch": 0.6552901023890785, + "grad_norm": 0.5674573183059692, + "learning_rate": 9.578965032316179e-06, + "loss": 0.4916, + "step": 2048 + }, + { + "epoch": 0.6556100682593856, + "grad_norm": 0.6029785871505737, + "learning_rate": 9.57821692395124e-06, + "loss": 0.5384, + "step": 2049 + }, + { + "epoch": 0.6559300341296929, + "grad_norm": 0.6047776341438293, + "learning_rate": 9.577468180811018e-06, + "loss": 0.6, + "step": 2050 + }, + { + "epoch": 0.65625, + "grad_norm": 0.6041454076766968, + "learning_rate": 9.576718802999323e-06, + "loss": 0.5408, + "step": 2051 + }, + { + "epoch": 0.6565699658703071, + "grad_norm": 0.6480724215507507, + "learning_rate": 9.575968790620062e-06, + "loss": 0.6052, + "step": 2052 + }, + { + "epoch": 0.6568899317406144, + "grad_norm": 0.6151446104049683, + "learning_rate": 9.575218143777223e-06, + "loss": 0.5528, + "step": 2053 + }, + { + "epoch": 0.6572098976109215, + "grad_norm": 0.608056366443634, + "learning_rate": 9.574466862574885e-06, + "loss": 0.5174, + "step": 2054 + }, + { + "epoch": 0.6575298634812287, + "grad_norm": 0.6194968819618225, + "learning_rate": 9.573714947117211e-06, + "loss": 0.5555, + "step": 2055 + }, + { + "epoch": 0.6578498293515358, + "grad_norm": 0.5183257460594177, + "learning_rate": 9.57296239750846e-06, + "loss": 0.5358, + "step": 2056 + }, + { + "epoch": 0.658169795221843, + "grad_norm": 0.6629782915115356, + "learning_rate": 9.57220921385297e-06, + "loss": 0.6058, + "step": 2057 + }, + { + "epoch": 0.6584897610921502, + "grad_norm": 0.6532469987869263, + "learning_rate": 9.571455396255172e-06, + "loss": 0.6021, + "step": 2058 + }, + { + "epoch": 0.6588097269624573, + "grad_norm": 0.6884270906448364, + "learning_rate": 9.570700944819584e-06, + "loss": 0.5639, + "step": 2059 + }, + { + "epoch": 0.6591296928327645, + "grad_norm": 0.5459734797477722, + "learning_rate": 9.56994585965081e-06, + "loss": 0.5185, + "step": 2060 + }, + { + "epoch": 0.6594496587030717, + "grad_norm": 0.5665500164031982, + "learning_rate": 9.569190140853547e-06, + "loss": 0.5767, + "step": 2061 + }, + { + "epoch": 0.6597696245733788, + "grad_norm": 0.5893674492835999, + "learning_rate": 9.568433788532571e-06, + "loss": 0.5901, + "step": 2062 + }, + { + "epoch": 0.660089590443686, + "grad_norm": 0.6321619749069214, + "learning_rate": 9.567676802792758e-06, + "loss": 0.556, + "step": 2063 + }, + { + "epoch": 0.6604095563139932, + "grad_norm": 0.6531881093978882, + "learning_rate": 9.566919183739058e-06, + "loss": 0.5742, + "step": 2064 + }, + { + "epoch": 0.6607295221843004, + "grad_norm": 0.7623351216316223, + "learning_rate": 9.566160931476522e-06, + "loss": 0.5853, + "step": 2065 + }, + { + "epoch": 0.6610494880546075, + "grad_norm": 0.5670693516731262, + "learning_rate": 9.565402046110277e-06, + "loss": 0.5645, + "step": 2066 + }, + { + "epoch": 0.6613694539249146, + "grad_norm": 0.5728259086608887, + "learning_rate": 9.564642527745546e-06, + "loss": 0.5466, + "step": 2067 + }, + { + "epoch": 0.6616894197952219, + "grad_norm": 0.625257134437561, + "learning_rate": 9.563882376487638e-06, + "loss": 0.5377, + "step": 2068 + }, + { + "epoch": 0.662009385665529, + "grad_norm": 0.6368067264556885, + "learning_rate": 9.563121592441949e-06, + "loss": 0.6021, + "step": 2069 + }, + { + "epoch": 0.6623293515358362, + "grad_norm": 0.6265316605567932, + "learning_rate": 9.562360175713962e-06, + "loss": 0.5594, + "step": 2070 + }, + { + "epoch": 0.6626493174061433, + "grad_norm": 0.550437867641449, + "learning_rate": 9.561598126409247e-06, + "loss": 0.5333, + "step": 2071 + }, + { + "epoch": 0.6629692832764505, + "grad_norm": 0.5962300300598145, + "learning_rate": 9.560835444633463e-06, + "loss": 0.5759, + "step": 2072 + }, + { + "epoch": 0.6632892491467577, + "grad_norm": 0.5362560153007507, + "learning_rate": 9.56007213049236e-06, + "loss": 0.5436, + "step": 2073 + }, + { + "epoch": 0.6636092150170648, + "grad_norm": 0.5743743181228638, + "learning_rate": 9.559308184091768e-06, + "loss": 0.5559, + "step": 2074 + }, + { + "epoch": 0.663929180887372, + "grad_norm": 0.5720234513282776, + "learning_rate": 9.55854360553761e-06, + "loss": 0.5366, + "step": 2075 + }, + { + "epoch": 0.6642491467576792, + "grad_norm": 0.6186221837997437, + "learning_rate": 9.557778394935901e-06, + "loss": 0.6093, + "step": 2076 + }, + { + "epoch": 0.6645691126279863, + "grad_norm": 0.5654708743095398, + "learning_rate": 9.557012552392733e-06, + "loss": 0.5436, + "step": 2077 + }, + { + "epoch": 0.6648890784982935, + "grad_norm": 0.6198699474334717, + "learning_rate": 9.55624607801429e-06, + "loss": 0.5633, + "step": 2078 + }, + { + "epoch": 0.6652090443686007, + "grad_norm": 0.5168883800506592, + "learning_rate": 9.55547897190685e-06, + "loss": 0.5425, + "step": 2079 + }, + { + "epoch": 0.6655290102389079, + "grad_norm": 0.5414618849754333, + "learning_rate": 9.55471123417677e-06, + "loss": 0.54, + "step": 2080 + }, + { + "epoch": 0.665848976109215, + "grad_norm": 0.5594053268432617, + "learning_rate": 9.553942864930496e-06, + "loss": 0.5526, + "step": 2081 + }, + { + "epoch": 0.6661689419795221, + "grad_norm": 0.5626882910728455, + "learning_rate": 9.553173864274567e-06, + "loss": 0.5354, + "step": 2082 + }, + { + "epoch": 0.6664889078498294, + "grad_norm": 0.5668383240699768, + "learning_rate": 9.552404232315602e-06, + "loss": 0.542, + "step": 2083 + }, + { + "epoch": 0.6668088737201365, + "grad_norm": 0.548861563205719, + "learning_rate": 9.551633969160316e-06, + "loss": 0.5673, + "step": 2084 + }, + { + "epoch": 0.6671288395904437, + "grad_norm": 0.5764604806900024, + "learning_rate": 9.550863074915505e-06, + "loss": 0.5761, + "step": 2085 + }, + { + "epoch": 0.6674488054607508, + "grad_norm": 0.5341320633888245, + "learning_rate": 9.550091549688049e-06, + "loss": 0.533, + "step": 2086 + }, + { + "epoch": 0.667768771331058, + "grad_norm": 0.5438635349273682, + "learning_rate": 9.54931939358493e-06, + "loss": 0.5435, + "step": 2087 + }, + { + "epoch": 0.6680887372013652, + "grad_norm": 0.5967687368392944, + "learning_rate": 9.548546606713203e-06, + "loss": 0.5621, + "step": 2088 + }, + { + "epoch": 0.6684087030716723, + "grad_norm": 0.5669074654579163, + "learning_rate": 9.547773189180018e-06, + "loss": 0.5309, + "step": 2089 + }, + { + "epoch": 0.6687286689419796, + "grad_norm": 0.5866972208023071, + "learning_rate": 9.546999141092608e-06, + "loss": 0.5896, + "step": 2090 + }, + { + "epoch": 0.6690486348122867, + "grad_norm": 0.5700693130493164, + "learning_rate": 9.546224462558297e-06, + "loss": 0.5485, + "step": 2091 + }, + { + "epoch": 0.6693686006825939, + "grad_norm": 0.566959798336029, + "learning_rate": 9.545449153684495e-06, + "loss": 0.5671, + "step": 2092 + }, + { + "epoch": 0.669688566552901, + "grad_norm": 0.5802577137947083, + "learning_rate": 9.544673214578699e-06, + "loss": 0.5132, + "step": 2093 + }, + { + "epoch": 0.6700085324232082, + "grad_norm": 0.6248685121536255, + "learning_rate": 9.543896645348496e-06, + "loss": 0.5444, + "step": 2094 + }, + { + "epoch": 0.6703284982935154, + "grad_norm": 0.5355051159858704, + "learning_rate": 9.543119446101556e-06, + "loss": 0.5325, + "step": 2095 + }, + { + "epoch": 0.6706484641638225, + "grad_norm": 0.5653332471847534, + "learning_rate": 9.542341616945638e-06, + "loss": 0.583, + "step": 2096 + }, + { + "epoch": 0.6709684300341296, + "grad_norm": 0.6538986563682556, + "learning_rate": 9.541563157988591e-06, + "loss": 0.5806, + "step": 2097 + }, + { + "epoch": 0.6712883959044369, + "grad_norm": 0.563230037689209, + "learning_rate": 9.540784069338349e-06, + "loss": 0.571, + "step": 2098 + }, + { + "epoch": 0.671608361774744, + "grad_norm": 0.6563031077384949, + "learning_rate": 9.540004351102933e-06, + "loss": 0.568, + "step": 2099 + }, + { + "epoch": 0.6719283276450512, + "grad_norm": 0.6009393334388733, + "learning_rate": 9.539224003390453e-06, + "loss": 0.5772, + "step": 2100 + }, + { + "epoch": 0.6722482935153583, + "grad_norm": 0.5868406891822815, + "learning_rate": 9.538443026309102e-06, + "loss": 0.5695, + "step": 2101 + }, + { + "epoch": 0.6725682593856656, + "grad_norm": 0.5810580253601074, + "learning_rate": 9.537661419967167e-06, + "loss": 0.5536, + "step": 2102 + }, + { + "epoch": 0.6728882252559727, + "grad_norm": 0.576774537563324, + "learning_rate": 9.536879184473016e-06, + "loss": 0.5695, + "step": 2103 + }, + { + "epoch": 0.6732081911262798, + "grad_norm": 0.5885899066925049, + "learning_rate": 9.536096319935109e-06, + "loss": 0.5558, + "step": 2104 + }, + { + "epoch": 0.6735281569965871, + "grad_norm": 0.548465371131897, + "learning_rate": 9.53531282646199e-06, + "loss": 0.558, + "step": 2105 + }, + { + "epoch": 0.6738481228668942, + "grad_norm": 0.6229990124702454, + "learning_rate": 9.53452870416229e-06, + "loss": 0.5522, + "step": 2106 + }, + { + "epoch": 0.6741680887372014, + "grad_norm": 0.6606165170669556, + "learning_rate": 9.533743953144733e-06, + "loss": 0.5813, + "step": 2107 + }, + { + "epoch": 0.6744880546075085, + "grad_norm": 0.5467153787612915, + "learning_rate": 9.532958573518121e-06, + "loss": 0.5552, + "step": 2108 + }, + { + "epoch": 0.6748080204778157, + "grad_norm": 0.6055881977081299, + "learning_rate": 9.532172565391348e-06, + "loss": 0.5708, + "step": 2109 + }, + { + "epoch": 0.6751279863481229, + "grad_norm": 0.656714916229248, + "learning_rate": 9.5313859288734e-06, + "loss": 0.6132, + "step": 2110 + }, + { + "epoch": 0.67544795221843, + "grad_norm": 0.6124818325042725, + "learning_rate": 9.530598664073339e-06, + "loss": 0.5514, + "step": 2111 + }, + { + "epoch": 0.6757679180887372, + "grad_norm": 0.6045283675193787, + "learning_rate": 9.529810771100323e-06, + "loss": 0.5938, + "step": 2112 + }, + { + "epoch": 0.6760878839590444, + "grad_norm": 0.5918228626251221, + "learning_rate": 9.529022250063596e-06, + "loss": 0.5707, + "step": 2113 + }, + { + "epoch": 0.6764078498293515, + "grad_norm": 0.6131752729415894, + "learning_rate": 9.528233101072484e-06, + "loss": 0.5446, + "step": 2114 + }, + { + "epoch": 0.6767278156996587, + "grad_norm": 0.5795077681541443, + "learning_rate": 9.527443324236403e-06, + "loss": 0.5858, + "step": 2115 + }, + { + "epoch": 0.6770477815699659, + "grad_norm": 0.5680365562438965, + "learning_rate": 9.526652919664862e-06, + "loss": 0.5822, + "step": 2116 + }, + { + "epoch": 0.6773677474402731, + "grad_norm": 0.5775081515312195, + "learning_rate": 9.525861887467445e-06, + "loss": 0.5118, + "step": 2117 + }, + { + "epoch": 0.6776877133105802, + "grad_norm": 0.5150967240333557, + "learning_rate": 9.525070227753835e-06, + "loss": 0.5428, + "step": 2118 + }, + { + "epoch": 0.6780076791808873, + "grad_norm": 0.5972749590873718, + "learning_rate": 9.524277940633793e-06, + "loss": 0.5869, + "step": 2119 + }, + { + "epoch": 0.6783276450511946, + "grad_norm": 0.5536697506904602, + "learning_rate": 9.523485026217173e-06, + "loss": 0.5436, + "step": 2120 + }, + { + "epoch": 0.6786476109215017, + "grad_norm": 0.5641371011734009, + "learning_rate": 9.52269148461391e-06, + "loss": 0.5613, + "step": 2121 + }, + { + "epoch": 0.6789675767918089, + "grad_norm": 0.5990756750106812, + "learning_rate": 9.521897315934032e-06, + "loss": 0.5683, + "step": 2122 + }, + { + "epoch": 0.679287542662116, + "grad_norm": 0.6080614924430847, + "learning_rate": 9.521102520287653e-06, + "loss": 0.5252, + "step": 2123 + }, + { + "epoch": 0.6796075085324232, + "grad_norm": 0.6255257725715637, + "learning_rate": 9.520307097784968e-06, + "loss": 0.5752, + "step": 2124 + }, + { + "epoch": 0.6799274744027304, + "grad_norm": 0.5562875866889954, + "learning_rate": 9.519511048536268e-06, + "loss": 0.5629, + "step": 2125 + }, + { + "epoch": 0.6802474402730375, + "grad_norm": 0.6114696860313416, + "learning_rate": 9.518714372651922e-06, + "loss": 0.5415, + "step": 2126 + }, + { + "epoch": 0.6805674061433447, + "grad_norm": 0.5472313761711121, + "learning_rate": 9.517917070242395e-06, + "loss": 0.5656, + "step": 2127 + }, + { + "epoch": 0.6808873720136519, + "grad_norm": 0.558124840259552, + "learning_rate": 9.51711914141823e-06, + "loss": 0.5919, + "step": 2128 + }, + { + "epoch": 0.681207337883959, + "grad_norm": 0.518718421459198, + "learning_rate": 9.51632058629006e-06, + "loss": 0.5501, + "step": 2129 + }, + { + "epoch": 0.6815273037542662, + "grad_norm": 0.6277896165847778, + "learning_rate": 9.515521404968608e-06, + "loss": 0.5744, + "step": 2130 + }, + { + "epoch": 0.6818472696245734, + "grad_norm": 0.5866926312446594, + "learning_rate": 9.514721597564683e-06, + "loss": 0.5584, + "step": 2131 + }, + { + "epoch": 0.6821672354948806, + "grad_norm": 0.6311371326446533, + "learning_rate": 9.513921164189174e-06, + "loss": 0.5514, + "step": 2132 + }, + { + "epoch": 0.6824872013651877, + "grad_norm": 0.5464537739753723, + "learning_rate": 9.513120104953068e-06, + "loss": 0.5536, + "step": 2133 + }, + { + "epoch": 0.6828071672354948, + "grad_norm": 0.6255695223808289, + "learning_rate": 9.512318419967427e-06, + "loss": 0.5714, + "step": 2134 + }, + { + "epoch": 0.6831271331058021, + "grad_norm": 0.6228389143943787, + "learning_rate": 9.511516109343413e-06, + "loss": 0.5756, + "step": 2135 + }, + { + "epoch": 0.6834470989761092, + "grad_norm": 0.5070335865020752, + "learning_rate": 9.51071317319226e-06, + "loss": 0.5195, + "step": 2136 + }, + { + "epoch": 0.6837670648464164, + "grad_norm": 0.6195150017738342, + "learning_rate": 9.509909611625298e-06, + "loss": 0.5495, + "step": 2137 + }, + { + "epoch": 0.6840870307167235, + "grad_norm": 0.5539630651473999, + "learning_rate": 9.509105424753945e-06, + "loss": 0.5532, + "step": 2138 + }, + { + "epoch": 0.6844069965870307, + "grad_norm": 0.5826463103294373, + "learning_rate": 9.5083006126897e-06, + "loss": 0.5402, + "step": 2139 + }, + { + "epoch": 0.6847269624573379, + "grad_norm": 0.5683773159980774, + "learning_rate": 9.507495175544151e-06, + "loss": 0.6041, + "step": 2140 + }, + { + "epoch": 0.685046928327645, + "grad_norm": 0.5582982301712036, + "learning_rate": 9.506689113428974e-06, + "loss": 0.5553, + "step": 2141 + }, + { + "epoch": 0.6853668941979523, + "grad_norm": 0.5302207469940186, + "learning_rate": 9.505882426455928e-06, + "loss": 0.5365, + "step": 2142 + }, + { + "epoch": 0.6856868600682594, + "grad_norm": 0.6130880117416382, + "learning_rate": 9.505075114736864e-06, + "loss": 0.5793, + "step": 2143 + }, + { + "epoch": 0.6860068259385665, + "grad_norm": 0.6494556665420532, + "learning_rate": 9.504267178383715e-06, + "loss": 0.5663, + "step": 2144 + }, + { + "epoch": 0.6863267918088737, + "grad_norm": 0.5702915787696838, + "learning_rate": 9.503458617508504e-06, + "loss": 0.5071, + "step": 2145 + }, + { + "epoch": 0.6866467576791809, + "grad_norm": 0.6269530057907104, + "learning_rate": 9.502649432223337e-06, + "loss": 0.5841, + "step": 2146 + }, + { + "epoch": 0.6869667235494881, + "grad_norm": 0.6300735473632812, + "learning_rate": 9.50183962264041e-06, + "loss": 0.5813, + "step": 2147 + }, + { + "epoch": 0.6872866894197952, + "grad_norm": 0.6130741238594055, + "learning_rate": 9.501029188872005e-06, + "loss": 0.5486, + "step": 2148 + }, + { + "epoch": 0.6876066552901023, + "grad_norm": 0.5809904932975769, + "learning_rate": 9.500218131030485e-06, + "loss": 0.5095, + "step": 2149 + }, + { + "epoch": 0.6879266211604096, + "grad_norm": 0.6523924469947815, + "learning_rate": 9.499406449228308e-06, + "loss": 0.5684, + "step": 2150 + }, + { + "epoch": 0.6882465870307167, + "grad_norm": 0.6566171050071716, + "learning_rate": 9.498594143578015e-06, + "loss": 0.6229, + "step": 2151 + }, + { + "epoch": 0.6885665529010239, + "grad_norm": 0.6797991394996643, + "learning_rate": 9.49778121419223e-06, + "loss": 0.5686, + "step": 2152 + }, + { + "epoch": 0.6888865187713311, + "grad_norm": 0.6882073283195496, + "learning_rate": 9.496967661183672e-06, + "loss": 0.5598, + "step": 2153 + }, + { + "epoch": 0.6892064846416383, + "grad_norm": 0.5378338098526001, + "learning_rate": 9.496153484665136e-06, + "loss": 0.5425, + "step": 2154 + }, + { + "epoch": 0.6895264505119454, + "grad_norm": 0.5711609125137329, + "learning_rate": 9.49533868474951e-06, + "loss": 0.5526, + "step": 2155 + }, + { + "epoch": 0.6898464163822525, + "grad_norm": 0.6440864205360413, + "learning_rate": 9.494523261549767e-06, + "loss": 0.5619, + "step": 2156 + }, + { + "epoch": 0.6901663822525598, + "grad_norm": 0.5898805856704712, + "learning_rate": 9.493707215178969e-06, + "loss": 0.5894, + "step": 2157 + }, + { + "epoch": 0.6904863481228669, + "grad_norm": 0.6344678997993469, + "learning_rate": 9.492890545750256e-06, + "loss": 0.6375, + "step": 2158 + }, + { + "epoch": 0.690806313993174, + "grad_norm": 0.6756842732429504, + "learning_rate": 9.492073253376865e-06, + "loss": 0.6162, + "step": 2159 + }, + { + "epoch": 0.6911262798634812, + "grad_norm": 0.5748379826545715, + "learning_rate": 9.491255338172116e-06, + "loss": 0.5633, + "step": 2160 + }, + { + "epoch": 0.6914462457337884, + "grad_norm": 0.48376211524009705, + "learning_rate": 9.490436800249408e-06, + "loss": 0.5135, + "step": 2161 + }, + { + "epoch": 0.6917662116040956, + "grad_norm": 0.5189322829246521, + "learning_rate": 9.489617639722238e-06, + "loss": 0.5291, + "step": 2162 + }, + { + "epoch": 0.6920861774744027, + "grad_norm": 0.6013830900192261, + "learning_rate": 9.48879785670418e-06, + "loss": 0.5762, + "step": 2163 + }, + { + "epoch": 0.6924061433447098, + "grad_norm": 0.5476438999176025, + "learning_rate": 9.487977451308899e-06, + "loss": 0.5396, + "step": 2164 + }, + { + "epoch": 0.6927261092150171, + "grad_norm": 0.6140022873878479, + "learning_rate": 9.487156423650145e-06, + "loss": 0.5812, + "step": 2165 + }, + { + "epoch": 0.6930460750853242, + "grad_norm": 0.6048492193222046, + "learning_rate": 9.486334773841754e-06, + "loss": 0.5541, + "step": 2166 + }, + { + "epoch": 0.6933660409556314, + "grad_norm": 0.5617876648902893, + "learning_rate": 9.48551250199765e-06, + "loss": 0.5525, + "step": 2167 + }, + { + "epoch": 0.6936860068259386, + "grad_norm": 0.5541356205940247, + "learning_rate": 9.484689608231841e-06, + "loss": 0.5662, + "step": 2168 + }, + { + "epoch": 0.6940059726962458, + "grad_norm": 0.5808799862861633, + "learning_rate": 9.483866092658424e-06, + "loss": 0.5356, + "step": 2169 + }, + { + "epoch": 0.6943259385665529, + "grad_norm": 0.5347186326980591, + "learning_rate": 9.48304195539158e-06, + "loss": 0.5437, + "step": 2170 + }, + { + "epoch": 0.69464590443686, + "grad_norm": 0.5472405552864075, + "learning_rate": 9.482217196545573e-06, + "loss": 0.5391, + "step": 2171 + }, + { + "epoch": 0.6949658703071673, + "grad_norm": 0.6138960123062134, + "learning_rate": 9.481391816234759e-06, + "loss": 0.5718, + "step": 2172 + }, + { + "epoch": 0.6952858361774744, + "grad_norm": 0.5508807301521301, + "learning_rate": 9.48056581457358e-06, + "loss": 0.5382, + "step": 2173 + }, + { + "epoch": 0.6956058020477816, + "grad_norm": 0.49969929456710815, + "learning_rate": 9.47973919167656e-06, + "loss": 0.525, + "step": 2174 + }, + { + "epoch": 0.6959257679180887, + "grad_norm": 0.5662696361541748, + "learning_rate": 9.478911947658312e-06, + "loss": 0.5612, + "step": 2175 + }, + { + "epoch": 0.6962457337883959, + "grad_norm": 0.6077730655670166, + "learning_rate": 9.478084082633535e-06, + "loss": 0.5865, + "step": 2176 + }, + { + "epoch": 0.6965656996587031, + "grad_norm": 0.5491179823875427, + "learning_rate": 9.477255596717012e-06, + "loss": 0.5485, + "step": 2177 + }, + { + "epoch": 0.6968856655290102, + "grad_norm": 0.571593165397644, + "learning_rate": 9.476426490023614e-06, + "loss": 0.5466, + "step": 2178 + }, + { + "epoch": 0.6972056313993175, + "grad_norm": 0.5554285049438477, + "learning_rate": 9.475596762668298e-06, + "loss": 0.527, + "step": 2179 + }, + { + "epoch": 0.6975255972696246, + "grad_norm": 0.5728554129600525, + "learning_rate": 9.474766414766108e-06, + "loss": 0.5425, + "step": 2180 + }, + { + "epoch": 0.6978455631399317, + "grad_norm": 0.5726357698440552, + "learning_rate": 9.473935446432169e-06, + "loss": 0.5428, + "step": 2181 + }, + { + "epoch": 0.6981655290102389, + "grad_norm": 0.5684338808059692, + "learning_rate": 9.4731038577817e-06, + "loss": 0.5561, + "step": 2182 + }, + { + "epoch": 0.6984854948805461, + "grad_norm": 0.632646381855011, + "learning_rate": 9.472271648929998e-06, + "loss": 0.5511, + "step": 2183 + }, + { + "epoch": 0.6988054607508533, + "grad_norm": 0.6783565878868103, + "learning_rate": 9.471438819992453e-06, + "loss": 0.5756, + "step": 2184 + }, + { + "epoch": 0.6991254266211604, + "grad_norm": 0.6838875412940979, + "learning_rate": 9.470605371084535e-06, + "loss": 0.5466, + "step": 2185 + }, + { + "epoch": 0.6994453924914675, + "grad_norm": 0.6280619502067566, + "learning_rate": 9.469771302321806e-06, + "loss": 0.5683, + "step": 2186 + }, + { + "epoch": 0.6997653583617748, + "grad_norm": 0.5844177007675171, + "learning_rate": 9.468936613819908e-06, + "loss": 0.5491, + "step": 2187 + }, + { + "epoch": 0.7000853242320819, + "grad_norm": 0.6486324667930603, + "learning_rate": 9.468101305694571e-06, + "loss": 0.5942, + "step": 2188 + }, + { + "epoch": 0.700405290102389, + "grad_norm": 0.723613440990448, + "learning_rate": 9.467265378061614e-06, + "loss": 0.5893, + "step": 2189 + }, + { + "epoch": 0.7007252559726962, + "grad_norm": 0.569793701171875, + "learning_rate": 9.466428831036938e-06, + "loss": 0.5499, + "step": 2190 + }, + { + "epoch": 0.7010452218430034, + "grad_norm": 0.6333669424057007, + "learning_rate": 9.46559166473653e-06, + "loss": 0.5642, + "step": 2191 + }, + { + "epoch": 0.7013651877133106, + "grad_norm": 0.5689230561256409, + "learning_rate": 9.464753879276467e-06, + "loss": 0.5523, + "step": 2192 + }, + { + "epoch": 0.7016851535836177, + "grad_norm": 0.612147867679596, + "learning_rate": 9.463915474772908e-06, + "loss": 0.5986, + "step": 2193 + }, + { + "epoch": 0.702005119453925, + "grad_norm": 0.6131147742271423, + "learning_rate": 9.463076451342096e-06, + "loss": 0.5693, + "step": 2194 + }, + { + "epoch": 0.7023250853242321, + "grad_norm": 0.583795964717865, + "learning_rate": 9.462236809100366e-06, + "loss": 0.5704, + "step": 2195 + }, + { + "epoch": 0.7026450511945392, + "grad_norm": 0.4957069754600525, + "learning_rate": 9.461396548164134e-06, + "loss": 0.5479, + "step": 2196 + }, + { + "epoch": 0.7029650170648464, + "grad_norm": 0.581490159034729, + "learning_rate": 9.460555668649902e-06, + "loss": 0.5683, + "step": 2197 + }, + { + "epoch": 0.7032849829351536, + "grad_norm": 0.5183387398719788, + "learning_rate": 9.459714170674264e-06, + "loss": 0.5238, + "step": 2198 + }, + { + "epoch": 0.7036049488054608, + "grad_norm": 0.559424102306366, + "learning_rate": 9.458872054353888e-06, + "loss": 0.561, + "step": 2199 + }, + { + "epoch": 0.7039249146757679, + "grad_norm": 0.5429365634918213, + "learning_rate": 9.458029319805539e-06, + "loss": 0.5593, + "step": 2200 + }, + { + "epoch": 0.704244880546075, + "grad_norm": 0.5553236603736877, + "learning_rate": 9.45718596714606e-06, + "loss": 0.5499, + "step": 2201 + }, + { + "epoch": 0.7045648464163823, + "grad_norm": 0.5307144522666931, + "learning_rate": 9.456341996492387e-06, + "loss": 0.5841, + "step": 2202 + }, + { + "epoch": 0.7048848122866894, + "grad_norm": 0.5576136708259583, + "learning_rate": 9.455497407961533e-06, + "loss": 0.5509, + "step": 2203 + }, + { + "epoch": 0.7052047781569966, + "grad_norm": 0.5416161417961121, + "learning_rate": 9.454652201670603e-06, + "loss": 0.5637, + "step": 2204 + }, + { + "epoch": 0.7055247440273038, + "grad_norm": 0.5229343175888062, + "learning_rate": 9.453806377736788e-06, + "loss": 0.5459, + "step": 2205 + }, + { + "epoch": 0.705844709897611, + "grad_norm": 0.5486166477203369, + "learning_rate": 9.452959936277359e-06, + "loss": 0.4968, + "step": 2206 + }, + { + "epoch": 0.7061646757679181, + "grad_norm": 0.578492283821106, + "learning_rate": 9.452112877409678e-06, + "loss": 0.5824, + "step": 2207 + }, + { + "epoch": 0.7064846416382252, + "grad_norm": 0.5441255569458008, + "learning_rate": 9.451265201251193e-06, + "loss": 0.5638, + "step": 2208 + }, + { + "epoch": 0.7068046075085325, + "grad_norm": 0.5505200028419495, + "learning_rate": 9.450416907919431e-06, + "loss": 0.5713, + "step": 2209 + }, + { + "epoch": 0.7071245733788396, + "grad_norm": 0.5709050893783569, + "learning_rate": 9.44956799753201e-06, + "loss": 0.5245, + "step": 2210 + }, + { + "epoch": 0.7074445392491467, + "grad_norm": 0.5804630517959595, + "learning_rate": 9.448718470206633e-06, + "loss": 0.5786, + "step": 2211 + }, + { + "epoch": 0.7077645051194539, + "grad_norm": 0.5691149830818176, + "learning_rate": 9.44786832606109e-06, + "loss": 0.5291, + "step": 2212 + }, + { + "epoch": 0.7080844709897611, + "grad_norm": 0.6125223636627197, + "learning_rate": 9.447017565213252e-06, + "loss": 0.5594, + "step": 2213 + }, + { + "epoch": 0.7084044368600683, + "grad_norm": 0.6221138834953308, + "learning_rate": 9.446166187781077e-06, + "loss": 0.5753, + "step": 2214 + }, + { + "epoch": 0.7087244027303754, + "grad_norm": 0.5770483016967773, + "learning_rate": 9.445314193882614e-06, + "loss": 0.5283, + "step": 2215 + }, + { + "epoch": 0.7090443686006825, + "grad_norm": 0.5975071787834167, + "learning_rate": 9.444461583635988e-06, + "loss": 0.5723, + "step": 2216 + }, + { + "epoch": 0.7093643344709898, + "grad_norm": 0.5773232579231262, + "learning_rate": 9.443608357159417e-06, + "loss": 0.5518, + "step": 2217 + }, + { + "epoch": 0.7096843003412969, + "grad_norm": 0.6178917288780212, + "learning_rate": 9.442754514571202e-06, + "loss": 0.5543, + "step": 2218 + }, + { + "epoch": 0.7100042662116041, + "grad_norm": 0.6687595248222351, + "learning_rate": 9.441900055989728e-06, + "loss": 0.5567, + "step": 2219 + }, + { + "epoch": 0.7103242320819113, + "grad_norm": 0.5844324231147766, + "learning_rate": 9.441044981533468e-06, + "loss": 0.5789, + "step": 2220 + }, + { + "epoch": 0.7106441979522184, + "grad_norm": 0.5794469714164734, + "learning_rate": 9.44018929132098e-06, + "loss": 0.5428, + "step": 2221 + }, + { + "epoch": 0.7109641638225256, + "grad_norm": 0.6569241881370544, + "learning_rate": 9.439332985470903e-06, + "loss": 0.53, + "step": 2222 + }, + { + "epoch": 0.7112841296928327, + "grad_norm": 0.5866938233375549, + "learning_rate": 9.438476064101968e-06, + "loss": 0.5437, + "step": 2223 + }, + { + "epoch": 0.71160409556314, + "grad_norm": 0.5498455762863159, + "learning_rate": 9.437618527332987e-06, + "loss": 0.5631, + "step": 2224 + }, + { + "epoch": 0.7119240614334471, + "grad_norm": 0.6395000219345093, + "learning_rate": 9.436760375282858e-06, + "loss": 0.5654, + "step": 2225 + }, + { + "epoch": 0.7122440273037542, + "grad_norm": 0.6275232434272766, + "learning_rate": 9.435901608070567e-06, + "loss": 0.5715, + "step": 2226 + }, + { + "epoch": 0.7125639931740614, + "grad_norm": 0.49817797541618347, + "learning_rate": 9.43504222581518e-06, + "loss": 0.5151, + "step": 2227 + }, + { + "epoch": 0.7128839590443686, + "grad_norm": 0.5261150598526001, + "learning_rate": 9.434182228635854e-06, + "loss": 0.5704, + "step": 2228 + }, + { + "epoch": 0.7132039249146758, + "grad_norm": 0.5595937967300415, + "learning_rate": 9.433321616651826e-06, + "loss": 0.5495, + "step": 2229 + }, + { + "epoch": 0.7135238907849829, + "grad_norm": 0.5996530055999756, + "learning_rate": 9.432460389982423e-06, + "loss": 0.5793, + "step": 2230 + }, + { + "epoch": 0.7138438566552902, + "grad_norm": 0.5601897835731506, + "learning_rate": 9.431598548747055e-06, + "loss": 0.5147, + "step": 2231 + }, + { + "epoch": 0.7141638225255973, + "grad_norm": 0.5980433225631714, + "learning_rate": 9.430736093065217e-06, + "loss": 0.5913, + "step": 2232 + }, + { + "epoch": 0.7144837883959044, + "grad_norm": 0.5890451669692993, + "learning_rate": 9.429873023056488e-06, + "loss": 0.5444, + "step": 2233 + }, + { + "epoch": 0.7148037542662116, + "grad_norm": 0.4962788224220276, + "learning_rate": 9.429009338840538e-06, + "loss": 0.5403, + "step": 2234 + }, + { + "epoch": 0.7151237201365188, + "grad_norm": 0.5949501395225525, + "learning_rate": 9.428145040537112e-06, + "loss": 0.5499, + "step": 2235 + }, + { + "epoch": 0.715443686006826, + "grad_norm": 0.7749754786491394, + "learning_rate": 9.427280128266049e-06, + "loss": 0.5708, + "step": 2236 + }, + { + "epoch": 0.7157636518771331, + "grad_norm": 0.5890344977378845, + "learning_rate": 9.426414602147272e-06, + "loss": 0.5684, + "step": 2237 + }, + { + "epoch": 0.7160836177474402, + "grad_norm": 0.5542466640472412, + "learning_rate": 9.425548462300784e-06, + "loss": 0.5601, + "step": 2238 + }, + { + "epoch": 0.7164035836177475, + "grad_norm": 0.5782498121261597, + "learning_rate": 9.424681708846679e-06, + "loss": 0.545, + "step": 2239 + }, + { + "epoch": 0.7167235494880546, + "grad_norm": 0.6824494004249573, + "learning_rate": 9.423814341905131e-06, + "loss": 0.5361, + "step": 2240 + }, + { + "epoch": 0.7170435153583617, + "grad_norm": 0.6495462656021118, + "learning_rate": 9.422946361596403e-06, + "loss": 0.5401, + "step": 2241 + }, + { + "epoch": 0.7173634812286689, + "grad_norm": 0.5456177592277527, + "learning_rate": 9.422077768040841e-06, + "loss": 0.5466, + "step": 2242 + }, + { + "epoch": 0.7176834470989761, + "grad_norm": 0.6046304106712341, + "learning_rate": 9.421208561358876e-06, + "loss": 0.5417, + "step": 2243 + }, + { + "epoch": 0.7180034129692833, + "grad_norm": 0.589859127998352, + "learning_rate": 9.420338741671025e-06, + "loss": 0.5742, + "step": 2244 + }, + { + "epoch": 0.7183233788395904, + "grad_norm": 0.6276467442512512, + "learning_rate": 9.41946830909789e-06, + "loss": 0.5462, + "step": 2245 + }, + { + "epoch": 0.7186433447098977, + "grad_norm": 0.6103612184524536, + "learning_rate": 9.418597263760156e-06, + "loss": 0.6013, + "step": 2246 + }, + { + "epoch": 0.7189633105802048, + "grad_norm": 0.5994572043418884, + "learning_rate": 9.417725605778599e-06, + "loss": 0.5452, + "step": 2247 + }, + { + "epoch": 0.7192832764505119, + "grad_norm": 0.6904470324516296, + "learning_rate": 9.416853335274068e-06, + "loss": 0.5627, + "step": 2248 + }, + { + "epoch": 0.7196032423208191, + "grad_norm": 0.5794340372085571, + "learning_rate": 9.415980452367511e-06, + "loss": 0.5332, + "step": 2249 + }, + { + "epoch": 0.7199232081911263, + "grad_norm": 0.6280128359794617, + "learning_rate": 9.415106957179952e-06, + "loss": 0.5445, + "step": 2250 + }, + { + "epoch": 0.7202431740614335, + "grad_norm": 0.6575672030448914, + "learning_rate": 9.414232849832501e-06, + "loss": 0.5578, + "step": 2251 + }, + { + "epoch": 0.7205631399317406, + "grad_norm": 0.6580616235733032, + "learning_rate": 9.413358130446356e-06, + "loss": 0.5861, + "step": 2252 + }, + { + "epoch": 0.7208831058020477, + "grad_norm": 0.5627638101577759, + "learning_rate": 9.412482799142795e-06, + "loss": 0.5496, + "step": 2253 + }, + { + "epoch": 0.721203071672355, + "grad_norm": 0.49028298258781433, + "learning_rate": 9.411606856043187e-06, + "loss": 0.5239, + "step": 2254 + }, + { + "epoch": 0.7215230375426621, + "grad_norm": 0.6451398730278015, + "learning_rate": 9.41073030126898e-06, + "loss": 0.5666, + "step": 2255 + }, + { + "epoch": 0.7218430034129693, + "grad_norm": 0.6588875651359558, + "learning_rate": 9.40985313494171e-06, + "loss": 0.5748, + "step": 2256 + }, + { + "epoch": 0.7221629692832765, + "grad_norm": 0.5587199926376343, + "learning_rate": 9.408975357182998e-06, + "loss": 0.5774, + "step": 2257 + }, + { + "epoch": 0.7224829351535836, + "grad_norm": 0.6241255402565002, + "learning_rate": 9.40809696811455e-06, + "loss": 0.5683, + "step": 2258 + }, + { + "epoch": 0.7228029010238908, + "grad_norm": 0.6500443816184998, + "learning_rate": 9.407217967858154e-06, + "loss": 0.5878, + "step": 2259 + }, + { + "epoch": 0.7231228668941979, + "grad_norm": 0.6004780530929565, + "learning_rate": 9.406338356535686e-06, + "loss": 0.5402, + "step": 2260 + }, + { + "epoch": 0.7234428327645052, + "grad_norm": 0.5574365258216858, + "learning_rate": 9.405458134269101e-06, + "loss": 0.564, + "step": 2261 + }, + { + "epoch": 0.7237627986348123, + "grad_norm": 0.5756843686103821, + "learning_rate": 9.404577301180447e-06, + "loss": 0.5, + "step": 2262 + }, + { + "epoch": 0.7240827645051194, + "grad_norm": 0.6266509294509888, + "learning_rate": 9.403695857391851e-06, + "loss": 0.5331, + "step": 2263 + }, + { + "epoch": 0.7244027303754266, + "grad_norm": 0.623327910900116, + "learning_rate": 9.402813803025526e-06, + "loss": 0.5532, + "step": 2264 + }, + { + "epoch": 0.7247226962457338, + "grad_norm": 0.5519808530807495, + "learning_rate": 9.401931138203771e-06, + "loss": 0.5492, + "step": 2265 + }, + { + "epoch": 0.725042662116041, + "grad_norm": 0.6137987375259399, + "learning_rate": 9.401047863048969e-06, + "loss": 0.5459, + "step": 2266 + }, + { + "epoch": 0.7253626279863481, + "grad_norm": 0.6169253587722778, + "learning_rate": 9.400163977683586e-06, + "loss": 0.5627, + "step": 2267 + }, + { + "epoch": 0.7256825938566553, + "grad_norm": 0.54496169090271, + "learning_rate": 9.399279482230174e-06, + "loss": 0.5469, + "step": 2268 + }, + { + "epoch": 0.7260025597269625, + "grad_norm": 0.5488664507865906, + "learning_rate": 9.398394376811368e-06, + "loss": 0.5397, + "step": 2269 + }, + { + "epoch": 0.7263225255972696, + "grad_norm": 0.6857829689979553, + "learning_rate": 9.397508661549892e-06, + "loss": 0.5724, + "step": 2270 + }, + { + "epoch": 0.7266424914675768, + "grad_norm": 0.5773136615753174, + "learning_rate": 9.396622336568548e-06, + "loss": 0.5493, + "step": 2271 + }, + { + "epoch": 0.726962457337884, + "grad_norm": 0.6079044938087463, + "learning_rate": 9.395735401990229e-06, + "loss": 0.5908, + "step": 2272 + }, + { + "epoch": 0.7272824232081911, + "grad_norm": 0.6028911471366882, + "learning_rate": 9.394847857937907e-06, + "loss": 0.5518, + "step": 2273 + }, + { + "epoch": 0.7276023890784983, + "grad_norm": 0.5717269778251648, + "learning_rate": 9.393959704534644e-06, + "loss": 0.5436, + "step": 2274 + }, + { + "epoch": 0.7279223549488054, + "grad_norm": 0.5948625206947327, + "learning_rate": 9.393070941903581e-06, + "loss": 0.5364, + "step": 2275 + }, + { + "epoch": 0.7282423208191127, + "grad_norm": 0.5648112893104553, + "learning_rate": 9.39218157016795e-06, + "loss": 0.5549, + "step": 2276 + }, + { + "epoch": 0.7285622866894198, + "grad_norm": 0.5189826488494873, + "learning_rate": 9.391291589451056e-06, + "loss": 0.5415, + "step": 2277 + }, + { + "epoch": 0.7288822525597269, + "grad_norm": 0.5410546660423279, + "learning_rate": 9.390400999876301e-06, + "loss": 0.5813, + "step": 2278 + }, + { + "epoch": 0.7292022184300341, + "grad_norm": 0.5291692614555359, + "learning_rate": 9.389509801567166e-06, + "loss": 0.5725, + "step": 2279 + }, + { + "epoch": 0.7295221843003413, + "grad_norm": 0.5540335178375244, + "learning_rate": 9.388617994647217e-06, + "loss": 0.535, + "step": 2280 + }, + { + "epoch": 0.7298421501706485, + "grad_norm": 0.5406812429428101, + "learning_rate": 9.387725579240103e-06, + "loss": 0.5461, + "step": 2281 + }, + { + "epoch": 0.7301621160409556, + "grad_norm": 0.694333553314209, + "learning_rate": 9.386832555469559e-06, + "loss": 0.5458, + "step": 2282 + }, + { + "epoch": 0.7304820819112628, + "grad_norm": 0.6259264349937439, + "learning_rate": 9.385938923459405e-06, + "loss": 0.5644, + "step": 2283 + }, + { + "epoch": 0.73080204778157, + "grad_norm": 0.5157386064529419, + "learning_rate": 9.385044683333542e-06, + "loss": 0.5576, + "step": 2284 + }, + { + "epoch": 0.7311220136518771, + "grad_norm": 0.6391353011131287, + "learning_rate": 9.384149835215956e-06, + "loss": 0.5555, + "step": 2285 + }, + { + "epoch": 0.7314419795221843, + "grad_norm": 0.5598930716514587, + "learning_rate": 9.383254379230724e-06, + "loss": 0.5507, + "step": 2286 + }, + { + "epoch": 0.7317619453924915, + "grad_norm": 0.5639938712120056, + "learning_rate": 9.382358315501999e-06, + "loss": 0.5908, + "step": 2287 + }, + { + "epoch": 0.7320819112627986, + "grad_norm": 0.5010425448417664, + "learning_rate": 9.38146164415402e-06, + "loss": 0.5409, + "step": 2288 + }, + { + "epoch": 0.7324018771331058, + "grad_norm": 0.5784039497375488, + "learning_rate": 9.380564365311113e-06, + "loss": 0.5346, + "step": 2289 + }, + { + "epoch": 0.7327218430034129, + "grad_norm": 0.5947515964508057, + "learning_rate": 9.379666479097688e-06, + "loss": 0.586, + "step": 2290 + }, + { + "epoch": 0.7330418088737202, + "grad_norm": 0.5534799098968506, + "learning_rate": 9.378767985638236e-06, + "loss": 0.5561, + "step": 2291 + }, + { + "epoch": 0.7333617747440273, + "grad_norm": 0.5672453045845032, + "learning_rate": 9.377868885057335e-06, + "loss": 0.5461, + "step": 2292 + }, + { + "epoch": 0.7336817406143344, + "grad_norm": 0.5631408095359802, + "learning_rate": 9.376969177479647e-06, + "loss": 0.5254, + "step": 2293 + }, + { + "epoch": 0.7340017064846417, + "grad_norm": 0.5236619710922241, + "learning_rate": 9.376068863029916e-06, + "loss": 0.5147, + "step": 2294 + }, + { + "epoch": 0.7343216723549488, + "grad_norm": 0.5404022336006165, + "learning_rate": 9.375167941832974e-06, + "loss": 0.4949, + "step": 2295 + }, + { + "epoch": 0.734641638225256, + "grad_norm": 0.5881881713867188, + "learning_rate": 9.374266414013731e-06, + "loss": 0.5972, + "step": 2296 + }, + { + "epoch": 0.7349616040955631, + "grad_norm": 0.4837624430656433, + "learning_rate": 9.37336427969719e-06, + "loss": 0.5481, + "step": 2297 + }, + { + "epoch": 0.7352815699658704, + "grad_norm": 0.5485632419586182, + "learning_rate": 9.372461539008429e-06, + "loss": 0.5446, + "step": 2298 + }, + { + "epoch": 0.7356015358361775, + "grad_norm": 0.587704598903656, + "learning_rate": 9.371558192072615e-06, + "loss": 0.5284, + "step": 2299 + }, + { + "epoch": 0.7359215017064846, + "grad_norm": 0.5816254615783691, + "learning_rate": 9.370654239014999e-06, + "loss": 0.5594, + "step": 2300 + }, + { + "epoch": 0.7362414675767918, + "grad_norm": 0.6278566122055054, + "learning_rate": 9.369749679960916e-06, + "loss": 0.5864, + "step": 2301 + }, + { + "epoch": 0.736561433447099, + "grad_norm": 0.5978617668151855, + "learning_rate": 9.36884451503578e-06, + "loss": 0.594, + "step": 2302 + }, + { + "epoch": 0.7368813993174061, + "grad_norm": 0.5780118703842163, + "learning_rate": 9.3679387443651e-06, + "loss": 0.5678, + "step": 2303 + }, + { + "epoch": 0.7372013651877133, + "grad_norm": 0.5417023301124573, + "learning_rate": 9.367032368074455e-06, + "loss": 0.5372, + "step": 2304 + }, + { + "epoch": 0.7375213310580204, + "grad_norm": 0.6295386552810669, + "learning_rate": 9.36612538628952e-06, + "loss": 0.5751, + "step": 2305 + }, + { + "epoch": 0.7378412969283277, + "grad_norm": 0.5656825304031372, + "learning_rate": 9.365217799136047e-06, + "loss": 0.556, + "step": 2306 + }, + { + "epoch": 0.7381612627986348, + "grad_norm": 0.5957443714141846, + "learning_rate": 9.364309606739875e-06, + "loss": 0.555, + "step": 2307 + }, + { + "epoch": 0.738481228668942, + "grad_norm": 0.5462073683738708, + "learning_rate": 9.363400809226925e-06, + "loss": 0.521, + "step": 2308 + }, + { + "epoch": 0.7388011945392492, + "grad_norm": 0.5976278185844421, + "learning_rate": 9.362491406723204e-06, + "loss": 0.5796, + "step": 2309 + }, + { + "epoch": 0.7391211604095563, + "grad_norm": 0.5917383432388306, + "learning_rate": 9.361581399354803e-06, + "loss": 0.5599, + "step": 2310 + }, + { + "epoch": 0.7394411262798635, + "grad_norm": 0.5707705020904541, + "learning_rate": 9.360670787247893e-06, + "loss": 0.5708, + "step": 2311 + }, + { + "epoch": 0.7397610921501706, + "grad_norm": 0.5376531481742859, + "learning_rate": 9.359759570528731e-06, + "loss": 0.5616, + "step": 2312 + }, + { + "epoch": 0.7400810580204779, + "grad_norm": 0.5556191205978394, + "learning_rate": 9.35884774932366e-06, + "loss": 0.5563, + "step": 2313 + }, + { + "epoch": 0.740401023890785, + "grad_norm": 0.6098505854606628, + "learning_rate": 9.357935323759104e-06, + "loss": 0.5677, + "step": 2314 + }, + { + "epoch": 0.7407209897610921, + "grad_norm": 0.5319908857345581, + "learning_rate": 9.357022293961574e-06, + "loss": 0.5497, + "step": 2315 + }, + { + "epoch": 0.7410409556313993, + "grad_norm": 0.5752114057540894, + "learning_rate": 9.356108660057662e-06, + "loss": 0.5249, + "step": 2316 + }, + { + "epoch": 0.7413609215017065, + "grad_norm": 0.5302920937538147, + "learning_rate": 9.355194422174042e-06, + "loss": 0.5319, + "step": 2317 + }, + { + "epoch": 0.7416808873720137, + "grad_norm": 0.5611221790313721, + "learning_rate": 9.354279580437476e-06, + "loss": 0.5687, + "step": 2318 + }, + { + "epoch": 0.7420008532423208, + "grad_norm": 0.5666431188583374, + "learning_rate": 9.35336413497481e-06, + "loss": 0.5119, + "step": 2319 + }, + { + "epoch": 0.742320819112628, + "grad_norm": 0.5902916789054871, + "learning_rate": 9.352448085912966e-06, + "loss": 0.5731, + "step": 2320 + }, + { + "epoch": 0.7426407849829352, + "grad_norm": 0.4963330626487732, + "learning_rate": 9.351531433378963e-06, + "loss": 0.5328, + "step": 2321 + }, + { + "epoch": 0.7429607508532423, + "grad_norm": 0.5711694955825806, + "learning_rate": 9.350614177499887e-06, + "loss": 0.5581, + "step": 2322 + }, + { + "epoch": 0.7432807167235495, + "grad_norm": 0.5014269948005676, + "learning_rate": 9.349696318402926e-06, + "loss": 0.5519, + "step": 2323 + }, + { + "epoch": 0.7436006825938567, + "grad_norm": 0.4947138726711273, + "learning_rate": 9.348777856215335e-06, + "loss": 0.5613, + "step": 2324 + }, + { + "epoch": 0.7439206484641638, + "grad_norm": 0.6116201281547546, + "learning_rate": 9.347858791064462e-06, + "loss": 0.5401, + "step": 2325 + }, + { + "epoch": 0.744240614334471, + "grad_norm": 0.5416233539581299, + "learning_rate": 9.346939123077738e-06, + "loss": 0.5463, + "step": 2326 + }, + { + "epoch": 0.7445605802047781, + "grad_norm": 0.5299511551856995, + "learning_rate": 9.346018852382676e-06, + "loss": 0.5404, + "step": 2327 + }, + { + "epoch": 0.7448805460750854, + "grad_norm": 0.5581720471382141, + "learning_rate": 9.34509797910687e-06, + "loss": 0.5422, + "step": 2328 + }, + { + "epoch": 0.7452005119453925, + "grad_norm": 0.6147540211677551, + "learning_rate": 9.344176503378003e-06, + "loss": 0.5996, + "step": 2329 + }, + { + "epoch": 0.7455204778156996, + "grad_norm": 0.5862179398536682, + "learning_rate": 9.343254425323837e-06, + "loss": 0.5432, + "step": 2330 + }, + { + "epoch": 0.7458404436860068, + "grad_norm": 0.6720561981201172, + "learning_rate": 9.34233174507222e-06, + "loss": 0.5424, + "step": 2331 + }, + { + "epoch": 0.746160409556314, + "grad_norm": 0.582324743270874, + "learning_rate": 9.341408462751084e-06, + "loss": 0.5591, + "step": 2332 + }, + { + "epoch": 0.7464803754266212, + "grad_norm": 0.5627602338790894, + "learning_rate": 9.34048457848844e-06, + "loss": 0.545, + "step": 2333 + }, + { + "epoch": 0.7468003412969283, + "grad_norm": 0.66405189037323, + "learning_rate": 9.339560092412387e-06, + "loss": 0.5512, + "step": 2334 + }, + { + "epoch": 0.7471203071672355, + "grad_norm": 0.6157020330429077, + "learning_rate": 9.338635004651108e-06, + "loss": 0.556, + "step": 2335 + }, + { + "epoch": 0.7474402730375427, + "grad_norm": 0.5703638792037964, + "learning_rate": 9.337709315332867e-06, + "loss": 0.5773, + "step": 2336 + }, + { + "epoch": 0.7477602389078498, + "grad_norm": 0.6717652082443237, + "learning_rate": 9.336783024586009e-06, + "loss": 0.5652, + "step": 2337 + }, + { + "epoch": 0.748080204778157, + "grad_norm": 0.6250460147857666, + "learning_rate": 9.335856132538968e-06, + "loss": 0.5611, + "step": 2338 + }, + { + "epoch": 0.7484001706484642, + "grad_norm": 0.5538761019706726, + "learning_rate": 9.33492863932026e-06, + "loss": 0.5491, + "step": 2339 + }, + { + "epoch": 0.7487201365187713, + "grad_norm": 0.5979110598564148, + "learning_rate": 9.334000545058479e-06, + "loss": 0.553, + "step": 2340 + }, + { + "epoch": 0.7490401023890785, + "grad_norm": 0.6450609564781189, + "learning_rate": 9.333071849882308e-06, + "loss": 0.5387, + "step": 2341 + }, + { + "epoch": 0.7493600682593856, + "grad_norm": 0.5491853952407837, + "learning_rate": 9.332142553920513e-06, + "loss": 0.5501, + "step": 2342 + }, + { + "epoch": 0.7496800341296929, + "grad_norm": 0.5262613296508789, + "learning_rate": 9.331212657301942e-06, + "loss": 0.5415, + "step": 2343 + }, + { + "epoch": 0.75, + "grad_norm": 0.6032441258430481, + "learning_rate": 9.330282160155524e-06, + "loss": 0.5409, + "step": 2344 + }, + { + "epoch": 0.7503199658703071, + "grad_norm": 0.5730162858963013, + "learning_rate": 9.329351062610277e-06, + "loss": 0.5634, + "step": 2345 + }, + { + "epoch": 0.7506399317406144, + "grad_norm": 0.673250138759613, + "learning_rate": 9.328419364795295e-06, + "loss": 0.5421, + "step": 2346 + }, + { + "epoch": 0.7509598976109215, + "grad_norm": 0.6810570955276489, + "learning_rate": 9.32748706683976e-06, + "loss": 0.6039, + "step": 2347 + }, + { + "epoch": 0.7512798634812287, + "grad_norm": 0.5421538352966309, + "learning_rate": 9.326554168872939e-06, + "loss": 0.535, + "step": 2348 + }, + { + "epoch": 0.7515998293515358, + "grad_norm": 0.5963729023933411, + "learning_rate": 9.325620671024179e-06, + "loss": 0.5228, + "step": 2349 + }, + { + "epoch": 0.751919795221843, + "grad_norm": 0.6112815141677856, + "learning_rate": 9.324686573422905e-06, + "loss": 0.5389, + "step": 2350 + }, + { + "epoch": 0.7522397610921502, + "grad_norm": 0.558083713054657, + "learning_rate": 9.323751876198638e-06, + "loss": 0.5221, + "step": 2351 + }, + { + "epoch": 0.7525597269624573, + "grad_norm": 0.6266518831253052, + "learning_rate": 9.32281657948097e-06, + "loss": 0.5705, + "step": 2352 + }, + { + "epoch": 0.7528796928327645, + "grad_norm": 0.5977521538734436, + "learning_rate": 9.321880683399585e-06, + "loss": 0.5114, + "step": 2353 + }, + { + "epoch": 0.7531996587030717, + "grad_norm": 0.5047455430030823, + "learning_rate": 9.320944188084241e-06, + "loss": 0.5552, + "step": 2354 + }, + { + "epoch": 0.7535196245733788, + "grad_norm": 0.6396503448486328, + "learning_rate": 9.32000709366479e-06, + "loss": 0.5598, + "step": 2355 + }, + { + "epoch": 0.753839590443686, + "grad_norm": 0.6784676313400269, + "learning_rate": 9.319069400271158e-06, + "loss": 0.6056, + "step": 2356 + }, + { + "epoch": 0.7541595563139932, + "grad_norm": 0.5095100402832031, + "learning_rate": 9.318131108033355e-06, + "loss": 0.5418, + "step": 2357 + }, + { + "epoch": 0.7544795221843004, + "grad_norm": 0.5125514268875122, + "learning_rate": 9.31719221708148e-06, + "loss": 0.532, + "step": 2358 + }, + { + "epoch": 0.7547994880546075, + "grad_norm": 0.6723603010177612, + "learning_rate": 9.316252727545712e-06, + "loss": 0.5809, + "step": 2359 + }, + { + "epoch": 0.7551194539249146, + "grad_norm": 0.6266522407531738, + "learning_rate": 9.31531263955631e-06, + "loss": 0.5725, + "step": 2360 + }, + { + "epoch": 0.7554394197952219, + "grad_norm": 0.7055239677429199, + "learning_rate": 9.314371953243619e-06, + "loss": 0.5895, + "step": 2361 + }, + { + "epoch": 0.755759385665529, + "grad_norm": 0.5571537613868713, + "learning_rate": 9.313430668738067e-06, + "loss": 0.5023, + "step": 2362 + }, + { + "epoch": 0.7560793515358362, + "grad_norm": 0.5370808839797974, + "learning_rate": 9.312488786170162e-06, + "loss": 0.5384, + "step": 2363 + }, + { + "epoch": 0.7563993174061433, + "grad_norm": 0.7225911021232605, + "learning_rate": 9.311546305670499e-06, + "loss": 0.5611, + "step": 2364 + }, + { + "epoch": 0.7567192832764505, + "grad_norm": 0.6097559928894043, + "learning_rate": 9.310603227369754e-06, + "loss": 0.5769, + "step": 2365 + }, + { + "epoch": 0.7570392491467577, + "grad_norm": 0.5765969753265381, + "learning_rate": 9.309659551398687e-06, + "loss": 0.5401, + "step": 2366 + }, + { + "epoch": 0.7573592150170648, + "grad_norm": 0.5765284299850464, + "learning_rate": 9.308715277888138e-06, + "loss": 0.5554, + "step": 2367 + }, + { + "epoch": 0.757679180887372, + "grad_norm": 0.5664956569671631, + "learning_rate": 9.307770406969032e-06, + "loss": 0.5521, + "step": 2368 + }, + { + "epoch": 0.7579991467576792, + "grad_norm": 0.6559858322143555, + "learning_rate": 9.306824938772376e-06, + "loss": 0.5975, + "step": 2369 + }, + { + "epoch": 0.7583191126279863, + "grad_norm": 0.5683647990226746, + "learning_rate": 9.305878873429262e-06, + "loss": 0.5589, + "step": 2370 + }, + { + "epoch": 0.7586390784982935, + "grad_norm": 0.5724915266036987, + "learning_rate": 9.30493221107086e-06, + "loss": 0.5845, + "step": 2371 + }, + { + "epoch": 0.7589590443686007, + "grad_norm": 0.5758845210075378, + "learning_rate": 9.303984951828432e-06, + "loss": 0.5192, + "step": 2372 + }, + { + "epoch": 0.7592790102389079, + "grad_norm": 0.6145362257957458, + "learning_rate": 9.30303709583331e-06, + "loss": 0.5919, + "step": 2373 + }, + { + "epoch": 0.759598976109215, + "grad_norm": 0.5690919756889343, + "learning_rate": 9.302088643216917e-06, + "loss": 0.5514, + "step": 2374 + }, + { + "epoch": 0.7599189419795221, + "grad_norm": 0.5775154232978821, + "learning_rate": 9.301139594110759e-06, + "loss": 0.5562, + "step": 2375 + }, + { + "epoch": 0.7602389078498294, + "grad_norm": 0.5660016536712646, + "learning_rate": 9.300189948646422e-06, + "loss": 0.5245, + "step": 2376 + }, + { + "epoch": 0.7605588737201365, + "grad_norm": 0.5814532041549683, + "learning_rate": 9.299239706955576e-06, + "loss": 0.567, + "step": 2377 + }, + { + "epoch": 0.7608788395904437, + "grad_norm": 0.567104160785675, + "learning_rate": 9.298288869169973e-06, + "loss": 0.5791, + "step": 2378 + }, + { + "epoch": 0.7611988054607508, + "grad_norm": 0.5601080060005188, + "learning_rate": 9.297337435421447e-06, + "loss": 0.5262, + "step": 2379 + }, + { + "epoch": 0.761518771331058, + "grad_norm": 0.6034749150276184, + "learning_rate": 9.296385405841915e-06, + "loss": 0.5797, + "step": 2380 + }, + { + "epoch": 0.7618387372013652, + "grad_norm": 0.5768266916275024, + "learning_rate": 9.295432780563378e-06, + "loss": 0.5407, + "step": 2381 + }, + { + "epoch": 0.7621587030716723, + "grad_norm": 0.641884446144104, + "learning_rate": 9.294479559717918e-06, + "loss": 0.5483, + "step": 2382 + }, + { + "epoch": 0.7624786689419796, + "grad_norm": 0.5506654381752014, + "learning_rate": 9.2935257434377e-06, + "loss": 0.5332, + "step": 2383 + }, + { + "epoch": 0.7627986348122867, + "grad_norm": 0.6289987564086914, + "learning_rate": 9.292571331854975e-06, + "loss": 0.5984, + "step": 2384 + }, + { + "epoch": 0.7631186006825939, + "grad_norm": 0.5896539092063904, + "learning_rate": 9.291616325102068e-06, + "loss": 0.5669, + "step": 2385 + }, + { + "epoch": 0.763438566552901, + "grad_norm": 0.5433265566825867, + "learning_rate": 9.290660723311396e-06, + "loss": 0.5449, + "step": 2386 + }, + { + "epoch": 0.7637585324232082, + "grad_norm": 0.6118497252464294, + "learning_rate": 9.289704526615453e-06, + "loss": 0.5488, + "step": 2387 + }, + { + "epoch": 0.7640784982935154, + "grad_norm": 0.5877235531806946, + "learning_rate": 9.288747735146817e-06, + "loss": 0.5455, + "step": 2388 + }, + { + "epoch": 0.7643984641638225, + "grad_norm": 0.5190421938896179, + "learning_rate": 9.287790349038148e-06, + "loss": 0.5325, + "step": 2389 + }, + { + "epoch": 0.7647184300341296, + "grad_norm": 0.6026919484138489, + "learning_rate": 9.286832368422188e-06, + "loss": 0.5318, + "step": 2390 + }, + { + "epoch": 0.7650383959044369, + "grad_norm": 0.5628033876419067, + "learning_rate": 9.285873793431764e-06, + "loss": 0.5363, + "step": 2391 + }, + { + "epoch": 0.765358361774744, + "grad_norm": 0.6197019219398499, + "learning_rate": 9.284914624199782e-06, + "loss": 0.5844, + "step": 2392 + }, + { + "epoch": 0.7656783276450512, + "grad_norm": 0.5734281539916992, + "learning_rate": 9.28395486085923e-06, + "loss": 0.5388, + "step": 2393 + }, + { + "epoch": 0.7659982935153583, + "grad_norm": 0.5641579627990723, + "learning_rate": 9.282994503543185e-06, + "loss": 0.5716, + "step": 2394 + }, + { + "epoch": 0.7663182593856656, + "grad_norm": 0.5311891436576843, + "learning_rate": 9.282033552384802e-06, + "loss": 0.5338, + "step": 2395 + }, + { + "epoch": 0.7666382252559727, + "grad_norm": 0.5733810067176819, + "learning_rate": 9.281072007517312e-06, + "loss": 0.589, + "step": 2396 + }, + { + "epoch": 0.7669581911262798, + "grad_norm": 0.5777380466461182, + "learning_rate": 9.28010986907404e-06, + "loss": 0.5742, + "step": 2397 + }, + { + "epoch": 0.7672781569965871, + "grad_norm": 0.547640860080719, + "learning_rate": 9.279147137188383e-06, + "loss": 0.5362, + "step": 2398 + }, + { + "epoch": 0.7675981228668942, + "grad_norm": 0.5929672718048096, + "learning_rate": 9.278183811993829e-06, + "loss": 0.5552, + "step": 2399 + }, + { + "epoch": 0.7679180887372014, + "grad_norm": 0.5406671762466431, + "learning_rate": 9.277219893623944e-06, + "loss": 0.5282, + "step": 2400 + }, + { + "epoch": 0.7682380546075085, + "grad_norm": 0.6314170360565186, + "learning_rate": 9.276255382212374e-06, + "loss": 0.5592, + "step": 2401 + }, + { + "epoch": 0.7685580204778157, + "grad_norm": 0.5819299817085266, + "learning_rate": 9.27529027789285e-06, + "loss": 0.55, + "step": 2402 + }, + { + "epoch": 0.7688779863481229, + "grad_norm": 0.5875648856163025, + "learning_rate": 9.274324580799187e-06, + "loss": 0.5232, + "step": 2403 + }, + { + "epoch": 0.76919795221843, + "grad_norm": 0.6291683912277222, + "learning_rate": 9.273358291065278e-06, + "loss": 0.581, + "step": 2404 + }, + { + "epoch": 0.7695179180887372, + "grad_norm": 0.5859588384628296, + "learning_rate": 9.2723914088251e-06, + "loss": 0.5647, + "step": 2405 + }, + { + "epoch": 0.7698378839590444, + "grad_norm": 0.6533288955688477, + "learning_rate": 9.271423934212716e-06, + "loss": 0.5455, + "step": 2406 + }, + { + "epoch": 0.7701578498293515, + "grad_norm": 0.6398971676826477, + "learning_rate": 9.270455867362262e-06, + "loss": 0.5864, + "step": 2407 + }, + { + "epoch": 0.7704778156996587, + "grad_norm": 0.651111900806427, + "learning_rate": 9.269487208407965e-06, + "loss": 0.5817, + "step": 2408 + }, + { + "epoch": 0.7707977815699659, + "grad_norm": 0.6151703000068665, + "learning_rate": 9.268517957484132e-06, + "loss": 0.5491, + "step": 2409 + }, + { + "epoch": 0.7711177474402731, + "grad_norm": 0.5498530864715576, + "learning_rate": 9.26754811472515e-06, + "loss": 0.5109, + "step": 2410 + }, + { + "epoch": 0.7714377133105802, + "grad_norm": 0.5751739144325256, + "learning_rate": 9.266577680265485e-06, + "loss": 0.5357, + "step": 2411 + }, + { + "epoch": 0.7717576791808873, + "grad_norm": 0.5931308269500732, + "learning_rate": 9.265606654239691e-06, + "loss": 0.5096, + "step": 2412 + }, + { + "epoch": 0.7720776450511946, + "grad_norm": 0.5704831480979919, + "learning_rate": 9.264635036782406e-06, + "loss": 0.5619, + "step": 2413 + }, + { + "epoch": 0.7723976109215017, + "grad_norm": 0.5751301646232605, + "learning_rate": 9.263662828028342e-06, + "loss": 0.564, + "step": 2414 + }, + { + "epoch": 0.7727175767918089, + "grad_norm": 0.6460949778556824, + "learning_rate": 9.262690028112296e-06, + "loss": 0.5846, + "step": 2415 + }, + { + "epoch": 0.773037542662116, + "grad_norm": 0.6896305680274963, + "learning_rate": 9.261716637169149e-06, + "loss": 0.5768, + "step": 2416 + }, + { + "epoch": 0.7733575085324232, + "grad_norm": 0.622558057308197, + "learning_rate": 9.260742655333866e-06, + "loss": 0.5444, + "step": 2417 + }, + { + "epoch": 0.7736774744027304, + "grad_norm": 0.5928201079368591, + "learning_rate": 9.259768082741487e-06, + "loss": 0.5549, + "step": 2418 + }, + { + "epoch": 0.7739974402730375, + "grad_norm": 0.601613461971283, + "learning_rate": 9.258792919527138e-06, + "loss": 0.5898, + "step": 2419 + }, + { + "epoch": 0.7743174061433447, + "grad_norm": 0.5642574429512024, + "learning_rate": 9.257817165826027e-06, + "loss": 0.5639, + "step": 2420 + }, + { + "epoch": 0.7746373720136519, + "grad_norm": 0.559229850769043, + "learning_rate": 9.256840821773444e-06, + "loss": 0.5087, + "step": 2421 + }, + { + "epoch": 0.774957337883959, + "grad_norm": 0.5684958100318909, + "learning_rate": 9.25586388750476e-06, + "loss": 0.5474, + "step": 2422 + }, + { + "epoch": 0.7752773037542662, + "grad_norm": 0.5171270370483398, + "learning_rate": 9.254886363155429e-06, + "loss": 0.4506, + "step": 2423 + }, + { + "epoch": 0.7755972696245734, + "grad_norm": 0.5402911305427551, + "learning_rate": 9.253908248860983e-06, + "loss": 0.5472, + "step": 2424 + }, + { + "epoch": 0.7759172354948806, + "grad_norm": 0.5100975632667542, + "learning_rate": 9.252929544757043e-06, + "loss": 0.5254, + "step": 2425 + }, + { + "epoch": 0.7762372013651877, + "grad_norm": 0.500893771648407, + "learning_rate": 9.251950250979303e-06, + "loss": 0.5522, + "step": 2426 + }, + { + "epoch": 0.7765571672354948, + "grad_norm": 0.5404722094535828, + "learning_rate": 9.250970367663548e-06, + "loss": 0.5315, + "step": 2427 + }, + { + "epoch": 0.7768771331058021, + "grad_norm": 0.5347235202789307, + "learning_rate": 9.249989894945635e-06, + "loss": 0.5537, + "step": 2428 + }, + { + "epoch": 0.7771970989761092, + "grad_norm": 0.5374565720558167, + "learning_rate": 9.249008832961512e-06, + "loss": 0.5449, + "step": 2429 + }, + { + "epoch": 0.7775170648464164, + "grad_norm": 0.5726111531257629, + "learning_rate": 9.248027181847202e-06, + "loss": 0.5544, + "step": 2430 + }, + { + "epoch": 0.7778370307167235, + "grad_norm": 0.6014326214790344, + "learning_rate": 9.247044941738812e-06, + "loss": 0.5675, + "step": 2431 + }, + { + "epoch": 0.7781569965870307, + "grad_norm": 0.6092072129249573, + "learning_rate": 9.246062112772532e-06, + "loss": 0.5617, + "step": 2432 + }, + { + "epoch": 0.7784769624573379, + "grad_norm": 0.5175899863243103, + "learning_rate": 9.245078695084632e-06, + "loss": 0.5597, + "step": 2433 + }, + { + "epoch": 0.778796928327645, + "grad_norm": 0.6158414483070374, + "learning_rate": 9.244094688811464e-06, + "loss": 0.5519, + "step": 2434 + }, + { + "epoch": 0.7791168941979523, + "grad_norm": 0.5460450053215027, + "learning_rate": 9.243110094089462e-06, + "loss": 0.5298, + "step": 2435 + }, + { + "epoch": 0.7794368600682594, + "grad_norm": 0.622840404510498, + "learning_rate": 9.242124911055141e-06, + "loss": 0.5494, + "step": 2436 + }, + { + "epoch": 0.7797568259385665, + "grad_norm": 0.6064757108688354, + "learning_rate": 9.241139139845098e-06, + "loss": 0.532, + "step": 2437 + }, + { + "epoch": 0.7800767918088737, + "grad_norm": 0.5535280704498291, + "learning_rate": 9.24015278059601e-06, + "loss": 0.5737, + "step": 2438 + }, + { + "epoch": 0.7803967576791809, + "grad_norm": 0.5520623922348022, + "learning_rate": 9.239165833444637e-06, + "loss": 0.5546, + "step": 2439 + }, + { + "epoch": 0.7807167235494881, + "grad_norm": 0.5182672739028931, + "learning_rate": 9.238178298527825e-06, + "loss": 0.5435, + "step": 2440 + }, + { + "epoch": 0.7810366894197952, + "grad_norm": 0.5491767525672913, + "learning_rate": 9.23719017598249e-06, + "loss": 0.5614, + "step": 2441 + }, + { + "epoch": 0.7813566552901023, + "grad_norm": 0.6395105123519897, + "learning_rate": 9.23620146594564e-06, + "loss": 0.5591, + "step": 2442 + }, + { + "epoch": 0.7816766211604096, + "grad_norm": 0.542748749256134, + "learning_rate": 9.235212168554361e-06, + "loss": 0.5607, + "step": 2443 + }, + { + "epoch": 0.7819965870307167, + "grad_norm": 0.5202548503875732, + "learning_rate": 9.234222283945819e-06, + "loss": 0.5218, + "step": 2444 + }, + { + "epoch": 0.7823165529010239, + "grad_norm": 0.5446274280548096, + "learning_rate": 9.233231812257266e-06, + "loss": 0.5106, + "step": 2445 + }, + { + "epoch": 0.7826365187713311, + "grad_norm": 0.575524091720581, + "learning_rate": 9.232240753626027e-06, + "loss": 0.5425, + "step": 2446 + }, + { + "epoch": 0.7829564846416383, + "grad_norm": 0.5616278052330017, + "learning_rate": 9.231249108189517e-06, + "loss": 0.5324, + "step": 2447 + }, + { + "epoch": 0.7832764505119454, + "grad_norm": 0.6591216325759888, + "learning_rate": 9.230256876085227e-06, + "loss": 0.5237, + "step": 2448 + }, + { + "epoch": 0.7835964163822525, + "grad_norm": 0.5643269419670105, + "learning_rate": 9.229264057450733e-06, + "loss": 0.5073, + "step": 2449 + }, + { + "epoch": 0.7839163822525598, + "grad_norm": 0.6130651831626892, + "learning_rate": 9.22827065242369e-06, + "loss": 0.5591, + "step": 2450 + }, + { + "epoch": 0.7842363481228669, + "grad_norm": 0.6077596545219421, + "learning_rate": 9.227276661141831e-06, + "loss": 0.5416, + "step": 2451 + }, + { + "epoch": 0.784556313993174, + "grad_norm": 0.5985930562019348, + "learning_rate": 9.22628208374298e-06, + "loss": 0.5855, + "step": 2452 + }, + { + "epoch": 0.7848762798634812, + "grad_norm": 0.5189453363418579, + "learning_rate": 9.225286920365034e-06, + "loss": 0.5481, + "step": 2453 + }, + { + "epoch": 0.7851962457337884, + "grad_norm": 0.5798327326774597, + "learning_rate": 9.224291171145973e-06, + "loss": 0.5645, + "step": 2454 + }, + { + "epoch": 0.7855162116040956, + "grad_norm": 0.6887683868408203, + "learning_rate": 9.223294836223859e-06, + "loss": 0.5703, + "step": 2455 + }, + { + "epoch": 0.7858361774744027, + "grad_norm": 0.4903722107410431, + "learning_rate": 9.222297915736835e-06, + "loss": 0.5485, + "step": 2456 + }, + { + "epoch": 0.7861561433447098, + "grad_norm": 0.48105481266975403, + "learning_rate": 9.221300409823127e-06, + "loss": 0.5185, + "step": 2457 + }, + { + "epoch": 0.7864761092150171, + "grad_norm": 0.6211410760879517, + "learning_rate": 9.220302318621036e-06, + "loss": 0.5486, + "step": 2458 + }, + { + "epoch": 0.7867960750853242, + "grad_norm": 0.5718651413917542, + "learning_rate": 9.219303642268953e-06, + "loss": 0.5382, + "step": 2459 + }, + { + "epoch": 0.7871160409556314, + "grad_norm": 0.5989018678665161, + "learning_rate": 9.218304380905343e-06, + "loss": 0.538, + "step": 2460 + }, + { + "epoch": 0.7874360068259386, + "grad_norm": 0.5832222104072571, + "learning_rate": 9.217304534668758e-06, + "loss": 0.5708, + "step": 2461 + }, + { + "epoch": 0.7877559726962458, + "grad_norm": 0.5775060653686523, + "learning_rate": 9.216304103697823e-06, + "loss": 0.5366, + "step": 2462 + }, + { + "epoch": 0.7880759385665529, + "grad_norm": 0.5242322087287903, + "learning_rate": 9.215303088131255e-06, + "loss": 0.5273, + "step": 2463 + }, + { + "epoch": 0.78839590443686, + "grad_norm": 0.5479870438575745, + "learning_rate": 9.21430148810784e-06, + "loss": 0.5312, + "step": 2464 + }, + { + "epoch": 0.7887158703071673, + "grad_norm": 0.6071968078613281, + "learning_rate": 9.213299303766455e-06, + "loss": 0.5651, + "step": 2465 + }, + { + "epoch": 0.7890358361774744, + "grad_norm": 0.5586081743240356, + "learning_rate": 9.212296535246053e-06, + "loss": 0.5379, + "step": 2466 + }, + { + "epoch": 0.7893558020477816, + "grad_norm": 0.6063216924667358, + "learning_rate": 9.211293182685668e-06, + "loss": 0.5698, + "step": 2467 + }, + { + "epoch": 0.7896757679180887, + "grad_norm": 0.5611611008644104, + "learning_rate": 9.210289246224416e-06, + "loss": 0.5793, + "step": 2468 + }, + { + "epoch": 0.7899957337883959, + "grad_norm": 0.557030200958252, + "learning_rate": 9.209284726001497e-06, + "loss": 0.544, + "step": 2469 + }, + { + "epoch": 0.7903156996587031, + "grad_norm": 0.5873385667800903, + "learning_rate": 9.208279622156187e-06, + "loss": 0.6026, + "step": 2470 + }, + { + "epoch": 0.7906356655290102, + "grad_norm": 0.5318636894226074, + "learning_rate": 9.207273934827843e-06, + "loss": 0.5339, + "step": 2471 + }, + { + "epoch": 0.7909556313993175, + "grad_norm": 0.5742864608764648, + "learning_rate": 9.206267664155906e-06, + "loss": 0.57, + "step": 2472 + }, + { + "epoch": 0.7912755972696246, + "grad_norm": 0.577001690864563, + "learning_rate": 9.205260810279898e-06, + "loss": 0.5645, + "step": 2473 + }, + { + "epoch": 0.7915955631399317, + "grad_norm": 0.5809705257415771, + "learning_rate": 9.20425337333942e-06, + "loss": 0.545, + "step": 2474 + }, + { + "epoch": 0.7919155290102389, + "grad_norm": 0.5854853391647339, + "learning_rate": 9.203245353474155e-06, + "loss": 0.613, + "step": 2475 + }, + { + "epoch": 0.7922354948805461, + "grad_norm": 0.5700336694717407, + "learning_rate": 9.202236750823863e-06, + "loss": 0.574, + "step": 2476 + }, + { + "epoch": 0.7925554607508533, + "grad_norm": 0.6192684769630432, + "learning_rate": 9.201227565528391e-06, + "loss": 0.5922, + "step": 2477 + }, + { + "epoch": 0.7928754266211604, + "grad_norm": 0.571351945400238, + "learning_rate": 9.200217797727663e-06, + "loss": 0.5931, + "step": 2478 + }, + { + "epoch": 0.7931953924914675, + "grad_norm": 0.6180151700973511, + "learning_rate": 9.199207447561685e-06, + "loss": 0.5079, + "step": 2479 + }, + { + "epoch": 0.7935153583617748, + "grad_norm": 0.5922881364822388, + "learning_rate": 9.198196515170541e-06, + "loss": 0.5711, + "step": 2480 + }, + { + "epoch": 0.7938353242320819, + "grad_norm": 0.5363336801528931, + "learning_rate": 9.1971850006944e-06, + "loss": 0.5138, + "step": 2481 + }, + { + "epoch": 0.794155290102389, + "grad_norm": 0.6847931742668152, + "learning_rate": 9.19617290427351e-06, + "loss": 0.5606, + "step": 2482 + }, + { + "epoch": 0.7944752559726962, + "grad_norm": 0.5617700815200806, + "learning_rate": 9.195160226048198e-06, + "loss": 0.5244, + "step": 2483 + }, + { + "epoch": 0.7947952218430034, + "grad_norm": 0.5189504027366638, + "learning_rate": 9.194146966158875e-06, + "loss": 0.5361, + "step": 2484 + }, + { + "epoch": 0.7951151877133106, + "grad_norm": 0.5778887271881104, + "learning_rate": 9.193133124746029e-06, + "loss": 0.5416, + "step": 2485 + }, + { + "epoch": 0.7954351535836177, + "grad_norm": 0.5868055820465088, + "learning_rate": 9.19211870195023e-06, + "loss": 0.543, + "step": 2486 + }, + { + "epoch": 0.795755119453925, + "grad_norm": 0.5262547731399536, + "learning_rate": 9.19110369791213e-06, + "loss": 0.5511, + "step": 2487 + }, + { + "epoch": 0.7960750853242321, + "grad_norm": 0.4826159179210663, + "learning_rate": 9.190088112772461e-06, + "loss": 0.5329, + "step": 2488 + }, + { + "epoch": 0.7963950511945392, + "grad_norm": 0.5700838565826416, + "learning_rate": 9.189071946672034e-06, + "loss": 0.5597, + "step": 2489 + }, + { + "epoch": 0.7967150170648464, + "grad_norm": 0.6034910082817078, + "learning_rate": 9.188055199751743e-06, + "loss": 0.5425, + "step": 2490 + }, + { + "epoch": 0.7970349829351536, + "grad_norm": 0.4894741475582123, + "learning_rate": 9.187037872152558e-06, + "loss": 0.5411, + "step": 2491 + }, + { + "epoch": 0.7973549488054608, + "grad_norm": 0.5268144607543945, + "learning_rate": 9.186019964015536e-06, + "loss": 0.5332, + "step": 2492 + }, + { + "epoch": 0.7976749146757679, + "grad_norm": 0.5782467126846313, + "learning_rate": 9.18500147548181e-06, + "loss": 0.5761, + "step": 2493 + }, + { + "epoch": 0.797994880546075, + "grad_norm": 0.6057249903678894, + "learning_rate": 9.183982406692593e-06, + "loss": 0.5691, + "step": 2494 + }, + { + "epoch": 0.7983148464163823, + "grad_norm": 0.5561552047729492, + "learning_rate": 9.182962757789184e-06, + "loss": 0.5671, + "step": 2495 + }, + { + "epoch": 0.7986348122866894, + "grad_norm": 0.6109516024589539, + "learning_rate": 9.181942528912957e-06, + "loss": 0.5615, + "step": 2496 + }, + { + "epoch": 0.7989547781569966, + "grad_norm": 0.6077998280525208, + "learning_rate": 9.180921720205365e-06, + "loss": 0.5502, + "step": 2497 + }, + { + "epoch": 0.7992747440273038, + "grad_norm": 0.6387701630592346, + "learning_rate": 9.179900331807949e-06, + "loss": 0.5795, + "step": 2498 + }, + { + "epoch": 0.799594709897611, + "grad_norm": 0.5575502514839172, + "learning_rate": 9.178878363862323e-06, + "loss": 0.5485, + "step": 2499 + }, + { + "epoch": 0.7999146757679181, + "grad_norm": 0.5736941695213318, + "learning_rate": 9.177855816510184e-06, + "loss": 0.539, + "step": 2500 + }, + { + "epoch": 0.8002346416382252, + "grad_norm": 0.5884807109832764, + "learning_rate": 9.176832689893308e-06, + "loss": 0.5695, + "step": 2501 + }, + { + "epoch": 0.8005546075085325, + "grad_norm": 0.6237932443618774, + "learning_rate": 9.175808984153558e-06, + "loss": 0.5824, + "step": 2502 + }, + { + "epoch": 0.8008745733788396, + "grad_norm": 0.6127126216888428, + "learning_rate": 9.174784699432869e-06, + "loss": 0.5423, + "step": 2503 + }, + { + "epoch": 0.8011945392491467, + "grad_norm": 0.5360780954360962, + "learning_rate": 9.173759835873257e-06, + "loss": 0.5472, + "step": 2504 + }, + { + "epoch": 0.8015145051194539, + "grad_norm": 0.5300256013870239, + "learning_rate": 9.172734393616823e-06, + "loss": 0.5536, + "step": 2505 + }, + { + "epoch": 0.8018344709897611, + "grad_norm": 0.548160970211029, + "learning_rate": 9.171708372805745e-06, + "loss": 0.5125, + "step": 2506 + }, + { + "epoch": 0.8021544368600683, + "grad_norm": 0.5152131915092468, + "learning_rate": 9.170681773582283e-06, + "loss": 0.518, + "step": 2507 + }, + { + "epoch": 0.8024744027303754, + "grad_norm": 0.5524023771286011, + "learning_rate": 9.169654596088775e-06, + "loss": 0.501, + "step": 2508 + }, + { + "epoch": 0.8027943686006825, + "grad_norm": 0.5697095394134521, + "learning_rate": 9.168626840467642e-06, + "loss": 0.5725, + "step": 2509 + }, + { + "epoch": 0.8031143344709898, + "grad_norm": 0.48423081636428833, + "learning_rate": 9.167598506861381e-06, + "loss": 0.5171, + "step": 2510 + }, + { + "epoch": 0.8034343003412969, + "grad_norm": 0.5414108037948608, + "learning_rate": 9.166569595412576e-06, + "loss": 0.5041, + "step": 2511 + }, + { + "epoch": 0.8037542662116041, + "grad_norm": 0.6320025324821472, + "learning_rate": 9.165540106263881e-06, + "loss": 0.547, + "step": 2512 + }, + { + "epoch": 0.8040742320819113, + "grad_norm": 0.6039816737174988, + "learning_rate": 9.16451003955804e-06, + "loss": 0.5548, + "step": 2513 + }, + { + "epoch": 0.8043941979522184, + "grad_norm": 0.6364385485649109, + "learning_rate": 9.163479395437874e-06, + "loss": 0.5447, + "step": 2514 + }, + { + "epoch": 0.8047141638225256, + "grad_norm": 0.5636100769042969, + "learning_rate": 9.16244817404628e-06, + "loss": 0.5541, + "step": 2515 + }, + { + "epoch": 0.8050341296928327, + "grad_norm": 0.5757590532302856, + "learning_rate": 9.161416375526239e-06, + "loss": 0.5308, + "step": 2516 + }, + { + "epoch": 0.80535409556314, + "grad_norm": 0.603026807308197, + "learning_rate": 9.16038400002081e-06, + "loss": 0.5636, + "step": 2517 + }, + { + "epoch": 0.8056740614334471, + "grad_norm": 0.5327634215354919, + "learning_rate": 9.159351047673135e-06, + "loss": 0.5385, + "step": 2518 + }, + { + "epoch": 0.8059940273037542, + "grad_norm": 0.5723426342010498, + "learning_rate": 9.158317518626434e-06, + "loss": 0.5588, + "step": 2519 + }, + { + "epoch": 0.8063139931740614, + "grad_norm": 0.5705613493919373, + "learning_rate": 9.157283413024007e-06, + "loss": 0.5524, + "step": 2520 + }, + { + "epoch": 0.8066339590443686, + "grad_norm": 0.5529534816741943, + "learning_rate": 9.156248731009232e-06, + "loss": 0.5022, + "step": 2521 + }, + { + "epoch": 0.8069539249146758, + "grad_norm": 0.49362483620643616, + "learning_rate": 9.155213472725572e-06, + "loss": 0.4861, + "step": 2522 + }, + { + "epoch": 0.8072738907849829, + "grad_norm": 0.5816895961761475, + "learning_rate": 9.154177638316563e-06, + "loss": 0.5934, + "step": 2523 + }, + { + "epoch": 0.8075938566552902, + "grad_norm": 0.6569143533706665, + "learning_rate": 9.153141227925828e-06, + "loss": 0.5722, + "step": 2524 + }, + { + "epoch": 0.8079138225255973, + "grad_norm": 0.5592034459114075, + "learning_rate": 9.152104241697066e-06, + "loss": 0.602, + "step": 2525 + }, + { + "epoch": 0.8082337883959044, + "grad_norm": 0.5290538668632507, + "learning_rate": 9.151066679774054e-06, + "loss": 0.5583, + "step": 2526 + }, + { + "epoch": 0.8085537542662116, + "grad_norm": 0.6192449927330017, + "learning_rate": 9.150028542300654e-06, + "loss": 0.5751, + "step": 2527 + }, + { + "epoch": 0.8088737201365188, + "grad_norm": 0.6423482894897461, + "learning_rate": 9.148989829420805e-06, + "loss": 0.5344, + "step": 2528 + }, + { + "epoch": 0.809193686006826, + "grad_norm": 0.5743381977081299, + "learning_rate": 9.147950541278523e-06, + "loss": 0.5625, + "step": 2529 + }, + { + "epoch": 0.8095136518771331, + "grad_norm": 0.6718060970306396, + "learning_rate": 9.14691067801791e-06, + "loss": 0.603, + "step": 2530 + }, + { + "epoch": 0.8098336177474402, + "grad_norm": 0.5860727429389954, + "learning_rate": 9.145870239783143e-06, + "loss": 0.5766, + "step": 2531 + }, + { + "epoch": 0.8101535836177475, + "grad_norm": 0.5706322193145752, + "learning_rate": 9.144829226718479e-06, + "loss": 0.5721, + "step": 2532 + }, + { + "epoch": 0.8104735494880546, + "grad_norm": 0.5903095006942749, + "learning_rate": 9.143787638968255e-06, + "loss": 0.5558, + "step": 2533 + }, + { + "epoch": 0.8107935153583617, + "grad_norm": 0.5164182186126709, + "learning_rate": 9.142745476676891e-06, + "loss": 0.5865, + "step": 2534 + }, + { + "epoch": 0.8111134812286689, + "grad_norm": 0.5815507173538208, + "learning_rate": 9.141702739988882e-06, + "loss": 0.5819, + "step": 2535 + }, + { + "epoch": 0.8114334470989761, + "grad_norm": 0.5640064477920532, + "learning_rate": 9.140659429048807e-06, + "loss": 0.5121, + "step": 2536 + }, + { + "epoch": 0.8117534129692833, + "grad_norm": 0.5752261281013489, + "learning_rate": 9.139615544001319e-06, + "loss": 0.5497, + "step": 2537 + }, + { + "epoch": 0.8120733788395904, + "grad_norm": 0.555300235748291, + "learning_rate": 9.138571084991157e-06, + "loss": 0.5443, + "step": 2538 + }, + { + "epoch": 0.8123933447098977, + "grad_norm": 0.6510416865348816, + "learning_rate": 9.137526052163135e-06, + "loss": 0.5879, + "step": 2539 + }, + { + "epoch": 0.8127133105802048, + "grad_norm": 0.5922650694847107, + "learning_rate": 9.136480445662147e-06, + "loss": 0.4972, + "step": 2540 + }, + { + "epoch": 0.8130332764505119, + "grad_norm": 0.5528703927993774, + "learning_rate": 9.13543426563317e-06, + "loss": 0.5555, + "step": 2541 + }, + { + "epoch": 0.8133532423208191, + "grad_norm": 0.5601704716682434, + "learning_rate": 9.134387512221258e-06, + "loss": 0.5401, + "step": 2542 + }, + { + "epoch": 0.8136732081911263, + "grad_norm": 0.5348582863807678, + "learning_rate": 9.133340185571541e-06, + "loss": 0.5659, + "step": 2543 + }, + { + "epoch": 0.8139931740614335, + "grad_norm": 0.578411877155304, + "learning_rate": 9.132292285829237e-06, + "loss": 0.583, + "step": 2544 + }, + { + "epoch": 0.8143131399317406, + "grad_norm": 0.5427941083908081, + "learning_rate": 9.131243813139636e-06, + "loss": 0.5231, + "step": 2545 + }, + { + "epoch": 0.8146331058020477, + "grad_norm": 0.5223431587219238, + "learning_rate": 9.130194767648111e-06, + "loss": 0.5417, + "step": 2546 + }, + { + "epoch": 0.814953071672355, + "grad_norm": 0.49521204829216003, + "learning_rate": 9.129145149500111e-06, + "loss": 0.5165, + "step": 2547 + }, + { + "epoch": 0.8152730375426621, + "grad_norm": 0.6073542237281799, + "learning_rate": 9.12809495884117e-06, + "loss": 0.5444, + "step": 2548 + }, + { + "epoch": 0.8155930034129693, + "grad_norm": 0.5415356159210205, + "learning_rate": 9.127044195816895e-06, + "loss": 0.5444, + "step": 2549 + }, + { + "epoch": 0.8159129692832765, + "grad_norm": 0.5234550833702087, + "learning_rate": 9.125992860572979e-06, + "loss": 0.5263, + "step": 2550 + }, + { + "epoch": 0.8162329351535836, + "grad_norm": 0.5407820343971252, + "learning_rate": 9.124940953255188e-06, + "loss": 0.5136, + "step": 2551 + }, + { + "epoch": 0.8165529010238908, + "grad_norm": 0.5394188761711121, + "learning_rate": 9.123888474009373e-06, + "loss": 0.5811, + "step": 2552 + }, + { + "epoch": 0.8168728668941979, + "grad_norm": 0.5949886441230774, + "learning_rate": 9.122835422981458e-06, + "loss": 0.6073, + "step": 2553 + }, + { + "epoch": 0.8171928327645052, + "grad_norm": 0.5322878956794739, + "learning_rate": 9.121781800317453e-06, + "loss": 0.5135, + "step": 2554 + }, + { + "epoch": 0.8175127986348123, + "grad_norm": 0.5521470308303833, + "learning_rate": 9.120727606163443e-06, + "loss": 0.5549, + "step": 2555 + }, + { + "epoch": 0.8178327645051194, + "grad_norm": 0.5179667472839355, + "learning_rate": 9.119672840665592e-06, + "loss": 0.5182, + "step": 2556 + }, + { + "epoch": 0.8181527303754266, + "grad_norm": 0.5399349331855774, + "learning_rate": 9.118617503970146e-06, + "loss": 0.5662, + "step": 2557 + }, + { + "epoch": 0.8184726962457338, + "grad_norm": 0.5927430391311646, + "learning_rate": 9.117561596223428e-06, + "loss": 0.5615, + "step": 2558 + }, + { + "epoch": 0.818792662116041, + "grad_norm": 0.5932854413986206, + "learning_rate": 9.116505117571841e-06, + "loss": 0.5695, + "step": 2559 + }, + { + "epoch": 0.8191126279863481, + "grad_norm": 0.5340254306793213, + "learning_rate": 9.115448068161867e-06, + "loss": 0.5032, + "step": 2560 + }, + { + "epoch": 0.8194325938566553, + "grad_norm": 0.5510215759277344, + "learning_rate": 9.114390448140068e-06, + "loss": 0.5229, + "step": 2561 + }, + { + "epoch": 0.8197525597269625, + "grad_norm": 0.5657440423965454, + "learning_rate": 9.113332257653083e-06, + "loss": 0.5447, + "step": 2562 + }, + { + "epoch": 0.8200725255972696, + "grad_norm": 0.5934768319129944, + "learning_rate": 9.112273496847633e-06, + "loss": 0.542, + "step": 2563 + }, + { + "epoch": 0.8203924914675768, + "grad_norm": 0.5292602777481079, + "learning_rate": 9.111214165870515e-06, + "loss": 0.5284, + "step": 2564 + }, + { + "epoch": 0.820712457337884, + "grad_norm": 0.6175116896629333, + "learning_rate": 9.110154264868607e-06, + "loss": 0.5362, + "step": 2565 + }, + { + "epoch": 0.8210324232081911, + "grad_norm": 0.645007312297821, + "learning_rate": 9.109093793988866e-06, + "loss": 0.5563, + "step": 2566 + }, + { + "epoch": 0.8213523890784983, + "grad_norm": 0.5482667088508606, + "learning_rate": 9.108032753378326e-06, + "loss": 0.5697, + "step": 2567 + }, + { + "epoch": 0.8216723549488054, + "grad_norm": 0.5520942807197571, + "learning_rate": 9.106971143184103e-06, + "loss": 0.538, + "step": 2568 + }, + { + "epoch": 0.8219923208191127, + "grad_norm": 0.6108903884887695, + "learning_rate": 9.10590896355339e-06, + "loss": 0.5506, + "step": 2569 + }, + { + "epoch": 0.8223122866894198, + "grad_norm": 0.6232043504714966, + "learning_rate": 9.10484621463346e-06, + "loss": 0.5474, + "step": 2570 + }, + { + "epoch": 0.8226322525597269, + "grad_norm": 0.6807348728179932, + "learning_rate": 9.103782896571664e-06, + "loss": 0.5816, + "step": 2571 + }, + { + "epoch": 0.8229522184300341, + "grad_norm": 0.5627373456954956, + "learning_rate": 9.102719009515433e-06, + "loss": 0.5305, + "step": 2572 + }, + { + "epoch": 0.8232721843003413, + "grad_norm": 0.5594233870506287, + "learning_rate": 9.101654553612276e-06, + "loss": 0.5761, + "step": 2573 + }, + { + "epoch": 0.8235921501706485, + "grad_norm": 0.5685380101203918, + "learning_rate": 9.100589529009781e-06, + "loss": 0.5471, + "step": 2574 + }, + { + "epoch": 0.8239121160409556, + "grad_norm": 0.5186659693717957, + "learning_rate": 9.099523935855612e-06, + "loss": 0.5139, + "step": 2575 + }, + { + "epoch": 0.8242320819112628, + "grad_norm": 0.5025923848152161, + "learning_rate": 9.09845777429752e-06, + "loss": 0.549, + "step": 2576 + }, + { + "epoch": 0.82455204778157, + "grad_norm": 0.560330331325531, + "learning_rate": 9.097391044483325e-06, + "loss": 0.5302, + "step": 2577 + }, + { + "epoch": 0.8248720136518771, + "grad_norm": 0.5975551009178162, + "learning_rate": 9.096323746560934e-06, + "loss": 0.5809, + "step": 2578 + }, + { + "epoch": 0.8251919795221843, + "grad_norm": 0.6584590077400208, + "learning_rate": 9.095255880678327e-06, + "loss": 0.566, + "step": 2579 + }, + { + "epoch": 0.8255119453924915, + "grad_norm": 0.5146849751472473, + "learning_rate": 9.094187446983567e-06, + "loss": 0.5378, + "step": 2580 + }, + { + "epoch": 0.8258319112627986, + "grad_norm": 0.5368896126747131, + "learning_rate": 9.09311844562479e-06, + "loss": 0.521, + "step": 2581 + }, + { + "epoch": 0.8261518771331058, + "grad_norm": 0.5094373822212219, + "learning_rate": 9.092048876750217e-06, + "loss": 0.5257, + "step": 2582 + }, + { + "epoch": 0.8264718430034129, + "grad_norm": 0.5936737060546875, + "learning_rate": 9.090978740508144e-06, + "loss": 0.5851, + "step": 2583 + }, + { + "epoch": 0.8267918088737202, + "grad_norm": 0.5526201128959656, + "learning_rate": 9.089908037046947e-06, + "loss": 0.5503, + "step": 2584 + }, + { + "epoch": 0.8271117747440273, + "grad_norm": 0.49123311042785645, + "learning_rate": 9.088836766515081e-06, + "loss": 0.5558, + "step": 2585 + }, + { + "epoch": 0.8274317406143344, + "grad_norm": 0.5265716910362244, + "learning_rate": 9.087764929061078e-06, + "loss": 0.5121, + "step": 2586 + }, + { + "epoch": 0.8277517064846417, + "grad_norm": 0.557817280292511, + "learning_rate": 9.086692524833551e-06, + "loss": 0.5553, + "step": 2587 + }, + { + "epoch": 0.8280716723549488, + "grad_norm": 0.5713567733764648, + "learning_rate": 9.085619553981186e-06, + "loss": 0.5635, + "step": 2588 + }, + { + "epoch": 0.828391638225256, + "grad_norm": 0.522912323474884, + "learning_rate": 9.084546016652758e-06, + "loss": 0.5343, + "step": 2589 + }, + { + "epoch": 0.8287116040955631, + "grad_norm": 0.5856422781944275, + "learning_rate": 9.08347191299711e-06, + "loss": 0.5829, + "step": 2590 + }, + { + "epoch": 0.8290315699658704, + "grad_norm": 0.49362123012542725, + "learning_rate": 9.082397243163167e-06, + "loss": 0.5586, + "step": 2591 + }, + { + "epoch": 0.8293515358361775, + "grad_norm": 0.5722553133964539, + "learning_rate": 9.081322007299937e-06, + "loss": 0.5474, + "step": 2592 + }, + { + "epoch": 0.8296715017064846, + "grad_norm": 0.5332806706428528, + "learning_rate": 9.0802462055565e-06, + "loss": 0.5405, + "step": 2593 + }, + { + "epoch": 0.8299914675767918, + "grad_norm": 0.6191956400871277, + "learning_rate": 9.079169838082018e-06, + "loss": 0.5523, + "step": 2594 + }, + { + "epoch": 0.830311433447099, + "grad_norm": 0.5539155602455139, + "learning_rate": 9.07809290502573e-06, + "loss": 0.5548, + "step": 2595 + }, + { + "epoch": 0.8306313993174061, + "grad_norm": 0.5115573406219482, + "learning_rate": 9.077015406536957e-06, + "loss": 0.5252, + "step": 2596 + }, + { + "epoch": 0.8309513651877133, + "grad_norm": 0.58519446849823, + "learning_rate": 9.075937342765091e-06, + "loss": 0.5527, + "step": 2597 + }, + { + "epoch": 0.8312713310580204, + "grad_norm": 0.5422502756118774, + "learning_rate": 9.074858713859609e-06, + "loss": 0.5694, + "step": 2598 + }, + { + "epoch": 0.8315912969283277, + "grad_norm": 0.6382519006729126, + "learning_rate": 9.073779519970066e-06, + "loss": 0.5818, + "step": 2599 + }, + { + "epoch": 0.8319112627986348, + "grad_norm": 0.5833666324615479, + "learning_rate": 9.072699761246088e-06, + "loss": 0.5681, + "step": 2600 + }, + { + "epoch": 0.832231228668942, + "grad_norm": 0.4834451675415039, + "learning_rate": 9.071619437837392e-06, + "loss": 0.5375, + "step": 2601 + }, + { + "epoch": 0.8325511945392492, + "grad_norm": 0.621788740158081, + "learning_rate": 9.070538549893762e-06, + "loss": 0.5985, + "step": 2602 + }, + { + "epoch": 0.8328711604095563, + "grad_norm": 0.5683957934379578, + "learning_rate": 9.069457097565064e-06, + "loss": 0.5245, + "step": 2603 + }, + { + "epoch": 0.8331911262798635, + "grad_norm": 0.5451465249061584, + "learning_rate": 9.068375081001243e-06, + "loss": 0.5411, + "step": 2604 + }, + { + "epoch": 0.8335110921501706, + "grad_norm": 0.5952214002609253, + "learning_rate": 9.067292500352325e-06, + "loss": 0.5121, + "step": 2605 + }, + { + "epoch": 0.8338310580204779, + "grad_norm": 0.5726862549781799, + "learning_rate": 9.066209355768408e-06, + "loss": 0.5588, + "step": 2606 + }, + { + "epoch": 0.834151023890785, + "grad_norm": 0.5016576051712036, + "learning_rate": 9.065125647399671e-06, + "loss": 0.5518, + "step": 2607 + }, + { + "epoch": 0.8344709897610921, + "grad_norm": 0.5270254611968994, + "learning_rate": 9.064041375396373e-06, + "loss": 0.5436, + "step": 2608 + }, + { + "epoch": 0.8347909556313993, + "grad_norm": 0.627449095249176, + "learning_rate": 9.062956539908849e-06, + "loss": 0.5619, + "step": 2609 + }, + { + "epoch": 0.8351109215017065, + "grad_norm": 0.6327446103096008, + "learning_rate": 9.061871141087514e-06, + "loss": 0.5936, + "step": 2610 + }, + { + "epoch": 0.8354308873720137, + "grad_norm": 0.5677685141563416, + "learning_rate": 9.060785179082859e-06, + "loss": 0.5432, + "step": 2611 + }, + { + "epoch": 0.8357508532423208, + "grad_norm": 0.6317307949066162, + "learning_rate": 9.059698654045452e-06, + "loss": 0.5711, + "step": 2612 + }, + { + "epoch": 0.836070819112628, + "grad_norm": 0.6746708750724792, + "learning_rate": 9.058611566125943e-06, + "loss": 0.5124, + "step": 2613 + }, + { + "epoch": 0.8363907849829352, + "grad_norm": 0.5662692189216614, + "learning_rate": 9.057523915475058e-06, + "loss": 0.522, + "step": 2614 + }, + { + "epoch": 0.8367107508532423, + "grad_norm": 0.6253296732902527, + "learning_rate": 9.056435702243601e-06, + "loss": 0.5421, + "step": 2615 + }, + { + "epoch": 0.8370307167235495, + "grad_norm": 0.6256198287010193, + "learning_rate": 9.055346926582453e-06, + "loss": 0.5609, + "step": 2616 + }, + { + "epoch": 0.8373506825938567, + "grad_norm": 0.616001307964325, + "learning_rate": 9.054257588642577e-06, + "loss": 0.5958, + "step": 2617 + }, + { + "epoch": 0.8376706484641638, + "grad_norm": 0.576518714427948, + "learning_rate": 9.053167688575007e-06, + "loss": 0.5546, + "step": 2618 + }, + { + "epoch": 0.837990614334471, + "grad_norm": 0.5338309407234192, + "learning_rate": 9.052077226530862e-06, + "loss": 0.5613, + "step": 2619 + }, + { + "epoch": 0.8383105802047781, + "grad_norm": 0.5424176454544067, + "learning_rate": 9.050986202661334e-06, + "loss": 0.5502, + "step": 2620 + }, + { + "epoch": 0.8386305460750854, + "grad_norm": 0.5512865781784058, + "learning_rate": 9.049894617117697e-06, + "loss": 0.4723, + "step": 2621 + }, + { + "epoch": 0.8389505119453925, + "grad_norm": 0.5608572363853455, + "learning_rate": 9.048802470051296e-06, + "loss": 0.5575, + "step": 2622 + }, + { + "epoch": 0.8392704778156996, + "grad_norm": 0.5869863033294678, + "learning_rate": 9.047709761613565e-06, + "loss": 0.5445, + "step": 2623 + }, + { + "epoch": 0.8395904436860068, + "grad_norm": 0.5191389918327332, + "learning_rate": 9.046616491956003e-06, + "loss": 0.5522, + "step": 2624 + }, + { + "epoch": 0.839910409556314, + "grad_norm": 0.5765413045883179, + "learning_rate": 9.045522661230199e-06, + "loss": 0.5665, + "step": 2625 + }, + { + "epoch": 0.8402303754266212, + "grad_norm": 0.6026390194892883, + "learning_rate": 9.044428269587811e-06, + "loss": 0.6141, + "step": 2626 + }, + { + "epoch": 0.8405503412969283, + "grad_norm": 0.6081013083457947, + "learning_rate": 9.043333317180576e-06, + "loss": 0.5959, + "step": 2627 + }, + { + "epoch": 0.8408703071672355, + "grad_norm": 0.5654779076576233, + "learning_rate": 9.042237804160313e-06, + "loss": 0.5416, + "step": 2628 + }, + { + "epoch": 0.8411902730375427, + "grad_norm": 0.5690180659294128, + "learning_rate": 9.041141730678916e-06, + "loss": 0.5304, + "step": 2629 + }, + { + "epoch": 0.8415102389078498, + "grad_norm": 0.5598645210266113, + "learning_rate": 9.040045096888356e-06, + "loss": 0.5438, + "step": 2630 + }, + { + "epoch": 0.841830204778157, + "grad_norm": 0.5394770503044128, + "learning_rate": 9.038947902940684e-06, + "loss": 0.5373, + "step": 2631 + }, + { + "epoch": 0.8421501706484642, + "grad_norm": 0.6138805150985718, + "learning_rate": 9.037850148988025e-06, + "loss": 0.5637, + "step": 2632 + }, + { + "epoch": 0.8424701365187713, + "grad_norm": 0.5667594075202942, + "learning_rate": 9.036751835182586e-06, + "loss": 0.513, + "step": 2633 + }, + { + "epoch": 0.8427901023890785, + "grad_norm": 0.5401478409767151, + "learning_rate": 9.035652961676649e-06, + "loss": 0.5349, + "step": 2634 + }, + { + "epoch": 0.8431100682593856, + "grad_norm": 0.5945031642913818, + "learning_rate": 9.034553528622572e-06, + "loss": 0.5904, + "step": 2635 + }, + { + "epoch": 0.8434300341296929, + "grad_norm": 0.5829552412033081, + "learning_rate": 9.033453536172796e-06, + "loss": 0.5686, + "step": 2636 + }, + { + "epoch": 0.84375, + "grad_norm": 0.5223750472068787, + "learning_rate": 9.032352984479836e-06, + "loss": 0.5471, + "step": 2637 + }, + { + "epoch": 0.8440699658703071, + "grad_norm": 0.5635274052619934, + "learning_rate": 9.031251873696281e-06, + "loss": 0.6, + "step": 2638 + }, + { + "epoch": 0.8443899317406144, + "grad_norm": 0.4674994647502899, + "learning_rate": 9.030150203974806e-06, + "loss": 0.5353, + "step": 2639 + }, + { + "epoch": 0.8447098976109215, + "grad_norm": 0.5379506945610046, + "learning_rate": 9.029047975468156e-06, + "loss": 0.5438, + "step": 2640 + }, + { + "epoch": 0.8450298634812287, + "grad_norm": 0.5450134873390198, + "learning_rate": 9.027945188329157e-06, + "loss": 0.5392, + "step": 2641 + }, + { + "epoch": 0.8453498293515358, + "grad_norm": 0.507026731967926, + "learning_rate": 9.026841842710711e-06, + "loss": 0.5031, + "step": 2642 + }, + { + "epoch": 0.845669795221843, + "grad_norm": 0.5834566950798035, + "learning_rate": 9.025737938765803e-06, + "loss": 0.5483, + "step": 2643 + }, + { + "epoch": 0.8459897610921502, + "grad_norm": 0.5272031426429749, + "learning_rate": 9.024633476647483e-06, + "loss": 0.5276, + "step": 2644 + }, + { + "epoch": 0.8463097269624573, + "grad_norm": 0.4700632691383362, + "learning_rate": 9.02352845650889e-06, + "loss": 0.4892, + "step": 2645 + }, + { + "epoch": 0.8466296928327645, + "grad_norm": 0.5325533747673035, + "learning_rate": 9.022422878503237e-06, + "loss": 0.5452, + "step": 2646 + }, + { + "epoch": 0.8469496587030717, + "grad_norm": 0.6028921008110046, + "learning_rate": 9.021316742783813e-06, + "loss": 0.562, + "step": 2647 + }, + { + "epoch": 0.8472696245733788, + "grad_norm": 0.5505808591842651, + "learning_rate": 9.020210049503985e-06, + "loss": 0.5695, + "step": 2648 + }, + { + "epoch": 0.847589590443686, + "grad_norm": 0.5052388906478882, + "learning_rate": 9.019102798817196e-06, + "loss": 0.5528, + "step": 2649 + }, + { + "epoch": 0.8479095563139932, + "grad_norm": 0.5818324685096741, + "learning_rate": 9.017994990876971e-06, + "loss": 0.5459, + "step": 2650 + }, + { + "epoch": 0.8482295221843004, + "grad_norm": 0.6578103303909302, + "learning_rate": 9.016886625836905e-06, + "loss": 0.5379, + "step": 2651 + }, + { + "epoch": 0.8485494880546075, + "grad_norm": 0.5725463628768921, + "learning_rate": 9.015777703850675e-06, + "loss": 0.5805, + "step": 2652 + }, + { + "epoch": 0.8488694539249146, + "grad_norm": 0.5812577605247498, + "learning_rate": 9.014668225072036e-06, + "loss": 0.5674, + "step": 2653 + }, + { + "epoch": 0.8491894197952219, + "grad_norm": 0.5736384987831116, + "learning_rate": 9.013558189654819e-06, + "loss": 0.5737, + "step": 2654 + }, + { + "epoch": 0.849509385665529, + "grad_norm": 0.5640150904655457, + "learning_rate": 9.01244759775293e-06, + "loss": 0.5291, + "step": 2655 + }, + { + "epoch": 0.8498293515358362, + "grad_norm": 0.5893073081970215, + "learning_rate": 9.011336449520354e-06, + "loss": 0.5637, + "step": 2656 + }, + { + "epoch": 0.8501493174061433, + "grad_norm": 0.566746175289154, + "learning_rate": 9.010224745111153e-06, + "loss": 0.5164, + "step": 2657 + }, + { + "epoch": 0.8504692832764505, + "grad_norm": 0.6648749113082886, + "learning_rate": 9.009112484679468e-06, + "loss": 0.5534, + "step": 2658 + }, + { + "epoch": 0.8507892491467577, + "grad_norm": 0.5884232521057129, + "learning_rate": 9.007999668379514e-06, + "loss": 0.5423, + "step": 2659 + }, + { + "epoch": 0.8511092150170648, + "grad_norm": 0.6019272208213806, + "learning_rate": 9.006886296365585e-06, + "loss": 0.5717, + "step": 2660 + }, + { + "epoch": 0.851429180887372, + "grad_norm": 0.5348987579345703, + "learning_rate": 9.005772368792048e-06, + "loss": 0.5093, + "step": 2661 + }, + { + "epoch": 0.8517491467576792, + "grad_norm": 0.6644459962844849, + "learning_rate": 9.004657885813354e-06, + "loss": 0.54, + "step": 2662 + }, + { + "epoch": 0.8520691126279863, + "grad_norm": 0.5964223146438599, + "learning_rate": 9.003542847584026e-06, + "loss": 0.5232, + "step": 2663 + }, + { + "epoch": 0.8523890784982935, + "grad_norm": 0.5980588793754578, + "learning_rate": 9.002427254258668e-06, + "loss": 0.5612, + "step": 2664 + }, + { + "epoch": 0.8527090443686007, + "grad_norm": 0.5639047026634216, + "learning_rate": 9.001311105991955e-06, + "loss": 0.5391, + "step": 2665 + }, + { + "epoch": 0.8530290102389079, + "grad_norm": 0.5928334593772888, + "learning_rate": 9.000194402938643e-06, + "loss": 0.5579, + "step": 2666 + }, + { + "epoch": 0.853348976109215, + "grad_norm": 0.6095545291900635, + "learning_rate": 8.999077145253564e-06, + "loss": 0.569, + "step": 2667 + }, + { + "epoch": 0.8536689419795221, + "grad_norm": 0.6068754196166992, + "learning_rate": 8.997959333091628e-06, + "loss": 0.5654, + "step": 2668 + }, + { + "epoch": 0.8539889078498294, + "grad_norm": 0.5735090970993042, + "learning_rate": 8.996840966607822e-06, + "loss": 0.5874, + "step": 2669 + }, + { + "epoch": 0.8543088737201365, + "grad_norm": 0.5521860718727112, + "learning_rate": 8.995722045957207e-06, + "loss": 0.541, + "step": 2670 + }, + { + "epoch": 0.8546288395904437, + "grad_norm": 0.5921656489372253, + "learning_rate": 8.994602571294923e-06, + "loss": 0.6061, + "step": 2671 + }, + { + "epoch": 0.8549488054607508, + "grad_norm": 0.4652414619922638, + "learning_rate": 8.993482542776189e-06, + "loss": 0.5335, + "step": 2672 + }, + { + "epoch": 0.855268771331058, + "grad_norm": 0.6044648885726929, + "learning_rate": 8.992361960556293e-06, + "loss": 0.581, + "step": 2673 + }, + { + "epoch": 0.8555887372013652, + "grad_norm": 0.5695050954818726, + "learning_rate": 8.99124082479061e-06, + "loss": 0.5629, + "step": 2674 + }, + { + "epoch": 0.8559087030716723, + "grad_norm": 0.5469666719436646, + "learning_rate": 8.990119135634585e-06, + "loss": 0.5034, + "step": 2675 + }, + { + "epoch": 0.8562286689419796, + "grad_norm": 0.5787575244903564, + "learning_rate": 8.988996893243742e-06, + "loss": 0.5616, + "step": 2676 + }, + { + "epoch": 0.8565486348122867, + "grad_norm": 0.56728196144104, + "learning_rate": 8.98787409777368e-06, + "loss": 0.5243, + "step": 2677 + }, + { + "epoch": 0.8568686006825939, + "grad_norm": 0.5428320169448853, + "learning_rate": 8.986750749380077e-06, + "loss": 0.4988, + "step": 2678 + }, + { + "epoch": 0.857188566552901, + "grad_norm": 0.5656844973564148, + "learning_rate": 8.985626848218685e-06, + "loss": 0.5216, + "step": 2679 + }, + { + "epoch": 0.8575085324232082, + "grad_norm": 0.5635480880737305, + "learning_rate": 8.984502394445338e-06, + "loss": 0.5733, + "step": 2680 + }, + { + "epoch": 0.8578284982935154, + "grad_norm": 0.6129047870635986, + "learning_rate": 8.98337738821594e-06, + "loss": 0.5393, + "step": 2681 + }, + { + "epoch": 0.8581484641638225, + "grad_norm": 0.5660448670387268, + "learning_rate": 8.982251829686474e-06, + "loss": 0.5313, + "step": 2682 + }, + { + "epoch": 0.8584684300341296, + "grad_norm": 0.5053395628929138, + "learning_rate": 8.981125719013002e-06, + "loss": 0.5444, + "step": 2683 + }, + { + "epoch": 0.8587883959044369, + "grad_norm": 0.5126214027404785, + "learning_rate": 8.97999905635166e-06, + "loss": 0.5096, + "step": 2684 + }, + { + "epoch": 0.859108361774744, + "grad_norm": 0.5766167044639587, + "learning_rate": 8.97887184185866e-06, + "loss": 0.5686, + "step": 2685 + }, + { + "epoch": 0.8594283276450512, + "grad_norm": 0.5058866739273071, + "learning_rate": 8.977744075690292e-06, + "loss": 0.5559, + "step": 2686 + }, + { + "epoch": 0.8597482935153583, + "grad_norm": 0.5399185419082642, + "learning_rate": 8.976615758002924e-06, + "loss": 0.5541, + "step": 2687 + }, + { + "epoch": 0.8600682593856656, + "grad_norm": 0.5576114058494568, + "learning_rate": 8.975486888952995e-06, + "loss": 0.5613, + "step": 2688 + }, + { + "epoch": 0.8603882252559727, + "grad_norm": 0.5862126350402832, + "learning_rate": 8.974357468697029e-06, + "loss": 0.5576, + "step": 2689 + }, + { + "epoch": 0.8607081911262798, + "grad_norm": 0.5256986021995544, + "learning_rate": 8.973227497391617e-06, + "loss": 0.5533, + "step": 2690 + }, + { + "epoch": 0.8610281569965871, + "grad_norm": 0.5526366829872131, + "learning_rate": 8.972096975193432e-06, + "loss": 0.5787, + "step": 2691 + }, + { + "epoch": 0.8613481228668942, + "grad_norm": 0.521307110786438, + "learning_rate": 8.970965902259225e-06, + "loss": 0.5153, + "step": 2692 + }, + { + "epoch": 0.8616680887372014, + "grad_norm": 0.5446743369102478, + "learning_rate": 8.969834278745817e-06, + "loss": 0.5052, + "step": 2693 + }, + { + "epoch": 0.8619880546075085, + "grad_norm": 0.5722483396530151, + "learning_rate": 8.96870210481011e-06, + "loss": 0.5212, + "step": 2694 + }, + { + "epoch": 0.8623080204778157, + "grad_norm": 0.6217164397239685, + "learning_rate": 8.967569380609082e-06, + "loss": 0.5632, + "step": 2695 + }, + { + "epoch": 0.8626279863481229, + "grad_norm": 0.5332825183868408, + "learning_rate": 8.966436106299787e-06, + "loss": 0.5366, + "step": 2696 + }, + { + "epoch": 0.86294795221843, + "grad_norm": 0.637496829032898, + "learning_rate": 8.965302282039355e-06, + "loss": 0.5572, + "step": 2697 + }, + { + "epoch": 0.8632679180887372, + "grad_norm": 0.5120847821235657, + "learning_rate": 8.964167907984989e-06, + "loss": 0.5332, + "step": 2698 + }, + { + "epoch": 0.8635878839590444, + "grad_norm": 0.5647338628768921, + "learning_rate": 8.963032984293974e-06, + "loss": 0.5273, + "step": 2699 + }, + { + "epoch": 0.8639078498293515, + "grad_norm": 0.5611434578895569, + "learning_rate": 8.961897511123668e-06, + "loss": 0.5539, + "step": 2700 + }, + { + "epoch": 0.8642278156996587, + "grad_norm": 0.5692500472068787, + "learning_rate": 8.960761488631507e-06, + "loss": 0.5587, + "step": 2701 + }, + { + "epoch": 0.8645477815699659, + "grad_norm": 0.5833654999732971, + "learning_rate": 8.959624916975001e-06, + "loss": 0.5765, + "step": 2702 + }, + { + "epoch": 0.8648677474402731, + "grad_norm": 0.5490713715553284, + "learning_rate": 8.958487796311735e-06, + "loss": 0.5488, + "step": 2703 + }, + { + "epoch": 0.8651877133105802, + "grad_norm": 0.5373678803443909, + "learning_rate": 8.957350126799373e-06, + "loss": 0.5698, + "step": 2704 + }, + { + "epoch": 0.8655076791808873, + "grad_norm": 0.514848530292511, + "learning_rate": 8.956211908595658e-06, + "loss": 0.5174, + "step": 2705 + }, + { + "epoch": 0.8658276450511946, + "grad_norm": 0.6037289500236511, + "learning_rate": 8.955073141858401e-06, + "loss": 0.5398, + "step": 2706 + }, + { + "epoch": 0.8661476109215017, + "grad_norm": 0.5680069327354431, + "learning_rate": 8.953933826745494e-06, + "loss": 0.5443, + "step": 2707 + }, + { + "epoch": 0.8664675767918089, + "grad_norm": 0.5605342984199524, + "learning_rate": 8.952793963414908e-06, + "loss": 0.5377, + "step": 2708 + }, + { + "epoch": 0.866787542662116, + "grad_norm": 0.5577074885368347, + "learning_rate": 8.951653552024681e-06, + "loss": 0.5356, + "step": 2709 + }, + { + "epoch": 0.8671075085324232, + "grad_norm": 0.623981237411499, + "learning_rate": 8.950512592732934e-06, + "loss": 0.6175, + "step": 2710 + }, + { + "epoch": 0.8674274744027304, + "grad_norm": 0.5811691880226135, + "learning_rate": 8.949371085697867e-06, + "loss": 0.5212, + "step": 2711 + }, + { + "epoch": 0.8677474402730375, + "grad_norm": 0.5302852392196655, + "learning_rate": 8.948229031077747e-06, + "loss": 0.5181, + "step": 2712 + }, + { + "epoch": 0.8680674061433447, + "grad_norm": 0.49263468384742737, + "learning_rate": 8.94708642903092e-06, + "loss": 0.5354, + "step": 2713 + }, + { + "epoch": 0.8683873720136519, + "grad_norm": 0.5899059176445007, + "learning_rate": 8.945943279715812e-06, + "loss": 0.5281, + "step": 2714 + }, + { + "epoch": 0.868707337883959, + "grad_norm": 0.552376389503479, + "learning_rate": 8.944799583290921e-06, + "loss": 0.5609, + "step": 2715 + }, + { + "epoch": 0.8690273037542662, + "grad_norm": 0.6187187433242798, + "learning_rate": 8.943655339914822e-06, + "loss": 0.5563, + "step": 2716 + }, + { + "epoch": 0.8693472696245734, + "grad_norm": 0.5890134572982788, + "learning_rate": 8.942510549746167e-06, + "loss": 0.553, + "step": 2717 + }, + { + "epoch": 0.8696672354948806, + "grad_norm": 0.513079047203064, + "learning_rate": 8.94136521294368e-06, + "loss": 0.5575, + "step": 2718 + }, + { + "epoch": 0.8699872013651877, + "grad_norm": 0.6054332852363586, + "learning_rate": 8.940219329666167e-06, + "loss": 0.581, + "step": 2719 + }, + { + "epoch": 0.8703071672354948, + "grad_norm": 0.6034533381462097, + "learning_rate": 8.939072900072501e-06, + "loss": 0.5318, + "step": 2720 + }, + { + "epoch": 0.8706271331058021, + "grad_norm": 0.5352435111999512, + "learning_rate": 8.93792592432164e-06, + "loss": 0.536, + "step": 2721 + }, + { + "epoch": 0.8709470989761092, + "grad_norm": 0.5317062735557556, + "learning_rate": 8.936778402572612e-06, + "loss": 0.5402, + "step": 2722 + }, + { + "epoch": 0.8712670648464164, + "grad_norm": 0.568576455116272, + "learning_rate": 8.935630334984522e-06, + "loss": 0.554, + "step": 2723 + }, + { + "epoch": 0.8715870307167235, + "grad_norm": 0.6381053924560547, + "learning_rate": 8.93448172171655e-06, + "loss": 0.5739, + "step": 2724 + }, + { + "epoch": 0.8719069965870307, + "grad_norm": 0.5327437520027161, + "learning_rate": 8.933332562927957e-06, + "loss": 0.5542, + "step": 2725 + }, + { + "epoch": 0.8722269624573379, + "grad_norm": 0.55152827501297, + "learning_rate": 8.932182858778068e-06, + "loss": 0.5835, + "step": 2726 + }, + { + "epoch": 0.872546928327645, + "grad_norm": 0.47426605224609375, + "learning_rate": 8.9310326094263e-06, + "loss": 0.5204, + "step": 2727 + }, + { + "epoch": 0.8728668941979523, + "grad_norm": 0.6053329706192017, + "learning_rate": 8.92988181503213e-06, + "loss": 0.5297, + "step": 2728 + }, + { + "epoch": 0.8731868600682594, + "grad_norm": 0.5693221092224121, + "learning_rate": 8.928730475755117e-06, + "loss": 0.5053, + "step": 2729 + }, + { + "epoch": 0.8735068259385665, + "grad_norm": 0.5186700820922852, + "learning_rate": 8.927578591754897e-06, + "loss": 0.5076, + "step": 2730 + }, + { + "epoch": 0.8738267918088737, + "grad_norm": 0.6362303495407104, + "learning_rate": 8.926426163191181e-06, + "loss": 0.5471, + "step": 2731 + }, + { + "epoch": 0.8741467576791809, + "grad_norm": 0.5959476232528687, + "learning_rate": 8.925273190223756e-06, + "loss": 0.5924, + "step": 2732 + }, + { + "epoch": 0.8744667235494881, + "grad_norm": 0.6008637547492981, + "learning_rate": 8.92411967301248e-06, + "loss": 0.5632, + "step": 2733 + }, + { + "epoch": 0.8747866894197952, + "grad_norm": 0.5581860542297363, + "learning_rate": 8.92296561171729e-06, + "loss": 0.5579, + "step": 2734 + }, + { + "epoch": 0.8751066552901023, + "grad_norm": 0.5555343627929688, + "learning_rate": 8.921811006498196e-06, + "loss": 0.5167, + "step": 2735 + }, + { + "epoch": 0.8754266211604096, + "grad_norm": 0.5411742329597473, + "learning_rate": 8.92065585751529e-06, + "loss": 0.5516, + "step": 2736 + }, + { + "epoch": 0.8757465870307167, + "grad_norm": 0.577962338924408, + "learning_rate": 8.919500164928734e-06, + "loss": 0.5955, + "step": 2737 + }, + { + "epoch": 0.8760665529010239, + "grad_norm": 0.5593159794807434, + "learning_rate": 8.918343928898763e-06, + "loss": 0.5599, + "step": 2738 + }, + { + "epoch": 0.8763865187713311, + "grad_norm": 0.5594756007194519, + "learning_rate": 8.917187149585693e-06, + "loss": 0.5772, + "step": 2739 + }, + { + "epoch": 0.8767064846416383, + "grad_norm": 0.5813189744949341, + "learning_rate": 8.916029827149913e-06, + "loss": 0.5239, + "step": 2740 + }, + { + "epoch": 0.8770264505119454, + "grad_norm": 0.47838127613067627, + "learning_rate": 8.914871961751885e-06, + "loss": 0.5125, + "step": 2741 + }, + { + "epoch": 0.8773464163822525, + "grad_norm": 0.6236173510551453, + "learning_rate": 8.91371355355215e-06, + "loss": 0.5797, + "step": 2742 + }, + { + "epoch": 0.8776663822525598, + "grad_norm": 0.5655630826950073, + "learning_rate": 8.912554602711322e-06, + "loss": 0.5769, + "step": 2743 + }, + { + "epoch": 0.8779863481228669, + "grad_norm": 0.5529646873474121, + "learning_rate": 8.911395109390092e-06, + "loss": 0.5223, + "step": 2744 + }, + { + "epoch": 0.878306313993174, + "grad_norm": 0.6270757913589478, + "learning_rate": 8.910235073749226e-06, + "loss": 0.5334, + "step": 2745 + }, + { + "epoch": 0.8786262798634812, + "grad_norm": 0.6319167613983154, + "learning_rate": 8.909074495949562e-06, + "loss": 0.5783, + "step": 2746 + }, + { + "epoch": 0.8789462457337884, + "grad_norm": 0.6389390826225281, + "learning_rate": 8.907913376152016e-06, + "loss": 0.5596, + "step": 2747 + }, + { + "epoch": 0.8792662116040956, + "grad_norm": 0.5657618045806885, + "learning_rate": 8.90675171451758e-06, + "loss": 0.5306, + "step": 2748 + }, + { + "epoch": 0.8795861774744027, + "grad_norm": 0.5682520270347595, + "learning_rate": 8.905589511207318e-06, + "loss": 0.5704, + "step": 2749 + }, + { + "epoch": 0.8799061433447098, + "grad_norm": 0.499196320772171, + "learning_rate": 8.904426766382372e-06, + "loss": 0.4701, + "step": 2750 + }, + { + "epoch": 0.8802261092150171, + "grad_norm": 0.5992483496665955, + "learning_rate": 8.903263480203956e-06, + "loss": 0.5276, + "step": 2751 + }, + { + "epoch": 0.8805460750853242, + "grad_norm": 0.5560103058815002, + "learning_rate": 8.902099652833364e-06, + "loss": 0.5456, + "step": 2752 + }, + { + "epoch": 0.8808660409556314, + "grad_norm": 0.5642792582511902, + "learning_rate": 8.900935284431962e-06, + "loss": 0.5321, + "step": 2753 + }, + { + "epoch": 0.8811860068259386, + "grad_norm": 0.516201376914978, + "learning_rate": 8.89977037516119e-06, + "loss": 0.55, + "step": 2754 + }, + { + "epoch": 0.8815059726962458, + "grad_norm": 0.5581544041633606, + "learning_rate": 8.89860492518256e-06, + "loss": 0.5317, + "step": 2755 + }, + { + "epoch": 0.8818259385665529, + "grad_norm": 0.5785426497459412, + "learning_rate": 8.897438934657672e-06, + "loss": 0.5578, + "step": 2756 + }, + { + "epoch": 0.88214590443686, + "grad_norm": 0.5539271831512451, + "learning_rate": 8.896272403748185e-06, + "loss": 0.5888, + "step": 2757 + }, + { + "epoch": 0.8824658703071673, + "grad_norm": 0.5707237124443054, + "learning_rate": 8.895105332615841e-06, + "loss": 0.5474, + "step": 2758 + }, + { + "epoch": 0.8827858361774744, + "grad_norm": 0.5254892706871033, + "learning_rate": 8.893937721422458e-06, + "loss": 0.5439, + "step": 2759 + }, + { + "epoch": 0.8831058020477816, + "grad_norm": 0.5878953337669373, + "learning_rate": 8.892769570329924e-06, + "loss": 0.5552, + "step": 2760 + }, + { + "epoch": 0.8834257679180887, + "grad_norm": 0.5976927280426025, + "learning_rate": 8.891600879500206e-06, + "loss": 0.5769, + "step": 2761 + }, + { + "epoch": 0.8837457337883959, + "grad_norm": 0.5668741464614868, + "learning_rate": 8.890431649095346e-06, + "loss": 0.5564, + "step": 2762 + }, + { + "epoch": 0.8840656996587031, + "grad_norm": 0.5591926574707031, + "learning_rate": 8.889261879277458e-06, + "loss": 0.5558, + "step": 2763 + }, + { + "epoch": 0.8843856655290102, + "grad_norm": 0.5671588182449341, + "learning_rate": 8.88809157020873e-06, + "loss": 0.5211, + "step": 2764 + }, + { + "epoch": 0.8847056313993175, + "grad_norm": 0.5909327864646912, + "learning_rate": 8.886920722051428e-06, + "loss": 0.5776, + "step": 2765 + }, + { + "epoch": 0.8850255972696246, + "grad_norm": 0.6187429428100586, + "learning_rate": 8.885749334967893e-06, + "loss": 0.5325, + "step": 2766 + }, + { + "epoch": 0.8853455631399317, + "grad_norm": 0.5392321348190308, + "learning_rate": 8.884577409120535e-06, + "loss": 0.5435, + "step": 2767 + }, + { + "epoch": 0.8856655290102389, + "grad_norm": 0.5093667507171631, + "learning_rate": 8.88340494467185e-06, + "loss": 0.5234, + "step": 2768 + }, + { + "epoch": 0.8859854948805461, + "grad_norm": 0.6120261549949646, + "learning_rate": 8.882231941784397e-06, + "loss": 0.5452, + "step": 2769 + }, + { + "epoch": 0.8863054607508533, + "grad_norm": 0.5771980285644531, + "learning_rate": 8.881058400620813e-06, + "loss": 0.5603, + "step": 2770 + }, + { + "epoch": 0.8866254266211604, + "grad_norm": 0.5893442034721375, + "learning_rate": 8.879884321343813e-06, + "loss": 0.545, + "step": 2771 + }, + { + "epoch": 0.8869453924914675, + "grad_norm": 0.52390456199646, + "learning_rate": 8.878709704116185e-06, + "loss": 0.5536, + "step": 2772 + }, + { + "epoch": 0.8872653583617748, + "grad_norm": 0.5993599891662598, + "learning_rate": 8.877534549100791e-06, + "loss": 0.5893, + "step": 2773 + }, + { + "epoch": 0.8875853242320819, + "grad_norm": 0.49362796545028687, + "learning_rate": 8.876358856460565e-06, + "loss": 0.5173, + "step": 2774 + }, + { + "epoch": 0.887905290102389, + "grad_norm": 0.5340424180030823, + "learning_rate": 8.87518262635852e-06, + "loss": 0.5477, + "step": 2775 + }, + { + "epoch": 0.8882252559726962, + "grad_norm": 0.5215820670127869, + "learning_rate": 8.874005858957743e-06, + "loss": 0.5443, + "step": 2776 + }, + { + "epoch": 0.8885452218430034, + "grad_norm": 0.5687538385391235, + "learning_rate": 8.872828554421393e-06, + "loss": 0.579, + "step": 2777 + }, + { + "epoch": 0.8888651877133106, + "grad_norm": 0.5075172185897827, + "learning_rate": 8.871650712912705e-06, + "loss": 0.5353, + "step": 2778 + }, + { + "epoch": 0.8891851535836177, + "grad_norm": 0.5134161114692688, + "learning_rate": 8.870472334594988e-06, + "loss": 0.5732, + "step": 2779 + }, + { + "epoch": 0.889505119453925, + "grad_norm": 0.53742516040802, + "learning_rate": 8.869293419631626e-06, + "loss": 0.5513, + "step": 2780 + }, + { + "epoch": 0.8898250853242321, + "grad_norm": 0.5549346804618835, + "learning_rate": 8.868113968186077e-06, + "loss": 0.5563, + "step": 2781 + }, + { + "epoch": 0.8901450511945392, + "grad_norm": 0.5468842387199402, + "learning_rate": 8.86693398042187e-06, + "loss": 0.5268, + "step": 2782 + }, + { + "epoch": 0.8904650170648464, + "grad_norm": 0.5684927105903625, + "learning_rate": 8.865753456502617e-06, + "loss": 0.5227, + "step": 2783 + }, + { + "epoch": 0.8907849829351536, + "grad_norm": 0.6799585819244385, + "learning_rate": 8.864572396591996e-06, + "loss": 0.5657, + "step": 2784 + }, + { + "epoch": 0.8911049488054608, + "grad_norm": 0.5659300684928894, + "learning_rate": 8.863390800853765e-06, + "loss": 0.5481, + "step": 2785 + }, + { + "epoch": 0.8914249146757679, + "grad_norm": 0.5571771860122681, + "learning_rate": 8.862208669451748e-06, + "loss": 0.5719, + "step": 2786 + }, + { + "epoch": 0.891744880546075, + "grad_norm": 0.6032619476318359, + "learning_rate": 8.861026002549855e-06, + "loss": 0.5787, + "step": 2787 + }, + { + "epoch": 0.8920648464163823, + "grad_norm": 0.5702298283576965, + "learning_rate": 8.85984280031206e-06, + "loss": 0.5556, + "step": 2788 + }, + { + "epoch": 0.8923848122866894, + "grad_norm": 0.5752246975898743, + "learning_rate": 8.85865906290242e-06, + "loss": 0.5418, + "step": 2789 + }, + { + "epoch": 0.8927047781569966, + "grad_norm": 0.6689168214797974, + "learning_rate": 8.857474790485057e-06, + "loss": 0.5781, + "step": 2790 + }, + { + "epoch": 0.8930247440273038, + "grad_norm": 0.5612762570381165, + "learning_rate": 8.856289983224173e-06, + "loss": 0.569, + "step": 2791 + }, + { + "epoch": 0.893344709897611, + "grad_norm": 0.5860694050788879, + "learning_rate": 8.855104641284045e-06, + "loss": 0.5601, + "step": 2792 + }, + { + "epoch": 0.8936646757679181, + "grad_norm": 0.5637890100479126, + "learning_rate": 8.853918764829019e-06, + "loss": 0.5481, + "step": 2793 + }, + { + "epoch": 0.8939846416382252, + "grad_norm": 0.6216111779212952, + "learning_rate": 8.85273235402352e-06, + "loss": 0.5817, + "step": 2794 + }, + { + "epoch": 0.8943046075085325, + "grad_norm": 0.5583057999610901, + "learning_rate": 8.851545409032047e-06, + "loss": 0.5643, + "step": 2795 + }, + { + "epoch": 0.8946245733788396, + "grad_norm": 0.5592483878135681, + "learning_rate": 8.850357930019169e-06, + "loss": 0.551, + "step": 2796 + }, + { + "epoch": 0.8949445392491467, + "grad_norm": 0.5292885303497314, + "learning_rate": 8.849169917149532e-06, + "loss": 0.5708, + "step": 2797 + }, + { + "epoch": 0.8952645051194539, + "grad_norm": 0.5390934944152832, + "learning_rate": 8.847981370587855e-06, + "loss": 0.5364, + "step": 2798 + }, + { + "epoch": 0.8955844709897611, + "grad_norm": 0.577752411365509, + "learning_rate": 8.84679229049893e-06, + "loss": 0.5575, + "step": 2799 + }, + { + "epoch": 0.8959044368600683, + "grad_norm": 0.5969122052192688, + "learning_rate": 8.845602677047629e-06, + "loss": 0.5527, + "step": 2800 + }, + { + "epoch": 0.8962244027303754, + "grad_norm": 0.5382528305053711, + "learning_rate": 8.84441253039889e-06, + "loss": 0.5714, + "step": 2801 + }, + { + "epoch": 0.8965443686006825, + "grad_norm": 0.5916634798049927, + "learning_rate": 8.843221850717726e-06, + "loss": 0.5521, + "step": 2802 + }, + { + "epoch": 0.8968643344709898, + "grad_norm": 0.554932177066803, + "learning_rate": 8.842030638169232e-06, + "loss": 0.5056, + "step": 2803 + }, + { + "epoch": 0.8971843003412969, + "grad_norm": 0.5452033281326294, + "learning_rate": 8.840838892918568e-06, + "loss": 0.5701, + "step": 2804 + }, + { + "epoch": 0.8975042662116041, + "grad_norm": 0.56942218542099, + "learning_rate": 8.83964661513097e-06, + "loss": 0.5709, + "step": 2805 + }, + { + "epoch": 0.8978242320819113, + "grad_norm": 0.5929978489875793, + "learning_rate": 8.838453804971747e-06, + "loss": 0.5407, + "step": 2806 + }, + { + "epoch": 0.8981441979522184, + "grad_norm": 0.6034539341926575, + "learning_rate": 8.83726046260629e-06, + "loss": 0.5651, + "step": 2807 + }, + { + "epoch": 0.8984641638225256, + "grad_norm": 0.6311516165733337, + "learning_rate": 8.836066588200052e-06, + "loss": 0.5282, + "step": 2808 + }, + { + "epoch": 0.8987841296928327, + "grad_norm": 0.5441495180130005, + "learning_rate": 8.834872181918566e-06, + "loss": 0.549, + "step": 2809 + }, + { + "epoch": 0.89910409556314, + "grad_norm": 0.5885206460952759, + "learning_rate": 8.833677243927439e-06, + "loss": 0.5435, + "step": 2810 + }, + { + "epoch": 0.8994240614334471, + "grad_norm": 0.5274768471717834, + "learning_rate": 8.832481774392349e-06, + "loss": 0.5029, + "step": 2811 + }, + { + "epoch": 0.8997440273037542, + "grad_norm": 0.5387116074562073, + "learning_rate": 8.831285773479053e-06, + "loss": 0.5111, + "step": 2812 + }, + { + "epoch": 0.9000639931740614, + "grad_norm": 0.5455957651138306, + "learning_rate": 8.830089241353375e-06, + "loss": 0.5528, + "step": 2813 + }, + { + "epoch": 0.9003839590443686, + "grad_norm": 0.6203275918960571, + "learning_rate": 8.828892178181215e-06, + "loss": 0.5325, + "step": 2814 + }, + { + "epoch": 0.9007039249146758, + "grad_norm": 0.5818467736244202, + "learning_rate": 8.827694584128547e-06, + "loss": 0.5351, + "step": 2815 + }, + { + "epoch": 0.9010238907849829, + "grad_norm": 0.5505111217498779, + "learning_rate": 8.82649645936142e-06, + "loss": 0.5425, + "step": 2816 + }, + { + "epoch": 0.9013438566552902, + "grad_norm": 0.6561931371688843, + "learning_rate": 8.825297804045958e-06, + "loss": 0.6024, + "step": 2817 + }, + { + "epoch": 0.9016638225255973, + "grad_norm": 0.5352081060409546, + "learning_rate": 8.824098618348353e-06, + "loss": 0.5039, + "step": 2818 + }, + { + "epoch": 0.9019837883959044, + "grad_norm": 0.5815563797950745, + "learning_rate": 8.822898902434873e-06, + "loss": 0.5709, + "step": 2819 + }, + { + "epoch": 0.9023037542662116, + "grad_norm": 0.5135438442230225, + "learning_rate": 8.821698656471863e-06, + "loss": 0.5081, + "step": 2820 + }, + { + "epoch": 0.9026237201365188, + "grad_norm": 0.6108431816101074, + "learning_rate": 8.820497880625733e-06, + "loss": 0.5468, + "step": 2821 + }, + { + "epoch": 0.902943686006826, + "grad_norm": 0.5745890140533447, + "learning_rate": 8.819296575062978e-06, + "loss": 0.5556, + "step": 2822 + }, + { + "epoch": 0.9032636518771331, + "grad_norm": 0.5635342001914978, + "learning_rate": 8.818094739950157e-06, + "loss": 0.524, + "step": 2823 + }, + { + "epoch": 0.9035836177474402, + "grad_norm": 0.5868473649024963, + "learning_rate": 8.81689237545391e-06, + "loss": 0.543, + "step": 2824 + }, + { + "epoch": 0.9039035836177475, + "grad_norm": 0.6234641671180725, + "learning_rate": 8.815689481740942e-06, + "loss": 0.5314, + "step": 2825 + }, + { + "epoch": 0.9042235494880546, + "grad_norm": 0.6335117816925049, + "learning_rate": 8.814486058978035e-06, + "loss": 0.545, + "step": 2826 + }, + { + "epoch": 0.9045435153583617, + "grad_norm": 0.5347116589546204, + "learning_rate": 8.81328210733205e-06, + "loss": 0.5594, + "step": 2827 + }, + { + "epoch": 0.9048634812286689, + "grad_norm": 0.560544490814209, + "learning_rate": 8.81207762696991e-06, + "loss": 0.5276, + "step": 2828 + }, + { + "epoch": 0.9051834470989761, + "grad_norm": 0.5208972692489624, + "learning_rate": 8.810872618058622e-06, + "loss": 0.5337, + "step": 2829 + }, + { + "epoch": 0.9055034129692833, + "grad_norm": 0.5120471715927124, + "learning_rate": 8.809667080765262e-06, + "loss": 0.5361, + "step": 2830 + }, + { + "epoch": 0.9058233788395904, + "grad_norm": 0.551671028137207, + "learning_rate": 8.808461015256976e-06, + "loss": 0.5768, + "step": 2831 + }, + { + "epoch": 0.9061433447098977, + "grad_norm": 0.6156903505325317, + "learning_rate": 8.807254421700991e-06, + "loss": 0.521, + "step": 2832 + }, + { + "epoch": 0.9064633105802048, + "grad_norm": 0.5635914206504822, + "learning_rate": 8.8060473002646e-06, + "loss": 0.5314, + "step": 2833 + }, + { + "epoch": 0.9067832764505119, + "grad_norm": 0.5528858304023743, + "learning_rate": 8.804839651115171e-06, + "loss": 0.547, + "step": 2834 + }, + { + "epoch": 0.9071032423208191, + "grad_norm": 0.5680487751960754, + "learning_rate": 8.803631474420146e-06, + "loss": 0.5885, + "step": 2835 + }, + { + "epoch": 0.9074232081911263, + "grad_norm": 0.6367501616477966, + "learning_rate": 8.802422770347044e-06, + "loss": 0.6022, + "step": 2836 + }, + { + "epoch": 0.9077431740614335, + "grad_norm": 0.5560848712921143, + "learning_rate": 8.801213539063448e-06, + "loss": 0.5123, + "step": 2837 + }, + { + "epoch": 0.9080631399317406, + "grad_norm": 0.5595853328704834, + "learning_rate": 8.800003780737024e-06, + "loss": 0.5648, + "step": 2838 + }, + { + "epoch": 0.9083831058020477, + "grad_norm": 0.5442648530006409, + "learning_rate": 8.798793495535503e-06, + "loss": 0.5222, + "step": 2839 + }, + { + "epoch": 0.908703071672355, + "grad_norm": 0.5956857204437256, + "learning_rate": 8.797582683626693e-06, + "loss": 0.5404, + "step": 2840 + }, + { + "epoch": 0.9090230375426621, + "grad_norm": 0.6557626128196716, + "learning_rate": 8.796371345178477e-06, + "loss": 0.5639, + "step": 2841 + }, + { + "epoch": 0.9093430034129693, + "grad_norm": 0.5646318793296814, + "learning_rate": 8.795159480358806e-06, + "loss": 0.5396, + "step": 2842 + }, + { + "epoch": 0.9096629692832765, + "grad_norm": 0.5170820951461792, + "learning_rate": 8.793947089335709e-06, + "loss": 0.5738, + "step": 2843 + }, + { + "epoch": 0.9099829351535836, + "grad_norm": 0.520595133304596, + "learning_rate": 8.792734172277282e-06, + "loss": 0.5519, + "step": 2844 + }, + { + "epoch": 0.9103029010238908, + "grad_norm": 0.562983512878418, + "learning_rate": 8.791520729351699e-06, + "loss": 0.5641, + "step": 2845 + }, + { + "epoch": 0.9106228668941979, + "grad_norm": 0.6088793873786926, + "learning_rate": 8.790306760727206e-06, + "loss": 0.5708, + "step": 2846 + }, + { + "epoch": 0.9109428327645052, + "grad_norm": 0.5645493268966675, + "learning_rate": 8.78909226657212e-06, + "loss": 0.5257, + "step": 2847 + }, + { + "epoch": 0.9112627986348123, + "grad_norm": 0.6144366264343262, + "learning_rate": 8.787877247054835e-06, + "loss": 0.5757, + "step": 2848 + }, + { + "epoch": 0.9115827645051194, + "grad_norm": 0.6192265152931213, + "learning_rate": 8.786661702343811e-06, + "loss": 0.5786, + "step": 2849 + }, + { + "epoch": 0.9119027303754266, + "grad_norm": 0.5839664340019226, + "learning_rate": 8.785445632607587e-06, + "loss": 0.5418, + "step": 2850 + }, + { + "epoch": 0.9122226962457338, + "grad_norm": 0.5896580219268799, + "learning_rate": 8.784229038014772e-06, + "loss": 0.5292, + "step": 2851 + }, + { + "epoch": 0.912542662116041, + "grad_norm": 0.594308614730835, + "learning_rate": 8.783011918734048e-06, + "loss": 0.5923, + "step": 2852 + }, + { + "epoch": 0.9128626279863481, + "grad_norm": 0.5671731233596802, + "learning_rate": 8.78179427493417e-06, + "loss": 0.5543, + "step": 2853 + }, + { + "epoch": 0.9131825938566553, + "grad_norm": 0.5983908772468567, + "learning_rate": 8.780576106783968e-06, + "loss": 0.5705, + "step": 2854 + }, + { + "epoch": 0.9135025597269625, + "grad_norm": 0.6325986385345459, + "learning_rate": 8.77935741445234e-06, + "loss": 0.5402, + "step": 2855 + }, + { + "epoch": 0.9138225255972696, + "grad_norm": 0.542965292930603, + "learning_rate": 8.778138198108259e-06, + "loss": 0.5163, + "step": 2856 + }, + { + "epoch": 0.9141424914675768, + "grad_norm": 0.5294291973114014, + "learning_rate": 8.776918457920772e-06, + "loss": 0.5433, + "step": 2857 + }, + { + "epoch": 0.914462457337884, + "grad_norm": 0.6218854784965515, + "learning_rate": 8.775698194058996e-06, + "loss": 0.5748, + "step": 2858 + }, + { + "epoch": 0.9147824232081911, + "grad_norm": 0.5364216566085815, + "learning_rate": 8.774477406692125e-06, + "loss": 0.5315, + "step": 2859 + }, + { + "epoch": 0.9151023890784983, + "grad_norm": 0.5388673543930054, + "learning_rate": 8.77325609598942e-06, + "loss": 0.569, + "step": 2860 + }, + { + "epoch": 0.9154223549488054, + "grad_norm": 0.5232530236244202, + "learning_rate": 8.772034262120219e-06, + "loss": 0.5583, + "step": 2861 + }, + { + "epoch": 0.9157423208191127, + "grad_norm": 0.5909742116928101, + "learning_rate": 8.770811905253929e-06, + "loss": 0.5538, + "step": 2862 + }, + { + "epoch": 0.9160622866894198, + "grad_norm": 0.547285795211792, + "learning_rate": 8.76958902556003e-06, + "loss": 0.5687, + "step": 2863 + }, + { + "epoch": 0.9163822525597269, + "grad_norm": 0.5807123780250549, + "learning_rate": 8.768365623208079e-06, + "loss": 0.5542, + "step": 2864 + }, + { + "epoch": 0.9167022184300341, + "grad_norm": 0.5130710601806641, + "learning_rate": 8.767141698367701e-06, + "loss": 0.5538, + "step": 2865 + }, + { + "epoch": 0.9170221843003413, + "grad_norm": 0.5519154071807861, + "learning_rate": 8.765917251208595e-06, + "loss": 0.5351, + "step": 2866 + }, + { + "epoch": 0.9173421501706485, + "grad_norm": 0.5024713277816772, + "learning_rate": 8.764692281900531e-06, + "loss": 0.5335, + "step": 2867 + }, + { + "epoch": 0.9176621160409556, + "grad_norm": 0.5541590452194214, + "learning_rate": 8.763466790613354e-06, + "loss": 0.5737, + "step": 2868 + }, + { + "epoch": 0.9179820819112628, + "grad_norm": 0.5962385535240173, + "learning_rate": 8.762240777516979e-06, + "loss": 0.5615, + "step": 2869 + }, + { + "epoch": 0.91830204778157, + "grad_norm": 0.6544550657272339, + "learning_rate": 8.761014242781392e-06, + "loss": 0.5484, + "step": 2870 + }, + { + "epoch": 0.9186220136518771, + "grad_norm": 0.5623205304145813, + "learning_rate": 8.759787186576659e-06, + "loss": 0.539, + "step": 2871 + }, + { + "epoch": 0.9189419795221843, + "grad_norm": 0.5766005516052246, + "learning_rate": 8.758559609072906e-06, + "loss": 0.5491, + "step": 2872 + }, + { + "epoch": 0.9192619453924915, + "grad_norm": 0.5268698334693909, + "learning_rate": 8.757331510440343e-06, + "loss": 0.5364, + "step": 2873 + }, + { + "epoch": 0.9195819112627986, + "grad_norm": 0.4956420660018921, + "learning_rate": 8.756102890849246e-06, + "loss": 0.5355, + "step": 2874 + }, + { + "epoch": 0.9199018771331058, + "grad_norm": 0.5091249942779541, + "learning_rate": 8.754873750469964e-06, + "loss": 0.5329, + "step": 2875 + }, + { + "epoch": 0.9202218430034129, + "grad_norm": 0.6004356145858765, + "learning_rate": 8.753644089472921e-06, + "loss": 0.5531, + "step": 2876 + }, + { + "epoch": 0.9205418088737202, + "grad_norm": 0.5854215621948242, + "learning_rate": 8.752413908028608e-06, + "loss": 0.5494, + "step": 2877 + }, + { + "epoch": 0.9208617747440273, + "grad_norm": 0.5781176090240479, + "learning_rate": 8.751183206307592e-06, + "loss": 0.5567, + "step": 2878 + }, + { + "epoch": 0.9211817406143344, + "grad_norm": 0.5860122442245483, + "learning_rate": 8.749951984480511e-06, + "loss": 0.5464, + "step": 2879 + }, + { + "epoch": 0.9215017064846417, + "grad_norm": 0.5667566061019897, + "learning_rate": 8.748720242718077e-06, + "loss": 0.5346, + "step": 2880 + }, + { + "epoch": 0.9218216723549488, + "grad_norm": 0.5337679386138916, + "learning_rate": 8.747487981191072e-06, + "loss": 0.5867, + "step": 2881 + }, + { + "epoch": 0.922141638225256, + "grad_norm": 0.5574522614479065, + "learning_rate": 8.74625520007035e-06, + "loss": 0.5468, + "step": 2882 + }, + { + "epoch": 0.9224616040955631, + "grad_norm": 0.5934003591537476, + "learning_rate": 8.745021899526836e-06, + "loss": 0.5639, + "step": 2883 + }, + { + "epoch": 0.9227815699658704, + "grad_norm": 0.5360896587371826, + "learning_rate": 8.743788079731533e-06, + "loss": 0.5547, + "step": 2884 + }, + { + "epoch": 0.9231015358361775, + "grad_norm": 0.544714629650116, + "learning_rate": 8.742553740855507e-06, + "loss": 0.5821, + "step": 2885 + }, + { + "epoch": 0.9234215017064846, + "grad_norm": 0.576442301273346, + "learning_rate": 8.741318883069903e-06, + "loss": 0.5365, + "step": 2886 + }, + { + "epoch": 0.9237414675767918, + "grad_norm": 0.508274257183075, + "learning_rate": 8.740083506545933e-06, + "loss": 0.5391, + "step": 2887 + }, + { + "epoch": 0.924061433447099, + "grad_norm": 0.6314295530319214, + "learning_rate": 8.738847611454887e-06, + "loss": 0.5923, + "step": 2888 + }, + { + "epoch": 0.9243813993174061, + "grad_norm": 0.5180608630180359, + "learning_rate": 8.737611197968123e-06, + "loss": 0.5535, + "step": 2889 + }, + { + "epoch": 0.9247013651877133, + "grad_norm": 0.5145260691642761, + "learning_rate": 8.736374266257069e-06, + "loss": 0.5231, + "step": 2890 + }, + { + "epoch": 0.9250213310580204, + "grad_norm": 0.5807346701622009, + "learning_rate": 8.735136816493227e-06, + "loss": 0.5227, + "step": 2891 + }, + { + "epoch": 0.9253412969283277, + "grad_norm": 0.5591275691986084, + "learning_rate": 8.733898848848172e-06, + "loss": 0.5787, + "step": 2892 + }, + { + "epoch": 0.9256612627986348, + "grad_norm": 0.5673110485076904, + "learning_rate": 8.732660363493551e-06, + "loss": 0.5172, + "step": 2893 + }, + { + "epoch": 0.925981228668942, + "grad_norm": 0.49727028608322144, + "learning_rate": 8.73142136060108e-06, + "loss": 0.4982, + "step": 2894 + }, + { + "epoch": 0.9263011945392492, + "grad_norm": 0.5635839700698853, + "learning_rate": 8.730181840342547e-06, + "loss": 0.5463, + "step": 2895 + }, + { + "epoch": 0.9266211604095563, + "grad_norm": 0.5192051529884338, + "learning_rate": 8.728941802889816e-06, + "loss": 0.5239, + "step": 2896 + }, + { + "epoch": 0.9269411262798635, + "grad_norm": 0.5240466594696045, + "learning_rate": 8.727701248414816e-06, + "loss": 0.5677, + "step": 2897 + }, + { + "epoch": 0.9272610921501706, + "grad_norm": 0.5680010318756104, + "learning_rate": 8.726460177089555e-06, + "loss": 0.5549, + "step": 2898 + }, + { + "epoch": 0.9275810580204779, + "grad_norm": 0.5307434797286987, + "learning_rate": 8.725218589086107e-06, + "loss": 0.5343, + "step": 2899 + }, + { + "epoch": 0.927901023890785, + "grad_norm": 0.5652339458465576, + "learning_rate": 8.72397648457662e-06, + "loss": 0.581, + "step": 2900 + }, + { + "epoch": 0.9282209897610921, + "grad_norm": 0.5568917393684387, + "learning_rate": 8.722733863733314e-06, + "loss": 0.5728, + "step": 2901 + }, + { + "epoch": 0.9285409556313993, + "grad_norm": 0.5474474430084229, + "learning_rate": 8.721490726728477e-06, + "loss": 0.525, + "step": 2902 + }, + { + "epoch": 0.9288609215017065, + "grad_norm": 0.5851140022277832, + "learning_rate": 8.720247073734477e-06, + "loss": 0.5398, + "step": 2903 + }, + { + "epoch": 0.9291808873720137, + "grad_norm": 0.5718165040016174, + "learning_rate": 8.719002904923742e-06, + "loss": 0.5565, + "step": 2904 + }, + { + "epoch": 0.9295008532423208, + "grad_norm": 0.5905211567878723, + "learning_rate": 8.717758220468781e-06, + "loss": 0.5636, + "step": 2905 + }, + { + "epoch": 0.929820819112628, + "grad_norm": 0.5033652186393738, + "learning_rate": 8.71651302054217e-06, + "loss": 0.5419, + "step": 2906 + }, + { + "epoch": 0.9301407849829352, + "grad_norm": 0.5867974162101746, + "learning_rate": 8.715267305316559e-06, + "loss": 0.5789, + "step": 2907 + }, + { + "epoch": 0.9304607508532423, + "grad_norm": 0.5967637300491333, + "learning_rate": 8.714021074964665e-06, + "loss": 0.5306, + "step": 2908 + }, + { + "epoch": 0.9307807167235495, + "grad_norm": 0.632807195186615, + "learning_rate": 8.712774329659282e-06, + "loss": 0.5807, + "step": 2909 + }, + { + "epoch": 0.9311006825938567, + "grad_norm": 0.6260687708854675, + "learning_rate": 8.71152706957327e-06, + "loss": 0.5246, + "step": 2910 + }, + { + "epoch": 0.9314206484641638, + "grad_norm": 0.5414279103279114, + "learning_rate": 8.710279294879565e-06, + "loss": 0.5595, + "step": 2911 + }, + { + "epoch": 0.931740614334471, + "grad_norm": 0.5461296439170837, + "learning_rate": 8.709031005751173e-06, + "loss": 0.5444, + "step": 2912 + }, + { + "epoch": 0.9320605802047781, + "grad_norm": 0.5828396677970886, + "learning_rate": 8.707782202361171e-06, + "loss": 0.5554, + "step": 2913 + }, + { + "epoch": 0.9323805460750854, + "grad_norm": 0.6433313488960266, + "learning_rate": 8.706532884882704e-06, + "loss": 0.5773, + "step": 2914 + }, + { + "epoch": 0.9327005119453925, + "grad_norm": 0.5411011576652527, + "learning_rate": 8.705283053488994e-06, + "loss": 0.5083, + "step": 2915 + }, + { + "epoch": 0.9330204778156996, + "grad_norm": 0.5453394055366516, + "learning_rate": 8.704032708353331e-06, + "loss": 0.5247, + "step": 2916 + }, + { + "epoch": 0.9333404436860068, + "grad_norm": 0.5410546064376831, + "learning_rate": 8.702781849649078e-06, + "loss": 0.5624, + "step": 2917 + }, + { + "epoch": 0.933660409556314, + "grad_norm": 0.644980788230896, + "learning_rate": 8.701530477549666e-06, + "loss": 0.5544, + "step": 2918 + }, + { + "epoch": 0.9339803754266212, + "grad_norm": 0.5591293573379517, + "learning_rate": 8.7002785922286e-06, + "loss": 0.5541, + "step": 2919 + }, + { + "epoch": 0.9343003412969283, + "grad_norm": 0.5224248766899109, + "learning_rate": 8.699026193859457e-06, + "loss": 0.5256, + "step": 2920 + }, + { + "epoch": 0.9346203071672355, + "grad_norm": 0.5552311539649963, + "learning_rate": 8.697773282615881e-06, + "loss": 0.5575, + "step": 2921 + }, + { + "epoch": 0.9349402730375427, + "grad_norm": 0.5218842029571533, + "learning_rate": 8.69651985867159e-06, + "loss": 0.4947, + "step": 2922 + }, + { + "epoch": 0.9352602389078498, + "grad_norm": 0.4850265681743622, + "learning_rate": 8.695265922200376e-06, + "loss": 0.5369, + "step": 2923 + }, + { + "epoch": 0.935580204778157, + "grad_norm": 0.48894792795181274, + "learning_rate": 8.694011473376094e-06, + "loss": 0.5039, + "step": 2924 + }, + { + "epoch": 0.9359001706484642, + "grad_norm": 0.4918539226055145, + "learning_rate": 8.69275651237268e-06, + "loss": 0.5237, + "step": 2925 + }, + { + "epoch": 0.9362201365187713, + "grad_norm": 0.5880904197692871, + "learning_rate": 8.69150103936413e-06, + "loss": 0.5804, + "step": 2926 + }, + { + "epoch": 0.9365401023890785, + "grad_norm": 0.4817546010017395, + "learning_rate": 8.690245054524522e-06, + "loss": 0.5256, + "step": 2927 + }, + { + "epoch": 0.9368600682593856, + "grad_norm": 0.5519566535949707, + "learning_rate": 8.688988558027997e-06, + "loss": 0.5757, + "step": 2928 + }, + { + "epoch": 0.9371800341296929, + "grad_norm": 0.5464499592781067, + "learning_rate": 8.68773155004877e-06, + "loss": 0.5341, + "step": 2929 + }, + { + "epoch": 0.9375, + "grad_norm": 0.517829954624176, + "learning_rate": 8.68647403076113e-06, + "loss": 0.5241, + "step": 2930 + }, + { + "epoch": 0.9378199658703071, + "grad_norm": 0.5223374366760254, + "learning_rate": 8.685216000339426e-06, + "loss": 0.5226, + "step": 2931 + }, + { + "epoch": 0.9381399317406144, + "grad_norm": 0.6255651116371155, + "learning_rate": 8.683957458958093e-06, + "loss": 0.5509, + "step": 2932 + }, + { + "epoch": 0.9384598976109215, + "grad_norm": 0.5552142858505249, + "learning_rate": 8.682698406791627e-06, + "loss": 0.5545, + "step": 2933 + }, + { + "epoch": 0.9387798634812287, + "grad_norm": 0.52936851978302, + "learning_rate": 8.681438844014595e-06, + "loss": 0.5503, + "step": 2934 + }, + { + "epoch": 0.9390998293515358, + "grad_norm": 0.5493532419204712, + "learning_rate": 8.680178770801639e-06, + "loss": 0.5101, + "step": 2935 + }, + { + "epoch": 0.939419795221843, + "grad_norm": 0.5778977870941162, + "learning_rate": 8.678918187327467e-06, + "loss": 0.5461, + "step": 2936 + }, + { + "epoch": 0.9397397610921502, + "grad_norm": 0.49943238496780396, + "learning_rate": 8.677657093766865e-06, + "loss": 0.5331, + "step": 2937 + }, + { + "epoch": 0.9400597269624573, + "grad_norm": 0.5420438051223755, + "learning_rate": 8.676395490294683e-06, + "loss": 0.5324, + "step": 2938 + }, + { + "epoch": 0.9403796928327645, + "grad_norm": 0.501508355140686, + "learning_rate": 8.675133377085842e-06, + "loss": 0.5687, + "step": 2939 + }, + { + "epoch": 0.9406996587030717, + "grad_norm": 0.5415540337562561, + "learning_rate": 8.673870754315336e-06, + "loss": 0.5755, + "step": 2940 + }, + { + "epoch": 0.9410196245733788, + "grad_norm": 0.5656829476356506, + "learning_rate": 8.672607622158232e-06, + "loss": 0.5624, + "step": 2941 + }, + { + "epoch": 0.941339590443686, + "grad_norm": 0.5423877239227295, + "learning_rate": 8.671343980789664e-06, + "loss": 0.5723, + "step": 2942 + }, + { + "epoch": 0.9416595563139932, + "grad_norm": 0.5485814809799194, + "learning_rate": 8.670079830384834e-06, + "loss": 0.5432, + "step": 2943 + }, + { + "epoch": 0.9419795221843004, + "grad_norm": 0.5483697652816772, + "learning_rate": 8.66881517111902e-06, + "loss": 0.5533, + "step": 2944 + }, + { + "epoch": 0.9422994880546075, + "grad_norm": 0.5534570217132568, + "learning_rate": 8.667550003167571e-06, + "loss": 0.5706, + "step": 2945 + }, + { + "epoch": 0.9426194539249146, + "grad_norm": 0.5758525133132935, + "learning_rate": 8.6662843267059e-06, + "loss": 0.5817, + "step": 2946 + }, + { + "epoch": 0.9429394197952219, + "grad_norm": 0.5041936039924622, + "learning_rate": 8.665018141909498e-06, + "loss": 0.5227, + "step": 2947 + }, + { + "epoch": 0.943259385665529, + "grad_norm": 0.5714983344078064, + "learning_rate": 8.663751448953921e-06, + "loss": 0.5125, + "step": 2948 + }, + { + "epoch": 0.9435793515358362, + "grad_norm": 0.5733338594436646, + "learning_rate": 8.662484248014798e-06, + "loss": 0.5499, + "step": 2949 + }, + { + "epoch": 0.9438993174061433, + "grad_norm": 0.5094056725502014, + "learning_rate": 8.661216539267827e-06, + "loss": 0.5274, + "step": 2950 + }, + { + "epoch": 0.9442192832764505, + "grad_norm": 0.49745941162109375, + "learning_rate": 8.659948322888778e-06, + "loss": 0.5243, + "step": 2951 + }, + { + "epoch": 0.9445392491467577, + "grad_norm": 0.5140569806098938, + "learning_rate": 8.658679599053493e-06, + "loss": 0.511, + "step": 2952 + }, + { + "epoch": 0.9448592150170648, + "grad_norm": 0.49003928899765015, + "learning_rate": 8.65741036793788e-06, + "loss": 0.5207, + "step": 2953 + }, + { + "epoch": 0.945179180887372, + "grad_norm": 0.5427951216697693, + "learning_rate": 8.656140629717918e-06, + "loss": 0.5698, + "step": 2954 + }, + { + "epoch": 0.9454991467576792, + "grad_norm": 0.5415076017379761, + "learning_rate": 8.65487038456966e-06, + "loss": 0.5584, + "step": 2955 + }, + { + "epoch": 0.9458191126279863, + "grad_norm": 0.5324149131774902, + "learning_rate": 8.653599632669225e-06, + "loss": 0.5395, + "step": 2956 + }, + { + "epoch": 0.9461390784982935, + "grad_norm": 0.538913369178772, + "learning_rate": 8.652328374192807e-06, + "loss": 0.5625, + "step": 2957 + }, + { + "epoch": 0.9464590443686007, + "grad_norm": 0.5423220992088318, + "learning_rate": 8.651056609316666e-06, + "loss": 0.5491, + "step": 2958 + }, + { + "epoch": 0.9467790102389079, + "grad_norm": 0.5532695651054382, + "learning_rate": 8.649784338217133e-06, + "loss": 0.545, + "step": 2959 + }, + { + "epoch": 0.947098976109215, + "grad_norm": 0.5493027567863464, + "learning_rate": 8.648511561070611e-06, + "loss": 0.5579, + "step": 2960 + }, + { + "epoch": 0.9474189419795221, + "grad_norm": 0.5765998363494873, + "learning_rate": 8.647238278053572e-06, + "loss": 0.5608, + "step": 2961 + }, + { + "epoch": 0.9477389078498294, + "grad_norm": 0.5316721796989441, + "learning_rate": 8.645964489342558e-06, + "loss": 0.5451, + "step": 2962 + }, + { + "epoch": 0.9480588737201365, + "grad_norm": 0.5517355799674988, + "learning_rate": 8.644690195114183e-06, + "loss": 0.5615, + "step": 2963 + }, + { + "epoch": 0.9483788395904437, + "grad_norm": 0.5976579189300537, + "learning_rate": 8.643415395545125e-06, + "loss": 0.5925, + "step": 2964 + }, + { + "epoch": 0.9486988054607508, + "grad_norm": 0.5386163592338562, + "learning_rate": 8.642140090812141e-06, + "loss": 0.5727, + "step": 2965 + }, + { + "epoch": 0.949018771331058, + "grad_norm": 0.6165236234664917, + "learning_rate": 8.640864281092051e-06, + "loss": 0.5379, + "step": 2966 + }, + { + "epoch": 0.9493387372013652, + "grad_norm": 0.5166876316070557, + "learning_rate": 8.639587966561748e-06, + "loss": 0.538, + "step": 2967 + }, + { + "epoch": 0.9496587030716723, + "grad_norm": 0.5584707856178284, + "learning_rate": 8.638311147398195e-06, + "loss": 0.5364, + "step": 2968 + }, + { + "epoch": 0.9499786689419796, + "grad_norm": 0.5661454796791077, + "learning_rate": 8.637033823778426e-06, + "loss": 0.5683, + "step": 2969 + }, + { + "epoch": 0.9502986348122867, + "grad_norm": 0.43857452273368835, + "learning_rate": 8.63575599587954e-06, + "loss": 0.5328, + "step": 2970 + }, + { + "epoch": 0.9506186006825939, + "grad_norm": 0.5182543396949768, + "learning_rate": 8.634477663878714e-06, + "loss": 0.5386, + "step": 2971 + }, + { + "epoch": 0.950938566552901, + "grad_norm": 0.6035263538360596, + "learning_rate": 8.633198827953185e-06, + "loss": 0.5329, + "step": 2972 + }, + { + "epoch": 0.9512585324232082, + "grad_norm": 0.6187506318092346, + "learning_rate": 8.631919488280267e-06, + "loss": 0.6131, + "step": 2973 + }, + { + "epoch": 0.9515784982935154, + "grad_norm": 0.6119416356086731, + "learning_rate": 8.630639645037345e-06, + "loss": 0.5745, + "step": 2974 + }, + { + "epoch": 0.9518984641638225, + "grad_norm": 0.6179429292678833, + "learning_rate": 8.629359298401866e-06, + "loss": 0.5084, + "step": 2975 + }, + { + "epoch": 0.9522184300341296, + "grad_norm": 0.5868006348609924, + "learning_rate": 8.628078448551355e-06, + "loss": 0.5159, + "step": 2976 + }, + { + "epoch": 0.9525383959044369, + "grad_norm": 0.5683136582374573, + "learning_rate": 8.626797095663403e-06, + "loss": 0.5664, + "step": 2977 + }, + { + "epoch": 0.952858361774744, + "grad_norm": 0.5581501722335815, + "learning_rate": 8.62551523991567e-06, + "loss": 0.5361, + "step": 2978 + }, + { + "epoch": 0.9531783276450512, + "grad_norm": 0.5560288429260254, + "learning_rate": 8.624232881485887e-06, + "loss": 0.5436, + "step": 2979 + }, + { + "epoch": 0.9534982935153583, + "grad_norm": 0.536558985710144, + "learning_rate": 8.622950020551857e-06, + "loss": 0.5327, + "step": 2980 + }, + { + "epoch": 0.9538182593856656, + "grad_norm": 0.5403125286102295, + "learning_rate": 8.621666657291446e-06, + "loss": 0.5641, + "step": 2981 + }, + { + "epoch": 0.9541382252559727, + "grad_norm": 0.5872126221656799, + "learning_rate": 8.620382791882597e-06, + "loss": 0.5457, + "step": 2982 + }, + { + "epoch": 0.9544581911262798, + "grad_norm": 0.6266401410102844, + "learning_rate": 8.619098424503318e-06, + "loss": 0.5543, + "step": 2983 + }, + { + "epoch": 0.9547781569965871, + "grad_norm": 0.5969432592391968, + "learning_rate": 8.61781355533169e-06, + "loss": 0.5553, + "step": 2984 + }, + { + "epoch": 0.9550981228668942, + "grad_norm": 0.5646101832389832, + "learning_rate": 8.616528184545858e-06, + "loss": 0.5469, + "step": 2985 + }, + { + "epoch": 0.9554180887372014, + "grad_norm": 0.5284569263458252, + "learning_rate": 8.615242312324043e-06, + "loss": 0.5331, + "step": 2986 + }, + { + "epoch": 0.9557380546075085, + "grad_norm": 0.5586323738098145, + "learning_rate": 8.613955938844533e-06, + "loss": 0.5167, + "step": 2987 + }, + { + "epoch": 0.9560580204778157, + "grad_norm": 0.612792432308197, + "learning_rate": 8.612669064285684e-06, + "loss": 0.5447, + "step": 2988 + }, + { + "epoch": 0.9563779863481229, + "grad_norm": 0.5168460011482239, + "learning_rate": 8.611381688825924e-06, + "loss": 0.524, + "step": 2989 + }, + { + "epoch": 0.95669795221843, + "grad_norm": 0.5096966028213501, + "learning_rate": 8.610093812643747e-06, + "loss": 0.527, + "step": 2990 + }, + { + "epoch": 0.9570179180887372, + "grad_norm": 0.5304121375083923, + "learning_rate": 8.60880543591772e-06, + "loss": 0.5298, + "step": 2991 + }, + { + "epoch": 0.9573378839590444, + "grad_norm": 0.5534718036651611, + "learning_rate": 8.607516558826477e-06, + "loss": 0.5516, + "step": 2992 + }, + { + "epoch": 0.9576578498293515, + "grad_norm": 0.5174388289451599, + "learning_rate": 8.606227181548725e-06, + "loss": 0.5036, + "step": 2993 + }, + { + "epoch": 0.9579778156996587, + "grad_norm": 0.6583805084228516, + "learning_rate": 8.604937304263234e-06, + "loss": 0.5903, + "step": 2994 + }, + { + "epoch": 0.9582977815699659, + "grad_norm": 0.5533691644668579, + "learning_rate": 8.60364692714885e-06, + "loss": 0.5468, + "step": 2995 + }, + { + "epoch": 0.9586177474402731, + "grad_norm": 0.6131950616836548, + "learning_rate": 8.602356050384483e-06, + "loss": 0.5772, + "step": 2996 + }, + { + "epoch": 0.9589377133105802, + "grad_norm": 0.5823208689689636, + "learning_rate": 8.601064674149118e-06, + "loss": 0.5806, + "step": 2997 + }, + { + "epoch": 0.9592576791808873, + "grad_norm": 0.4916262626647949, + "learning_rate": 8.599772798621804e-06, + "loss": 0.516, + "step": 2998 + }, + { + "epoch": 0.9595776450511946, + "grad_norm": 0.6060745120048523, + "learning_rate": 8.598480423981658e-06, + "loss": 0.5884, + "step": 2999 + }, + { + "epoch": 0.9598976109215017, + "grad_norm": 0.618219792842865, + "learning_rate": 8.597187550407875e-06, + "loss": 0.5709, + "step": 3000 + }, + { + "epoch": 0.9602175767918089, + "grad_norm": 0.5556377172470093, + "learning_rate": 8.59589417807971e-06, + "loss": 0.5586, + "step": 3001 + }, + { + "epoch": 0.960537542662116, + "grad_norm": 0.5143526196479797, + "learning_rate": 8.594600307176489e-06, + "loss": 0.5287, + "step": 3002 + }, + { + "epoch": 0.9608575085324232, + "grad_norm": 0.6290923357009888, + "learning_rate": 8.593305937877614e-06, + "loss": 0.5302, + "step": 3003 + }, + { + "epoch": 0.9611774744027304, + "grad_norm": 0.5228386521339417, + "learning_rate": 8.592011070362546e-06, + "loss": 0.5407, + "step": 3004 + }, + { + "epoch": 0.9614974402730375, + "grad_norm": 0.579010009765625, + "learning_rate": 8.590715704810823e-06, + "loss": 0.5843, + "step": 3005 + }, + { + "epoch": 0.9618174061433447, + "grad_norm": 0.5646539926528931, + "learning_rate": 8.589419841402046e-06, + "loss": 0.551, + "step": 3006 + }, + { + "epoch": 0.9621373720136519, + "grad_norm": 0.5978478193283081, + "learning_rate": 8.58812348031589e-06, + "loss": 0.5851, + "step": 3007 + }, + { + "epoch": 0.962457337883959, + "grad_norm": 0.5651780962944031, + "learning_rate": 8.586826621732099e-06, + "loss": 0.5137, + "step": 3008 + }, + { + "epoch": 0.9627773037542662, + "grad_norm": 0.4923568665981293, + "learning_rate": 8.58552926583048e-06, + "loss": 0.5288, + "step": 3009 + }, + { + "epoch": 0.9630972696245734, + "grad_norm": 0.5681024789810181, + "learning_rate": 8.584231412790915e-06, + "loss": 0.5577, + "step": 3010 + }, + { + "epoch": 0.9634172354948806, + "grad_norm": 0.6538505554199219, + "learning_rate": 8.582933062793353e-06, + "loss": 0.5595, + "step": 3011 + }, + { + "epoch": 0.9637372013651877, + "grad_norm": 0.6075664758682251, + "learning_rate": 8.581634216017812e-06, + "loss": 0.585, + "step": 3012 + }, + { + "epoch": 0.9640571672354948, + "grad_norm": 0.5595016479492188, + "learning_rate": 8.580334872644379e-06, + "loss": 0.5538, + "step": 3013 + }, + { + "epoch": 0.9643771331058021, + "grad_norm": 0.532981812953949, + "learning_rate": 8.579035032853207e-06, + "loss": 0.5603, + "step": 3014 + }, + { + "epoch": 0.9646970989761092, + "grad_norm": 0.6255946159362793, + "learning_rate": 8.577734696824523e-06, + "loss": 0.5478, + "step": 3015 + }, + { + "epoch": 0.9650170648464164, + "grad_norm": 0.5752984881401062, + "learning_rate": 8.576433864738618e-06, + "loss": 0.5203, + "step": 3016 + }, + { + "epoch": 0.9653370307167235, + "grad_norm": 0.5297114849090576, + "learning_rate": 8.575132536775854e-06, + "loss": 0.5227, + "step": 3017 + }, + { + "epoch": 0.9656569965870307, + "grad_norm": 0.581031322479248, + "learning_rate": 8.573830713116663e-06, + "loss": 0.5667, + "step": 3018 + }, + { + "epoch": 0.9659769624573379, + "grad_norm": 0.5882019996643066, + "learning_rate": 8.572528393941547e-06, + "loss": 0.5482, + "step": 3019 + }, + { + "epoch": 0.966296928327645, + "grad_norm": 0.5825169086456299, + "learning_rate": 8.57122557943107e-06, + "loss": 0.5323, + "step": 3020 + }, + { + "epoch": 0.9666168941979523, + "grad_norm": 0.5378094911575317, + "learning_rate": 8.56992226976587e-06, + "loss": 0.536, + "step": 3021 + }, + { + "epoch": 0.9669368600682594, + "grad_norm": 0.5341657400131226, + "learning_rate": 8.568618465126653e-06, + "loss": 0.5173, + "step": 3022 + }, + { + "epoch": 0.9672568259385665, + "grad_norm": 0.5237258076667786, + "learning_rate": 8.567314165694192e-06, + "loss": 0.5211, + "step": 3023 + }, + { + "epoch": 0.9675767918088737, + "grad_norm": 0.5663238763809204, + "learning_rate": 8.566009371649331e-06, + "loss": 0.5642, + "step": 3024 + }, + { + "epoch": 0.9678967576791809, + "grad_norm": 0.5667411684989929, + "learning_rate": 8.56470408317298e-06, + "loss": 0.5888, + "step": 3025 + }, + { + "epoch": 0.9682167235494881, + "grad_norm": 0.5694569945335388, + "learning_rate": 8.56339830044612e-06, + "loss": 0.5472, + "step": 3026 + }, + { + "epoch": 0.9685366894197952, + "grad_norm": 0.5204916596412659, + "learning_rate": 8.562092023649797e-06, + "loss": 0.5311, + "step": 3027 + }, + { + "epoch": 0.9688566552901023, + "grad_norm": 0.5056095123291016, + "learning_rate": 8.560785252965131e-06, + "loss": 0.4955, + "step": 3028 + }, + { + "epoch": 0.9691766211604096, + "grad_norm": 0.5372504591941833, + "learning_rate": 8.559477988573305e-06, + "loss": 0.5407, + "step": 3029 + }, + { + "epoch": 0.9694965870307167, + "grad_norm": 0.5474112033843994, + "learning_rate": 8.558170230655576e-06, + "loss": 0.5968, + "step": 3030 + }, + { + "epoch": 0.9698165529010239, + "grad_norm": 0.5350422859191895, + "learning_rate": 8.556861979393263e-06, + "loss": 0.5182, + "step": 3031 + }, + { + "epoch": 0.9701365187713311, + "grad_norm": 0.546489417552948, + "learning_rate": 8.555553234967757e-06, + "loss": 0.5249, + "step": 3032 + }, + { + "epoch": 0.9704564846416383, + "grad_norm": 0.5644913911819458, + "learning_rate": 8.554243997560517e-06, + "loss": 0.5295, + "step": 3033 + }, + { + "epoch": 0.9707764505119454, + "grad_norm": 0.673241913318634, + "learning_rate": 8.552934267353072e-06, + "loss": 0.541, + "step": 3034 + }, + { + "epoch": 0.9710964163822525, + "grad_norm": 0.6079564094543457, + "learning_rate": 8.551624044527016e-06, + "loss": 0.5735, + "step": 3035 + }, + { + "epoch": 0.9714163822525598, + "grad_norm": 0.5430492162704468, + "learning_rate": 8.550313329264015e-06, + "loss": 0.5315, + "step": 3036 + }, + { + "epoch": 0.9717363481228669, + "grad_norm": 0.5448910593986511, + "learning_rate": 8.549002121745798e-06, + "loss": 0.5193, + "step": 3037 + }, + { + "epoch": 0.972056313993174, + "grad_norm": 0.568170964717865, + "learning_rate": 8.547690422154167e-06, + "loss": 0.5585, + "step": 3038 + }, + { + "epoch": 0.9723762798634812, + "grad_norm": 0.536199152469635, + "learning_rate": 8.546378230670992e-06, + "loss": 0.5278, + "step": 3039 + }, + { + "epoch": 0.9726962457337884, + "grad_norm": 0.6083328723907471, + "learning_rate": 8.545065547478209e-06, + "loss": 0.5808, + "step": 3040 + }, + { + "epoch": 0.9730162116040956, + "grad_norm": 0.5674001574516296, + "learning_rate": 8.543752372757822e-06, + "loss": 0.5648, + "step": 3041 + }, + { + "epoch": 0.9733361774744027, + "grad_norm": 0.5693421363830566, + "learning_rate": 8.542438706691906e-06, + "loss": 0.5582, + "step": 3042 + }, + { + "epoch": 0.9736561433447098, + "grad_norm": 0.44867393374443054, + "learning_rate": 8.541124549462601e-06, + "loss": 0.5147, + "step": 3043 + }, + { + "epoch": 0.9739761092150171, + "grad_norm": 0.6155474185943604, + "learning_rate": 8.539809901252118e-06, + "loss": 0.5725, + "step": 3044 + }, + { + "epoch": 0.9742960750853242, + "grad_norm": 0.4862614870071411, + "learning_rate": 8.538494762242733e-06, + "loss": 0.558, + "step": 3045 + }, + { + "epoch": 0.9746160409556314, + "grad_norm": 0.45874154567718506, + "learning_rate": 8.537179132616794e-06, + "loss": 0.5087, + "step": 3046 + }, + { + "epoch": 0.9749360068259386, + "grad_norm": 0.5131855607032776, + "learning_rate": 8.53586301255671e-06, + "loss": 0.49, + "step": 3047 + }, + { + "epoch": 0.9752559726962458, + "grad_norm": 0.5835806727409363, + "learning_rate": 8.534546402244968e-06, + "loss": 0.5597, + "step": 3048 + }, + { + "epoch": 0.9755759385665529, + "grad_norm": 0.559389054775238, + "learning_rate": 8.533229301864114e-06, + "loss": 0.5037, + "step": 3049 + }, + { + "epoch": 0.97589590443686, + "grad_norm": 0.5053697824478149, + "learning_rate": 8.531911711596767e-06, + "loss": 0.5171, + "step": 3050 + }, + { + "epoch": 0.9762158703071673, + "grad_norm": 0.586311936378479, + "learning_rate": 8.530593631625611e-06, + "loss": 0.5929, + "step": 3051 + }, + { + "epoch": 0.9765358361774744, + "grad_norm": 0.5995824933052063, + "learning_rate": 8.529275062133404e-06, + "loss": 0.5303, + "step": 3052 + }, + { + "epoch": 0.9768558020477816, + "grad_norm": 0.5481569170951843, + "learning_rate": 8.527956003302961e-06, + "loss": 0.5559, + "step": 3053 + }, + { + "epoch": 0.9771757679180887, + "grad_norm": 0.5199539065361023, + "learning_rate": 8.526636455317174e-06, + "loss": 0.5406, + "step": 3054 + }, + { + "epoch": 0.9774957337883959, + "grad_norm": 0.5552107095718384, + "learning_rate": 8.525316418359e-06, + "loss": 0.5756, + "step": 3055 + }, + { + "epoch": 0.9778156996587031, + "grad_norm": 0.47679877281188965, + "learning_rate": 8.523995892611465e-06, + "loss": 0.5354, + "step": 3056 + }, + { + "epoch": 0.9781356655290102, + "grad_norm": 0.5298394560813904, + "learning_rate": 8.522674878257658e-06, + "loss": 0.5243, + "step": 3057 + }, + { + "epoch": 0.9784556313993175, + "grad_norm": 0.5738407373428345, + "learning_rate": 8.521353375480743e-06, + "loss": 0.5903, + "step": 3058 + }, + { + "epoch": 0.9787755972696246, + "grad_norm": 0.6375570893287659, + "learning_rate": 8.520031384463945e-06, + "loss": 0.6056, + "step": 3059 + }, + { + "epoch": 0.9790955631399317, + "grad_norm": 0.512746274471283, + "learning_rate": 8.518708905390562e-06, + "loss": 0.5555, + "step": 3060 + }, + { + "epoch": 0.9794155290102389, + "grad_norm": 0.544510006904602, + "learning_rate": 8.517385938443955e-06, + "loss": 0.5579, + "step": 3061 + }, + { + "epoch": 0.9797354948805461, + "grad_norm": 0.5213636755943298, + "learning_rate": 8.516062483807556e-06, + "loss": 0.5501, + "step": 3062 + }, + { + "epoch": 0.9800554607508533, + "grad_norm": 0.5775883793830872, + "learning_rate": 8.514738541664865e-06, + "loss": 0.5589, + "step": 3063 + }, + { + "epoch": 0.9803754266211604, + "grad_norm": 0.5627555251121521, + "learning_rate": 8.513414112199445e-06, + "loss": 0.5033, + "step": 3064 + }, + { + "epoch": 0.9806953924914675, + "grad_norm": 0.5240724682807922, + "learning_rate": 8.512089195594933e-06, + "loss": 0.5415, + "step": 3065 + }, + { + "epoch": 0.9810153583617748, + "grad_norm": 0.4943949282169342, + "learning_rate": 8.510763792035029e-06, + "loss": 0.5319, + "step": 3066 + }, + { + "epoch": 0.9813353242320819, + "grad_norm": 0.5783194303512573, + "learning_rate": 8.509437901703501e-06, + "loss": 0.5346, + "step": 3067 + }, + { + "epoch": 0.981655290102389, + "grad_norm": 0.5179439187049866, + "learning_rate": 8.508111524784186e-06, + "loss": 0.52, + "step": 3068 + }, + { + "epoch": 0.9819752559726962, + "grad_norm": 0.557386577129364, + "learning_rate": 8.506784661460987e-06, + "loss": 0.5643, + "step": 3069 + }, + { + "epoch": 0.9822952218430034, + "grad_norm": 0.5388240814208984, + "learning_rate": 8.505457311917878e-06, + "loss": 0.5375, + "step": 3070 + }, + { + "epoch": 0.9826151877133106, + "grad_norm": 0.5537760257720947, + "learning_rate": 8.504129476338893e-06, + "loss": 0.588, + "step": 3071 + }, + { + "epoch": 0.9829351535836177, + "grad_norm": 0.5869219303131104, + "learning_rate": 8.502801154908142e-06, + "loss": 0.6043, + "step": 3072 + }, + { + "epoch": 0.983255119453925, + "grad_norm": 0.5651918649673462, + "learning_rate": 8.501472347809799e-06, + "loss": 0.5446, + "step": 3073 + }, + { + "epoch": 0.9835750853242321, + "grad_norm": 0.5868792533874512, + "learning_rate": 8.500143055228098e-06, + "loss": 0.5518, + "step": 3074 + }, + { + "epoch": 0.9838950511945392, + "grad_norm": 0.5043837428092957, + "learning_rate": 8.498813277347355e-06, + "loss": 0.5071, + "step": 3075 + }, + { + "epoch": 0.9842150170648464, + "grad_norm": 0.5354411602020264, + "learning_rate": 8.497483014351941e-06, + "loss": 0.5535, + "step": 3076 + }, + { + "epoch": 0.9845349829351536, + "grad_norm": 0.5286612510681152, + "learning_rate": 8.4961522664263e-06, + "loss": 0.5606, + "step": 3077 + }, + { + "epoch": 0.9848549488054608, + "grad_norm": 0.5994420647621155, + "learning_rate": 8.49482103375494e-06, + "loss": 0.5705, + "step": 3078 + }, + { + "epoch": 0.9851749146757679, + "grad_norm": 0.602558970451355, + "learning_rate": 8.49348931652244e-06, + "loss": 0.5416, + "step": 3079 + }, + { + "epoch": 0.985494880546075, + "grad_norm": 0.556376039981842, + "learning_rate": 8.492157114913444e-06, + "loss": 0.5471, + "step": 3080 + }, + { + "epoch": 0.9858148464163823, + "grad_norm": 0.5554435849189758, + "learning_rate": 8.490824429112664e-06, + "loss": 0.508, + "step": 3081 + }, + { + "epoch": 0.9861348122866894, + "grad_norm": 0.540307343006134, + "learning_rate": 8.489491259304874e-06, + "loss": 0.5097, + "step": 3082 + }, + { + "epoch": 0.9864547781569966, + "grad_norm": 0.5240446925163269, + "learning_rate": 8.488157605674924e-06, + "loss": 0.5212, + "step": 3083 + }, + { + "epoch": 0.9867747440273038, + "grad_norm": 0.5979939699172974, + "learning_rate": 8.486823468407727e-06, + "loss": 0.5393, + "step": 3084 + }, + { + "epoch": 0.987094709897611, + "grad_norm": 0.5482702851295471, + "learning_rate": 8.48548884768826e-06, + "loss": 0.5371, + "step": 3085 + }, + { + "epoch": 0.9874146757679181, + "grad_norm": 0.5796502828598022, + "learning_rate": 8.484153743701572e-06, + "loss": 0.5188, + "step": 3086 + }, + { + "epoch": 0.9877346416382252, + "grad_norm": 0.5995625853538513, + "learning_rate": 8.482818156632776e-06, + "loss": 0.5373, + "step": 3087 + }, + { + "epoch": 0.9880546075085325, + "grad_norm": 0.6242316365242004, + "learning_rate": 8.481482086667052e-06, + "loss": 0.5614, + "step": 3088 + }, + { + "epoch": 0.9883745733788396, + "grad_norm": 0.5484504103660583, + "learning_rate": 8.480145533989648e-06, + "loss": 0.55, + "step": 3089 + }, + { + "epoch": 0.9886945392491467, + "grad_norm": 0.5537928342819214, + "learning_rate": 8.478808498785879e-06, + "loss": 0.4946, + "step": 3090 + }, + { + "epoch": 0.9890145051194539, + "grad_norm": 0.5466088056564331, + "learning_rate": 8.47747098124113e-06, + "loss": 0.5628, + "step": 3091 + }, + { + "epoch": 0.9893344709897611, + "grad_norm": 0.6240554451942444, + "learning_rate": 8.476132981540842e-06, + "loss": 0.5386, + "step": 3092 + }, + { + "epoch": 0.9896544368600683, + "grad_norm": 0.6110613346099854, + "learning_rate": 8.474794499870535e-06, + "loss": 0.5548, + "step": 3093 + }, + { + "epoch": 0.9899744027303754, + "grad_norm": 0.5479440689086914, + "learning_rate": 8.47345553641579e-06, + "loss": 0.5283, + "step": 3094 + }, + { + "epoch": 0.9902943686006825, + "grad_norm": 0.4982355237007141, + "learning_rate": 8.472116091362256e-06, + "loss": 0.5143, + "step": 3095 + }, + { + "epoch": 0.9906143344709898, + "grad_norm": 0.577256441116333, + "learning_rate": 8.47077616489565e-06, + "loss": 0.5033, + "step": 3096 + }, + { + "epoch": 0.9909343003412969, + "grad_norm": 0.6515082120895386, + "learning_rate": 8.469435757201754e-06, + "loss": 0.6009, + "step": 3097 + }, + { + "epoch": 0.9912542662116041, + "grad_norm": 0.5402945280075073, + "learning_rate": 8.468094868466416e-06, + "loss": 0.542, + "step": 3098 + }, + { + "epoch": 0.9915742320819113, + "grad_norm": 0.5361272096633911, + "learning_rate": 8.466753498875551e-06, + "loss": 0.5378, + "step": 3099 + }, + { + "epoch": 0.9918941979522184, + "grad_norm": 0.6215559244155884, + "learning_rate": 8.465411648615144e-06, + "loss": 0.5756, + "step": 3100 + }, + { + "epoch": 0.9922141638225256, + "grad_norm": 0.540403425693512, + "learning_rate": 8.464069317871242e-06, + "loss": 0.5224, + "step": 3101 + }, + { + "epoch": 0.9925341296928327, + "grad_norm": 0.6040542125701904, + "learning_rate": 8.46272650682996e-06, + "loss": 0.5539, + "step": 3102 + }, + { + "epoch": 0.99285409556314, + "grad_norm": 0.5860587954521179, + "learning_rate": 8.461383215677486e-06, + "loss": 0.5314, + "step": 3103 + }, + { + "epoch": 0.9931740614334471, + "grad_norm": 0.5857371687889099, + "learning_rate": 8.460039444600063e-06, + "loss": 0.564, + "step": 3104 + }, + { + "epoch": 0.9934940273037542, + "grad_norm": 0.5509449243545532, + "learning_rate": 8.45869519378401e-06, + "loss": 0.5033, + "step": 3105 + }, + { + "epoch": 0.9938139931740614, + "grad_norm": 0.5352454781532288, + "learning_rate": 8.457350463415706e-06, + "loss": 0.5561, + "step": 3106 + }, + { + "epoch": 0.9941339590443686, + "grad_norm": 0.5762051343917847, + "learning_rate": 8.456005253681601e-06, + "loss": 0.5428, + "step": 3107 + }, + { + "epoch": 0.9944539249146758, + "grad_norm": 0.5870938897132874, + "learning_rate": 8.45465956476821e-06, + "loss": 0.507, + "step": 3108 + }, + { + "epoch": 0.9947738907849829, + "grad_norm": 0.5523510575294495, + "learning_rate": 8.453313396862113e-06, + "loss": 0.5233, + "step": 3109 + }, + { + "epoch": 0.9950938566552902, + "grad_norm": 0.5425226092338562, + "learning_rate": 8.45196675014996e-06, + "loss": 0.5139, + "step": 3110 + }, + { + "epoch": 0.9954138225255973, + "grad_norm": 0.6074098348617554, + "learning_rate": 8.450619624818465e-06, + "loss": 0.5366, + "step": 3111 + }, + { + "epoch": 0.9957337883959044, + "grad_norm": 0.6013935208320618, + "learning_rate": 8.449272021054407e-06, + "loss": 0.5685, + "step": 3112 + }, + { + "epoch": 0.9960537542662116, + "grad_norm": 0.5533202290534973, + "learning_rate": 8.447923939044636e-06, + "loss": 0.506, + "step": 3113 + }, + { + "epoch": 0.9963737201365188, + "grad_norm": 0.5740543603897095, + "learning_rate": 8.44657537897606e-06, + "loss": 0.5556, + "step": 3114 + }, + { + "epoch": 0.996693686006826, + "grad_norm": 0.539481520652771, + "learning_rate": 8.445226341035662e-06, + "loss": 0.5381, + "step": 3115 + }, + { + "epoch": 0.9970136518771331, + "grad_norm": 0.5156455636024475, + "learning_rate": 8.443876825410488e-06, + "loss": 0.5027, + "step": 3116 + }, + { + "epoch": 0.9973336177474402, + "grad_norm": 0.5651331543922424, + "learning_rate": 8.442526832287649e-06, + "loss": 0.5773, + "step": 3117 + }, + { + "epoch": 0.9976535836177475, + "grad_norm": 0.5460453629493713, + "learning_rate": 8.441176361854322e-06, + "loss": 0.5315, + "step": 3118 + }, + { + "epoch": 0.9979735494880546, + "grad_norm": 0.5436340570449829, + "learning_rate": 8.439825414297755e-06, + "loss": 0.5222, + "step": 3119 + }, + { + "epoch": 0.9982935153583617, + "grad_norm": 0.5273920893669128, + "learning_rate": 8.438473989805253e-06, + "loss": 0.5286, + "step": 3120 + }, + { + "epoch": 0.9986134812286689, + "grad_norm": 0.5575355291366577, + "learning_rate": 8.437122088564197e-06, + "loss": 0.5435, + "step": 3121 + }, + { + "epoch": 0.9989334470989761, + "grad_norm": 0.5628740191459656, + "learning_rate": 8.43576971076203e-06, + "loss": 0.5183, + "step": 3122 + }, + { + "epoch": 0.9992534129692833, + "grad_norm": 0.48700088262557983, + "learning_rate": 8.434416856586258e-06, + "loss": 0.5431, + "step": 3123 + }, + { + "epoch": 0.9995733788395904, + "grad_norm": 0.5751681923866272, + "learning_rate": 8.433063526224456e-06, + "loss": 0.5581, + "step": 3124 + }, + { + "epoch": 0.9998933447098977, + "grad_norm": 0.5555768609046936, + "learning_rate": 8.431709719864268e-06, + "loss": 0.5391, + "step": 3125 + }, + { + "epoch": 1.0002133105802047, + "grad_norm": 1.2219892740249634, + "learning_rate": 8.430355437693398e-06, + "loss": 0.9491, + "step": 3126 + }, + { + "epoch": 1.000533276450512, + "grad_norm": 0.4970657229423523, + "learning_rate": 8.42900067989962e-06, + "loss": 0.4525, + "step": 3127 + }, + { + "epoch": 1.0008532423208192, + "grad_norm": 0.5354446172714233, + "learning_rate": 8.427645446670772e-06, + "loss": 0.5274, + "step": 3128 + }, + { + "epoch": 1.0011732081911262, + "grad_norm": 0.5338842272758484, + "learning_rate": 8.426289738194759e-06, + "loss": 0.5127, + "step": 3129 + }, + { + "epoch": 1.0014931740614335, + "grad_norm": 0.5222037434577942, + "learning_rate": 8.424933554659554e-06, + "loss": 0.4923, + "step": 3130 + }, + { + "epoch": 1.0018131399317407, + "grad_norm": 0.5694612264633179, + "learning_rate": 8.42357689625319e-06, + "loss": 0.505, + "step": 3131 + }, + { + "epoch": 1.0021331058020477, + "grad_norm": 0.5298541784286499, + "learning_rate": 8.422219763163769e-06, + "loss": 0.4549, + "step": 3132 + }, + { + "epoch": 1.002453071672355, + "grad_norm": 0.5489261150360107, + "learning_rate": 8.420862155579463e-06, + "loss": 0.5095, + "step": 3133 + }, + { + "epoch": 1.0027730375426622, + "grad_norm": 0.5825126767158508, + "learning_rate": 8.419504073688503e-06, + "loss": 0.455, + "step": 3134 + }, + { + "epoch": 1.0030930034129693, + "grad_norm": 0.5706313848495483, + "learning_rate": 8.418145517679188e-06, + "loss": 0.532, + "step": 3135 + }, + { + "epoch": 1.0034129692832765, + "grad_norm": 0.5857523083686829, + "learning_rate": 8.416786487739888e-06, + "loss": 0.4992, + "step": 3136 + }, + { + "epoch": 1.0037329351535835, + "grad_norm": 0.5802491903305054, + "learning_rate": 8.41542698405903e-06, + "loss": 0.4995, + "step": 3137 + }, + { + "epoch": 1.0040529010238908, + "grad_norm": 0.5400286316871643, + "learning_rate": 8.414067006825108e-06, + "loss": 0.5103, + "step": 3138 + }, + { + "epoch": 1.004372866894198, + "grad_norm": 0.5207788944244385, + "learning_rate": 8.412706556226694e-06, + "loss": 0.4338, + "step": 3139 + }, + { + "epoch": 1.004692832764505, + "grad_norm": 0.5851981043815613, + "learning_rate": 8.411345632452405e-06, + "loss": 0.5164, + "step": 3140 + }, + { + "epoch": 1.0050127986348123, + "grad_norm": 0.6366373300552368, + "learning_rate": 8.409984235690945e-06, + "loss": 0.5192, + "step": 3141 + }, + { + "epoch": 1.0053327645051195, + "grad_norm": 0.5763211846351624, + "learning_rate": 8.408622366131067e-06, + "loss": 0.5273, + "step": 3142 + }, + { + "epoch": 1.0056527303754266, + "grad_norm": 0.6516720652580261, + "learning_rate": 8.407260023961594e-06, + "loss": 0.5475, + "step": 3143 + }, + { + "epoch": 1.0059726962457338, + "grad_norm": 0.5470252633094788, + "learning_rate": 8.405897209371424e-06, + "loss": 0.477, + "step": 3144 + }, + { + "epoch": 1.0062926621160408, + "grad_norm": 0.5717004537582397, + "learning_rate": 8.404533922549506e-06, + "loss": 0.4527, + "step": 3145 + }, + { + "epoch": 1.006612627986348, + "grad_norm": 0.614112913608551, + "learning_rate": 8.403170163684864e-06, + "loss": 0.4917, + "step": 3146 + }, + { + "epoch": 1.0069325938566553, + "grad_norm": 0.5276660919189453, + "learning_rate": 8.401805932966585e-06, + "loss": 0.5293, + "step": 3147 + }, + { + "epoch": 1.0072525597269624, + "grad_norm": 0.5499048829078674, + "learning_rate": 8.400441230583822e-06, + "loss": 0.4858, + "step": 3148 + }, + { + "epoch": 1.0075725255972696, + "grad_norm": 0.6219374537467957, + "learning_rate": 8.39907605672579e-06, + "loss": 0.5287, + "step": 3149 + }, + { + "epoch": 1.0078924914675769, + "grad_norm": 0.5626710057258606, + "learning_rate": 8.397710411581774e-06, + "loss": 0.5166, + "step": 3150 + }, + { + "epoch": 1.008212457337884, + "grad_norm": 0.5865727663040161, + "learning_rate": 8.396344295341124e-06, + "loss": 0.5228, + "step": 3151 + }, + { + "epoch": 1.0085324232081911, + "grad_norm": 0.6077318787574768, + "learning_rate": 8.394977708193248e-06, + "loss": 0.5257, + "step": 3152 + }, + { + "epoch": 1.0088523890784984, + "grad_norm": 0.4857766628265381, + "learning_rate": 8.39361065032763e-06, + "loss": 0.4495, + "step": 3153 + }, + { + "epoch": 1.0091723549488054, + "grad_norm": 0.5425410866737366, + "learning_rate": 8.392243121933815e-06, + "loss": 0.5265, + "step": 3154 + }, + { + "epoch": 1.0094923208191127, + "grad_norm": 0.5726216435432434, + "learning_rate": 8.390875123201408e-06, + "loss": 0.4532, + "step": 3155 + }, + { + "epoch": 1.0098122866894197, + "grad_norm": 0.5400950908660889, + "learning_rate": 8.389506654320085e-06, + "loss": 0.5149, + "step": 3156 + }, + { + "epoch": 1.010132252559727, + "grad_norm": 0.5174579620361328, + "learning_rate": 8.388137715479587e-06, + "loss": 0.5041, + "step": 3157 + }, + { + "epoch": 1.0104522184300342, + "grad_norm": 0.5398826599121094, + "learning_rate": 8.38676830686972e-06, + "loss": 0.524, + "step": 3158 + }, + { + "epoch": 1.0107721843003412, + "grad_norm": 0.5821525454521179, + "learning_rate": 8.385398428680353e-06, + "loss": 0.4714, + "step": 3159 + }, + { + "epoch": 1.0110921501706485, + "grad_norm": 0.5623863935470581, + "learning_rate": 8.384028081101419e-06, + "loss": 0.5151, + "step": 3160 + }, + { + "epoch": 1.0114121160409557, + "grad_norm": 0.5194283127784729, + "learning_rate": 8.382657264322924e-06, + "loss": 0.4981, + "step": 3161 + }, + { + "epoch": 1.0117320819112627, + "grad_norm": 0.5321395397186279, + "learning_rate": 8.381285978534925e-06, + "loss": 0.4672, + "step": 3162 + }, + { + "epoch": 1.01205204778157, + "grad_norm": 0.48655349016189575, + "learning_rate": 8.379914223927562e-06, + "loss": 0.4755, + "step": 3163 + }, + { + "epoch": 1.0123720136518772, + "grad_norm": 0.5503526329994202, + "learning_rate": 8.378542000691024e-06, + "loss": 0.5549, + "step": 3164 + }, + { + "epoch": 1.0126919795221843, + "grad_norm": 0.5182765126228333, + "learning_rate": 8.377169309015571e-06, + "loss": 0.4444, + "step": 3165 + }, + { + "epoch": 1.0130119453924915, + "grad_norm": 0.5637997984886169, + "learning_rate": 8.375796149091534e-06, + "loss": 0.5394, + "step": 3166 + }, + { + "epoch": 1.0133319112627985, + "grad_norm": 0.5851975083351135, + "learning_rate": 8.374422521109298e-06, + "loss": 0.5391, + "step": 3167 + }, + { + "epoch": 1.0136518771331058, + "grad_norm": 0.5644164681434631, + "learning_rate": 8.373048425259319e-06, + "loss": 0.4577, + "step": 3168 + }, + { + "epoch": 1.013971843003413, + "grad_norm": 0.5628954768180847, + "learning_rate": 8.371673861732119e-06, + "loss": 0.4758, + "step": 3169 + }, + { + "epoch": 1.01429180887372, + "grad_norm": 0.5550352931022644, + "learning_rate": 8.370298830718283e-06, + "loss": 0.5186, + "step": 3170 + }, + { + "epoch": 1.0146117747440273, + "grad_norm": 0.5640912652015686, + "learning_rate": 8.368923332408458e-06, + "loss": 0.487, + "step": 3171 + }, + { + "epoch": 1.0149317406143346, + "grad_norm": 0.6037182211875916, + "learning_rate": 8.36754736699336e-06, + "loss": 0.5129, + "step": 3172 + }, + { + "epoch": 1.0152517064846416, + "grad_norm": 0.6077131628990173, + "learning_rate": 8.366170934663768e-06, + "loss": 0.5372, + "step": 3173 + }, + { + "epoch": 1.0155716723549488, + "grad_norm": 0.5096414089202881, + "learning_rate": 8.364794035610527e-06, + "loss": 0.5031, + "step": 3174 + }, + { + "epoch": 1.015891638225256, + "grad_norm": 0.5736592411994934, + "learning_rate": 8.363416670024546e-06, + "loss": 0.4667, + "step": 3175 + }, + { + "epoch": 1.016211604095563, + "grad_norm": 0.6692733764648438, + "learning_rate": 8.362038838096797e-06, + "loss": 0.5971, + "step": 3176 + }, + { + "epoch": 1.0165315699658704, + "grad_norm": 0.555644154548645, + "learning_rate": 8.360660540018317e-06, + "loss": 0.4759, + "step": 3177 + }, + { + "epoch": 1.0168515358361774, + "grad_norm": 0.5540698766708374, + "learning_rate": 8.35928177598021e-06, + "loss": 0.4867, + "step": 3178 + }, + { + "epoch": 1.0171715017064846, + "grad_norm": 0.5888321399688721, + "learning_rate": 8.357902546173645e-06, + "loss": 0.5535, + "step": 3179 + }, + { + "epoch": 1.0174914675767919, + "grad_norm": 0.516242265701294, + "learning_rate": 8.356522850789852e-06, + "loss": 0.4709, + "step": 3180 + }, + { + "epoch": 1.017811433447099, + "grad_norm": 0.5233179926872253, + "learning_rate": 8.355142690020128e-06, + "loss": 0.5062, + "step": 3181 + }, + { + "epoch": 1.0181313993174061, + "grad_norm": 0.5436132550239563, + "learning_rate": 8.353762064055832e-06, + "loss": 0.535, + "step": 3182 + }, + { + "epoch": 1.0184513651877134, + "grad_norm": 0.5038325190544128, + "learning_rate": 8.352380973088394e-06, + "loss": 0.4532, + "step": 3183 + }, + { + "epoch": 1.0187713310580204, + "grad_norm": 0.5422745943069458, + "learning_rate": 8.350999417309298e-06, + "loss": 0.4596, + "step": 3184 + }, + { + "epoch": 1.0190912969283277, + "grad_norm": 0.5560320019721985, + "learning_rate": 8.349617396910104e-06, + "loss": 0.4887, + "step": 3185 + }, + { + "epoch": 1.019411262798635, + "grad_norm": 0.5979790687561035, + "learning_rate": 8.348234912082427e-06, + "loss": 0.4871, + "step": 3186 + }, + { + "epoch": 1.019731228668942, + "grad_norm": 0.5235424637794495, + "learning_rate": 8.346851963017952e-06, + "loss": 0.4893, + "step": 3187 + }, + { + "epoch": 1.0200511945392492, + "grad_norm": 0.5454728007316589, + "learning_rate": 8.345468549908425e-06, + "loss": 0.5642, + "step": 3188 + }, + { + "epoch": 1.0203711604095562, + "grad_norm": 0.46851009130477905, + "learning_rate": 8.344084672945659e-06, + "loss": 0.5052, + "step": 3189 + }, + { + "epoch": 1.0206911262798635, + "grad_norm": 0.531676709651947, + "learning_rate": 8.342700332321531e-06, + "loss": 0.4974, + "step": 3190 + }, + { + "epoch": 1.0210110921501707, + "grad_norm": 0.5864760279655457, + "learning_rate": 8.34131552822798e-06, + "loss": 0.5055, + "step": 3191 + }, + { + "epoch": 1.0213310580204777, + "grad_norm": 0.5224139094352722, + "learning_rate": 8.339930260857011e-06, + "loss": 0.5447, + "step": 3192 + }, + { + "epoch": 1.021651023890785, + "grad_norm": 0.5960366725921631, + "learning_rate": 8.338544530400693e-06, + "loss": 0.5089, + "step": 3193 + }, + { + "epoch": 1.0219709897610922, + "grad_norm": 0.5359147787094116, + "learning_rate": 8.337158337051161e-06, + "loss": 0.5099, + "step": 3194 + }, + { + "epoch": 1.0222909556313993, + "grad_norm": 0.48335564136505127, + "learning_rate": 8.33577168100061e-06, + "loss": 0.4905, + "step": 3195 + }, + { + "epoch": 1.0226109215017065, + "grad_norm": 0.5697606801986694, + "learning_rate": 8.334384562441302e-06, + "loss": 0.47, + "step": 3196 + }, + { + "epoch": 1.0229308873720138, + "grad_norm": 0.5768377780914307, + "learning_rate": 8.332996981565564e-06, + "loss": 0.4789, + "step": 3197 + }, + { + "epoch": 1.0232508532423208, + "grad_norm": 0.5663846731185913, + "learning_rate": 8.331608938565782e-06, + "loss": 0.5589, + "step": 3198 + }, + { + "epoch": 1.023570819112628, + "grad_norm": 0.5420123934745789, + "learning_rate": 8.330220433634416e-06, + "loss": 0.5203, + "step": 3199 + }, + { + "epoch": 1.023890784982935, + "grad_norm": 0.49670645594596863, + "learning_rate": 8.32883146696398e-06, + "loss": 0.4672, + "step": 3200 + }, + { + "epoch": 1.0242107508532423, + "grad_norm": 0.5433145761489868, + "learning_rate": 8.327442038747055e-06, + "loss": 0.4974, + "step": 3201 + }, + { + "epoch": 1.0245307167235496, + "grad_norm": 0.567774772644043, + "learning_rate": 8.32605214917629e-06, + "loss": 0.5517, + "step": 3202 + }, + { + "epoch": 1.0248506825938566, + "grad_norm": 0.5511053204536438, + "learning_rate": 8.324661798444391e-06, + "loss": 0.4643, + "step": 3203 + }, + { + "epoch": 1.0251706484641638, + "grad_norm": 0.5236759781837463, + "learning_rate": 8.323270986744137e-06, + "loss": 0.5015, + "step": 3204 + }, + { + "epoch": 1.025490614334471, + "grad_norm": 0.5212015509605408, + "learning_rate": 8.321879714268361e-06, + "loss": 0.5431, + "step": 3205 + }, + { + "epoch": 1.025810580204778, + "grad_norm": 0.5333899855613708, + "learning_rate": 8.320487981209966e-06, + "loss": 0.4933, + "step": 3206 + }, + { + "epoch": 1.0261305460750854, + "grad_norm": 0.5734988451004028, + "learning_rate": 8.319095787761918e-06, + "loss": 0.5144, + "step": 3207 + }, + { + "epoch": 1.0264505119453924, + "grad_norm": 0.5054695010185242, + "learning_rate": 8.317703134117245e-06, + "loss": 0.4651, + "step": 3208 + }, + { + "epoch": 1.0267704778156996, + "grad_norm": 0.5281773209571838, + "learning_rate": 8.316310020469043e-06, + "loss": 0.5086, + "step": 3209 + }, + { + "epoch": 1.0270904436860069, + "grad_norm": 0.556444525718689, + "learning_rate": 8.314916447010467e-06, + "loss": 0.472, + "step": 3210 + }, + { + "epoch": 1.027410409556314, + "grad_norm": 0.4962175488471985, + "learning_rate": 8.313522413934736e-06, + "loss": 0.4704, + "step": 3211 + }, + { + "epoch": 1.0277303754266212, + "grad_norm": 0.589185357093811, + "learning_rate": 8.312127921435137e-06, + "loss": 0.5101, + "step": 3212 + }, + { + "epoch": 1.0280503412969284, + "grad_norm": 0.5513185858726501, + "learning_rate": 8.310732969705018e-06, + "loss": 0.4972, + "step": 3213 + }, + { + "epoch": 1.0283703071672354, + "grad_norm": 0.5592458844184875, + "learning_rate": 8.309337558937789e-06, + "loss": 0.4905, + "step": 3214 + }, + { + "epoch": 1.0286902730375427, + "grad_norm": 0.5382454991340637, + "learning_rate": 8.307941689326926e-06, + "loss": 0.4726, + "step": 3215 + }, + { + "epoch": 1.02901023890785, + "grad_norm": 0.6426695585250854, + "learning_rate": 8.306545361065968e-06, + "loss": 0.5357, + "step": 3216 + }, + { + "epoch": 1.029330204778157, + "grad_norm": 0.5265448093414307, + "learning_rate": 8.305148574348519e-06, + "loss": 0.4265, + "step": 3217 + }, + { + "epoch": 1.0296501706484642, + "grad_norm": 0.5393176078796387, + "learning_rate": 8.303751329368242e-06, + "loss": 0.5238, + "step": 3218 + }, + { + "epoch": 1.0299701365187712, + "grad_norm": 0.5781740546226501, + "learning_rate": 8.30235362631887e-06, + "loss": 0.5115, + "step": 3219 + }, + { + "epoch": 1.0302901023890785, + "grad_norm": 0.5424593687057495, + "learning_rate": 8.300955465394197e-06, + "loss": 0.4784, + "step": 3220 + }, + { + "epoch": 1.0306100682593857, + "grad_norm": 0.5782645344734192, + "learning_rate": 8.299556846788074e-06, + "loss": 0.5157, + "step": 3221 + }, + { + "epoch": 1.0309300341296928, + "grad_norm": 0.5042377710342407, + "learning_rate": 8.298157770694427e-06, + "loss": 0.4234, + "step": 3222 + }, + { + "epoch": 1.03125, + "grad_norm": 0.5630826354026794, + "learning_rate": 8.296758237307234e-06, + "loss": 0.5474, + "step": 3223 + }, + { + "epoch": 1.0315699658703072, + "grad_norm": 0.5753233432769775, + "learning_rate": 8.295358246820548e-06, + "loss": 0.4971, + "step": 3224 + }, + { + "epoch": 1.0318899317406143, + "grad_norm": 0.5464937090873718, + "learning_rate": 8.293957799428477e-06, + "loss": 0.4847, + "step": 3225 + }, + { + "epoch": 1.0322098976109215, + "grad_norm": 0.5086600184440613, + "learning_rate": 8.292556895325195e-06, + "loss": 0.5046, + "step": 3226 + }, + { + "epoch": 1.0325298634812288, + "grad_norm": 0.5309995412826538, + "learning_rate": 8.291155534704937e-06, + "loss": 0.4632, + "step": 3227 + }, + { + "epoch": 1.0328498293515358, + "grad_norm": 0.5247173309326172, + "learning_rate": 8.289753717762004e-06, + "loss": 0.4877, + "step": 3228 + }, + { + "epoch": 1.033169795221843, + "grad_norm": 0.5032480955123901, + "learning_rate": 8.28835144469076e-06, + "loss": 0.5124, + "step": 3229 + }, + { + "epoch": 1.03348976109215, + "grad_norm": 0.5265656113624573, + "learning_rate": 8.286948715685635e-06, + "loss": 0.5028, + "step": 3230 + }, + { + "epoch": 1.0338097269624573, + "grad_norm": 0.5498801469802856, + "learning_rate": 8.285545530941114e-06, + "loss": 0.4715, + "step": 3231 + }, + { + "epoch": 1.0341296928327646, + "grad_norm": 0.5443584322929382, + "learning_rate": 8.284141890651754e-06, + "loss": 0.4765, + "step": 3232 + }, + { + "epoch": 1.0344496587030716, + "grad_norm": 0.5465605854988098, + "learning_rate": 8.282737795012169e-06, + "loss": 0.4875, + "step": 3233 + }, + { + "epoch": 1.0347696245733788, + "grad_norm": 0.5705323815345764, + "learning_rate": 8.281333244217041e-06, + "loss": 0.5012, + "step": 3234 + }, + { + "epoch": 1.035089590443686, + "grad_norm": 0.5770228505134583, + "learning_rate": 8.279928238461108e-06, + "loss": 0.5451, + "step": 3235 + }, + { + "epoch": 1.0354095563139931, + "grad_norm": 0.5116875767707825, + "learning_rate": 8.278522777939181e-06, + "loss": 0.423, + "step": 3236 + }, + { + "epoch": 1.0357295221843004, + "grad_norm": 0.5620558857917786, + "learning_rate": 8.277116862846127e-06, + "loss": 0.5358, + "step": 3237 + }, + { + "epoch": 1.0360494880546076, + "grad_norm": 0.5042481422424316, + "learning_rate": 8.275710493376876e-06, + "loss": 0.4488, + "step": 3238 + }, + { + "epoch": 1.0363694539249146, + "grad_norm": 0.5645015239715576, + "learning_rate": 8.274303669726427e-06, + "loss": 0.51, + "step": 3239 + }, + { + "epoch": 1.036689419795222, + "grad_norm": 0.5250040888786316, + "learning_rate": 8.272896392089833e-06, + "loss": 0.5062, + "step": 3240 + }, + { + "epoch": 1.037009385665529, + "grad_norm": 0.528319239616394, + "learning_rate": 8.271488660662217e-06, + "loss": 0.4766, + "step": 3241 + }, + { + "epoch": 1.0373293515358362, + "grad_norm": 0.5562976598739624, + "learning_rate": 8.270080475638762e-06, + "loss": 0.4982, + "step": 3242 + }, + { + "epoch": 1.0376493174061434, + "grad_norm": 0.5111921429634094, + "learning_rate": 8.268671837214717e-06, + "loss": 0.472, + "step": 3243 + }, + { + "epoch": 1.0379692832764504, + "grad_norm": 0.5360894203186035, + "learning_rate": 8.267262745585387e-06, + "loss": 0.5228, + "step": 3244 + }, + { + "epoch": 1.0382892491467577, + "grad_norm": 0.543817937374115, + "learning_rate": 8.265853200946147e-06, + "loss": 0.4887, + "step": 3245 + }, + { + "epoch": 1.038609215017065, + "grad_norm": 0.5790839791297913, + "learning_rate": 8.264443203492435e-06, + "loss": 0.4692, + "step": 3246 + }, + { + "epoch": 1.038929180887372, + "grad_norm": 0.5253699421882629, + "learning_rate": 8.263032753419741e-06, + "loss": 0.5429, + "step": 3247 + }, + { + "epoch": 1.0392491467576792, + "grad_norm": 0.48694974184036255, + "learning_rate": 8.261621850923634e-06, + "loss": 0.5038, + "step": 3248 + }, + { + "epoch": 1.0395691126279862, + "grad_norm": 0.5342028737068176, + "learning_rate": 8.260210496199732e-06, + "loss": 0.4863, + "step": 3249 + }, + { + "epoch": 1.0398890784982935, + "grad_norm": 0.5062534213066101, + "learning_rate": 8.258798689443724e-06, + "loss": 0.4391, + "step": 3250 + }, + { + "epoch": 1.0402090443686007, + "grad_norm": 0.613639771938324, + "learning_rate": 8.257386430851355e-06, + "loss": 0.5303, + "step": 3251 + }, + { + "epoch": 1.0405290102389078, + "grad_norm": 0.5553945302963257, + "learning_rate": 8.255973720618438e-06, + "loss": 0.5305, + "step": 3252 + }, + { + "epoch": 1.040848976109215, + "grad_norm": 0.5541419982910156, + "learning_rate": 8.25456055894085e-06, + "loss": 0.5362, + "step": 3253 + }, + { + "epoch": 1.0411689419795223, + "grad_norm": 0.552528977394104, + "learning_rate": 8.253146946014525e-06, + "loss": 0.499, + "step": 3254 + }, + { + "epoch": 1.0414889078498293, + "grad_norm": 0.5524160861968994, + "learning_rate": 8.25173288203546e-06, + "loss": 0.5571, + "step": 3255 + }, + { + "epoch": 1.0418088737201365, + "grad_norm": 0.5329165458679199, + "learning_rate": 8.250318367199722e-06, + "loss": 0.4503, + "step": 3256 + }, + { + "epoch": 1.0421288395904438, + "grad_norm": 0.5469471216201782, + "learning_rate": 8.24890340170343e-06, + "loss": 0.473, + "step": 3257 + }, + { + "epoch": 1.0424488054607508, + "grad_norm": 0.5729274749755859, + "learning_rate": 8.247487985742774e-06, + "loss": 0.4979, + "step": 3258 + }, + { + "epoch": 1.042768771331058, + "grad_norm": 0.5663588047027588, + "learning_rate": 8.246072119514002e-06, + "loss": 0.5106, + "step": 3259 + }, + { + "epoch": 1.0430887372013653, + "grad_norm": 0.5898635983467102, + "learning_rate": 8.244655803213425e-06, + "loss": 0.4891, + "step": 3260 + }, + { + "epoch": 1.0434087030716723, + "grad_norm": 0.5540909171104431, + "learning_rate": 8.243239037037418e-06, + "loss": 0.5196, + "step": 3261 + }, + { + "epoch": 1.0437286689419796, + "grad_norm": 0.5461093783378601, + "learning_rate": 8.241821821182417e-06, + "loss": 0.4959, + "step": 3262 + }, + { + "epoch": 1.0440486348122866, + "grad_norm": 0.5747105479240417, + "learning_rate": 8.240404155844919e-06, + "loss": 0.4654, + "step": 3263 + }, + { + "epoch": 1.0443686006825939, + "grad_norm": 0.5178041458129883, + "learning_rate": 8.238986041221485e-06, + "loss": 0.4613, + "step": 3264 + }, + { + "epoch": 1.044688566552901, + "grad_norm": 0.5070105791091919, + "learning_rate": 8.237567477508744e-06, + "loss": 0.4415, + "step": 3265 + }, + { + "epoch": 1.0450085324232081, + "grad_norm": 0.5689228177070618, + "learning_rate": 8.236148464903374e-06, + "loss": 0.4882, + "step": 3266 + }, + { + "epoch": 1.0453284982935154, + "grad_norm": 0.6333630084991455, + "learning_rate": 8.234729003602128e-06, + "loss": 0.5739, + "step": 3267 + }, + { + "epoch": 1.0456484641638226, + "grad_norm": 0.5439204573631287, + "learning_rate": 8.233309093801815e-06, + "loss": 0.5014, + "step": 3268 + }, + { + "epoch": 1.0459684300341296, + "grad_norm": 0.673212468624115, + "learning_rate": 8.231888735699305e-06, + "loss": 0.501, + "step": 3269 + }, + { + "epoch": 1.046288395904437, + "grad_norm": 0.5605301260948181, + "learning_rate": 8.230467929491533e-06, + "loss": 0.4604, + "step": 3270 + }, + { + "epoch": 1.046608361774744, + "grad_norm": 0.49820882081985474, + "learning_rate": 8.229046675375498e-06, + "loss": 0.4852, + "step": 3271 + }, + { + "epoch": 1.0469283276450512, + "grad_norm": 0.5804089307785034, + "learning_rate": 8.227624973548256e-06, + "loss": 0.486, + "step": 3272 + }, + { + "epoch": 1.0472482935153584, + "grad_norm": 0.6037084460258484, + "learning_rate": 8.226202824206929e-06, + "loss": 0.4991, + "step": 3273 + }, + { + "epoch": 1.0475682593856654, + "grad_norm": 0.567973792552948, + "learning_rate": 8.224780227548698e-06, + "loss": 0.5378, + "step": 3274 + }, + { + "epoch": 1.0478882252559727, + "grad_norm": 0.5975142121315002, + "learning_rate": 8.22335718377081e-06, + "loss": 0.4967, + "step": 3275 + }, + { + "epoch": 1.04820819112628, + "grad_norm": 0.5783261060714722, + "learning_rate": 8.221933693070569e-06, + "loss": 0.4535, + "step": 3276 + }, + { + "epoch": 1.048528156996587, + "grad_norm": 0.5568932890892029, + "learning_rate": 8.220509755645348e-06, + "loss": 0.5058, + "step": 3277 + }, + { + "epoch": 1.0488481228668942, + "grad_norm": 0.524492084980011, + "learning_rate": 8.219085371692573e-06, + "loss": 0.4632, + "step": 3278 + }, + { + "epoch": 1.0491680887372015, + "grad_norm": 0.534509003162384, + "learning_rate": 8.21766054140974e-06, + "loss": 0.4929, + "step": 3279 + }, + { + "epoch": 1.0494880546075085, + "grad_norm": 0.5752255916595459, + "learning_rate": 8.216235264994402e-06, + "loss": 0.544, + "step": 3280 + }, + { + "epoch": 1.0498080204778157, + "grad_norm": 0.5233433842658997, + "learning_rate": 8.214809542644173e-06, + "loss": 0.4828, + "step": 3281 + }, + { + "epoch": 1.0501279863481228, + "grad_norm": 0.6086320281028748, + "learning_rate": 8.213383374556736e-06, + "loss": 0.4538, + "step": 3282 + }, + { + "epoch": 1.05044795221843, + "grad_norm": 0.609158456325531, + "learning_rate": 8.211956760929827e-06, + "loss": 0.4731, + "step": 3283 + }, + { + "epoch": 1.0507679180887373, + "grad_norm": 0.5908544659614563, + "learning_rate": 8.210529701961248e-06, + "loss": 0.5263, + "step": 3284 + }, + { + "epoch": 1.0510878839590443, + "grad_norm": 0.6520845293998718, + "learning_rate": 8.209102197848866e-06, + "loss": 0.5137, + "step": 3285 + }, + { + "epoch": 1.0514078498293515, + "grad_norm": 0.6237879395484924, + "learning_rate": 8.207674248790602e-06, + "loss": 0.4624, + "step": 3286 + }, + { + "epoch": 1.0517278156996588, + "grad_norm": 0.5166109800338745, + "learning_rate": 8.206245854984445e-06, + "loss": 0.5023, + "step": 3287 + }, + { + "epoch": 1.0520477815699658, + "grad_norm": 0.5419546961784363, + "learning_rate": 8.20481701662844e-06, + "loss": 0.4643, + "step": 3288 + }, + { + "epoch": 1.052367747440273, + "grad_norm": 0.6240190863609314, + "learning_rate": 8.203387733920704e-06, + "loss": 0.5036, + "step": 3289 + }, + { + "epoch": 1.0526877133105803, + "grad_norm": 0.6003880500793457, + "learning_rate": 8.201958007059405e-06, + "loss": 0.5088, + "step": 3290 + }, + { + "epoch": 1.0530076791808873, + "grad_norm": 0.5665394067764282, + "learning_rate": 8.200527836242775e-06, + "loss": 0.5574, + "step": 3291 + }, + { + "epoch": 1.0533276450511946, + "grad_norm": 0.4948168098926544, + "learning_rate": 8.19909722166911e-06, + "loss": 0.4787, + "step": 3292 + }, + { + "epoch": 1.0536476109215016, + "grad_norm": 0.5346024632453918, + "learning_rate": 8.197666163536768e-06, + "loss": 0.468, + "step": 3293 + }, + { + "epoch": 1.0539675767918089, + "grad_norm": 0.535876989364624, + "learning_rate": 8.196234662044164e-06, + "loss": 0.4379, + "step": 3294 + }, + { + "epoch": 1.054287542662116, + "grad_norm": 0.5247263312339783, + "learning_rate": 8.19480271738978e-06, + "loss": 0.5081, + "step": 3295 + }, + { + "epoch": 1.0546075085324231, + "grad_norm": 0.5306622385978699, + "learning_rate": 8.193370329772154e-06, + "loss": 0.4969, + "step": 3296 + }, + { + "epoch": 1.0549274744027304, + "grad_norm": 0.49779438972473145, + "learning_rate": 8.191937499389892e-06, + "loss": 0.4608, + "step": 3297 + }, + { + "epoch": 1.0552474402730376, + "grad_norm": 0.555260181427002, + "learning_rate": 8.190504226441654e-06, + "loss": 0.5152, + "step": 3298 + }, + { + "epoch": 1.0555674061433447, + "grad_norm": 0.5562401413917542, + "learning_rate": 8.18907051112617e-06, + "loss": 0.5202, + "step": 3299 + }, + { + "epoch": 1.055887372013652, + "grad_norm": 0.49704766273498535, + "learning_rate": 8.187636353642218e-06, + "loss": 0.4844, + "step": 3300 + }, + { + "epoch": 1.0562073378839592, + "grad_norm": 0.5293060541152954, + "learning_rate": 8.186201754188655e-06, + "loss": 0.4755, + "step": 3301 + }, + { + "epoch": 1.0565273037542662, + "grad_norm": 0.5252447128295898, + "learning_rate": 8.184766712964385e-06, + "loss": 0.5046, + "step": 3302 + }, + { + "epoch": 1.0568472696245734, + "grad_norm": 0.4727516174316406, + "learning_rate": 8.183331230168376e-06, + "loss": 0.4682, + "step": 3303 + }, + { + "epoch": 1.0571672354948805, + "grad_norm": 0.5746709108352661, + "learning_rate": 8.181895305999665e-06, + "loss": 0.5164, + "step": 3304 + }, + { + "epoch": 1.0574872013651877, + "grad_norm": 0.5763015747070312, + "learning_rate": 8.180458940657343e-06, + "loss": 0.5427, + "step": 3305 + }, + { + "epoch": 1.057807167235495, + "grad_norm": 0.5482608079910278, + "learning_rate": 8.17902213434056e-06, + "loss": 0.463, + "step": 3306 + }, + { + "epoch": 1.058127133105802, + "grad_norm": 0.5401831269264221, + "learning_rate": 8.177584887248536e-06, + "loss": 0.5213, + "step": 3307 + }, + { + "epoch": 1.0584470989761092, + "grad_norm": 0.5248244404792786, + "learning_rate": 8.176147199580542e-06, + "loss": 0.5038, + "step": 3308 + }, + { + "epoch": 1.0587670648464165, + "grad_norm": 0.5097712278366089, + "learning_rate": 8.17470907153592e-06, + "loss": 0.4701, + "step": 3309 + }, + { + "epoch": 1.0590870307167235, + "grad_norm": 0.47569549083709717, + "learning_rate": 8.173270503314063e-06, + "loss": 0.4556, + "step": 3310 + }, + { + "epoch": 1.0594069965870307, + "grad_norm": 0.5200427770614624, + "learning_rate": 8.171831495114436e-06, + "loss": 0.4838, + "step": 3311 + }, + { + "epoch": 1.0597269624573378, + "grad_norm": 0.5401687026023865, + "learning_rate": 8.170392047136555e-06, + "loss": 0.5279, + "step": 3312 + }, + { + "epoch": 1.060046928327645, + "grad_norm": 0.5490127205848694, + "learning_rate": 8.168952159580004e-06, + "loss": 0.5255, + "step": 3313 + }, + { + "epoch": 1.0603668941979523, + "grad_norm": 0.5770564675331116, + "learning_rate": 8.167511832644423e-06, + "loss": 0.4994, + "step": 3314 + }, + { + "epoch": 1.0606868600682593, + "grad_norm": 0.5831413269042969, + "learning_rate": 8.166071066529515e-06, + "loss": 0.5379, + "step": 3315 + }, + { + "epoch": 1.0610068259385665, + "grad_norm": 0.5414169430732727, + "learning_rate": 8.164629861435047e-06, + "loss": 0.4899, + "step": 3316 + }, + { + "epoch": 1.0613267918088738, + "grad_norm": 0.5539081692695618, + "learning_rate": 8.16318821756084e-06, + "loss": 0.4502, + "step": 3317 + }, + { + "epoch": 1.0616467576791808, + "grad_norm": 0.5647209882736206, + "learning_rate": 8.16174613510678e-06, + "loss": 0.5359, + "step": 3318 + }, + { + "epoch": 1.061966723549488, + "grad_norm": 0.49853506684303284, + "learning_rate": 8.160303614272817e-06, + "loss": 0.4405, + "step": 3319 + }, + { + "epoch": 1.0622866894197953, + "grad_norm": 0.483504056930542, + "learning_rate": 8.158860655258955e-06, + "loss": 0.4475, + "step": 3320 + }, + { + "epoch": 1.0626066552901023, + "grad_norm": 0.5061872005462646, + "learning_rate": 8.157417258265263e-06, + "loss": 0.4942, + "step": 3321 + }, + { + "epoch": 1.0629266211604096, + "grad_norm": 0.5668355226516724, + "learning_rate": 8.155973423491868e-06, + "loss": 0.5019, + "step": 3322 + }, + { + "epoch": 1.0632465870307168, + "grad_norm": 0.6143890023231506, + "learning_rate": 8.154529151138964e-06, + "loss": 0.5261, + "step": 3323 + }, + { + "epoch": 1.0635665529010239, + "grad_norm": 0.4885425269603729, + "learning_rate": 8.153084441406797e-06, + "loss": 0.4456, + "step": 3324 + }, + { + "epoch": 1.0638865187713311, + "grad_norm": 0.5878362059593201, + "learning_rate": 8.151639294495678e-06, + "loss": 0.5305, + "step": 3325 + }, + { + "epoch": 1.0642064846416381, + "grad_norm": 0.5964034199714661, + "learning_rate": 8.15019371060598e-06, + "loss": 0.5334, + "step": 3326 + }, + { + "epoch": 1.0645264505119454, + "grad_norm": 0.5398755669593811, + "learning_rate": 8.148747689938135e-06, + "loss": 0.4528, + "step": 3327 + }, + { + "epoch": 1.0648464163822526, + "grad_norm": 0.5904126167297363, + "learning_rate": 8.147301232692634e-06, + "loss": 0.5541, + "step": 3328 + }, + { + "epoch": 1.0651663822525597, + "grad_norm": 0.54847651720047, + "learning_rate": 8.14585433907003e-06, + "loss": 0.536, + "step": 3329 + }, + { + "epoch": 1.065486348122867, + "grad_norm": 0.5282127857208252, + "learning_rate": 8.144407009270939e-06, + "loss": 0.5024, + "step": 3330 + }, + { + "epoch": 1.0658063139931742, + "grad_norm": 0.5941870808601379, + "learning_rate": 8.142959243496032e-06, + "loss": 0.4766, + "step": 3331 + }, + { + "epoch": 1.0661262798634812, + "grad_norm": 0.519917368888855, + "learning_rate": 8.141511041946046e-06, + "loss": 0.5628, + "step": 3332 + }, + { + "epoch": 1.0664462457337884, + "grad_norm": 0.5521838665008545, + "learning_rate": 8.140062404821773e-06, + "loss": 0.5019, + "step": 3333 + }, + { + "epoch": 1.0667662116040955, + "grad_norm": 0.643412709236145, + "learning_rate": 8.138613332324072e-06, + "loss": 0.5178, + "step": 3334 + }, + { + "epoch": 1.0670861774744027, + "grad_norm": 0.6367138028144836, + "learning_rate": 8.137163824653855e-06, + "loss": 0.4894, + "step": 3335 + }, + { + "epoch": 1.06740614334471, + "grad_norm": 0.5044212341308594, + "learning_rate": 8.135713882012102e-06, + "loss": 0.4632, + "step": 3336 + }, + { + "epoch": 1.067726109215017, + "grad_norm": 0.602066159248352, + "learning_rate": 8.134263504599844e-06, + "loss": 0.4942, + "step": 3337 + }, + { + "epoch": 1.0680460750853242, + "grad_norm": 0.6120120286941528, + "learning_rate": 8.132812692618181e-06, + "loss": 0.5182, + "step": 3338 + }, + { + "epoch": 1.0683660409556315, + "grad_norm": 0.5152328014373779, + "learning_rate": 8.131361446268271e-06, + "loss": 0.4354, + "step": 3339 + }, + { + "epoch": 1.0686860068259385, + "grad_norm": 0.5362467169761658, + "learning_rate": 8.129909765751326e-06, + "loss": 0.5225, + "step": 3340 + }, + { + "epoch": 1.0690059726962458, + "grad_norm": 0.5970868468284607, + "learning_rate": 8.128457651268631e-06, + "loss": 0.5388, + "step": 3341 + }, + { + "epoch": 1.069325938566553, + "grad_norm": 0.53805011510849, + "learning_rate": 8.127005103021516e-06, + "loss": 0.4494, + "step": 3342 + }, + { + "epoch": 1.06964590443686, + "grad_norm": 0.5005940794944763, + "learning_rate": 8.125552121211385e-06, + "loss": 0.5064, + "step": 3343 + }, + { + "epoch": 1.0699658703071673, + "grad_norm": 0.537519097328186, + "learning_rate": 8.124098706039687e-06, + "loss": 0.5224, + "step": 3344 + }, + { + "epoch": 1.0702858361774743, + "grad_norm": 0.5686160922050476, + "learning_rate": 8.122644857707949e-06, + "loss": 0.5255, + "step": 3345 + }, + { + "epoch": 1.0706058020477816, + "grad_norm": 0.5152590870857239, + "learning_rate": 8.121190576417744e-06, + "loss": 0.5161, + "step": 3346 + }, + { + "epoch": 1.0709257679180888, + "grad_norm": 0.5477904081344604, + "learning_rate": 8.11973586237071e-06, + "loss": 0.5362, + "step": 3347 + }, + { + "epoch": 1.0712457337883958, + "grad_norm": 0.5535791516304016, + "learning_rate": 8.118280715768546e-06, + "loss": 0.4732, + "step": 3348 + }, + { + "epoch": 1.071565699658703, + "grad_norm": 0.5651800632476807, + "learning_rate": 8.116825136813009e-06, + "loss": 0.5736, + "step": 3349 + }, + { + "epoch": 1.0718856655290103, + "grad_norm": 0.5503290295600891, + "learning_rate": 8.115369125705919e-06, + "loss": 0.4364, + "step": 3350 + }, + { + "epoch": 1.0722056313993173, + "grad_norm": 0.5717608332633972, + "learning_rate": 8.11391268264915e-06, + "loss": 0.4905, + "step": 3351 + }, + { + "epoch": 1.0725255972696246, + "grad_norm": 0.6119275689125061, + "learning_rate": 8.112455807844642e-06, + "loss": 0.4943, + "step": 3352 + }, + { + "epoch": 1.0728455631399316, + "grad_norm": 0.5548709034919739, + "learning_rate": 8.110998501494391e-06, + "loss": 0.4723, + "step": 3353 + }, + { + "epoch": 1.0731655290102389, + "grad_norm": 0.5616167783737183, + "learning_rate": 8.109540763800454e-06, + "loss": 0.5413, + "step": 3354 + }, + { + "epoch": 1.0734854948805461, + "grad_norm": 0.5117178559303284, + "learning_rate": 8.108082594964951e-06, + "loss": 0.4356, + "step": 3355 + }, + { + "epoch": 1.0738054607508531, + "grad_norm": 0.6019683480262756, + "learning_rate": 8.106623995190058e-06, + "loss": 0.5237, + "step": 3356 + }, + { + "epoch": 1.0741254266211604, + "grad_norm": 0.5925198197364807, + "learning_rate": 8.105164964678009e-06, + "loss": 0.6085, + "step": 3357 + }, + { + "epoch": 1.0744453924914676, + "grad_norm": 0.5321702361106873, + "learning_rate": 8.103705503631103e-06, + "loss": 0.4576, + "step": 3358 + }, + { + "epoch": 1.0747653583617747, + "grad_norm": 0.578241229057312, + "learning_rate": 8.102245612251696e-06, + "loss": 0.5489, + "step": 3359 + }, + { + "epoch": 1.075085324232082, + "grad_norm": 0.609102189540863, + "learning_rate": 8.100785290742202e-06, + "loss": 0.5153, + "step": 3360 + }, + { + "epoch": 1.0754052901023892, + "grad_norm": 0.5429154634475708, + "learning_rate": 8.099324539305096e-06, + "loss": 0.5183, + "step": 3361 + }, + { + "epoch": 1.0757252559726962, + "grad_norm": 0.5318170785903931, + "learning_rate": 8.097863358142914e-06, + "loss": 0.4384, + "step": 3362 + }, + { + "epoch": 1.0760452218430034, + "grad_norm": 0.5602610111236572, + "learning_rate": 8.096401747458253e-06, + "loss": 0.4774, + "step": 3363 + }, + { + "epoch": 1.0763651877133107, + "grad_norm": 0.6025781631469727, + "learning_rate": 8.094939707453762e-06, + "loss": 0.5688, + "step": 3364 + }, + { + "epoch": 1.0766851535836177, + "grad_norm": 0.5115510821342468, + "learning_rate": 8.09347723833216e-06, + "loss": 0.4436, + "step": 3365 + }, + { + "epoch": 1.077005119453925, + "grad_norm": 0.5866360664367676, + "learning_rate": 8.092014340296215e-06, + "loss": 0.5131, + "step": 3366 + }, + { + "epoch": 1.077325085324232, + "grad_norm": 0.5072371363639832, + "learning_rate": 8.090551013548762e-06, + "loss": 0.4412, + "step": 3367 + }, + { + "epoch": 1.0776450511945392, + "grad_norm": 0.525693953037262, + "learning_rate": 8.089087258292693e-06, + "loss": 0.4848, + "step": 3368 + }, + { + "epoch": 1.0779650170648465, + "grad_norm": 0.5036314129829407, + "learning_rate": 8.08762307473096e-06, + "loss": 0.5591, + "step": 3369 + }, + { + "epoch": 1.0782849829351535, + "grad_norm": 0.5551751852035522, + "learning_rate": 8.086158463066574e-06, + "loss": 0.4882, + "step": 3370 + }, + { + "epoch": 1.0786049488054608, + "grad_norm": 0.6021180748939514, + "learning_rate": 8.084693423502602e-06, + "loss": 0.5101, + "step": 3371 + }, + { + "epoch": 1.078924914675768, + "grad_norm": 0.5439380407333374, + "learning_rate": 8.083227956242179e-06, + "loss": 0.4716, + "step": 3372 + }, + { + "epoch": 1.079244880546075, + "grad_norm": 0.5912663340568542, + "learning_rate": 8.081762061488489e-06, + "loss": 0.515, + "step": 3373 + }, + { + "epoch": 1.0795648464163823, + "grad_norm": 0.5152432322502136, + "learning_rate": 8.080295739444782e-06, + "loss": 0.4317, + "step": 3374 + }, + { + "epoch": 1.0798848122866893, + "grad_norm": 0.5591632127761841, + "learning_rate": 8.078828990314364e-06, + "loss": 0.5434, + "step": 3375 + }, + { + "epoch": 1.0802047781569966, + "grad_norm": 0.5563610792160034, + "learning_rate": 8.077361814300602e-06, + "loss": 0.5485, + "step": 3376 + }, + { + "epoch": 1.0805247440273038, + "grad_norm": 0.5657890439033508, + "learning_rate": 8.075894211606925e-06, + "loss": 0.4926, + "step": 3377 + }, + { + "epoch": 1.0808447098976108, + "grad_norm": 0.5712653994560242, + "learning_rate": 8.074426182436812e-06, + "loss": 0.4851, + "step": 3378 + }, + { + "epoch": 1.081164675767918, + "grad_norm": 0.6388885378837585, + "learning_rate": 8.072957726993813e-06, + "loss": 0.55, + "step": 3379 + }, + { + "epoch": 1.0814846416382253, + "grad_norm": 0.5559564232826233, + "learning_rate": 8.071488845481528e-06, + "loss": 0.5251, + "step": 3380 + }, + { + "epoch": 1.0818046075085324, + "grad_norm": 0.48583507537841797, + "learning_rate": 8.070019538103617e-06, + "loss": 0.4192, + "step": 3381 + }, + { + "epoch": 1.0821245733788396, + "grad_norm": 0.6477534174919128, + "learning_rate": 8.068549805063806e-06, + "loss": 0.519, + "step": 3382 + }, + { + "epoch": 1.0824445392491469, + "grad_norm": 0.6126707196235657, + "learning_rate": 8.067079646565872e-06, + "loss": 0.52, + "step": 3383 + }, + { + "epoch": 1.0827645051194539, + "grad_norm": 0.5714107155799866, + "learning_rate": 8.065609062813657e-06, + "loss": 0.4332, + "step": 3384 + }, + { + "epoch": 1.0830844709897611, + "grad_norm": 0.5787184238433838, + "learning_rate": 8.064138054011056e-06, + "loss": 0.4626, + "step": 3385 + }, + { + "epoch": 1.0834044368600684, + "grad_norm": 0.5252422094345093, + "learning_rate": 8.062666620362028e-06, + "loss": 0.545, + "step": 3386 + }, + { + "epoch": 1.0837244027303754, + "grad_norm": 0.5058318972587585, + "learning_rate": 8.061194762070588e-06, + "loss": 0.494, + "step": 3387 + }, + { + "epoch": 1.0840443686006827, + "grad_norm": 0.6317284107208252, + "learning_rate": 8.05972247934081e-06, + "loss": 0.5148, + "step": 3388 + }, + { + "epoch": 1.0843643344709897, + "grad_norm": 0.5603262782096863, + "learning_rate": 8.058249772376831e-06, + "loss": 0.4689, + "step": 3389 + }, + { + "epoch": 1.084684300341297, + "grad_norm": 0.5218910574913025, + "learning_rate": 8.056776641382844e-06, + "loss": 0.495, + "step": 3390 + }, + { + "epoch": 1.0850042662116042, + "grad_norm": 0.5703060626983643, + "learning_rate": 8.055303086563095e-06, + "loss": 0.5145, + "step": 3391 + }, + { + "epoch": 1.0853242320819112, + "grad_norm": 0.5872288942337036, + "learning_rate": 8.0538291081219e-06, + "loss": 0.5014, + "step": 3392 + }, + { + "epoch": 1.0856441979522184, + "grad_norm": 0.5883344411849976, + "learning_rate": 8.052354706263624e-06, + "loss": 0.5359, + "step": 3393 + }, + { + "epoch": 1.0859641638225257, + "grad_norm": 0.5529842376708984, + "learning_rate": 8.050879881192695e-06, + "loss": 0.4936, + "step": 3394 + }, + { + "epoch": 1.0862841296928327, + "grad_norm": 0.5603022575378418, + "learning_rate": 8.0494046331136e-06, + "loss": 0.5348, + "step": 3395 + }, + { + "epoch": 1.08660409556314, + "grad_norm": 0.5294363498687744, + "learning_rate": 8.047928962230886e-06, + "loss": 0.4998, + "step": 3396 + }, + { + "epoch": 1.086924061433447, + "grad_norm": 0.5693865418434143, + "learning_rate": 8.046452868749155e-06, + "loss": 0.4972, + "step": 3397 + }, + { + "epoch": 1.0872440273037542, + "grad_norm": 0.5111364722251892, + "learning_rate": 8.044976352873066e-06, + "loss": 0.4359, + "step": 3398 + }, + { + "epoch": 1.0875639931740615, + "grad_norm": 0.5233534574508667, + "learning_rate": 8.043499414807344e-06, + "loss": 0.5259, + "step": 3399 + }, + { + "epoch": 1.0878839590443685, + "grad_norm": 0.4962422549724579, + "learning_rate": 8.042022054756765e-06, + "loss": 0.4379, + "step": 3400 + }, + { + "epoch": 1.0882039249146758, + "grad_norm": 0.5874216556549072, + "learning_rate": 8.040544272926171e-06, + "loss": 0.5349, + "step": 3401 + }, + { + "epoch": 1.088523890784983, + "grad_norm": 0.6073318123817444, + "learning_rate": 8.039066069520455e-06, + "loss": 0.5709, + "step": 3402 + }, + { + "epoch": 1.08884385665529, + "grad_norm": 0.5323451161384583, + "learning_rate": 8.037587444744573e-06, + "loss": 0.4819, + "step": 3403 + }, + { + "epoch": 1.0891638225255973, + "grad_norm": 0.5540988445281982, + "learning_rate": 8.036108398803536e-06, + "loss": 0.4548, + "step": 3404 + }, + { + "epoch": 1.0894837883959045, + "grad_norm": 0.545438289642334, + "learning_rate": 8.034628931902419e-06, + "loss": 0.4836, + "step": 3405 + }, + { + "epoch": 1.0898037542662116, + "grad_norm": 0.5775328278541565, + "learning_rate": 8.033149044246348e-06, + "loss": 0.5394, + "step": 3406 + }, + { + "epoch": 1.0901237201365188, + "grad_norm": 0.49468064308166504, + "learning_rate": 8.031668736040515e-06, + "loss": 0.4658, + "step": 3407 + }, + { + "epoch": 1.0904436860068258, + "grad_norm": 0.5362755656242371, + "learning_rate": 8.030188007490164e-06, + "loss": 0.5005, + "step": 3408 + }, + { + "epoch": 1.090763651877133, + "grad_norm": 0.5736452341079712, + "learning_rate": 8.028706858800598e-06, + "loss": 0.5454, + "step": 3409 + }, + { + "epoch": 1.0910836177474403, + "grad_norm": 0.5717540383338928, + "learning_rate": 8.027225290177182e-06, + "loss": 0.4439, + "step": 3410 + }, + { + "epoch": 1.0914035836177474, + "grad_norm": 0.5134483575820923, + "learning_rate": 8.025743301825342e-06, + "loss": 0.4817, + "step": 3411 + }, + { + "epoch": 1.0917235494880546, + "grad_norm": 0.488443523645401, + "learning_rate": 8.02426089395055e-06, + "loss": 0.4982, + "step": 3412 + }, + { + "epoch": 1.0920435153583619, + "grad_norm": 0.5470604300498962, + "learning_rate": 8.022778066758348e-06, + "loss": 0.503, + "step": 3413 + }, + { + "epoch": 1.0923634812286689, + "grad_norm": 0.5137178301811218, + "learning_rate": 8.02129482045433e-06, + "loss": 0.5184, + "step": 3414 + }, + { + "epoch": 1.0926834470989761, + "grad_norm": 0.5206054449081421, + "learning_rate": 8.01981115524415e-06, + "loss": 0.5315, + "step": 3415 + }, + { + "epoch": 1.0930034129692832, + "grad_norm": 0.49802935123443604, + "learning_rate": 8.018327071333521e-06, + "loss": 0.4993, + "step": 3416 + }, + { + "epoch": 1.0933233788395904, + "grad_norm": 0.4849085509777069, + "learning_rate": 8.016842568928213e-06, + "loss": 0.4391, + "step": 3417 + }, + { + "epoch": 1.0936433447098977, + "grad_norm": 0.5910800695419312, + "learning_rate": 8.015357648234052e-06, + "loss": 0.5363, + "step": 3418 + }, + { + "epoch": 1.0939633105802047, + "grad_norm": 0.6028000712394714, + "learning_rate": 8.013872309456928e-06, + "loss": 0.507, + "step": 3419 + }, + { + "epoch": 1.094283276450512, + "grad_norm": 0.5333077907562256, + "learning_rate": 8.01238655280278e-06, + "loss": 0.5316, + "step": 3420 + }, + { + "epoch": 1.0946032423208192, + "grad_norm": 0.5566905736923218, + "learning_rate": 8.010900378477612e-06, + "loss": 0.5112, + "step": 3421 + }, + { + "epoch": 1.0949232081911262, + "grad_norm": 0.5991218686103821, + "learning_rate": 8.009413786687486e-06, + "loss": 0.481, + "step": 3422 + }, + { + "epoch": 1.0952431740614335, + "grad_norm": 0.5846459865570068, + "learning_rate": 8.007926777638519e-06, + "loss": 0.5164, + "step": 3423 + }, + { + "epoch": 1.0955631399317407, + "grad_norm": 0.603992760181427, + "learning_rate": 8.006439351536884e-06, + "loss": 0.51, + "step": 3424 + }, + { + "epoch": 1.0958831058020477, + "grad_norm": 0.5503547787666321, + "learning_rate": 8.004951508588817e-06, + "loss": 0.4707, + "step": 3425 + }, + { + "epoch": 1.096203071672355, + "grad_norm": 0.5133087635040283, + "learning_rate": 8.003463249000608e-06, + "loss": 0.5167, + "step": 3426 + }, + { + "epoch": 1.0965230375426622, + "grad_norm": 0.5910055637359619, + "learning_rate": 8.001974572978605e-06, + "loss": 0.485, + "step": 3427 + }, + { + "epoch": 1.0968430034129693, + "grad_norm": 0.5813104510307312, + "learning_rate": 8.000485480729217e-06, + "loss": 0.4728, + "step": 3428 + }, + { + "epoch": 1.0971629692832765, + "grad_norm": 0.5279648303985596, + "learning_rate": 7.998995972458907e-06, + "loss": 0.4882, + "step": 3429 + }, + { + "epoch": 1.0974829351535835, + "grad_norm": 0.5567468404769897, + "learning_rate": 7.997506048374198e-06, + "loss": 0.4974, + "step": 3430 + }, + { + "epoch": 1.0978029010238908, + "grad_norm": 0.5733569264411926, + "learning_rate": 7.996015708681671e-06, + "loss": 0.5255, + "step": 3431 + }, + { + "epoch": 1.098122866894198, + "grad_norm": 0.46804070472717285, + "learning_rate": 7.994524953587959e-06, + "loss": 0.4432, + "step": 3432 + }, + { + "epoch": 1.098442832764505, + "grad_norm": 0.5812718272209167, + "learning_rate": 7.99303378329976e-06, + "loss": 0.5642, + "step": 3433 + }, + { + "epoch": 1.0987627986348123, + "grad_norm": 0.5907002687454224, + "learning_rate": 7.991542198023827e-06, + "loss": 0.5194, + "step": 3434 + }, + { + "epoch": 1.0990827645051195, + "grad_norm": 0.5492275357246399, + "learning_rate": 7.99005019796697e-06, + "loss": 0.4919, + "step": 3435 + }, + { + "epoch": 1.0994027303754266, + "grad_norm": 0.586290717124939, + "learning_rate": 7.988557783336055e-06, + "loss": 0.4975, + "step": 3436 + }, + { + "epoch": 1.0997226962457338, + "grad_norm": 0.6038715839385986, + "learning_rate": 7.987064954338007e-06, + "loss": 0.5042, + "step": 3437 + }, + { + "epoch": 1.1000426621160408, + "grad_norm": 0.5693135857582092, + "learning_rate": 7.985571711179812e-06, + "loss": 0.5071, + "step": 3438 + }, + { + "epoch": 1.100362627986348, + "grad_norm": 0.5785577297210693, + "learning_rate": 7.984078054068505e-06, + "loss": 0.4957, + "step": 3439 + }, + { + "epoch": 1.1006825938566553, + "grad_norm": 0.5374515056610107, + "learning_rate": 7.982583983211187e-06, + "loss": 0.5037, + "step": 3440 + }, + { + "epoch": 1.1010025597269624, + "grad_norm": 0.5297296643257141, + "learning_rate": 7.981089498815013e-06, + "loss": 0.4822, + "step": 3441 + }, + { + "epoch": 1.1013225255972696, + "grad_norm": 0.5651179552078247, + "learning_rate": 7.979594601087191e-06, + "loss": 0.496, + "step": 3442 + }, + { + "epoch": 1.1016424914675769, + "grad_norm": 0.5637895464897156, + "learning_rate": 7.978099290234995e-06, + "loss": 0.4916, + "step": 3443 + }, + { + "epoch": 1.101962457337884, + "grad_norm": 0.5521149039268494, + "learning_rate": 7.976603566465747e-06, + "loss": 0.4709, + "step": 3444 + }, + { + "epoch": 1.1022824232081911, + "grad_norm": 0.4833112359046936, + "learning_rate": 7.975107429986836e-06, + "loss": 0.4566, + "step": 3445 + }, + { + "epoch": 1.1026023890784984, + "grad_norm": 0.5734673142433167, + "learning_rate": 7.973610881005702e-06, + "loss": 0.518, + "step": 3446 + }, + { + "epoch": 1.1029223549488054, + "grad_norm": 0.5697218775749207, + "learning_rate": 7.97211391972984e-06, + "loss": 0.515, + "step": 3447 + }, + { + "epoch": 1.1032423208191127, + "grad_norm": 0.5977779030799866, + "learning_rate": 7.970616546366809e-06, + "loss": 0.5046, + "step": 3448 + }, + { + "epoch": 1.1035622866894197, + "grad_norm": 0.5142936110496521, + "learning_rate": 7.969118761124217e-06, + "loss": 0.4758, + "step": 3449 + }, + { + "epoch": 1.103882252559727, + "grad_norm": 0.5793342590332031, + "learning_rate": 7.96762056420974e-06, + "loss": 0.5176, + "step": 3450 + }, + { + "epoch": 1.1042022184300342, + "grad_norm": 0.5544641613960266, + "learning_rate": 7.9661219558311e-06, + "loss": 0.4681, + "step": 3451 + }, + { + "epoch": 1.1045221843003412, + "grad_norm": 0.5324788689613342, + "learning_rate": 7.964622936196082e-06, + "loss": 0.4927, + "step": 3452 + }, + { + "epoch": 1.1048421501706485, + "grad_norm": 0.5800503492355347, + "learning_rate": 7.963123505512529e-06, + "loss": 0.501, + "step": 3453 + }, + { + "epoch": 1.1051621160409557, + "grad_norm": 0.6092519164085388, + "learning_rate": 7.961623663988336e-06, + "loss": 0.4742, + "step": 3454 + }, + { + "epoch": 1.1054820819112627, + "grad_norm": 0.5879743695259094, + "learning_rate": 7.960123411831459e-06, + "loss": 0.6033, + "step": 3455 + }, + { + "epoch": 1.10580204778157, + "grad_norm": 0.5198939442634583, + "learning_rate": 7.958622749249911e-06, + "loss": 0.4904, + "step": 3456 + }, + { + "epoch": 1.1061220136518772, + "grad_norm": 0.48006105422973633, + "learning_rate": 7.95712167645176e-06, + "loss": 0.4164, + "step": 3457 + }, + { + "epoch": 1.1064419795221843, + "grad_norm": 0.5257875323295593, + "learning_rate": 7.95562019364513e-06, + "loss": 0.4833, + "step": 3458 + }, + { + "epoch": 1.1067619453924915, + "grad_norm": 0.5582365989685059, + "learning_rate": 7.954118301038203e-06, + "loss": 0.521, + "step": 3459 + }, + { + "epoch": 1.1070819112627985, + "grad_norm": 0.5165651440620422, + "learning_rate": 7.952615998839222e-06, + "loss": 0.5118, + "step": 3460 + }, + { + "epoch": 1.1074018771331058, + "grad_norm": 0.5487995147705078, + "learning_rate": 7.95111328725648e-06, + "loss": 0.4862, + "step": 3461 + }, + { + "epoch": 1.107721843003413, + "grad_norm": 0.5739307999610901, + "learning_rate": 7.949610166498329e-06, + "loss": 0.5425, + "step": 3462 + }, + { + "epoch": 1.10804180887372, + "grad_norm": 0.5141011476516724, + "learning_rate": 7.948106636773184e-06, + "loss": 0.4642, + "step": 3463 + }, + { + "epoch": 1.1083617747440273, + "grad_norm": 0.4784911274909973, + "learning_rate": 7.946602698289503e-06, + "loss": 0.4384, + "step": 3464 + }, + { + "epoch": 1.1086817406143346, + "grad_norm": 0.5456899404525757, + "learning_rate": 7.945098351255814e-06, + "loss": 0.5388, + "step": 3465 + }, + { + "epoch": 1.1090017064846416, + "grad_norm": 0.520692765712738, + "learning_rate": 7.943593595880696e-06, + "loss": 0.5118, + "step": 3466 + }, + { + "epoch": 1.1093216723549488, + "grad_norm": 0.5141918063163757, + "learning_rate": 7.942088432372784e-06, + "loss": 0.4171, + "step": 3467 + }, + { + "epoch": 1.109641638225256, + "grad_norm": 0.5195292234420776, + "learning_rate": 7.940582860940771e-06, + "loss": 0.5243, + "step": 3468 + }, + { + "epoch": 1.109961604095563, + "grad_norm": 0.5189877152442932, + "learning_rate": 7.939076881793408e-06, + "loss": 0.4453, + "step": 3469 + }, + { + "epoch": 1.1102815699658704, + "grad_norm": 0.5618454813957214, + "learning_rate": 7.937570495139499e-06, + "loss": 0.5297, + "step": 3470 + }, + { + "epoch": 1.1106015358361774, + "grad_norm": 0.509519636631012, + "learning_rate": 7.936063701187908e-06, + "loss": 0.4521, + "step": 3471 + }, + { + "epoch": 1.1109215017064846, + "grad_norm": 0.5387877225875854, + "learning_rate": 7.934556500147553e-06, + "loss": 0.5053, + "step": 3472 + }, + { + "epoch": 1.1112414675767919, + "grad_norm": 0.5411096811294556, + "learning_rate": 7.933048892227406e-06, + "loss": 0.516, + "step": 3473 + }, + { + "epoch": 1.111561433447099, + "grad_norm": 0.6244606375694275, + "learning_rate": 7.931540877636503e-06, + "loss": 0.5462, + "step": 3474 + }, + { + "epoch": 1.1118813993174061, + "grad_norm": 0.5785732865333557, + "learning_rate": 7.930032456583931e-06, + "loss": 0.4832, + "step": 3475 + }, + { + "epoch": 1.1122013651877134, + "grad_norm": 0.5496439337730408, + "learning_rate": 7.928523629278835e-06, + "loss": 0.5213, + "step": 3476 + }, + { + "epoch": 1.1125213310580204, + "grad_norm": 0.5574624538421631, + "learning_rate": 7.927014395930412e-06, + "loss": 0.4723, + "step": 3477 + }, + { + "epoch": 1.1128412969283277, + "grad_norm": 0.5714210271835327, + "learning_rate": 7.925504756747924e-06, + "loss": 0.5104, + "step": 3478 + }, + { + "epoch": 1.1131612627986347, + "grad_norm": 0.5857492685317993, + "learning_rate": 7.923994711940682e-06, + "loss": 0.6138, + "step": 3479 + }, + { + "epoch": 1.113481228668942, + "grad_norm": 0.5344730615615845, + "learning_rate": 7.922484261718055e-06, + "loss": 0.4384, + "step": 3480 + }, + { + "epoch": 1.1138011945392492, + "grad_norm": 0.528183102607727, + "learning_rate": 7.920973406289469e-06, + "loss": 0.5313, + "step": 3481 + }, + { + "epoch": 1.1141211604095562, + "grad_norm": 0.6133708953857422, + "learning_rate": 7.919462145864405e-06, + "loss": 0.509, + "step": 3482 + }, + { + "epoch": 1.1144411262798635, + "grad_norm": 0.5474709868431091, + "learning_rate": 7.917950480652403e-06, + "loss": 0.4831, + "step": 3483 + }, + { + "epoch": 1.1147610921501707, + "grad_norm": 0.6181241869926453, + "learning_rate": 7.916438410863058e-06, + "loss": 0.5362, + "step": 3484 + }, + { + "epoch": 1.1150810580204777, + "grad_norm": 0.5898253321647644, + "learning_rate": 7.914925936706018e-06, + "loss": 0.49, + "step": 3485 + }, + { + "epoch": 1.115401023890785, + "grad_norm": 0.619894802570343, + "learning_rate": 7.913413058390989e-06, + "loss": 0.504, + "step": 3486 + }, + { + "epoch": 1.1157209897610922, + "grad_norm": 0.5624905228614807, + "learning_rate": 7.911899776127736e-06, + "loss": 0.523, + "step": 3487 + }, + { + "epoch": 1.1160409556313993, + "grad_norm": 0.5253486037254333, + "learning_rate": 7.910386090126076e-06, + "loss": 0.4276, + "step": 3488 + }, + { + "epoch": 1.1163609215017065, + "grad_norm": 0.6467494964599609, + "learning_rate": 7.908872000595882e-06, + "loss": 0.5575, + "step": 3489 + }, + { + "epoch": 1.1166808873720138, + "grad_norm": 0.5251191854476929, + "learning_rate": 7.907357507747087e-06, + "loss": 0.4818, + "step": 3490 + }, + { + "epoch": 1.1170008532423208, + "grad_norm": 0.636448860168457, + "learning_rate": 7.905842611789675e-06, + "loss": 0.5822, + "step": 3491 + }, + { + "epoch": 1.117320819112628, + "grad_norm": 0.5673524141311646, + "learning_rate": 7.904327312933688e-06, + "loss": 0.522, + "step": 3492 + }, + { + "epoch": 1.117640784982935, + "grad_norm": 0.569740355014801, + "learning_rate": 7.902811611389227e-06, + "loss": 0.4544, + "step": 3493 + }, + { + "epoch": 1.1179607508532423, + "grad_norm": 0.613383412361145, + "learning_rate": 7.901295507366445e-06, + "loss": 0.5084, + "step": 3494 + }, + { + "epoch": 1.1182807167235496, + "grad_norm": 0.5107038617134094, + "learning_rate": 7.899779001075547e-06, + "loss": 0.4043, + "step": 3495 + }, + { + "epoch": 1.1186006825938566, + "grad_norm": 0.5148925185203552, + "learning_rate": 7.898262092726805e-06, + "loss": 0.5596, + "step": 3496 + }, + { + "epoch": 1.1189206484641638, + "grad_norm": 0.5359509587287903, + "learning_rate": 7.896744782530535e-06, + "loss": 0.5055, + "step": 3497 + }, + { + "epoch": 1.119240614334471, + "grad_norm": 0.5600014328956604, + "learning_rate": 7.895227070697118e-06, + "loss": 0.4898, + "step": 3498 + }, + { + "epoch": 1.119560580204778, + "grad_norm": 0.5266281962394714, + "learning_rate": 7.893708957436982e-06, + "loss": 0.4846, + "step": 3499 + }, + { + "epoch": 1.1198805460750854, + "grad_norm": 0.5474272966384888, + "learning_rate": 7.892190442960621e-06, + "loss": 0.4776, + "step": 3500 + }, + { + "epoch": 1.1202005119453924, + "grad_norm": 0.5400792956352234, + "learning_rate": 7.890671527478575e-06, + "loss": 0.5099, + "step": 3501 + }, + { + "epoch": 1.1205204778156996, + "grad_norm": 0.528954267501831, + "learning_rate": 7.889152211201442e-06, + "loss": 0.5321, + "step": 3502 + }, + { + "epoch": 1.1208404436860069, + "grad_norm": 0.5811314582824707, + "learning_rate": 7.887632494339881e-06, + "loss": 0.4551, + "step": 3503 + }, + { + "epoch": 1.121160409556314, + "grad_norm": 0.5416238903999329, + "learning_rate": 7.8861123771046e-06, + "loss": 0.4848, + "step": 3504 + }, + { + "epoch": 1.1214803754266212, + "grad_norm": 0.5491629838943481, + "learning_rate": 7.884591859706365e-06, + "loss": 0.5012, + "step": 3505 + }, + { + "epoch": 1.1218003412969284, + "grad_norm": 0.5045695900917053, + "learning_rate": 7.883070942356001e-06, + "loss": 0.4729, + "step": 3506 + }, + { + "epoch": 1.1221203071672354, + "grad_norm": 0.5594595074653625, + "learning_rate": 7.88154962526438e-06, + "loss": 0.524, + "step": 3507 + }, + { + "epoch": 1.1224402730375427, + "grad_norm": 0.5319141745567322, + "learning_rate": 7.880027908642439e-06, + "loss": 0.5206, + "step": 3508 + }, + { + "epoch": 1.12276023890785, + "grad_norm": 0.5062196850776672, + "learning_rate": 7.878505792701162e-06, + "loss": 0.5039, + "step": 3509 + }, + { + "epoch": 1.123080204778157, + "grad_norm": 0.5017806887626648, + "learning_rate": 7.876983277651592e-06, + "loss": 0.4911, + "step": 3510 + }, + { + "epoch": 1.1234001706484642, + "grad_norm": 0.45875391364097595, + "learning_rate": 7.875460363704834e-06, + "loss": 0.4715, + "step": 3511 + }, + { + "epoch": 1.1237201365187712, + "grad_norm": 0.4955386817455292, + "learning_rate": 7.873937051072037e-06, + "loss": 0.5038, + "step": 3512 + }, + { + "epoch": 1.1240401023890785, + "grad_norm": 0.555153489112854, + "learning_rate": 7.872413339964407e-06, + "loss": 0.4824, + "step": 3513 + }, + { + "epoch": 1.1243600682593857, + "grad_norm": 0.5983140468597412, + "learning_rate": 7.870889230593213e-06, + "loss": 0.5636, + "step": 3514 + }, + { + "epoch": 1.1246800341296928, + "grad_norm": 0.5900697112083435, + "learning_rate": 7.869364723169774e-06, + "loss": 0.5056, + "step": 3515 + }, + { + "epoch": 1.125, + "grad_norm": 0.5284932851791382, + "learning_rate": 7.867839817905466e-06, + "loss": 0.4837, + "step": 3516 + }, + { + "epoch": 1.1253199658703072, + "grad_norm": 0.6077491641044617, + "learning_rate": 7.866314515011713e-06, + "loss": 0.5123, + "step": 3517 + }, + { + "epoch": 1.1256399317406143, + "grad_norm": 0.5993654131889343, + "learning_rate": 7.864788814700006e-06, + "loss": 0.5094, + "step": 3518 + }, + { + "epoch": 1.1259598976109215, + "grad_norm": 0.6185998320579529, + "learning_rate": 7.863262717181884e-06, + "loss": 0.548, + "step": 3519 + }, + { + "epoch": 1.1262798634812285, + "grad_norm": 0.5267618894577026, + "learning_rate": 7.861736222668943e-06, + "loss": 0.542, + "step": 3520 + }, + { + "epoch": 1.1265998293515358, + "grad_norm": 0.5616210103034973, + "learning_rate": 7.86020933137283e-06, + "loss": 0.5353, + "step": 3521 + }, + { + "epoch": 1.126919795221843, + "grad_norm": 0.5184577703475952, + "learning_rate": 7.858682043505253e-06, + "loss": 0.4946, + "step": 3522 + }, + { + "epoch": 1.12723976109215, + "grad_norm": 0.5412434339523315, + "learning_rate": 7.857154359277972e-06, + "loss": 0.512, + "step": 3523 + }, + { + "epoch": 1.1275597269624573, + "grad_norm": 0.5075342655181885, + "learning_rate": 7.855626278902799e-06, + "loss": 0.4858, + "step": 3524 + }, + { + "epoch": 1.1278796928327646, + "grad_norm": 0.6242489814758301, + "learning_rate": 7.85409780259161e-06, + "loss": 0.5285, + "step": 3525 + }, + { + "epoch": 1.1281996587030716, + "grad_norm": 0.5738124847412109, + "learning_rate": 7.852568930556328e-06, + "loss": 0.4731, + "step": 3526 + }, + { + "epoch": 1.1285196245733788, + "grad_norm": 0.551703691482544, + "learning_rate": 7.851039663008929e-06, + "loss": 0.5013, + "step": 3527 + }, + { + "epoch": 1.128839590443686, + "grad_norm": 0.5525521039962769, + "learning_rate": 7.849510000161454e-06, + "loss": 0.5382, + "step": 3528 + }, + { + "epoch": 1.1291595563139931, + "grad_norm": 0.5638213157653809, + "learning_rate": 7.847979942225985e-06, + "loss": 0.4689, + "step": 3529 + }, + { + "epoch": 1.1294795221843004, + "grad_norm": 0.5558946132659912, + "learning_rate": 7.846449489414675e-06, + "loss": 0.4681, + "step": 3530 + }, + { + "epoch": 1.1297994880546076, + "grad_norm": 0.583815336227417, + "learning_rate": 7.844918641939717e-06, + "loss": 0.5684, + "step": 3531 + }, + { + "epoch": 1.1301194539249146, + "grad_norm": 0.6019387245178223, + "learning_rate": 7.843387400013367e-06, + "loss": 0.4637, + "step": 3532 + }, + { + "epoch": 1.130439419795222, + "grad_norm": 0.5469242334365845, + "learning_rate": 7.841855763847934e-06, + "loss": 0.5238, + "step": 3533 + }, + { + "epoch": 1.130759385665529, + "grad_norm": 0.5761325359344482, + "learning_rate": 7.84032373365578e-06, + "loss": 0.508, + "step": 3534 + }, + { + "epoch": 1.1310793515358362, + "grad_norm": 0.6610817909240723, + "learning_rate": 7.838791309649324e-06, + "loss": 0.5173, + "step": 3535 + }, + { + "epoch": 1.1313993174061434, + "grad_norm": 0.588883101940155, + "learning_rate": 7.837258492041036e-06, + "loss": 0.4501, + "step": 3536 + }, + { + "epoch": 1.1317192832764504, + "grad_norm": 0.5269465446472168, + "learning_rate": 7.835725281043447e-06, + "loss": 0.5181, + "step": 3537 + }, + { + "epoch": 1.1320392491467577, + "grad_norm": 0.6146968007087708, + "learning_rate": 7.834191676869135e-06, + "loss": 0.5121, + "step": 3538 + }, + { + "epoch": 1.132359215017065, + "grad_norm": 0.568915843963623, + "learning_rate": 7.832657679730738e-06, + "loss": 0.4973, + "step": 3539 + }, + { + "epoch": 1.132679180887372, + "grad_norm": 0.5202858448028564, + "learning_rate": 7.831123289840946e-06, + "loss": 0.4665, + "step": 3540 + }, + { + "epoch": 1.1329991467576792, + "grad_norm": 0.5461797118186951, + "learning_rate": 7.829588507412503e-06, + "loss": 0.4784, + "step": 3541 + }, + { + "epoch": 1.1333191126279862, + "grad_norm": 0.5784962773323059, + "learning_rate": 7.82805333265821e-06, + "loss": 0.4863, + "step": 3542 + }, + { + "epoch": 1.1336390784982935, + "grad_norm": 0.5158717036247253, + "learning_rate": 7.826517765790919e-06, + "loss": 0.4341, + "step": 3543 + }, + { + "epoch": 1.1339590443686007, + "grad_norm": 0.6058448553085327, + "learning_rate": 7.82498180702354e-06, + "loss": 0.5447, + "step": 3544 + }, + { + "epoch": 1.1342790102389078, + "grad_norm": 0.545375645160675, + "learning_rate": 7.823445456569036e-06, + "loss": 0.4691, + "step": 3545 + }, + { + "epoch": 1.134598976109215, + "grad_norm": 0.6026395559310913, + "learning_rate": 7.821908714640421e-06, + "loss": 0.5708, + "step": 3546 + }, + { + "epoch": 1.1349189419795223, + "grad_norm": 0.5096762776374817, + "learning_rate": 7.82037158145077e-06, + "loss": 0.446, + "step": 3547 + }, + { + "epoch": 1.1352389078498293, + "grad_norm": 0.5593512058258057, + "learning_rate": 7.818834057213205e-06, + "loss": 0.4804, + "step": 3548 + }, + { + "epoch": 1.1355588737201365, + "grad_norm": 0.5331721901893616, + "learning_rate": 7.817296142140907e-06, + "loss": 0.526, + "step": 3549 + }, + { + "epoch": 1.1358788395904438, + "grad_norm": 0.5482791066169739, + "learning_rate": 7.81575783644711e-06, + "loss": 0.499, + "step": 3550 + }, + { + "epoch": 1.1361988054607508, + "grad_norm": 0.48665091395378113, + "learning_rate": 7.8142191403451e-06, + "loss": 0.4608, + "step": 3551 + }, + { + "epoch": 1.136518771331058, + "grad_norm": 0.5491358041763306, + "learning_rate": 7.812680054048223e-06, + "loss": 0.5322, + "step": 3552 + }, + { + "epoch": 1.1368387372013653, + "grad_norm": 0.5035708546638489, + "learning_rate": 7.81114057776987e-06, + "loss": 0.4729, + "step": 3553 + }, + { + "epoch": 1.1371587030716723, + "grad_norm": 0.5579684972763062, + "learning_rate": 7.809600711723497e-06, + "loss": 0.533, + "step": 3554 + }, + { + "epoch": 1.1374786689419796, + "grad_norm": 0.5616936683654785, + "learning_rate": 7.808060456122603e-06, + "loss": 0.499, + "step": 3555 + }, + { + "epoch": 1.1377986348122866, + "grad_norm": 0.5179516077041626, + "learning_rate": 7.80651981118075e-06, + "loss": 0.4782, + "step": 3556 + }, + { + "epoch": 1.1381186006825939, + "grad_norm": 0.5304527878761292, + "learning_rate": 7.804978777111549e-06, + "loss": 0.4879, + "step": 3557 + }, + { + "epoch": 1.138438566552901, + "grad_norm": 0.5091323256492615, + "learning_rate": 7.803437354128665e-06, + "loss": 0.4899, + "step": 3558 + }, + { + "epoch": 1.1387585324232081, + "grad_norm": 0.5082446336746216, + "learning_rate": 7.801895542445819e-06, + "loss": 0.5065, + "step": 3559 + }, + { + "epoch": 1.1390784982935154, + "grad_norm": 0.4766141176223755, + "learning_rate": 7.800353342276787e-06, + "loss": 0.469, + "step": 3560 + }, + { + "epoch": 1.1393984641638226, + "grad_norm": 0.5594205856323242, + "learning_rate": 7.798810753835392e-06, + "loss": 0.5247, + "step": 3561 + }, + { + "epoch": 1.1397184300341296, + "grad_norm": 0.5137618780136108, + "learning_rate": 7.797267777335523e-06, + "loss": 0.4822, + "step": 3562 + }, + { + "epoch": 1.140038395904437, + "grad_norm": 0.5178728699684143, + "learning_rate": 7.795724412991108e-06, + "loss": 0.5045, + "step": 3563 + }, + { + "epoch": 1.140358361774744, + "grad_norm": 0.5358498692512512, + "learning_rate": 7.794180661016143e-06, + "loss": 0.4849, + "step": 3564 + }, + { + "epoch": 1.1406783276450512, + "grad_norm": 0.4935292601585388, + "learning_rate": 7.792636521624665e-06, + "loss": 0.4566, + "step": 3565 + }, + { + "epoch": 1.1409982935153584, + "grad_norm": 0.5339186191558838, + "learning_rate": 7.791091995030775e-06, + "loss": 0.5159, + "step": 3566 + }, + { + "epoch": 1.1413182593856654, + "grad_norm": 0.5401326417922974, + "learning_rate": 7.789547081448622e-06, + "loss": 0.4636, + "step": 3567 + }, + { + "epoch": 1.1416382252559727, + "grad_norm": 0.6003397703170776, + "learning_rate": 7.78800178109241e-06, + "loss": 0.5024, + "step": 3568 + }, + { + "epoch": 1.14195819112628, + "grad_norm": 0.5978726744651794, + "learning_rate": 7.786456094176398e-06, + "loss": 0.4962, + "step": 3569 + }, + { + "epoch": 1.142278156996587, + "grad_norm": 0.6008307933807373, + "learning_rate": 7.784910020914895e-06, + "loss": 0.5261, + "step": 3570 + }, + { + "epoch": 1.1425981228668942, + "grad_norm": 0.5893910527229309, + "learning_rate": 7.783363561522269e-06, + "loss": 0.5164, + "step": 3571 + }, + { + "epoch": 1.1429180887372015, + "grad_norm": 0.5838620662689209, + "learning_rate": 7.781816716212935e-06, + "loss": 0.5484, + "step": 3572 + }, + { + "epoch": 1.1432380546075085, + "grad_norm": 0.5458717942237854, + "learning_rate": 7.780269485201366e-06, + "loss": 0.4492, + "step": 3573 + }, + { + "epoch": 1.1435580204778157, + "grad_norm": 0.5298654437065125, + "learning_rate": 7.778721868702091e-06, + "loss": 0.4869, + "step": 3574 + }, + { + "epoch": 1.143877986348123, + "grad_norm": 0.6417195796966553, + "learning_rate": 7.777173866929682e-06, + "loss": 0.5718, + "step": 3575 + }, + { + "epoch": 1.14419795221843, + "grad_norm": 0.5706183910369873, + "learning_rate": 7.775625480098778e-06, + "loss": 0.4595, + "step": 3576 + }, + { + "epoch": 1.1445179180887373, + "grad_norm": 0.5745950937271118, + "learning_rate": 7.774076708424062e-06, + "loss": 0.5629, + "step": 3577 + }, + { + "epoch": 1.1448378839590443, + "grad_norm": 0.5032234787940979, + "learning_rate": 7.772527552120274e-06, + "loss": 0.4597, + "step": 3578 + }, + { + "epoch": 1.1451578498293515, + "grad_norm": 0.5396010279655457, + "learning_rate": 7.770978011402204e-06, + "loss": 0.5109, + "step": 3579 + }, + { + "epoch": 1.1454778156996588, + "grad_norm": 0.5810889601707458, + "learning_rate": 7.769428086484699e-06, + "loss": 0.4488, + "step": 3580 + }, + { + "epoch": 1.1457977815699658, + "grad_norm": 0.6284360289573669, + "learning_rate": 7.767877777582659e-06, + "loss": 0.524, + "step": 3581 + }, + { + "epoch": 1.146117747440273, + "grad_norm": 0.545884370803833, + "learning_rate": 7.766327084911034e-06, + "loss": 0.5034, + "step": 3582 + }, + { + "epoch": 1.14643771331058, + "grad_norm": 0.5152574777603149, + "learning_rate": 7.764776008684832e-06, + "loss": 0.4454, + "step": 3583 + }, + { + "epoch": 1.1467576791808873, + "grad_norm": 0.5437384843826294, + "learning_rate": 7.763224549119111e-06, + "loss": 0.4885, + "step": 3584 + }, + { + "epoch": 1.1470776450511946, + "grad_norm": 0.5401154160499573, + "learning_rate": 7.76167270642898e-06, + "loss": 0.4683, + "step": 3585 + }, + { + "epoch": 1.1473976109215016, + "grad_norm": 0.6209553480148315, + "learning_rate": 7.760120480829606e-06, + "loss": 0.5572, + "step": 3586 + }, + { + "epoch": 1.1477175767918089, + "grad_norm": 0.5519247055053711, + "learning_rate": 7.758567872536208e-06, + "loss": 0.4868, + "step": 3587 + }, + { + "epoch": 1.148037542662116, + "grad_norm": 0.5403838157653809, + "learning_rate": 7.757014881764056e-06, + "loss": 0.4766, + "step": 3588 + }, + { + "epoch": 1.1483575085324231, + "grad_norm": 0.5646221041679382, + "learning_rate": 7.755461508728473e-06, + "loss": 0.5081, + "step": 3589 + }, + { + "epoch": 1.1486774744027304, + "grad_norm": 0.5375046730041504, + "learning_rate": 7.753907753644835e-06, + "loss": 0.4697, + "step": 3590 + }, + { + "epoch": 1.1489974402730376, + "grad_norm": 0.5387952327728271, + "learning_rate": 7.752353616728579e-06, + "loss": 0.4507, + "step": 3591 + }, + { + "epoch": 1.1493174061433447, + "grad_norm": 0.5580173134803772, + "learning_rate": 7.750799098195179e-06, + "loss": 0.51, + "step": 3592 + }, + { + "epoch": 1.149637372013652, + "grad_norm": 0.5314825773239136, + "learning_rate": 7.749244198260175e-06, + "loss": 0.5039, + "step": 3593 + }, + { + "epoch": 1.1499573378839592, + "grad_norm": 0.6043481826782227, + "learning_rate": 7.747688917139158e-06, + "loss": 0.5187, + "step": 3594 + }, + { + "epoch": 1.1502773037542662, + "grad_norm": 0.5296799540519714, + "learning_rate": 7.746133255047764e-06, + "loss": 0.4991, + "step": 3595 + }, + { + "epoch": 1.1505972696245734, + "grad_norm": 0.5214899182319641, + "learning_rate": 7.744577212201692e-06, + "loss": 0.4695, + "step": 3596 + }, + { + "epoch": 1.1509172354948805, + "grad_norm": 0.5552440881729126, + "learning_rate": 7.743020788816688e-06, + "loss": 0.5271, + "step": 3597 + }, + { + "epoch": 1.1512372013651877, + "grad_norm": 0.6425376534461975, + "learning_rate": 7.741463985108551e-06, + "loss": 0.5198, + "step": 3598 + }, + { + "epoch": 1.151557167235495, + "grad_norm": 0.5188428163528442, + "learning_rate": 7.739906801293136e-06, + "loss": 0.4517, + "step": 3599 + }, + { + "epoch": 1.151877133105802, + "grad_norm": 0.5039048790931702, + "learning_rate": 7.738349237586344e-06, + "loss": 0.4634, + "step": 3600 + }, + { + "epoch": 1.1521970989761092, + "grad_norm": 0.6108112931251526, + "learning_rate": 7.736791294204138e-06, + "loss": 0.5629, + "step": 3601 + }, + { + "epoch": 1.1525170648464165, + "grad_norm": 0.6057872176170349, + "learning_rate": 7.735232971362525e-06, + "loss": 0.4689, + "step": 3602 + }, + { + "epoch": 1.1528370307167235, + "grad_norm": 0.6644954681396484, + "learning_rate": 7.733674269277572e-06, + "loss": 0.5443, + "step": 3603 + }, + { + "epoch": 1.1531569965870307, + "grad_norm": 0.5029540061950684, + "learning_rate": 7.73211518816539e-06, + "loss": 0.4546, + "step": 3604 + }, + { + "epoch": 1.1534769624573378, + "grad_norm": 0.6114902496337891, + "learning_rate": 7.730555728242154e-06, + "loss": 0.522, + "step": 3605 + }, + { + "epoch": 1.153796928327645, + "grad_norm": 0.5663495063781738, + "learning_rate": 7.728995889724078e-06, + "loss": 0.5292, + "step": 3606 + }, + { + "epoch": 1.1541168941979523, + "grad_norm": 0.5311685800552368, + "learning_rate": 7.72743567282744e-06, + "loss": 0.4955, + "step": 3607 + }, + { + "epoch": 1.1544368600682593, + "grad_norm": 0.5237186551094055, + "learning_rate": 7.725875077768565e-06, + "loss": 0.483, + "step": 3608 + }, + { + "epoch": 1.1547568259385665, + "grad_norm": 0.5001371502876282, + "learning_rate": 7.724314104763832e-06, + "loss": 0.4528, + "step": 3609 + }, + { + "epoch": 1.1550767918088738, + "grad_norm": 0.556122362613678, + "learning_rate": 7.72275275402967e-06, + "loss": 0.4903, + "step": 3610 + }, + { + "epoch": 1.1553967576791808, + "grad_norm": 0.5611037611961365, + "learning_rate": 7.721191025782563e-06, + "loss": 0.4774, + "step": 3611 + }, + { + "epoch": 1.155716723549488, + "grad_norm": 0.5195589661598206, + "learning_rate": 7.719628920239047e-06, + "loss": 0.4798, + "step": 3612 + }, + { + "epoch": 1.1560366894197953, + "grad_norm": 0.5992692112922668, + "learning_rate": 7.71806643761571e-06, + "loss": 0.5276, + "step": 3613 + }, + { + "epoch": 1.1563566552901023, + "grad_norm": 0.604767918586731, + "learning_rate": 7.71650357812919e-06, + "loss": 0.532, + "step": 3614 + }, + { + "epoch": 1.1566766211604096, + "grad_norm": 0.5662189722061157, + "learning_rate": 7.714940341996182e-06, + "loss": 0.4664, + "step": 3615 + }, + { + "epoch": 1.1569965870307168, + "grad_norm": 0.5372429490089417, + "learning_rate": 7.71337672943343e-06, + "loss": 0.4649, + "step": 3616 + }, + { + "epoch": 1.1573165529010239, + "grad_norm": 0.5533605217933655, + "learning_rate": 7.71181274065773e-06, + "loss": 0.4805, + "step": 3617 + }, + { + "epoch": 1.1576365187713311, + "grad_norm": 0.7256330251693726, + "learning_rate": 7.71024837588593e-06, + "loss": 0.5416, + "step": 3618 + }, + { + "epoch": 1.1579564846416381, + "grad_norm": 0.5054036378860474, + "learning_rate": 7.70868363533493e-06, + "loss": 0.4353, + "step": 3619 + }, + { + "epoch": 1.1582764505119454, + "grad_norm": 0.6028924584388733, + "learning_rate": 7.70711851922169e-06, + "loss": 0.5202, + "step": 3620 + }, + { + "epoch": 1.1585964163822526, + "grad_norm": 0.5548515319824219, + "learning_rate": 7.705553027763208e-06, + "loss": 0.5081, + "step": 3621 + }, + { + "epoch": 1.1589163822525597, + "grad_norm": 0.576492965221405, + "learning_rate": 7.703987161176545e-06, + "loss": 0.4899, + "step": 3622 + }, + { + "epoch": 1.159236348122867, + "grad_norm": 0.5903343558311462, + "learning_rate": 7.702420919678808e-06, + "loss": 0.5091, + "step": 3623 + }, + { + "epoch": 1.159556313993174, + "grad_norm": 0.5502042770385742, + "learning_rate": 7.70085430348716e-06, + "loss": 0.5349, + "step": 3624 + }, + { + "epoch": 1.1598762798634812, + "grad_norm": 0.5461248159408569, + "learning_rate": 7.699287312818812e-06, + "loss": 0.4802, + "step": 3625 + }, + { + "epoch": 1.1601962457337884, + "grad_norm": 0.5498080253601074, + "learning_rate": 7.697719947891032e-06, + "loss": 0.4375, + "step": 3626 + }, + { + "epoch": 1.1605162116040955, + "grad_norm": 0.5582046508789062, + "learning_rate": 7.696152208921137e-06, + "loss": 0.5167, + "step": 3627 + }, + { + "epoch": 1.1608361774744027, + "grad_norm": 0.5449654459953308, + "learning_rate": 7.694584096126493e-06, + "loss": 0.4915, + "step": 3628 + }, + { + "epoch": 1.16115614334471, + "grad_norm": 0.5606305003166199, + "learning_rate": 7.693015609724524e-06, + "loss": 0.5151, + "step": 3629 + }, + { + "epoch": 1.161476109215017, + "grad_norm": 0.5650143623352051, + "learning_rate": 7.6914467499327e-06, + "loss": 0.481, + "step": 3630 + }, + { + "epoch": 1.1617960750853242, + "grad_norm": 0.5574379563331604, + "learning_rate": 7.689877516968546e-06, + "loss": 0.5309, + "step": 3631 + }, + { + "epoch": 1.1621160409556315, + "grad_norm": 0.5204976797103882, + "learning_rate": 7.688307911049639e-06, + "loss": 0.4789, + "step": 3632 + }, + { + "epoch": 1.1624360068259385, + "grad_norm": 0.5226881504058838, + "learning_rate": 7.686737932393606e-06, + "loss": 0.4807, + "step": 3633 + }, + { + "epoch": 1.1627559726962458, + "grad_norm": 0.5622045993804932, + "learning_rate": 7.685167581218125e-06, + "loss": 0.4955, + "step": 3634 + }, + { + "epoch": 1.163075938566553, + "grad_norm": 0.4869825839996338, + "learning_rate": 7.683596857740929e-06, + "loss": 0.49, + "step": 3635 + }, + { + "epoch": 1.16339590443686, + "grad_norm": 0.5154707431793213, + "learning_rate": 7.682025762179801e-06, + "loss": 0.4621, + "step": 3636 + }, + { + "epoch": 1.1637158703071673, + "grad_norm": 0.5349481105804443, + "learning_rate": 7.680454294752574e-06, + "loss": 0.5108, + "step": 3637 + }, + { + "epoch": 1.1640358361774745, + "grad_norm": 0.5748913288116455, + "learning_rate": 7.678882455677135e-06, + "loss": 0.5004, + "step": 3638 + }, + { + "epoch": 1.1643558020477816, + "grad_norm": 0.5858158469200134, + "learning_rate": 7.67731024517142e-06, + "loss": 0.4905, + "step": 3639 + }, + { + "epoch": 1.1646757679180888, + "grad_norm": 0.5105371475219727, + "learning_rate": 7.67573766345342e-06, + "loss": 0.4657, + "step": 3640 + }, + { + "epoch": 1.1649957337883958, + "grad_norm": 0.5406975150108337, + "learning_rate": 7.674164710741172e-06, + "loss": 0.4788, + "step": 3641 + }, + { + "epoch": 1.165315699658703, + "grad_norm": 0.6753333806991577, + "learning_rate": 7.672591387252773e-06, + "loss": 0.5787, + "step": 3642 + }, + { + "epoch": 1.1656356655290103, + "grad_norm": 0.529922366142273, + "learning_rate": 7.671017693206358e-06, + "loss": 0.5105, + "step": 3643 + }, + { + "epoch": 1.1659556313993173, + "grad_norm": 0.6094779968261719, + "learning_rate": 7.66944362882013e-06, + "loss": 0.474, + "step": 3644 + }, + { + "epoch": 1.1662755972696246, + "grad_norm": 0.5302444100379944, + "learning_rate": 7.667869194312331e-06, + "loss": 0.4518, + "step": 3645 + }, + { + "epoch": 1.1665955631399316, + "grad_norm": 0.5726618766784668, + "learning_rate": 7.666294389901258e-06, + "loss": 0.4765, + "step": 3646 + }, + { + "epoch": 1.1669155290102389, + "grad_norm": 0.5927723050117493, + "learning_rate": 7.66471921580526e-06, + "loss": 0.4839, + "step": 3647 + }, + { + "epoch": 1.1672354948805461, + "grad_norm": 0.5504419207572937, + "learning_rate": 7.663143672242739e-06, + "loss": 0.5065, + "step": 3648 + }, + { + "epoch": 1.1675554607508531, + "grad_norm": 0.5538895726203918, + "learning_rate": 7.661567759432142e-06, + "loss": 0.5142, + "step": 3649 + }, + { + "epoch": 1.1678754266211604, + "grad_norm": 0.550895094871521, + "learning_rate": 7.659991477591973e-06, + "loss": 0.4485, + "step": 3650 + }, + { + "epoch": 1.1681953924914676, + "grad_norm": 0.496063768863678, + "learning_rate": 7.658414826940788e-06, + "loss": 0.4729, + "step": 3651 + }, + { + "epoch": 1.1685153583617747, + "grad_norm": 0.5706762075424194, + "learning_rate": 7.656837807697187e-06, + "loss": 0.5289, + "step": 3652 + }, + { + "epoch": 1.168835324232082, + "grad_norm": 0.5495587587356567, + "learning_rate": 7.655260420079827e-06, + "loss": 0.4959, + "step": 3653 + }, + { + "epoch": 1.1691552901023892, + "grad_norm": 0.5080969929695129, + "learning_rate": 7.653682664307419e-06, + "loss": 0.452, + "step": 3654 + }, + { + "epoch": 1.1694752559726962, + "grad_norm": 0.531419038772583, + "learning_rate": 7.652104540598712e-06, + "loss": 0.4954, + "step": 3655 + }, + { + "epoch": 1.1697952218430034, + "grad_norm": 0.5374448895454407, + "learning_rate": 7.650526049172524e-06, + "loss": 0.4881, + "step": 3656 + }, + { + "epoch": 1.1701151877133107, + "grad_norm": 0.5262492895126343, + "learning_rate": 7.648947190247709e-06, + "loss": 0.4666, + "step": 3657 + }, + { + "epoch": 1.1704351535836177, + "grad_norm": 0.5371428728103638, + "learning_rate": 7.64736796404318e-06, + "loss": 0.5373, + "step": 3658 + }, + { + "epoch": 1.170755119453925, + "grad_norm": 0.5738182663917542, + "learning_rate": 7.645788370777898e-06, + "loss": 0.5619, + "step": 3659 + }, + { + "epoch": 1.171075085324232, + "grad_norm": 0.6016698479652405, + "learning_rate": 7.644208410670874e-06, + "loss": 0.4897, + "step": 3660 + }, + { + "epoch": 1.1713950511945392, + "grad_norm": 0.5825162529945374, + "learning_rate": 7.642628083941172e-06, + "loss": 0.5264, + "step": 3661 + }, + { + "epoch": 1.1717150170648465, + "grad_norm": 0.5533202290534973, + "learning_rate": 7.64104739080791e-06, + "loss": 0.4943, + "step": 3662 + }, + { + "epoch": 1.1720349829351535, + "grad_norm": 0.5478944778442383, + "learning_rate": 7.639466331490248e-06, + "loss": 0.5156, + "step": 3663 + }, + { + "epoch": 1.1723549488054608, + "grad_norm": 0.5522754788398743, + "learning_rate": 7.637884906207406e-06, + "loss": 0.4601, + "step": 3664 + }, + { + "epoch": 1.172674914675768, + "grad_norm": 0.5325472950935364, + "learning_rate": 7.636303115178646e-06, + "loss": 0.4643, + "step": 3665 + }, + { + "epoch": 1.172994880546075, + "grad_norm": 0.5285465717315674, + "learning_rate": 7.634720958623287e-06, + "loss": 0.4952, + "step": 3666 + }, + { + "epoch": 1.1733148464163823, + "grad_norm": 0.5523001551628113, + "learning_rate": 7.633138436760702e-06, + "loss": 0.5429, + "step": 3667 + }, + { + "epoch": 1.1736348122866893, + "grad_norm": 0.5258068442344666, + "learning_rate": 7.6315555498103e-06, + "loss": 0.474, + "step": 3668 + }, + { + "epoch": 1.1739547781569966, + "grad_norm": 0.5632274746894836, + "learning_rate": 7.629972297991558e-06, + "loss": 0.5199, + "step": 3669 + }, + { + "epoch": 1.1742747440273038, + "grad_norm": 0.5483708381652832, + "learning_rate": 7.62838868152399e-06, + "loss": 0.5031, + "step": 3670 + }, + { + "epoch": 1.1745947098976108, + "grad_norm": 0.5653208494186401, + "learning_rate": 7.626804700627172e-06, + "loss": 0.4494, + "step": 3671 + }, + { + "epoch": 1.174914675767918, + "grad_norm": 0.530471920967102, + "learning_rate": 7.625220355520722e-06, + "loss": 0.521, + "step": 3672 + }, + { + "epoch": 1.1752346416382253, + "grad_norm": 0.47522956132888794, + "learning_rate": 7.62363564642431e-06, + "loss": 0.4676, + "step": 3673 + }, + { + "epoch": 1.1755546075085324, + "grad_norm": 0.583890438079834, + "learning_rate": 7.62205057355766e-06, + "loss": 0.5447, + "step": 3674 + }, + { + "epoch": 1.1758745733788396, + "grad_norm": 0.5340621471405029, + "learning_rate": 7.620465137140543e-06, + "loss": 0.4847, + "step": 3675 + }, + { + "epoch": 1.1761945392491469, + "grad_norm": 0.5879509449005127, + "learning_rate": 7.6188793373927835e-06, + "loss": 0.4956, + "step": 3676 + }, + { + "epoch": 1.1765145051194539, + "grad_norm": 0.5591239333152771, + "learning_rate": 7.617293174534252e-06, + "loss": 0.5683, + "step": 3677 + }, + { + "epoch": 1.1768344709897611, + "grad_norm": 0.5671486258506775, + "learning_rate": 7.6157066487848745e-06, + "loss": 0.4741, + "step": 3678 + }, + { + "epoch": 1.1771544368600684, + "grad_norm": 0.5293091535568237, + "learning_rate": 7.6141197603646225e-06, + "loss": 0.438, + "step": 3679 + }, + { + "epoch": 1.1774744027303754, + "grad_norm": 0.5160754323005676, + "learning_rate": 7.612532509493521e-06, + "loss": 0.4938, + "step": 3680 + }, + { + "epoch": 1.1777943686006827, + "grad_norm": 0.5765207409858704, + "learning_rate": 7.610944896391644e-06, + "loss": 0.5045, + "step": 3681 + }, + { + "epoch": 1.1781143344709897, + "grad_norm": 0.48545050621032715, + "learning_rate": 7.609356921279116e-06, + "loss": 0.4778, + "step": 3682 + }, + { + "epoch": 1.178434300341297, + "grad_norm": 0.5499753355979919, + "learning_rate": 7.607768584376112e-06, + "loss": 0.5101, + "step": 3683 + }, + { + "epoch": 1.1787542662116042, + "grad_norm": 0.6321439146995544, + "learning_rate": 7.606179885902858e-06, + "loss": 0.4998, + "step": 3684 + }, + { + "epoch": 1.1790742320819112, + "grad_norm": 0.5480753779411316, + "learning_rate": 7.604590826079625e-06, + "loss": 0.4849, + "step": 3685 + }, + { + "epoch": 1.1793941979522184, + "grad_norm": 0.5435191988945007, + "learning_rate": 7.603001405126742e-06, + "loss": 0.5338, + "step": 3686 + }, + { + "epoch": 1.1797141638225255, + "grad_norm": 0.5904504060745239, + "learning_rate": 7.601411623264582e-06, + "loss": 0.4498, + "step": 3687 + }, + { + "epoch": 1.1800341296928327, + "grad_norm": 0.5386573076248169, + "learning_rate": 7.599821480713571e-06, + "loss": 0.4663, + "step": 3688 + }, + { + "epoch": 1.18035409556314, + "grad_norm": 0.5477933287620544, + "learning_rate": 7.598230977694185e-06, + "loss": 0.5051, + "step": 3689 + }, + { + "epoch": 1.180674061433447, + "grad_norm": 0.5635031461715698, + "learning_rate": 7.596640114426948e-06, + "loss": 0.4759, + "step": 3690 + }, + { + "epoch": 1.1809940273037542, + "grad_norm": 0.5476839542388916, + "learning_rate": 7.5950488911324325e-06, + "loss": 0.4915, + "step": 3691 + }, + { + "epoch": 1.1813139931740615, + "grad_norm": 0.5442044734954834, + "learning_rate": 7.593457308031269e-06, + "loss": 0.515, + "step": 3692 + }, + { + "epoch": 1.1816339590443685, + "grad_norm": 0.5077284574508667, + "learning_rate": 7.591865365344129e-06, + "loss": 0.4912, + "step": 3693 + }, + { + "epoch": 1.1819539249146758, + "grad_norm": 0.46649423241615295, + "learning_rate": 7.5902730632917395e-06, + "loss": 0.4025, + "step": 3694 + }, + { + "epoch": 1.182273890784983, + "grad_norm": 0.5748319029808044, + "learning_rate": 7.5886804020948725e-06, + "loss": 0.5741, + "step": 3695 + }, + { + "epoch": 1.18259385665529, + "grad_norm": 0.5408945679664612, + "learning_rate": 7.587087381974353e-06, + "loss": 0.5067, + "step": 3696 + }, + { + "epoch": 1.1829138225255973, + "grad_norm": 0.5399074554443359, + "learning_rate": 7.585494003151055e-06, + "loss": 0.4849, + "step": 3697 + }, + { + "epoch": 1.1832337883959045, + "grad_norm": 0.5231366753578186, + "learning_rate": 7.583900265845905e-06, + "loss": 0.5007, + "step": 3698 + }, + { + "epoch": 1.1835537542662116, + "grad_norm": 0.590194582939148, + "learning_rate": 7.582306170279873e-06, + "loss": 0.4894, + "step": 3699 + }, + { + "epoch": 1.1838737201365188, + "grad_norm": 0.5386995673179626, + "learning_rate": 7.580711716673985e-06, + "loss": 0.5246, + "step": 3700 + }, + { + "epoch": 1.184193686006826, + "grad_norm": 0.514801025390625, + "learning_rate": 7.579116905249313e-06, + "loss": 0.4177, + "step": 3701 + }, + { + "epoch": 1.184513651877133, + "grad_norm": 0.5471378564834595, + "learning_rate": 7.5775217362269785e-06, + "loss": 0.5378, + "step": 3702 + }, + { + "epoch": 1.1848336177474403, + "grad_norm": 0.5445446372032166, + "learning_rate": 7.575926209828154e-06, + "loss": 0.5192, + "step": 3703 + }, + { + "epoch": 1.1851535836177474, + "grad_norm": 0.4958280324935913, + "learning_rate": 7.574330326274062e-06, + "loss": 0.5271, + "step": 3704 + }, + { + "epoch": 1.1854735494880546, + "grad_norm": 0.5592751502990723, + "learning_rate": 7.572734085785973e-06, + "loss": 0.5534, + "step": 3705 + }, + { + "epoch": 1.1857935153583619, + "grad_norm": 0.48439303040504456, + "learning_rate": 7.57113748858521e-06, + "loss": 0.4323, + "step": 3706 + }, + { + "epoch": 1.1861134812286689, + "grad_norm": 0.5728739500045776, + "learning_rate": 7.569540534893139e-06, + "loss": 0.5274, + "step": 3707 + }, + { + "epoch": 1.1864334470989761, + "grad_norm": 0.5586733222007751, + "learning_rate": 7.567943224931182e-06, + "loss": 0.484, + "step": 3708 + }, + { + "epoch": 1.1867534129692832, + "grad_norm": 0.6303009986877441, + "learning_rate": 7.566345558920807e-06, + "loss": 0.4873, + "step": 3709 + }, + { + "epoch": 1.1870733788395904, + "grad_norm": 0.5872066020965576, + "learning_rate": 7.564747537083533e-06, + "loss": 0.5474, + "step": 3710 + }, + { + "epoch": 1.1873933447098977, + "grad_norm": 0.5944347977638245, + "learning_rate": 7.563149159640929e-06, + "loss": 0.5085, + "step": 3711 + }, + { + "epoch": 1.1877133105802047, + "grad_norm": 0.5409294366836548, + "learning_rate": 7.56155042681461e-06, + "loss": 0.4931, + "step": 3712 + }, + { + "epoch": 1.188033276450512, + "grad_norm": 0.49820324778556824, + "learning_rate": 7.559951338826242e-06, + "loss": 0.4691, + "step": 3713 + }, + { + "epoch": 1.1883532423208192, + "grad_norm": 0.560949981212616, + "learning_rate": 7.558351895897541e-06, + "loss": 0.4733, + "step": 3714 + }, + { + "epoch": 1.1886732081911262, + "grad_norm": 0.5346348285675049, + "learning_rate": 7.556752098250272e-06, + "loss": 0.5654, + "step": 3715 + }, + { + "epoch": 1.1889931740614335, + "grad_norm": 0.5067862868309021, + "learning_rate": 7.555151946106249e-06, + "loss": 0.4771, + "step": 3716 + }, + { + "epoch": 1.1893131399317407, + "grad_norm": 0.47606727480888367, + "learning_rate": 7.553551439687335e-06, + "loss": 0.4608, + "step": 3717 + }, + { + "epoch": 1.1896331058020477, + "grad_norm": 0.6058302521705627, + "learning_rate": 7.551950579215441e-06, + "loss": 0.5519, + "step": 3718 + }, + { + "epoch": 1.189953071672355, + "grad_norm": 0.5280526280403137, + "learning_rate": 7.55034936491253e-06, + "loss": 0.4647, + "step": 3719 + }, + { + "epoch": 1.1902730375426622, + "grad_norm": 0.5691516995429993, + "learning_rate": 7.548747797000611e-06, + "loss": 0.529, + "step": 3720 + }, + { + "epoch": 1.1905930034129693, + "grad_norm": 0.5301852226257324, + "learning_rate": 7.547145875701744e-06, + "loss": 0.489, + "step": 3721 + }, + { + "epoch": 1.1909129692832765, + "grad_norm": 0.5888407230377197, + "learning_rate": 7.545543601238037e-06, + "loss": 0.49, + "step": 3722 + }, + { + "epoch": 1.1912329351535835, + "grad_norm": 0.5133174061775208, + "learning_rate": 7.543940973831647e-06, + "loss": 0.5068, + "step": 3723 + }, + { + "epoch": 1.1915529010238908, + "grad_norm": 0.4624699354171753, + "learning_rate": 7.54233799370478e-06, + "loss": 0.4747, + "step": 3724 + }, + { + "epoch": 1.191872866894198, + "grad_norm": 0.6100415587425232, + "learning_rate": 7.540734661079694e-06, + "loss": 0.5976, + "step": 3725 + }, + { + "epoch": 1.192192832764505, + "grad_norm": 0.5251874923706055, + "learning_rate": 7.539130976178688e-06, + "loss": 0.4889, + "step": 3726 + }, + { + "epoch": 1.1925127986348123, + "grad_norm": 0.5343571901321411, + "learning_rate": 7.5375269392241205e-06, + "loss": 0.5279, + "step": 3727 + }, + { + "epoch": 1.1928327645051195, + "grad_norm": 0.5854262113571167, + "learning_rate": 7.535922550438387e-06, + "loss": 0.4543, + "step": 3728 + }, + { + "epoch": 1.1931527303754266, + "grad_norm": 0.5820112824440002, + "learning_rate": 7.534317810043943e-06, + "loss": 0.5291, + "step": 3729 + }, + { + "epoch": 1.1934726962457338, + "grad_norm": 0.4850411117076874, + "learning_rate": 7.532712718263285e-06, + "loss": 0.478, + "step": 3730 + }, + { + "epoch": 1.1937926621160408, + "grad_norm": 0.515326738357544, + "learning_rate": 7.5311072753189604e-06, + "loss": 0.4698, + "step": 3731 + }, + { + "epoch": 1.194112627986348, + "grad_norm": 0.5624679923057556, + "learning_rate": 7.52950148143357e-06, + "loss": 0.5063, + "step": 3732 + }, + { + "epoch": 1.1944325938566553, + "grad_norm": 0.556932270526886, + "learning_rate": 7.527895336829754e-06, + "loss": 0.5218, + "step": 3733 + }, + { + "epoch": 1.1947525597269624, + "grad_norm": 0.565220296382904, + "learning_rate": 7.526288841730209e-06, + "loss": 0.4744, + "step": 3734 + }, + { + "epoch": 1.1950725255972696, + "grad_norm": 0.5757796168327332, + "learning_rate": 7.524681996357677e-06, + "loss": 0.4932, + "step": 3735 + }, + { + "epoch": 1.1953924914675769, + "grad_norm": 0.5540634989738464, + "learning_rate": 7.523074800934948e-06, + "loss": 0.53, + "step": 3736 + }, + { + "epoch": 1.195712457337884, + "grad_norm": 0.518076479434967, + "learning_rate": 7.5214672556848645e-06, + "loss": 0.495, + "step": 3737 + }, + { + "epoch": 1.1960324232081911, + "grad_norm": 0.5275093913078308, + "learning_rate": 7.5198593608303106e-06, + "loss": 0.5026, + "step": 3738 + }, + { + "epoch": 1.1963523890784984, + "grad_norm": 0.5814018249511719, + "learning_rate": 7.5182511165942264e-06, + "loss": 0.5303, + "step": 3739 + }, + { + "epoch": 1.1966723549488054, + "grad_norm": 0.5568544268608093, + "learning_rate": 7.516642523199595e-06, + "loss": 0.4707, + "step": 3740 + }, + { + "epoch": 1.1969923208191127, + "grad_norm": 0.4932362139225006, + "learning_rate": 7.51503358086945e-06, + "loss": 0.4604, + "step": 3741 + }, + { + "epoch": 1.19731228668942, + "grad_norm": 0.5840187072753906, + "learning_rate": 7.513424289826874e-06, + "loss": 0.5553, + "step": 3742 + }, + { + "epoch": 1.197632252559727, + "grad_norm": 0.5801661610603333, + "learning_rate": 7.511814650294994e-06, + "loss": 0.5175, + "step": 3743 + }, + { + "epoch": 1.1979522184300342, + "grad_norm": 0.5177812576293945, + "learning_rate": 7.510204662496994e-06, + "loss": 0.4402, + "step": 3744 + }, + { + "epoch": 1.1982721843003412, + "grad_norm": 0.5612710118293762, + "learning_rate": 7.508594326656096e-06, + "loss": 0.5158, + "step": 3745 + }, + { + "epoch": 1.1985921501706485, + "grad_norm": 0.5415658950805664, + "learning_rate": 7.506983642995576e-06, + "loss": 0.4528, + "step": 3746 + }, + { + "epoch": 1.1989121160409557, + "grad_norm": 0.5858944058418274, + "learning_rate": 7.50537261173876e-06, + "loss": 0.5192, + "step": 3747 + }, + { + "epoch": 1.1992320819112627, + "grad_norm": 0.5906977653503418, + "learning_rate": 7.503761233109015e-06, + "loss": 0.5243, + "step": 3748 + }, + { + "epoch": 1.19955204778157, + "grad_norm": 0.5039230585098267, + "learning_rate": 7.502149507329766e-06, + "loss": 0.4875, + "step": 3749 + }, + { + "epoch": 1.199872013651877, + "grad_norm": 0.5558406710624695, + "learning_rate": 7.5005374346244755e-06, + "loss": 0.4915, + "step": 3750 + }, + { + "epoch": 1.2001919795221843, + "grad_norm": 0.5710980296134949, + "learning_rate": 7.498925015216662e-06, + "loss": 0.5284, + "step": 3751 + }, + { + "epoch": 1.2005119453924915, + "grad_norm": 0.5781745910644531, + "learning_rate": 7.497312249329889e-06, + "loss": 0.4762, + "step": 3752 + }, + { + "epoch": 1.2008319112627985, + "grad_norm": 0.5416224002838135, + "learning_rate": 7.495699137187767e-06, + "loss": 0.4787, + "step": 3753 + }, + { + "epoch": 1.2011518771331058, + "grad_norm": 0.531855046749115, + "learning_rate": 7.494085679013959e-06, + "loss": 0.5433, + "step": 3754 + }, + { + "epoch": 1.201471843003413, + "grad_norm": 0.5295029878616333, + "learning_rate": 7.49247187503217e-06, + "loss": 0.5452, + "step": 3755 + }, + { + "epoch": 1.20179180887372, + "grad_norm": 0.5794286131858826, + "learning_rate": 7.490857725466157e-06, + "loss": 0.4521, + "step": 3756 + }, + { + "epoch": 1.2021117747440273, + "grad_norm": 0.587912917137146, + "learning_rate": 7.489243230539726e-06, + "loss": 0.5056, + "step": 3757 + }, + { + "epoch": 1.2024317406143346, + "grad_norm": 0.5290972590446472, + "learning_rate": 7.487628390476724e-06, + "loss": 0.5188, + "step": 3758 + }, + { + "epoch": 1.2027517064846416, + "grad_norm": 0.5166901350021362, + "learning_rate": 7.486013205501053e-06, + "loss": 0.4563, + "step": 3759 + }, + { + "epoch": 1.2030716723549488, + "grad_norm": 0.5781247615814209, + "learning_rate": 7.484397675836661e-06, + "loss": 0.5591, + "step": 3760 + }, + { + "epoch": 1.203391638225256, + "grad_norm": 0.4947353005409241, + "learning_rate": 7.482781801707543e-06, + "loss": 0.4994, + "step": 3761 + }, + { + "epoch": 1.203711604095563, + "grad_norm": 0.6305955648422241, + "learning_rate": 7.481165583337741e-06, + "loss": 0.5184, + "step": 3762 + }, + { + "epoch": 1.2040315699658704, + "grad_norm": 0.5540664792060852, + "learning_rate": 7.479549020951345e-06, + "loss": 0.5269, + "step": 3763 + }, + { + "epoch": 1.2043515358361774, + "grad_norm": 0.5958874225616455, + "learning_rate": 7.477932114772495e-06, + "loss": 0.4759, + "step": 3764 + }, + { + "epoch": 1.2046715017064846, + "grad_norm": 0.6359137296676636, + "learning_rate": 7.476314865025376e-06, + "loss": 0.509, + "step": 3765 + }, + { + "epoch": 1.2049914675767919, + "grad_norm": 0.5720431208610535, + "learning_rate": 7.474697271934221e-06, + "loss": 0.5066, + "step": 3766 + }, + { + "epoch": 1.205311433447099, + "grad_norm": 0.5656705498695374, + "learning_rate": 7.473079335723314e-06, + "loss": 0.5232, + "step": 3767 + }, + { + "epoch": 1.2056313993174061, + "grad_norm": 0.5155625939369202, + "learning_rate": 7.471461056616981e-06, + "loss": 0.4721, + "step": 3768 + }, + { + "epoch": 1.2059513651877134, + "grad_norm": 0.5236969590187073, + "learning_rate": 7.469842434839601e-06, + "loss": 0.5029, + "step": 3769 + }, + { + "epoch": 1.2062713310580204, + "grad_norm": 0.518295168876648, + "learning_rate": 7.468223470615593e-06, + "loss": 0.4919, + "step": 3770 + }, + { + "epoch": 1.2065912969283277, + "grad_norm": 0.5845280289649963, + "learning_rate": 7.466604164169433e-06, + "loss": 0.514, + "step": 3771 + }, + { + "epoch": 1.2069112627986347, + "grad_norm": 0.559059202671051, + "learning_rate": 7.464984515725638e-06, + "loss": 0.5455, + "step": 3772 + }, + { + "epoch": 1.207231228668942, + "grad_norm": 0.588559091091156, + "learning_rate": 7.463364525508775e-06, + "loss": 0.5324, + "step": 3773 + }, + { + "epoch": 1.2075511945392492, + "grad_norm": 0.5640997886657715, + "learning_rate": 7.461744193743457e-06, + "loss": 0.486, + "step": 3774 + }, + { + "epoch": 1.2078711604095562, + "grad_norm": 0.5955579876899719, + "learning_rate": 7.460123520654343e-06, + "loss": 0.4913, + "step": 3775 + }, + { + "epoch": 1.2081911262798635, + "grad_norm": 0.576140820980072, + "learning_rate": 7.458502506466146e-06, + "loss": 0.5175, + "step": 3776 + }, + { + "epoch": 1.2085110921501707, + "grad_norm": 0.5625115633010864, + "learning_rate": 7.456881151403618e-06, + "loss": 0.4463, + "step": 3777 + }, + { + "epoch": 1.2088310580204777, + "grad_norm": 0.7174860835075378, + "learning_rate": 7.455259455691562e-06, + "loss": 0.5819, + "step": 3778 + }, + { + "epoch": 1.209151023890785, + "grad_norm": 0.5854070782661438, + "learning_rate": 7.45363741955483e-06, + "loss": 0.5497, + "step": 3779 + }, + { + "epoch": 1.2094709897610922, + "grad_norm": 0.5234859585762024, + "learning_rate": 7.4520150432183156e-06, + "loss": 0.4448, + "step": 3780 + }, + { + "epoch": 1.2097909556313993, + "grad_norm": 0.5454064011573792, + "learning_rate": 7.4503923269069675e-06, + "loss": 0.4655, + "step": 3781 + }, + { + "epoch": 1.2101109215017065, + "grad_norm": 0.6177207827568054, + "learning_rate": 7.448769270845776e-06, + "loss": 0.5141, + "step": 3782 + }, + { + "epoch": 1.2104308873720138, + "grad_norm": 0.5759219527244568, + "learning_rate": 7.447145875259778e-06, + "loss": 0.3989, + "step": 3783 + }, + { + "epoch": 1.2107508532423208, + "grad_norm": 0.589008092880249, + "learning_rate": 7.4455221403740595e-06, + "loss": 0.5068, + "step": 3784 + }, + { + "epoch": 1.211070819112628, + "grad_norm": 0.6944248080253601, + "learning_rate": 7.443898066413755e-06, + "loss": 0.5725, + "step": 3785 + }, + { + "epoch": 1.211390784982935, + "grad_norm": 0.6175182461738586, + "learning_rate": 7.442273653604045e-06, + "loss": 0.5015, + "step": 3786 + }, + { + "epoch": 1.2117107508532423, + "grad_norm": 0.5488205552101135, + "learning_rate": 7.440648902170153e-06, + "loss": 0.498, + "step": 3787 + }, + { + "epoch": 1.2120307167235496, + "grad_norm": 0.5426703095436096, + "learning_rate": 7.439023812337356e-06, + "loss": 0.5147, + "step": 3788 + }, + { + "epoch": 1.2123506825938566, + "grad_norm": 0.6002992987632751, + "learning_rate": 7.4373983843309716e-06, + "loss": 0.51, + "step": 3789 + }, + { + "epoch": 1.2126706484641638, + "grad_norm": 0.5468200445175171, + "learning_rate": 7.435772618376369e-06, + "loss": 0.4511, + "step": 3790 + }, + { + "epoch": 1.212990614334471, + "grad_norm": 0.5139322280883789, + "learning_rate": 7.4341465146989645e-06, + "loss": 0.5221, + "step": 3791 + }, + { + "epoch": 1.213310580204778, + "grad_norm": 0.5981809496879578, + "learning_rate": 7.432520073524217e-06, + "loss": 0.504, + "step": 3792 + }, + { + "epoch": 1.2136305460750854, + "grad_norm": 0.6306090950965881, + "learning_rate": 7.4308932950776325e-06, + "loss": 0.5581, + "step": 3793 + }, + { + "epoch": 1.2139505119453924, + "grad_norm": 0.526813805103302, + "learning_rate": 7.42926617958477e-06, + "loss": 0.4759, + "step": 3794 + }, + { + "epoch": 1.2142704778156996, + "grad_norm": 0.46746328473091125, + "learning_rate": 7.4276387272712295e-06, + "loss": 0.4742, + "step": 3795 + }, + { + "epoch": 1.2145904436860069, + "grad_norm": 0.5398551821708679, + "learning_rate": 7.426010938362657e-06, + "loss": 0.4754, + "step": 3796 + }, + { + "epoch": 1.214910409556314, + "grad_norm": 0.6205205917358398, + "learning_rate": 7.424382813084751e-06, + "loss": 0.5797, + "step": 3797 + }, + { + "epoch": 1.2152303754266212, + "grad_norm": 0.4543765187263489, + "learning_rate": 7.422754351663252e-06, + "loss": 0.4051, + "step": 3798 + }, + { + "epoch": 1.2155503412969284, + "grad_norm": 0.5468537211418152, + "learning_rate": 7.421125554323945e-06, + "loss": 0.5062, + "step": 3799 + }, + { + "epoch": 1.2158703071672354, + "grad_norm": 0.5804492831230164, + "learning_rate": 7.419496421292669e-06, + "loss": 0.4949, + "step": 3800 + }, + { + "epoch": 1.2161902730375427, + "grad_norm": 0.5590612292289734, + "learning_rate": 7.417866952795304e-06, + "loss": 0.5176, + "step": 3801 + }, + { + "epoch": 1.21651023890785, + "grad_norm": 0.6057924032211304, + "learning_rate": 7.416237149057776e-06, + "loss": 0.5109, + "step": 3802 + }, + { + "epoch": 1.216830204778157, + "grad_norm": 0.5891520380973816, + "learning_rate": 7.414607010306062e-06, + "loss": 0.4933, + "step": 3803 + }, + { + "epoch": 1.2171501706484642, + "grad_norm": 0.5428804159164429, + "learning_rate": 7.41297653676618e-06, + "loss": 0.4571, + "step": 3804 + }, + { + "epoch": 1.2174701365187715, + "grad_norm": 0.5450373888015747, + "learning_rate": 7.411345728664198e-06, + "loss": 0.4901, + "step": 3805 + }, + { + "epoch": 1.2177901023890785, + "grad_norm": 0.5642934441566467, + "learning_rate": 7.4097145862262295e-06, + "loss": 0.5008, + "step": 3806 + }, + { + "epoch": 1.2181100682593857, + "grad_norm": 0.5875202417373657, + "learning_rate": 7.408083109678437e-06, + "loss": 0.4655, + "step": 3807 + }, + { + "epoch": 1.2184300341296928, + "grad_norm": 0.6120683550834656, + "learning_rate": 7.406451299247022e-06, + "loss": 0.5183, + "step": 3808 + }, + { + "epoch": 1.21875, + "grad_norm": 0.5873705744743347, + "learning_rate": 7.40481915515824e-06, + "loss": 0.5272, + "step": 3809 + }, + { + "epoch": 1.2190699658703072, + "grad_norm": 0.633432924747467, + "learning_rate": 7.40318667763839e-06, + "loss": 0.5113, + "step": 3810 + }, + { + "epoch": 1.2193899317406143, + "grad_norm": 0.5159578323364258, + "learning_rate": 7.4015538669138144e-06, + "loss": 0.4249, + "step": 3811 + }, + { + "epoch": 1.2197098976109215, + "grad_norm": 0.6399247646331787, + "learning_rate": 7.399920723210907e-06, + "loss": 0.5455, + "step": 3812 + }, + { + "epoch": 1.2200298634812285, + "grad_norm": 0.5966647267341614, + "learning_rate": 7.398287246756104e-06, + "loss": 0.5282, + "step": 3813 + }, + { + "epoch": 1.2203498293515358, + "grad_norm": 0.6498501896858215, + "learning_rate": 7.396653437775889e-06, + "loss": 0.5312, + "step": 3814 + }, + { + "epoch": 1.220669795221843, + "grad_norm": 0.5518560409545898, + "learning_rate": 7.395019296496792e-06, + "loss": 0.4957, + "step": 3815 + }, + { + "epoch": 1.22098976109215, + "grad_norm": 0.5691536068916321, + "learning_rate": 7.393384823145386e-06, + "loss": 0.4695, + "step": 3816 + }, + { + "epoch": 1.2213097269624573, + "grad_norm": 0.506877064704895, + "learning_rate": 7.391750017948298e-06, + "loss": 0.4619, + "step": 3817 + }, + { + "epoch": 1.2216296928327646, + "grad_norm": 0.5100228786468506, + "learning_rate": 7.390114881132192e-06, + "loss": 0.4562, + "step": 3818 + }, + { + "epoch": 1.2219496587030716, + "grad_norm": 0.5075163841247559, + "learning_rate": 7.388479412923782e-06, + "loss": 0.5437, + "step": 3819 + }, + { + "epoch": 1.2222696245733788, + "grad_norm": 0.5212882161140442, + "learning_rate": 7.386843613549828e-06, + "loss": 0.5016, + "step": 3820 + }, + { + "epoch": 1.222589590443686, + "grad_norm": 0.5214887857437134, + "learning_rate": 7.385207483237135e-06, + "loss": 0.4902, + "step": 3821 + }, + { + "epoch": 1.2229095563139931, + "grad_norm": 0.5080110430717468, + "learning_rate": 7.383571022212555e-06, + "loss": 0.4779, + "step": 3822 + }, + { + "epoch": 1.2232295221843004, + "grad_norm": 0.5242145657539368, + "learning_rate": 7.381934230702985e-06, + "loss": 0.5419, + "step": 3823 + }, + { + "epoch": 1.2235494880546076, + "grad_norm": 0.5515463948249817, + "learning_rate": 7.3802971089353696e-06, + "loss": 0.4978, + "step": 3824 + }, + { + "epoch": 1.2238694539249146, + "grad_norm": 0.4667942523956299, + "learning_rate": 7.378659657136696e-06, + "loss": 0.4749, + "step": 3825 + }, + { + "epoch": 1.224189419795222, + "grad_norm": 0.49958017468452454, + "learning_rate": 7.377021875533998e-06, + "loss": 0.4846, + "step": 3826 + }, + { + "epoch": 1.224509385665529, + "grad_norm": 0.49523359537124634, + "learning_rate": 7.37538376435436e-06, + "loss": 0.4785, + "step": 3827 + }, + { + "epoch": 1.2248293515358362, + "grad_norm": 0.5057116150856018, + "learning_rate": 7.373745323824902e-06, + "loss": 0.4996, + "step": 3828 + }, + { + "epoch": 1.2251493174061434, + "grad_norm": 0.5707506537437439, + "learning_rate": 7.372106554172802e-06, + "loss": 0.4677, + "step": 3829 + }, + { + "epoch": 1.2254692832764504, + "grad_norm": 0.5567911863327026, + "learning_rate": 7.370467455625274e-06, + "loss": 0.5527, + "step": 3830 + }, + { + "epoch": 1.2257892491467577, + "grad_norm": 0.539019763469696, + "learning_rate": 7.368828028409581e-06, + "loss": 0.5028, + "step": 3831 + }, + { + "epoch": 1.226109215017065, + "grad_norm": 0.5459343194961548, + "learning_rate": 7.367188272753033e-06, + "loss": 0.5102, + "step": 3832 + }, + { + "epoch": 1.226429180887372, + "grad_norm": 0.550582766532898, + "learning_rate": 7.365548188882982e-06, + "loss": 0.5098, + "step": 3833 + }, + { + "epoch": 1.2267491467576792, + "grad_norm": 0.5117180347442627, + "learning_rate": 7.363907777026829e-06, + "loss": 0.4085, + "step": 3834 + }, + { + "epoch": 1.2270691126279862, + "grad_norm": 0.6057801842689514, + "learning_rate": 7.36226703741202e-06, + "loss": 0.6115, + "step": 3835 + }, + { + "epoch": 1.2273890784982935, + "grad_norm": 0.5283210277557373, + "learning_rate": 7.360625970266042e-06, + "loss": 0.5066, + "step": 3836 + }, + { + "epoch": 1.2277090443686007, + "grad_norm": 0.5724843740463257, + "learning_rate": 7.358984575816437e-06, + "loss": 0.5366, + "step": 3837 + }, + { + "epoch": 1.2280290102389078, + "grad_norm": 0.49431341886520386, + "learning_rate": 7.3573428542907784e-06, + "loss": 0.4815, + "step": 3838 + }, + { + "epoch": 1.228348976109215, + "grad_norm": 0.5193344950675964, + "learning_rate": 7.3557008059167e-06, + "loss": 0.4899, + "step": 3839 + }, + { + "epoch": 1.2286689419795223, + "grad_norm": 0.5036170482635498, + "learning_rate": 7.35405843092187e-06, + "loss": 0.4829, + "step": 3840 + }, + { + "epoch": 1.2289889078498293, + "grad_norm": 0.5176819562911987, + "learning_rate": 7.352415729534006e-06, + "loss": 0.5814, + "step": 3841 + }, + { + "epoch": 1.2293088737201365, + "grad_norm": 0.46437740325927734, + "learning_rate": 7.350772701980873e-06, + "loss": 0.467, + "step": 3842 + }, + { + "epoch": 1.2296288395904438, + "grad_norm": 0.520893394947052, + "learning_rate": 7.3491293484902735e-06, + "loss": 0.4686, + "step": 3843 + }, + { + "epoch": 1.2299488054607508, + "grad_norm": 0.5439760088920593, + "learning_rate": 7.347485669290067e-06, + "loss": 0.5304, + "step": 3844 + }, + { + "epoch": 1.230268771331058, + "grad_norm": 0.5253164768218994, + "learning_rate": 7.345841664608146e-06, + "loss": 0.4339, + "step": 3845 + }, + { + "epoch": 1.2305887372013653, + "grad_norm": 0.5876121520996094, + "learning_rate": 7.344197334672458e-06, + "loss": 0.4885, + "step": 3846 + }, + { + "epoch": 1.2309087030716723, + "grad_norm": 0.5066995024681091, + "learning_rate": 7.34255267971099e-06, + "loss": 0.5047, + "step": 3847 + }, + { + "epoch": 1.2312286689419796, + "grad_norm": 0.530464231967926, + "learning_rate": 7.3409076999517715e-06, + "loss": 0.5129, + "step": 3848 + }, + { + "epoch": 1.2315486348122866, + "grad_norm": 0.5482179522514343, + "learning_rate": 7.339262395622887e-06, + "loss": 0.4811, + "step": 3849 + }, + { + "epoch": 1.2318686006825939, + "grad_norm": 0.5003216862678528, + "learning_rate": 7.337616766952455e-06, + "loss": 0.4204, + "step": 3850 + }, + { + "epoch": 1.232188566552901, + "grad_norm": 0.5469177961349487, + "learning_rate": 7.3359708141686484e-06, + "loss": 0.5048, + "step": 3851 + }, + { + "epoch": 1.2325085324232081, + "grad_norm": 0.6300505995750427, + "learning_rate": 7.334324537499677e-06, + "loss": 0.467, + "step": 3852 + }, + { + "epoch": 1.2328284982935154, + "grad_norm": 0.5806970000267029, + "learning_rate": 7.3326779371738e-06, + "loss": 0.5221, + "step": 3853 + }, + { + "epoch": 1.2331484641638226, + "grad_norm": 0.4897526502609253, + "learning_rate": 7.3310310134193215e-06, + "loss": 0.472, + "step": 3854 + }, + { + "epoch": 1.2334684300341296, + "grad_norm": 0.5181820392608643, + "learning_rate": 7.329383766464588e-06, + "loss": 0.4576, + "step": 3855 + }, + { + "epoch": 1.233788395904437, + "grad_norm": 0.60699862241745, + "learning_rate": 7.3277361965379936e-06, + "loss": 0.472, + "step": 3856 + }, + { + "epoch": 1.234108361774744, + "grad_norm": 0.5165917873382568, + "learning_rate": 7.326088303867974e-06, + "loss": 0.4934, + "step": 3857 + }, + { + "epoch": 1.2344283276450512, + "grad_norm": 0.4755984842777252, + "learning_rate": 7.324440088683014e-06, + "loss": 0.4642, + "step": 3858 + }, + { + "epoch": 1.2347482935153584, + "grad_norm": 0.5469703078269958, + "learning_rate": 7.322791551211639e-06, + "loss": 0.4932, + "step": 3859 + }, + { + "epoch": 1.2350682593856654, + "grad_norm": 0.5659158825874329, + "learning_rate": 7.321142691682421e-06, + "loss": 0.4837, + "step": 3860 + }, + { + "epoch": 1.2353882252559727, + "grad_norm": 0.5701580047607422, + "learning_rate": 7.319493510323976e-06, + "loss": 0.493, + "step": 3861 + }, + { + "epoch": 1.23570819112628, + "grad_norm": 0.5682664513587952, + "learning_rate": 7.317844007364966e-06, + "loss": 0.5328, + "step": 3862 + }, + { + "epoch": 1.236028156996587, + "grad_norm": 0.5998944044113159, + "learning_rate": 7.316194183034096e-06, + "loss": 0.4956, + "step": 3863 + }, + { + "epoch": 1.2363481228668942, + "grad_norm": 0.5686836242675781, + "learning_rate": 7.314544037560115e-06, + "loss": 0.5208, + "step": 3864 + }, + { + "epoch": 1.2366680887372015, + "grad_norm": 0.5470593571662903, + "learning_rate": 7.3128935711718185e-06, + "loss": 0.5055, + "step": 3865 + }, + { + "epoch": 1.2369880546075085, + "grad_norm": 0.5307252407073975, + "learning_rate": 7.311242784098047e-06, + "loss": 0.5105, + "step": 3866 + }, + { + "epoch": 1.2373080204778157, + "grad_norm": 0.544973611831665, + "learning_rate": 7.309591676567683e-06, + "loss": 0.4639, + "step": 3867 + }, + { + "epoch": 1.237627986348123, + "grad_norm": 0.6316813230514526, + "learning_rate": 7.307940248809655e-06, + "loss": 0.567, + "step": 3868 + }, + { + "epoch": 1.23794795221843, + "grad_norm": 0.5119458436965942, + "learning_rate": 7.306288501052934e-06, + "loss": 0.4561, + "step": 3869 + }, + { + "epoch": 1.2382679180887373, + "grad_norm": 0.5519663095474243, + "learning_rate": 7.3046364335265375e-06, + "loss": 0.532, + "step": 3870 + }, + { + "epoch": 1.2385878839590443, + "grad_norm": 0.568614661693573, + "learning_rate": 7.302984046459528e-06, + "loss": 0.4663, + "step": 3871 + }, + { + "epoch": 1.2389078498293515, + "grad_norm": 0.6015342473983765, + "learning_rate": 7.301331340081009e-06, + "loss": 0.5595, + "step": 3872 + }, + { + "epoch": 1.2392278156996588, + "grad_norm": 0.5088731050491333, + "learning_rate": 7.299678314620132e-06, + "loss": 0.4932, + "step": 3873 + }, + { + "epoch": 1.2395477815699658, + "grad_norm": 0.5520064234733582, + "learning_rate": 7.298024970306089e-06, + "loss": 0.5142, + "step": 3874 + }, + { + "epoch": 1.239867747440273, + "grad_norm": 0.5585522055625916, + "learning_rate": 7.29637130736812e-06, + "loss": 0.4941, + "step": 3875 + }, + { + "epoch": 1.24018771331058, + "grad_norm": 0.5988473892211914, + "learning_rate": 7.294717326035508e-06, + "loss": 0.4555, + "step": 3876 + }, + { + "epoch": 1.2405076791808873, + "grad_norm": 0.5895554423332214, + "learning_rate": 7.293063026537575e-06, + "loss": 0.536, + "step": 3877 + }, + { + "epoch": 1.2408276450511946, + "grad_norm": 0.5260162353515625, + "learning_rate": 7.291408409103698e-06, + "loss": 0.5072, + "step": 3878 + }, + { + "epoch": 1.2411476109215016, + "grad_norm": 0.5081711411476135, + "learning_rate": 7.289753473963286e-06, + "loss": 0.4796, + "step": 3879 + }, + { + "epoch": 1.2414675767918089, + "grad_norm": 0.5044803023338318, + "learning_rate": 7.2880982213458e-06, + "loss": 0.5022, + "step": 3880 + }, + { + "epoch": 1.241787542662116, + "grad_norm": 0.5551183223724365, + "learning_rate": 7.2864426514807455e-06, + "loss": 0.5554, + "step": 3881 + }, + { + "epoch": 1.2421075085324231, + "grad_norm": 0.5451104044914246, + "learning_rate": 7.2847867645976656e-06, + "loss": 0.4554, + "step": 3882 + }, + { + "epoch": 1.2424274744027304, + "grad_norm": 0.5335094928741455, + "learning_rate": 7.2831305609261514e-06, + "loss": 0.5072, + "step": 3883 + }, + { + "epoch": 1.2427474402730376, + "grad_norm": 0.5673237442970276, + "learning_rate": 7.281474040695839e-06, + "loss": 0.4995, + "step": 3884 + }, + { + "epoch": 1.2430674061433447, + "grad_norm": 0.5887155532836914, + "learning_rate": 7.279817204136405e-06, + "loss": 0.5045, + "step": 3885 + }, + { + "epoch": 1.243387372013652, + "grad_norm": 0.5778360366821289, + "learning_rate": 7.278160051477574e-06, + "loss": 0.4836, + "step": 3886 + }, + { + "epoch": 1.2437073378839592, + "grad_norm": 0.5329510569572449, + "learning_rate": 7.2765025829491105e-06, + "loss": 0.5502, + "step": 3887 + }, + { + "epoch": 1.2440273037542662, + "grad_norm": 0.5407595038414001, + "learning_rate": 7.274844798780826e-06, + "loss": 0.4718, + "step": 3888 + }, + { + "epoch": 1.2443472696245734, + "grad_norm": 0.5253113508224487, + "learning_rate": 7.273186699202572e-06, + "loss": 0.4936, + "step": 3889 + }, + { + "epoch": 1.2446672354948805, + "grad_norm": 0.5383972525596619, + "learning_rate": 7.27152828444425e-06, + "loss": 0.5175, + "step": 3890 + }, + { + "epoch": 1.2449872013651877, + "grad_norm": 0.5071119666099548, + "learning_rate": 7.269869554735796e-06, + "loss": 0.5104, + "step": 3891 + }, + { + "epoch": 1.245307167235495, + "grad_norm": 0.5174180865287781, + "learning_rate": 7.2682105103072e-06, + "loss": 0.5488, + "step": 3892 + }, + { + "epoch": 1.245627133105802, + "grad_norm": 0.45115968585014343, + "learning_rate": 7.266551151388485e-06, + "loss": 0.4492, + "step": 3893 + }, + { + "epoch": 1.2459470989761092, + "grad_norm": 0.5595500469207764, + "learning_rate": 7.264891478209729e-06, + "loss": 0.533, + "step": 3894 + }, + { + "epoch": 1.2462670648464165, + "grad_norm": 0.5710468292236328, + "learning_rate": 7.263231491001044e-06, + "loss": 0.4815, + "step": 3895 + }, + { + "epoch": 1.2465870307167235, + "grad_norm": 0.5594239830970764, + "learning_rate": 7.261571189992589e-06, + "loss": 0.4924, + "step": 3896 + }, + { + "epoch": 1.2469069965870307, + "grad_norm": 0.5303643345832825, + "learning_rate": 7.259910575414569e-06, + "loss": 0.468, + "step": 3897 + }, + { + "epoch": 1.2472269624573378, + "grad_norm": 0.5882477760314941, + "learning_rate": 7.258249647497228e-06, + "loss": 0.498, + "step": 3898 + }, + { + "epoch": 1.247546928327645, + "grad_norm": 0.578035831451416, + "learning_rate": 7.256588406470857e-06, + "loss": 0.4829, + "step": 3899 + }, + { + "epoch": 1.2478668941979523, + "grad_norm": 0.4835994243621826, + "learning_rate": 7.254926852565789e-06, + "loss": 0.486, + "step": 3900 + }, + { + "epoch": 1.2481868600682593, + "grad_norm": 0.5448126196861267, + "learning_rate": 7.253264986012399e-06, + "loss": 0.468, + "step": 3901 + }, + { + "epoch": 1.2485068259385665, + "grad_norm": 0.6015421152114868, + "learning_rate": 7.251602807041111e-06, + "loss": 0.54, + "step": 3902 + }, + { + "epoch": 1.2488267918088738, + "grad_norm": 0.5436009168624878, + "learning_rate": 7.249940315882382e-06, + "loss": 0.5052, + "step": 3903 + }, + { + "epoch": 1.2491467576791808, + "grad_norm": 0.595392644405365, + "learning_rate": 7.248277512766722e-06, + "loss": 0.5412, + "step": 3904 + }, + { + "epoch": 1.249466723549488, + "grad_norm": 0.5022367238998413, + "learning_rate": 7.24661439792468e-06, + "loss": 0.4956, + "step": 3905 + }, + { + "epoch": 1.2497866894197953, + "grad_norm": 0.5141124725341797, + "learning_rate": 7.244950971586849e-06, + "loss": 0.4603, + "step": 3906 + }, + { + "epoch": 1.2501066552901023, + "grad_norm": 0.5934104323387146, + "learning_rate": 7.243287233983866e-06, + "loss": 0.5731, + "step": 3907 + }, + { + "epoch": 1.2504266211604096, + "grad_norm": 0.5501481294631958, + "learning_rate": 7.241623185346409e-06, + "loss": 0.4424, + "step": 3908 + }, + { + "epoch": 1.2507465870307168, + "grad_norm": 0.5349768400192261, + "learning_rate": 7.239958825905201e-06, + "loss": 0.5224, + "step": 3909 + }, + { + "epoch": 1.2510665529010239, + "grad_norm": 0.5530389547348022, + "learning_rate": 7.238294155891006e-06, + "loss": 0.5149, + "step": 3910 + }, + { + "epoch": 1.2513865187713311, + "grad_norm": 0.5477389693260193, + "learning_rate": 7.2366291755346344e-06, + "loss": 0.4818, + "step": 3911 + }, + { + "epoch": 1.2517064846416384, + "grad_norm": 0.5573341250419617, + "learning_rate": 7.234963885066937e-06, + "loss": 0.4673, + "step": 3912 + }, + { + "epoch": 1.2520264505119454, + "grad_norm": 0.5481014251708984, + "learning_rate": 7.233298284718806e-06, + "loss": 0.5064, + "step": 3913 + }, + { + "epoch": 1.2523464163822526, + "grad_norm": 0.5816885828971863, + "learning_rate": 7.231632374721184e-06, + "loss": 0.4764, + "step": 3914 + }, + { + "epoch": 1.2526663822525597, + "grad_norm": 0.6380035281181335, + "learning_rate": 7.2299661553050474e-06, + "loss": 0.5305, + "step": 3915 + }, + { + "epoch": 1.252986348122867, + "grad_norm": 0.571526825428009, + "learning_rate": 7.22829962670142e-06, + "loss": 0.5471, + "step": 3916 + }, + { + "epoch": 1.253306313993174, + "grad_norm": 0.5838145613670349, + "learning_rate": 7.2266327891413705e-06, + "loss": 0.503, + "step": 3917 + }, + { + "epoch": 1.2536262798634812, + "grad_norm": 0.5403713583946228, + "learning_rate": 7.224965642856003e-06, + "loss": 0.4958, + "step": 3918 + }, + { + "epoch": 1.2539462457337884, + "grad_norm": 0.5514960289001465, + "learning_rate": 7.223298188076475e-06, + "loss": 0.4782, + "step": 3919 + }, + { + "epoch": 1.2542662116040955, + "grad_norm": 0.6443050503730774, + "learning_rate": 7.221630425033977e-06, + "loss": 0.501, + "step": 3920 + }, + { + "epoch": 1.2545861774744027, + "grad_norm": 0.5583950281143188, + "learning_rate": 7.219962353959749e-06, + "loss": 0.4591, + "step": 3921 + }, + { + "epoch": 1.25490614334471, + "grad_norm": 0.5959684252738953, + "learning_rate": 7.218293975085071e-06, + "loss": 0.5533, + "step": 3922 + }, + { + "epoch": 1.255226109215017, + "grad_norm": 0.5673621296882629, + "learning_rate": 7.2166252886412614e-06, + "loss": 0.4859, + "step": 3923 + }, + { + "epoch": 1.2555460750853242, + "grad_norm": 0.5439527034759521, + "learning_rate": 7.2149562948596914e-06, + "loss": 0.445, + "step": 3924 + }, + { + "epoch": 1.2558660409556315, + "grad_norm": 0.6172842383384705, + "learning_rate": 7.213286993971764e-06, + "loss": 0.5214, + "step": 3925 + }, + { + "epoch": 1.2561860068259385, + "grad_norm": 0.5660737156867981, + "learning_rate": 7.211617386208936e-06, + "loss": 0.5096, + "step": 3926 + }, + { + "epoch": 1.2565059726962458, + "grad_norm": 0.554425060749054, + "learning_rate": 7.209947471802694e-06, + "loss": 0.5024, + "step": 3927 + }, + { + "epoch": 1.256825938566553, + "grad_norm": 0.6696584224700928, + "learning_rate": 7.208277250984577e-06, + "loss": 0.5449, + "step": 3928 + }, + { + "epoch": 1.25714590443686, + "grad_norm": 0.5374483466148376, + "learning_rate": 7.206606723986164e-06, + "loss": 0.483, + "step": 3929 + }, + { + "epoch": 1.2574658703071673, + "grad_norm": 0.5405775308609009, + "learning_rate": 7.204935891039071e-06, + "loss": 0.4617, + "step": 3930 + }, + { + "epoch": 1.2577858361774745, + "grad_norm": 0.636476993560791, + "learning_rate": 7.203264752374968e-06, + "loss": 0.5509, + "step": 3931 + }, + { + "epoch": 1.2581058020477816, + "grad_norm": 0.5011911392211914, + "learning_rate": 7.201593308225554e-06, + "loss": 0.4196, + "step": 3932 + }, + { + "epoch": 1.2584257679180888, + "grad_norm": 0.5440027117729187, + "learning_rate": 7.199921558822579e-06, + "loss": 0.5234, + "step": 3933 + }, + { + "epoch": 1.2587457337883958, + "grad_norm": 0.6314454674720764, + "learning_rate": 7.198249504397834e-06, + "loss": 0.4913, + "step": 3934 + }, + { + "epoch": 1.259065699658703, + "grad_norm": 0.5794150233268738, + "learning_rate": 7.19657714518315e-06, + "loss": 0.5514, + "step": 3935 + }, + { + "epoch": 1.25938566552901, + "grad_norm": 0.5919234156608582, + "learning_rate": 7.194904481410403e-06, + "loss": 0.5185, + "step": 3936 + }, + { + "epoch": 1.2597056313993173, + "grad_norm": 0.5429485440254211, + "learning_rate": 7.193231513311509e-06, + "loss": 0.5403, + "step": 3937 + }, + { + "epoch": 1.2600255972696246, + "grad_norm": 0.5316311717033386, + "learning_rate": 7.191558241118427e-06, + "loss": 0.4117, + "step": 3938 + }, + { + "epoch": 1.2603455631399316, + "grad_norm": 0.5937775373458862, + "learning_rate": 7.18988466506316e-06, + "loss": 0.5259, + "step": 3939 + }, + { + "epoch": 1.2606655290102389, + "grad_norm": 0.5185807943344116, + "learning_rate": 7.188210785377748e-06, + "loss": 0.5076, + "step": 3940 + }, + { + "epoch": 1.2609854948805461, + "grad_norm": 0.5359899401664734, + "learning_rate": 7.186536602294278e-06, + "loss": 0.5209, + "step": 3941 + }, + { + "epoch": 1.2613054607508531, + "grad_norm": 0.5379454493522644, + "learning_rate": 7.184862116044877e-06, + "loss": 0.5135, + "step": 3942 + }, + { + "epoch": 1.2616254266211604, + "grad_norm": 0.5110536217689514, + "learning_rate": 7.183187326861717e-06, + "loss": 0.4563, + "step": 3943 + }, + { + "epoch": 1.2619453924914676, + "grad_norm": 0.5227304697036743, + "learning_rate": 7.1815122349770075e-06, + "loss": 0.4812, + "step": 3944 + }, + { + "epoch": 1.2622653583617747, + "grad_norm": 0.5106444954872131, + "learning_rate": 7.1798368406230004e-06, + "loss": 0.4683, + "step": 3945 + }, + { + "epoch": 1.262585324232082, + "grad_norm": 0.5713802576065063, + "learning_rate": 7.178161144031994e-06, + "loss": 0.5206, + "step": 3946 + }, + { + "epoch": 1.2629052901023892, + "grad_norm": 0.6216138601303101, + "learning_rate": 7.176485145436325e-06, + "loss": 0.4972, + "step": 3947 + }, + { + "epoch": 1.2632252559726962, + "grad_norm": 0.5910817980766296, + "learning_rate": 7.174808845068372e-06, + "loss": 0.4889, + "step": 3948 + }, + { + "epoch": 1.2635452218430034, + "grad_norm": 0.5474295020103455, + "learning_rate": 7.173132243160557e-06, + "loss": 0.5213, + "step": 3949 + }, + { + "epoch": 1.2638651877133107, + "grad_norm": 0.48133257031440735, + "learning_rate": 7.171455339945339e-06, + "loss": 0.4447, + "step": 3950 + }, + { + "epoch": 1.2641851535836177, + "grad_norm": 0.5086634755134583, + "learning_rate": 7.1697781356552295e-06, + "loss": 0.4936, + "step": 3951 + }, + { + "epoch": 1.264505119453925, + "grad_norm": 0.5170502662658691, + "learning_rate": 7.168100630522769e-06, + "loss": 0.4732, + "step": 3952 + }, + { + "epoch": 1.2648250853242322, + "grad_norm": 0.5468535423278809, + "learning_rate": 7.166422824780549e-06, + "loss": 0.499, + "step": 3953 + }, + { + "epoch": 1.2651450511945392, + "grad_norm": 0.5571072101593018, + "learning_rate": 7.164744718661198e-06, + "loss": 0.4899, + "step": 3954 + }, + { + "epoch": 1.2654650170648465, + "grad_norm": 0.5490705966949463, + "learning_rate": 7.163066312397386e-06, + "loss": 0.492, + "step": 3955 + }, + { + "epoch": 1.2657849829351535, + "grad_norm": 0.567050039768219, + "learning_rate": 7.16138760622183e-06, + "loss": 0.5362, + "step": 3956 + }, + { + "epoch": 1.2661049488054608, + "grad_norm": 0.5776382684707642, + "learning_rate": 7.15970860036728e-06, + "loss": 0.466, + "step": 3957 + }, + { + "epoch": 1.2664249146757678, + "grad_norm": 0.4942425787448883, + "learning_rate": 7.158029295066535e-06, + "loss": 0.4487, + "step": 3958 + }, + { + "epoch": 1.266744880546075, + "grad_norm": 0.4956960082054138, + "learning_rate": 7.156349690552433e-06, + "loss": 0.5155, + "step": 3959 + }, + { + "epoch": 1.2670648464163823, + "grad_norm": 0.5051069259643555, + "learning_rate": 7.154669787057851e-06, + "loss": 0.4963, + "step": 3960 + }, + { + "epoch": 1.2673848122866893, + "grad_norm": 0.5262790322303772, + "learning_rate": 7.152989584815712e-06, + "loss": 0.4625, + "step": 3961 + }, + { + "epoch": 1.2677047781569966, + "grad_norm": 0.592454731464386, + "learning_rate": 7.151309084058976e-06, + "loss": 0.5454, + "step": 3962 + }, + { + "epoch": 1.2680247440273038, + "grad_norm": 0.5562735199928284, + "learning_rate": 7.149628285020647e-06, + "loss": 0.5272, + "step": 3963 + }, + { + "epoch": 1.2683447098976108, + "grad_norm": 0.5365687608718872, + "learning_rate": 7.147947187933772e-06, + "loss": 0.5179, + "step": 3964 + }, + { + "epoch": 1.268664675767918, + "grad_norm": 0.49221140146255493, + "learning_rate": 7.146265793031434e-06, + "loss": 0.4755, + "step": 3965 + }, + { + "epoch": 1.2689846416382253, + "grad_norm": 0.5778467655181885, + "learning_rate": 7.144584100546762e-06, + "loss": 0.4927, + "step": 3966 + }, + { + "epoch": 1.2693046075085324, + "grad_norm": 0.5484925508499146, + "learning_rate": 7.142902110712925e-06, + "loss": 0.4975, + "step": 3967 + }, + { + "epoch": 1.2696245733788396, + "grad_norm": 0.5711588263511658, + "learning_rate": 7.141219823763132e-06, + "loss": 0.513, + "step": 3968 + }, + { + "epoch": 1.2699445392491469, + "grad_norm": 0.4926115572452545, + "learning_rate": 7.139537239930634e-06, + "loss": 0.462, + "step": 3969 + }, + { + "epoch": 1.2702645051194539, + "grad_norm": 0.5405696034431458, + "learning_rate": 7.137854359448723e-06, + "loss": 0.5335, + "step": 3970 + }, + { + "epoch": 1.2705844709897611, + "grad_norm": 0.493924617767334, + "learning_rate": 7.136171182550736e-06, + "loss": 0.4639, + "step": 3971 + }, + { + "epoch": 1.2709044368600684, + "grad_norm": 0.5897502303123474, + "learning_rate": 7.1344877094700425e-06, + "loss": 0.518, + "step": 3972 + }, + { + "epoch": 1.2712244027303754, + "grad_norm": 0.5166211128234863, + "learning_rate": 7.132803940440061e-06, + "loss": 0.5369, + "step": 3973 + }, + { + "epoch": 1.2715443686006827, + "grad_norm": 0.5344966053962708, + "learning_rate": 7.131119875694246e-06, + "loss": 0.4817, + "step": 3974 + }, + { + "epoch": 1.27186433447099, + "grad_norm": 0.5253636240959167, + "learning_rate": 7.129435515466098e-06, + "loss": 0.4572, + "step": 3975 + }, + { + "epoch": 1.272184300341297, + "grad_norm": 0.5483711361885071, + "learning_rate": 7.127750859989154e-06, + "loss": 0.5079, + "step": 3976 + }, + { + "epoch": 1.2725042662116042, + "grad_norm": 0.5688130855560303, + "learning_rate": 7.126065909496993e-06, + "loss": 0.5126, + "step": 3977 + }, + { + "epoch": 1.2728242320819112, + "grad_norm": 0.5963906049728394, + "learning_rate": 7.124380664223236e-06, + "loss": 0.5238, + "step": 3978 + }, + { + "epoch": 1.2731441979522184, + "grad_norm": 0.6051427721977234, + "learning_rate": 7.122695124401544e-06, + "loss": 0.4873, + "step": 3979 + }, + { + "epoch": 1.2734641638225255, + "grad_norm": 0.5014509558677673, + "learning_rate": 7.121009290265619e-06, + "loss": 0.4829, + "step": 3980 + }, + { + "epoch": 1.2737841296928327, + "grad_norm": 0.551353931427002, + "learning_rate": 7.1193231620492055e-06, + "loss": 0.5161, + "step": 3981 + }, + { + "epoch": 1.27410409556314, + "grad_norm": 0.5647012591362, + "learning_rate": 7.117636739986086e-06, + "loss": 0.4926, + "step": 3982 + }, + { + "epoch": 1.274424061433447, + "grad_norm": 0.5722877383232117, + "learning_rate": 7.115950024310083e-06, + "loss": 0.5451, + "step": 3983 + }, + { + "epoch": 1.2747440273037542, + "grad_norm": 0.5415180325508118, + "learning_rate": 7.114263015255065e-06, + "loss": 0.5067, + "step": 3984 + }, + { + "epoch": 1.2750639931740615, + "grad_norm": 0.5215157866477966, + "learning_rate": 7.112575713054937e-06, + "loss": 0.4927, + "step": 3985 + }, + { + "epoch": 1.2753839590443685, + "grad_norm": 0.490604043006897, + "learning_rate": 7.110888117943643e-06, + "loss": 0.481, + "step": 3986 + }, + { + "epoch": 1.2757039249146758, + "grad_norm": 0.5944197177886963, + "learning_rate": 7.109200230155173e-06, + "loss": 0.4671, + "step": 3987 + }, + { + "epoch": 1.276023890784983, + "grad_norm": 0.5736401677131653, + "learning_rate": 7.107512049923555e-06, + "loss": 0.5181, + "step": 3988 + }, + { + "epoch": 1.27634385665529, + "grad_norm": 0.5174470543861389, + "learning_rate": 7.105823577482853e-06, + "loss": 0.5242, + "step": 3989 + }, + { + "epoch": 1.2766638225255973, + "grad_norm": 0.48873624205589294, + "learning_rate": 7.10413481306718e-06, + "loss": 0.4692, + "step": 3990 + }, + { + "epoch": 1.2769837883959045, + "grad_norm": 0.5451329350471497, + "learning_rate": 7.102445756910685e-06, + "loss": 0.5023, + "step": 3991 + }, + { + "epoch": 1.2773037542662116, + "grad_norm": 0.6022747159004211, + "learning_rate": 7.100756409247556e-06, + "loss": 0.521, + "step": 3992 + }, + { + "epoch": 1.2776237201365188, + "grad_norm": 0.5710732936859131, + "learning_rate": 7.099066770312023e-06, + "loss": 0.4571, + "step": 3993 + }, + { + "epoch": 1.277943686006826, + "grad_norm": 0.5511924624443054, + "learning_rate": 7.097376840338357e-06, + "loss": 0.5359, + "step": 3994 + }, + { + "epoch": 1.278263651877133, + "grad_norm": 0.5261993408203125, + "learning_rate": 7.095686619560868e-06, + "loss": 0.4652, + "step": 3995 + }, + { + "epoch": 1.2785836177474403, + "grad_norm": 0.5718681216239929, + "learning_rate": 7.09399610821391e-06, + "loss": 0.5136, + "step": 3996 + }, + { + "epoch": 1.2789035836177474, + "grad_norm": 0.5797334909439087, + "learning_rate": 7.092305306531872e-06, + "loss": 0.4918, + "step": 3997 + }, + { + "epoch": 1.2792235494880546, + "grad_norm": 0.5252419710159302, + "learning_rate": 7.090614214749185e-06, + "loss": 0.4283, + "step": 3998 + }, + { + "epoch": 1.2795435153583616, + "grad_norm": 0.576372504234314, + "learning_rate": 7.0889228331003236e-06, + "loss": 0.5202, + "step": 3999 + }, + { + "epoch": 1.2798634812286689, + "grad_norm": 0.5281257629394531, + "learning_rate": 7.087231161819796e-06, + "loss": 0.5165, + "step": 4000 + }, + { + "epoch": 1.2801834470989761, + "grad_norm": 0.5718040466308594, + "learning_rate": 7.085539201142159e-06, + "loss": 0.4998, + "step": 4001 + }, + { + "epoch": 1.2805034129692832, + "grad_norm": 0.5610496401786804, + "learning_rate": 7.083846951302002e-06, + "loss": 0.4913, + "step": 4002 + }, + { + "epoch": 1.2808233788395904, + "grad_norm": 0.48426106572151184, + "learning_rate": 7.0821544125339585e-06, + "loss": 0.4525, + "step": 4003 + }, + { + "epoch": 1.2811433447098977, + "grad_norm": 0.563601553440094, + "learning_rate": 7.080461585072701e-06, + "loss": 0.5032, + "step": 4004 + }, + { + "epoch": 1.2814633105802047, + "grad_norm": 0.5950374603271484, + "learning_rate": 7.078768469152941e-06, + "loss": 0.5532, + "step": 4005 + }, + { + "epoch": 1.281783276450512, + "grad_norm": 0.5124794244766235, + "learning_rate": 7.0770750650094335e-06, + "loss": 0.4315, + "step": 4006 + }, + { + "epoch": 1.2821032423208192, + "grad_norm": 0.5625861883163452, + "learning_rate": 7.075381372876969e-06, + "loss": 0.4925, + "step": 4007 + }, + { + "epoch": 1.2824232081911262, + "grad_norm": 0.5750341415405273, + "learning_rate": 7.073687392990379e-06, + "loss": 0.5136, + "step": 4008 + }, + { + "epoch": 1.2827431740614335, + "grad_norm": 0.519352376461029, + "learning_rate": 7.071993125584537e-06, + "loss": 0.5323, + "step": 4009 + }, + { + "epoch": 1.2830631399317407, + "grad_norm": 0.5940561294555664, + "learning_rate": 7.070298570894357e-06, + "loss": 0.5413, + "step": 4010 + }, + { + "epoch": 1.2833831058020477, + "grad_norm": 0.5244735479354858, + "learning_rate": 7.068603729154789e-06, + "loss": 0.458, + "step": 4011 + }, + { + "epoch": 1.283703071672355, + "grad_norm": 0.48761364817619324, + "learning_rate": 7.066908600600826e-06, + "loss": 0.4501, + "step": 4012 + }, + { + "epoch": 1.2840230375426622, + "grad_norm": 0.5354412794113159, + "learning_rate": 7.065213185467497e-06, + "loss": 0.4823, + "step": 4013 + }, + { + "epoch": 1.2843430034129693, + "grad_norm": 0.5936446785926819, + "learning_rate": 7.063517483989879e-06, + "loss": 0.5013, + "step": 4014 + }, + { + "epoch": 1.2846629692832765, + "grad_norm": 0.5855966210365295, + "learning_rate": 7.061821496403075e-06, + "loss": 0.5154, + "step": 4015 + }, + { + "epoch": 1.2849829351535837, + "grad_norm": 0.5463988780975342, + "learning_rate": 7.0601252229422435e-06, + "loss": 0.5012, + "step": 4016 + }, + { + "epoch": 1.2853029010238908, + "grad_norm": 0.6146540641784668, + "learning_rate": 7.05842866384257e-06, + "loss": 0.5459, + "step": 4017 + }, + { + "epoch": 1.285622866894198, + "grad_norm": 0.5338971018791199, + "learning_rate": 7.056731819339287e-06, + "loss": 0.4941, + "step": 4018 + }, + { + "epoch": 1.285942832764505, + "grad_norm": 0.5357498526573181, + "learning_rate": 7.055034689667661e-06, + "loss": 0.4963, + "step": 4019 + }, + { + "epoch": 1.2862627986348123, + "grad_norm": 0.5381874442100525, + "learning_rate": 7.053337275063004e-06, + "loss": 0.4578, + "step": 4020 + }, + { + "epoch": 1.2865827645051193, + "grad_norm": 0.642769455909729, + "learning_rate": 7.051639575760664e-06, + "loss": 0.5292, + "step": 4021 + }, + { + "epoch": 1.2869027303754266, + "grad_norm": 0.5198947191238403, + "learning_rate": 7.0499415919960276e-06, + "loss": 0.4965, + "step": 4022 + }, + { + "epoch": 1.2872226962457338, + "grad_norm": 0.5457560420036316, + "learning_rate": 7.048243324004525e-06, + "loss": 0.4504, + "step": 4023 + }, + { + "epoch": 1.2875426621160408, + "grad_norm": 0.5783088207244873, + "learning_rate": 7.046544772021621e-06, + "loss": 0.4804, + "step": 4024 + }, + { + "epoch": 1.287862627986348, + "grad_norm": 0.5254973769187927, + "learning_rate": 7.0448459362828206e-06, + "loss": 0.4466, + "step": 4025 + }, + { + "epoch": 1.2881825938566553, + "grad_norm": 0.5481970906257629, + "learning_rate": 7.043146817023673e-06, + "loss": 0.5046, + "step": 4026 + }, + { + "epoch": 1.2885025597269624, + "grad_norm": 0.579662024974823, + "learning_rate": 7.04144741447976e-06, + "loss": 0.4715, + "step": 4027 + }, + { + "epoch": 1.2888225255972696, + "grad_norm": 0.5224183797836304, + "learning_rate": 7.039747728886708e-06, + "loss": 0.5318, + "step": 4028 + }, + { + "epoch": 1.2891424914675769, + "grad_norm": 0.4587222635746002, + "learning_rate": 7.0380477604801786e-06, + "loss": 0.3968, + "step": 4029 + }, + { + "epoch": 1.289462457337884, + "grad_norm": 0.6167943477630615, + "learning_rate": 7.036347509495875e-06, + "loss": 0.5936, + "step": 4030 + }, + { + "epoch": 1.2897824232081911, + "grad_norm": 0.5406517386436462, + "learning_rate": 7.034646976169541e-06, + "loss": 0.4896, + "step": 4031 + }, + { + "epoch": 1.2901023890784984, + "grad_norm": 0.4828827381134033, + "learning_rate": 7.032946160736956e-06, + "loss": 0.4759, + "step": 4032 + }, + { + "epoch": 1.2904223549488054, + "grad_norm": 0.5524978041648865, + "learning_rate": 7.031245063433938e-06, + "loss": 0.5368, + "step": 4033 + }, + { + "epoch": 1.2907423208191127, + "grad_norm": 0.5164719223976135, + "learning_rate": 7.029543684496352e-06, + "loss": 0.4773, + "step": 4034 + }, + { + "epoch": 1.29106228668942, + "grad_norm": 0.5399884581565857, + "learning_rate": 7.027842024160087e-06, + "loss": 0.4917, + "step": 4035 + }, + { + "epoch": 1.291382252559727, + "grad_norm": 0.5169853568077087, + "learning_rate": 7.0261400826610904e-06, + "loss": 0.4577, + "step": 4036 + }, + { + "epoch": 1.2917022184300342, + "grad_norm": 0.576972246170044, + "learning_rate": 7.0244378602353334e-06, + "loss": 0.5175, + "step": 4037 + }, + { + "epoch": 1.2920221843003412, + "grad_norm": 0.6214761137962341, + "learning_rate": 7.02273535711883e-06, + "loss": 0.5466, + "step": 4038 + }, + { + "epoch": 1.2923421501706485, + "grad_norm": 0.5590415000915527, + "learning_rate": 7.021032573547637e-06, + "loss": 0.4417, + "step": 4039 + }, + { + "epoch": 1.2926621160409557, + "grad_norm": 0.5316540598869324, + "learning_rate": 7.019329509757845e-06, + "loss": 0.4974, + "step": 4040 + }, + { + "epoch": 1.2929820819112627, + "grad_norm": 0.5390881896018982, + "learning_rate": 7.017626165985589e-06, + "loss": 0.4569, + "step": 4041 + }, + { + "epoch": 1.29330204778157, + "grad_norm": 0.5673624873161316, + "learning_rate": 7.015922542467038e-06, + "loss": 0.5331, + "step": 4042 + }, + { + "epoch": 1.293622013651877, + "grad_norm": 0.5370771884918213, + "learning_rate": 7.014218639438399e-06, + "loss": 0.506, + "step": 4043 + }, + { + "epoch": 1.2939419795221843, + "grad_norm": 0.5172004699707031, + "learning_rate": 7.012514457135924e-06, + "loss": 0.4921, + "step": 4044 + }, + { + "epoch": 1.2942619453924915, + "grad_norm": 0.5819820761680603, + "learning_rate": 7.010809995795897e-06, + "loss": 0.487, + "step": 4045 + }, + { + "epoch": 1.2945819112627985, + "grad_norm": 0.5423453450202942, + "learning_rate": 7.009105255654647e-06, + "loss": 0.4607, + "step": 4046 + }, + { + "epoch": 1.2949018771331058, + "grad_norm": 0.5723846554756165, + "learning_rate": 7.007400236948536e-06, + "loss": 0.5675, + "step": 4047 + }, + { + "epoch": 1.295221843003413, + "grad_norm": 0.5019378066062927, + "learning_rate": 7.005694939913966e-06, + "loss": 0.4632, + "step": 4048 + }, + { + "epoch": 1.29554180887372, + "grad_norm": 0.6267055869102478, + "learning_rate": 7.00398936478738e-06, + "loss": 0.5521, + "step": 4049 + }, + { + "epoch": 1.2958617747440273, + "grad_norm": 0.573279082775116, + "learning_rate": 7.002283511805258e-06, + "loss": 0.5095, + "step": 4050 + }, + { + "epoch": 1.2961817406143346, + "grad_norm": 0.5108147859573364, + "learning_rate": 7.000577381204119e-06, + "loss": 0.5046, + "step": 4051 + }, + { + "epoch": 1.2965017064846416, + "grad_norm": 0.4877328872680664, + "learning_rate": 6.998870973220519e-06, + "loss": 0.4508, + "step": 4052 + }, + { + "epoch": 1.2968216723549488, + "grad_norm": 0.5888170599937439, + "learning_rate": 6.997164288091053e-06, + "loss": 0.5318, + "step": 4053 + }, + { + "epoch": 1.297141638225256, + "grad_norm": 0.5399531722068787, + "learning_rate": 6.995457326052355e-06, + "loss": 0.5082, + "step": 4054 + }, + { + "epoch": 1.297461604095563, + "grad_norm": 0.5535011291503906, + "learning_rate": 6.9937500873411e-06, + "loss": 0.53, + "step": 4055 + }, + { + "epoch": 1.2977815699658704, + "grad_norm": 0.522777795791626, + "learning_rate": 6.992042572193997e-06, + "loss": 0.4785, + "step": 4056 + }, + { + "epoch": 1.2981015358361776, + "grad_norm": 0.5063053369522095, + "learning_rate": 6.990334780847794e-06, + "loss": 0.4694, + "step": 4057 + }, + { + "epoch": 1.2984215017064846, + "grad_norm": 0.5899012088775635, + "learning_rate": 6.98862671353928e-06, + "loss": 0.5359, + "step": 4058 + }, + { + "epoch": 1.2987414675767919, + "grad_norm": 0.5376660823822021, + "learning_rate": 6.9869183705052775e-06, + "loss": 0.4811, + "step": 4059 + }, + { + "epoch": 1.299061433447099, + "grad_norm": 0.5922868251800537, + "learning_rate": 6.985209751982655e-06, + "loss": 0.5346, + "step": 4060 + }, + { + "epoch": 1.2993813993174061, + "grad_norm": 0.5054578185081482, + "learning_rate": 6.983500858208312e-06, + "loss": 0.4288, + "step": 4061 + }, + { + "epoch": 1.2997013651877132, + "grad_norm": 0.5970536470413208, + "learning_rate": 6.981791689419187e-06, + "loss": 0.5166, + "step": 4062 + }, + { + "epoch": 1.3000213310580204, + "grad_norm": 0.5270934700965881, + "learning_rate": 6.980082245852262e-06, + "loss": 0.4704, + "step": 4063 + }, + { + "epoch": 1.3003412969283277, + "grad_norm": 0.5640825033187866, + "learning_rate": 6.97837252774455e-06, + "loss": 0.4985, + "step": 4064 + }, + { + "epoch": 1.3006612627986347, + "grad_norm": 0.5541390776634216, + "learning_rate": 6.976662535333107e-06, + "loss": 0.5034, + "step": 4065 + }, + { + "epoch": 1.300981228668942, + "grad_norm": 0.5418012738227844, + "learning_rate": 6.974952268855028e-06, + "loss": 0.4515, + "step": 4066 + }, + { + "epoch": 1.3013011945392492, + "grad_norm": 0.5639212727546692, + "learning_rate": 6.9732417285474376e-06, + "loss": 0.482, + "step": 4067 + }, + { + "epoch": 1.3016211604095562, + "grad_norm": 0.5673478245735168, + "learning_rate": 6.971530914647511e-06, + "loss": 0.5271, + "step": 4068 + }, + { + "epoch": 1.3019411262798635, + "grad_norm": 0.5235918760299683, + "learning_rate": 6.969819827392449e-06, + "loss": 0.4988, + "step": 4069 + }, + { + "epoch": 1.3022610921501707, + "grad_norm": 0.4916918873786926, + "learning_rate": 6.9681084670194984e-06, + "loss": 0.4936, + "step": 4070 + }, + { + "epoch": 1.3025810580204777, + "grad_norm": 0.5602163672447205, + "learning_rate": 6.966396833765941e-06, + "loss": 0.5123, + "step": 4071 + }, + { + "epoch": 1.302901023890785, + "grad_norm": 0.5419959425926208, + "learning_rate": 6.964684927869098e-06, + "loss": 0.4499, + "step": 4072 + }, + { + "epoch": 1.3032209897610922, + "grad_norm": 0.5200892686843872, + "learning_rate": 6.9629727495663265e-06, + "loss": 0.4945, + "step": 4073 + }, + { + "epoch": 1.3035409556313993, + "grad_norm": 0.5808398127555847, + "learning_rate": 6.961260299095021e-06, + "loss": 0.4693, + "step": 4074 + }, + { + "epoch": 1.3038609215017065, + "grad_norm": 0.6151460409164429, + "learning_rate": 6.959547576692616e-06, + "loss": 0.5331, + "step": 4075 + }, + { + "epoch": 1.3041808873720138, + "grad_norm": 0.608762264251709, + "learning_rate": 6.9578345825965805e-06, + "loss": 0.52, + "step": 4076 + }, + { + "epoch": 1.3045008532423208, + "grad_norm": 0.5830143094062805, + "learning_rate": 6.956121317044426e-06, + "loss": 0.5071, + "step": 4077 + }, + { + "epoch": 1.304820819112628, + "grad_norm": 0.5641476511955261, + "learning_rate": 6.954407780273697e-06, + "loss": 0.5199, + "step": 4078 + }, + { + "epoch": 1.3051407849829353, + "grad_norm": 0.5512235164642334, + "learning_rate": 6.95269397252198e-06, + "loss": 0.4253, + "step": 4079 + }, + { + "epoch": 1.3054607508532423, + "grad_norm": 0.5953153371810913, + "learning_rate": 6.950979894026893e-06, + "loss": 0.5585, + "step": 4080 + }, + { + "epoch": 1.3057807167235496, + "grad_norm": 0.5572149157524109, + "learning_rate": 6.949265545026096e-06, + "loss": 0.4667, + "step": 4081 + }, + { + "epoch": 1.3061006825938566, + "grad_norm": 0.5659805536270142, + "learning_rate": 6.947550925757287e-06, + "loss": 0.5031, + "step": 4082 + }, + { + "epoch": 1.3064206484641638, + "grad_norm": 0.4794667363166809, + "learning_rate": 6.9458360364582e-06, + "loss": 0.4365, + "step": 4083 + }, + { + "epoch": 1.3067406143344709, + "grad_norm": 0.5645291209220886, + "learning_rate": 6.944120877366605e-06, + "loss": 0.6142, + "step": 4084 + }, + { + "epoch": 1.307060580204778, + "grad_norm": 0.5863152742385864, + "learning_rate": 6.942405448720311e-06, + "loss": 0.501, + "step": 4085 + }, + { + "epoch": 1.3073805460750854, + "grad_norm": 0.5941131711006165, + "learning_rate": 6.940689750757163e-06, + "loss": 0.5084, + "step": 4086 + }, + { + "epoch": 1.3077005119453924, + "grad_norm": 0.5305261611938477, + "learning_rate": 6.938973783715049e-06, + "loss": 0.5205, + "step": 4087 + }, + { + "epoch": 1.3080204778156996, + "grad_norm": 0.5063480138778687, + "learning_rate": 6.937257547831887e-06, + "loss": 0.4882, + "step": 4088 + }, + { + "epoch": 1.3083404436860069, + "grad_norm": 0.5177851915359497, + "learning_rate": 6.935541043345636e-06, + "loss": 0.5213, + "step": 4089 + }, + { + "epoch": 1.308660409556314, + "grad_norm": 0.5293291807174683, + "learning_rate": 6.933824270494289e-06, + "loss": 0.4641, + "step": 4090 + }, + { + "epoch": 1.3089803754266212, + "grad_norm": 0.5412163734436035, + "learning_rate": 6.932107229515881e-06, + "loss": 0.5234, + "step": 4091 + }, + { + "epoch": 1.3093003412969284, + "grad_norm": 0.5220082998275757, + "learning_rate": 6.930389920648481e-06, + "loss": 0.4499, + "step": 4092 + }, + { + "epoch": 1.3096203071672354, + "grad_norm": 0.631682276725769, + "learning_rate": 6.928672344130197e-06, + "loss": 0.5726, + "step": 4093 + }, + { + "epoch": 1.3099402730375427, + "grad_norm": 0.5830767154693604, + "learning_rate": 6.9269545001991735e-06, + "loss": 0.5176, + "step": 4094 + }, + { + "epoch": 1.31026023890785, + "grad_norm": 0.5310470461845398, + "learning_rate": 6.925236389093588e-06, + "loss": 0.4576, + "step": 4095 + }, + { + "epoch": 1.310580204778157, + "grad_norm": 0.5761337876319885, + "learning_rate": 6.923518011051662e-06, + "loss": 0.5753, + "step": 4096 + }, + { + "epoch": 1.3109001706484642, + "grad_norm": 0.5443764328956604, + "learning_rate": 6.92179936631165e-06, + "loss": 0.4763, + "step": 4097 + }, + { + "epoch": 1.3112201365187715, + "grad_norm": 0.5663328766822815, + "learning_rate": 6.920080455111843e-06, + "loss": 0.5491, + "step": 4098 + }, + { + "epoch": 1.3115401023890785, + "grad_norm": 0.540813148021698, + "learning_rate": 6.918361277690572e-06, + "loss": 0.4574, + "step": 4099 + }, + { + "epoch": 1.3118600682593857, + "grad_norm": 0.4971042573451996, + "learning_rate": 6.916641834286202e-06, + "loss": 0.4582, + "step": 4100 + }, + { + "epoch": 1.3121800341296928, + "grad_norm": 0.5601802468299866, + "learning_rate": 6.914922125137135e-06, + "loss": 0.5056, + "step": 4101 + }, + { + "epoch": 1.3125, + "grad_norm": 0.582165539264679, + "learning_rate": 6.9132021504818115e-06, + "loss": 0.5102, + "step": 4102 + }, + { + "epoch": 1.3128199658703072, + "grad_norm": 0.5567464232444763, + "learning_rate": 6.9114819105587085e-06, + "loss": 0.5152, + "step": 4103 + }, + { + "epoch": 1.3131399317406143, + "grad_norm": 0.5217399597167969, + "learning_rate": 6.90976140560634e-06, + "loss": 0.4812, + "step": 4104 + }, + { + "epoch": 1.3134598976109215, + "grad_norm": 0.5689828395843506, + "learning_rate": 6.9080406358632535e-06, + "loss": 0.4781, + "step": 4105 + }, + { + "epoch": 1.3137798634812285, + "grad_norm": 0.5900806784629822, + "learning_rate": 6.906319601568039e-06, + "loss": 0.5043, + "step": 4106 + }, + { + "epoch": 1.3140998293515358, + "grad_norm": 0.49958544969558716, + "learning_rate": 6.904598302959318e-06, + "loss": 0.4601, + "step": 4107 + }, + { + "epoch": 1.314419795221843, + "grad_norm": 0.5210117101669312, + "learning_rate": 6.90287674027575e-06, + "loss": 0.5583, + "step": 4108 + }, + { + "epoch": 1.31473976109215, + "grad_norm": 0.5850756764411926, + "learning_rate": 6.901154913756034e-06, + "loss": 0.5359, + "step": 4109 + }, + { + "epoch": 1.3150597269624573, + "grad_norm": 0.5848261713981628, + "learning_rate": 6.8994328236389006e-06, + "loss": 0.4849, + "step": 4110 + }, + { + "epoch": 1.3153796928327646, + "grad_norm": 0.5249413251876831, + "learning_rate": 6.8977104701631226e-06, + "loss": 0.5132, + "step": 4111 + }, + { + "epoch": 1.3156996587030716, + "grad_norm": 0.5254254937171936, + "learning_rate": 6.895987853567504e-06, + "loss": 0.4501, + "step": 4112 + }, + { + "epoch": 1.3160196245733788, + "grad_norm": 0.6095935702323914, + "learning_rate": 6.894264974090889e-06, + "loss": 0.5633, + "step": 4113 + }, + { + "epoch": 1.316339590443686, + "grad_norm": 0.5580556392669678, + "learning_rate": 6.892541831972157e-06, + "loss": 0.5085, + "step": 4114 + }, + { + "epoch": 1.3166595563139931, + "grad_norm": 0.5535349249839783, + "learning_rate": 6.890818427450221e-06, + "loss": 0.5184, + "step": 4115 + }, + { + "epoch": 1.3169795221843004, + "grad_norm": 0.5419041514396667, + "learning_rate": 6.8890947607640376e-06, + "loss": 0.4825, + "step": 4116 + }, + { + "epoch": 1.3172994880546076, + "grad_norm": 0.5677555799484253, + "learning_rate": 6.8873708321525925e-06, + "loss": 0.5134, + "step": 4117 + }, + { + "epoch": 1.3176194539249146, + "grad_norm": 0.5081667900085449, + "learning_rate": 6.88564664185491e-06, + "loss": 0.5181, + "step": 4118 + }, + { + "epoch": 1.317939419795222, + "grad_norm": 0.449253648519516, + "learning_rate": 6.883922190110054e-06, + "loss": 0.4367, + "step": 4119 + }, + { + "epoch": 1.3182593856655291, + "grad_norm": 0.5432609915733337, + "learning_rate": 6.882197477157118e-06, + "loss": 0.5034, + "step": 4120 + }, + { + "epoch": 1.3185793515358362, + "grad_norm": 0.5242462754249573, + "learning_rate": 6.880472503235238e-06, + "loss": 0.4599, + "step": 4121 + }, + { + "epoch": 1.3188993174061434, + "grad_norm": 0.5869811177253723, + "learning_rate": 6.878747268583584e-06, + "loss": 0.5204, + "step": 4122 + }, + { + "epoch": 1.3192192832764504, + "grad_norm": 0.6002717018127441, + "learning_rate": 6.8770217734413606e-06, + "loss": 0.5206, + "step": 4123 + }, + { + "epoch": 1.3195392491467577, + "grad_norm": 0.5224626064300537, + "learning_rate": 6.87529601804781e-06, + "loss": 0.4443, + "step": 4124 + }, + { + "epoch": 1.3198592150170647, + "grad_norm": 0.5698513984680176, + "learning_rate": 6.873570002642209e-06, + "loss": 0.5324, + "step": 4125 + }, + { + "epoch": 1.320179180887372, + "grad_norm": 0.555054247379303, + "learning_rate": 6.871843727463876e-06, + "loss": 0.4941, + "step": 4126 + }, + { + "epoch": 1.3204991467576792, + "grad_norm": 0.6183741092681885, + "learning_rate": 6.870117192752157e-06, + "loss": 0.5522, + "step": 4127 + }, + { + "epoch": 1.3208191126279862, + "grad_norm": 0.5069617629051208, + "learning_rate": 6.86839039874644e-06, + "loss": 0.4479, + "step": 4128 + }, + { + "epoch": 1.3211390784982935, + "grad_norm": 0.5739780068397522, + "learning_rate": 6.866663345686147e-06, + "loss": 0.4713, + "step": 4129 + }, + { + "epoch": 1.3214590443686007, + "grad_norm": 0.6382153034210205, + "learning_rate": 6.8649360338107315e-06, + "loss": 0.5124, + "step": 4130 + }, + { + "epoch": 1.3217790102389078, + "grad_norm": 0.614399254322052, + "learning_rate": 6.8632084633596964e-06, + "loss": 0.5227, + "step": 4131 + }, + { + "epoch": 1.322098976109215, + "grad_norm": 0.5141477584838867, + "learning_rate": 6.861480634572564e-06, + "loss": 0.4562, + "step": 4132 + }, + { + "epoch": 1.3224189419795223, + "grad_norm": 0.5629687309265137, + "learning_rate": 6.859752547688902e-06, + "loss": 0.4857, + "step": 4133 + }, + { + "epoch": 1.3227389078498293, + "grad_norm": 0.607196569442749, + "learning_rate": 6.858024202948312e-06, + "loss": 0.5035, + "step": 4134 + }, + { + "epoch": 1.3230588737201365, + "grad_norm": 0.5607357621192932, + "learning_rate": 6.85629560059043e-06, + "loss": 0.4983, + "step": 4135 + }, + { + "epoch": 1.3233788395904438, + "grad_norm": 0.5918465852737427, + "learning_rate": 6.854566740854932e-06, + "loss": 0.5253, + "step": 4136 + }, + { + "epoch": 1.3236988054607508, + "grad_norm": 0.5611516237258911, + "learning_rate": 6.852837623981523e-06, + "loss": 0.465, + "step": 4137 + }, + { + "epoch": 1.324018771331058, + "grad_norm": 0.5750157833099365, + "learning_rate": 6.851108250209949e-06, + "loss": 0.531, + "step": 4138 + }, + { + "epoch": 1.3243387372013653, + "grad_norm": 0.5478917360305786, + "learning_rate": 6.849378619779989e-06, + "loss": 0.4976, + "step": 4139 + }, + { + "epoch": 1.3246587030716723, + "grad_norm": 0.4693235456943512, + "learning_rate": 6.847648732931458e-06, + "loss": 0.4736, + "step": 4140 + }, + { + "epoch": 1.3249786689419796, + "grad_norm": 0.5775055289268494, + "learning_rate": 6.845918589904209e-06, + "loss": 0.4965, + "step": 4141 + }, + { + "epoch": 1.3252986348122868, + "grad_norm": 0.6024038195610046, + "learning_rate": 6.844188190938125e-06, + "loss": 0.496, + "step": 4142 + }, + { + "epoch": 1.3256186006825939, + "grad_norm": 0.6416028738021851, + "learning_rate": 6.84245753627313e-06, + "loss": 0.494, + "step": 4143 + }, + { + "epoch": 1.325938566552901, + "grad_norm": 0.5104652643203735, + "learning_rate": 6.840726626149182e-06, + "loss": 0.4953, + "step": 4144 + }, + { + "epoch": 1.3262585324232081, + "grad_norm": 0.5593465566635132, + "learning_rate": 6.838995460806272e-06, + "loss": 0.4754, + "step": 4145 + }, + { + "epoch": 1.3265784982935154, + "grad_norm": 0.5791028141975403, + "learning_rate": 6.837264040484431e-06, + "loss": 0.4951, + "step": 4146 + }, + { + "epoch": 1.3268984641638224, + "grad_norm": 0.5533764958381653, + "learning_rate": 6.835532365423716e-06, + "loss": 0.5162, + "step": 4147 + }, + { + "epoch": 1.3272184300341296, + "grad_norm": 0.5973724722862244, + "learning_rate": 6.833800435864235e-06, + "loss": 0.506, + "step": 4148 + }, + { + "epoch": 1.327538395904437, + "grad_norm": 0.5919985771179199, + "learning_rate": 6.832068252046116e-06, + "loss": 0.4638, + "step": 4149 + }, + { + "epoch": 1.327858361774744, + "grad_norm": 0.5745266079902649, + "learning_rate": 6.830335814209527e-06, + "loss": 0.481, + "step": 4150 + }, + { + "epoch": 1.3281783276450512, + "grad_norm": 0.5681392550468445, + "learning_rate": 6.828603122594679e-06, + "loss": 0.4785, + "step": 4151 + }, + { + "epoch": 1.3284982935153584, + "grad_norm": 0.5746228098869324, + "learning_rate": 6.826870177441807e-06, + "loss": 0.5011, + "step": 4152 + }, + { + "epoch": 1.3288182593856654, + "grad_norm": 0.5923866629600525, + "learning_rate": 6.825136978991187e-06, + "loss": 0.5005, + "step": 4153 + }, + { + "epoch": 1.3291382252559727, + "grad_norm": 0.5296605825424194, + "learning_rate": 6.82340352748313e-06, + "loss": 0.5353, + "step": 4154 + }, + { + "epoch": 1.32945819112628, + "grad_norm": 0.570958137512207, + "learning_rate": 6.821669823157979e-06, + "loss": 0.4754, + "step": 4155 + }, + { + "epoch": 1.329778156996587, + "grad_norm": 0.5462219715118408, + "learning_rate": 6.819935866256116e-06, + "loss": 0.479, + "step": 4156 + }, + { + "epoch": 1.3300981228668942, + "grad_norm": 0.5859329700469971, + "learning_rate": 6.818201657017955e-06, + "loss": 0.5163, + "step": 4157 + }, + { + "epoch": 1.3304180887372015, + "grad_norm": 0.5124255418777466, + "learning_rate": 6.816467195683949e-06, + "loss": 0.4648, + "step": 4158 + }, + { + "epoch": 1.3307380546075085, + "grad_norm": 0.520322859287262, + "learning_rate": 6.814732482494579e-06, + "loss": 0.5181, + "step": 4159 + }, + { + "epoch": 1.3310580204778157, + "grad_norm": 0.5678321719169617, + "learning_rate": 6.812997517690368e-06, + "loss": 0.5017, + "step": 4160 + }, + { + "epoch": 1.331377986348123, + "grad_norm": 0.5411324501037598, + "learning_rate": 6.8112623015118695e-06, + "loss": 0.4702, + "step": 4161 + }, + { + "epoch": 1.33169795221843, + "grad_norm": 0.5200678706169128, + "learning_rate": 6.809526834199675e-06, + "loss": 0.491, + "step": 4162 + }, + { + "epoch": 1.3320179180887373, + "grad_norm": 0.4977427124977112, + "learning_rate": 6.807791115994408e-06, + "loss": 0.5364, + "step": 4163 + }, + { + "epoch": 1.3323378839590443, + "grad_norm": 0.5493211150169373, + "learning_rate": 6.806055147136727e-06, + "loss": 0.5077, + "step": 4164 + }, + { + "epoch": 1.3326578498293515, + "grad_norm": 0.5668837428092957, + "learning_rate": 6.804318927867327e-06, + "loss": 0.4819, + "step": 4165 + }, + { + "epoch": 1.3329778156996588, + "grad_norm": 0.6170305609703064, + "learning_rate": 6.802582458426937e-06, + "loss": 0.5639, + "step": 4166 + }, + { + "epoch": 1.3332977815699658, + "grad_norm": 0.46503856778144836, + "learning_rate": 6.800845739056321e-06, + "loss": 0.4498, + "step": 4167 + }, + { + "epoch": 1.333617747440273, + "grad_norm": 0.4986816942691803, + "learning_rate": 6.799108769996279e-06, + "loss": 0.5055, + "step": 4168 + }, + { + "epoch": 1.33393771331058, + "grad_norm": 0.5867021083831787, + "learning_rate": 6.797371551487638e-06, + "loss": 0.5206, + "step": 4169 + }, + { + "epoch": 1.3342576791808873, + "grad_norm": 0.5245837569236755, + "learning_rate": 6.795634083771271e-06, + "loss": 0.4289, + "step": 4170 + }, + { + "epoch": 1.3345776450511946, + "grad_norm": 0.5654990077018738, + "learning_rate": 6.793896367088077e-06, + "loss": 0.53, + "step": 4171 + }, + { + "epoch": 1.3348976109215016, + "grad_norm": 0.5152663588523865, + "learning_rate": 6.7921584016789945e-06, + "loss": 0.5477, + "step": 4172 + }, + { + "epoch": 1.3352175767918089, + "grad_norm": 0.5061206221580505, + "learning_rate": 6.790420187784993e-06, + "loss": 0.4932, + "step": 4173 + }, + { + "epoch": 1.335537542662116, + "grad_norm": 0.5259354710578918, + "learning_rate": 6.788681725647079e-06, + "loss": 0.5231, + "step": 4174 + }, + { + "epoch": 1.3358575085324231, + "grad_norm": 0.45490285754203796, + "learning_rate": 6.786943015506292e-06, + "loss": 0.4411, + "step": 4175 + }, + { + "epoch": 1.3361774744027304, + "grad_norm": 0.5072759389877319, + "learning_rate": 6.785204057603706e-06, + "loss": 0.4343, + "step": 4176 + }, + { + "epoch": 1.3364974402730376, + "grad_norm": 0.5490756630897522, + "learning_rate": 6.78346485218043e-06, + "loss": 0.4856, + "step": 4177 + }, + { + "epoch": 1.3368174061433447, + "grad_norm": 0.5347392559051514, + "learning_rate": 6.781725399477606e-06, + "loss": 0.5403, + "step": 4178 + }, + { + "epoch": 1.337137372013652, + "grad_norm": 0.48878544569015503, + "learning_rate": 6.779985699736413e-06, + "loss": 0.4763, + "step": 4179 + }, + { + "epoch": 1.3374573378839592, + "grad_norm": 0.5018070340156555, + "learning_rate": 6.778245753198061e-06, + "loss": 0.5074, + "step": 4180 + }, + { + "epoch": 1.3377773037542662, + "grad_norm": 0.5557435750961304, + "learning_rate": 6.776505560103795e-06, + "loss": 0.453, + "step": 4181 + }, + { + "epoch": 1.3380972696245734, + "grad_norm": 0.6862342357635498, + "learning_rate": 6.774765120694896e-06, + "loss": 0.5549, + "step": 4182 + }, + { + "epoch": 1.3384172354948807, + "grad_norm": 0.5224938988685608, + "learning_rate": 6.773024435212678e-06, + "loss": 0.4618, + "step": 4183 + }, + { + "epoch": 1.3387372013651877, + "grad_norm": 0.5360044240951538, + "learning_rate": 6.771283503898492e-06, + "loss": 0.5176, + "step": 4184 + }, + { + "epoch": 1.339057167235495, + "grad_norm": 0.5369170904159546, + "learning_rate": 6.769542326993713e-06, + "loss": 0.534, + "step": 4185 + }, + { + "epoch": 1.339377133105802, + "grad_norm": 0.48003724217414856, + "learning_rate": 6.767800904739764e-06, + "loss": 0.4751, + "step": 4186 + }, + { + "epoch": 1.3396970989761092, + "grad_norm": 0.5560697913169861, + "learning_rate": 6.766059237378092e-06, + "loss": 0.4935, + "step": 4187 + }, + { + "epoch": 1.3400170648464163, + "grad_norm": 0.5447647571563721, + "learning_rate": 6.764317325150183e-06, + "loss": 0.5242, + "step": 4188 + }, + { + "epoch": 1.3403370307167235, + "grad_norm": 0.4936341345310211, + "learning_rate": 6.762575168297554e-06, + "loss": 0.4576, + "step": 4189 + }, + { + "epoch": 1.3406569965870307, + "grad_norm": 0.6130675077438354, + "learning_rate": 6.760832767061756e-06, + "loss": 0.5732, + "step": 4190 + }, + { + "epoch": 1.3409769624573378, + "grad_norm": 0.5237568616867065, + "learning_rate": 6.7590901216843775e-06, + "loss": 0.4388, + "step": 4191 + }, + { + "epoch": 1.341296928327645, + "grad_norm": 0.5610059499740601, + "learning_rate": 6.757347232407037e-06, + "loss": 0.5656, + "step": 4192 + }, + { + "epoch": 1.3416168941979523, + "grad_norm": 0.5444913506507874, + "learning_rate": 6.75560409947139e-06, + "loss": 0.466, + "step": 4193 + }, + { + "epoch": 1.3419368600682593, + "grad_norm": 0.5468993186950684, + "learning_rate": 6.753860723119122e-06, + "loss": 0.5077, + "step": 4194 + }, + { + "epoch": 1.3422568259385665, + "grad_norm": 0.5734399557113647, + "learning_rate": 6.752117103591953e-06, + "loss": 0.5201, + "step": 4195 + }, + { + "epoch": 1.3425767918088738, + "grad_norm": 0.491445392370224, + "learning_rate": 6.750373241131643e-06, + "loss": 0.4913, + "step": 4196 + }, + { + "epoch": 1.3428967576791808, + "grad_norm": 0.5285682082176208, + "learning_rate": 6.748629135979975e-06, + "loss": 0.5144, + "step": 4197 + }, + { + "epoch": 1.343216723549488, + "grad_norm": 0.47706079483032227, + "learning_rate": 6.746884788378774e-06, + "loss": 0.475, + "step": 4198 + }, + { + "epoch": 1.3435366894197953, + "grad_norm": 0.5360599160194397, + "learning_rate": 6.745140198569897e-06, + "loss": 0.5166, + "step": 4199 + }, + { + "epoch": 1.3438566552901023, + "grad_norm": 0.4963415265083313, + "learning_rate": 6.743395366795229e-06, + "loss": 0.4649, + "step": 4200 + }, + { + "epoch": 1.3441766211604096, + "grad_norm": 0.4736952483654022, + "learning_rate": 6.7416502932967e-06, + "loss": 0.4616, + "step": 4201 + }, + { + "epoch": 1.3444965870307168, + "grad_norm": 0.5033711194992065, + "learning_rate": 6.73990497831626e-06, + "loss": 0.4607, + "step": 4202 + }, + { + "epoch": 1.3448165529010239, + "grad_norm": 0.515525221824646, + "learning_rate": 6.738159422095902e-06, + "loss": 0.5365, + "step": 4203 + }, + { + "epoch": 1.3451365187713311, + "grad_norm": 0.5089691281318665, + "learning_rate": 6.736413624877649e-06, + "loss": 0.4572, + "step": 4204 + }, + { + "epoch": 1.3454564846416384, + "grad_norm": 0.5544673204421997, + "learning_rate": 6.734667586903557e-06, + "loss": 0.495, + "step": 4205 + }, + { + "epoch": 1.3457764505119454, + "grad_norm": 0.533311128616333, + "learning_rate": 6.732921308415719e-06, + "loss": 0.5134, + "step": 4206 + }, + { + "epoch": 1.3460964163822526, + "grad_norm": 0.549479067325592, + "learning_rate": 6.7311747896562565e-06, + "loss": 0.4981, + "step": 4207 + }, + { + "epoch": 1.3464163822525597, + "grad_norm": 0.5087676048278809, + "learning_rate": 6.729428030867325e-06, + "loss": 0.4681, + "step": 4208 + }, + { + "epoch": 1.346736348122867, + "grad_norm": 0.5456358790397644, + "learning_rate": 6.72768103229112e-06, + "loss": 0.4963, + "step": 4209 + }, + { + "epoch": 1.347056313993174, + "grad_norm": 0.539222776889801, + "learning_rate": 6.725933794169858e-06, + "loss": 0.4599, + "step": 4210 + }, + { + "epoch": 1.3473762798634812, + "grad_norm": 0.5186460018157959, + "learning_rate": 6.724186316745801e-06, + "loss": 0.5047, + "step": 4211 + }, + { + "epoch": 1.3476962457337884, + "grad_norm": 0.5008431673049927, + "learning_rate": 6.722438600261234e-06, + "loss": 0.4849, + "step": 4212 + }, + { + "epoch": 1.3480162116040955, + "grad_norm": 0.525508463382721, + "learning_rate": 6.7206906449584875e-06, + "loss": 0.4877, + "step": 4213 + }, + { + "epoch": 1.3483361774744027, + "grad_norm": 0.522044837474823, + "learning_rate": 6.718942451079911e-06, + "loss": 0.5136, + "step": 4214 + }, + { + "epoch": 1.34865614334471, + "grad_norm": 0.4838278293609619, + "learning_rate": 6.7171940188678945e-06, + "loss": 0.4731, + "step": 4215 + }, + { + "epoch": 1.348976109215017, + "grad_norm": 0.5309022068977356, + "learning_rate": 6.715445348564863e-06, + "loss": 0.5491, + "step": 4216 + }, + { + "epoch": 1.3492960750853242, + "grad_norm": 0.5318649411201477, + "learning_rate": 6.713696440413269e-06, + "loss": 0.5208, + "step": 4217 + }, + { + "epoch": 1.3496160409556315, + "grad_norm": 0.5618005394935608, + "learning_rate": 6.711947294655603e-06, + "loss": 0.4938, + "step": 4218 + }, + { + "epoch": 1.3499360068259385, + "grad_norm": 0.5245568752288818, + "learning_rate": 6.710197911534384e-06, + "loss": 0.4842, + "step": 4219 + }, + { + "epoch": 1.3502559726962458, + "grad_norm": 0.5300405025482178, + "learning_rate": 6.708448291292168e-06, + "loss": 0.4699, + "step": 4220 + }, + { + "epoch": 1.350575938566553, + "grad_norm": 0.5263601541519165, + "learning_rate": 6.706698434171542e-06, + "loss": 0.4887, + "step": 4221 + }, + { + "epoch": 1.35089590443686, + "grad_norm": 0.5631449818611145, + "learning_rate": 6.704948340415122e-06, + "loss": 0.4885, + "step": 4222 + }, + { + "epoch": 1.3512158703071673, + "grad_norm": 0.5387476682662964, + "learning_rate": 6.703198010265566e-06, + "loss": 0.4892, + "step": 4223 + }, + { + "epoch": 1.3515358361774745, + "grad_norm": 0.5017451643943787, + "learning_rate": 6.701447443965557e-06, + "loss": 0.5033, + "step": 4224 + }, + { + "epoch": 1.3518558020477816, + "grad_norm": 0.48120734095573425, + "learning_rate": 6.699696641757811e-06, + "loss": 0.5109, + "step": 4225 + }, + { + "epoch": 1.3521757679180888, + "grad_norm": 0.48019859194755554, + "learning_rate": 6.697945603885082e-06, + "loss": 0.4445, + "step": 4226 + }, + { + "epoch": 1.3524957337883958, + "grad_norm": 0.529334545135498, + "learning_rate": 6.6961943305901515e-06, + "loss": 0.507, + "step": 4227 + }, + { + "epoch": 1.352815699658703, + "grad_norm": 0.5923269391059875, + "learning_rate": 6.694442822115839e-06, + "loss": 0.5348, + "step": 4228 + }, + { + "epoch": 1.35313566552901, + "grad_norm": 0.5207008123397827, + "learning_rate": 6.692691078704988e-06, + "loss": 0.5012, + "step": 4229 + }, + { + "epoch": 1.3534556313993173, + "grad_norm": 0.5076130032539368, + "learning_rate": 6.690939100600483e-06, + "loss": 0.477, + "step": 4230 + }, + { + "epoch": 1.3537755972696246, + "grad_norm": 0.5776696801185608, + "learning_rate": 6.689186888045239e-06, + "loss": 0.4944, + "step": 4231 + }, + { + "epoch": 1.3540955631399316, + "grad_norm": 0.549407422542572, + "learning_rate": 6.687434441282199e-06, + "loss": 0.4745, + "step": 4232 + }, + { + "epoch": 1.3544155290102389, + "grad_norm": 0.5382245182991028, + "learning_rate": 6.685681760554346e-06, + "loss": 0.5273, + "step": 4233 + }, + { + "epoch": 1.3547354948805461, + "grad_norm": 0.5502383708953857, + "learning_rate": 6.683928846104687e-06, + "loss": 0.4917, + "step": 4234 + }, + { + "epoch": 1.3550554607508531, + "grad_norm": 0.5238730311393738, + "learning_rate": 6.6821756981762674e-06, + "loss": 0.5307, + "step": 4235 + }, + { + "epoch": 1.3553754266211604, + "grad_norm": 0.5505203008651733, + "learning_rate": 6.680422317012165e-06, + "loss": 0.4522, + "step": 4236 + }, + { + "epoch": 1.3556953924914676, + "grad_norm": 0.5909242630004883, + "learning_rate": 6.678668702855485e-06, + "loss": 0.5243, + "step": 4237 + }, + { + "epoch": 1.3560153583617747, + "grad_norm": 0.5503251552581787, + "learning_rate": 6.676914855949372e-06, + "loss": 0.5046, + "step": 4238 + }, + { + "epoch": 1.356335324232082, + "grad_norm": 0.516450822353363, + "learning_rate": 6.675160776536996e-06, + "loss": 0.4225, + "step": 4239 + }, + { + "epoch": 1.3566552901023892, + "grad_norm": 0.5583712458610535, + "learning_rate": 6.673406464861563e-06, + "loss": 0.5473, + "step": 4240 + }, + { + "epoch": 1.3569752559726962, + "grad_norm": 0.4808334708213806, + "learning_rate": 6.671651921166311e-06, + "loss": 0.4481, + "step": 4241 + }, + { + "epoch": 1.3572952218430034, + "grad_norm": 0.5560965538024902, + "learning_rate": 6.669897145694507e-06, + "loss": 0.5281, + "step": 4242 + }, + { + "epoch": 1.3576151877133107, + "grad_norm": 0.5860428810119629, + "learning_rate": 6.668142138689458e-06, + "loss": 0.5322, + "step": 4243 + }, + { + "epoch": 1.3579351535836177, + "grad_norm": 0.5544723272323608, + "learning_rate": 6.666386900394493e-06, + "loss": 0.4332, + "step": 4244 + }, + { + "epoch": 1.358255119453925, + "grad_norm": 0.5531538128852844, + "learning_rate": 6.664631431052978e-06, + "loss": 0.5327, + "step": 4245 + }, + { + "epoch": 1.3585750853242322, + "grad_norm": 0.5179780125617981, + "learning_rate": 6.662875730908313e-06, + "loss": 0.4247, + "step": 4246 + }, + { + "epoch": 1.3588950511945392, + "grad_norm": 0.6032951474189758, + "learning_rate": 6.661119800203929e-06, + "loss": 0.5362, + "step": 4247 + }, + { + "epoch": 1.3592150170648465, + "grad_norm": 0.6043846011161804, + "learning_rate": 6.659363639183287e-06, + "loss": 0.5117, + "step": 4248 + }, + { + "epoch": 1.3595349829351535, + "grad_norm": 0.5354533195495605, + "learning_rate": 6.657607248089879e-06, + "loss": 0.4898, + "step": 4249 + }, + { + "epoch": 1.3598549488054608, + "grad_norm": 0.541419506072998, + "learning_rate": 6.65585062716723e-06, + "loss": 0.4887, + "step": 4250 + }, + { + "epoch": 1.3601749146757678, + "grad_norm": 0.5310930609703064, + "learning_rate": 6.654093776658902e-06, + "loss": 0.452, + "step": 4251 + }, + { + "epoch": 1.360494880546075, + "grad_norm": 0.6092569828033447, + "learning_rate": 6.652336696808481e-06, + "loss": 0.5665, + "step": 4252 + }, + { + "epoch": 1.3608148464163823, + "grad_norm": 0.5411043167114258, + "learning_rate": 6.65057938785959e-06, + "loss": 0.4336, + "step": 4253 + }, + { + "epoch": 1.3611348122866893, + "grad_norm": 0.627030074596405, + "learning_rate": 6.648821850055881e-06, + "loss": 0.5466, + "step": 4254 + }, + { + "epoch": 1.3614547781569966, + "grad_norm": 0.5255821943283081, + "learning_rate": 6.6470640836410385e-06, + "loss": 0.4861, + "step": 4255 + }, + { + "epoch": 1.3617747440273038, + "grad_norm": 0.5354859232902527, + "learning_rate": 6.6453060888587785e-06, + "loss": 0.5151, + "step": 4256 + }, + { + "epoch": 1.3620947098976108, + "grad_norm": 0.47319358587265015, + "learning_rate": 6.6435478659528505e-06, + "loss": 0.4744, + "step": 4257 + }, + { + "epoch": 1.362414675767918, + "grad_norm": 0.4874507486820221, + "learning_rate": 6.641789415167035e-06, + "loss": 0.4122, + "step": 4258 + }, + { + "epoch": 1.3627346416382253, + "grad_norm": 0.679909348487854, + "learning_rate": 6.64003073674514e-06, + "loss": 0.6077, + "step": 4259 + }, + { + "epoch": 1.3630546075085324, + "grad_norm": 0.57915860414505, + "learning_rate": 6.638271830931011e-06, + "loss": 0.4758, + "step": 4260 + }, + { + "epoch": 1.3633745733788396, + "grad_norm": 0.5016781091690063, + "learning_rate": 6.6365126979685226e-06, + "loss": 0.4749, + "step": 4261 + }, + { + "epoch": 1.3636945392491469, + "grad_norm": 0.5453657507896423, + "learning_rate": 6.634753338101578e-06, + "loss": 0.5459, + "step": 4262 + }, + { + "epoch": 1.3640145051194539, + "grad_norm": 0.5285664796829224, + "learning_rate": 6.632993751574119e-06, + "loss": 0.4907, + "step": 4263 + }, + { + "epoch": 1.3643344709897611, + "grad_norm": 0.5597354173660278, + "learning_rate": 6.6312339386301106e-06, + "loss": 0.5013, + "step": 4264 + }, + { + "epoch": 1.3646544368600684, + "grad_norm": 0.47672930359840393, + "learning_rate": 6.6294738995135545e-06, + "loss": 0.4441, + "step": 4265 + }, + { + "epoch": 1.3649744027303754, + "grad_norm": 0.5141209959983826, + "learning_rate": 6.62771363446848e-06, + "loss": 0.4799, + "step": 4266 + }, + { + "epoch": 1.3652943686006827, + "grad_norm": 0.5273243188858032, + "learning_rate": 6.625953143738954e-06, + "loss": 0.4501, + "step": 4267 + }, + { + "epoch": 1.36561433447099, + "grad_norm": 0.5881574749946594, + "learning_rate": 6.624192427569067e-06, + "loss": 0.5606, + "step": 4268 + }, + { + "epoch": 1.365934300341297, + "grad_norm": 0.5605153441429138, + "learning_rate": 6.622431486202948e-06, + "loss": 0.505, + "step": 4269 + }, + { + "epoch": 1.3662542662116042, + "grad_norm": 0.5217699408531189, + "learning_rate": 6.62067031988475e-06, + "loss": 0.5574, + "step": 4270 + }, + { + "epoch": 1.3665742320819112, + "grad_norm": 0.5430951714515686, + "learning_rate": 6.618908928858663e-06, + "loss": 0.4656, + "step": 4271 + }, + { + "epoch": 1.3668941979522184, + "grad_norm": 0.5407613515853882, + "learning_rate": 6.6171473133689054e-06, + "loss": 0.4657, + "step": 4272 + }, + { + "epoch": 1.3672141638225255, + "grad_norm": 0.5902913212776184, + "learning_rate": 6.615385473659726e-06, + "loss": 0.5697, + "step": 4273 + }, + { + "epoch": 1.3675341296928327, + "grad_norm": 0.4799916744232178, + "learning_rate": 6.61362340997541e-06, + "loss": 0.4782, + "step": 4274 + }, + { + "epoch": 1.36785409556314, + "grad_norm": 0.48387208580970764, + "learning_rate": 6.611861122560266e-06, + "loss": 0.4736, + "step": 4275 + }, + { + "epoch": 1.368174061433447, + "grad_norm": 0.5645774006843567, + "learning_rate": 6.610098611658637e-06, + "loss": 0.4813, + "step": 4276 + }, + { + "epoch": 1.3684940273037542, + "grad_norm": 0.6046881079673767, + "learning_rate": 6.6083358775149e-06, + "loss": 0.5181, + "step": 4277 + }, + { + "epoch": 1.3688139931740615, + "grad_norm": 0.48757874965667725, + "learning_rate": 6.606572920373457e-06, + "loss": 0.4181, + "step": 4278 + }, + { + "epoch": 1.3691339590443685, + "grad_norm": 0.5753948092460632, + "learning_rate": 6.604809740478748e-06, + "loss": 0.5632, + "step": 4279 + }, + { + "epoch": 1.3694539249146758, + "grad_norm": 0.5969665050506592, + "learning_rate": 6.6030463380752355e-06, + "loss": 0.5201, + "step": 4280 + }, + { + "epoch": 1.369773890784983, + "grad_norm": 0.5898136496543884, + "learning_rate": 6.601282713407421e-06, + "loss": 0.5166, + "step": 4281 + }, + { + "epoch": 1.37009385665529, + "grad_norm": 0.5735839605331421, + "learning_rate": 6.599518866719832e-06, + "loss": 0.5284, + "step": 4282 + }, + { + "epoch": 1.3704138225255973, + "grad_norm": 0.49882665276527405, + "learning_rate": 6.597754798257028e-06, + "loss": 0.4944, + "step": 4283 + }, + { + "epoch": 1.3707337883959045, + "grad_norm": 0.6059930324554443, + "learning_rate": 6.595990508263599e-06, + "loss": 0.5447, + "step": 4284 + }, + { + "epoch": 1.3710537542662116, + "grad_norm": 0.48686307668685913, + "learning_rate": 6.594225996984165e-06, + "loss": 0.4422, + "step": 4285 + }, + { + "epoch": 1.3713737201365188, + "grad_norm": 0.492951899766922, + "learning_rate": 6.592461264663379e-06, + "loss": 0.4841, + "step": 4286 + }, + { + "epoch": 1.371693686006826, + "grad_norm": 0.5390515327453613, + "learning_rate": 6.590696311545923e-06, + "loss": 0.5077, + "step": 4287 + }, + { + "epoch": 1.372013651877133, + "grad_norm": 0.5391637086868286, + "learning_rate": 6.588931137876511e-06, + "loss": 0.4964, + "step": 4288 + }, + { + "epoch": 1.3723336177474403, + "grad_norm": 0.5298265218734741, + "learning_rate": 6.587165743899884e-06, + "loss": 0.4985, + "step": 4289 + }, + { + "epoch": 1.3726535836177474, + "grad_norm": 0.5959964990615845, + "learning_rate": 6.585400129860816e-06, + "loss": 0.535, + "step": 4290 + }, + { + "epoch": 1.3729735494880546, + "grad_norm": 0.4907185733318329, + "learning_rate": 6.583634296004116e-06, + "loss": 0.4545, + "step": 4291 + }, + { + "epoch": 1.3732935153583616, + "grad_norm": 0.4811200499534607, + "learning_rate": 6.581868242574613e-06, + "loss": 0.4626, + "step": 4292 + }, + { + "epoch": 1.3736134812286689, + "grad_norm": 0.5177947282791138, + "learning_rate": 6.580101969817176e-06, + "loss": 0.4521, + "step": 4293 + }, + { + "epoch": 1.3739334470989761, + "grad_norm": 0.5208389163017273, + "learning_rate": 6.578335477976699e-06, + "loss": 0.4762, + "step": 4294 + }, + { + "epoch": 1.3742534129692832, + "grad_norm": 0.6004898548126221, + "learning_rate": 6.5765687672981104e-06, + "loss": 0.4973, + "step": 4295 + }, + { + "epoch": 1.3745733788395904, + "grad_norm": 0.5464395880699158, + "learning_rate": 6.574801838026366e-06, + "loss": 0.5322, + "step": 4296 + }, + { + "epoch": 1.3748933447098977, + "grad_norm": 0.5355855822563171, + "learning_rate": 6.57303469040645e-06, + "loss": 0.4759, + "step": 4297 + }, + { + "epoch": 1.3752133105802047, + "grad_norm": 0.6042685508728027, + "learning_rate": 6.571267324683383e-06, + "loss": 0.5253, + "step": 4298 + }, + { + "epoch": 1.375533276450512, + "grad_norm": 0.5859010219573975, + "learning_rate": 6.5694997411022125e-06, + "loss": 0.4887, + "step": 4299 + }, + { + "epoch": 1.3758532423208192, + "grad_norm": 0.5045570731163025, + "learning_rate": 6.567731939908013e-06, + "loss": 0.4559, + "step": 4300 + }, + { + "epoch": 1.3761732081911262, + "grad_norm": 0.5356442332267761, + "learning_rate": 6.565963921345896e-06, + "loss": 0.5476, + "step": 4301 + }, + { + "epoch": 1.3764931740614335, + "grad_norm": 0.4713646173477173, + "learning_rate": 6.564195685660994e-06, + "loss": 0.4164, + "step": 4302 + }, + { + "epoch": 1.3768131399317407, + "grad_norm": 0.5834218263626099, + "learning_rate": 6.562427233098482e-06, + "loss": 0.536, + "step": 4303 + }, + { + "epoch": 1.3771331058020477, + "grad_norm": 0.4851571023464203, + "learning_rate": 6.560658563903552e-06, + "loss": 0.3982, + "step": 4304 + }, + { + "epoch": 1.377453071672355, + "grad_norm": 0.6995872259140015, + "learning_rate": 6.558889678321436e-06, + "loss": 0.534, + "step": 4305 + }, + { + "epoch": 1.3777730375426622, + "grad_norm": 0.5719860792160034, + "learning_rate": 6.557120576597391e-06, + "loss": 0.516, + "step": 4306 + }, + { + "epoch": 1.3780930034129693, + "grad_norm": 0.5261143445968628, + "learning_rate": 6.5553512589767035e-06, + "loss": 0.495, + "step": 4307 + }, + { + "epoch": 1.3784129692832765, + "grad_norm": 0.5705453157424927, + "learning_rate": 6.553581725704694e-06, + "loss": 0.4526, + "step": 4308 + }, + { + "epoch": 1.3787329351535837, + "grad_norm": 0.5274971127510071, + "learning_rate": 6.551811977026707e-06, + "loss": 0.4554, + "step": 4309 + }, + { + "epoch": 1.3790529010238908, + "grad_norm": 0.47913479804992676, + "learning_rate": 6.550042013188125e-06, + "loss": 0.4548, + "step": 4310 + }, + { + "epoch": 1.379372866894198, + "grad_norm": 0.6112393140792847, + "learning_rate": 6.548271834434352e-06, + "loss": 0.5369, + "step": 4311 + }, + { + "epoch": 1.379692832764505, + "grad_norm": 0.5392977595329285, + "learning_rate": 6.546501441010826e-06, + "loss": 0.4944, + "step": 4312 + }, + { + "epoch": 1.3800127986348123, + "grad_norm": 0.6074196696281433, + "learning_rate": 6.544730833163016e-06, + "loss": 0.4915, + "step": 4313 + }, + { + "epoch": 1.3803327645051193, + "grad_norm": 0.5954985618591309, + "learning_rate": 6.542960011136418e-06, + "loss": 0.4676, + "step": 4314 + }, + { + "epoch": 1.3806527303754266, + "grad_norm": 0.5220882296562195, + "learning_rate": 6.541188975176557e-06, + "loss": 0.488, + "step": 4315 + }, + { + "epoch": 1.3809726962457338, + "grad_norm": 0.49383068084716797, + "learning_rate": 6.539417725528992e-06, + "loss": 0.447, + "step": 4316 + }, + { + "epoch": 1.3812926621160408, + "grad_norm": 0.5188145041465759, + "learning_rate": 6.5376462624393045e-06, + "loss": 0.4711, + "step": 4317 + }, + { + "epoch": 1.381612627986348, + "grad_norm": 0.5797480344772339, + "learning_rate": 6.535874586153115e-06, + "loss": 0.5739, + "step": 4318 + }, + { + "epoch": 1.3819325938566553, + "grad_norm": 0.5276670455932617, + "learning_rate": 6.534102696916066e-06, + "loss": 0.4939, + "step": 4319 + }, + { + "epoch": 1.3822525597269624, + "grad_norm": 0.5280518531799316, + "learning_rate": 6.532330594973833e-06, + "loss": 0.5221, + "step": 4320 + }, + { + "epoch": 1.3825725255972696, + "grad_norm": 0.4720649719238281, + "learning_rate": 6.530558280572119e-06, + "loss": 0.4313, + "step": 4321 + }, + { + "epoch": 1.3828924914675769, + "grad_norm": 0.524074375629425, + "learning_rate": 6.528785753956657e-06, + "loss": 0.5351, + "step": 4322 + }, + { + "epoch": 1.383212457337884, + "grad_norm": 0.543278157711029, + "learning_rate": 6.5270130153732135e-06, + "loss": 0.486, + "step": 4323 + }, + { + "epoch": 1.3835324232081911, + "grad_norm": 0.5051714181900024, + "learning_rate": 6.525240065067576e-06, + "loss": 0.4621, + "step": 4324 + }, + { + "epoch": 1.3838523890784984, + "grad_norm": 0.5485959649085999, + "learning_rate": 6.523466903285571e-06, + "loss": 0.5552, + "step": 4325 + }, + { + "epoch": 1.3841723549488054, + "grad_norm": 0.47958627343177795, + "learning_rate": 6.521693530273046e-06, + "loss": 0.4453, + "step": 4326 + }, + { + "epoch": 1.3844923208191127, + "grad_norm": 0.5448989272117615, + "learning_rate": 6.519919946275881e-06, + "loss": 0.504, + "step": 4327 + }, + { + "epoch": 1.38481228668942, + "grad_norm": 0.48901090025901794, + "learning_rate": 6.518146151539988e-06, + "loss": 0.474, + "step": 4328 + }, + { + "epoch": 1.385132252559727, + "grad_norm": 0.48510274291038513, + "learning_rate": 6.516372146311305e-06, + "loss": 0.4333, + "step": 4329 + }, + { + "epoch": 1.3854522184300342, + "grad_norm": 0.5269479155540466, + "learning_rate": 6.5145979308358e-06, + "loss": 0.5249, + "step": 4330 + }, + { + "epoch": 1.3857721843003412, + "grad_norm": 0.5366206765174866, + "learning_rate": 6.512823505359469e-06, + "loss": 0.4811, + "step": 4331 + }, + { + "epoch": 1.3860921501706485, + "grad_norm": 0.5333788394927979, + "learning_rate": 6.511048870128338e-06, + "loss": 0.527, + "step": 4332 + }, + { + "epoch": 1.3864121160409557, + "grad_norm": 0.5139570832252502, + "learning_rate": 6.509274025388466e-06, + "loss": 0.4685, + "step": 4333 + }, + { + "epoch": 1.3867320819112627, + "grad_norm": 0.5171332955360413, + "learning_rate": 6.5074989713859325e-06, + "loss": 0.4854, + "step": 4334 + }, + { + "epoch": 1.38705204778157, + "grad_norm": 0.5386876463890076, + "learning_rate": 6.505723708366854e-06, + "loss": 0.5118, + "step": 4335 + }, + { + "epoch": 1.387372013651877, + "grad_norm": 0.5003767013549805, + "learning_rate": 6.503948236577371e-06, + "loss": 0.4569, + "step": 4336 + }, + { + "epoch": 1.3876919795221843, + "grad_norm": 0.5198277235031128, + "learning_rate": 6.502172556263656e-06, + "loss": 0.5561, + "step": 4337 + }, + { + "epoch": 1.3880119453924915, + "grad_norm": 0.48306384682655334, + "learning_rate": 6.500396667671911e-06, + "loss": 0.4224, + "step": 4338 + }, + { + "epoch": 1.3883319112627985, + "grad_norm": 0.5466201305389404, + "learning_rate": 6.498620571048361e-06, + "loss": 0.5258, + "step": 4339 + }, + { + "epoch": 1.3886518771331058, + "grad_norm": 0.5272469520568848, + "learning_rate": 6.496844266639266e-06, + "loss": 0.5222, + "step": 4340 + }, + { + "epoch": 1.388971843003413, + "grad_norm": 0.519819974899292, + "learning_rate": 6.4950677546909125e-06, + "loss": 0.5233, + "step": 4341 + }, + { + "epoch": 1.38929180887372, + "grad_norm": 0.5432008504867554, + "learning_rate": 6.493291035449616e-06, + "loss": 0.5145, + "step": 4342 + }, + { + "epoch": 1.3896117747440273, + "grad_norm": 0.44557735323905945, + "learning_rate": 6.491514109161723e-06, + "loss": 0.4508, + "step": 4343 + }, + { + "epoch": 1.3899317406143346, + "grad_norm": 0.5467312335968018, + "learning_rate": 6.489736976073603e-06, + "loss": 0.5767, + "step": 4344 + }, + { + "epoch": 1.3902517064846416, + "grad_norm": 0.5047017335891724, + "learning_rate": 6.487959636431659e-06, + "loss": 0.4873, + "step": 4345 + }, + { + "epoch": 1.3905716723549488, + "grad_norm": 0.46947747468948364, + "learning_rate": 6.486182090482322e-06, + "loss": 0.4477, + "step": 4346 + }, + { + "epoch": 1.390891638225256, + "grad_norm": 0.5482416152954102, + "learning_rate": 6.484404338472049e-06, + "loss": 0.5166, + "step": 4347 + }, + { + "epoch": 1.391211604095563, + "grad_norm": 0.4624686539173126, + "learning_rate": 6.48262638064733e-06, + "loss": 0.4461, + "step": 4348 + }, + { + "epoch": 1.3915315699658704, + "grad_norm": 0.5751667618751526, + "learning_rate": 6.480848217254678e-06, + "loss": 0.5293, + "step": 4349 + }, + { + "epoch": 1.3918515358361776, + "grad_norm": 0.5793061852455139, + "learning_rate": 6.479069848540641e-06, + "loss": 0.5203, + "step": 4350 + }, + { + "epoch": 1.3921715017064846, + "grad_norm": 0.5687735676765442, + "learning_rate": 6.477291274751789e-06, + "loss": 0.5057, + "step": 4351 + }, + { + "epoch": 1.3924914675767919, + "grad_norm": 0.5311597585678101, + "learning_rate": 6.4755124961347245e-06, + "loss": 0.5907, + "step": 4352 + }, + { + "epoch": 1.392811433447099, + "grad_norm": 0.5056808590888977, + "learning_rate": 6.473733512936078e-06, + "loss": 0.4645, + "step": 4353 + }, + { + "epoch": 1.3931313993174061, + "grad_norm": 0.5219675302505493, + "learning_rate": 6.471954325402508e-06, + "loss": 0.5018, + "step": 4354 + }, + { + "epoch": 1.3934513651877132, + "grad_norm": 0.5257090330123901, + "learning_rate": 6.4701749337806975e-06, + "loss": 0.4795, + "step": 4355 + }, + { + "epoch": 1.3937713310580204, + "grad_norm": 0.52811199426651, + "learning_rate": 6.468395338317365e-06, + "loss": 0.5357, + "step": 4356 + }, + { + "epoch": 1.3940912969283277, + "grad_norm": 0.5537838339805603, + "learning_rate": 6.466615539259252e-06, + "loss": 0.4896, + "step": 4357 + }, + { + "epoch": 1.3944112627986347, + "grad_norm": 0.5262822508811951, + "learning_rate": 6.46483553685313e-06, + "loss": 0.526, + "step": 4358 + }, + { + "epoch": 1.394731228668942, + "grad_norm": 0.5014103055000305, + "learning_rate": 6.4630553313457985e-06, + "loss": 0.4792, + "step": 4359 + }, + { + "epoch": 1.3950511945392492, + "grad_norm": 0.4720224142074585, + "learning_rate": 6.461274922984087e-06, + "loss": 0.4431, + "step": 4360 + }, + { + "epoch": 1.3953711604095562, + "grad_norm": 0.536048173904419, + "learning_rate": 6.459494312014848e-06, + "loss": 0.535, + "step": 4361 + }, + { + "epoch": 1.3956911262798635, + "grad_norm": 0.5579946637153625, + "learning_rate": 6.457713498684966e-06, + "loss": 0.5141, + "step": 4362 + }, + { + "epoch": 1.3960110921501707, + "grad_norm": 0.5562567710876465, + "learning_rate": 6.455932483241355e-06, + "loss": 0.4696, + "step": 4363 + }, + { + "epoch": 1.3963310580204777, + "grad_norm": 0.5763883590698242, + "learning_rate": 6.454151265930953e-06, + "loss": 0.5313, + "step": 4364 + }, + { + "epoch": 1.396651023890785, + "grad_norm": 0.5138348937034607, + "learning_rate": 6.45236984700073e-06, + "loss": 0.4556, + "step": 4365 + }, + { + "epoch": 1.3969709897610922, + "grad_norm": 0.5289807319641113, + "learning_rate": 6.450588226697679e-06, + "loss": 0.4584, + "step": 4366 + }, + { + "epoch": 1.3972909556313993, + "grad_norm": 0.6209538578987122, + "learning_rate": 6.448806405268826e-06, + "loss": 0.6022, + "step": 4367 + }, + { + "epoch": 1.3976109215017065, + "grad_norm": 0.5078195333480835, + "learning_rate": 6.4470243829612215e-06, + "loss": 0.4854, + "step": 4368 + }, + { + "epoch": 1.3979308873720138, + "grad_norm": 0.504351019859314, + "learning_rate": 6.445242160021945e-06, + "loss": 0.512, + "step": 4369 + }, + { + "epoch": 1.3982508532423208, + "grad_norm": 0.5405347943305969, + "learning_rate": 6.443459736698106e-06, + "loss": 0.4986, + "step": 4370 + }, + { + "epoch": 1.398570819112628, + "grad_norm": 0.5360117554664612, + "learning_rate": 6.441677113236837e-06, + "loss": 0.4989, + "step": 4371 + }, + { + "epoch": 1.3988907849829353, + "grad_norm": 0.4768439829349518, + "learning_rate": 6.439894289885303e-06, + "loss": 0.5027, + "step": 4372 + }, + { + "epoch": 1.3992107508532423, + "grad_norm": 0.533896267414093, + "learning_rate": 6.438111266890691e-06, + "loss": 0.472, + "step": 4373 + }, + { + "epoch": 1.3995307167235496, + "grad_norm": 0.6232820749282837, + "learning_rate": 6.436328044500223e-06, + "loss": 0.5015, + "step": 4374 + }, + { + "epoch": 1.3998506825938566, + "grad_norm": 0.6136919856071472, + "learning_rate": 6.434544622961143e-06, + "loss": 0.4755, + "step": 4375 + }, + { + "epoch": 1.4001706484641638, + "grad_norm": 0.5707723498344421, + "learning_rate": 6.4327610025207265e-06, + "loss": 0.4811, + "step": 4376 + }, + { + "epoch": 1.4004906143344709, + "grad_norm": 0.5362030267715454, + "learning_rate": 6.430977183426273e-06, + "loss": 0.5322, + "step": 4377 + }, + { + "epoch": 1.400810580204778, + "grad_norm": 0.5937615036964417, + "learning_rate": 6.4291931659251115e-06, + "loss": 0.4952, + "step": 4378 + }, + { + "epoch": 1.4011305460750854, + "grad_norm": 0.5517301559448242, + "learning_rate": 6.427408950264597e-06, + "loss": 0.5075, + "step": 4379 + }, + { + "epoch": 1.4014505119453924, + "grad_norm": 0.5565688610076904, + "learning_rate": 6.425624536692116e-06, + "loss": 0.4979, + "step": 4380 + }, + { + "epoch": 1.4017704778156996, + "grad_norm": 0.5186704993247986, + "learning_rate": 6.423839925455078e-06, + "loss": 0.4758, + "step": 4381 + }, + { + "epoch": 1.4020904436860069, + "grad_norm": 0.4568119943141937, + "learning_rate": 6.422055116800921e-06, + "loss": 0.4475, + "step": 4382 + }, + { + "epoch": 1.402410409556314, + "grad_norm": 0.5236671566963196, + "learning_rate": 6.4202701109771105e-06, + "loss": 0.4868, + "step": 4383 + }, + { + "epoch": 1.4027303754266212, + "grad_norm": 0.4925619661808014, + "learning_rate": 6.41848490823114e-06, + "loss": 0.4861, + "step": 4384 + }, + { + "epoch": 1.4030503412969284, + "grad_norm": 0.5599465370178223, + "learning_rate": 6.416699508810531e-06, + "loss": 0.4979, + "step": 4385 + }, + { + "epoch": 1.4033703071672354, + "grad_norm": 0.5127217173576355, + "learning_rate": 6.414913912962832e-06, + "loss": 0.4933, + "step": 4386 + }, + { + "epoch": 1.4036902730375427, + "grad_norm": 0.49806344509124756, + "learning_rate": 6.413128120935616e-06, + "loss": 0.4744, + "step": 4387 + }, + { + "epoch": 1.40401023890785, + "grad_norm": 0.5179780125617981, + "learning_rate": 6.411342132976486e-06, + "loss": 0.5146, + "step": 4388 + }, + { + "epoch": 1.404330204778157, + "grad_norm": 0.49689823389053345, + "learning_rate": 6.409555949333072e-06, + "loss": 0.517, + "step": 4389 + }, + { + "epoch": 1.4046501706484642, + "grad_norm": 0.52426677942276, + "learning_rate": 6.4077695702530285e-06, + "loss": 0.4991, + "step": 4390 + }, + { + "epoch": 1.4049701365187715, + "grad_norm": 0.5141168236732483, + "learning_rate": 6.405982995984043e-06, + "loss": 0.4354, + "step": 4391 + }, + { + "epoch": 1.4052901023890785, + "grad_norm": 0.565462052822113, + "learning_rate": 6.40419622677382e-06, + "loss": 0.5294, + "step": 4392 + }, + { + "epoch": 1.4056100682593857, + "grad_norm": 0.5194235444068909, + "learning_rate": 6.402409262870104e-06, + "loss": 0.5367, + "step": 4393 + }, + { + "epoch": 1.4059300341296928, + "grad_norm": 0.5440438389778137, + "learning_rate": 6.400622104520655e-06, + "loss": 0.5077, + "step": 4394 + }, + { + "epoch": 1.40625, + "grad_norm": 0.44700324535369873, + "learning_rate": 6.398834751973266e-06, + "loss": 0.4344, + "step": 4395 + }, + { + "epoch": 1.4065699658703072, + "grad_norm": 0.4888385236263275, + "learning_rate": 6.397047205475757e-06, + "loss": 0.4749, + "step": 4396 + }, + { + "epoch": 1.4068899317406143, + "grad_norm": 0.5607296824455261, + "learning_rate": 6.395259465275971e-06, + "loss": 0.5707, + "step": 4397 + }, + { + "epoch": 1.4072098976109215, + "grad_norm": 0.5562505722045898, + "learning_rate": 6.393471531621783e-06, + "loss": 0.5076, + "step": 4398 + }, + { + "epoch": 1.4075298634812285, + "grad_norm": 0.5209715366363525, + "learning_rate": 6.39168340476109e-06, + "loss": 0.4721, + "step": 4399 + }, + { + "epoch": 1.4078498293515358, + "grad_norm": 0.5225011706352234, + "learning_rate": 6.3898950849418184e-06, + "loss": 0.4992, + "step": 4400 + }, + { + "epoch": 1.408169795221843, + "grad_norm": 0.4573118984699249, + "learning_rate": 6.388106572411922e-06, + "loss": 0.4788, + "step": 4401 + }, + { + "epoch": 1.40848976109215, + "grad_norm": 0.5134564638137817, + "learning_rate": 6.386317867419379e-06, + "loss": 0.479, + "step": 4402 + }, + { + "epoch": 1.4088097269624573, + "grad_norm": 0.5742200016975403, + "learning_rate": 6.384528970212196e-06, + "loss": 0.5452, + "step": 4403 + }, + { + "epoch": 1.4091296928327646, + "grad_norm": 0.4976794123649597, + "learning_rate": 6.3827398810384045e-06, + "loss": 0.4799, + "step": 4404 + }, + { + "epoch": 1.4094496587030716, + "grad_norm": 0.5459603667259216, + "learning_rate": 6.380950600146066e-06, + "loss": 0.4634, + "step": 4405 + }, + { + "epoch": 1.4097696245733788, + "grad_norm": 0.5783873200416565, + "learning_rate": 6.379161127783267e-06, + "loss": 0.536, + "step": 4406 + }, + { + "epoch": 1.410089590443686, + "grad_norm": 0.5288960337638855, + "learning_rate": 6.377371464198114e-06, + "loss": 0.4809, + "step": 4407 + }, + { + "epoch": 1.4104095563139931, + "grad_norm": 0.5651437044143677, + "learning_rate": 6.375581609638754e-06, + "loss": 0.5454, + "step": 4408 + }, + { + "epoch": 1.4107295221843004, + "grad_norm": 0.4710017144680023, + "learning_rate": 6.3737915643533484e-06, + "loss": 0.45, + "step": 4409 + }, + { + "epoch": 1.4110494880546076, + "grad_norm": 0.4859638512134552, + "learning_rate": 6.3720013285900905e-06, + "loss": 0.4643, + "step": 4410 + }, + { + "epoch": 1.4113694539249146, + "grad_norm": 0.5344699621200562, + "learning_rate": 6.370210902597198e-06, + "loss": 0.5429, + "step": 4411 + }, + { + "epoch": 1.411689419795222, + "grad_norm": 0.5420651435852051, + "learning_rate": 6.368420286622913e-06, + "loss": 0.4771, + "step": 4412 + }, + { + "epoch": 1.4120093856655291, + "grad_norm": 0.5548089146614075, + "learning_rate": 6.366629480915512e-06, + "loss": 0.527, + "step": 4413 + }, + { + "epoch": 1.4123293515358362, + "grad_norm": 0.48039743304252625, + "learning_rate": 6.364838485723286e-06, + "loss": 0.492, + "step": 4414 + }, + { + "epoch": 1.4126493174061434, + "grad_norm": 0.513484537601471, + "learning_rate": 6.363047301294564e-06, + "loss": 0.5029, + "step": 4415 + }, + { + "epoch": 1.4129692832764504, + "grad_norm": 0.5703129768371582, + "learning_rate": 6.361255927877694e-06, + "loss": 0.5472, + "step": 4416 + }, + { + "epoch": 1.4132892491467577, + "grad_norm": 0.5068913102149963, + "learning_rate": 6.359464365721051e-06, + "loss": 0.4668, + "step": 4417 + }, + { + "epoch": 1.4136092150170647, + "grad_norm": 0.517008364200592, + "learning_rate": 6.357672615073039e-06, + "loss": 0.5086, + "step": 4418 + }, + { + "epoch": 1.413929180887372, + "grad_norm": 0.49030840396881104, + "learning_rate": 6.355880676182086e-06, + "loss": 0.4621, + "step": 4419 + }, + { + "epoch": 1.4142491467576792, + "grad_norm": 0.5618411302566528, + "learning_rate": 6.354088549296645e-06, + "loss": 0.5257, + "step": 4420 + }, + { + "epoch": 1.4145691126279862, + "grad_norm": 0.5886921286582947, + "learning_rate": 6.3522962346651985e-06, + "loss": 0.5134, + "step": 4421 + }, + { + "epoch": 1.4148890784982935, + "grad_norm": 0.5557438731193542, + "learning_rate": 6.3505037325362515e-06, + "loss": 0.4778, + "step": 4422 + }, + { + "epoch": 1.4152090443686007, + "grad_norm": 0.5813466310501099, + "learning_rate": 6.348711043158339e-06, + "loss": 0.558, + "step": 4423 + }, + { + "epoch": 1.4155290102389078, + "grad_norm": 0.5316801071166992, + "learning_rate": 6.346918166780016e-06, + "loss": 0.5242, + "step": 4424 + }, + { + "epoch": 1.415848976109215, + "grad_norm": 0.4986865520477295, + "learning_rate": 6.34512510364987e-06, + "loss": 0.4553, + "step": 4425 + }, + { + "epoch": 1.4161689419795223, + "grad_norm": 0.6381187438964844, + "learning_rate": 6.3433318540165074e-06, + "loss": 0.5365, + "step": 4426 + }, + { + "epoch": 1.4164889078498293, + "grad_norm": 0.5615751147270203, + "learning_rate": 6.34153841812857e-06, + "loss": 0.5152, + "step": 4427 + }, + { + "epoch": 1.4168088737201365, + "grad_norm": 0.5620642900466919, + "learning_rate": 6.339744796234717e-06, + "loss": 0.5363, + "step": 4428 + }, + { + "epoch": 1.4171288395904438, + "grad_norm": 0.5495388507843018, + "learning_rate": 6.337950988583634e-06, + "loss": 0.4252, + "step": 4429 + }, + { + "epoch": 1.4174488054607508, + "grad_norm": 0.5447618961334229, + "learning_rate": 6.336156995424039e-06, + "loss": 0.4448, + "step": 4430 + }, + { + "epoch": 1.417768771331058, + "grad_norm": 0.5106366872787476, + "learning_rate": 6.334362817004669e-06, + "loss": 0.4755, + "step": 4431 + }, + { + "epoch": 1.4180887372013653, + "grad_norm": 0.553810179233551, + "learning_rate": 6.332568453574288e-06, + "loss": 0.5131, + "step": 4432 + }, + { + "epoch": 1.4184087030716723, + "grad_norm": 0.48646387457847595, + "learning_rate": 6.330773905381689e-06, + "loss": 0.5283, + "step": 4433 + }, + { + "epoch": 1.4187286689419796, + "grad_norm": 0.529070258140564, + "learning_rate": 6.328979172675685e-06, + "loss": 0.4845, + "step": 4434 + }, + { + "epoch": 1.4190486348122868, + "grad_norm": 0.5212682485580444, + "learning_rate": 6.327184255705123e-06, + "loss": 0.4631, + "step": 4435 + }, + { + "epoch": 1.4193686006825939, + "grad_norm": 0.5363638401031494, + "learning_rate": 6.325389154718865e-06, + "loss": 0.5055, + "step": 4436 + }, + { + "epoch": 1.419688566552901, + "grad_norm": 0.5298949480056763, + "learning_rate": 6.323593869965807e-06, + "loss": 0.4998, + "step": 4437 + }, + { + "epoch": 1.4200085324232081, + "grad_norm": 0.5294914245605469, + "learning_rate": 6.3217984016948675e-06, + "loss": 0.4782, + "step": 4438 + }, + { + "epoch": 1.4203284982935154, + "grad_norm": 0.5212501287460327, + "learning_rate": 6.320002750154988e-06, + "loss": 0.4807, + "step": 4439 + }, + { + "epoch": 1.4206484641638224, + "grad_norm": 0.5917192697525024, + "learning_rate": 6.3182069155951394e-06, + "loss": 0.5038, + "step": 4440 + }, + { + "epoch": 1.4209684300341296, + "grad_norm": 0.5383630394935608, + "learning_rate": 6.3164108982643155e-06, + "loss": 0.5086, + "step": 4441 + }, + { + "epoch": 1.421288395904437, + "grad_norm": 0.5691002011299133, + "learning_rate": 6.314614698411539e-06, + "loss": 0.4683, + "step": 4442 + }, + { + "epoch": 1.421608361774744, + "grad_norm": 0.5580256581306458, + "learning_rate": 6.3128183162858515e-06, + "loss": 0.5137, + "step": 4443 + }, + { + "epoch": 1.4219283276450512, + "grad_norm": 0.5336185693740845, + "learning_rate": 6.311021752136325e-06, + "loss": 0.5186, + "step": 4444 + }, + { + "epoch": 1.4222482935153584, + "grad_norm": 0.567459225654602, + "learning_rate": 6.309225006212056e-06, + "loss": 0.535, + "step": 4445 + }, + { + "epoch": 1.4225682593856654, + "grad_norm": 0.5339890122413635, + "learning_rate": 6.307428078762165e-06, + "loss": 0.4494, + "step": 4446 + }, + { + "epoch": 1.4228882252559727, + "grad_norm": 0.49709126353263855, + "learning_rate": 6.305630970035796e-06, + "loss": 0.4704, + "step": 4447 + }, + { + "epoch": 1.42320819112628, + "grad_norm": 0.6274470090866089, + "learning_rate": 6.303833680282125e-06, + "loss": 0.5234, + "step": 4448 + }, + { + "epoch": 1.423528156996587, + "grad_norm": 0.5845829844474792, + "learning_rate": 6.302036209750344e-06, + "loss": 0.4932, + "step": 4449 + }, + { + "epoch": 1.4238481228668942, + "grad_norm": 0.5564723014831543, + "learning_rate": 6.300238558689678e-06, + "loss": 0.4898, + "step": 4450 + }, + { + "epoch": 1.4241680887372015, + "grad_norm": 0.5889071226119995, + "learning_rate": 6.2984407273493695e-06, + "loss": 0.4758, + "step": 4451 + }, + { + "epoch": 1.4244880546075085, + "grad_norm": 0.6283670663833618, + "learning_rate": 6.296642715978693e-06, + "loss": 0.5183, + "step": 4452 + }, + { + "epoch": 1.4248080204778157, + "grad_norm": 0.5578067302703857, + "learning_rate": 6.294844524826944e-06, + "loss": 0.476, + "step": 4453 + }, + { + "epoch": 1.425127986348123, + "grad_norm": 0.54828280210495, + "learning_rate": 6.2930461541434426e-06, + "loss": 0.526, + "step": 4454 + }, + { + "epoch": 1.42544795221843, + "grad_norm": 0.5079402327537537, + "learning_rate": 6.291247604177539e-06, + "loss": 0.4554, + "step": 4455 + }, + { + "epoch": 1.4257679180887373, + "grad_norm": 0.5570620894432068, + "learning_rate": 6.2894488751786e-06, + "loss": 0.515, + "step": 4456 + }, + { + "epoch": 1.4260878839590443, + "grad_norm": 0.5727689266204834, + "learning_rate": 6.287649967396024e-06, + "loss": 0.4815, + "step": 4457 + }, + { + "epoch": 1.4264078498293515, + "grad_norm": 0.5417753458023071, + "learning_rate": 6.285850881079229e-06, + "loss": 0.4616, + "step": 4458 + }, + { + "epoch": 1.4267278156996588, + "grad_norm": 0.5604990720748901, + "learning_rate": 6.284051616477664e-06, + "loss": 0.5118, + "step": 4459 + }, + { + "epoch": 1.4270477815699658, + "grad_norm": 0.48993706703186035, + "learning_rate": 6.282252173840797e-06, + "loss": 0.5136, + "step": 4460 + }, + { + "epoch": 1.427367747440273, + "grad_norm": 0.5203538537025452, + "learning_rate": 6.280452553418126e-06, + "loss": 0.5236, + "step": 4461 + }, + { + "epoch": 1.42768771331058, + "grad_norm": 0.5383204817771912, + "learning_rate": 6.278652755459165e-06, + "loss": 0.4484, + "step": 4462 + }, + { + "epoch": 1.4280076791808873, + "grad_norm": 0.541811466217041, + "learning_rate": 6.276852780213463e-06, + "loss": 0.5496, + "step": 4463 + }, + { + "epoch": 1.4283276450511946, + "grad_norm": 0.5643213987350464, + "learning_rate": 6.275052627930585e-06, + "loss": 0.4532, + "step": 4464 + }, + { + "epoch": 1.4286476109215016, + "grad_norm": 0.5392845273017883, + "learning_rate": 6.273252298860127e-06, + "loss": 0.4213, + "step": 4465 + }, + { + "epoch": 1.4289675767918089, + "grad_norm": 0.5217435956001282, + "learning_rate": 6.271451793251707e-06, + "loss": 0.5221, + "step": 4466 + }, + { + "epoch": 1.429287542662116, + "grad_norm": 0.5222741365432739, + "learning_rate": 6.269651111354967e-06, + "loss": 0.4767, + "step": 4467 + }, + { + "epoch": 1.4296075085324231, + "grad_norm": 0.5568684339523315, + "learning_rate": 6.267850253419571e-06, + "loss": 0.4401, + "step": 4468 + }, + { + "epoch": 1.4299274744027304, + "grad_norm": 0.6233351230621338, + "learning_rate": 6.266049219695211e-06, + "loss": 0.5179, + "step": 4469 + }, + { + "epoch": 1.4302474402730376, + "grad_norm": 0.5469030141830444, + "learning_rate": 6.264248010431606e-06, + "loss": 0.5201, + "step": 4470 + }, + { + "epoch": 1.4305674061433447, + "grad_norm": 0.5641871094703674, + "learning_rate": 6.262446625878493e-06, + "loss": 0.4582, + "step": 4471 + }, + { + "epoch": 1.430887372013652, + "grad_norm": 0.5134510397911072, + "learning_rate": 6.260645066285637e-06, + "loss": 0.459, + "step": 4472 + }, + { + "epoch": 1.4312073378839592, + "grad_norm": 0.5164428353309631, + "learning_rate": 6.258843331902824e-06, + "loss": 0.4997, + "step": 4473 + }, + { + "epoch": 1.4315273037542662, + "grad_norm": 0.5034913420677185, + "learning_rate": 6.257041422979871e-06, + "loss": 0.445, + "step": 4474 + }, + { + "epoch": 1.4318472696245734, + "grad_norm": 0.549512505531311, + "learning_rate": 6.25523933976661e-06, + "loss": 0.5495, + "step": 4475 + }, + { + "epoch": 1.4321672354948807, + "grad_norm": 0.5255251526832581, + "learning_rate": 6.253437082512908e-06, + "loss": 0.4362, + "step": 4476 + }, + { + "epoch": 1.4324872013651877, + "grad_norm": 0.5761882066726685, + "learning_rate": 6.251634651468642e-06, + "loss": 0.5355, + "step": 4477 + }, + { + "epoch": 1.432807167235495, + "grad_norm": 0.4902086555957794, + "learning_rate": 6.249832046883729e-06, + "loss": 0.4468, + "step": 4478 + }, + { + "epoch": 1.433127133105802, + "grad_norm": 0.5528424382209778, + "learning_rate": 6.248029269008098e-06, + "loss": 0.523, + "step": 4479 + }, + { + "epoch": 1.4334470989761092, + "grad_norm": 0.563441276550293, + "learning_rate": 6.246226318091708e-06, + "loss": 0.504, + "step": 4480 + }, + { + "epoch": 1.4337670648464163, + "grad_norm": 0.5314658880233765, + "learning_rate": 6.24442319438454e-06, + "loss": 0.5354, + "step": 4481 + }, + { + "epoch": 1.4340870307167235, + "grad_norm": 0.5264476537704468, + "learning_rate": 6.242619898136597e-06, + "loss": 0.4423, + "step": 4482 + }, + { + "epoch": 1.4344069965870307, + "grad_norm": 0.5332199335098267, + "learning_rate": 6.240816429597913e-06, + "loss": 0.5381, + "step": 4483 + }, + { + "epoch": 1.4347269624573378, + "grad_norm": 0.563339114189148, + "learning_rate": 6.239012789018537e-06, + "loss": 0.5371, + "step": 4484 + }, + { + "epoch": 1.435046928327645, + "grad_norm": 0.5181301832199097, + "learning_rate": 6.237208976648547e-06, + "loss": 0.4461, + "step": 4485 + }, + { + "epoch": 1.4353668941979523, + "grad_norm": 0.5529221892356873, + "learning_rate": 6.235404992738044e-06, + "loss": 0.5006, + "step": 4486 + }, + { + "epoch": 1.4356868600682593, + "grad_norm": 0.5713469386100769, + "learning_rate": 6.233600837537153e-06, + "loss": 0.547, + "step": 4487 + }, + { + "epoch": 1.4360068259385665, + "grad_norm": 0.46673092246055603, + "learning_rate": 6.231796511296022e-06, + "loss": 0.4204, + "step": 4488 + }, + { + "epoch": 1.4363267918088738, + "grad_norm": 0.5404127240180969, + "learning_rate": 6.229992014264822e-06, + "loss": 0.4984, + "step": 4489 + }, + { + "epoch": 1.4366467576791808, + "grad_norm": 0.5437692403793335, + "learning_rate": 6.228187346693749e-06, + "loss": 0.5257, + "step": 4490 + }, + { + "epoch": 1.436966723549488, + "grad_norm": 0.57470703125, + "learning_rate": 6.2263825088330254e-06, + "loss": 0.5077, + "step": 4491 + }, + { + "epoch": 1.4372866894197953, + "grad_norm": 0.528522789478302, + "learning_rate": 6.224577500932888e-06, + "loss": 0.4906, + "step": 4492 + }, + { + "epoch": 1.4376066552901023, + "grad_norm": 0.4943845272064209, + "learning_rate": 6.22277232324361e-06, + "loss": 0.4605, + "step": 4493 + }, + { + "epoch": 1.4379266211604096, + "grad_norm": 0.5202118754386902, + "learning_rate": 6.220966976015477e-06, + "loss": 0.572, + "step": 4494 + }, + { + "epoch": 1.4382465870307168, + "grad_norm": 0.4969080090522766, + "learning_rate": 6.219161459498804e-06, + "loss": 0.4913, + "step": 4495 + }, + { + "epoch": 1.4385665529010239, + "grad_norm": 0.5071524381637573, + "learning_rate": 6.217355773943928e-06, + "loss": 0.4422, + "step": 4496 + }, + { + "epoch": 1.4388865187713311, + "grad_norm": 0.47950607538223267, + "learning_rate": 6.215549919601209e-06, + "loss": 0.4634, + "step": 4497 + }, + { + "epoch": 1.4392064846416384, + "grad_norm": 0.5288113355636597, + "learning_rate": 6.213743896721033e-06, + "loss": 0.4738, + "step": 4498 + }, + { + "epoch": 1.4395264505119454, + "grad_norm": 0.5222241282463074, + "learning_rate": 6.2119377055538025e-06, + "loss": 0.4789, + "step": 4499 + }, + { + "epoch": 1.4398464163822526, + "grad_norm": 0.5527712106704712, + "learning_rate": 6.210131346349953e-06, + "loss": 0.5543, + "step": 4500 + }, + { + "epoch": 1.4401663822525597, + "grad_norm": 0.48620420694351196, + "learning_rate": 6.208324819359939e-06, + "loss": 0.4742, + "step": 4501 + }, + { + "epoch": 1.440486348122867, + "grad_norm": 0.542361855506897, + "learning_rate": 6.206518124834231e-06, + "loss": 0.5524, + "step": 4502 + }, + { + "epoch": 1.440806313993174, + "grad_norm": 0.5217463374137878, + "learning_rate": 6.204711263023335e-06, + "loss": 0.4876, + "step": 4503 + }, + { + "epoch": 1.4411262798634812, + "grad_norm": 0.5138661861419678, + "learning_rate": 6.202904234177772e-06, + "loss": 0.4902, + "step": 4504 + }, + { + "epoch": 1.4414462457337884, + "grad_norm": 0.48745661973953247, + "learning_rate": 6.201097038548092e-06, + "loss": 0.4861, + "step": 4505 + }, + { + "epoch": 1.4417662116040955, + "grad_norm": 0.5008990168571472, + "learning_rate": 6.199289676384863e-06, + "loss": 0.4723, + "step": 4506 + }, + { + "epoch": 1.4420861774744027, + "grad_norm": 0.565268874168396, + "learning_rate": 6.197482147938675e-06, + "loss": 0.5103, + "step": 4507 + }, + { + "epoch": 1.44240614334471, + "grad_norm": 0.49539247155189514, + "learning_rate": 6.195674453460148e-06, + "loss": 0.4721, + "step": 4508 + }, + { + "epoch": 1.442726109215017, + "grad_norm": 0.5051780343055725, + "learning_rate": 6.193866593199919e-06, + "loss": 0.5116, + "step": 4509 + }, + { + "epoch": 1.4430460750853242, + "grad_norm": 0.4699634909629822, + "learning_rate": 6.192058567408652e-06, + "loss": 0.4729, + "step": 4510 + }, + { + "epoch": 1.4433660409556315, + "grad_norm": 0.5447596311569214, + "learning_rate": 6.190250376337028e-06, + "loss": 0.5723, + "step": 4511 + }, + { + "epoch": 1.4436860068259385, + "grad_norm": 0.5196283459663391, + "learning_rate": 6.18844202023576e-06, + "loss": 0.4707, + "step": 4512 + }, + { + "epoch": 1.4440059726962458, + "grad_norm": 0.5027531981468201, + "learning_rate": 6.186633499355576e-06, + "loss": 0.4685, + "step": 4513 + }, + { + "epoch": 1.444325938566553, + "grad_norm": 0.5171422362327576, + "learning_rate": 6.1848248139472265e-06, + "loss": 0.5018, + "step": 4514 + }, + { + "epoch": 1.44464590443686, + "grad_norm": 0.5533860921859741, + "learning_rate": 6.183015964261495e-06, + "loss": 0.5076, + "step": 4515 + }, + { + "epoch": 1.4449658703071673, + "grad_norm": 0.4939151108264923, + "learning_rate": 6.181206950549175e-06, + "loss": 0.4772, + "step": 4516 + }, + { + "epoch": 1.4452858361774745, + "grad_norm": 0.5183248519897461, + "learning_rate": 6.179397773061091e-06, + "loss": 0.5044, + "step": 4517 + }, + { + "epoch": 1.4456058020477816, + "grad_norm": 0.5419015884399414, + "learning_rate": 6.177588432048087e-06, + "loss": 0.4741, + "step": 4518 + }, + { + "epoch": 1.4459257679180888, + "grad_norm": 0.570713460445404, + "learning_rate": 6.175778927761029e-06, + "loss": 0.4863, + "step": 4519 + }, + { + "epoch": 1.4462457337883958, + "grad_norm": 0.49967116117477417, + "learning_rate": 6.173969260450809e-06, + "loss": 0.4739, + "step": 4520 + }, + { + "epoch": 1.446565699658703, + "grad_norm": 0.5816405415534973, + "learning_rate": 6.172159430368339e-06, + "loss": 0.5197, + "step": 4521 + }, + { + "epoch": 1.44688566552901, + "grad_norm": 0.5697169899940491, + "learning_rate": 6.170349437764554e-06, + "loss": 0.5279, + "step": 4522 + }, + { + "epoch": 1.4472056313993173, + "grad_norm": 0.5516278743743896, + "learning_rate": 6.1685392828904115e-06, + "loss": 0.5231, + "step": 4523 + }, + { + "epoch": 1.4475255972696246, + "grad_norm": 0.4786688983440399, + "learning_rate": 6.16672896599689e-06, + "loss": 0.43, + "step": 4524 + }, + { + "epoch": 1.4478455631399316, + "grad_norm": 0.5326284170150757, + "learning_rate": 6.1649184873349955e-06, + "loss": 0.5344, + "step": 4525 + }, + { + "epoch": 1.4481655290102389, + "grad_norm": 0.6142365336418152, + "learning_rate": 6.16310784715575e-06, + "loss": 0.5301, + "step": 4526 + }, + { + "epoch": 1.4484854948805461, + "grad_norm": 0.47505971789360046, + "learning_rate": 6.161297045710203e-06, + "loss": 0.4368, + "step": 4527 + }, + { + "epoch": 1.4488054607508531, + "grad_norm": 0.5444201827049255, + "learning_rate": 6.159486083249424e-06, + "loss": 0.5113, + "step": 4528 + }, + { + "epoch": 1.4491254266211604, + "grad_norm": 0.5231147408485413, + "learning_rate": 6.157674960024503e-06, + "loss": 0.5127, + "step": 4529 + }, + { + "epoch": 1.4494453924914676, + "grad_norm": 0.5051755309104919, + "learning_rate": 6.155863676286559e-06, + "loss": 0.4394, + "step": 4530 + }, + { + "epoch": 1.4497653583617747, + "grad_norm": 0.5654588937759399, + "learning_rate": 6.154052232286724e-06, + "loss": 0.6037, + "step": 4531 + }, + { + "epoch": 1.450085324232082, + "grad_norm": 0.44511279463768005, + "learning_rate": 6.152240628276158e-06, + "loss": 0.4571, + "step": 4532 + }, + { + "epoch": 1.4504052901023892, + "grad_norm": 0.49236172437667847, + "learning_rate": 6.150428864506043e-06, + "loss": 0.4829, + "step": 4533 + }, + { + "epoch": 1.4507252559726962, + "grad_norm": 0.5119214057922363, + "learning_rate": 6.148616941227582e-06, + "loss": 0.5184, + "step": 4534 + }, + { + "epoch": 1.4510452218430034, + "grad_norm": 0.5311828851699829, + "learning_rate": 6.146804858692001e-06, + "loss": 0.4858, + "step": 4535 + }, + { + "epoch": 1.4513651877133107, + "grad_norm": 0.5040050745010376, + "learning_rate": 6.144992617150547e-06, + "loss": 0.5039, + "step": 4536 + }, + { + "epoch": 1.4516851535836177, + "grad_norm": 0.4733341932296753, + "learning_rate": 6.143180216854488e-06, + "loss": 0.4451, + "step": 4537 + }, + { + "epoch": 1.452005119453925, + "grad_norm": 0.5829097628593445, + "learning_rate": 6.141367658055116e-06, + "loss": 0.5119, + "step": 4538 + }, + { + "epoch": 1.4523250853242322, + "grad_norm": 0.533114492893219, + "learning_rate": 6.139554941003747e-06, + "loss": 0.498, + "step": 4539 + }, + { + "epoch": 1.4526450511945392, + "grad_norm": 0.5461969971656799, + "learning_rate": 6.137742065951714e-06, + "loss": 0.5466, + "step": 4540 + }, + { + "epoch": 1.4529650170648465, + "grad_norm": 0.4805932641029358, + "learning_rate": 6.135929033150374e-06, + "loss": 0.4623, + "step": 4541 + }, + { + "epoch": 1.4532849829351535, + "grad_norm": 0.5171344876289368, + "learning_rate": 6.134115842851107e-06, + "loss": 0.4539, + "step": 4542 + }, + { + "epoch": 1.4536049488054608, + "grad_norm": 0.5133865475654602, + "learning_rate": 6.1323024953053155e-06, + "loss": 0.4817, + "step": 4543 + }, + { + "epoch": 1.4539249146757678, + "grad_norm": 0.5248680710792542, + "learning_rate": 6.130488990764419e-06, + "loss": 0.492, + "step": 4544 + }, + { + "epoch": 1.454244880546075, + "grad_norm": 0.5106906294822693, + "learning_rate": 6.128675329479865e-06, + "loss": 0.4955, + "step": 4545 + }, + { + "epoch": 1.4545648464163823, + "grad_norm": 0.5252261757850647, + "learning_rate": 6.126861511703119e-06, + "loss": 0.4828, + "step": 4546 + }, + { + "epoch": 1.4548848122866893, + "grad_norm": 0.5696649551391602, + "learning_rate": 6.125047537685668e-06, + "loss": 0.4814, + "step": 4547 + }, + { + "epoch": 1.4552047781569966, + "grad_norm": 0.6102914214134216, + "learning_rate": 6.123233407679022e-06, + "loss": 0.5212, + "step": 4548 + }, + { + "epoch": 1.4555247440273038, + "grad_norm": 0.5728787183761597, + "learning_rate": 6.121419121934714e-06, + "loss": 0.485, + "step": 4549 + }, + { + "epoch": 1.4558447098976108, + "grad_norm": 0.5707157850265503, + "learning_rate": 6.119604680704295e-06, + "loss": 0.4785, + "step": 4550 + }, + { + "epoch": 1.456164675767918, + "grad_norm": 0.5385659337043762, + "learning_rate": 6.117790084239341e-06, + "loss": 0.5202, + "step": 4551 + }, + { + "epoch": 1.4564846416382253, + "grad_norm": 0.5896328687667847, + "learning_rate": 6.115975332791446e-06, + "loss": 0.515, + "step": 4552 + }, + { + "epoch": 1.4568046075085324, + "grad_norm": 0.510565996170044, + "learning_rate": 6.114160426612229e-06, + "loss": 0.4589, + "step": 4553 + }, + { + "epoch": 1.4571245733788396, + "grad_norm": 0.5548239946365356, + "learning_rate": 6.112345365953327e-06, + "loss": 0.5438, + "step": 4554 + }, + { + "epoch": 1.4574445392491469, + "grad_norm": 0.4849468171596527, + "learning_rate": 6.110530151066404e-06, + "loss": 0.4236, + "step": 4555 + }, + { + "epoch": 1.4577645051194539, + "grad_norm": 0.5229912996292114, + "learning_rate": 6.108714782203139e-06, + "loss": 0.4821, + "step": 4556 + }, + { + "epoch": 1.4580844709897611, + "grad_norm": 0.4885832965373993, + "learning_rate": 6.1068992596152355e-06, + "loss": 0.4836, + "step": 4557 + }, + { + "epoch": 1.4584044368600684, + "grad_norm": 0.47455790638923645, + "learning_rate": 6.1050835835544174e-06, + "loss": 0.4684, + "step": 4558 + }, + { + "epoch": 1.4587244027303754, + "grad_norm": 0.583892822265625, + "learning_rate": 6.103267754272431e-06, + "loss": 0.5626, + "step": 4559 + }, + { + "epoch": 1.4590443686006827, + "grad_norm": 0.5615590214729309, + "learning_rate": 6.1014517720210444e-06, + "loss": 0.4889, + "step": 4560 + }, + { + "epoch": 1.45936433447099, + "grad_norm": 0.5332754254341125, + "learning_rate": 6.099635637052044e-06, + "loss": 0.5188, + "step": 4561 + }, + { + "epoch": 1.459684300341297, + "grad_norm": 0.5477079153060913, + "learning_rate": 6.097819349617239e-06, + "loss": 0.5068, + "step": 4562 + }, + { + "epoch": 1.4600042662116042, + "grad_norm": 0.5035220980644226, + "learning_rate": 6.096002909968461e-06, + "loss": 0.4824, + "step": 4563 + }, + { + "epoch": 1.4603242320819112, + "grad_norm": 0.5831959843635559, + "learning_rate": 6.094186318357561e-06, + "loss": 0.5456, + "step": 4564 + }, + { + "epoch": 1.4606441979522184, + "grad_norm": 0.46603646874427795, + "learning_rate": 6.092369575036411e-06, + "loss": 0.4154, + "step": 4565 + }, + { + "epoch": 1.4609641638225255, + "grad_norm": 0.5741982460021973, + "learning_rate": 6.090552680256907e-06, + "loss": 0.5277, + "step": 4566 + }, + { + "epoch": 1.4612841296928327, + "grad_norm": 0.4673265814781189, + "learning_rate": 6.088735634270959e-06, + "loss": 0.4391, + "step": 4567 + }, + { + "epoch": 1.46160409556314, + "grad_norm": 0.4824434816837311, + "learning_rate": 6.086918437330508e-06, + "loss": 0.48, + "step": 4568 + }, + { + "epoch": 1.461924061433447, + "grad_norm": 0.5617666244506836, + "learning_rate": 6.0851010896875074e-06, + "loss": 0.4755, + "step": 4569 + }, + { + "epoch": 1.4622440273037542, + "grad_norm": 0.5845708250999451, + "learning_rate": 6.0832835915939346e-06, + "loss": 0.5971, + "step": 4570 + }, + { + "epoch": 1.4625639931740615, + "grad_norm": 0.5152300596237183, + "learning_rate": 6.08146594330179e-06, + "loss": 0.5032, + "step": 4571 + }, + { + "epoch": 1.4628839590443685, + "grad_norm": 0.4916346073150635, + "learning_rate": 6.079648145063089e-06, + "loss": 0.4346, + "step": 4572 + }, + { + "epoch": 1.4632039249146758, + "grad_norm": 0.5176913738250732, + "learning_rate": 6.077830197129877e-06, + "loss": 0.5062, + "step": 4573 + }, + { + "epoch": 1.463523890784983, + "grad_norm": 0.5129452347755432, + "learning_rate": 6.076012099754209e-06, + "loss": 0.4645, + "step": 4574 + }, + { + "epoch": 1.46384385665529, + "grad_norm": 0.5412490963935852, + "learning_rate": 6.074193853188171e-06, + "loss": 0.5391, + "step": 4575 + }, + { + "epoch": 1.4641638225255973, + "grad_norm": 0.534808337688446, + "learning_rate": 6.072375457683861e-06, + "loss": 0.4695, + "step": 4576 + }, + { + "epoch": 1.4644837883959045, + "grad_norm": 0.5926088094711304, + "learning_rate": 6.070556913493404e-06, + "loss": 0.566, + "step": 4577 + }, + { + "epoch": 1.4648037542662116, + "grad_norm": 0.5333240032196045, + "learning_rate": 6.068738220868944e-06, + "loss": 0.4734, + "step": 4578 + }, + { + "epoch": 1.4651237201365188, + "grad_norm": 0.5552252531051636, + "learning_rate": 6.066919380062644e-06, + "loss": 0.5371, + "step": 4579 + }, + { + "epoch": 1.465443686006826, + "grad_norm": 0.5706639289855957, + "learning_rate": 6.065100391326687e-06, + "loss": 0.4958, + "step": 4580 + }, + { + "epoch": 1.465763651877133, + "grad_norm": 0.557990550994873, + "learning_rate": 6.06328125491328e-06, + "loss": 0.4884, + "step": 4581 + }, + { + "epoch": 1.4660836177474403, + "grad_norm": 0.4787043035030365, + "learning_rate": 6.061461971074648e-06, + "loss": 0.4394, + "step": 4582 + }, + { + "epoch": 1.4664035836177474, + "grad_norm": 0.5325111746788025, + "learning_rate": 6.059642540063038e-06, + "loss": 0.5368, + "step": 4583 + }, + { + "epoch": 1.4667235494880546, + "grad_norm": 0.5462748408317566, + "learning_rate": 6.0578229621307115e-06, + "loss": 0.4885, + "step": 4584 + }, + { + "epoch": 1.4670435153583616, + "grad_norm": 0.5821309089660645, + "learning_rate": 6.056003237529963e-06, + "loss": 0.514, + "step": 4585 + }, + { + "epoch": 1.4673634812286689, + "grad_norm": 0.5710198879241943, + "learning_rate": 6.054183366513092e-06, + "loss": 0.4664, + "step": 4586 + }, + { + "epoch": 1.4676834470989761, + "grad_norm": 0.5251044631004333, + "learning_rate": 6.0523633493324306e-06, + "loss": 0.4987, + "step": 4587 + }, + { + "epoch": 1.4680034129692832, + "grad_norm": 0.4900965392589569, + "learning_rate": 6.050543186240325e-06, + "loss": 0.4147, + "step": 4588 + }, + { + "epoch": 1.4683233788395904, + "grad_norm": 0.556492030620575, + "learning_rate": 6.048722877489141e-06, + "loss": 0.5776, + "step": 4589 + }, + { + "epoch": 1.4686433447098977, + "grad_norm": 0.5008504390716553, + "learning_rate": 6.046902423331271e-06, + "loss": 0.4207, + "step": 4590 + }, + { + "epoch": 1.4689633105802047, + "grad_norm": 0.5561898350715637, + "learning_rate": 6.045081824019119e-06, + "loss": 0.5348, + "step": 4591 + }, + { + "epoch": 1.469283276450512, + "grad_norm": 0.5290460586547852, + "learning_rate": 6.043261079805115e-06, + "loss": 0.4541, + "step": 4592 + }, + { + "epoch": 1.4696032423208192, + "grad_norm": 0.5425384640693665, + "learning_rate": 6.041440190941707e-06, + "loss": 0.4747, + "step": 4593 + }, + { + "epoch": 1.4699232081911262, + "grad_norm": 0.5726215243339539, + "learning_rate": 6.039619157681363e-06, + "loss": 0.5583, + "step": 4594 + }, + { + "epoch": 1.4702431740614335, + "grad_norm": 0.5443554520606995, + "learning_rate": 6.037797980276574e-06, + "loss": 0.4511, + "step": 4595 + }, + { + "epoch": 1.4705631399317407, + "grad_norm": 0.550445020198822, + "learning_rate": 6.035976658979846e-06, + "loss": 0.5567, + "step": 4596 + }, + { + "epoch": 1.4708831058020477, + "grad_norm": 0.5292520523071289, + "learning_rate": 6.034155194043707e-06, + "loss": 0.4503, + "step": 4597 + }, + { + "epoch": 1.471203071672355, + "grad_norm": 0.5710574984550476, + "learning_rate": 6.032333585720707e-06, + "loss": 0.4981, + "step": 4598 + }, + { + "epoch": 1.4715230375426622, + "grad_norm": 0.5461126565933228, + "learning_rate": 6.0305118342634125e-06, + "loss": 0.5255, + "step": 4599 + }, + { + "epoch": 1.4718430034129693, + "grad_norm": 0.5360023975372314, + "learning_rate": 6.028689939924413e-06, + "loss": 0.4961, + "step": 4600 + }, + { + "epoch": 1.4721629692832765, + "grad_norm": 0.6214020252227783, + "learning_rate": 6.026867902956317e-06, + "loss": 0.4996, + "step": 4601 + }, + { + "epoch": 1.4724829351535837, + "grad_norm": 0.5595095157623291, + "learning_rate": 6.025045723611749e-06, + "loss": 0.5351, + "step": 4602 + }, + { + "epoch": 1.4728029010238908, + "grad_norm": 0.5649869441986084, + "learning_rate": 6.02322340214336e-06, + "loss": 0.4554, + "step": 4603 + }, + { + "epoch": 1.473122866894198, + "grad_norm": 0.6052053570747375, + "learning_rate": 6.021400938803813e-06, + "loss": 0.5152, + "step": 4604 + }, + { + "epoch": 1.473442832764505, + "grad_norm": 0.5253227949142456, + "learning_rate": 6.019578333845799e-06, + "loss": 0.5051, + "step": 4605 + }, + { + "epoch": 1.4737627986348123, + "grad_norm": 0.5365368723869324, + "learning_rate": 6.0177555875220215e-06, + "loss": 0.4785, + "step": 4606 + }, + { + "epoch": 1.4740827645051193, + "grad_norm": 0.6048970818519592, + "learning_rate": 6.015932700085207e-06, + "loss": 0.485, + "step": 4607 + }, + { + "epoch": 1.4744027303754266, + "grad_norm": 0.5239445567131042, + "learning_rate": 6.014109671788105e-06, + "loss": 0.4881, + "step": 4608 + }, + { + "epoch": 1.4747226962457338, + "grad_norm": 0.5779982209205627, + "learning_rate": 6.012286502883472e-06, + "loss": 0.5032, + "step": 4609 + }, + { + "epoch": 1.4750426621160408, + "grad_norm": 0.6136361360549927, + "learning_rate": 6.010463193624101e-06, + "loss": 0.5191, + "step": 4610 + }, + { + "epoch": 1.475362627986348, + "grad_norm": 0.5763174295425415, + "learning_rate": 6.008639744262791e-06, + "loss": 0.4517, + "step": 4611 + }, + { + "epoch": 1.4756825938566553, + "grad_norm": 0.6773867607116699, + "learning_rate": 6.006816155052366e-06, + "loss": 0.5122, + "step": 4612 + }, + { + "epoch": 1.4760025597269624, + "grad_norm": 0.5779669284820557, + "learning_rate": 6.004992426245673e-06, + "loss": 0.4881, + "step": 4613 + }, + { + "epoch": 1.4763225255972696, + "grad_norm": 0.5112670063972473, + "learning_rate": 6.003168558095568e-06, + "loss": 0.4937, + "step": 4614 + }, + { + "epoch": 1.4766424914675769, + "grad_norm": 0.5309809446334839, + "learning_rate": 6.001344550854937e-06, + "loss": 0.4664, + "step": 4615 + }, + { + "epoch": 1.476962457337884, + "grad_norm": 0.6619424223899841, + "learning_rate": 5.99952040477668e-06, + "loss": 0.5895, + "step": 4616 + }, + { + "epoch": 1.4772824232081911, + "grad_norm": 0.5281962752342224, + "learning_rate": 5.9976961201137155e-06, + "loss": 0.4549, + "step": 4617 + }, + { + "epoch": 1.4776023890784984, + "grad_norm": 0.504805326461792, + "learning_rate": 5.995871697118983e-06, + "loss": 0.5001, + "step": 4618 + }, + { + "epoch": 1.4779223549488054, + "grad_norm": 0.5093720555305481, + "learning_rate": 5.994047136045442e-06, + "loss": 0.4665, + "step": 4619 + }, + { + "epoch": 1.4782423208191127, + "grad_norm": 0.4796576499938965, + "learning_rate": 5.992222437146072e-06, + "loss": 0.5012, + "step": 4620 + }, + { + "epoch": 1.47856228668942, + "grad_norm": 0.505854606628418, + "learning_rate": 5.990397600673865e-06, + "loss": 0.4341, + "step": 4621 + }, + { + "epoch": 1.478882252559727, + "grad_norm": 0.5118453502655029, + "learning_rate": 5.9885726268818385e-06, + "loss": 0.5302, + "step": 4622 + }, + { + "epoch": 1.4792022184300342, + "grad_norm": 0.5388093590736389, + "learning_rate": 5.986747516023031e-06, + "loss": 0.4738, + "step": 4623 + }, + { + "epoch": 1.4795221843003412, + "grad_norm": 0.529425859451294, + "learning_rate": 5.98492226835049e-06, + "loss": 0.5502, + "step": 4624 + }, + { + "epoch": 1.4798421501706485, + "grad_norm": 0.5454614758491516, + "learning_rate": 5.983096884117294e-06, + "loss": 0.4309, + "step": 4625 + }, + { + "epoch": 1.4801621160409557, + "grad_norm": 0.5218547582626343, + "learning_rate": 5.9812713635765315e-06, + "loss": 0.5107, + "step": 4626 + }, + { + "epoch": 1.4804820819112627, + "grad_norm": 0.5063565969467163, + "learning_rate": 5.979445706981314e-06, + "loss": 0.4839, + "step": 4627 + }, + { + "epoch": 1.48080204778157, + "grad_norm": 0.5817762613296509, + "learning_rate": 5.977619914584771e-06, + "loss": 0.4742, + "step": 4628 + }, + { + "epoch": 1.481122013651877, + "grad_norm": 0.5883211493492126, + "learning_rate": 5.975793986640051e-06, + "loss": 0.5371, + "step": 4629 + }, + { + "epoch": 1.4814419795221843, + "grad_norm": 0.5384506583213806, + "learning_rate": 5.973967923400321e-06, + "loss": 0.5137, + "step": 4630 + }, + { + "epoch": 1.4817619453924915, + "grad_norm": 0.47395968437194824, + "learning_rate": 5.9721417251187655e-06, + "loss": 0.4571, + "step": 4631 + }, + { + "epoch": 1.4820819112627985, + "grad_norm": 0.5172684192657471, + "learning_rate": 5.970315392048591e-06, + "loss": 0.5435, + "step": 4632 + }, + { + "epoch": 1.4824018771331058, + "grad_norm": 0.5081861615180969, + "learning_rate": 5.96848892444302e-06, + "loss": 0.4908, + "step": 4633 + }, + { + "epoch": 1.482721843003413, + "grad_norm": 0.5542958378791809, + "learning_rate": 5.966662322555295e-06, + "loss": 0.4779, + "step": 4634 + }, + { + "epoch": 1.48304180887372, + "grad_norm": 0.4761788547039032, + "learning_rate": 5.964835586638674e-06, + "loss": 0.4499, + "step": 4635 + }, + { + "epoch": 1.4833617747440273, + "grad_norm": 0.4623819887638092, + "learning_rate": 5.963008716946441e-06, + "loss": 0.4911, + "step": 4636 + }, + { + "epoch": 1.4836817406143346, + "grad_norm": 0.4585903584957123, + "learning_rate": 5.96118171373189e-06, + "loss": 0.4528, + "step": 4637 + }, + { + "epoch": 1.4840017064846416, + "grad_norm": 0.5143521428108215, + "learning_rate": 5.959354577248337e-06, + "loss": 0.5496, + "step": 4638 + }, + { + "epoch": 1.4843216723549488, + "grad_norm": 0.4974868595600128, + "learning_rate": 5.957527307749118e-06, + "loss": 0.4704, + "step": 4639 + }, + { + "epoch": 1.484641638225256, + "grad_norm": 0.5384208559989929, + "learning_rate": 5.9556999054875864e-06, + "loss": 0.4347, + "step": 4640 + }, + { + "epoch": 1.484961604095563, + "grad_norm": 0.528374433517456, + "learning_rate": 5.953872370717114e-06, + "loss": 0.5509, + "step": 4641 + }, + { + "epoch": 1.4852815699658704, + "grad_norm": 0.45676034688949585, + "learning_rate": 5.952044703691089e-06, + "loss": 0.4405, + "step": 4642 + }, + { + "epoch": 1.4856015358361776, + "grad_norm": 0.5470004677772522, + "learning_rate": 5.95021690466292e-06, + "loss": 0.5111, + "step": 4643 + }, + { + "epoch": 1.4859215017064846, + "grad_norm": 0.5926286578178406, + "learning_rate": 5.948388973886035e-06, + "loss": 0.4974, + "step": 4644 + }, + { + "epoch": 1.4862414675767919, + "grad_norm": 0.5255135893821716, + "learning_rate": 5.946560911613877e-06, + "loss": 0.5402, + "step": 4645 + }, + { + "epoch": 1.486561433447099, + "grad_norm": 0.49791911244392395, + "learning_rate": 5.94473271809991e-06, + "loss": 0.4569, + "step": 4646 + }, + { + "epoch": 1.4868813993174061, + "grad_norm": 0.45523884892463684, + "learning_rate": 5.942904393597616e-06, + "loss": 0.4204, + "step": 4647 + }, + { + "epoch": 1.4872013651877132, + "grad_norm": 0.5056770443916321, + "learning_rate": 5.9410759383604925e-06, + "loss": 0.55, + "step": 4648 + }, + { + "epoch": 1.4875213310580204, + "grad_norm": 0.49918273091316223, + "learning_rate": 5.939247352642058e-06, + "loss": 0.5166, + "step": 4649 + }, + { + "epoch": 1.4878412969283277, + "grad_norm": 0.512938916683197, + "learning_rate": 5.937418636695849e-06, + "loss": 0.5038, + "step": 4650 + }, + { + "epoch": 1.4881612627986347, + "grad_norm": 0.5014441013336182, + "learning_rate": 5.935589790775417e-06, + "loss": 0.4865, + "step": 4651 + }, + { + "epoch": 1.488481228668942, + "grad_norm": 0.4981122314929962, + "learning_rate": 5.933760815134335e-06, + "loss": 0.4903, + "step": 4652 + }, + { + "epoch": 1.4888011945392492, + "grad_norm": 0.552680253982544, + "learning_rate": 5.9319317100261944e-06, + "loss": 0.5292, + "step": 4653 + }, + { + "epoch": 1.4891211604095562, + "grad_norm": 0.4756479561328888, + "learning_rate": 5.9301024757046e-06, + "loss": 0.4578, + "step": 4654 + }, + { + "epoch": 1.4894411262798635, + "grad_norm": 0.511592447757721, + "learning_rate": 5.928273112423177e-06, + "loss": 0.481, + "step": 4655 + }, + { + "epoch": 1.4897610921501707, + "grad_norm": 0.5142168402671814, + "learning_rate": 5.926443620435572e-06, + "loss": 0.5207, + "step": 4656 + }, + { + "epoch": 1.4900810580204777, + "grad_norm": 0.5315825343132019, + "learning_rate": 5.924613999995443e-06, + "loss": 0.4931, + "step": 4657 + }, + { + "epoch": 1.490401023890785, + "grad_norm": 0.5098580121994019, + "learning_rate": 5.922784251356474e-06, + "loss": 0.4662, + "step": 4658 + }, + { + "epoch": 1.4907209897610922, + "grad_norm": 0.5362950563430786, + "learning_rate": 5.920954374772356e-06, + "loss": 0.5303, + "step": 4659 + }, + { + "epoch": 1.4910409556313993, + "grad_norm": 0.5265049338340759, + "learning_rate": 5.919124370496805e-06, + "loss": 0.4547, + "step": 4660 + }, + { + "epoch": 1.4913609215017065, + "grad_norm": 0.4737069010734558, + "learning_rate": 5.917294238783555e-06, + "loss": 0.4993, + "step": 4661 + }, + { + "epoch": 1.4916808873720138, + "grad_norm": 0.4527972936630249, + "learning_rate": 5.915463979886356e-06, + "loss": 0.4522, + "step": 4662 + }, + { + "epoch": 1.4920008532423208, + "grad_norm": 0.5019689798355103, + "learning_rate": 5.913633594058975e-06, + "loss": 0.4868, + "step": 4663 + }, + { + "epoch": 1.492320819112628, + "grad_norm": 0.530006468296051, + "learning_rate": 5.911803081555197e-06, + "loss": 0.4841, + "step": 4664 + }, + { + "epoch": 1.4926407849829353, + "grad_norm": 0.5467323064804077, + "learning_rate": 5.909972442628823e-06, + "loss": 0.449, + "step": 4665 + }, + { + "epoch": 1.4929607508532423, + "grad_norm": 0.5686629414558411, + "learning_rate": 5.908141677533676e-06, + "loss": 0.508, + "step": 4666 + }, + { + "epoch": 1.4932807167235496, + "grad_norm": 0.5066972970962524, + "learning_rate": 5.906310786523594e-06, + "loss": 0.4984, + "step": 4667 + }, + { + "epoch": 1.4936006825938566, + "grad_norm": 0.5746550559997559, + "learning_rate": 5.90447976985243e-06, + "loss": 0.4575, + "step": 4668 + }, + { + "epoch": 1.4939206484641638, + "grad_norm": 0.6012100577354431, + "learning_rate": 5.902648627774059e-06, + "loss": 0.5184, + "step": 4669 + }, + { + "epoch": 1.4942406143344709, + "grad_norm": 0.5028852820396423, + "learning_rate": 5.900817360542369e-06, + "loss": 0.4661, + "step": 4670 + }, + { + "epoch": 1.494560580204778, + "grad_norm": 0.5793988108634949, + "learning_rate": 5.898985968411268e-06, + "loss": 0.5398, + "step": 4671 + }, + { + "epoch": 1.4948805460750854, + "grad_norm": 0.5978303551673889, + "learning_rate": 5.897154451634681e-06, + "loss": 0.505, + "step": 4672 + }, + { + "epoch": 1.4952005119453924, + "grad_norm": 0.49366962909698486, + "learning_rate": 5.8953228104665505e-06, + "loss": 0.4814, + "step": 4673 + }, + { + "epoch": 1.4955204778156996, + "grad_norm": 0.4858211278915405, + "learning_rate": 5.893491045160834e-06, + "loss": 0.4624, + "step": 4674 + }, + { + "epoch": 1.4958404436860069, + "grad_norm": 0.5152391195297241, + "learning_rate": 5.891659155971512e-06, + "loss": 0.503, + "step": 4675 + }, + { + "epoch": 1.496160409556314, + "grad_norm": 0.5337923169136047, + "learning_rate": 5.8898271431525725e-06, + "loss": 0.4369, + "step": 4676 + }, + { + "epoch": 1.4964803754266212, + "grad_norm": 0.5717154741287231, + "learning_rate": 5.887995006958031e-06, + "loss": 0.4911, + "step": 4677 + }, + { + "epoch": 1.4968003412969284, + "grad_norm": 0.6406721472740173, + "learning_rate": 5.886162747641912e-06, + "loss": 0.5007, + "step": 4678 + }, + { + "epoch": 1.4971203071672354, + "grad_norm": 0.5949901938438416, + "learning_rate": 5.884330365458261e-06, + "loss": 0.5153, + "step": 4679 + }, + { + "epoch": 1.4974402730375427, + "grad_norm": 0.5068642497062683, + "learning_rate": 5.882497860661145e-06, + "loss": 0.4907, + "step": 4680 + }, + { + "epoch": 1.49776023890785, + "grad_norm": 0.5365095138549805, + "learning_rate": 5.880665233504636e-06, + "loss": 0.4697, + "step": 4681 + }, + { + "epoch": 1.498080204778157, + "grad_norm": 0.5669816732406616, + "learning_rate": 5.878832484242833e-06, + "loss": 0.5335, + "step": 4682 + }, + { + "epoch": 1.4984001706484642, + "grad_norm": 0.5329282879829407, + "learning_rate": 5.87699961312985e-06, + "loss": 0.4687, + "step": 4683 + }, + { + "epoch": 1.4987201365187715, + "grad_norm": 0.5123113393783569, + "learning_rate": 5.875166620419813e-06, + "loss": 0.4799, + "step": 4684 + }, + { + "epoch": 1.4990401023890785, + "grad_norm": 0.5470655560493469, + "learning_rate": 5.873333506366874e-06, + "loss": 0.5002, + "step": 4685 + }, + { + "epoch": 1.4993600682593857, + "grad_norm": 0.5343730449676514, + "learning_rate": 5.871500271225193e-06, + "loss": 0.5105, + "step": 4686 + }, + { + "epoch": 1.4996800341296928, + "grad_norm": 0.5782073736190796, + "learning_rate": 5.869666915248951e-06, + "loss": 0.5231, + "step": 4687 + }, + { + "epoch": 1.5, + "grad_norm": 0.6019355058670044, + "learning_rate": 5.867833438692346e-06, + "loss": 0.4944, + "step": 4688 + }, + { + "epoch": 1.500319965870307, + "grad_norm": 0.552795946598053, + "learning_rate": 5.8659998418095865e-06, + "loss": 0.421, + "step": 4689 + }, + { + "epoch": 1.5006399317406145, + "grad_norm": 0.5298478603363037, + "learning_rate": 5.864166124854911e-06, + "loss": 0.5113, + "step": 4690 + }, + { + "epoch": 1.5009598976109215, + "grad_norm": 0.5174188613891602, + "learning_rate": 5.862332288082562e-06, + "loss": 0.4433, + "step": 4691 + }, + { + "epoch": 1.5012798634812285, + "grad_norm": 0.559489905834198, + "learning_rate": 5.8604983317468036e-06, + "loss": 0.4896, + "step": 4692 + }, + { + "epoch": 1.5015998293515358, + "grad_norm": 0.5089955925941467, + "learning_rate": 5.858664256101917e-06, + "loss": 0.4669, + "step": 4693 + }, + { + "epoch": 1.501919795221843, + "grad_norm": 0.5707961916923523, + "learning_rate": 5.856830061402196e-06, + "loss": 0.513, + "step": 4694 + }, + { + "epoch": 1.50223976109215, + "grad_norm": 0.5894507765769958, + "learning_rate": 5.854995747901958e-06, + "loss": 0.5176, + "step": 4695 + }, + { + "epoch": 1.5025597269624573, + "grad_norm": 0.5667989253997803, + "learning_rate": 5.853161315855531e-06, + "loss": 0.4725, + "step": 4696 + }, + { + "epoch": 1.5028796928327646, + "grad_norm": 0.5133673548698425, + "learning_rate": 5.85132676551726e-06, + "loss": 0.4423, + "step": 4697 + }, + { + "epoch": 1.5031996587030716, + "grad_norm": 0.5565043091773987, + "learning_rate": 5.8494920971415105e-06, + "loss": 0.5359, + "step": 4698 + }, + { + "epoch": 1.5035196245733788, + "grad_norm": 0.5063019394874573, + "learning_rate": 5.847657310982658e-06, + "loss": 0.5081, + "step": 4699 + }, + { + "epoch": 1.503839590443686, + "grad_norm": 0.58235102891922, + "learning_rate": 5.845822407295101e-06, + "loss": 0.5608, + "step": 4700 + }, + { + "epoch": 1.5041595563139931, + "grad_norm": 0.4960373044013977, + "learning_rate": 5.843987386333248e-06, + "loss": 0.4317, + "step": 4701 + }, + { + "epoch": 1.5044795221843004, + "grad_norm": 0.5047836303710938, + "learning_rate": 5.842152248351531e-06, + "loss": 0.4673, + "step": 4702 + }, + { + "epoch": 1.5047994880546076, + "grad_norm": 0.5804183483123779, + "learning_rate": 5.84031699360439e-06, + "loss": 0.5219, + "step": 4703 + }, + { + "epoch": 1.5051194539249146, + "grad_norm": 0.5353643298149109, + "learning_rate": 5.838481622346286e-06, + "loss": 0.5067, + "step": 4704 + }, + { + "epoch": 1.505439419795222, + "grad_norm": 0.5168746113777161, + "learning_rate": 5.836646134831698e-06, + "loss": 0.4733, + "step": 4705 + }, + { + "epoch": 1.5057593856655291, + "grad_norm": 0.5259865522384644, + "learning_rate": 5.834810531315117e-06, + "loss": 0.4797, + "step": 4706 + }, + { + "epoch": 1.5060793515358362, + "grad_norm": 0.5396210551261902, + "learning_rate": 5.832974812051051e-06, + "loss": 0.463, + "step": 4707 + }, + { + "epoch": 1.5063993174061432, + "grad_norm": 0.5567886233329773, + "learning_rate": 5.831138977294025e-06, + "loss": 0.6018, + "step": 4708 + }, + { + "epoch": 1.5067192832764507, + "grad_norm": 0.48316898941993713, + "learning_rate": 5.8293030272985794e-06, + "loss": 0.4108, + "step": 4709 + }, + { + "epoch": 1.5070392491467577, + "grad_norm": 0.5766081213951111, + "learning_rate": 5.827466962319274e-06, + "loss": 0.5038, + "step": 4710 + }, + { + "epoch": 1.5073592150170647, + "grad_norm": 0.5991988778114319, + "learning_rate": 5.825630782610676e-06, + "loss": 0.477, + "step": 4711 + }, + { + "epoch": 1.507679180887372, + "grad_norm": 0.5347451567649841, + "learning_rate": 5.823794488427378e-06, + "loss": 0.4746, + "step": 4712 + }, + { + "epoch": 1.5079991467576792, + "grad_norm": 0.5310901403427124, + "learning_rate": 5.821958080023982e-06, + "loss": 0.5233, + "step": 4713 + }, + { + "epoch": 1.5083191126279862, + "grad_norm": 0.5356836915016174, + "learning_rate": 5.820121557655109e-06, + "loss": 0.4868, + "step": 4714 + }, + { + "epoch": 1.5086390784982935, + "grad_norm": 0.5663773417472839, + "learning_rate": 5.818284921575398e-06, + "loss": 0.5044, + "step": 4715 + }, + { + "epoch": 1.5089590443686007, + "grad_norm": 0.5553083419799805, + "learning_rate": 5.816448172039497e-06, + "loss": 0.4643, + "step": 4716 + }, + { + "epoch": 1.5092790102389078, + "grad_norm": 0.5697829723358154, + "learning_rate": 5.814611309302074e-06, + "loss": 0.5406, + "step": 4717 + }, + { + "epoch": 1.509598976109215, + "grad_norm": 0.5327697992324829, + "learning_rate": 5.8127743336178136e-06, + "loss": 0.4588, + "step": 4718 + }, + { + "epoch": 1.5099189419795223, + "grad_norm": 0.5630555748939514, + "learning_rate": 5.810937245241414e-06, + "loss": 0.4749, + "step": 4719 + }, + { + "epoch": 1.5102389078498293, + "grad_norm": 0.5737113952636719, + "learning_rate": 5.8091000444275904e-06, + "loss": 0.505, + "step": 4720 + }, + { + "epoch": 1.5105588737201365, + "grad_norm": 0.5312773585319519, + "learning_rate": 5.807262731431069e-06, + "loss": 0.4828, + "step": 4721 + }, + { + "epoch": 1.5108788395904438, + "grad_norm": 0.5531858205795288, + "learning_rate": 5.805425306506602e-06, + "loss": 0.5074, + "step": 4722 + }, + { + "epoch": 1.5111988054607508, + "grad_norm": 0.48062023520469666, + "learning_rate": 5.8035877699089475e-06, + "loss": 0.484, + "step": 4723 + }, + { + "epoch": 1.511518771331058, + "grad_norm": 0.46315205097198486, + "learning_rate": 5.801750121892881e-06, + "loss": 0.4692, + "step": 4724 + }, + { + "epoch": 1.5118387372013653, + "grad_norm": 0.5213401913642883, + "learning_rate": 5.799912362713195e-06, + "loss": 0.4811, + "step": 4725 + }, + { + "epoch": 1.5121587030716723, + "grad_norm": 0.535945475101471, + "learning_rate": 5.798074492624698e-06, + "loss": 0.4338, + "step": 4726 + }, + { + "epoch": 1.5124786689419796, + "grad_norm": 0.5230036973953247, + "learning_rate": 5.796236511882214e-06, + "loss": 0.4526, + "step": 4727 + }, + { + "epoch": 1.5127986348122868, + "grad_norm": 0.5411539077758789, + "learning_rate": 5.7943984207405776e-06, + "loss": 0.5429, + "step": 4728 + }, + { + "epoch": 1.5131186006825939, + "grad_norm": 0.5598258376121521, + "learning_rate": 5.792560219454645e-06, + "loss": 0.478, + "step": 4729 + }, + { + "epoch": 1.5134385665529009, + "grad_norm": 0.5569031834602356, + "learning_rate": 5.790721908279284e-06, + "loss": 0.504, + "step": 4730 + }, + { + "epoch": 1.5137585324232083, + "grad_norm": 0.5236627459526062, + "learning_rate": 5.78888348746938e-06, + "loss": 0.472, + "step": 4731 + }, + { + "epoch": 1.5140784982935154, + "grad_norm": 0.5903027057647705, + "learning_rate": 5.787044957279833e-06, + "loss": 0.5809, + "step": 4732 + }, + { + "epoch": 1.5143984641638224, + "grad_norm": 0.4684828519821167, + "learning_rate": 5.785206317965554e-06, + "loss": 0.4436, + "step": 4733 + }, + { + "epoch": 1.5147184300341296, + "grad_norm": 0.5297166109085083, + "learning_rate": 5.783367569781474e-06, + "loss": 0.4934, + "step": 4734 + }, + { + "epoch": 1.515038395904437, + "grad_norm": 0.5144308805465698, + "learning_rate": 5.781528712982538e-06, + "loss": 0.4815, + "step": 4735 + }, + { + "epoch": 1.515358361774744, + "grad_norm": 0.5451584458351135, + "learning_rate": 5.779689747823707e-06, + "loss": 0.466, + "step": 4736 + }, + { + "epoch": 1.5156783276450512, + "grad_norm": 0.509016215801239, + "learning_rate": 5.777850674559954e-06, + "loss": 0.4894, + "step": 4737 + }, + { + "epoch": 1.5159982935153584, + "grad_norm": 0.5265939831733704, + "learning_rate": 5.776011493446269e-06, + "loss": 0.5545, + "step": 4738 + }, + { + "epoch": 1.5163182593856654, + "grad_norm": 0.5136027932167053, + "learning_rate": 5.7741722047376565e-06, + "loss": 0.4837, + "step": 4739 + }, + { + "epoch": 1.5166382252559727, + "grad_norm": 0.5121413469314575, + "learning_rate": 5.772332808689136e-06, + "loss": 0.4503, + "step": 4740 + }, + { + "epoch": 1.51695819112628, + "grad_norm": 0.6478096842765808, + "learning_rate": 5.770493305555743e-06, + "loss": 0.5672, + "step": 4741 + }, + { + "epoch": 1.517278156996587, + "grad_norm": 0.514471709728241, + "learning_rate": 5.768653695592527e-06, + "loss": 0.5323, + "step": 4742 + }, + { + "epoch": 1.5175981228668942, + "grad_norm": 0.46245503425598145, + "learning_rate": 5.766813979054552e-06, + "loss": 0.4682, + "step": 4743 + }, + { + "epoch": 1.5179180887372015, + "grad_norm": 0.5243746638298035, + "learning_rate": 5.764974156196895e-06, + "loss": 0.4612, + "step": 4744 + }, + { + "epoch": 1.5182380546075085, + "grad_norm": 0.5565612316131592, + "learning_rate": 5.76313422727465e-06, + "loss": 0.4659, + "step": 4745 + }, + { + "epoch": 1.5185580204778157, + "grad_norm": 0.48753592371940613, + "learning_rate": 5.761294192542928e-06, + "loss": 0.4612, + "step": 4746 + }, + { + "epoch": 1.518877986348123, + "grad_norm": 0.5521674156188965, + "learning_rate": 5.7594540522568495e-06, + "loss": 0.4912, + "step": 4747 + }, + { + "epoch": 1.51919795221843, + "grad_norm": 0.5125954151153564, + "learning_rate": 5.757613806671555e-06, + "loss": 0.5305, + "step": 4748 + }, + { + "epoch": 1.519517918088737, + "grad_norm": 0.474223256111145, + "learning_rate": 5.755773456042193e-06, + "loss": 0.4656, + "step": 4749 + }, + { + "epoch": 1.5198378839590445, + "grad_norm": 0.496297687292099, + "learning_rate": 5.753933000623932e-06, + "loss": 0.4613, + "step": 4750 + }, + { + "epoch": 1.5201578498293515, + "grad_norm": 0.5795848369598389, + "learning_rate": 5.752092440671952e-06, + "loss": 0.4818, + "step": 4751 + }, + { + "epoch": 1.5204778156996586, + "grad_norm": 0.6200262308120728, + "learning_rate": 5.750251776441453e-06, + "loss": 0.5102, + "step": 4752 + }, + { + "epoch": 1.520797781569966, + "grad_norm": 0.5648042559623718, + "learning_rate": 5.748411008187642e-06, + "loss": 0.5391, + "step": 4753 + }, + { + "epoch": 1.521117747440273, + "grad_norm": 0.5272075533866882, + "learning_rate": 5.746570136165744e-06, + "loss": 0.4633, + "step": 4754 + }, + { + "epoch": 1.52143771331058, + "grad_norm": 0.5459337830543518, + "learning_rate": 5.744729160630998e-06, + "loss": 0.5066, + "step": 4755 + }, + { + "epoch": 1.5217576791808873, + "grad_norm": 0.5689736604690552, + "learning_rate": 5.7428880818386574e-06, + "loss": 0.5233, + "step": 4756 + }, + { + "epoch": 1.5220776450511946, + "grad_norm": 0.4891079366207123, + "learning_rate": 5.741046900043992e-06, + "loss": 0.439, + "step": 4757 + }, + { + "epoch": 1.5223976109215016, + "grad_norm": 0.5367847084999084, + "learning_rate": 5.7392056155022835e-06, + "loss": 0.5036, + "step": 4758 + }, + { + "epoch": 1.5227175767918089, + "grad_norm": 0.5277146100997925, + "learning_rate": 5.737364228468823e-06, + "loss": 0.4871, + "step": 4759 + }, + { + "epoch": 1.523037542662116, + "grad_norm": 0.5709060430526733, + "learning_rate": 5.73552273919893e-06, + "loss": 0.5384, + "step": 4760 + }, + { + "epoch": 1.5233575085324231, + "grad_norm": 0.5281832218170166, + "learning_rate": 5.7336811479479215e-06, + "loss": 0.469, + "step": 4761 + }, + { + "epoch": 1.5236774744027304, + "grad_norm": 0.5688793659210205, + "learning_rate": 5.7318394549711395e-06, + "loss": 0.4974, + "step": 4762 + }, + { + "epoch": 1.5239974402730376, + "grad_norm": 0.586227297782898, + "learning_rate": 5.729997660523938e-06, + "loss": 0.5188, + "step": 4763 + }, + { + "epoch": 1.5243174061433447, + "grad_norm": 0.5579555034637451, + "learning_rate": 5.72815576486168e-06, + "loss": 0.5151, + "step": 4764 + }, + { + "epoch": 1.524637372013652, + "grad_norm": 0.5547084808349609, + "learning_rate": 5.7263137682397515e-06, + "loss": 0.5007, + "step": 4765 + }, + { + "epoch": 1.5249573378839592, + "grad_norm": 0.5869400501251221, + "learning_rate": 5.724471670913545e-06, + "loss": 0.508, + "step": 4766 + }, + { + "epoch": 1.5252773037542662, + "grad_norm": 0.5429580211639404, + "learning_rate": 5.722629473138469e-06, + "loss": 0.4988, + "step": 4767 + }, + { + "epoch": 1.5255972696245734, + "grad_norm": 0.528658390045166, + "learning_rate": 5.720787175169949e-06, + "loss": 0.5182, + "step": 4768 + }, + { + "epoch": 1.5259172354948807, + "grad_norm": 0.5507444143295288, + "learning_rate": 5.718944777263417e-06, + "loss": 0.5175, + "step": 4769 + }, + { + "epoch": 1.5262372013651877, + "grad_norm": 0.48885491490364075, + "learning_rate": 5.717102279674329e-06, + "loss": 0.4476, + "step": 4770 + }, + { + "epoch": 1.5265571672354947, + "grad_norm": 0.5946596264839172, + "learning_rate": 5.715259682658145e-06, + "loss": 0.5351, + "step": 4771 + }, + { + "epoch": 1.5268771331058022, + "grad_norm": 0.5150769948959351, + "learning_rate": 5.713416986470348e-06, + "loss": 0.4542, + "step": 4772 + }, + { + "epoch": 1.5271970989761092, + "grad_norm": 0.5974960327148438, + "learning_rate": 5.711574191366427e-06, + "loss": 0.5476, + "step": 4773 + }, + { + "epoch": 1.5275170648464163, + "grad_norm": 0.545831024646759, + "learning_rate": 5.709731297601887e-06, + "loss": 0.4823, + "step": 4774 + }, + { + "epoch": 1.5278370307167235, + "grad_norm": 0.552084743976593, + "learning_rate": 5.70788830543225e-06, + "loss": 0.4665, + "step": 4775 + }, + { + "epoch": 1.5281569965870307, + "grad_norm": 0.5725771188735962, + "learning_rate": 5.706045215113047e-06, + "loss": 0.4726, + "step": 4776 + }, + { + "epoch": 1.5284769624573378, + "grad_norm": 0.5605619549751282, + "learning_rate": 5.704202026899827e-06, + "loss": 0.4549, + "step": 4777 + }, + { + "epoch": 1.528796928327645, + "grad_norm": 0.5407191514968872, + "learning_rate": 5.702358741048149e-06, + "loss": 0.4926, + "step": 4778 + }, + { + "epoch": 1.5291168941979523, + "grad_norm": 0.573394775390625, + "learning_rate": 5.700515357813583e-06, + "loss": 0.5661, + "step": 4779 + }, + { + "epoch": 1.5294368600682593, + "grad_norm": 0.5355685949325562, + "learning_rate": 5.698671877451723e-06, + "loss": 0.5147, + "step": 4780 + }, + { + "epoch": 1.5297568259385665, + "grad_norm": 0.5256187915802002, + "learning_rate": 5.696828300218167e-06, + "loss": 0.4632, + "step": 4781 + }, + { + "epoch": 1.5300767918088738, + "grad_norm": 0.45056822896003723, + "learning_rate": 5.694984626368527e-06, + "loss": 0.4563, + "step": 4782 + }, + { + "epoch": 1.5303967576791808, + "grad_norm": 0.5010212659835815, + "learning_rate": 5.693140856158435e-06, + "loss": 0.4869, + "step": 4783 + }, + { + "epoch": 1.530716723549488, + "grad_norm": 0.4796273410320282, + "learning_rate": 5.691296989843525e-06, + "loss": 0.4452, + "step": 4784 + }, + { + "epoch": 1.5310366894197953, + "grad_norm": 0.5340262055397034, + "learning_rate": 5.6894530276794605e-06, + "loss": 0.5197, + "step": 4785 + }, + { + "epoch": 1.5313566552901023, + "grad_norm": 0.4710986912250519, + "learning_rate": 5.6876089699219016e-06, + "loss": 0.4281, + "step": 4786 + }, + { + "epoch": 1.5316766211604096, + "grad_norm": 0.5724782943725586, + "learning_rate": 5.685764816826532e-06, + "loss": 0.5371, + "step": 4787 + }, + { + "epoch": 1.5319965870307168, + "grad_norm": 0.4858018457889557, + "learning_rate": 5.6839205686490474e-06, + "loss": 0.4442, + "step": 4788 + }, + { + "epoch": 1.5323165529010239, + "grad_norm": 0.4786800444126129, + "learning_rate": 5.68207622564515e-06, + "loss": 0.4393, + "step": 4789 + }, + { + "epoch": 1.5326365187713311, + "grad_norm": 0.5682559609413147, + "learning_rate": 5.680231788070566e-06, + "loss": 0.4934, + "step": 4790 + }, + { + "epoch": 1.5329564846416384, + "grad_norm": 0.5499330759048462, + "learning_rate": 5.678387256181024e-06, + "loss": 0.5285, + "step": 4791 + }, + { + "epoch": 1.5332764505119454, + "grad_norm": 0.5110037922859192, + "learning_rate": 5.6765426302322755e-06, + "loss": 0.462, + "step": 4792 + }, + { + "epoch": 1.5335964163822524, + "grad_norm": 0.5054475665092468, + "learning_rate": 5.674697910480077e-06, + "loss": 0.4424, + "step": 4793 + }, + { + "epoch": 1.5339163822525599, + "grad_norm": 0.46804389357566833, + "learning_rate": 5.6728530971801986e-06, + "loss": 0.4438, + "step": 4794 + }, + { + "epoch": 1.534236348122867, + "grad_norm": 0.5104069113731384, + "learning_rate": 5.671008190588432e-06, + "loss": 0.5263, + "step": 4795 + }, + { + "epoch": 1.534556313993174, + "grad_norm": 0.4755001664161682, + "learning_rate": 5.6691631909605705e-06, + "loss": 0.4567, + "step": 4796 + }, + { + "epoch": 1.5348762798634812, + "grad_norm": 0.6045056581497192, + "learning_rate": 5.667318098552429e-06, + "loss": 0.5137, + "step": 4797 + }, + { + "epoch": 1.5351962457337884, + "grad_norm": 0.5286843776702881, + "learning_rate": 5.66547291361983e-06, + "loss": 0.4719, + "step": 4798 + }, + { + "epoch": 1.5355162116040955, + "grad_norm": 0.5271816849708557, + "learning_rate": 5.663627636418611e-06, + "loss": 0.5051, + "step": 4799 + }, + { + "epoch": 1.5358361774744027, + "grad_norm": 0.5123486518859863, + "learning_rate": 5.661782267204622e-06, + "loss": 0.4481, + "step": 4800 + }, + { + "epoch": 1.53615614334471, + "grad_norm": 0.589840292930603, + "learning_rate": 5.659936806233723e-06, + "loss": 0.5574, + "step": 4801 + }, + { + "epoch": 1.536476109215017, + "grad_norm": 0.5192856788635254, + "learning_rate": 5.658091253761795e-06, + "loss": 0.4439, + "step": 4802 + }, + { + "epoch": 1.5367960750853242, + "grad_norm": 0.546211838722229, + "learning_rate": 5.656245610044722e-06, + "loss": 0.5547, + "step": 4803 + }, + { + "epoch": 1.5371160409556315, + "grad_norm": 0.49008166790008545, + "learning_rate": 5.654399875338406e-06, + "loss": 0.4649, + "step": 4804 + }, + { + "epoch": 1.5374360068259385, + "grad_norm": 0.5199187397956848, + "learning_rate": 5.65255404989876e-06, + "loss": 0.4941, + "step": 4805 + }, + { + "epoch": 1.5377559726962458, + "grad_norm": 0.5492860078811646, + "learning_rate": 5.650708133981708e-06, + "loss": 0.5224, + "step": 4806 + }, + { + "epoch": 1.538075938566553, + "grad_norm": 0.5844526886940002, + "learning_rate": 5.648862127843193e-06, + "loss": 0.5619, + "step": 4807 + }, + { + "epoch": 1.53839590443686, + "grad_norm": 0.5239481925964355, + "learning_rate": 5.647016031739162e-06, + "loss": 0.4662, + "step": 4808 + }, + { + "epoch": 1.5387158703071673, + "grad_norm": 0.49354660511016846, + "learning_rate": 5.6451698459255795e-06, + "loss": 0.458, + "step": 4809 + }, + { + "epoch": 1.5390358361774745, + "grad_norm": 0.5162451863288879, + "learning_rate": 5.6433235706584245e-06, + "loss": 0.4691, + "step": 4810 + }, + { + "epoch": 1.5393558020477816, + "grad_norm": 0.5093298554420471, + "learning_rate": 5.641477206193678e-06, + "loss": 0.4694, + "step": 4811 + }, + { + "epoch": 1.5396757679180886, + "grad_norm": 0.6124956607818604, + "learning_rate": 5.639630752787349e-06, + "loss": 0.5314, + "step": 4812 + }, + { + "epoch": 1.539995733788396, + "grad_norm": 0.5342500805854797, + "learning_rate": 5.637784210695445e-06, + "loss": 0.4845, + "step": 4813 + }, + { + "epoch": 1.540315699658703, + "grad_norm": 0.5140454769134521, + "learning_rate": 5.6359375801739925e-06, + "loss": 0.504, + "step": 4814 + }, + { + "epoch": 1.54063566552901, + "grad_norm": 0.5607699155807495, + "learning_rate": 5.63409086147903e-06, + "loss": 0.4849, + "step": 4815 + }, + { + "epoch": 1.5409556313993176, + "grad_norm": 0.5525639057159424, + "learning_rate": 5.632244054866607e-06, + "loss": 0.5025, + "step": 4816 + }, + { + "epoch": 1.5412755972696246, + "grad_norm": 0.5072596073150635, + "learning_rate": 5.630397160592785e-06, + "loss": 0.4693, + "step": 4817 + }, + { + "epoch": 1.5415955631399316, + "grad_norm": 0.47427666187286377, + "learning_rate": 5.628550178913639e-06, + "loss": 0.4643, + "step": 4818 + }, + { + "epoch": 1.5419155290102389, + "grad_norm": 0.5358697772026062, + "learning_rate": 5.626703110085253e-06, + "loss": 0.589, + "step": 4819 + }, + { + "epoch": 1.5422354948805461, + "grad_norm": 0.5243690609931946, + "learning_rate": 5.624855954363727e-06, + "loss": 0.4273, + "step": 4820 + }, + { + "epoch": 1.5425554607508531, + "grad_norm": 0.568854570388794, + "learning_rate": 5.6230087120051724e-06, + "loss": 0.5026, + "step": 4821 + }, + { + "epoch": 1.5428754266211604, + "grad_norm": 0.6009626984596252, + "learning_rate": 5.62116138326571e-06, + "loss": 0.4867, + "step": 4822 + }, + { + "epoch": 1.5431953924914676, + "grad_norm": 0.5987993478775024, + "learning_rate": 5.6193139684014745e-06, + "loss": 0.47, + "step": 4823 + }, + { + "epoch": 1.5435153583617747, + "grad_norm": 0.5260308980941772, + "learning_rate": 5.617466467668612e-06, + "loss": 0.5007, + "step": 4824 + }, + { + "epoch": 1.543835324232082, + "grad_norm": 0.5159297585487366, + "learning_rate": 5.6156188813232806e-06, + "loss": 0.4762, + "step": 4825 + }, + { + "epoch": 1.5441552901023892, + "grad_norm": 0.4811621606349945, + "learning_rate": 5.6137712096216515e-06, + "loss": 0.5078, + "step": 4826 + }, + { + "epoch": 1.5444752559726962, + "grad_norm": 0.4753594696521759, + "learning_rate": 5.611923452819905e-06, + "loss": 0.443, + "step": 4827 + }, + { + "epoch": 1.5447952218430034, + "grad_norm": 0.5283935070037842, + "learning_rate": 5.610075611174236e-06, + "loss": 0.4821, + "step": 4828 + }, + { + "epoch": 1.5451151877133107, + "grad_norm": 0.5373314023017883, + "learning_rate": 5.60822768494085e-06, + "loss": 0.5089, + "step": 4829 + }, + { + "epoch": 1.5454351535836177, + "grad_norm": 0.5004515647888184, + "learning_rate": 5.606379674375962e-06, + "loss": 0.4773, + "step": 4830 + }, + { + "epoch": 1.545755119453925, + "grad_norm": 0.490653395652771, + "learning_rate": 5.6045315797358046e-06, + "loss": 0.4468, + "step": 4831 + }, + { + "epoch": 1.5460750853242322, + "grad_norm": 0.5463707447052002, + "learning_rate": 5.6026834012766155e-06, + "loss": 0.5, + "step": 4832 + }, + { + "epoch": 1.5463950511945392, + "grad_norm": 0.4810127317905426, + "learning_rate": 5.600835139254649e-06, + "loss": 0.479, + "step": 4833 + }, + { + "epoch": 1.5467150170648463, + "grad_norm": 0.5389829277992249, + "learning_rate": 5.598986793926167e-06, + "loss": 0.4864, + "step": 4834 + }, + { + "epoch": 1.5470349829351537, + "grad_norm": 0.5847005844116211, + "learning_rate": 5.5971383655474445e-06, + "loss": 0.4849, + "step": 4835 + }, + { + "epoch": 1.5473549488054608, + "grad_norm": 0.5487697720527649, + "learning_rate": 5.595289854374769e-06, + "loss": 0.4645, + "step": 4836 + }, + { + "epoch": 1.5476749146757678, + "grad_norm": 0.5846667289733887, + "learning_rate": 5.5934412606644404e-06, + "loss": 0.5447, + "step": 4837 + }, + { + "epoch": 1.547994880546075, + "grad_norm": 0.5189180970191956, + "learning_rate": 5.591592584672767e-06, + "loss": 0.398, + "step": 4838 + }, + { + "epoch": 1.5483148464163823, + "grad_norm": 0.5232536196708679, + "learning_rate": 5.58974382665607e-06, + "loss": 0.4891, + "step": 4839 + }, + { + "epoch": 1.5486348122866893, + "grad_norm": 0.5266755223274231, + "learning_rate": 5.587894986870681e-06, + "loss": 0.4772, + "step": 4840 + }, + { + "epoch": 1.5489547781569966, + "grad_norm": 0.45549291372299194, + "learning_rate": 5.586046065572945e-06, + "loss": 0.424, + "step": 4841 + }, + { + "epoch": 1.5492747440273038, + "grad_norm": 0.5046857595443726, + "learning_rate": 5.5841970630192165e-06, + "loss": 0.4693, + "step": 4842 + }, + { + "epoch": 1.5495947098976108, + "grad_norm": 0.6059446930885315, + "learning_rate": 5.582347979465864e-06, + "loss": 0.5427, + "step": 4843 + }, + { + "epoch": 1.549914675767918, + "grad_norm": 0.5254539251327515, + "learning_rate": 5.580498815169263e-06, + "loss": 0.4544, + "step": 4844 + }, + { + "epoch": 1.5502346416382253, + "grad_norm": 0.49290019273757935, + "learning_rate": 5.578649570385803e-06, + "loss": 0.4965, + "step": 4845 + }, + { + "epoch": 1.5505546075085324, + "grad_norm": 0.4934821128845215, + "learning_rate": 5.576800245371883e-06, + "loss": 0.4897, + "step": 4846 + }, + { + "epoch": 1.5508745733788396, + "grad_norm": 0.5148047208786011, + "learning_rate": 5.5749508403839155e-06, + "loss": 0.4829, + "step": 4847 + }, + { + "epoch": 1.5511945392491469, + "grad_norm": 0.48020872473716736, + "learning_rate": 5.5731013556783235e-06, + "loss": 0.4699, + "step": 4848 + }, + { + "epoch": 1.5515145051194539, + "grad_norm": 0.5449939370155334, + "learning_rate": 5.571251791511537e-06, + "loss": 0.4994, + "step": 4849 + }, + { + "epoch": 1.5518344709897611, + "grad_norm": 0.5520308017730713, + "learning_rate": 5.569402148140004e-06, + "loss": 0.5094, + "step": 4850 + }, + { + "epoch": 1.5521544368600684, + "grad_norm": 0.5400788187980652, + "learning_rate": 5.567552425820177e-06, + "loss": 0.4565, + "step": 4851 + }, + { + "epoch": 1.5524744027303754, + "grad_norm": 0.6016430258750916, + "learning_rate": 5.565702624808524e-06, + "loss": 0.5279, + "step": 4852 + }, + { + "epoch": 1.5527943686006824, + "grad_norm": 0.49213698506355286, + "learning_rate": 5.563852745361523e-06, + "loss": 0.4936, + "step": 4853 + }, + { + "epoch": 1.55311433447099, + "grad_norm": 0.501552164554596, + "learning_rate": 5.562002787735658e-06, + "loss": 0.5142, + "step": 4854 + }, + { + "epoch": 1.553434300341297, + "grad_norm": 0.48596420884132385, + "learning_rate": 5.560152752187432e-06, + "loss": 0.4898, + "step": 4855 + }, + { + "epoch": 1.553754266211604, + "grad_norm": 0.4873291850090027, + "learning_rate": 5.558302638973352e-06, + "loss": 0.5229, + "step": 4856 + }, + { + "epoch": 1.5540742320819114, + "grad_norm": 0.5241739749908447, + "learning_rate": 5.55645244834994e-06, + "loss": 0.4728, + "step": 4857 + }, + { + "epoch": 1.5543941979522184, + "grad_norm": 0.5372526049613953, + "learning_rate": 5.554602180573727e-06, + "loss": 0.5489, + "step": 4858 + }, + { + "epoch": 1.5547141638225255, + "grad_norm": 0.5021660327911377, + "learning_rate": 5.5527518359012535e-06, + "loss": 0.4192, + "step": 4859 + }, + { + "epoch": 1.5550341296928327, + "grad_norm": 0.5167604684829712, + "learning_rate": 5.550901414589074e-06, + "loss": 0.4931, + "step": 4860 + }, + { + "epoch": 1.55535409556314, + "grad_norm": 0.5241142511367798, + "learning_rate": 5.54905091689375e-06, + "loss": 0.5093, + "step": 4861 + }, + { + "epoch": 1.555674061433447, + "grad_norm": 0.49749425053596497, + "learning_rate": 5.547200343071856e-06, + "loss": 0.4647, + "step": 4862 + }, + { + "epoch": 1.5559940273037542, + "grad_norm": 0.5043935179710388, + "learning_rate": 5.545349693379976e-06, + "loss": 0.4992, + "step": 4863 + }, + { + "epoch": 1.5563139931740615, + "grad_norm": 0.5102501511573792, + "learning_rate": 5.543498968074704e-06, + "loss": 0.4872, + "step": 4864 + }, + { + "epoch": 1.5566339590443685, + "grad_norm": 0.5491971969604492, + "learning_rate": 5.541648167412648e-06, + "loss": 0.5917, + "step": 4865 + }, + { + "epoch": 1.5569539249146758, + "grad_norm": 0.4569743573665619, + "learning_rate": 5.539797291650419e-06, + "loss": 0.4059, + "step": 4866 + }, + { + "epoch": 1.557273890784983, + "grad_norm": 0.48071160912513733, + "learning_rate": 5.53794634104465e-06, + "loss": 0.4606, + "step": 4867 + }, + { + "epoch": 1.55759385665529, + "grad_norm": 0.5465747117996216, + "learning_rate": 5.536095315851972e-06, + "loss": 0.5211, + "step": 4868 + }, + { + "epoch": 1.5579138225255973, + "grad_norm": 0.4880690276622772, + "learning_rate": 5.534244216329033e-06, + "loss": 0.4539, + "step": 4869 + }, + { + "epoch": 1.5582337883959045, + "grad_norm": 0.5803085565567017, + "learning_rate": 5.532393042732493e-06, + "loss": 0.5402, + "step": 4870 + }, + { + "epoch": 1.5585537542662116, + "grad_norm": 0.5311477184295654, + "learning_rate": 5.530541795319013e-06, + "loss": 0.4667, + "step": 4871 + }, + { + "epoch": 1.5588737201365188, + "grad_norm": 0.5335061550140381, + "learning_rate": 5.528690474345279e-06, + "loss": 0.5134, + "step": 4872 + }, + { + "epoch": 1.559193686006826, + "grad_norm": 0.5214332342147827, + "learning_rate": 5.5268390800679715e-06, + "loss": 0.5388, + "step": 4873 + }, + { + "epoch": 1.559513651877133, + "grad_norm": 0.5451050996780396, + "learning_rate": 5.524987612743793e-06, + "loss": 0.4695, + "step": 4874 + }, + { + "epoch": 1.5598336177474401, + "grad_norm": 0.5352328419685364, + "learning_rate": 5.5231360726294515e-06, + "loss": 0.5356, + "step": 4875 + }, + { + "epoch": 1.5601535836177476, + "grad_norm": 0.4510428309440613, + "learning_rate": 5.521284459981662e-06, + "loss": 0.4489, + "step": 4876 + }, + { + "epoch": 1.5604735494880546, + "grad_norm": 0.4923088848590851, + "learning_rate": 5.519432775057158e-06, + "loss": 0.4708, + "step": 4877 + }, + { + "epoch": 1.5607935153583616, + "grad_norm": 0.4845147430896759, + "learning_rate": 5.517581018112672e-06, + "loss": 0.4851, + "step": 4878 + }, + { + "epoch": 1.5611134812286689, + "grad_norm": 0.5657869577407837, + "learning_rate": 5.515729189404956e-06, + "loss": 0.5487, + "step": 4879 + }, + { + "epoch": 1.5614334470989761, + "grad_norm": 0.5293117165565491, + "learning_rate": 5.513877289190769e-06, + "loss": 0.4913, + "step": 4880 + }, + { + "epoch": 1.5617534129692832, + "grad_norm": 0.5773677825927734, + "learning_rate": 5.512025317726875e-06, + "loss": 0.4695, + "step": 4881 + }, + { + "epoch": 1.5620733788395904, + "grad_norm": 0.5001729726791382, + "learning_rate": 5.510173275270057e-06, + "loss": 0.5551, + "step": 4882 + }, + { + "epoch": 1.5623933447098977, + "grad_norm": 0.493765264749527, + "learning_rate": 5.508321162077098e-06, + "loss": 0.4966, + "step": 4883 + }, + { + "epoch": 1.5627133105802047, + "grad_norm": 0.46060910820961, + "learning_rate": 5.506468978404801e-06, + "loss": 0.3832, + "step": 4884 + }, + { + "epoch": 1.563033276450512, + "grad_norm": 0.5179235339164734, + "learning_rate": 5.5046167245099705e-06, + "loss": 0.5098, + "step": 4885 + }, + { + "epoch": 1.5633532423208192, + "grad_norm": 0.5208465456962585, + "learning_rate": 5.5027644006494215e-06, + "loss": 0.5181, + "step": 4886 + }, + { + "epoch": 1.5636732081911262, + "grad_norm": 0.5684758424758911, + "learning_rate": 5.500912007079987e-06, + "loss": 0.4807, + "step": 4887 + }, + { + "epoch": 1.5639931740614335, + "grad_norm": 0.5089905858039856, + "learning_rate": 5.499059544058498e-06, + "loss": 0.5215, + "step": 4888 + }, + { + "epoch": 1.5643131399317407, + "grad_norm": 0.4783798158168793, + "learning_rate": 5.497207011841803e-06, + "loss": 0.3939, + "step": 4889 + }, + { + "epoch": 1.5646331058020477, + "grad_norm": 0.5981779098510742, + "learning_rate": 5.4953544106867594e-06, + "loss": 0.5299, + "step": 4890 + }, + { + "epoch": 1.564953071672355, + "grad_norm": 0.48870930075645447, + "learning_rate": 5.493501740850228e-06, + "loss": 0.4722, + "step": 4891 + }, + { + "epoch": 1.5652730375426622, + "grad_norm": 0.5528561472892761, + "learning_rate": 5.491649002589089e-06, + "loss": 0.5449, + "step": 4892 + }, + { + "epoch": 1.5655930034129693, + "grad_norm": 0.5057819485664368, + "learning_rate": 5.489796196160223e-06, + "loss": 0.4901, + "step": 4893 + }, + { + "epoch": 1.5659129692832765, + "grad_norm": 0.5546565055847168, + "learning_rate": 5.4879433218205246e-06, + "loss": 0.4618, + "step": 4894 + }, + { + "epoch": 1.5662329351535837, + "grad_norm": 0.6191630363464355, + "learning_rate": 5.4860903798269e-06, + "loss": 0.55, + "step": 4895 + }, + { + "epoch": 1.5665529010238908, + "grad_norm": 0.5262441039085388, + "learning_rate": 5.4842373704362565e-06, + "loss": 0.4451, + "step": 4896 + }, + { + "epoch": 1.5668728668941978, + "grad_norm": 0.5814364552497864, + "learning_rate": 5.482384293905521e-06, + "loss": 0.5156, + "step": 4897 + }, + { + "epoch": 1.5671928327645053, + "grad_norm": 0.552703857421875, + "learning_rate": 5.480531150491622e-06, + "loss": 0.5127, + "step": 4898 + }, + { + "epoch": 1.5675127986348123, + "grad_norm": 0.5163992643356323, + "learning_rate": 5.478677940451501e-06, + "loss": 0.4666, + "step": 4899 + }, + { + "epoch": 1.5678327645051193, + "grad_norm": 0.5370565056800842, + "learning_rate": 5.476824664042109e-06, + "loss": 0.4769, + "step": 4900 + }, + { + "epoch": 1.5681527303754266, + "grad_norm": 0.5487856864929199, + "learning_rate": 5.4749713215204015e-06, + "loss": 0.5001, + "step": 4901 + }, + { + "epoch": 1.5684726962457338, + "grad_norm": 0.547643780708313, + "learning_rate": 5.473117913143351e-06, + "loss": 0.5623, + "step": 4902 + }, + { + "epoch": 1.5687926621160408, + "grad_norm": 0.48040950298309326, + "learning_rate": 5.471264439167932e-06, + "loss": 0.48, + "step": 4903 + }, + { + "epoch": 1.569112627986348, + "grad_norm": 0.49863874912261963, + "learning_rate": 5.4694108998511296e-06, + "loss": 0.5297, + "step": 4904 + }, + { + "epoch": 1.5694325938566553, + "grad_norm": 0.4937259554862976, + "learning_rate": 5.467557295449944e-06, + "loss": 0.4536, + "step": 4905 + }, + { + "epoch": 1.5697525597269624, + "grad_norm": 0.5208419561386108, + "learning_rate": 5.465703626221375e-06, + "loss": 0.491, + "step": 4906 + }, + { + "epoch": 1.5700725255972696, + "grad_norm": 0.5429254770278931, + "learning_rate": 5.463849892422439e-06, + "loss": 0.5094, + "step": 4907 + }, + { + "epoch": 1.5703924914675769, + "grad_norm": 0.4844743311405182, + "learning_rate": 5.461996094310157e-06, + "loss": 0.4475, + "step": 4908 + }, + { + "epoch": 1.570712457337884, + "grad_norm": 0.4967704117298126, + "learning_rate": 5.460142232141561e-06, + "loss": 0.4937, + "step": 4909 + }, + { + "epoch": 1.5710324232081911, + "grad_norm": 0.5399122834205627, + "learning_rate": 5.4582883061736905e-06, + "loss": 0.4866, + "step": 4910 + }, + { + "epoch": 1.5713523890784984, + "grad_norm": 0.5367668867111206, + "learning_rate": 5.456434316663595e-06, + "loss": 0.5012, + "step": 4911 + }, + { + "epoch": 1.5716723549488054, + "grad_norm": 0.4544510841369629, + "learning_rate": 5.454580263868333e-06, + "loss": 0.4756, + "step": 4912 + }, + { + "epoch": 1.5719923208191127, + "grad_norm": 0.49972185492515564, + "learning_rate": 5.45272614804497e-06, + "loss": 0.5177, + "step": 4913 + }, + { + "epoch": 1.57231228668942, + "grad_norm": 0.4960210919380188, + "learning_rate": 5.450871969450582e-06, + "loss": 0.4862, + "step": 4914 + }, + { + "epoch": 1.572632252559727, + "grad_norm": 0.4943525493144989, + "learning_rate": 5.449017728342253e-06, + "loss": 0.461, + "step": 4915 + }, + { + "epoch": 1.572952218430034, + "grad_norm": 0.4947192072868347, + "learning_rate": 5.447163424977076e-06, + "loss": 0.4634, + "step": 4916 + }, + { + "epoch": 1.5732721843003414, + "grad_norm": 0.5367088317871094, + "learning_rate": 5.4453090596121525e-06, + "loss": 0.5331, + "step": 4917 + }, + { + "epoch": 1.5735921501706485, + "grad_norm": 0.5815829038619995, + "learning_rate": 5.443454632504592e-06, + "loss": 0.4941, + "step": 4918 + }, + { + "epoch": 1.5739121160409555, + "grad_norm": 0.624888002872467, + "learning_rate": 5.441600143911513e-06, + "loss": 0.5256, + "step": 4919 + }, + { + "epoch": 1.574232081911263, + "grad_norm": 0.49638548493385315, + "learning_rate": 5.439745594090042e-06, + "loss": 0.4947, + "step": 4920 + }, + { + "epoch": 1.57455204778157, + "grad_norm": 0.48923373222351074, + "learning_rate": 5.437890983297316e-06, + "loss": 0.4701, + "step": 4921 + }, + { + "epoch": 1.574872013651877, + "grad_norm": 0.5430967807769775, + "learning_rate": 5.436036311790479e-06, + "loss": 0.486, + "step": 4922 + }, + { + "epoch": 1.5751919795221843, + "grad_norm": 0.5682241916656494, + "learning_rate": 5.434181579826683e-06, + "loss": 0.5104, + "step": 4923 + }, + { + "epoch": 1.5755119453924915, + "grad_norm": 0.5226086378097534, + "learning_rate": 5.432326787663088e-06, + "loss": 0.492, + "step": 4924 + }, + { + "epoch": 1.5758319112627985, + "grad_norm": 0.4683378040790558, + "learning_rate": 5.4304719355568635e-06, + "loss": 0.4571, + "step": 4925 + }, + { + "epoch": 1.5761518771331058, + "grad_norm": 0.6027663946151733, + "learning_rate": 5.428617023765187e-06, + "loss": 0.5227, + "step": 4926 + }, + { + "epoch": 1.576471843003413, + "grad_norm": 0.5830893516540527, + "learning_rate": 5.426762052545245e-06, + "loss": 0.4977, + "step": 4927 + }, + { + "epoch": 1.57679180887372, + "grad_norm": 0.521967887878418, + "learning_rate": 5.424907022154231e-06, + "loss": 0.506, + "step": 4928 + }, + { + "epoch": 1.5771117747440273, + "grad_norm": 0.5393236875534058, + "learning_rate": 5.423051932849348e-06, + "loss": 0.4989, + "step": 4929 + }, + { + "epoch": 1.5774317406143346, + "grad_norm": 0.4632890820503235, + "learning_rate": 5.421196784887804e-06, + "loss": 0.4158, + "step": 4930 + }, + { + "epoch": 1.5777517064846416, + "grad_norm": 0.6038793325424194, + "learning_rate": 5.41934157852682e-06, + "loss": 0.535, + "step": 4931 + }, + { + "epoch": 1.5780716723549488, + "grad_norm": 0.5062716007232666, + "learning_rate": 5.4174863140236215e-06, + "loss": 0.4804, + "step": 4932 + }, + { + "epoch": 1.578391638225256, + "grad_norm": 0.5247980952262878, + "learning_rate": 5.415630991635443e-06, + "loss": 0.5311, + "step": 4933 + }, + { + "epoch": 1.578711604095563, + "grad_norm": 0.48763102293014526, + "learning_rate": 5.413775611619529e-06, + "loss": 0.4794, + "step": 4934 + }, + { + "epoch": 1.5790315699658704, + "grad_norm": 0.5592321157455444, + "learning_rate": 5.411920174233126e-06, + "loss": 0.5148, + "step": 4935 + }, + { + "epoch": 1.5793515358361776, + "grad_norm": 0.5766536593437195, + "learning_rate": 5.410064679733498e-06, + "loss": 0.5538, + "step": 4936 + }, + { + "epoch": 1.5796715017064846, + "grad_norm": 0.5071977376937866, + "learning_rate": 5.408209128377907e-06, + "loss": 0.4534, + "step": 4937 + }, + { + "epoch": 1.5799914675767917, + "grad_norm": 0.5171620845794678, + "learning_rate": 5.40635352042363e-06, + "loss": 0.5179, + "step": 4938 + }, + { + "epoch": 1.5803114334470991, + "grad_norm": 0.5436701774597168, + "learning_rate": 5.404497856127947e-06, + "loss": 0.5014, + "step": 4939 + }, + { + "epoch": 1.5806313993174061, + "grad_norm": 0.494022399187088, + "learning_rate": 5.402642135748152e-06, + "loss": 0.467, + "step": 4940 + }, + { + "epoch": 1.5809513651877132, + "grad_norm": 0.5031909942626953, + "learning_rate": 5.40078635954154e-06, + "loss": 0.4756, + "step": 4941 + }, + { + "epoch": 1.5812713310580204, + "grad_norm": 0.5093727707862854, + "learning_rate": 5.398930527765416e-06, + "loss": 0.4848, + "step": 4942 + }, + { + "epoch": 1.5815912969283277, + "grad_norm": 0.5057315826416016, + "learning_rate": 5.397074640677094e-06, + "loss": 0.4867, + "step": 4943 + }, + { + "epoch": 1.5819112627986347, + "grad_norm": 0.49545595049858093, + "learning_rate": 5.395218698533897e-06, + "loss": 0.5062, + "step": 4944 + }, + { + "epoch": 1.582231228668942, + "grad_norm": 0.5084137916564941, + "learning_rate": 5.393362701593153e-06, + "loss": 0.5009, + "step": 4945 + }, + { + "epoch": 1.5825511945392492, + "grad_norm": 0.5302429795265198, + "learning_rate": 5.391506650112195e-06, + "loss": 0.471, + "step": 4946 + }, + { + "epoch": 1.5828711604095562, + "grad_norm": 0.5519916415214539, + "learning_rate": 5.389650544348369e-06, + "loss": 0.5387, + "step": 4947 + }, + { + "epoch": 1.5831911262798635, + "grad_norm": 0.5033243298530579, + "learning_rate": 5.387794384559027e-06, + "loss": 0.4859, + "step": 4948 + }, + { + "epoch": 1.5835110921501707, + "grad_norm": 0.462449848651886, + "learning_rate": 5.3859381710015256e-06, + "loss": 0.435, + "step": 4949 + }, + { + "epoch": 1.5838310580204777, + "grad_norm": 0.5657418966293335, + "learning_rate": 5.384081903933235e-06, + "loss": 0.5097, + "step": 4950 + }, + { + "epoch": 1.584151023890785, + "grad_norm": 0.5815718173980713, + "learning_rate": 5.382225583611525e-06, + "loss": 0.5071, + "step": 4951 + }, + { + "epoch": 1.5844709897610922, + "grad_norm": 0.5068009495735168, + "learning_rate": 5.380369210293777e-06, + "loss": 0.5294, + "step": 4952 + }, + { + "epoch": 1.5847909556313993, + "grad_norm": 0.4753865897655487, + "learning_rate": 5.378512784237382e-06, + "loss": 0.4522, + "step": 4953 + }, + { + "epoch": 1.5851109215017065, + "grad_norm": 0.5203690528869629, + "learning_rate": 5.376656305699733e-06, + "loss": 0.4568, + "step": 4954 + }, + { + "epoch": 1.5854308873720138, + "grad_norm": 0.5779047012329102, + "learning_rate": 5.374799774938236e-06, + "loss": 0.5153, + "step": 4955 + }, + { + "epoch": 1.5857508532423208, + "grad_norm": 0.5613170862197876, + "learning_rate": 5.372943192210297e-06, + "loss": 0.5389, + "step": 4956 + }, + { + "epoch": 1.586070819112628, + "grad_norm": 0.5037094950675964, + "learning_rate": 5.371086557773337e-06, + "loss": 0.4675, + "step": 4957 + }, + { + "epoch": 1.5863907849829353, + "grad_norm": 0.5480237007141113, + "learning_rate": 5.369229871884779e-06, + "loss": 0.5099, + "step": 4958 + }, + { + "epoch": 1.5867107508532423, + "grad_norm": 0.4700008034706116, + "learning_rate": 5.367373134802055e-06, + "loss": 0.4598, + "step": 4959 + }, + { + "epoch": 1.5870307167235493, + "grad_norm": 0.5223121047019958, + "learning_rate": 5.365516346782605e-06, + "loss": 0.5131, + "step": 4960 + }, + { + "epoch": 1.5873506825938568, + "grad_norm": 0.48291903734207153, + "learning_rate": 5.363659508083872e-06, + "loss": 0.4667, + "step": 4961 + }, + { + "epoch": 1.5876706484641638, + "grad_norm": 0.5335783362388611, + "learning_rate": 5.361802618963312e-06, + "loss": 0.4978, + "step": 4962 + }, + { + "epoch": 1.5879906143344709, + "grad_norm": 0.5844630599021912, + "learning_rate": 5.359945679678383e-06, + "loss": 0.5726, + "step": 4963 + }, + { + "epoch": 1.588310580204778, + "grad_norm": 0.5081358551979065, + "learning_rate": 5.3580886904865525e-06, + "loss": 0.4611, + "step": 4964 + }, + { + "epoch": 1.5886305460750854, + "grad_norm": 0.5709756016731262, + "learning_rate": 5.356231651645295e-06, + "loss": 0.4868, + "step": 4965 + }, + { + "epoch": 1.5889505119453924, + "grad_norm": 0.5915698409080505, + "learning_rate": 5.354374563412089e-06, + "loss": 0.5062, + "step": 4966 + }, + { + "epoch": 1.5892704778156996, + "grad_norm": 0.576073169708252, + "learning_rate": 5.3525174260444255e-06, + "loss": 0.4828, + "step": 4967 + }, + { + "epoch": 1.5895904436860069, + "grad_norm": 0.5176089406013489, + "learning_rate": 5.350660239799795e-06, + "loss": 0.484, + "step": 4968 + }, + { + "epoch": 1.589910409556314, + "grad_norm": 0.5273805260658264, + "learning_rate": 5.348803004935701e-06, + "loss": 0.4837, + "step": 4969 + }, + { + "epoch": 1.5902303754266212, + "grad_norm": 0.5675475597381592, + "learning_rate": 5.346945721709652e-06, + "loss": 0.5104, + "step": 4970 + }, + { + "epoch": 1.5905503412969284, + "grad_norm": 0.5363828539848328, + "learning_rate": 5.3450883903791585e-06, + "loss": 0.4595, + "step": 4971 + }, + { + "epoch": 1.5908703071672354, + "grad_norm": 0.525600790977478, + "learning_rate": 5.343231011201747e-06, + "loss": 0.4865, + "step": 4972 + }, + { + "epoch": 1.5911902730375427, + "grad_norm": 0.5501222014427185, + "learning_rate": 5.341373584434941e-06, + "loss": 0.4936, + "step": 4973 + }, + { + "epoch": 1.59151023890785, + "grad_norm": 0.4852727949619293, + "learning_rate": 5.339516110336278e-06, + "loss": 0.427, + "step": 4974 + }, + { + "epoch": 1.591830204778157, + "grad_norm": 0.5601401925086975, + "learning_rate": 5.3376585891633e-06, + "loss": 0.5627, + "step": 4975 + }, + { + "epoch": 1.5921501706484642, + "grad_norm": 0.4850602447986603, + "learning_rate": 5.335801021173548e-06, + "loss": 0.4997, + "step": 4976 + }, + { + "epoch": 1.5924701365187715, + "grad_norm": 0.5282145738601685, + "learning_rate": 5.3339434066245835e-06, + "loss": 0.482, + "step": 4977 + }, + { + "epoch": 1.5927901023890785, + "grad_norm": 0.5373655557632446, + "learning_rate": 5.3320857457739615e-06, + "loss": 0.5199, + "step": 4978 + }, + { + "epoch": 1.5931100682593855, + "grad_norm": 0.5853095054626465, + "learning_rate": 5.330228038879253e-06, + "loss": 0.5109, + "step": 4979 + }, + { + "epoch": 1.593430034129693, + "grad_norm": 0.5338366031646729, + "learning_rate": 5.32837028619803e-06, + "loss": 0.5002, + "step": 4980 + }, + { + "epoch": 1.59375, + "grad_norm": 0.5058368444442749, + "learning_rate": 5.326512487987871e-06, + "loss": 0.5277, + "step": 4981 + }, + { + "epoch": 1.594069965870307, + "grad_norm": 0.5702630281448364, + "learning_rate": 5.324654644506363e-06, + "loss": 0.5024, + "step": 4982 + }, + { + "epoch": 1.5943899317406145, + "grad_norm": 0.46419402956962585, + "learning_rate": 5.3227967560110984e-06, + "loss": 0.4306, + "step": 4983 + }, + { + "epoch": 1.5947098976109215, + "grad_norm": 0.45496001839637756, + "learning_rate": 5.320938822759674e-06, + "loss": 0.5151, + "step": 4984 + }, + { + "epoch": 1.5950298634812285, + "grad_norm": 0.5695767998695374, + "learning_rate": 5.319080845009697e-06, + "loss": 0.5443, + "step": 4985 + }, + { + "epoch": 1.5953498293515358, + "grad_norm": 0.5018442273139954, + "learning_rate": 5.317222823018775e-06, + "loss": 0.4859, + "step": 4986 + }, + { + "epoch": 1.595669795221843, + "grad_norm": 0.5421056151390076, + "learning_rate": 5.31536475704453e-06, + "loss": 0.4921, + "step": 4987 + }, + { + "epoch": 1.59598976109215, + "grad_norm": 0.4972304105758667, + "learning_rate": 5.31350664734458e-06, + "loss": 0.5217, + "step": 4988 + }, + { + "epoch": 1.5963097269624573, + "grad_norm": 0.5531447529792786, + "learning_rate": 5.3116484941765564e-06, + "loss": 0.4992, + "step": 4989 + }, + { + "epoch": 1.5966296928327646, + "grad_norm": 0.557141125202179, + "learning_rate": 5.309790297798094e-06, + "loss": 0.4901, + "step": 4990 + }, + { + "epoch": 1.5969496587030716, + "grad_norm": 0.600490391254425, + "learning_rate": 5.307932058466835e-06, + "loss": 0.5142, + "step": 4991 + }, + { + "epoch": 1.5972696245733788, + "grad_norm": 0.5329923629760742, + "learning_rate": 5.3060737764404254e-06, + "loss": 0.5022, + "step": 4992 + }, + { + "epoch": 1.597589590443686, + "grad_norm": 0.4775228500366211, + "learning_rate": 5.304215451976517e-06, + "loss": 0.4576, + "step": 4993 + }, + { + "epoch": 1.5979095563139931, + "grad_norm": 0.5137864351272583, + "learning_rate": 5.3023570853327725e-06, + "loss": 0.5079, + "step": 4994 + }, + { + "epoch": 1.5982295221843004, + "grad_norm": 0.48908084630966187, + "learning_rate": 5.300498676766853e-06, + "loss": 0.4567, + "step": 4995 + }, + { + "epoch": 1.5985494880546076, + "grad_norm": 0.5525791645050049, + "learning_rate": 5.29864022653643e-06, + "loss": 0.5698, + "step": 4996 + }, + { + "epoch": 1.5988694539249146, + "grad_norm": 0.5054503679275513, + "learning_rate": 5.296781734899182e-06, + "loss": 0.4754, + "step": 4997 + }, + { + "epoch": 1.599189419795222, + "grad_norm": 0.5379093885421753, + "learning_rate": 5.294923202112788e-06, + "loss": 0.506, + "step": 4998 + }, + { + "epoch": 1.5995093856655291, + "grad_norm": 0.4838557839393616, + "learning_rate": 5.293064628434939e-06, + "loss": 0.4752, + "step": 4999 + }, + { + "epoch": 1.5998293515358362, + "grad_norm": 0.4621947407722473, + "learning_rate": 5.291206014123325e-06, + "loss": 0.4347, + "step": 5000 + }, + { + "epoch": 1.6001493174061432, + "grad_norm": 0.5238200426101685, + "learning_rate": 5.289347359435646e-06, + "loss": 0.4992, + "step": 5001 + }, + { + "epoch": 1.6004692832764507, + "grad_norm": 0.5577722787857056, + "learning_rate": 5.28748866462961e-06, + "loss": 0.5308, + "step": 5002 + }, + { + "epoch": 1.6007892491467577, + "grad_norm": 0.5425005555152893, + "learning_rate": 5.2856299299629235e-06, + "loss": 0.5073, + "step": 5003 + }, + { + "epoch": 1.6011092150170647, + "grad_norm": 0.49496665596961975, + "learning_rate": 5.283771155693304e-06, + "loss": 0.4808, + "step": 5004 + }, + { + "epoch": 1.601429180887372, + "grad_norm": 0.5529054403305054, + "learning_rate": 5.281912342078471e-06, + "loss": 0.4747, + "step": 5005 + }, + { + "epoch": 1.6017491467576792, + "grad_norm": 0.5010500550270081, + "learning_rate": 5.280053489376154e-06, + "loss": 0.425, + "step": 5006 + }, + { + "epoch": 1.6020691126279862, + "grad_norm": 0.5153359770774841, + "learning_rate": 5.278194597844083e-06, + "loss": 0.5108, + "step": 5007 + }, + { + "epoch": 1.6023890784982935, + "grad_norm": 0.5091795325279236, + "learning_rate": 5.276335667739998e-06, + "loss": 0.514, + "step": 5008 + }, + { + "epoch": 1.6027090443686007, + "grad_norm": 0.5603525638580322, + "learning_rate": 5.274476699321638e-06, + "loss": 0.5229, + "step": 5009 + }, + { + "epoch": 1.6030290102389078, + "grad_norm": 0.5608372092247009, + "learning_rate": 5.272617692846755e-06, + "loss": 0.5504, + "step": 5010 + }, + { + "epoch": 1.603348976109215, + "grad_norm": 0.556414008140564, + "learning_rate": 5.2707586485731e-06, + "loss": 0.4701, + "step": 5011 + }, + { + "epoch": 1.6036689419795223, + "grad_norm": 0.4858776330947876, + "learning_rate": 5.268899566758432e-06, + "loss": 0.4416, + "step": 5012 + }, + { + "epoch": 1.6039889078498293, + "grad_norm": 0.5496314764022827, + "learning_rate": 5.267040447660517e-06, + "loss": 0.5484, + "step": 5013 + }, + { + "epoch": 1.6043088737201365, + "grad_norm": 0.5210900902748108, + "learning_rate": 5.265181291537122e-06, + "loss": 0.5136, + "step": 5014 + }, + { + "epoch": 1.6046288395904438, + "grad_norm": 0.5538973212242126, + "learning_rate": 5.263322098646023e-06, + "loss": 0.4832, + "step": 5015 + }, + { + "epoch": 1.6049488054607508, + "grad_norm": 0.5693376064300537, + "learning_rate": 5.261462869244997e-06, + "loss": 0.5117, + "step": 5016 + }, + { + "epoch": 1.605268771331058, + "grad_norm": 0.5388249754905701, + "learning_rate": 5.25960360359183e-06, + "loss": 0.5164, + "step": 5017 + }, + { + "epoch": 1.6055887372013653, + "grad_norm": 0.5378933548927307, + "learning_rate": 5.257744301944311e-06, + "loss": 0.4877, + "step": 5018 + }, + { + "epoch": 1.6059087030716723, + "grad_norm": 0.533751904964447, + "learning_rate": 5.255884964560235e-06, + "loss": 0.4989, + "step": 5019 + }, + { + "epoch": 1.6062286689419796, + "grad_norm": 0.4835195243358612, + "learning_rate": 5.2540255916974005e-06, + "loss": 0.4758, + "step": 5020 + }, + { + "epoch": 1.6065486348122868, + "grad_norm": 0.5001780390739441, + "learning_rate": 5.252166183613612e-06, + "loss": 0.511, + "step": 5021 + }, + { + "epoch": 1.6068686006825939, + "grad_norm": 0.5064964890480042, + "learning_rate": 5.2503067405666776e-06, + "loss": 0.5093, + "step": 5022 + }, + { + "epoch": 1.6071885665529009, + "grad_norm": 0.5077722668647766, + "learning_rate": 5.248447262814413e-06, + "loss": 0.4866, + "step": 5023 + }, + { + "epoch": 1.6075085324232083, + "grad_norm": 0.5091267824172974, + "learning_rate": 5.246587750614638e-06, + "loss": 0.4883, + "step": 5024 + }, + { + "epoch": 1.6078284982935154, + "grad_norm": 0.4769721031188965, + "learning_rate": 5.244728204225175e-06, + "loss": 0.4887, + "step": 5025 + }, + { + "epoch": 1.6081484641638224, + "grad_norm": 0.5393322706222534, + "learning_rate": 5.242868623903851e-06, + "loss": 0.4896, + "step": 5026 + }, + { + "epoch": 1.6084684300341296, + "grad_norm": 0.539297878742218, + "learning_rate": 5.241009009908499e-06, + "loss": 0.5276, + "step": 5027 + }, + { + "epoch": 1.608788395904437, + "grad_norm": 0.4828234910964966, + "learning_rate": 5.239149362496959e-06, + "loss": 0.5081, + "step": 5028 + }, + { + "epoch": 1.609108361774744, + "grad_norm": 0.47531136870384216, + "learning_rate": 5.237289681927072e-06, + "loss": 0.4619, + "step": 5029 + }, + { + "epoch": 1.6094283276450512, + "grad_norm": 0.6127607226371765, + "learning_rate": 5.235429968456686e-06, + "loss": 0.5101, + "step": 5030 + }, + { + "epoch": 1.6097482935153584, + "grad_norm": 0.5446188449859619, + "learning_rate": 5.2335702223436505e-06, + "loss": 0.5345, + "step": 5031 + }, + { + "epoch": 1.6100682593856654, + "grad_norm": 0.5053405165672302, + "learning_rate": 5.2317104438458235e-06, + "loss": 0.4653, + "step": 5032 + }, + { + "epoch": 1.6103882252559727, + "grad_norm": 0.5763628482818604, + "learning_rate": 5.229850633221063e-06, + "loss": 0.495, + "step": 5033 + }, + { + "epoch": 1.61070819112628, + "grad_norm": 0.5162618160247803, + "learning_rate": 5.2279907907272375e-06, + "loss": 0.5125, + "step": 5034 + }, + { + "epoch": 1.611028156996587, + "grad_norm": 0.43388307094573975, + "learning_rate": 5.226130916622214e-06, + "loss": 0.4038, + "step": 5035 + }, + { + "epoch": 1.6113481228668942, + "grad_norm": 0.5374667048454285, + "learning_rate": 5.224271011163867e-06, + "loss": 0.5445, + "step": 5036 + }, + { + "epoch": 1.6116680887372015, + "grad_norm": 0.5645819902420044, + "learning_rate": 5.222411074610074e-06, + "loss": 0.5007, + "step": 5037 + }, + { + "epoch": 1.6119880546075085, + "grad_norm": 0.5632933378219604, + "learning_rate": 5.2205511072187185e-06, + "loss": 0.5081, + "step": 5038 + }, + { + "epoch": 1.6123080204778157, + "grad_norm": 0.5653529167175293, + "learning_rate": 5.218691109247686e-06, + "loss": 0.5217, + "step": 5039 + }, + { + "epoch": 1.612627986348123, + "grad_norm": 0.5255849361419678, + "learning_rate": 5.216831080954869e-06, + "loss": 0.4302, + "step": 5040 + }, + { + "epoch": 1.61294795221843, + "grad_norm": 0.6479395627975464, + "learning_rate": 5.214971022598162e-06, + "loss": 0.571, + "step": 5041 + }, + { + "epoch": 1.613267918088737, + "grad_norm": 0.5265260934829712, + "learning_rate": 5.213110934435463e-06, + "loss": 0.5054, + "step": 5042 + }, + { + "epoch": 1.6135878839590445, + "grad_norm": 0.5972551107406616, + "learning_rate": 5.211250816724676e-06, + "loss": 0.4689, + "step": 5043 + }, + { + "epoch": 1.6139078498293515, + "grad_norm": 0.4919164776802063, + "learning_rate": 5.2093906697237104e-06, + "loss": 0.4354, + "step": 5044 + }, + { + "epoch": 1.6142278156996586, + "grad_norm": 0.5815161466598511, + "learning_rate": 5.207530493690478e-06, + "loss": 0.5371, + "step": 5045 + }, + { + "epoch": 1.614547781569966, + "grad_norm": 0.5273357033729553, + "learning_rate": 5.205670288882889e-06, + "loss": 0.4503, + "step": 5046 + }, + { + "epoch": 1.614867747440273, + "grad_norm": 0.5358985662460327, + "learning_rate": 5.203810055558871e-06, + "loss": 0.5407, + "step": 5047 + }, + { + "epoch": 1.61518771331058, + "grad_norm": 0.5318018198013306, + "learning_rate": 5.201949793976342e-06, + "loss": 0.4957, + "step": 5048 + }, + { + "epoch": 1.6155076791808873, + "grad_norm": 0.499167799949646, + "learning_rate": 5.2000895043932296e-06, + "loss": 0.4988, + "step": 5049 + }, + { + "epoch": 1.6158276450511946, + "grad_norm": 0.5340614914894104, + "learning_rate": 5.198229187067469e-06, + "loss": 0.5203, + "step": 5050 + }, + { + "epoch": 1.6161476109215016, + "grad_norm": 0.5454068183898926, + "learning_rate": 5.196368842256992e-06, + "loss": 0.5007, + "step": 5051 + }, + { + "epoch": 1.6164675767918089, + "grad_norm": 0.49962106347084045, + "learning_rate": 5.194508470219739e-06, + "loss": 0.4042, + "step": 5052 + }, + { + "epoch": 1.616787542662116, + "grad_norm": 0.540557861328125, + "learning_rate": 5.192648071213653e-06, + "loss": 0.5337, + "step": 5053 + }, + { + "epoch": 1.6171075085324231, + "grad_norm": 0.49432751536369324, + "learning_rate": 5.19078764549668e-06, + "loss": 0.4718, + "step": 5054 + }, + { + "epoch": 1.6174274744027304, + "grad_norm": 0.5162655115127563, + "learning_rate": 5.18892719332677e-06, + "loss": 0.4571, + "step": 5055 + }, + { + "epoch": 1.6177474402730376, + "grad_norm": 0.5410633683204651, + "learning_rate": 5.187066714961877e-06, + "loss": 0.5203, + "step": 5056 + }, + { + "epoch": 1.6180674061433447, + "grad_norm": 0.4666568636894226, + "learning_rate": 5.185206210659961e-06, + "loss": 0.4568, + "step": 5057 + }, + { + "epoch": 1.618387372013652, + "grad_norm": 0.5138685703277588, + "learning_rate": 5.183345680678979e-06, + "loss": 0.4949, + "step": 5058 + }, + { + "epoch": 1.6187073378839592, + "grad_norm": 0.5632014274597168, + "learning_rate": 5.181485125276898e-06, + "loss": 0.4725, + "step": 5059 + }, + { + "epoch": 1.6190273037542662, + "grad_norm": 0.5428216457366943, + "learning_rate": 5.179624544711688e-06, + "loss": 0.5151, + "step": 5060 + }, + { + "epoch": 1.6193472696245734, + "grad_norm": 0.47719550132751465, + "learning_rate": 5.177763939241315e-06, + "loss": 0.4976, + "step": 5061 + }, + { + "epoch": 1.6196672354948807, + "grad_norm": 0.5554200410842896, + "learning_rate": 5.175903309123761e-06, + "loss": 0.4945, + "step": 5062 + }, + { + "epoch": 1.6199872013651877, + "grad_norm": 0.49648791551589966, + "learning_rate": 5.174042654617001e-06, + "loss": 0.5075, + "step": 5063 + }, + { + "epoch": 1.6203071672354947, + "grad_norm": 0.49515655636787415, + "learning_rate": 5.172181975979017e-06, + "loss": 0.4577, + "step": 5064 + }, + { + "epoch": 1.6206271331058022, + "grad_norm": 0.4947766661643982, + "learning_rate": 5.170321273467797e-06, + "loss": 0.4943, + "step": 5065 + }, + { + "epoch": 1.6209470989761092, + "grad_norm": 0.49154120683670044, + "learning_rate": 5.1684605473413245e-06, + "loss": 0.4443, + "step": 5066 + }, + { + "epoch": 1.6212670648464163, + "grad_norm": 0.5223211646080017, + "learning_rate": 5.166599797857595e-06, + "loss": 0.5108, + "step": 5067 + }, + { + "epoch": 1.6215870307167235, + "grad_norm": 0.5061191320419312, + "learning_rate": 5.164739025274604e-06, + "loss": 0.4437, + "step": 5068 + }, + { + "epoch": 1.6219069965870307, + "grad_norm": 0.4762420356273651, + "learning_rate": 5.162878229850349e-06, + "loss": 0.4674, + "step": 5069 + }, + { + "epoch": 1.6222269624573378, + "grad_norm": 0.5315890312194824, + "learning_rate": 5.1610174118428325e-06, + "loss": 0.5548, + "step": 5070 + }, + { + "epoch": 1.622546928327645, + "grad_norm": 0.510001540184021, + "learning_rate": 5.159156571510056e-06, + "loss": 0.4548, + "step": 5071 + }, + { + "epoch": 1.6228668941979523, + "grad_norm": 0.48377159237861633, + "learning_rate": 5.157295709110031e-06, + "loss": 0.5017, + "step": 5072 + }, + { + "epoch": 1.6231868600682593, + "grad_norm": 0.5535182952880859, + "learning_rate": 5.155434824900764e-06, + "loss": 0.5205, + "step": 5073 + }, + { + "epoch": 1.6235068259385665, + "grad_norm": 0.48844847083091736, + "learning_rate": 5.153573919140275e-06, + "loss": 0.4537, + "step": 5074 + }, + { + "epoch": 1.6238267918088738, + "grad_norm": 0.5575363039970398, + "learning_rate": 5.151712992086576e-06, + "loss": 0.5385, + "step": 5075 + }, + { + "epoch": 1.6241467576791808, + "grad_norm": 0.4875611960887909, + "learning_rate": 5.149852043997687e-06, + "loss": 0.4025, + "step": 5076 + }, + { + "epoch": 1.624466723549488, + "grad_norm": 0.5373698472976685, + "learning_rate": 5.1479910751316344e-06, + "loss": 0.5013, + "step": 5077 + }, + { + "epoch": 1.6247866894197953, + "grad_norm": 0.5397643446922302, + "learning_rate": 5.146130085746439e-06, + "loss": 0.4812, + "step": 5078 + }, + { + "epoch": 1.6251066552901023, + "grad_norm": 0.5343145132064819, + "learning_rate": 5.144269076100131e-06, + "loss": 0.5401, + "step": 5079 + }, + { + "epoch": 1.6254266211604096, + "grad_norm": 0.5181583166122437, + "learning_rate": 5.142408046450743e-06, + "loss": 0.4916, + "step": 5080 + }, + { + "epoch": 1.6257465870307168, + "grad_norm": 0.5414060950279236, + "learning_rate": 5.1405469970563085e-06, + "loss": 0.4682, + "step": 5081 + }, + { + "epoch": 1.6260665529010239, + "grad_norm": 0.5570537447929382, + "learning_rate": 5.138685928174863e-06, + "loss": 0.493, + "step": 5082 + }, + { + "epoch": 1.6263865187713311, + "grad_norm": 0.5074068903923035, + "learning_rate": 5.136824840064447e-06, + "loss": 0.5192, + "step": 5083 + }, + { + "epoch": 1.6267064846416384, + "grad_norm": 0.5079135298728943, + "learning_rate": 5.134963732983104e-06, + "loss": 0.4646, + "step": 5084 + }, + { + "epoch": 1.6270264505119454, + "grad_norm": 0.5378671884536743, + "learning_rate": 5.133102607188875e-06, + "loss": 0.5366, + "step": 5085 + }, + { + "epoch": 1.6273464163822524, + "grad_norm": 0.45548543334007263, + "learning_rate": 5.131241462939809e-06, + "loss": 0.4907, + "step": 5086 + }, + { + "epoch": 1.6276663822525599, + "grad_norm": 0.5167208909988403, + "learning_rate": 5.129380300493958e-06, + "loss": 0.464, + "step": 5087 + }, + { + "epoch": 1.627986348122867, + "grad_norm": 0.5381355881690979, + "learning_rate": 5.127519120109371e-06, + "loss": 0.486, + "step": 5088 + }, + { + "epoch": 1.628306313993174, + "grad_norm": 0.4981718063354492, + "learning_rate": 5.125657922044106e-06, + "loss": 0.4635, + "step": 5089 + }, + { + "epoch": 1.6286262798634812, + "grad_norm": 0.5355318784713745, + "learning_rate": 5.123796706556217e-06, + "loss": 0.4921, + "step": 5090 + }, + { + "epoch": 1.6289462457337884, + "grad_norm": 0.5649579167366028, + "learning_rate": 5.121935473903768e-06, + "loss": 0.5134, + "step": 5091 + }, + { + "epoch": 1.6292662116040955, + "grad_norm": 0.5361959338188171, + "learning_rate": 5.120074224344819e-06, + "loss": 0.4934, + "step": 5092 + }, + { + "epoch": 1.6295861774744027, + "grad_norm": 0.510814905166626, + "learning_rate": 5.118212958137432e-06, + "loss": 0.4841, + "step": 5093 + }, + { + "epoch": 1.62990614334471, + "grad_norm": 0.5256816148757935, + "learning_rate": 5.116351675539678e-06, + "loss": 0.4785, + "step": 5094 + }, + { + "epoch": 1.630226109215017, + "grad_norm": 0.5437941551208496, + "learning_rate": 5.1144903768096235e-06, + "loss": 0.4538, + "step": 5095 + }, + { + "epoch": 1.6305460750853242, + "grad_norm": 0.5067790150642395, + "learning_rate": 5.112629062205341e-06, + "loss": 0.5508, + "step": 5096 + }, + { + "epoch": 1.6308660409556315, + "grad_norm": 0.5657281875610352, + "learning_rate": 5.110767731984902e-06, + "loss": 0.4965, + "step": 5097 + }, + { + "epoch": 1.6311860068259385, + "grad_norm": 0.48917561769485474, + "learning_rate": 5.108906386406385e-06, + "loss": 0.444, + "step": 5098 + }, + { + "epoch": 1.6315059726962458, + "grad_norm": 0.5413578748703003, + "learning_rate": 5.107045025727867e-06, + "loss": 0.529, + "step": 5099 + }, + { + "epoch": 1.631825938566553, + "grad_norm": 0.5355716943740845, + "learning_rate": 5.105183650207427e-06, + "loss": 0.4952, + "step": 5100 + }, + { + "epoch": 1.63214590443686, + "grad_norm": 0.5102506279945374, + "learning_rate": 5.103322260103147e-06, + "loss": 0.4945, + "step": 5101 + }, + { + "epoch": 1.6324658703071673, + "grad_norm": 0.49773868918418884, + "learning_rate": 5.101460855673112e-06, + "loss": 0.4505, + "step": 5102 + }, + { + "epoch": 1.6327858361774745, + "grad_norm": 0.5560070872306824, + "learning_rate": 5.099599437175405e-06, + "loss": 0.5011, + "step": 5103 + }, + { + "epoch": 1.6331058020477816, + "grad_norm": 0.556323230266571, + "learning_rate": 5.097738004868119e-06, + "loss": 0.4982, + "step": 5104 + }, + { + "epoch": 1.6334257679180886, + "grad_norm": 0.5519665479660034, + "learning_rate": 5.095876559009341e-06, + "loss": 0.5599, + "step": 5105 + }, + { + "epoch": 1.633745733788396, + "grad_norm": 0.501521110534668, + "learning_rate": 5.0940150998571614e-06, + "loss": 0.5081, + "step": 5106 + }, + { + "epoch": 1.634065699658703, + "grad_norm": 0.537854015827179, + "learning_rate": 5.092153627669675e-06, + "loss": 0.5022, + "step": 5107 + }, + { + "epoch": 1.63438566552901, + "grad_norm": 0.5112746953964233, + "learning_rate": 5.090292142704978e-06, + "loss": 0.53, + "step": 5108 + }, + { + "epoch": 1.6347056313993176, + "grad_norm": 0.429195761680603, + "learning_rate": 5.0884306452211675e-06, + "loss": 0.4135, + "step": 5109 + }, + { + "epoch": 1.6350255972696246, + "grad_norm": 0.5935645699501038, + "learning_rate": 5.08656913547634e-06, + "loss": 0.5464, + "step": 5110 + }, + { + "epoch": 1.6353455631399316, + "grad_norm": 0.5522560477256775, + "learning_rate": 5.084707613728598e-06, + "loss": 0.4994, + "step": 5111 + }, + { + "epoch": 1.6356655290102389, + "grad_norm": 0.49781811237335205, + "learning_rate": 5.082846080236044e-06, + "loss": 0.4707, + "step": 5112 + }, + { + "epoch": 1.6359854948805461, + "grad_norm": 0.5011031627655029, + "learning_rate": 5.080984535256781e-06, + "loss": 0.5452, + "step": 5113 + }, + { + "epoch": 1.6363054607508531, + "grad_norm": 0.47108280658721924, + "learning_rate": 5.0791229790489134e-06, + "loss": 0.4407, + "step": 5114 + }, + { + "epoch": 1.6366254266211604, + "grad_norm": 0.5003146529197693, + "learning_rate": 5.077261411870553e-06, + "loss": 0.5277, + "step": 5115 + }, + { + "epoch": 1.6369453924914676, + "grad_norm": 0.4633661210536957, + "learning_rate": 5.075399833979803e-06, + "loss": 0.4814, + "step": 5116 + }, + { + "epoch": 1.6372653583617747, + "grad_norm": 0.47853884100914, + "learning_rate": 5.0735382456347745e-06, + "loss": 0.462, + "step": 5117 + }, + { + "epoch": 1.637585324232082, + "grad_norm": 0.5123675465583801, + "learning_rate": 5.071676647093581e-06, + "loss": 0.4885, + "step": 5118 + }, + { + "epoch": 1.6379052901023892, + "grad_norm": 0.547149121761322, + "learning_rate": 5.069815038614333e-06, + "loss": 0.535, + "step": 5119 + }, + { + "epoch": 1.6382252559726962, + "grad_norm": 0.5156007409095764, + "learning_rate": 5.067953420455148e-06, + "loss": 0.4916, + "step": 5120 + }, + { + "epoch": 1.6385452218430034, + "grad_norm": 0.5121249556541443, + "learning_rate": 5.066091792874138e-06, + "loss": 0.4676, + "step": 5121 + }, + { + "epoch": 1.6388651877133107, + "grad_norm": 0.5301448702812195, + "learning_rate": 5.064230156129422e-06, + "loss": 0.4829, + "step": 5122 + }, + { + "epoch": 1.6391851535836177, + "grad_norm": 0.5886031985282898, + "learning_rate": 5.062368510479116e-06, + "loss": 0.4933, + "step": 5123 + }, + { + "epoch": 1.639505119453925, + "grad_norm": 0.4933142066001892, + "learning_rate": 5.060506856181342e-06, + "loss": 0.5095, + "step": 5124 + }, + { + "epoch": 1.6398250853242322, + "grad_norm": 0.4435693919658661, + "learning_rate": 5.05864519349422e-06, + "loss": 0.4523, + "step": 5125 + }, + { + "epoch": 1.6401450511945392, + "grad_norm": 0.5381711721420288, + "learning_rate": 5.056783522675871e-06, + "loss": 0.5406, + "step": 5126 + }, + { + "epoch": 1.6404650170648463, + "grad_norm": 0.4852531850337982, + "learning_rate": 5.0549218439844185e-06, + "loss": 0.4794, + "step": 5127 + }, + { + "epoch": 1.6407849829351537, + "grad_norm": 0.4958217740058899, + "learning_rate": 5.053060157677986e-06, + "loss": 0.4866, + "step": 5128 + }, + { + "epoch": 1.6411049488054608, + "grad_norm": 0.543684184551239, + "learning_rate": 5.051198464014698e-06, + "loss": 0.5488, + "step": 5129 + }, + { + "epoch": 1.6414249146757678, + "grad_norm": 0.5541886687278748, + "learning_rate": 5.049336763252683e-06, + "loss": 0.4791, + "step": 5130 + }, + { + "epoch": 1.641744880546075, + "grad_norm": 0.4899916648864746, + "learning_rate": 5.047475055650065e-06, + "loss": 0.4622, + "step": 5131 + }, + { + "epoch": 1.6420648464163823, + "grad_norm": 0.4611685276031494, + "learning_rate": 5.045613341464976e-06, + "loss": 0.4939, + "step": 5132 + }, + { + "epoch": 1.6423848122866893, + "grad_norm": 0.5908498167991638, + "learning_rate": 5.043751620955539e-06, + "loss": 0.558, + "step": 5133 + }, + { + "epoch": 1.6427047781569966, + "grad_norm": 0.4951009154319763, + "learning_rate": 5.041889894379889e-06, + "loss": 0.4257, + "step": 5134 + }, + { + "epoch": 1.6430247440273038, + "grad_norm": 0.5275335907936096, + "learning_rate": 5.040028161996156e-06, + "loss": 0.5045, + "step": 5135 + }, + { + "epoch": 1.6433447098976108, + "grad_norm": 0.5004864931106567, + "learning_rate": 5.038166424062466e-06, + "loss": 0.4902, + "step": 5136 + }, + { + "epoch": 1.643664675767918, + "grad_norm": 0.4917805790901184, + "learning_rate": 5.036304680836959e-06, + "loss": 0.4812, + "step": 5137 + }, + { + "epoch": 1.6439846416382253, + "grad_norm": 0.5122542381286621, + "learning_rate": 5.034442932577764e-06, + "loss": 0.5423, + "step": 5138 + }, + { + "epoch": 1.6443046075085324, + "grad_norm": 0.5459406971931458, + "learning_rate": 5.032581179543014e-06, + "loss": 0.4673, + "step": 5139 + }, + { + "epoch": 1.6446245733788396, + "grad_norm": 0.5206505656242371, + "learning_rate": 5.030719421990845e-06, + "loss": 0.5056, + "step": 5140 + }, + { + "epoch": 1.6449445392491469, + "grad_norm": 0.5480968356132507, + "learning_rate": 5.028857660179389e-06, + "loss": 0.5183, + "step": 5141 + }, + { + "epoch": 1.6452645051194539, + "grad_norm": 0.6217779517173767, + "learning_rate": 5.0269958943667865e-06, + "loss": 0.5225, + "step": 5142 + }, + { + "epoch": 1.6455844709897611, + "grad_norm": 0.4600312411785126, + "learning_rate": 5.02513412481117e-06, + "loss": 0.4457, + "step": 5143 + }, + { + "epoch": 1.6459044368600684, + "grad_norm": 0.5307644605636597, + "learning_rate": 5.023272351770675e-06, + "loss": 0.4937, + "step": 5144 + }, + { + "epoch": 1.6462244027303754, + "grad_norm": 0.5613411664962769, + "learning_rate": 5.021410575503442e-06, + "loss": 0.546, + "step": 5145 + }, + { + "epoch": 1.6465443686006824, + "grad_norm": 0.5683817863464355, + "learning_rate": 5.019548796267607e-06, + "loss": 0.4823, + "step": 5146 + }, + { + "epoch": 1.64686433447099, + "grad_norm": 0.6270329356193542, + "learning_rate": 5.017687014321308e-06, + "loss": 0.5589, + "step": 5147 + }, + { + "epoch": 1.647184300341297, + "grad_norm": 0.4889225363731384, + "learning_rate": 5.0158252299226826e-06, + "loss": 0.4477, + "step": 5148 + }, + { + "epoch": 1.647504266211604, + "grad_norm": 0.5191217064857483, + "learning_rate": 5.0139634433298705e-06, + "loss": 0.47, + "step": 5149 + }, + { + "epoch": 1.6478242320819114, + "grad_norm": 0.6270520091056824, + "learning_rate": 5.01210165480101e-06, + "loss": 0.5093, + "step": 5150 + }, + { + "epoch": 1.6481441979522184, + "grad_norm": 0.5916458964347839, + "learning_rate": 5.010239864594241e-06, + "loss": 0.4977, + "step": 5151 + }, + { + "epoch": 1.6484641638225255, + "grad_norm": 0.5480887293815613, + "learning_rate": 5.008378072967702e-06, + "loss": 0.4963, + "step": 5152 + }, + { + "epoch": 1.6487841296928327, + "grad_norm": 0.4946962296962738, + "learning_rate": 5.006516280179532e-06, + "loss": 0.4925, + "step": 5153 + }, + { + "epoch": 1.64910409556314, + "grad_norm": 0.521058201789856, + "learning_rate": 5.004654486487873e-06, + "loss": 0.4757, + "step": 5154 + }, + { + "epoch": 1.649424061433447, + "grad_norm": 0.5836396813392639, + "learning_rate": 5.002792692150865e-06, + "loss": 0.5098, + "step": 5155 + }, + { + "epoch": 1.6497440273037542, + "grad_norm": 0.5659805536270142, + "learning_rate": 5.000930897426645e-06, + "loss": 0.5355, + "step": 5156 + }, + { + "epoch": 1.6500639931740615, + "grad_norm": 0.4996427297592163, + "learning_rate": 4.999069102573356e-06, + "loss": 0.4866, + "step": 5157 + }, + { + "epoch": 1.6503839590443685, + "grad_norm": 0.48106181621551514, + "learning_rate": 4.997207307849136e-06, + "loss": 0.4804, + "step": 5158 + }, + { + "epoch": 1.6507039249146758, + "grad_norm": 0.4977264404296875, + "learning_rate": 4.995345513512128e-06, + "loss": 0.4477, + "step": 5159 + }, + { + "epoch": 1.651023890784983, + "grad_norm": 0.5068228244781494, + "learning_rate": 4.99348371982047e-06, + "loss": 0.5494, + "step": 5160 + }, + { + "epoch": 1.65134385665529, + "grad_norm": 0.46656063199043274, + "learning_rate": 4.991621927032301e-06, + "loss": 0.4418, + "step": 5161 + }, + { + "epoch": 1.6516638225255973, + "grad_norm": 0.4818781912326813, + "learning_rate": 4.98976013540576e-06, + "loss": 0.4838, + "step": 5162 + }, + { + "epoch": 1.6519837883959045, + "grad_norm": 0.5414600372314453, + "learning_rate": 4.9878983451989904e-06, + "loss": 0.503, + "step": 5163 + }, + { + "epoch": 1.6523037542662116, + "grad_norm": 0.49550461769104004, + "learning_rate": 4.98603655667013e-06, + "loss": 0.4731, + "step": 5164 + }, + { + "epoch": 1.6526237201365188, + "grad_norm": 0.4871700406074524, + "learning_rate": 4.984174770077318e-06, + "loss": 0.4822, + "step": 5165 + }, + { + "epoch": 1.652943686006826, + "grad_norm": 0.47365880012512207, + "learning_rate": 4.982312985678694e-06, + "loss": 0.4956, + "step": 5166 + }, + { + "epoch": 1.653263651877133, + "grad_norm": 0.5273645520210266, + "learning_rate": 4.980451203732394e-06, + "loss": 0.5123, + "step": 5167 + }, + { + "epoch": 1.6535836177474401, + "grad_norm": 0.4631376564502716, + "learning_rate": 4.9785894244965585e-06, + "loss": 0.4321, + "step": 5168 + }, + { + "epoch": 1.6539035836177476, + "grad_norm": 0.5108820199966431, + "learning_rate": 4.976727648229326e-06, + "loss": 0.5083, + "step": 5169 + }, + { + "epoch": 1.6542235494880546, + "grad_norm": 0.542654812335968, + "learning_rate": 4.974865875188832e-06, + "loss": 0.4916, + "step": 5170 + }, + { + "epoch": 1.6545435153583616, + "grad_norm": 0.4643901288509369, + "learning_rate": 4.973004105633216e-06, + "loss": 0.4344, + "step": 5171 + }, + { + "epoch": 1.6548634812286689, + "grad_norm": 0.4893297255039215, + "learning_rate": 4.9711423398206115e-06, + "loss": 0.4908, + "step": 5172 + }, + { + "epoch": 1.6551834470989761, + "grad_norm": 0.4885755181312561, + "learning_rate": 4.969280578009157e-06, + "loss": 0.4866, + "step": 5173 + }, + { + "epoch": 1.6555034129692832, + "grad_norm": 0.5066401362419128, + "learning_rate": 4.967418820456989e-06, + "loss": 0.5061, + "step": 5174 + }, + { + "epoch": 1.6558233788395904, + "grad_norm": 0.5579888224601746, + "learning_rate": 4.965557067422238e-06, + "loss": 0.4974, + "step": 5175 + }, + { + "epoch": 1.6561433447098977, + "grad_norm": 0.5364543199539185, + "learning_rate": 4.963695319163041e-06, + "loss": 0.526, + "step": 5176 + }, + { + "epoch": 1.6564633105802047, + "grad_norm": 0.5210397243499756, + "learning_rate": 4.961833575937534e-06, + "loss": 0.498, + "step": 5177 + }, + { + "epoch": 1.656783276450512, + "grad_norm": 0.5044063329696655, + "learning_rate": 4.959971838003846e-06, + "loss": 0.447, + "step": 5178 + }, + { + "epoch": 1.6571032423208192, + "grad_norm": 0.5081261396408081, + "learning_rate": 4.9581101056201125e-06, + "loss": 0.4988, + "step": 5179 + }, + { + "epoch": 1.6574232081911262, + "grad_norm": 0.4985758364200592, + "learning_rate": 4.956248379044462e-06, + "loss": 0.4952, + "step": 5180 + }, + { + "epoch": 1.6577431740614335, + "grad_norm": 0.5072740912437439, + "learning_rate": 4.954386658535025e-06, + "loss": 0.4663, + "step": 5181 + }, + { + "epoch": 1.6580631399317407, + "grad_norm": 0.5251668095588684, + "learning_rate": 4.952524944349935e-06, + "loss": 0.5379, + "step": 5182 + }, + { + "epoch": 1.6583831058020477, + "grad_norm": 0.4959995746612549, + "learning_rate": 4.950663236747318e-06, + "loss": 0.5143, + "step": 5183 + }, + { + "epoch": 1.658703071672355, + "grad_norm": 0.4989515542984009, + "learning_rate": 4.948801535985303e-06, + "loss": 0.4442, + "step": 5184 + }, + { + "epoch": 1.6590230375426622, + "grad_norm": 0.5242576599121094, + "learning_rate": 4.946939842322015e-06, + "loss": 0.5116, + "step": 5185 + }, + { + "epoch": 1.6593430034129693, + "grad_norm": 0.5447114109992981, + "learning_rate": 4.945078156015582e-06, + "loss": 0.5209, + "step": 5186 + }, + { + "epoch": 1.6596629692832765, + "grad_norm": 0.4931800663471222, + "learning_rate": 4.94321647732413e-06, + "loss": 0.4093, + "step": 5187 + }, + { + "epoch": 1.6599829351535837, + "grad_norm": 0.4911198616027832, + "learning_rate": 4.941354806505781e-06, + "loss": 0.531, + "step": 5188 + }, + { + "epoch": 1.6603029010238908, + "grad_norm": 0.47494980692863464, + "learning_rate": 4.939493143818659e-06, + "loss": 0.4545, + "step": 5189 + }, + { + "epoch": 1.6606228668941978, + "grad_norm": 0.5940124988555908, + "learning_rate": 4.937631489520886e-06, + "loss": 0.544, + "step": 5190 + }, + { + "epoch": 1.6609428327645053, + "grad_norm": 0.5294722318649292, + "learning_rate": 4.935769843870579e-06, + "loss": 0.5287, + "step": 5191 + }, + { + "epoch": 1.6612627986348123, + "grad_norm": 0.4588381350040436, + "learning_rate": 4.933908207125864e-06, + "loss": 0.4448, + "step": 5192 + }, + { + "epoch": 1.6615827645051193, + "grad_norm": 0.6200984120368958, + "learning_rate": 4.932046579544853e-06, + "loss": 0.561, + "step": 5193 + }, + { + "epoch": 1.6619027303754266, + "grad_norm": 0.4721447229385376, + "learning_rate": 4.930184961385668e-06, + "loss": 0.4833, + "step": 5194 + }, + { + "epoch": 1.6622226962457338, + "grad_norm": 0.5099005699157715, + "learning_rate": 4.928323352906421e-06, + "loss": 0.4971, + "step": 5195 + }, + { + "epoch": 1.6625426621160408, + "grad_norm": 0.49571260809898376, + "learning_rate": 4.9264617543652255e-06, + "loss": 0.4256, + "step": 5196 + }, + { + "epoch": 1.662862627986348, + "grad_norm": 0.5441554188728333, + "learning_rate": 4.9246001660201986e-06, + "loss": 0.5296, + "step": 5197 + }, + { + "epoch": 1.6631825938566553, + "grad_norm": 0.5576848983764648, + "learning_rate": 4.922738588129449e-06, + "loss": 0.5088, + "step": 5198 + }, + { + "epoch": 1.6635025597269624, + "grad_norm": 0.5072476863861084, + "learning_rate": 4.920877020951087e-06, + "loss": 0.4697, + "step": 5199 + }, + { + "epoch": 1.6638225255972696, + "grad_norm": 0.5600112676620483, + "learning_rate": 4.919015464743221e-06, + "loss": 0.4881, + "step": 5200 + }, + { + "epoch": 1.6641424914675769, + "grad_norm": 0.5764448642730713, + "learning_rate": 4.917153919763956e-06, + "loss": 0.4724, + "step": 5201 + }, + { + "epoch": 1.664462457337884, + "grad_norm": 0.49314001202583313, + "learning_rate": 4.915292386271403e-06, + "loss": 0.5051, + "step": 5202 + }, + { + "epoch": 1.6647824232081911, + "grad_norm": 0.5252131819725037, + "learning_rate": 4.913430864523662e-06, + "loss": 0.5285, + "step": 5203 + }, + { + "epoch": 1.6651023890784984, + "grad_norm": 0.5187550783157349, + "learning_rate": 4.911569354778836e-06, + "loss": 0.4487, + "step": 5204 + }, + { + "epoch": 1.6654223549488054, + "grad_norm": 0.4986821711063385, + "learning_rate": 4.909707857295024e-06, + "loss": 0.4566, + "step": 5205 + }, + { + "epoch": 1.6657423208191127, + "grad_norm": 0.4897250831127167, + "learning_rate": 4.907846372330326e-06, + "loss": 0.487, + "step": 5206 + }, + { + "epoch": 1.66606228668942, + "grad_norm": 0.5298472046852112, + "learning_rate": 4.90598490014284e-06, + "loss": 0.4788, + "step": 5207 + }, + { + "epoch": 1.666382252559727, + "grad_norm": 0.5566151142120361, + "learning_rate": 4.9041234409906615e-06, + "loss": 0.5047, + "step": 5208 + }, + { + "epoch": 1.666702218430034, + "grad_norm": 0.5224797129631042, + "learning_rate": 4.902261995131883e-06, + "loss": 0.4653, + "step": 5209 + }, + { + "epoch": 1.6670221843003414, + "grad_norm": 0.5086098909378052, + "learning_rate": 4.900400562824595e-06, + "loss": 0.5158, + "step": 5210 + }, + { + "epoch": 1.6673421501706485, + "grad_norm": 0.522621750831604, + "learning_rate": 4.898539144326889e-06, + "loss": 0.5033, + "step": 5211 + }, + { + "epoch": 1.6676621160409555, + "grad_norm": 0.5157283544540405, + "learning_rate": 4.8966777398968545e-06, + "loss": 0.5352, + "step": 5212 + }, + { + "epoch": 1.667982081911263, + "grad_norm": 0.5230862498283386, + "learning_rate": 4.894816349792575e-06, + "loss": 0.4778, + "step": 5213 + }, + { + "epoch": 1.66830204778157, + "grad_norm": 0.5108460187911987, + "learning_rate": 4.892954974272135e-06, + "loss": 0.4818, + "step": 5214 + }, + { + "epoch": 1.668622013651877, + "grad_norm": 0.5653759837150574, + "learning_rate": 4.891093613593615e-06, + "loss": 0.462, + "step": 5215 + }, + { + "epoch": 1.6689419795221843, + "grad_norm": 0.5857290029525757, + "learning_rate": 4.889232268015098e-06, + "loss": 0.5238, + "step": 5216 + }, + { + "epoch": 1.6692619453924915, + "grad_norm": 0.5446109771728516, + "learning_rate": 4.887370937794661e-06, + "loss": 0.5031, + "step": 5217 + }, + { + "epoch": 1.6695819112627985, + "grad_norm": 0.4403999447822571, + "learning_rate": 4.885509623190378e-06, + "loss": 0.4663, + "step": 5218 + }, + { + "epoch": 1.6699018771331058, + "grad_norm": 0.5504063367843628, + "learning_rate": 4.883648324460325e-06, + "loss": 0.5099, + "step": 5219 + }, + { + "epoch": 1.670221843003413, + "grad_norm": 0.5568899512290955, + "learning_rate": 4.881787041862569e-06, + "loss": 0.4942, + "step": 5220 + }, + { + "epoch": 1.67054180887372, + "grad_norm": 0.45460644364356995, + "learning_rate": 4.879925775655183e-06, + "loss": 0.4857, + "step": 5221 + }, + { + "epoch": 1.6708617747440273, + "grad_norm": 0.48582178354263306, + "learning_rate": 4.878064526096233e-06, + "loss": 0.4674, + "step": 5222 + }, + { + "epoch": 1.6711817406143346, + "grad_norm": 0.5432444214820862, + "learning_rate": 4.876203293443783e-06, + "loss": 0.5085, + "step": 5223 + }, + { + "epoch": 1.6715017064846416, + "grad_norm": 0.5521763563156128, + "learning_rate": 4.874342077955896e-06, + "loss": 0.4981, + "step": 5224 + }, + { + "epoch": 1.6718216723549488, + "grad_norm": 0.5622273683547974, + "learning_rate": 4.87248087989063e-06, + "loss": 0.5365, + "step": 5225 + }, + { + "epoch": 1.672141638225256, + "grad_norm": 0.48718908429145813, + "learning_rate": 4.870619699506044e-06, + "loss": 0.5037, + "step": 5226 + }, + { + "epoch": 1.672461604095563, + "grad_norm": 0.4818768799304962, + "learning_rate": 4.868758537060192e-06, + "loss": 0.51, + "step": 5227 + }, + { + "epoch": 1.6727815699658704, + "grad_norm": 0.5165402293205261, + "learning_rate": 4.866897392811127e-06, + "loss": 0.4729, + "step": 5228 + }, + { + "epoch": 1.6731015358361776, + "grad_norm": 0.5203539729118347, + "learning_rate": 4.8650362670168995e-06, + "loss": 0.5086, + "step": 5229 + }, + { + "epoch": 1.6734215017064846, + "grad_norm": 0.5446471571922302, + "learning_rate": 4.8631751599355535e-06, + "loss": 0.4446, + "step": 5230 + }, + { + "epoch": 1.6737414675767917, + "grad_norm": 0.5485162734985352, + "learning_rate": 4.861314071825137e-06, + "loss": 0.4937, + "step": 5231 + }, + { + "epoch": 1.6740614334470991, + "grad_norm": 0.5686075091362, + "learning_rate": 4.859453002943693e-06, + "loss": 0.4762, + "step": 5232 + }, + { + "epoch": 1.6743813993174061, + "grad_norm": 0.5025482773780823, + "learning_rate": 4.857591953549258e-06, + "loss": 0.5068, + "step": 5233 + }, + { + "epoch": 1.6747013651877132, + "grad_norm": 0.5636971592903137, + "learning_rate": 4.855730923899871e-06, + "loss": 0.5251, + "step": 5234 + }, + { + "epoch": 1.6750213310580204, + "grad_norm": 0.545708954334259, + "learning_rate": 4.853869914253563e-06, + "loss": 0.515, + "step": 5235 + }, + { + "epoch": 1.6753412969283277, + "grad_norm": 0.5094165205955505, + "learning_rate": 4.852008924868368e-06, + "loss": 0.4409, + "step": 5236 + }, + { + "epoch": 1.6756612627986347, + "grad_norm": 0.5279160141944885, + "learning_rate": 4.850147956002314e-06, + "loss": 0.4593, + "step": 5237 + }, + { + "epoch": 1.675981228668942, + "grad_norm": 0.5463525056838989, + "learning_rate": 4.8482870079134266e-06, + "loss": 0.4775, + "step": 5238 + }, + { + "epoch": 1.6763011945392492, + "grad_norm": 0.5538739562034607, + "learning_rate": 4.846426080859728e-06, + "loss": 0.4966, + "step": 5239 + }, + { + "epoch": 1.6766211604095562, + "grad_norm": 0.5735815167427063, + "learning_rate": 4.844565175099236e-06, + "loss": 0.4424, + "step": 5240 + }, + { + "epoch": 1.6769411262798635, + "grad_norm": 0.601814866065979, + "learning_rate": 4.842704290889971e-06, + "loss": 0.5584, + "step": 5241 + }, + { + "epoch": 1.6772610921501707, + "grad_norm": 0.5215892195701599, + "learning_rate": 4.840843428489947e-06, + "loss": 0.415, + "step": 5242 + }, + { + "epoch": 1.6775810580204777, + "grad_norm": 0.5272678732872009, + "learning_rate": 4.838982588157171e-06, + "loss": 0.4728, + "step": 5243 + }, + { + "epoch": 1.677901023890785, + "grad_norm": 0.5651388764381409, + "learning_rate": 4.837121770149654e-06, + "loss": 0.505, + "step": 5244 + }, + { + "epoch": 1.6782209897610922, + "grad_norm": 0.49386903643608093, + "learning_rate": 4.835260974725397e-06, + "loss": 0.4553, + "step": 5245 + }, + { + "epoch": 1.6785409556313993, + "grad_norm": 0.4737684726715088, + "learning_rate": 4.8334002021424056e-06, + "loss": 0.4795, + "step": 5246 + }, + { + "epoch": 1.6788609215017065, + "grad_norm": 0.5495589375495911, + "learning_rate": 4.831539452658678e-06, + "loss": 0.5644, + "step": 5247 + }, + { + "epoch": 1.6791808873720138, + "grad_norm": 0.5037122368812561, + "learning_rate": 4.829678726532207e-06, + "loss": 0.4522, + "step": 5248 + }, + { + "epoch": 1.6795008532423208, + "grad_norm": 0.5186344981193542, + "learning_rate": 4.8278180240209835e-06, + "loss": 0.5104, + "step": 5249 + }, + { + "epoch": 1.679820819112628, + "grad_norm": 0.48219427466392517, + "learning_rate": 4.825957345383e-06, + "loss": 0.5191, + "step": 5250 + }, + { + "epoch": 1.6801407849829353, + "grad_norm": 0.45844265818595886, + "learning_rate": 4.82409669087624e-06, + "loss": 0.4388, + "step": 5251 + }, + { + "epoch": 1.6804607508532423, + "grad_norm": 0.552278995513916, + "learning_rate": 4.822236060758686e-06, + "loss": 0.5429, + "step": 5252 + }, + { + "epoch": 1.6807807167235493, + "grad_norm": 0.5578888654708862, + "learning_rate": 4.820375455288316e-06, + "loss": 0.4918, + "step": 5253 + }, + { + "epoch": 1.6811006825938568, + "grad_norm": 0.49955689907073975, + "learning_rate": 4.818514874723103e-06, + "loss": 0.4605, + "step": 5254 + }, + { + "epoch": 1.6814206484641638, + "grad_norm": 0.5871486663818359, + "learning_rate": 4.816654319321022e-06, + "loss": 0.5048, + "step": 5255 + }, + { + "epoch": 1.6817406143344709, + "grad_norm": 0.5110973119735718, + "learning_rate": 4.814793789340041e-06, + "loss": 0.4682, + "step": 5256 + }, + { + "epoch": 1.682060580204778, + "grad_norm": 0.5397192239761353, + "learning_rate": 4.812933285038125e-06, + "loss": 0.5696, + "step": 5257 + }, + { + "epoch": 1.6823805460750854, + "grad_norm": 0.54073566198349, + "learning_rate": 4.811072806673231e-06, + "loss": 0.4941, + "step": 5258 + }, + { + "epoch": 1.6827005119453924, + "grad_norm": 0.4927590787410736, + "learning_rate": 4.809212354503321e-06, + "loss": 0.4296, + "step": 5259 + }, + { + "epoch": 1.6830204778156996, + "grad_norm": 0.4850064218044281, + "learning_rate": 4.807351928786348e-06, + "loss": 0.4729, + "step": 5260 + }, + { + "epoch": 1.6833404436860069, + "grad_norm": 0.48674237728118896, + "learning_rate": 4.805491529780261e-06, + "loss": 0.4501, + "step": 5261 + }, + { + "epoch": 1.683660409556314, + "grad_norm": 0.4843261241912842, + "learning_rate": 4.803631157743009e-06, + "loss": 0.4757, + "step": 5262 + }, + { + "epoch": 1.6839803754266212, + "grad_norm": 0.48974671959877014, + "learning_rate": 4.801770812932533e-06, + "loss": 0.4378, + "step": 5263 + }, + { + "epoch": 1.6843003412969284, + "grad_norm": 0.5552740693092346, + "learning_rate": 4.7999104956067696e-06, + "loss": 0.524, + "step": 5264 + }, + { + "epoch": 1.6846203071672354, + "grad_norm": 0.5226063132286072, + "learning_rate": 4.79805020602366e-06, + "loss": 0.4503, + "step": 5265 + }, + { + "epoch": 1.6849402730375427, + "grad_norm": 0.5147300362586975, + "learning_rate": 4.796189944441131e-06, + "loss": 0.4908, + "step": 5266 + }, + { + "epoch": 1.68526023890785, + "grad_norm": 0.5338008999824524, + "learning_rate": 4.7943297111171115e-06, + "loss": 0.4623, + "step": 5267 + }, + { + "epoch": 1.685580204778157, + "grad_norm": 0.5334160923957825, + "learning_rate": 4.792469506309525e-06, + "loss": 0.5154, + "step": 5268 + }, + { + "epoch": 1.6859001706484642, + "grad_norm": 0.5443989038467407, + "learning_rate": 4.7906093302762895e-06, + "loss": 0.4864, + "step": 5269 + }, + { + "epoch": 1.6862201365187715, + "grad_norm": 0.5522932410240173, + "learning_rate": 4.7887491832753245e-06, + "loss": 0.4796, + "step": 5270 + }, + { + "epoch": 1.6865401023890785, + "grad_norm": 0.5522817373275757, + "learning_rate": 4.7868890655645385e-06, + "loss": 0.5353, + "step": 5271 + }, + { + "epoch": 1.6868600682593855, + "grad_norm": 0.5342884063720703, + "learning_rate": 4.78502897740184e-06, + "loss": 0.4993, + "step": 5272 + }, + { + "epoch": 1.687180034129693, + "grad_norm": 0.5327131152153015, + "learning_rate": 4.783168919045133e-06, + "loss": 0.4668, + "step": 5273 + }, + { + "epoch": 1.6875, + "grad_norm": 0.5152961611747742, + "learning_rate": 4.7813088907523145e-06, + "loss": 0.5055, + "step": 5274 + }, + { + "epoch": 1.687819965870307, + "grad_norm": 0.5315884947776794, + "learning_rate": 4.779448892781283e-06, + "loss": 0.5097, + "step": 5275 + }, + { + "epoch": 1.6881399317406145, + "grad_norm": 0.4565909802913666, + "learning_rate": 4.7775889253899276e-06, + "loss": 0.4346, + "step": 5276 + }, + { + "epoch": 1.6884598976109215, + "grad_norm": 0.5067062973976135, + "learning_rate": 4.7757289888361354e-06, + "loss": 0.4961, + "step": 5277 + }, + { + "epoch": 1.6887798634812285, + "grad_norm": 0.5119850635528564, + "learning_rate": 4.773869083377788e-06, + "loss": 0.5255, + "step": 5278 + }, + { + "epoch": 1.6890998293515358, + "grad_norm": 0.4938701391220093, + "learning_rate": 4.772009209272763e-06, + "loss": 0.4959, + "step": 5279 + }, + { + "epoch": 1.689419795221843, + "grad_norm": 0.5242329835891724, + "learning_rate": 4.770149366778938e-06, + "loss": 0.4476, + "step": 5280 + }, + { + "epoch": 1.68973976109215, + "grad_norm": 0.5618206858634949, + "learning_rate": 4.768289556154179e-06, + "loss": 0.4743, + "step": 5281 + }, + { + "epoch": 1.6900597269624573, + "grad_norm": 0.5645880103111267, + "learning_rate": 4.766429777656352e-06, + "loss": 0.497, + "step": 5282 + }, + { + "epoch": 1.6903796928327646, + "grad_norm": 0.46402737498283386, + "learning_rate": 4.764570031543316e-06, + "loss": 0.5025, + "step": 5283 + }, + { + "epoch": 1.6906996587030716, + "grad_norm": 0.44544854760169983, + "learning_rate": 4.762710318072929e-06, + "loss": 0.4349, + "step": 5284 + }, + { + "epoch": 1.6910196245733788, + "grad_norm": 0.4936232566833496, + "learning_rate": 4.760850637503043e-06, + "loss": 0.4617, + "step": 5285 + }, + { + "epoch": 1.691339590443686, + "grad_norm": 0.5291861295700073, + "learning_rate": 4.758990990091502e-06, + "loss": 0.5592, + "step": 5286 + }, + { + "epoch": 1.6916595563139931, + "grad_norm": 0.5856315493583679, + "learning_rate": 4.757131376096153e-06, + "loss": 0.4911, + "step": 5287 + }, + { + "epoch": 1.6919795221843004, + "grad_norm": 0.4970265030860901, + "learning_rate": 4.755271795774827e-06, + "loss": 0.4808, + "step": 5288 + }, + { + "epoch": 1.6922994880546076, + "grad_norm": 0.48993000388145447, + "learning_rate": 4.753412249385363e-06, + "loss": 0.5011, + "step": 5289 + }, + { + "epoch": 1.6926194539249146, + "grad_norm": 0.49721401929855347, + "learning_rate": 4.751552737185588e-06, + "loss": 0.4306, + "step": 5290 + }, + { + "epoch": 1.692939419795222, + "grad_norm": 0.46261000633239746, + "learning_rate": 4.749693259433323e-06, + "loss": 0.4482, + "step": 5291 + }, + { + "epoch": 1.6932593856655291, + "grad_norm": 0.5152326226234436, + "learning_rate": 4.747833816386391e-06, + "loss": 0.5524, + "step": 5292 + }, + { + "epoch": 1.6935793515358362, + "grad_norm": 0.5263875126838684, + "learning_rate": 4.745974408302602e-06, + "loss": 0.5055, + "step": 5293 + }, + { + "epoch": 1.6938993174061432, + "grad_norm": 0.5017513632774353, + "learning_rate": 4.744115035439766e-06, + "loss": 0.4955, + "step": 5294 + }, + { + "epoch": 1.6942192832764507, + "grad_norm": 0.5186927914619446, + "learning_rate": 4.742255698055691e-06, + "loss": 0.5308, + "step": 5295 + }, + { + "epoch": 1.6945392491467577, + "grad_norm": 0.5493725538253784, + "learning_rate": 4.740396396408172e-06, + "loss": 0.4872, + "step": 5296 + }, + { + "epoch": 1.6948592150170647, + "grad_norm": 0.5045214891433716, + "learning_rate": 4.738537130755006e-06, + "loss": 0.4415, + "step": 5297 + }, + { + "epoch": 1.695179180887372, + "grad_norm": 0.5636321306228638, + "learning_rate": 4.736677901353978e-06, + "loss": 0.4903, + "step": 5298 + }, + { + "epoch": 1.6954991467576792, + "grad_norm": 0.5362817645072937, + "learning_rate": 4.734818708462879e-06, + "loss": 0.4758, + "step": 5299 + }, + { + "epoch": 1.6958191126279862, + "grad_norm": 0.517052412033081, + "learning_rate": 4.732959552339485e-06, + "loss": 0.5091, + "step": 5300 + }, + { + "epoch": 1.6961390784982935, + "grad_norm": 0.5431403517723083, + "learning_rate": 4.731100433241569e-06, + "loss": 0.4677, + "step": 5301 + }, + { + "epoch": 1.6964590443686007, + "grad_norm": 0.5283934473991394, + "learning_rate": 4.729241351426903e-06, + "loss": 0.5165, + "step": 5302 + }, + { + "epoch": 1.6967790102389078, + "grad_norm": 0.5136128067970276, + "learning_rate": 4.727382307153246e-06, + "loss": 0.4849, + "step": 5303 + }, + { + "epoch": 1.697098976109215, + "grad_norm": 0.5218483209609985, + "learning_rate": 4.7255233006783626e-06, + "loss": 0.5034, + "step": 5304 + }, + { + "epoch": 1.6974189419795223, + "grad_norm": 0.52696293592453, + "learning_rate": 4.723664332260004e-06, + "loss": 0.4628, + "step": 5305 + }, + { + "epoch": 1.6977389078498293, + "grad_norm": 0.4817391335964203, + "learning_rate": 4.721805402155919e-06, + "loss": 0.4892, + "step": 5306 + }, + { + "epoch": 1.6980588737201365, + "grad_norm": 0.49527376890182495, + "learning_rate": 4.719946510623849e-06, + "loss": 0.4826, + "step": 5307 + }, + { + "epoch": 1.6983788395904438, + "grad_norm": 0.5037565231323242, + "learning_rate": 4.718087657921529e-06, + "loss": 0.4797, + "step": 5308 + }, + { + "epoch": 1.6986988054607508, + "grad_norm": 0.5579425096511841, + "learning_rate": 4.716228844306699e-06, + "loss": 0.4991, + "step": 5309 + }, + { + "epoch": 1.699018771331058, + "grad_norm": 0.5586383938789368, + "learning_rate": 4.714370070037078e-06, + "loss": 0.4815, + "step": 5310 + }, + { + "epoch": 1.6993387372013653, + "grad_norm": 0.48886433243751526, + "learning_rate": 4.712511335370393e-06, + "loss": 0.4813, + "step": 5311 + }, + { + "epoch": 1.6996587030716723, + "grad_norm": 0.5777683258056641, + "learning_rate": 4.710652640564355e-06, + "loss": 0.5003, + "step": 5312 + }, + { + "epoch": 1.6999786689419796, + "grad_norm": 0.4665423035621643, + "learning_rate": 4.708793985876676e-06, + "loss": 0.4176, + "step": 5313 + }, + { + "epoch": 1.7002986348122868, + "grad_norm": 0.514366626739502, + "learning_rate": 4.706935371565064e-06, + "loss": 0.5199, + "step": 5314 + }, + { + "epoch": 1.7006186006825939, + "grad_norm": 0.5152296423912048, + "learning_rate": 4.7050767978872124e-06, + "loss": 0.4805, + "step": 5315 + }, + { + "epoch": 1.7009385665529009, + "grad_norm": 0.5062317252159119, + "learning_rate": 4.7032182651008204e-06, + "loss": 0.4758, + "step": 5316 + }, + { + "epoch": 1.7012585324232083, + "grad_norm": 0.5014505982398987, + "learning_rate": 4.701359773463569e-06, + "loss": 0.5328, + "step": 5317 + }, + { + "epoch": 1.7015784982935154, + "grad_norm": 0.5042576193809509, + "learning_rate": 4.699501323233148e-06, + "loss": 0.4486, + "step": 5318 + }, + { + "epoch": 1.7018984641638224, + "grad_norm": 0.49599841237068176, + "learning_rate": 4.697642914667229e-06, + "loss": 0.5328, + "step": 5319 + }, + { + "epoch": 1.7022184300341296, + "grad_norm": 0.556387722492218, + "learning_rate": 4.695784548023483e-06, + "loss": 0.5067, + "step": 5320 + }, + { + "epoch": 1.702538395904437, + "grad_norm": 0.5078489780426025, + "learning_rate": 4.693926223559577e-06, + "loss": 0.4924, + "step": 5321 + }, + { + "epoch": 1.702858361774744, + "grad_norm": 0.540920078754425, + "learning_rate": 4.692067941533166e-06, + "loss": 0.5612, + "step": 5322 + }, + { + "epoch": 1.7031783276450512, + "grad_norm": 0.43708547949790955, + "learning_rate": 4.690209702201906e-06, + "loss": 0.4392, + "step": 5323 + }, + { + "epoch": 1.7034982935153584, + "grad_norm": 0.5136210918426514, + "learning_rate": 4.688351505823445e-06, + "loss": 0.4489, + "step": 5324 + }, + { + "epoch": 1.7038182593856654, + "grad_norm": 0.5493529438972473, + "learning_rate": 4.686493352655421e-06, + "loss": 0.5417, + "step": 5325 + }, + { + "epoch": 1.7041382252559727, + "grad_norm": 0.5002792477607727, + "learning_rate": 4.684635242955473e-06, + "loss": 0.5055, + "step": 5326 + }, + { + "epoch": 1.70445819112628, + "grad_norm": 0.48529189825057983, + "learning_rate": 4.682777176981225e-06, + "loss": 0.4326, + "step": 5327 + }, + { + "epoch": 1.704778156996587, + "grad_norm": 0.482902854681015, + "learning_rate": 4.6809191549903045e-06, + "loss": 0.4539, + "step": 5328 + }, + { + "epoch": 1.7050981228668942, + "grad_norm": 0.5050659775733948, + "learning_rate": 4.679061177240328e-06, + "loss": 0.5154, + "step": 5329 + }, + { + "epoch": 1.7054180887372015, + "grad_norm": 0.5438242554664612, + "learning_rate": 4.677203243988904e-06, + "loss": 0.5037, + "step": 5330 + }, + { + "epoch": 1.7057380546075085, + "grad_norm": 0.4972843825817108, + "learning_rate": 4.675345355493639e-06, + "loss": 0.4288, + "step": 5331 + }, + { + "epoch": 1.7060580204778157, + "grad_norm": 0.51446533203125, + "learning_rate": 4.67348751201213e-06, + "loss": 0.502, + "step": 5332 + }, + { + "epoch": 1.706377986348123, + "grad_norm": 0.5037707090377808, + "learning_rate": 4.671629713801971e-06, + "loss": 0.5552, + "step": 5333 + }, + { + "epoch": 1.70669795221843, + "grad_norm": 0.5276395082473755, + "learning_rate": 4.669771961120748e-06, + "loss": 0.4873, + "step": 5334 + }, + { + "epoch": 1.707017918088737, + "grad_norm": 0.5363672375679016, + "learning_rate": 4.667914254226039e-06, + "loss": 0.4185, + "step": 5335 + }, + { + "epoch": 1.7073378839590445, + "grad_norm": 0.5792633891105652, + "learning_rate": 4.66605659337542e-06, + "loss": 0.5128, + "step": 5336 + }, + { + "epoch": 1.7076578498293515, + "grad_norm": 0.5017244219779968, + "learning_rate": 4.664198978826453e-06, + "loss": 0.5228, + "step": 5337 + }, + { + "epoch": 1.7079778156996586, + "grad_norm": 0.5084531903266907, + "learning_rate": 4.662341410836703e-06, + "loss": 0.476, + "step": 5338 + }, + { + "epoch": 1.708297781569966, + "grad_norm": 0.5443561673164368, + "learning_rate": 4.660483889663723e-06, + "loss": 0.492, + "step": 5339 + }, + { + "epoch": 1.708617747440273, + "grad_norm": 0.4886680841445923, + "learning_rate": 4.6586264155650594e-06, + "loss": 0.4512, + "step": 5340 + }, + { + "epoch": 1.70893771331058, + "grad_norm": 0.49199721217155457, + "learning_rate": 4.6567689887982555e-06, + "loss": 0.4372, + "step": 5341 + }, + { + "epoch": 1.7092576791808873, + "grad_norm": 0.5576441884040833, + "learning_rate": 4.654911609620842e-06, + "loss": 0.5568, + "step": 5342 + }, + { + "epoch": 1.7095776450511946, + "grad_norm": 0.5655575394630432, + "learning_rate": 4.65305427829035e-06, + "loss": 0.5534, + "step": 5343 + }, + { + "epoch": 1.7098976109215016, + "grad_norm": 0.5352281332015991, + "learning_rate": 4.651196995064301e-06, + "loss": 0.462, + "step": 5344 + }, + { + "epoch": 1.7102175767918089, + "grad_norm": 0.5099778175354004, + "learning_rate": 4.649339760200206e-06, + "loss": 0.4943, + "step": 5345 + }, + { + "epoch": 1.710537542662116, + "grad_norm": 0.49300286173820496, + "learning_rate": 4.647482573955577e-06, + "loss": 0.4141, + "step": 5346 + }, + { + "epoch": 1.7108575085324231, + "grad_norm": 0.5802779197692871, + "learning_rate": 4.645625436587911e-06, + "loss": 0.5397, + "step": 5347 + }, + { + "epoch": 1.7111774744027304, + "grad_norm": 0.5505796074867249, + "learning_rate": 4.6437683483547055e-06, + "loss": 0.5048, + "step": 5348 + }, + { + "epoch": 1.7114974402730376, + "grad_norm": 0.5246226191520691, + "learning_rate": 4.641911309513449e-06, + "loss": 0.4853, + "step": 5349 + }, + { + "epoch": 1.7118174061433447, + "grad_norm": 0.475522518157959, + "learning_rate": 4.640054320321618e-06, + "loss": 0.4368, + "step": 5350 + }, + { + "epoch": 1.712137372013652, + "grad_norm": 0.5598135590553284, + "learning_rate": 4.6381973810366905e-06, + "loss": 0.4944, + "step": 5351 + }, + { + "epoch": 1.7124573378839592, + "grad_norm": 0.5293557643890381, + "learning_rate": 4.636340491916128e-06, + "loss": 0.4788, + "step": 5352 + }, + { + "epoch": 1.7127773037542662, + "grad_norm": 0.4790298044681549, + "learning_rate": 4.634483653217396e-06, + "loss": 0.4699, + "step": 5353 + }, + { + "epoch": 1.7130972696245734, + "grad_norm": 0.5051926970481873, + "learning_rate": 4.632626865197946e-06, + "loss": 0.4542, + "step": 5354 + }, + { + "epoch": 1.7134172354948807, + "grad_norm": 0.48448479175567627, + "learning_rate": 4.630770128115222e-06, + "loss": 0.4639, + "step": 5355 + }, + { + "epoch": 1.7137372013651877, + "grad_norm": 0.5340156555175781, + "learning_rate": 4.628913442226663e-06, + "loss": 0.5082, + "step": 5356 + }, + { + "epoch": 1.7140571672354947, + "grad_norm": 0.5201172232627869, + "learning_rate": 4.627056807789703e-06, + "loss": 0.4907, + "step": 5357 + }, + { + "epoch": 1.7143771331058022, + "grad_norm": 0.4603649973869324, + "learning_rate": 4.625200225061765e-06, + "loss": 0.4678, + "step": 5358 + }, + { + "epoch": 1.7146970989761092, + "grad_norm": 0.5088764429092407, + "learning_rate": 4.623343694300268e-06, + "loss": 0.4958, + "step": 5359 + }, + { + "epoch": 1.7150170648464163, + "grad_norm": 0.5996106863021851, + "learning_rate": 4.621487215762619e-06, + "loss": 0.5432, + "step": 5360 + }, + { + "epoch": 1.7153370307167235, + "grad_norm": 0.4862479269504547, + "learning_rate": 4.619630789706223e-06, + "loss": 0.492, + "step": 5361 + }, + { + "epoch": 1.7156569965870307, + "grad_norm": 0.5018873810768127, + "learning_rate": 4.617774416388476e-06, + "loss": 0.4865, + "step": 5362 + }, + { + "epoch": 1.7159769624573378, + "grad_norm": 0.56586754322052, + "learning_rate": 4.615918096066766e-06, + "loss": 0.5417, + "step": 5363 + }, + { + "epoch": 1.716296928327645, + "grad_norm": 0.486603707075119, + "learning_rate": 4.614061828998475e-06, + "loss": 0.4384, + "step": 5364 + }, + { + "epoch": 1.7166168941979523, + "grad_norm": 0.485688179731369, + "learning_rate": 4.612205615440975e-06, + "loss": 0.502, + "step": 5365 + }, + { + "epoch": 1.7169368600682593, + "grad_norm": 0.529779314994812, + "learning_rate": 4.610349455651632e-06, + "loss": 0.4902, + "step": 5366 + }, + { + "epoch": 1.7172568259385665, + "grad_norm": 0.4693937599658966, + "learning_rate": 4.608493349887807e-06, + "loss": 0.4407, + "step": 5367 + }, + { + "epoch": 1.7175767918088738, + "grad_norm": 0.49611836671829224, + "learning_rate": 4.606637298406849e-06, + "loss": 0.529, + "step": 5368 + }, + { + "epoch": 1.7178967576791808, + "grad_norm": 0.5174630284309387, + "learning_rate": 4.604781301466105e-06, + "loss": 0.4793, + "step": 5369 + }, + { + "epoch": 1.718216723549488, + "grad_norm": 0.4785979390144348, + "learning_rate": 4.6029253593229065e-06, + "loss": 0.4402, + "step": 5370 + }, + { + "epoch": 1.7185366894197953, + "grad_norm": 0.4945693910121918, + "learning_rate": 4.601069472234584e-06, + "loss": 0.4642, + "step": 5371 + }, + { + "epoch": 1.7188566552901023, + "grad_norm": 0.5145953893661499, + "learning_rate": 4.599213640458462e-06, + "loss": 0.4756, + "step": 5372 + }, + { + "epoch": 1.7191766211604096, + "grad_norm": 0.5472128987312317, + "learning_rate": 4.597357864251849e-06, + "loss": 0.5126, + "step": 5373 + }, + { + "epoch": 1.7194965870307168, + "grad_norm": 0.4303402304649353, + "learning_rate": 4.5955021438720544e-06, + "loss": 0.4392, + "step": 5374 + }, + { + "epoch": 1.7198165529010239, + "grad_norm": 0.544348955154419, + "learning_rate": 4.593646479576372e-06, + "loss": 0.4915, + "step": 5375 + }, + { + "epoch": 1.7201365187713311, + "grad_norm": 0.5345340967178345, + "learning_rate": 4.5917908716220934e-06, + "loss": 0.4787, + "step": 5376 + }, + { + "epoch": 1.7204564846416384, + "grad_norm": 0.5413882732391357, + "learning_rate": 4.589935320266505e-06, + "loss": 0.4689, + "step": 5377 + }, + { + "epoch": 1.7207764505119454, + "grad_norm": 0.5263100862503052, + "learning_rate": 4.588079825766875e-06, + "loss": 0.4881, + "step": 5378 + }, + { + "epoch": 1.7210964163822524, + "grad_norm": 0.5405126214027405, + "learning_rate": 4.586224388380474e-06, + "loss": 0.4922, + "step": 5379 + }, + { + "epoch": 1.7214163822525599, + "grad_norm": 0.5615954995155334, + "learning_rate": 4.584369008364559e-06, + "loss": 0.5068, + "step": 5380 + }, + { + "epoch": 1.721736348122867, + "grad_norm": 0.5604912638664246, + "learning_rate": 4.582513685976379e-06, + "loss": 0.501, + "step": 5381 + }, + { + "epoch": 1.722056313993174, + "grad_norm": 0.5413294434547424, + "learning_rate": 4.580658421473182e-06, + "loss": 0.5194, + "step": 5382 + }, + { + "epoch": 1.7223762798634812, + "grad_norm": 0.5368830561637878, + "learning_rate": 4.5788032151121976e-06, + "loss": 0.5022, + "step": 5383 + }, + { + "epoch": 1.7226962457337884, + "grad_norm": 0.5325060486793518, + "learning_rate": 4.576948067150655e-06, + "loss": 0.4813, + "step": 5384 + }, + { + "epoch": 1.7230162116040955, + "grad_norm": 0.5181230902671814, + "learning_rate": 4.575092977845771e-06, + "loss": 0.4177, + "step": 5385 + }, + { + "epoch": 1.7233361774744027, + "grad_norm": 0.5556210875511169, + "learning_rate": 4.573237947454755e-06, + "loss": 0.4621, + "step": 5386 + }, + { + "epoch": 1.72365614334471, + "grad_norm": 0.5162448287010193, + "learning_rate": 4.5713829762348145e-06, + "loss": 0.4613, + "step": 5387 + }, + { + "epoch": 1.723976109215017, + "grad_norm": 0.5468537211418152, + "learning_rate": 4.569528064443138e-06, + "loss": 0.5083, + "step": 5388 + }, + { + "epoch": 1.7242960750853242, + "grad_norm": 0.5611181259155273, + "learning_rate": 4.567673212336915e-06, + "loss": 0.5377, + "step": 5389 + }, + { + "epoch": 1.7246160409556315, + "grad_norm": 0.5240436792373657, + "learning_rate": 4.565818420173319e-06, + "loss": 0.4401, + "step": 5390 + }, + { + "epoch": 1.7249360068259385, + "grad_norm": 0.5515648126602173, + "learning_rate": 4.5639636882095215e-06, + "loss": 0.568, + "step": 5391 + }, + { + "epoch": 1.7252559726962458, + "grad_norm": 0.5274022221565247, + "learning_rate": 4.5621090167026846e-06, + "loss": 0.4569, + "step": 5392 + }, + { + "epoch": 1.725575938566553, + "grad_norm": 0.5471010804176331, + "learning_rate": 4.560254405909959e-06, + "loss": 0.4855, + "step": 5393 + }, + { + "epoch": 1.72589590443686, + "grad_norm": 0.5593867897987366, + "learning_rate": 4.558399856088489e-06, + "loss": 0.4829, + "step": 5394 + }, + { + "epoch": 1.7262158703071673, + "grad_norm": 0.5176609754562378, + "learning_rate": 4.5565453674954095e-06, + "loss": 0.4627, + "step": 5395 + }, + { + "epoch": 1.7265358361774745, + "grad_norm": 0.5561937093734741, + "learning_rate": 4.554690940387848e-06, + "loss": 0.5149, + "step": 5396 + }, + { + "epoch": 1.7268558020477816, + "grad_norm": 0.5249950289726257, + "learning_rate": 4.552836575022925e-06, + "loss": 0.5108, + "step": 5397 + }, + { + "epoch": 1.7271757679180886, + "grad_norm": 0.46012160181999207, + "learning_rate": 4.550982271657748e-06, + "loss": 0.4368, + "step": 5398 + }, + { + "epoch": 1.727495733788396, + "grad_norm": 0.4988340139389038, + "learning_rate": 4.54912803054942e-06, + "loss": 0.4959, + "step": 5399 + }, + { + "epoch": 1.727815699658703, + "grad_norm": 0.5390985608100891, + "learning_rate": 4.547273851955031e-06, + "loss": 0.4913, + "step": 5400 + }, + { + "epoch": 1.72813566552901, + "grad_norm": 0.5166676044464111, + "learning_rate": 4.545419736131668e-06, + "loss": 0.46, + "step": 5401 + }, + { + "epoch": 1.7284556313993176, + "grad_norm": 0.47667017579078674, + "learning_rate": 4.543565683336407e-06, + "loss": 0.476, + "step": 5402 + }, + { + "epoch": 1.7287755972696246, + "grad_norm": 0.46250563859939575, + "learning_rate": 4.541711693826311e-06, + "loss": 0.5012, + "step": 5403 + }, + { + "epoch": 1.7290955631399316, + "grad_norm": 0.4555782675743103, + "learning_rate": 4.539857767858442e-06, + "loss": 0.4517, + "step": 5404 + }, + { + "epoch": 1.7294155290102389, + "grad_norm": 0.49214455485343933, + "learning_rate": 4.538003905689844e-06, + "loss": 0.4959, + "step": 5405 + }, + { + "epoch": 1.7297354948805461, + "grad_norm": 0.5046555995941162, + "learning_rate": 4.536150107577562e-06, + "loss": 0.4794, + "step": 5406 + }, + { + "epoch": 1.7300554607508531, + "grad_norm": 0.4979184865951538, + "learning_rate": 4.534296373778627e-06, + "loss": 0.4775, + "step": 5407 + }, + { + "epoch": 1.7303754266211604, + "grad_norm": 0.46941980719566345, + "learning_rate": 4.532442704550058e-06, + "loss": 0.498, + "step": 5408 + }, + { + "epoch": 1.7306953924914676, + "grad_norm": 0.41967618465423584, + "learning_rate": 4.530589100148872e-06, + "loss": 0.4439, + "step": 5409 + }, + { + "epoch": 1.7310153583617747, + "grad_norm": 0.5239220857620239, + "learning_rate": 4.528735560832071e-06, + "loss": 0.5585, + "step": 5410 + }, + { + "epoch": 1.731335324232082, + "grad_norm": 0.5077087879180908, + "learning_rate": 4.52688208685665e-06, + "loss": 0.4779, + "step": 5411 + }, + { + "epoch": 1.7316552901023892, + "grad_norm": 0.5944514870643616, + "learning_rate": 4.5250286784796e-06, + "loss": 0.531, + "step": 5412 + }, + { + "epoch": 1.7319752559726962, + "grad_norm": 0.5323827266693115, + "learning_rate": 4.523175335957894e-06, + "loss": 0.4265, + "step": 5413 + }, + { + "epoch": 1.7322952218430034, + "grad_norm": 0.5563470721244812, + "learning_rate": 4.521322059548501e-06, + "loss": 0.5252, + "step": 5414 + }, + { + "epoch": 1.7326151877133107, + "grad_norm": 0.5009718537330627, + "learning_rate": 4.519468849508379e-06, + "loss": 0.5364, + "step": 5415 + }, + { + "epoch": 1.7329351535836177, + "grad_norm": 0.5379970669746399, + "learning_rate": 4.51761570609448e-06, + "loss": 0.4852, + "step": 5416 + }, + { + "epoch": 1.733255119453925, + "grad_norm": 0.5320088267326355, + "learning_rate": 4.515762629563745e-06, + "loss": 0.4876, + "step": 5417 + }, + { + "epoch": 1.7335750853242322, + "grad_norm": 0.5413721203804016, + "learning_rate": 4.513909620173103e-06, + "loss": 0.4788, + "step": 5418 + }, + { + "epoch": 1.7338950511945392, + "grad_norm": 0.5124156475067139, + "learning_rate": 4.512056678179477e-06, + "loss": 0.4605, + "step": 5419 + }, + { + "epoch": 1.7342150170648463, + "grad_norm": 0.5723578333854675, + "learning_rate": 4.510203803839779e-06, + "loss": 0.5124, + "step": 5420 + }, + { + "epoch": 1.7345349829351537, + "grad_norm": 0.528305172920227, + "learning_rate": 4.508350997410913e-06, + "loss": 0.4558, + "step": 5421 + }, + { + "epoch": 1.7348549488054608, + "grad_norm": 0.4989233613014221, + "learning_rate": 4.506498259149774e-06, + "loss": 0.4747, + "step": 5422 + }, + { + "epoch": 1.7351749146757678, + "grad_norm": 0.5517019629478455, + "learning_rate": 4.504645589313243e-06, + "loss": 0.4991, + "step": 5423 + }, + { + "epoch": 1.735494880546075, + "grad_norm": 0.5232354402542114, + "learning_rate": 4.502792988158197e-06, + "loss": 0.5079, + "step": 5424 + }, + { + "epoch": 1.7358148464163823, + "grad_norm": 0.5111263990402222, + "learning_rate": 4.500940455941504e-06, + "loss": 0.4625, + "step": 5425 + }, + { + "epoch": 1.7361348122866893, + "grad_norm": 0.4850997030735016, + "learning_rate": 4.4990879929200145e-06, + "loss": 0.4769, + "step": 5426 + }, + { + "epoch": 1.7364547781569966, + "grad_norm": 0.5141841769218445, + "learning_rate": 4.497235599350579e-06, + "loss": 0.5197, + "step": 5427 + }, + { + "epoch": 1.7367747440273038, + "grad_norm": 0.5438143610954285, + "learning_rate": 4.495383275490032e-06, + "loss": 0.4813, + "step": 5428 + }, + { + "epoch": 1.7370947098976108, + "grad_norm": 0.5233721733093262, + "learning_rate": 4.4935310215952e-06, + "loss": 0.4919, + "step": 5429 + }, + { + "epoch": 1.737414675767918, + "grad_norm": 0.502494752407074, + "learning_rate": 4.4916788379229024e-06, + "loss": 0.4678, + "step": 5430 + }, + { + "epoch": 1.7377346416382253, + "grad_norm": 0.49350664019584656, + "learning_rate": 4.489826724729946e-06, + "loss": 0.5335, + "step": 5431 + }, + { + "epoch": 1.7380546075085324, + "grad_norm": 0.49686992168426514, + "learning_rate": 4.487974682273128e-06, + "loss": 0.4986, + "step": 5432 + }, + { + "epoch": 1.7383745733788396, + "grad_norm": 0.5533159971237183, + "learning_rate": 4.486122710809234e-06, + "loss": 0.513, + "step": 5433 + }, + { + "epoch": 1.7386945392491469, + "grad_norm": 0.5231044292449951, + "learning_rate": 4.484270810595044e-06, + "loss": 0.4997, + "step": 5434 + }, + { + "epoch": 1.7390145051194539, + "grad_norm": 0.520380973815918, + "learning_rate": 4.4824189818873295e-06, + "loss": 0.4561, + "step": 5435 + }, + { + "epoch": 1.7393344709897611, + "grad_norm": 0.4811047315597534, + "learning_rate": 4.480567224942845e-06, + "loss": 0.4668, + "step": 5436 + }, + { + "epoch": 1.7396544368600684, + "grad_norm": 0.5269007086753845, + "learning_rate": 4.47871554001834e-06, + "loss": 0.5144, + "step": 5437 + }, + { + "epoch": 1.7399744027303754, + "grad_norm": 0.503526508808136, + "learning_rate": 4.476863927370551e-06, + "loss": 0.4572, + "step": 5438 + }, + { + "epoch": 1.7402943686006824, + "grad_norm": 0.4947238266468048, + "learning_rate": 4.475012387256208e-06, + "loss": 0.5347, + "step": 5439 + }, + { + "epoch": 1.74061433447099, + "grad_norm": 0.5225102305412292, + "learning_rate": 4.473160919932029e-06, + "loss": 0.4947, + "step": 5440 + }, + { + "epoch": 1.740934300341297, + "grad_norm": 0.568291425704956, + "learning_rate": 4.471309525654723e-06, + "loss": 0.5045, + "step": 5441 + }, + { + "epoch": 1.741254266211604, + "grad_norm": 0.4956567585468292, + "learning_rate": 4.469458204680988e-06, + "loss": 0.4392, + "step": 5442 + }, + { + "epoch": 1.7415742320819114, + "grad_norm": 0.5219036340713501, + "learning_rate": 4.46760695726751e-06, + "loss": 0.4645, + "step": 5443 + }, + { + "epoch": 1.7418941979522184, + "grad_norm": 0.5405728220939636, + "learning_rate": 4.465755783670968e-06, + "loss": 0.5228, + "step": 5444 + }, + { + "epoch": 1.7422141638225255, + "grad_norm": 0.5067600011825562, + "learning_rate": 4.4639046841480285e-06, + "loss": 0.5074, + "step": 5445 + }, + { + "epoch": 1.7425341296928327, + "grad_norm": 0.5223138332366943, + "learning_rate": 4.462053658955352e-06, + "loss": 0.4536, + "step": 5446 + }, + { + "epoch": 1.74285409556314, + "grad_norm": 0.5472766757011414, + "learning_rate": 4.460202708349582e-06, + "loss": 0.4719, + "step": 5447 + }, + { + "epoch": 1.743174061433447, + "grad_norm": 0.5742120742797852, + "learning_rate": 4.458351832587354e-06, + "loss": 0.4926, + "step": 5448 + }, + { + "epoch": 1.7434940273037542, + "grad_norm": 0.548699140548706, + "learning_rate": 4.456501031925297e-06, + "loss": 0.5002, + "step": 5449 + }, + { + "epoch": 1.7438139931740615, + "grad_norm": 0.4848234951496124, + "learning_rate": 4.454650306620025e-06, + "loss": 0.4795, + "step": 5450 + }, + { + "epoch": 1.7441339590443685, + "grad_norm": 0.4963657855987549, + "learning_rate": 4.452799656928146e-06, + "loss": 0.5057, + "step": 5451 + }, + { + "epoch": 1.7444539249146758, + "grad_norm": 0.49005699157714844, + "learning_rate": 4.450949083106253e-06, + "loss": 0.4562, + "step": 5452 + }, + { + "epoch": 1.744773890784983, + "grad_norm": 0.5466716289520264, + "learning_rate": 4.4490985854109285e-06, + "loss": 0.5327, + "step": 5453 + }, + { + "epoch": 1.74509385665529, + "grad_norm": 0.49012893438339233, + "learning_rate": 4.447248164098747e-06, + "loss": 0.4744, + "step": 5454 + }, + { + "epoch": 1.7454138225255973, + "grad_norm": 0.4600718319416046, + "learning_rate": 4.445397819426274e-06, + "loss": 0.4915, + "step": 5455 + }, + { + "epoch": 1.7457337883959045, + "grad_norm": 0.5044729113578796, + "learning_rate": 4.4435475516500615e-06, + "loss": 0.5106, + "step": 5456 + }, + { + "epoch": 1.7460537542662116, + "grad_norm": 0.5379459857940674, + "learning_rate": 4.441697361026649e-06, + "loss": 0.5507, + "step": 5457 + }, + { + "epoch": 1.7463737201365188, + "grad_norm": 0.5367534756660461, + "learning_rate": 4.439847247812568e-06, + "loss": 0.444, + "step": 5458 + }, + { + "epoch": 1.746693686006826, + "grad_norm": 0.5278060436248779, + "learning_rate": 4.437997212264343e-06, + "loss": 0.5187, + "step": 5459 + }, + { + "epoch": 1.747013651877133, + "grad_norm": 0.6021057367324829, + "learning_rate": 4.4361472546384785e-06, + "loss": 0.5194, + "step": 5460 + }, + { + "epoch": 1.7473336177474401, + "grad_norm": 0.5518205761909485, + "learning_rate": 4.434297375191477e-06, + "loss": 0.4924, + "step": 5461 + }, + { + "epoch": 1.7476535836177476, + "grad_norm": 0.5334278345108032, + "learning_rate": 4.4324475741798235e-06, + "loss": 0.4701, + "step": 5462 + }, + { + "epoch": 1.7479735494880546, + "grad_norm": 0.538500964641571, + "learning_rate": 4.430597851859996e-06, + "loss": 0.47, + "step": 5463 + }, + { + "epoch": 1.7482935153583616, + "grad_norm": 0.5172466039657593, + "learning_rate": 4.428748208488464e-06, + "loss": 0.4833, + "step": 5464 + }, + { + "epoch": 1.7486134812286689, + "grad_norm": 0.4630873501300812, + "learning_rate": 4.426898644321677e-06, + "loss": 0.4668, + "step": 5465 + }, + { + "epoch": 1.7489334470989761, + "grad_norm": 0.5098137259483337, + "learning_rate": 4.425049159616085e-06, + "loss": 0.4796, + "step": 5466 + }, + { + "epoch": 1.7492534129692832, + "grad_norm": 0.5252134799957275, + "learning_rate": 4.423199754628118e-06, + "loss": 0.5101, + "step": 5467 + }, + { + "epoch": 1.7495733788395904, + "grad_norm": 0.48943209648132324, + "learning_rate": 4.421350429614197e-06, + "loss": 0.4858, + "step": 5468 + }, + { + "epoch": 1.7498933447098977, + "grad_norm": 0.4845833480358124, + "learning_rate": 4.4195011848307376e-06, + "loss": 0.4504, + "step": 5469 + }, + { + "epoch": 1.7502133105802047, + "grad_norm": 0.5351172089576721, + "learning_rate": 4.4176520205341365e-06, + "loss": 0.542, + "step": 5470 + }, + { + "epoch": 1.750533276450512, + "grad_norm": 0.496385395526886, + "learning_rate": 4.415802936980784e-06, + "loss": 0.4326, + "step": 5471 + }, + { + "epoch": 1.7508532423208192, + "grad_norm": 0.5513273477554321, + "learning_rate": 4.413953934427057e-06, + "loss": 0.5309, + "step": 5472 + }, + { + "epoch": 1.7511732081911262, + "grad_norm": 0.48049625754356384, + "learning_rate": 4.41210501312932e-06, + "loss": 0.4664, + "step": 5473 + }, + { + "epoch": 1.7514931740614335, + "grad_norm": 0.5158618688583374, + "learning_rate": 4.410256173343932e-06, + "loss": 0.5327, + "step": 5474 + }, + { + "epoch": 1.7518131399317407, + "grad_norm": 0.4715256094932556, + "learning_rate": 4.4084074153272346e-06, + "loss": 0.4593, + "step": 5475 + }, + { + "epoch": 1.7521331058020477, + "grad_norm": 0.5609480738639832, + "learning_rate": 4.406558739335561e-06, + "loss": 0.5126, + "step": 5476 + }, + { + "epoch": 1.752453071672355, + "grad_norm": 0.5760226249694824, + "learning_rate": 4.4047101456252325e-06, + "loss": 0.5278, + "step": 5477 + }, + { + "epoch": 1.7527730375426622, + "grad_norm": 0.5125241875648499, + "learning_rate": 4.402861634452556e-06, + "loss": 0.4898, + "step": 5478 + }, + { + "epoch": 1.7530930034129693, + "grad_norm": 0.5064290165901184, + "learning_rate": 4.4010132060738356e-06, + "loss": 0.4935, + "step": 5479 + }, + { + "epoch": 1.7534129692832765, + "grad_norm": 0.49492013454437256, + "learning_rate": 4.399164860745353e-06, + "loss": 0.5455, + "step": 5480 + }, + { + "epoch": 1.7537329351535837, + "grad_norm": 0.4533080756664276, + "learning_rate": 4.397316598723385e-06, + "loss": 0.3917, + "step": 5481 + }, + { + "epoch": 1.7540529010238908, + "grad_norm": 0.48983830213546753, + "learning_rate": 4.395468420264197e-06, + "loss": 0.5256, + "step": 5482 + }, + { + "epoch": 1.7543728668941978, + "grad_norm": 0.4834545850753784, + "learning_rate": 4.393620325624038e-06, + "loss": 0.4618, + "step": 5483 + }, + { + "epoch": 1.7546928327645053, + "grad_norm": 0.48400387167930603, + "learning_rate": 4.391772315059152e-06, + "loss": 0.4857, + "step": 5484 + }, + { + "epoch": 1.7550127986348123, + "grad_norm": 0.4745435416698456, + "learning_rate": 4.389924388825766e-06, + "loss": 0.5117, + "step": 5485 + }, + { + "epoch": 1.7553327645051193, + "grad_norm": 0.48883339762687683, + "learning_rate": 4.388076547180097e-06, + "loss": 0.4828, + "step": 5486 + }, + { + "epoch": 1.7556527303754266, + "grad_norm": 0.5395194888114929, + "learning_rate": 4.386228790378351e-06, + "loss": 0.4743, + "step": 5487 + }, + { + "epoch": 1.7559726962457338, + "grad_norm": 0.5261681079864502, + "learning_rate": 4.38438111867672e-06, + "loss": 0.4712, + "step": 5488 + }, + { + "epoch": 1.7562926621160408, + "grad_norm": 0.5287356972694397, + "learning_rate": 4.38253353233139e-06, + "loss": 0.5399, + "step": 5489 + }, + { + "epoch": 1.756612627986348, + "grad_norm": 0.5015385150909424, + "learning_rate": 4.380686031598527e-06, + "loss": 0.464, + "step": 5490 + }, + { + "epoch": 1.7569325938566553, + "grad_norm": 0.49232450127601624, + "learning_rate": 4.378838616734292e-06, + "loss": 0.4641, + "step": 5491 + }, + { + "epoch": 1.7572525597269624, + "grad_norm": 0.49756553769111633, + "learning_rate": 4.376991287994829e-06, + "loss": 0.4978, + "step": 5492 + }, + { + "epoch": 1.7575725255972696, + "grad_norm": 0.48109307885169983, + "learning_rate": 4.375144045636273e-06, + "loss": 0.4733, + "step": 5493 + }, + { + "epoch": 1.7578924914675769, + "grad_norm": 0.4919877052307129, + "learning_rate": 4.373296889914748e-06, + "loss": 0.5169, + "step": 5494 + }, + { + "epoch": 1.758212457337884, + "grad_norm": 0.46694469451904297, + "learning_rate": 4.371449821086363e-06, + "loss": 0.4253, + "step": 5495 + }, + { + "epoch": 1.7585324232081911, + "grad_norm": 0.5384265184402466, + "learning_rate": 4.369602839407217e-06, + "loss": 0.4927, + "step": 5496 + }, + { + "epoch": 1.7588523890784984, + "grad_norm": 0.4768705666065216, + "learning_rate": 4.367755945133394e-06, + "loss": 0.4495, + "step": 5497 + }, + { + "epoch": 1.7591723549488054, + "grad_norm": 0.6496914625167847, + "learning_rate": 4.36590913852097e-06, + "loss": 0.5469, + "step": 5498 + }, + { + "epoch": 1.7594923208191127, + "grad_norm": 0.47362634539604187, + "learning_rate": 4.364062419826008e-06, + "loss": 0.4203, + "step": 5499 + }, + { + "epoch": 1.75981228668942, + "grad_norm": 0.5249969959259033, + "learning_rate": 4.3622157893045565e-06, + "loss": 0.5064, + "step": 5500 + }, + { + "epoch": 1.760132252559727, + "grad_norm": 0.4841233789920807, + "learning_rate": 4.360369247212653e-06, + "loss": 0.4478, + "step": 5501 + }, + { + "epoch": 1.760452218430034, + "grad_norm": 0.5113783478736877, + "learning_rate": 4.358522793806322e-06, + "loss": 0.5385, + "step": 5502 + }, + { + "epoch": 1.7607721843003414, + "grad_norm": 0.4811757504940033, + "learning_rate": 4.356676429341577e-06, + "loss": 0.4544, + "step": 5503 + }, + { + "epoch": 1.7610921501706485, + "grad_norm": 0.48280641436576843, + "learning_rate": 4.354830154074421e-06, + "loss": 0.4919, + "step": 5504 + }, + { + "epoch": 1.7614121160409555, + "grad_norm": 0.4716619849205017, + "learning_rate": 4.352983968260839e-06, + "loss": 0.4455, + "step": 5505 + }, + { + "epoch": 1.761732081911263, + "grad_norm": 0.5332163572311401, + "learning_rate": 4.351137872156809e-06, + "loss": 0.513, + "step": 5506 + }, + { + "epoch": 1.76205204778157, + "grad_norm": 0.526752769947052, + "learning_rate": 4.349291866018292e-06, + "loss": 0.4645, + "step": 5507 + }, + { + "epoch": 1.762372013651877, + "grad_norm": 0.5340533256530762, + "learning_rate": 4.347445950101241e-06, + "loss": 0.5284, + "step": 5508 + }, + { + "epoch": 1.7626919795221843, + "grad_norm": 0.4629971385002136, + "learning_rate": 4.345600124661596e-06, + "loss": 0.4387, + "step": 5509 + }, + { + "epoch": 1.7630119453924915, + "grad_norm": 0.4559417963027954, + "learning_rate": 4.343754389955279e-06, + "loss": 0.5109, + "step": 5510 + }, + { + "epoch": 1.7633319112627985, + "grad_norm": 0.49508291482925415, + "learning_rate": 4.341908746238207e-06, + "loss": 0.4781, + "step": 5511 + }, + { + "epoch": 1.7636518771331058, + "grad_norm": 0.4971843361854553, + "learning_rate": 4.340063193766277e-06, + "loss": 0.4451, + "step": 5512 + }, + { + "epoch": 1.763971843003413, + "grad_norm": 0.5565189719200134, + "learning_rate": 4.33821773279538e-06, + "loss": 0.5169, + "step": 5513 + }, + { + "epoch": 1.76429180887372, + "grad_norm": 0.47722023725509644, + "learning_rate": 4.336372363581391e-06, + "loss": 0.4604, + "step": 5514 + }, + { + "epoch": 1.7646117747440273, + "grad_norm": 0.504019021987915, + "learning_rate": 4.334527086380171e-06, + "loss": 0.4904, + "step": 5515 + }, + { + "epoch": 1.7649317406143346, + "grad_norm": 0.5100414156913757, + "learning_rate": 4.3326819014475726e-06, + "loss": 0.4528, + "step": 5516 + }, + { + "epoch": 1.7652517064846416, + "grad_norm": 0.55582195520401, + "learning_rate": 4.3308368090394295e-06, + "loss": 0.5074, + "step": 5517 + }, + { + "epoch": 1.7655716723549488, + "grad_norm": 0.530032753944397, + "learning_rate": 4.328991809411569e-06, + "loss": 0.4683, + "step": 5518 + }, + { + "epoch": 1.765891638225256, + "grad_norm": 0.5319012403488159, + "learning_rate": 4.327146902819802e-06, + "loss": 0.4514, + "step": 5519 + }, + { + "epoch": 1.766211604095563, + "grad_norm": 0.5432710647583008, + "learning_rate": 4.325302089519926e-06, + "loss": 0.4615, + "step": 5520 + }, + { + "epoch": 1.7665315699658704, + "grad_norm": 0.5205046534538269, + "learning_rate": 4.323457369767727e-06, + "loss": 0.479, + "step": 5521 + }, + { + "epoch": 1.7668515358361776, + "grad_norm": 0.6415116786956787, + "learning_rate": 4.321612743818976e-06, + "loss": 0.5481, + "step": 5522 + }, + { + "epoch": 1.7671715017064846, + "grad_norm": 0.5402652621269226, + "learning_rate": 4.319768211929435e-06, + "loss": 0.4699, + "step": 5523 + }, + { + "epoch": 1.7674914675767917, + "grad_norm": 0.5080769062042236, + "learning_rate": 4.3179237743548516e-06, + "loss": 0.5002, + "step": 5524 + }, + { + "epoch": 1.7678114334470991, + "grad_norm": 0.49535703659057617, + "learning_rate": 4.316079431350956e-06, + "loss": 0.4754, + "step": 5525 + }, + { + "epoch": 1.7681313993174061, + "grad_norm": 0.527343213558197, + "learning_rate": 4.31423518317347e-06, + "loss": 0.4869, + "step": 5526 + }, + { + "epoch": 1.7684513651877132, + "grad_norm": 0.48264944553375244, + "learning_rate": 4.3123910300781e-06, + "loss": 0.4852, + "step": 5527 + }, + { + "epoch": 1.7687713310580204, + "grad_norm": 0.5463948845863342, + "learning_rate": 4.310546972320542e-06, + "loss": 0.4725, + "step": 5528 + }, + { + "epoch": 1.7690912969283277, + "grad_norm": 0.598416805267334, + "learning_rate": 4.308703010156476e-06, + "loss": 0.5052, + "step": 5529 + }, + { + "epoch": 1.7694112627986347, + "grad_norm": 0.5130217671394348, + "learning_rate": 4.306859143841569e-06, + "loss": 0.4952, + "step": 5530 + }, + { + "epoch": 1.769731228668942, + "grad_norm": 0.5175359845161438, + "learning_rate": 4.305015373631474e-06, + "loss": 0.5155, + "step": 5531 + }, + { + "epoch": 1.7700511945392492, + "grad_norm": 0.5057743787765503, + "learning_rate": 4.303171699781835e-06, + "loss": 0.451, + "step": 5532 + }, + { + "epoch": 1.7703711604095562, + "grad_norm": 0.6024937629699707, + "learning_rate": 4.301328122548278e-06, + "loss": 0.447, + "step": 5533 + }, + { + "epoch": 1.7706911262798635, + "grad_norm": 0.5402462482452393, + "learning_rate": 4.299484642186418e-06, + "loss": 0.5155, + "step": 5534 + }, + { + "epoch": 1.7710110921501707, + "grad_norm": 0.5450391173362732, + "learning_rate": 4.297641258951855e-06, + "loss": 0.5274, + "step": 5535 + }, + { + "epoch": 1.7713310580204777, + "grad_norm": 0.52930748462677, + "learning_rate": 4.295797973100174e-06, + "loss": 0.4358, + "step": 5536 + }, + { + "epoch": 1.771651023890785, + "grad_norm": 0.5490062236785889, + "learning_rate": 4.293954784886954e-06, + "loss": 0.52, + "step": 5537 + }, + { + "epoch": 1.7719709897610922, + "grad_norm": 0.5440360307693481, + "learning_rate": 4.292111694567751e-06, + "loss": 0.4968, + "step": 5538 + }, + { + "epoch": 1.7722909556313993, + "grad_norm": 0.5466500520706177, + "learning_rate": 4.290268702398115e-06, + "loss": 0.446, + "step": 5539 + }, + { + "epoch": 1.7726109215017065, + "grad_norm": 0.49099576473236084, + "learning_rate": 4.2884258086335755e-06, + "loss": 0.5082, + "step": 5540 + }, + { + "epoch": 1.7729308873720138, + "grad_norm": 0.5124669075012207, + "learning_rate": 4.286583013529653e-06, + "loss": 0.4779, + "step": 5541 + }, + { + "epoch": 1.7732508532423208, + "grad_norm": 0.5103096961975098, + "learning_rate": 4.284740317341856e-06, + "loss": 0.4809, + "step": 5542 + }, + { + "epoch": 1.773570819112628, + "grad_norm": 0.5411840081214905, + "learning_rate": 4.282897720325673e-06, + "loss": 0.4795, + "step": 5543 + }, + { + "epoch": 1.7738907849829353, + "grad_norm": 0.540686309337616, + "learning_rate": 4.281055222736586e-06, + "loss": 0.4575, + "step": 5544 + }, + { + "epoch": 1.7742107508532423, + "grad_norm": 0.5696617960929871, + "learning_rate": 4.2792128248300546e-06, + "loss": 0.5115, + "step": 5545 + }, + { + "epoch": 1.7745307167235493, + "grad_norm": 0.5039750337600708, + "learning_rate": 4.2773705268615315e-06, + "loss": 0.4749, + "step": 5546 + }, + { + "epoch": 1.7748506825938568, + "grad_norm": 0.522910475730896, + "learning_rate": 4.275528329086457e-06, + "loss": 0.4636, + "step": 5547 + }, + { + "epoch": 1.7751706484641638, + "grad_norm": 0.5602678060531616, + "learning_rate": 4.273686231760249e-06, + "loss": 0.5559, + "step": 5548 + }, + { + "epoch": 1.7754906143344709, + "grad_norm": 0.5650404691696167, + "learning_rate": 4.271844235138321e-06, + "loss": 0.5355, + "step": 5549 + }, + { + "epoch": 1.775810580204778, + "grad_norm": 0.4923083782196045, + "learning_rate": 4.270002339476065e-06, + "loss": 0.4698, + "step": 5550 + }, + { + "epoch": 1.7761305460750854, + "grad_norm": 0.5023771524429321, + "learning_rate": 4.268160545028861e-06, + "loss": 0.4528, + "step": 5551 + }, + { + "epoch": 1.7764505119453924, + "grad_norm": 0.529879629611969, + "learning_rate": 4.26631885205208e-06, + "loss": 0.5279, + "step": 5552 + }, + { + "epoch": 1.7767704778156996, + "grad_norm": 0.4917384386062622, + "learning_rate": 4.264477260801072e-06, + "loss": 0.4707, + "step": 5553 + }, + { + "epoch": 1.7770904436860069, + "grad_norm": 0.5691450834274292, + "learning_rate": 4.262635771531178e-06, + "loss": 0.5206, + "step": 5554 + }, + { + "epoch": 1.777410409556314, + "grad_norm": 0.5118280649185181, + "learning_rate": 4.26079438449772e-06, + "loss": 0.4956, + "step": 5555 + }, + { + "epoch": 1.7777303754266212, + "grad_norm": 0.49388378858566284, + "learning_rate": 4.258953099956008e-06, + "loss": 0.4807, + "step": 5556 + }, + { + "epoch": 1.7780503412969284, + "grad_norm": 0.4365639388561249, + "learning_rate": 4.257111918161343e-06, + "loss": 0.4138, + "step": 5557 + }, + { + "epoch": 1.7783703071672354, + "grad_norm": 0.5118427872657776, + "learning_rate": 4.2552708393690035e-06, + "loss": 0.5282, + "step": 5558 + }, + { + "epoch": 1.7786902730375427, + "grad_norm": 0.5046088695526123, + "learning_rate": 4.253429863834258e-06, + "loss": 0.5192, + "step": 5559 + }, + { + "epoch": 1.77901023890785, + "grad_norm": 0.5091738700866699, + "learning_rate": 4.25158899181236e-06, + "loss": 0.5138, + "step": 5560 + }, + { + "epoch": 1.779330204778157, + "grad_norm": 12.748663902282715, + "learning_rate": 4.249748223558547e-06, + "loss": 0.4696, + "step": 5561 + }, + { + "epoch": 1.7796501706484642, + "grad_norm": 0.5930684208869934, + "learning_rate": 4.247907559328049e-06, + "loss": 0.5821, + "step": 5562 + }, + { + "epoch": 1.7799701365187715, + "grad_norm": 0.5774797201156616, + "learning_rate": 4.24606699937607e-06, + "loss": 0.5031, + "step": 5563 + }, + { + "epoch": 1.7802901023890785, + "grad_norm": 0.5613293051719666, + "learning_rate": 4.24422654395781e-06, + "loss": 0.4981, + "step": 5564 + }, + { + "epoch": 1.7806100682593855, + "grad_norm": 0.5332891345024109, + "learning_rate": 4.242386193328447e-06, + "loss": 0.4651, + "step": 5565 + }, + { + "epoch": 1.780930034129693, + "grad_norm": 0.564367949962616, + "learning_rate": 4.2405459477431505e-06, + "loss": 0.5202, + "step": 5566 + }, + { + "epoch": 1.78125, + "grad_norm": 0.5317816734313965, + "learning_rate": 4.238705807457073e-06, + "loss": 0.4937, + "step": 5567 + }, + { + "epoch": 1.781569965870307, + "grad_norm": 0.46839404106140137, + "learning_rate": 4.236865772725351e-06, + "loss": 0.4636, + "step": 5568 + }, + { + "epoch": 1.7818899317406145, + "grad_norm": 0.45433491468429565, + "learning_rate": 4.2350258438031075e-06, + "loss": 0.4566, + "step": 5569 + }, + { + "epoch": 1.7822098976109215, + "grad_norm": 0.5349555611610413, + "learning_rate": 4.23318602094545e-06, + "loss": 0.4865, + "step": 5570 + }, + { + "epoch": 1.7825298634812285, + "grad_norm": 0.5639216303825378, + "learning_rate": 4.2313463044074735e-06, + "loss": 0.4934, + "step": 5571 + }, + { + "epoch": 1.7828498293515358, + "grad_norm": 0.5030721426010132, + "learning_rate": 4.229506694444258e-06, + "loss": 0.471, + "step": 5572 + }, + { + "epoch": 1.783169795221843, + "grad_norm": 0.5446768999099731, + "learning_rate": 4.227667191310865e-06, + "loss": 0.5168, + "step": 5573 + }, + { + "epoch": 1.78348976109215, + "grad_norm": 0.5222355723381042, + "learning_rate": 4.225827795262346e-06, + "loss": 0.4883, + "step": 5574 + }, + { + "epoch": 1.7838097269624573, + "grad_norm": 0.5582959055900574, + "learning_rate": 4.223988506553732e-06, + "loss": 0.4787, + "step": 5575 + }, + { + "epoch": 1.7841296928327646, + "grad_norm": 0.5071340799331665, + "learning_rate": 4.222149325440048e-06, + "loss": 0.4875, + "step": 5576 + }, + { + "epoch": 1.7844496587030716, + "grad_norm": 0.5143452882766724, + "learning_rate": 4.220310252176295e-06, + "loss": 0.4675, + "step": 5577 + }, + { + "epoch": 1.7847696245733788, + "grad_norm": 0.4680158793926239, + "learning_rate": 4.218471287017463e-06, + "loss": 0.5007, + "step": 5578 + }, + { + "epoch": 1.785089590443686, + "grad_norm": 0.5283663868904114, + "learning_rate": 4.216632430218528e-06, + "loss": 0.5296, + "step": 5579 + }, + { + "epoch": 1.7854095563139931, + "grad_norm": 0.4833642244338989, + "learning_rate": 4.2147936820344484e-06, + "loss": 0.4334, + "step": 5580 + }, + { + "epoch": 1.7857295221843004, + "grad_norm": 0.5667260885238647, + "learning_rate": 4.212955042720169e-06, + "loss": 0.5489, + "step": 5581 + }, + { + "epoch": 1.7860494880546076, + "grad_norm": 0.5367039442062378, + "learning_rate": 4.211116512530621e-06, + "loss": 0.4817, + "step": 5582 + }, + { + "epoch": 1.7863694539249146, + "grad_norm": 0.5276321172714233, + "learning_rate": 4.2092780917207165e-06, + "loss": 0.4729, + "step": 5583 + }, + { + "epoch": 1.786689419795222, + "grad_norm": 0.4945152997970581, + "learning_rate": 4.2074397805453575e-06, + "loss": 0.4959, + "step": 5584 + }, + { + "epoch": 1.7870093856655291, + "grad_norm": 0.5138298273086548, + "learning_rate": 4.205601579259424e-06, + "loss": 0.4802, + "step": 5585 + }, + { + "epoch": 1.7873293515358362, + "grad_norm": 0.5320633053779602, + "learning_rate": 4.203763488117788e-06, + "loss": 0.4885, + "step": 5586 + }, + { + "epoch": 1.7876493174061432, + "grad_norm": 0.5252824425697327, + "learning_rate": 4.201925507375303e-06, + "loss": 0.5041, + "step": 5587 + }, + { + "epoch": 1.7879692832764507, + "grad_norm": 0.5192841291427612, + "learning_rate": 4.200087637286807e-06, + "loss": 0.5156, + "step": 5588 + }, + { + "epoch": 1.7882892491467577, + "grad_norm": 0.5447317957878113, + "learning_rate": 4.198249878107122e-06, + "loss": 0.4994, + "step": 5589 + }, + { + "epoch": 1.7886092150170647, + "grad_norm": 0.48571932315826416, + "learning_rate": 4.196412230091053e-06, + "loss": 0.4869, + "step": 5590 + }, + { + "epoch": 1.788929180887372, + "grad_norm": 0.5020157098770142, + "learning_rate": 4.194574693493398e-06, + "loss": 0.4807, + "step": 5591 + }, + { + "epoch": 1.7892491467576792, + "grad_norm": 0.5085739493370056, + "learning_rate": 4.1927372685689315e-06, + "loss": 0.4823, + "step": 5592 + }, + { + "epoch": 1.7895691126279862, + "grad_norm": 0.567353367805481, + "learning_rate": 4.190899955572413e-06, + "loss": 0.5069, + "step": 5593 + }, + { + "epoch": 1.7898890784982935, + "grad_norm": 0.5270597338676453, + "learning_rate": 4.18906275475859e-06, + "loss": 0.5283, + "step": 5594 + }, + { + "epoch": 1.7902090443686007, + "grad_norm": 0.46188998222351074, + "learning_rate": 4.187225666382187e-06, + "loss": 0.4108, + "step": 5595 + }, + { + "epoch": 1.7905290102389078, + "grad_norm": 0.607689380645752, + "learning_rate": 4.185388690697927e-06, + "loss": 0.5774, + "step": 5596 + }, + { + "epoch": 1.790848976109215, + "grad_norm": 0.615105926990509, + "learning_rate": 4.183551827960505e-06, + "loss": 0.4953, + "step": 5597 + }, + { + "epoch": 1.7911689419795223, + "grad_norm": 0.5350819230079651, + "learning_rate": 4.181715078424604e-06, + "loss": 0.5159, + "step": 5598 + }, + { + "epoch": 1.7914889078498293, + "grad_norm": 0.5074175000190735, + "learning_rate": 4.179878442344892e-06, + "loss": 0.4562, + "step": 5599 + }, + { + "epoch": 1.7918088737201365, + "grad_norm": 0.49207979440689087, + "learning_rate": 4.178041919976018e-06, + "loss": 0.4636, + "step": 5600 + }, + { + "epoch": 1.7921288395904438, + "grad_norm": 0.523507297039032, + "learning_rate": 4.176205511572624e-06, + "loss": 0.5079, + "step": 5601 + }, + { + "epoch": 1.7924488054607508, + "grad_norm": 0.5207238793373108, + "learning_rate": 4.174369217389326e-06, + "loss": 0.4361, + "step": 5602 + }, + { + "epoch": 1.792768771331058, + "grad_norm": 0.5948736667633057, + "learning_rate": 4.17253303768073e-06, + "loss": 0.5183, + "step": 5603 + }, + { + "epoch": 1.7930887372013653, + "grad_norm": 0.5301239490509033, + "learning_rate": 4.1706969727014205e-06, + "loss": 0.5135, + "step": 5604 + }, + { + "epoch": 1.7934087030716723, + "grad_norm": 0.5359353423118591, + "learning_rate": 4.168861022705976e-06, + "loss": 0.4835, + "step": 5605 + }, + { + "epoch": 1.7937286689419796, + "grad_norm": 0.5441157817840576, + "learning_rate": 4.1670251879489506e-06, + "loss": 0.5016, + "step": 5606 + }, + { + "epoch": 1.7940486348122868, + "grad_norm": 0.5022502541542053, + "learning_rate": 4.1651894686848844e-06, + "loss": 0.5005, + "step": 5607 + }, + { + "epoch": 1.7943686006825939, + "grad_norm": 0.5129440426826477, + "learning_rate": 4.163353865168303e-06, + "loss": 0.5304, + "step": 5608 + }, + { + "epoch": 1.7946885665529009, + "grad_norm": 0.46050283312797546, + "learning_rate": 4.161518377653714e-06, + "loss": 0.4459, + "step": 5609 + }, + { + "epoch": 1.7950085324232083, + "grad_norm": 0.5066482424736023, + "learning_rate": 4.159683006395611e-06, + "loss": 0.4895, + "step": 5610 + }, + { + "epoch": 1.7953284982935154, + "grad_norm": 0.5382094979286194, + "learning_rate": 4.157847751648471e-06, + "loss": 0.5432, + "step": 5611 + }, + { + "epoch": 1.7956484641638224, + "grad_norm": 0.4516243636608124, + "learning_rate": 4.156012613666753e-06, + "loss": 0.4866, + "step": 5612 + }, + { + "epoch": 1.7959684300341296, + "grad_norm": 0.5349157452583313, + "learning_rate": 4.154177592704902e-06, + "loss": 0.5145, + "step": 5613 + }, + { + "epoch": 1.796288395904437, + "grad_norm": 0.4918440580368042, + "learning_rate": 4.152342689017342e-06, + "loss": 0.4627, + "step": 5614 + }, + { + "epoch": 1.796608361774744, + "grad_norm": 0.5058891177177429, + "learning_rate": 4.15050790285849e-06, + "loss": 0.4573, + "step": 5615 + }, + { + "epoch": 1.7969283276450512, + "grad_norm": 0.5324491262435913, + "learning_rate": 4.148673234482741e-06, + "loss": 0.4959, + "step": 5616 + }, + { + "epoch": 1.7972482935153584, + "grad_norm": 0.5199218988418579, + "learning_rate": 4.146838684144471e-06, + "loss": 0.5029, + "step": 5617 + }, + { + "epoch": 1.7975682593856654, + "grad_norm": 0.5540761947631836, + "learning_rate": 4.145004252098044e-06, + "loss": 0.5378, + "step": 5618 + }, + { + "epoch": 1.7978882252559727, + "grad_norm": 0.4679339528083801, + "learning_rate": 4.143169938597805e-06, + "loss": 0.4652, + "step": 5619 + }, + { + "epoch": 1.79820819112628, + "grad_norm": 0.5310577154159546, + "learning_rate": 4.141335743898085e-06, + "loss": 0.4744, + "step": 5620 + }, + { + "epoch": 1.798528156996587, + "grad_norm": 0.502953290939331, + "learning_rate": 4.139501668253199e-06, + "loss": 0.468, + "step": 5621 + }, + { + "epoch": 1.7988481228668942, + "grad_norm": 0.4934243857860565, + "learning_rate": 4.13766771191744e-06, + "loss": 0.5008, + "step": 5622 + }, + { + "epoch": 1.7991680887372015, + "grad_norm": 0.4875839054584503, + "learning_rate": 4.135833875145091e-06, + "loss": 0.5144, + "step": 5623 + }, + { + "epoch": 1.7994880546075085, + "grad_norm": 0.4996601343154907, + "learning_rate": 4.1340001581904135e-06, + "loss": 0.463, + "step": 5624 + }, + { + "epoch": 1.7998080204778157, + "grad_norm": 0.49659842252731323, + "learning_rate": 4.132166561307657e-06, + "loss": 0.4526, + "step": 5625 + }, + { + "epoch": 1.800127986348123, + "grad_norm": 0.5389164686203003, + "learning_rate": 4.130333084751051e-06, + "loss": 0.4994, + "step": 5626 + }, + { + "epoch": 1.80044795221843, + "grad_norm": 0.5688105821609497, + "learning_rate": 4.128499728774808e-06, + "loss": 0.5709, + "step": 5627 + }, + { + "epoch": 1.800767918088737, + "grad_norm": 0.5026649236679077, + "learning_rate": 4.126666493633128e-06, + "loss": 0.4336, + "step": 5628 + }, + { + "epoch": 1.8010878839590445, + "grad_norm": 0.612531304359436, + "learning_rate": 4.124833379580186e-06, + "loss": 0.5483, + "step": 5629 + }, + { + "epoch": 1.8014078498293515, + "grad_norm": 0.554692804813385, + "learning_rate": 4.1230003868701515e-06, + "loss": 0.4851, + "step": 5630 + }, + { + "epoch": 1.8017278156996586, + "grad_norm": 0.5775287747383118, + "learning_rate": 4.121167515757168e-06, + "loss": 0.5384, + "step": 5631 + }, + { + "epoch": 1.802047781569966, + "grad_norm": 0.46787363290786743, + "learning_rate": 4.119334766495366e-06, + "loss": 0.4737, + "step": 5632 + }, + { + "epoch": 1.802367747440273, + "grad_norm": 0.4719943404197693, + "learning_rate": 4.117502139338858e-06, + "loss": 0.4557, + "step": 5633 + }, + { + "epoch": 1.80268771331058, + "grad_norm": 0.5867651700973511, + "learning_rate": 4.1156696345417386e-06, + "loss": 0.515, + "step": 5634 + }, + { + "epoch": 1.8030076791808873, + "grad_norm": 0.47377723455429077, + "learning_rate": 4.113837252358089e-06, + "loss": 0.4255, + "step": 5635 + }, + { + "epoch": 1.8033276450511946, + "grad_norm": 0.5013930797576904, + "learning_rate": 4.112004993041972e-06, + "loss": 0.4913, + "step": 5636 + }, + { + "epoch": 1.8036476109215016, + "grad_norm": 0.46539056301116943, + "learning_rate": 4.110172856847428e-06, + "loss": 0.4621, + "step": 5637 + }, + { + "epoch": 1.8039675767918089, + "grad_norm": 0.5124224424362183, + "learning_rate": 4.108340844028489e-06, + "loss": 0.5124, + "step": 5638 + }, + { + "epoch": 1.804287542662116, + "grad_norm": 0.4970622658729553, + "learning_rate": 4.106508954839166e-06, + "loss": 0.489, + "step": 5639 + }, + { + "epoch": 1.8046075085324231, + "grad_norm": 0.4849103093147278, + "learning_rate": 4.10467718953345e-06, + "loss": 0.4426, + "step": 5640 + }, + { + "epoch": 1.8049274744027304, + "grad_norm": 0.4756026566028595, + "learning_rate": 4.102845548365321e-06, + "loss": 0.4956, + "step": 5641 + }, + { + "epoch": 1.8052474402730376, + "grad_norm": 0.42802461981773376, + "learning_rate": 4.101014031588734e-06, + "loss": 0.4411, + "step": 5642 + }, + { + "epoch": 1.8055674061433447, + "grad_norm": 0.5087606906890869, + "learning_rate": 4.0991826394576315e-06, + "loss": 0.4988, + "step": 5643 + }, + { + "epoch": 1.805887372013652, + "grad_norm": 0.5063179731369019, + "learning_rate": 4.097351372225943e-06, + "loss": 0.4359, + "step": 5644 + }, + { + "epoch": 1.8062073378839592, + "grad_norm": 0.5513647198677063, + "learning_rate": 4.095520230147571e-06, + "loss": 0.5395, + "step": 5645 + }, + { + "epoch": 1.8065273037542662, + "grad_norm": 0.5584523677825928, + "learning_rate": 4.093689213476408e-06, + "loss": 0.5034, + "step": 5646 + }, + { + "epoch": 1.8068472696245734, + "grad_norm": 0.46291977167129517, + "learning_rate": 4.091858322466325e-06, + "loss": 0.4515, + "step": 5647 + }, + { + "epoch": 1.8071672354948807, + "grad_norm": 0.5163590312004089, + "learning_rate": 4.090027557371177e-06, + "loss": 0.4652, + "step": 5648 + }, + { + "epoch": 1.8074872013651877, + "grad_norm": 0.5587126612663269, + "learning_rate": 4.088196918444806e-06, + "loss": 0.5111, + "step": 5649 + }, + { + "epoch": 1.8078071672354947, + "grad_norm": 0.46405214071273804, + "learning_rate": 4.086366405941027e-06, + "loss": 0.3938, + "step": 5650 + }, + { + "epoch": 1.8081271331058022, + "grad_norm": 0.48423051834106445, + "learning_rate": 4.084536020113646e-06, + "loss": 0.462, + "step": 5651 + }, + { + "epoch": 1.8084470989761092, + "grad_norm": 0.5340949296951294, + "learning_rate": 4.0827057612164465e-06, + "loss": 0.5212, + "step": 5652 + }, + { + "epoch": 1.8087670648464163, + "grad_norm": 0.5507909655570984, + "learning_rate": 4.080875629503196e-06, + "loss": 0.5558, + "step": 5653 + }, + { + "epoch": 1.8090870307167235, + "grad_norm": 0.5128883123397827, + "learning_rate": 4.079045625227646e-06, + "loss": 0.4978, + "step": 5654 + }, + { + "epoch": 1.8094069965870307, + "grad_norm": 0.48712173104286194, + "learning_rate": 4.077215748643528e-06, + "loss": 0.4804, + "step": 5655 + }, + { + "epoch": 1.8097269624573378, + "grad_norm": 0.5109673142433167, + "learning_rate": 4.075386000004558e-06, + "loss": 0.5142, + "step": 5656 + }, + { + "epoch": 1.810046928327645, + "grad_norm": 0.5423082709312439, + "learning_rate": 4.073556379564429e-06, + "loss": 0.4911, + "step": 5657 + }, + { + "epoch": 1.8103668941979523, + "grad_norm": 0.564723551273346, + "learning_rate": 4.071726887576823e-06, + "loss": 0.5238, + "step": 5658 + }, + { + "epoch": 1.8106868600682593, + "grad_norm": 0.47076481580734253, + "learning_rate": 4.069897524295402e-06, + "loss": 0.4368, + "step": 5659 + }, + { + "epoch": 1.8110068259385665, + "grad_norm": 0.5020716786384583, + "learning_rate": 4.068068289973807e-06, + "loss": 0.4936, + "step": 5660 + }, + { + "epoch": 1.8113267918088738, + "grad_norm": 0.5677059292793274, + "learning_rate": 4.066239184865666e-06, + "loss": 0.5229, + "step": 5661 + }, + { + "epoch": 1.8116467576791808, + "grad_norm": 0.4805668592453003, + "learning_rate": 4.0644102092245845e-06, + "loss": 0.45, + "step": 5662 + }, + { + "epoch": 1.811966723549488, + "grad_norm": 0.5515758395195007, + "learning_rate": 4.062581363304152e-06, + "loss": 0.5542, + "step": 5663 + }, + { + "epoch": 1.8122866894197953, + "grad_norm": 0.4808933138847351, + "learning_rate": 4.0607526473579434e-06, + "loss": 0.4161, + "step": 5664 + }, + { + "epoch": 1.8126066552901023, + "grad_norm": 0.5447078943252563, + "learning_rate": 4.058924061639508e-06, + "loss": 0.51, + "step": 5665 + }, + { + "epoch": 1.8129266211604096, + "grad_norm": 0.5344134569168091, + "learning_rate": 4.057095606402387e-06, + "loss": 0.473, + "step": 5666 + }, + { + "epoch": 1.8132465870307168, + "grad_norm": 0.565631628036499, + "learning_rate": 4.055267281900091e-06, + "loss": 0.5364, + "step": 5667 + }, + { + "epoch": 1.8135665529010239, + "grad_norm": 0.5139793157577515, + "learning_rate": 4.053439088386124e-06, + "loss": 0.4596, + "step": 5668 + }, + { + "epoch": 1.8138865187713311, + "grad_norm": 0.5374452471733093, + "learning_rate": 4.051611026113967e-06, + "loss": 0.5074, + "step": 5669 + }, + { + "epoch": 1.8142064846416384, + "grad_norm": 0.5051067471504211, + "learning_rate": 4.049783095337081e-06, + "loss": 0.5007, + "step": 5670 + }, + { + "epoch": 1.8145264505119454, + "grad_norm": 0.4644712805747986, + "learning_rate": 4.047955296308913e-06, + "loss": 0.4286, + "step": 5671 + }, + { + "epoch": 1.8148464163822524, + "grad_norm": 0.5553407669067383, + "learning_rate": 4.046127629282887e-06, + "loss": 0.4873, + "step": 5672 + }, + { + "epoch": 1.8151663822525599, + "grad_norm": 0.566815197467804, + "learning_rate": 4.0443000945124135e-06, + "loss": 0.5101, + "step": 5673 + }, + { + "epoch": 1.815486348122867, + "grad_norm": 0.505651593208313, + "learning_rate": 4.042472692250883e-06, + "loss": 0.4675, + "step": 5674 + }, + { + "epoch": 1.815806313993174, + "grad_norm": 0.609451413154602, + "learning_rate": 4.0406454227516635e-06, + "loss": 0.5255, + "step": 5675 + }, + { + "epoch": 1.8161262798634812, + "grad_norm": 0.5524954795837402, + "learning_rate": 4.038818286268113e-06, + "loss": 0.5007, + "step": 5676 + }, + { + "epoch": 1.8164462457337884, + "grad_norm": 0.5468558669090271, + "learning_rate": 4.03699128305356e-06, + "loss": 0.4939, + "step": 5677 + }, + { + "epoch": 1.8167662116040955, + "grad_norm": 0.5983974933624268, + "learning_rate": 4.035164413361325e-06, + "loss": 0.4809, + "step": 5678 + }, + { + "epoch": 1.8170861774744027, + "grad_norm": 0.4928704500198364, + "learning_rate": 4.033337677444708e-06, + "loss": 0.5119, + "step": 5679 + }, + { + "epoch": 1.81740614334471, + "grad_norm": 0.4960091710090637, + "learning_rate": 4.031511075556982e-06, + "loss": 0.4016, + "step": 5680 + }, + { + "epoch": 1.817726109215017, + "grad_norm": 0.5329631567001343, + "learning_rate": 4.0296846079514114e-06, + "loss": 0.5223, + "step": 5681 + }, + { + "epoch": 1.8180460750853242, + "grad_norm": 0.5038538575172424, + "learning_rate": 4.027858274881235e-06, + "loss": 0.4667, + "step": 5682 + }, + { + "epoch": 1.8183660409556315, + "grad_norm": 0.5360050201416016, + "learning_rate": 4.026032076599681e-06, + "loss": 0.4993, + "step": 5683 + }, + { + "epoch": 1.8186860068259385, + "grad_norm": 0.4917638599872589, + "learning_rate": 4.024206013359952e-06, + "loss": 0.4794, + "step": 5684 + }, + { + "epoch": 1.8190059726962458, + "grad_norm": 0.6024853587150574, + "learning_rate": 4.02238008541523e-06, + "loss": 0.5482, + "step": 5685 + }, + { + "epoch": 1.819325938566553, + "grad_norm": 0.5780578255653381, + "learning_rate": 4.020554293018688e-06, + "loss": 0.4987, + "step": 5686 + }, + { + "epoch": 1.81964590443686, + "grad_norm": 0.5435405969619751, + "learning_rate": 4.018728636423469e-06, + "loss": 0.452, + "step": 5687 + }, + { + "epoch": 1.8199658703071673, + "grad_norm": 0.5363346338272095, + "learning_rate": 4.016903115882707e-06, + "loss": 0.4646, + "step": 5688 + }, + { + "epoch": 1.8202858361774745, + "grad_norm": 0.5494429469108582, + "learning_rate": 4.015077731649511e-06, + "loss": 0.5383, + "step": 5689 + }, + { + "epoch": 1.8206058020477816, + "grad_norm": 0.5325194597244263, + "learning_rate": 4.013252483976972e-06, + "loss": 0.5186, + "step": 5690 + }, + { + "epoch": 1.8209257679180886, + "grad_norm": 0.5208801627159119, + "learning_rate": 4.011427373118163e-06, + "loss": 0.512, + "step": 5691 + }, + { + "epoch": 1.821245733788396, + "grad_norm": 0.4978253245353699, + "learning_rate": 4.009602399326137e-06, + "loss": 0.4225, + "step": 5692 + }, + { + "epoch": 1.821565699658703, + "grad_norm": 0.48743247985839844, + "learning_rate": 4.00777756285393e-06, + "loss": 0.4746, + "step": 5693 + }, + { + "epoch": 1.82188566552901, + "grad_norm": 0.5696543455123901, + "learning_rate": 4.005952863954559e-06, + "loss": 0.5713, + "step": 5694 + }, + { + "epoch": 1.8222056313993176, + "grad_norm": 0.4884602725505829, + "learning_rate": 4.004128302881018e-06, + "loss": 0.469, + "step": 5695 + }, + { + "epoch": 1.8225255972696246, + "grad_norm": 0.5204737782478333, + "learning_rate": 4.002303879886288e-06, + "loss": 0.4507, + "step": 5696 + }, + { + "epoch": 1.8228455631399316, + "grad_norm": 0.49271154403686523, + "learning_rate": 4.000479595223322e-06, + "loss": 0.4515, + "step": 5697 + }, + { + "epoch": 1.8231655290102389, + "grad_norm": 0.519041895866394, + "learning_rate": 3.998655449145064e-06, + "loss": 0.5166, + "step": 5698 + }, + { + "epoch": 1.8234854948805461, + "grad_norm": 0.5471581816673279, + "learning_rate": 3.9968314419044334e-06, + "loss": 0.4911, + "step": 5699 + }, + { + "epoch": 1.8238054607508531, + "grad_norm": 0.6149153113365173, + "learning_rate": 3.9950075737543296e-06, + "loss": 0.5139, + "step": 5700 + }, + { + "epoch": 1.8241254266211604, + "grad_norm": 0.5749152302742004, + "learning_rate": 3.9931838449476355e-06, + "loss": 0.426, + "step": 5701 + }, + { + "epoch": 1.8244453924914676, + "grad_norm": 0.5508496165275574, + "learning_rate": 3.991360255737211e-06, + "loss": 0.5323, + "step": 5702 + }, + { + "epoch": 1.8247653583617747, + "grad_norm": 0.5065162181854248, + "learning_rate": 3.989536806375901e-06, + "loss": 0.4695, + "step": 5703 + }, + { + "epoch": 1.825085324232082, + "grad_norm": 0.5152005553245544, + "learning_rate": 3.98771349711653e-06, + "loss": 0.4943, + "step": 5704 + }, + { + "epoch": 1.8254052901023892, + "grad_norm": 0.521493136882782, + "learning_rate": 3.985890328211899e-06, + "loss": 0.4214, + "step": 5705 + }, + { + "epoch": 1.8257252559726962, + "grad_norm": 0.5740010142326355, + "learning_rate": 3.984067299914792e-06, + "loss": 0.5061, + "step": 5706 + }, + { + "epoch": 1.8260452218430034, + "grad_norm": 0.5458466410636902, + "learning_rate": 3.982244412477979e-06, + "loss": 0.4617, + "step": 5707 + }, + { + "epoch": 1.8263651877133107, + "grad_norm": 0.5201452374458313, + "learning_rate": 3.980421666154202e-06, + "loss": 0.4686, + "step": 5708 + }, + { + "epoch": 1.8266851535836177, + "grad_norm": 0.530581533908844, + "learning_rate": 3.978599061196188e-06, + "loss": 0.4927, + "step": 5709 + }, + { + "epoch": 1.827005119453925, + "grad_norm": 0.4961286187171936, + "learning_rate": 3.976776597856643e-06, + "loss": 0.4274, + "step": 5710 + }, + { + "epoch": 1.8273250853242322, + "grad_norm": 0.5156742930412292, + "learning_rate": 3.974954276388251e-06, + "loss": 0.5662, + "step": 5711 + }, + { + "epoch": 1.8276450511945392, + "grad_norm": 0.5241951942443848, + "learning_rate": 3.973132097043685e-06, + "loss": 0.4626, + "step": 5712 + }, + { + "epoch": 1.8279650170648463, + "grad_norm": 0.5877885818481445, + "learning_rate": 3.971310060075588e-06, + "loss": 0.5165, + "step": 5713 + }, + { + "epoch": 1.8282849829351537, + "grad_norm": 0.5485692024230957, + "learning_rate": 3.969488165736589e-06, + "loss": 0.4743, + "step": 5714 + }, + { + "epoch": 1.8286049488054608, + "grad_norm": 0.598365843296051, + "learning_rate": 3.967666414279294e-06, + "loss": 0.5029, + "step": 5715 + }, + { + "epoch": 1.8289249146757678, + "grad_norm": 0.5118568539619446, + "learning_rate": 3.9658448059562934e-06, + "loss": 0.5022, + "step": 5716 + }, + { + "epoch": 1.829244880546075, + "grad_norm": 0.5245277881622314, + "learning_rate": 3.9640233410201555e-06, + "loss": 0.4935, + "step": 5717 + }, + { + "epoch": 1.8295648464163823, + "grad_norm": 0.49036654829978943, + "learning_rate": 3.962202019723427e-06, + "loss": 0.4344, + "step": 5718 + }, + { + "epoch": 1.8298848122866893, + "grad_norm": 0.5216358304023743, + "learning_rate": 3.960380842318638e-06, + "loss": 0.545, + "step": 5719 + }, + { + "epoch": 1.8302047781569966, + "grad_norm": 0.5128059387207031, + "learning_rate": 3.958559809058294e-06, + "loss": 0.4777, + "step": 5720 + }, + { + "epoch": 1.8305247440273038, + "grad_norm": 0.543674111366272, + "learning_rate": 3.956738920194886e-06, + "loss": 0.4862, + "step": 5721 + }, + { + "epoch": 1.8308447098976108, + "grad_norm": 0.5424734950065613, + "learning_rate": 3.954918175980882e-06, + "loss": 0.4828, + "step": 5722 + }, + { + "epoch": 1.831164675767918, + "grad_norm": 0.553587019443512, + "learning_rate": 3.95309757666873e-06, + "loss": 0.5023, + "step": 5723 + }, + { + "epoch": 1.8314846416382253, + "grad_norm": 0.48538005352020264, + "learning_rate": 3.9512771225108595e-06, + "loss": 0.442, + "step": 5724 + }, + { + "epoch": 1.8318046075085324, + "grad_norm": 0.5084359049797058, + "learning_rate": 3.949456813759677e-06, + "loss": 0.4686, + "step": 5725 + }, + { + "epoch": 1.8321245733788396, + "grad_norm": 0.5093129277229309, + "learning_rate": 3.947636650667569e-06, + "loss": 0.5013, + "step": 5726 + }, + { + "epoch": 1.8324445392491469, + "grad_norm": 0.5247644186019897, + "learning_rate": 3.9458166334869085e-06, + "loss": 0.5435, + "step": 5727 + }, + { + "epoch": 1.8327645051194539, + "grad_norm": 0.5353206396102905, + "learning_rate": 3.943996762470039e-06, + "loss": 0.4698, + "step": 5728 + }, + { + "epoch": 1.8330844709897611, + "grad_norm": 0.5606423020362854, + "learning_rate": 3.942177037869289e-06, + "loss": 0.4671, + "step": 5729 + }, + { + "epoch": 1.8334044368600684, + "grad_norm": 0.5476802587509155, + "learning_rate": 3.9403574599369645e-06, + "loss": 0.5291, + "step": 5730 + }, + { + "epoch": 1.8337244027303754, + "grad_norm": 0.5585899353027344, + "learning_rate": 3.938538028925353e-06, + "loss": 0.5022, + "step": 5731 + }, + { + "epoch": 1.8340443686006824, + "grad_norm": 0.49875614047050476, + "learning_rate": 3.936718745086721e-06, + "loss": 0.5021, + "step": 5732 + }, + { + "epoch": 1.83436433447099, + "grad_norm": 0.4922947883605957, + "learning_rate": 3.934899608673315e-06, + "loss": 0.458, + "step": 5733 + }, + { + "epoch": 1.834684300341297, + "grad_norm": 0.5266348719596863, + "learning_rate": 3.9330806199373595e-06, + "loss": 0.4972, + "step": 5734 + }, + { + "epoch": 1.835004266211604, + "grad_norm": 0.45093604922294617, + "learning_rate": 3.931261779131058e-06, + "loss": 0.4192, + "step": 5735 + }, + { + "epoch": 1.8353242320819114, + "grad_norm": 0.5231369137763977, + "learning_rate": 3.929443086506597e-06, + "loss": 0.4642, + "step": 5736 + }, + { + "epoch": 1.8356441979522184, + "grad_norm": 0.5847294330596924, + "learning_rate": 3.9276245423161396e-06, + "loss": 0.5085, + "step": 5737 + }, + { + "epoch": 1.8359641638225255, + "grad_norm": 0.5863973498344421, + "learning_rate": 3.925806146811832e-06, + "loss": 0.5286, + "step": 5738 + }, + { + "epoch": 1.8362841296928327, + "grad_norm": 0.5376518368721008, + "learning_rate": 3.9239879002457935e-06, + "loss": 0.4939, + "step": 5739 + }, + { + "epoch": 1.83660409556314, + "grad_norm": 0.5330129861831665, + "learning_rate": 3.9221698028701254e-06, + "loss": 0.5003, + "step": 5740 + }, + { + "epoch": 1.836924061433447, + "grad_norm": 0.5517658591270447, + "learning_rate": 3.920351854936911e-06, + "loss": 0.484, + "step": 5741 + }, + { + "epoch": 1.8372440273037542, + "grad_norm": 0.49880000948905945, + "learning_rate": 3.918534056698211e-06, + "loss": 0.4407, + "step": 5742 + }, + { + "epoch": 1.8375639931740615, + "grad_norm": 0.5193721652030945, + "learning_rate": 3.916716408406066e-06, + "loss": 0.45, + "step": 5743 + }, + { + "epoch": 1.8378839590443685, + "grad_norm": 0.5570769906044006, + "learning_rate": 3.914898910312495e-06, + "loss": 0.5672, + "step": 5744 + }, + { + "epoch": 1.8382039249146758, + "grad_norm": 0.45104196667671204, + "learning_rate": 3.913081562669492e-06, + "loss": 0.4154, + "step": 5745 + }, + { + "epoch": 1.838523890784983, + "grad_norm": 0.5590640902519226, + "learning_rate": 3.911264365729041e-06, + "loss": 0.5161, + "step": 5746 + }, + { + "epoch": 1.83884385665529, + "grad_norm": 0.49155643582344055, + "learning_rate": 3.909447319743095e-06, + "loss": 0.4712, + "step": 5747 + }, + { + "epoch": 1.8391638225255973, + "grad_norm": 0.5074732303619385, + "learning_rate": 3.9076304249635905e-06, + "loss": 0.5029, + "step": 5748 + }, + { + "epoch": 1.8394837883959045, + "grad_norm": 0.49026942253112793, + "learning_rate": 3.9058136816424405e-06, + "loss": 0.4676, + "step": 5749 + }, + { + "epoch": 1.8398037542662116, + "grad_norm": 0.5693995952606201, + "learning_rate": 3.90399709003154e-06, + "loss": 0.4981, + "step": 5750 + }, + { + "epoch": 1.8401237201365188, + "grad_norm": 0.6498878002166748, + "learning_rate": 3.902180650382762e-06, + "loss": 0.5269, + "step": 5751 + }, + { + "epoch": 1.840443686006826, + "grad_norm": 0.5374650955200195, + "learning_rate": 3.900364362947957e-06, + "loss": 0.4717, + "step": 5752 + }, + { + "epoch": 1.840763651877133, + "grad_norm": 0.4944017827510834, + "learning_rate": 3.898548227978958e-06, + "loss": 0.5311, + "step": 5753 + }, + { + "epoch": 1.8410836177474401, + "grad_norm": 0.48711761832237244, + "learning_rate": 3.89673224572757e-06, + "loss": 0.4029, + "step": 5754 + }, + { + "epoch": 1.8414035836177476, + "grad_norm": 0.5856059193611145, + "learning_rate": 3.8949164164455825e-06, + "loss": 0.517, + "step": 5755 + }, + { + "epoch": 1.8417235494880546, + "grad_norm": 0.5151647329330444, + "learning_rate": 3.893100740384766e-06, + "loss": 0.4727, + "step": 5756 + }, + { + "epoch": 1.8420435153583616, + "grad_norm": 0.5710545778274536, + "learning_rate": 3.891285217796862e-06, + "loss": 0.5521, + "step": 5757 + }, + { + "epoch": 1.8423634812286689, + "grad_norm": 0.5346451997756958, + "learning_rate": 3.889469848933598e-06, + "loss": 0.4912, + "step": 5758 + }, + { + "epoch": 1.8426834470989761, + "grad_norm": 0.5003721117973328, + "learning_rate": 3.887654634046674e-06, + "loss": 0.4756, + "step": 5759 + }, + { + "epoch": 1.8430034129692832, + "grad_norm": 0.593546986579895, + "learning_rate": 3.885839573387772e-06, + "loss": 0.47, + "step": 5760 + }, + { + "epoch": 1.8433233788395904, + "grad_norm": 0.5888879299163818, + "learning_rate": 3.884024667208556e-06, + "loss": 0.4706, + "step": 5761 + }, + { + "epoch": 1.8436433447098977, + "grad_norm": 0.5248382687568665, + "learning_rate": 3.8822099157606605e-06, + "loss": 0.5013, + "step": 5762 + }, + { + "epoch": 1.8439633105802047, + "grad_norm": 0.48654475808143616, + "learning_rate": 3.880395319295707e-06, + "loss": 0.4886, + "step": 5763 + }, + { + "epoch": 1.844283276450512, + "grad_norm": 0.5649202466011047, + "learning_rate": 3.8785808780652875e-06, + "loss": 0.5276, + "step": 5764 + }, + { + "epoch": 1.8446032423208192, + "grad_norm": 0.5007580518722534, + "learning_rate": 3.876766592320977e-06, + "loss": 0.4502, + "step": 5765 + }, + { + "epoch": 1.8449232081911262, + "grad_norm": 0.5063193440437317, + "learning_rate": 3.874952462314333e-06, + "loss": 0.4951, + "step": 5766 + }, + { + "epoch": 1.8452431740614335, + "grad_norm": 0.5258777737617493, + "learning_rate": 3.873138488296883e-06, + "loss": 0.4848, + "step": 5767 + }, + { + "epoch": 1.8455631399317407, + "grad_norm": 0.5359113216400146, + "learning_rate": 3.871324670520137e-06, + "loss": 0.553, + "step": 5768 + }, + { + "epoch": 1.8458831058020477, + "grad_norm": 0.4946892559528351, + "learning_rate": 3.869511009235582e-06, + "loss": 0.536, + "step": 5769 + }, + { + "epoch": 1.846203071672355, + "grad_norm": 0.49602842330932617, + "learning_rate": 3.867697504694685e-06, + "loss": 0.4712, + "step": 5770 + }, + { + "epoch": 1.8465230375426622, + "grad_norm": 0.4690818190574646, + "learning_rate": 3.865884157148894e-06, + "loss": 0.4463, + "step": 5771 + }, + { + "epoch": 1.8468430034129693, + "grad_norm": 0.5733300447463989, + "learning_rate": 3.8640709668496265e-06, + "loss": 0.5591, + "step": 5772 + }, + { + "epoch": 1.8471629692832765, + "grad_norm": 0.45847299695014954, + "learning_rate": 3.862257934048289e-06, + "loss": 0.4772, + "step": 5773 + }, + { + "epoch": 1.8474829351535837, + "grad_norm": 0.5056205987930298, + "learning_rate": 3.860445058996255e-06, + "loss": 0.5491, + "step": 5774 + }, + { + "epoch": 1.8478029010238908, + "grad_norm": 0.5060153603553772, + "learning_rate": 3.858632341944884e-06, + "loss": 0.4892, + "step": 5775 + }, + { + "epoch": 1.8481228668941978, + "grad_norm": 0.4729257822036743, + "learning_rate": 3.856819783145514e-06, + "loss": 0.4729, + "step": 5776 + }, + { + "epoch": 1.8484428327645053, + "grad_norm": 0.5032588243484497, + "learning_rate": 3.855007382849455e-06, + "loss": 0.5282, + "step": 5777 + }, + { + "epoch": 1.8487627986348123, + "grad_norm": 0.5291029214859009, + "learning_rate": 3.853195141308001e-06, + "loss": 0.4616, + "step": 5778 + }, + { + "epoch": 1.8490827645051193, + "grad_norm": 0.5785648822784424, + "learning_rate": 3.851383058772418e-06, + "loss": 0.5161, + "step": 5779 + }, + { + "epoch": 1.8494027303754266, + "grad_norm": 0.4976552426815033, + "learning_rate": 3.849571135493958e-06, + "loss": 0.4668, + "step": 5780 + }, + { + "epoch": 1.8497226962457338, + "grad_norm": 0.5364990234375, + "learning_rate": 3.847759371723844e-06, + "loss": 0.4914, + "step": 5781 + }, + { + "epoch": 1.8500426621160408, + "grad_norm": 0.5276072025299072, + "learning_rate": 3.845947767713278e-06, + "loss": 0.4959, + "step": 5782 + }, + { + "epoch": 1.850362627986348, + "grad_norm": 0.5432918071746826, + "learning_rate": 3.844136323713444e-06, + "loss": 0.5032, + "step": 5783 + }, + { + "epoch": 1.8506825938566553, + "grad_norm": 0.5424608588218689, + "learning_rate": 3.842325039975497e-06, + "loss": 0.4803, + "step": 5784 + }, + { + "epoch": 1.8510025597269624, + "grad_norm": 0.5053834319114685, + "learning_rate": 3.840513916750577e-06, + "loss": 0.4765, + "step": 5785 + }, + { + "epoch": 1.8513225255972696, + "grad_norm": 0.5041948556900024, + "learning_rate": 3.838702954289798e-06, + "loss": 0.4887, + "step": 5786 + }, + { + "epoch": 1.8516424914675769, + "grad_norm": 0.5387802720069885, + "learning_rate": 3.836892152844251e-06, + "loss": 0.4919, + "step": 5787 + }, + { + "epoch": 1.851962457337884, + "grad_norm": 0.49143558740615845, + "learning_rate": 3.835081512665007e-06, + "loss": 0.437, + "step": 5788 + }, + { + "epoch": 1.8522824232081911, + "grad_norm": 0.50626140832901, + "learning_rate": 3.833271034003111e-06, + "loss": 0.4643, + "step": 5789 + }, + { + "epoch": 1.8526023890784984, + "grad_norm": 0.4904909133911133, + "learning_rate": 3.83146071710959e-06, + "loss": 0.4946, + "step": 5790 + }, + { + "epoch": 1.8529223549488054, + "grad_norm": 0.5600085854530334, + "learning_rate": 3.829650562235448e-06, + "loss": 0.5578, + "step": 5791 + }, + { + "epoch": 1.8532423208191127, + "grad_norm": 0.5587226152420044, + "learning_rate": 3.827840569631662e-06, + "loss": 0.4908, + "step": 5792 + }, + { + "epoch": 1.85356228668942, + "grad_norm": 0.49955758452415466, + "learning_rate": 3.826030739549193e-06, + "loss": 0.4517, + "step": 5793 + }, + { + "epoch": 1.853882252559727, + "grad_norm": 0.5089410543441772, + "learning_rate": 3.824221072238972e-06, + "loss": 0.5098, + "step": 5794 + }, + { + "epoch": 1.854202218430034, + "grad_norm": 0.5281291007995605, + "learning_rate": 3.8224115679519144e-06, + "loss": 0.4656, + "step": 5795 + }, + { + "epoch": 1.8545221843003414, + "grad_norm": 0.4958762526512146, + "learning_rate": 3.820602226938911e-06, + "loss": 0.5027, + "step": 5796 + }, + { + "epoch": 1.8548421501706485, + "grad_norm": 0.4767692983150482, + "learning_rate": 3.818793049450826e-06, + "loss": 0.4702, + "step": 5797 + }, + { + "epoch": 1.8551621160409555, + "grad_norm": 0.45611903071403503, + "learning_rate": 3.8169840357385075e-06, + "loss": 0.4344, + "step": 5798 + }, + { + "epoch": 1.855482081911263, + "grad_norm": 0.5477572679519653, + "learning_rate": 3.8151751860527734e-06, + "loss": 0.4854, + "step": 5799 + }, + { + "epoch": 1.85580204778157, + "grad_norm": 0.5560097098350525, + "learning_rate": 3.813366500644426e-06, + "loss": 0.5266, + "step": 5800 + }, + { + "epoch": 1.856122013651877, + "grad_norm": 0.5477871894836426, + "learning_rate": 3.811557979764242e-06, + "loss": 0.5198, + "step": 5801 + }, + { + "epoch": 1.8564419795221843, + "grad_norm": 0.5169461965560913, + "learning_rate": 3.809749623662973e-06, + "loss": 0.4519, + "step": 5802 + }, + { + "epoch": 1.8567619453924915, + "grad_norm": 0.544291615486145, + "learning_rate": 3.807941432591351e-06, + "loss": 0.5343, + "step": 5803 + }, + { + "epoch": 1.8570819112627985, + "grad_norm": 0.5299692153930664, + "learning_rate": 3.806133406800082e-06, + "loss": 0.5507, + "step": 5804 + }, + { + "epoch": 1.8574018771331058, + "grad_norm": 0.5153736472129822, + "learning_rate": 3.8043255465398533e-06, + "loss": 0.4455, + "step": 5805 + }, + { + "epoch": 1.857721843003413, + "grad_norm": 0.49252840876579285, + "learning_rate": 3.802517852061327e-06, + "loss": 0.5033, + "step": 5806 + }, + { + "epoch": 1.85804180887372, + "grad_norm": 0.4893689453601837, + "learning_rate": 3.80071032361514e-06, + "loss": 0.4719, + "step": 5807 + }, + { + "epoch": 1.8583617747440273, + "grad_norm": 0.5370251536369324, + "learning_rate": 3.79890296145191e-06, + "loss": 0.4967, + "step": 5808 + }, + { + "epoch": 1.8586817406143346, + "grad_norm": 0.5664129853248596, + "learning_rate": 3.797095765822228e-06, + "loss": 0.514, + "step": 5809 + }, + { + "epoch": 1.8590017064846416, + "grad_norm": 0.5876678824424744, + "learning_rate": 3.795288736976666e-06, + "loss": 0.5098, + "step": 5810 + }, + { + "epoch": 1.8593216723549488, + "grad_norm": 0.5216983556747437, + "learning_rate": 3.793481875165771e-06, + "loss": 0.461, + "step": 5811 + }, + { + "epoch": 1.859641638225256, + "grad_norm": 0.4685456156730652, + "learning_rate": 3.791675180640064e-06, + "loss": 0.4429, + "step": 5812 + }, + { + "epoch": 1.859961604095563, + "grad_norm": 0.5082848072052002, + "learning_rate": 3.789868653650046e-06, + "loss": 0.5158, + "step": 5813 + }, + { + "epoch": 1.8602815699658704, + "grad_norm": 0.4984305500984192, + "learning_rate": 3.7880622944461975e-06, + "loss": 0.4712, + "step": 5814 + }, + { + "epoch": 1.8606015358361776, + "grad_norm": 0.5433828830718994, + "learning_rate": 3.7862561032789692e-06, + "loss": 0.5411, + "step": 5815 + }, + { + "epoch": 1.8609215017064846, + "grad_norm": 0.5483769774436951, + "learning_rate": 3.7844500803987925e-06, + "loss": 0.42, + "step": 5816 + }, + { + "epoch": 1.8612414675767917, + "grad_norm": 0.5509781241416931, + "learning_rate": 3.7826442260560737e-06, + "loss": 0.5298, + "step": 5817 + }, + { + "epoch": 1.8615614334470991, + "grad_norm": 0.5578811168670654, + "learning_rate": 3.7808385405011965e-06, + "loss": 0.5022, + "step": 5818 + }, + { + "epoch": 1.8618813993174061, + "grad_norm": 0.5305883884429932, + "learning_rate": 3.779033023984524e-06, + "loss": 0.4844, + "step": 5819 + }, + { + "epoch": 1.8622013651877132, + "grad_norm": 0.500226616859436, + "learning_rate": 3.7772276767563916e-06, + "loss": 0.4359, + "step": 5820 + }, + { + "epoch": 1.8625213310580204, + "grad_norm": 0.5373531579971313, + "learning_rate": 3.775422499067113e-06, + "loss": 0.5137, + "step": 5821 + }, + { + "epoch": 1.8628412969283277, + "grad_norm": 0.5030234456062317, + "learning_rate": 3.7736174911669775e-06, + "loss": 0.4959, + "step": 5822 + }, + { + "epoch": 1.8631612627986347, + "grad_norm": 0.49924901127815247, + "learning_rate": 3.7718126533062504e-06, + "loss": 0.5021, + "step": 5823 + }, + { + "epoch": 1.863481228668942, + "grad_norm": 0.5491090416908264, + "learning_rate": 3.7700079857351797e-06, + "loss": 0.5029, + "step": 5824 + }, + { + "epoch": 1.8638011945392492, + "grad_norm": 0.5774345993995667, + "learning_rate": 3.76820348870398e-06, + "loss": 0.5108, + "step": 5825 + }, + { + "epoch": 1.8641211604095562, + "grad_norm": 0.5084494948387146, + "learning_rate": 3.7663991624628495e-06, + "loss": 0.448, + "step": 5826 + }, + { + "epoch": 1.8644411262798635, + "grad_norm": 0.5092008113861084, + "learning_rate": 3.7645950072619573e-06, + "loss": 0.4599, + "step": 5827 + }, + { + "epoch": 1.8647610921501707, + "grad_norm": 0.5794777274131775, + "learning_rate": 3.762791023351454e-06, + "loss": 0.542, + "step": 5828 + }, + { + "epoch": 1.8650810580204777, + "grad_norm": 0.4939880967140198, + "learning_rate": 3.7609872109814648e-06, + "loss": 0.4244, + "step": 5829 + }, + { + "epoch": 1.865401023890785, + "grad_norm": 0.4751640558242798, + "learning_rate": 3.7591835704020885e-06, + "loss": 0.4604, + "step": 5830 + }, + { + "epoch": 1.8657209897610922, + "grad_norm": 0.4862428903579712, + "learning_rate": 3.7573801018634043e-06, + "loss": 0.4439, + "step": 5831 + }, + { + "epoch": 1.8660409556313993, + "grad_norm": 0.5053917169570923, + "learning_rate": 3.7555768056154623e-06, + "loss": 0.5173, + "step": 5832 + }, + { + "epoch": 1.8663609215017065, + "grad_norm": 0.47630491852760315, + "learning_rate": 3.7537736819082926e-06, + "loss": 0.44, + "step": 5833 + }, + { + "epoch": 1.8666808873720138, + "grad_norm": 0.5234889388084412, + "learning_rate": 3.751970730991903e-06, + "loss": 0.5028, + "step": 5834 + }, + { + "epoch": 1.8670008532423208, + "grad_norm": 0.532590389251709, + "learning_rate": 3.750167953116272e-06, + "loss": 0.4931, + "step": 5835 + }, + { + "epoch": 1.867320819112628, + "grad_norm": 0.48530295491218567, + "learning_rate": 3.748365348531359e-06, + "loss": 0.4657, + "step": 5836 + }, + { + "epoch": 1.8676407849829353, + "grad_norm": 0.551383376121521, + "learning_rate": 3.7465629174870955e-06, + "loss": 0.51, + "step": 5837 + }, + { + "epoch": 1.8679607508532423, + "grad_norm": 0.4987407624721527, + "learning_rate": 3.7447606602333895e-06, + "loss": 0.4767, + "step": 5838 + }, + { + "epoch": 1.8682807167235493, + "grad_norm": 0.49918514490127563, + "learning_rate": 3.7429585770201314e-06, + "loss": 0.4366, + "step": 5839 + }, + { + "epoch": 1.8686006825938568, + "grad_norm": 0.5327936410903931, + "learning_rate": 3.7411566680971765e-06, + "loss": 0.4906, + "step": 5840 + }, + { + "epoch": 1.8689206484641638, + "grad_norm": 0.49172770977020264, + "learning_rate": 3.739354933714366e-06, + "loss": 0.4405, + "step": 5841 + }, + { + "epoch": 1.8692406143344709, + "grad_norm": 0.5563710927963257, + "learning_rate": 3.7375533741215086e-06, + "loss": 0.5039, + "step": 5842 + }, + { + "epoch": 1.869560580204778, + "grad_norm": 0.5064439177513123, + "learning_rate": 3.7357519895683948e-06, + "loss": 0.4778, + "step": 5843 + }, + { + "epoch": 1.8698805460750854, + "grad_norm": 0.5443150401115417, + "learning_rate": 3.73395078030479e-06, + "loss": 0.5132, + "step": 5844 + }, + { + "epoch": 1.8702005119453924, + "grad_norm": 0.5381937623023987, + "learning_rate": 3.732149746580431e-06, + "loss": 0.5251, + "step": 5845 + }, + { + "epoch": 1.8705204778156996, + "grad_norm": 0.5324935913085938, + "learning_rate": 3.7303488886450367e-06, + "loss": 0.4619, + "step": 5846 + }, + { + "epoch": 1.8708404436860069, + "grad_norm": 0.49195364117622375, + "learning_rate": 3.728548206748295e-06, + "loss": 0.5433, + "step": 5847 + }, + { + "epoch": 1.871160409556314, + "grad_norm": 0.49643033742904663, + "learning_rate": 3.7267477011398735e-06, + "loss": 0.426, + "step": 5848 + }, + { + "epoch": 1.8714803754266212, + "grad_norm": 0.5382016897201538, + "learning_rate": 3.724947372069416e-06, + "loss": 0.4352, + "step": 5849 + }, + { + "epoch": 1.8718003412969284, + "grad_norm": 0.5211527943611145, + "learning_rate": 3.7231472197865394e-06, + "loss": 0.46, + "step": 5850 + }, + { + "epoch": 1.8721203071672354, + "grad_norm": 0.5157228708267212, + "learning_rate": 3.721347244540837e-06, + "loss": 0.5262, + "step": 5851 + }, + { + "epoch": 1.8724402730375427, + "grad_norm": 0.48551246523857117, + "learning_rate": 3.719547446581876e-06, + "loss": 0.4761, + "step": 5852 + }, + { + "epoch": 1.87276023890785, + "grad_norm": 0.47866228222846985, + "learning_rate": 3.7177478261592024e-06, + "loss": 0.4987, + "step": 5853 + }, + { + "epoch": 1.873080204778157, + "grad_norm": 0.4556387960910797, + "learning_rate": 3.7159483835223374e-06, + "loss": 0.4724, + "step": 5854 + }, + { + "epoch": 1.8734001706484642, + "grad_norm": 0.5026599168777466, + "learning_rate": 3.7141491189207716e-06, + "loss": 0.47, + "step": 5855 + }, + { + "epoch": 1.8737201365187715, + "grad_norm": 0.5129727721214294, + "learning_rate": 3.712350032603979e-06, + "loss": 0.5134, + "step": 5856 + }, + { + "epoch": 1.8740401023890785, + "grad_norm": 0.4955362379550934, + "learning_rate": 3.710551124821401e-06, + "loss": 0.4959, + "step": 5857 + }, + { + "epoch": 1.8743600682593855, + "grad_norm": 0.4802592396736145, + "learning_rate": 3.7087523958224624e-06, + "loss": 0.499, + "step": 5858 + }, + { + "epoch": 1.874680034129693, + "grad_norm": 0.5008774995803833, + "learning_rate": 3.706953845856558e-06, + "loss": 0.4934, + "step": 5859 + }, + { + "epoch": 1.875, + "grad_norm": 0.5445345044136047, + "learning_rate": 3.705155475173058e-06, + "loss": 0.4794, + "step": 5860 + }, + { + "epoch": 1.875319965870307, + "grad_norm": 0.5107492208480835, + "learning_rate": 3.7033572840213093e-06, + "loss": 0.4789, + "step": 5861 + }, + { + "epoch": 1.8756399317406145, + "grad_norm": 0.5111907124519348, + "learning_rate": 3.7015592726506318e-06, + "loss": 0.4625, + "step": 5862 + }, + { + "epoch": 1.8759598976109215, + "grad_norm": 0.5461314916610718, + "learning_rate": 3.699761441310324e-06, + "loss": 0.5222, + "step": 5863 + }, + { + "epoch": 1.8762798634812285, + "grad_norm": 0.4889307916164398, + "learning_rate": 3.697963790249657e-06, + "loss": 0.4288, + "step": 5864 + }, + { + "epoch": 1.8765998293515358, + "grad_norm": 0.4892370104789734, + "learning_rate": 3.6961663197178767e-06, + "loss": 0.4727, + "step": 5865 + }, + { + "epoch": 1.876919795221843, + "grad_norm": 0.4976430833339691, + "learning_rate": 3.6943690299642055e-06, + "loss": 0.5502, + "step": 5866 + }, + { + "epoch": 1.87723976109215, + "grad_norm": 0.5494089722633362, + "learning_rate": 3.6925719212378364e-06, + "loss": 0.4605, + "step": 5867 + }, + { + "epoch": 1.8775597269624573, + "grad_norm": 0.5490610003471375, + "learning_rate": 3.690774993787945e-06, + "loss": 0.5349, + "step": 5868 + }, + { + "epoch": 1.8778796928327646, + "grad_norm": 0.5048588514328003, + "learning_rate": 3.6889782478636773e-06, + "loss": 0.4975, + "step": 5869 + }, + { + "epoch": 1.8781996587030716, + "grad_norm": 0.5531996488571167, + "learning_rate": 3.6871816837141506e-06, + "loss": 0.4456, + "step": 5870 + }, + { + "epoch": 1.8785196245733788, + "grad_norm": 0.5299012660980225, + "learning_rate": 3.6853853015884645e-06, + "loss": 0.4258, + "step": 5871 + }, + { + "epoch": 1.878839590443686, + "grad_norm": 0.5071423053741455, + "learning_rate": 3.683589101735685e-06, + "loss": 0.5059, + "step": 5872 + }, + { + "epoch": 1.8791595563139931, + "grad_norm": 0.5356152057647705, + "learning_rate": 3.6817930844048622e-06, + "loss": 0.4685, + "step": 5873 + }, + { + "epoch": 1.8794795221843004, + "grad_norm": 0.5635944604873657, + "learning_rate": 3.6799972498450145e-06, + "loss": 0.5245, + "step": 5874 + }, + { + "epoch": 1.8797994880546076, + "grad_norm": 0.5151473879814148, + "learning_rate": 3.678201598305135e-06, + "loss": 0.4866, + "step": 5875 + }, + { + "epoch": 1.8801194539249146, + "grad_norm": 0.4848635792732239, + "learning_rate": 3.6764061300341956e-06, + "loss": 0.4681, + "step": 5876 + }, + { + "epoch": 1.880439419795222, + "grad_norm": 0.5050166845321655, + "learning_rate": 3.6746108452811347e-06, + "loss": 0.5087, + "step": 5877 + }, + { + "epoch": 1.8807593856655291, + "grad_norm": 0.48006412386894226, + "learning_rate": 3.6728157442948786e-06, + "loss": 0.4778, + "step": 5878 + }, + { + "epoch": 1.8810793515358362, + "grad_norm": 0.5760488510131836, + "learning_rate": 3.6710208273243157e-06, + "loss": 0.5432, + "step": 5879 + }, + { + "epoch": 1.8813993174061432, + "grad_norm": 0.5436517000198364, + "learning_rate": 3.6692260946183133e-06, + "loss": 0.4493, + "step": 5880 + }, + { + "epoch": 1.8817192832764507, + "grad_norm": 0.5405221581459045, + "learning_rate": 3.6674315464257142e-06, + "loss": 0.4942, + "step": 5881 + }, + { + "epoch": 1.8820392491467577, + "grad_norm": 0.49750787019729614, + "learning_rate": 3.665637182995332e-06, + "loss": 0.4563, + "step": 5882 + }, + { + "epoch": 1.8823592150170647, + "grad_norm": 0.539986252784729, + "learning_rate": 3.663843004575962e-06, + "loss": 0.5238, + "step": 5883 + }, + { + "epoch": 1.882679180887372, + "grad_norm": 0.5363343358039856, + "learning_rate": 3.662049011416368e-06, + "loss": 0.5187, + "step": 5884 + }, + { + "epoch": 1.8829991467576792, + "grad_norm": 0.5319176316261292, + "learning_rate": 3.6602552037652855e-06, + "loss": 0.4631, + "step": 5885 + }, + { + "epoch": 1.8833191126279862, + "grad_norm": 0.4889177680015564, + "learning_rate": 3.658461581871431e-06, + "loss": 0.4641, + "step": 5886 + }, + { + "epoch": 1.8836390784982935, + "grad_norm": 0.5030321478843689, + "learning_rate": 3.6566681459834917e-06, + "loss": 0.4517, + "step": 5887 + }, + { + "epoch": 1.8839590443686007, + "grad_norm": 0.5236909985542297, + "learning_rate": 3.6548748963501324e-06, + "loss": 0.4772, + "step": 5888 + }, + { + "epoch": 1.8842790102389078, + "grad_norm": 0.5497915148735046, + "learning_rate": 3.6530818332199858e-06, + "loss": 0.5134, + "step": 5889 + }, + { + "epoch": 1.884598976109215, + "grad_norm": 0.563951849937439, + "learning_rate": 3.6512889568416643e-06, + "loss": 0.5396, + "step": 5890 + }, + { + "epoch": 1.8849189419795223, + "grad_norm": 0.5152337551116943, + "learning_rate": 3.649496267463749e-06, + "loss": 0.4871, + "step": 5891 + }, + { + "epoch": 1.8852389078498293, + "grad_norm": 0.44863662123680115, + "learning_rate": 3.6477037653348023e-06, + "loss": 0.3912, + "step": 5892 + }, + { + "epoch": 1.8855588737201365, + "grad_norm": 0.5549201965332031, + "learning_rate": 3.645911450703356e-06, + "loss": 0.5718, + "step": 5893 + }, + { + "epoch": 1.8858788395904438, + "grad_norm": 0.48657435178756714, + "learning_rate": 3.6441193238179152e-06, + "loss": 0.483, + "step": 5894 + }, + { + "epoch": 1.8861988054607508, + "grad_norm": 0.5181356072425842, + "learning_rate": 3.6423273849269624e-06, + "loss": 0.4646, + "step": 5895 + }, + { + "epoch": 1.886518771331058, + "grad_norm": 0.49020054936408997, + "learning_rate": 3.6405356342789496e-06, + "loss": 0.479, + "step": 5896 + }, + { + "epoch": 1.8868387372013653, + "grad_norm": 0.42209747433662415, + "learning_rate": 3.6387440721223067e-06, + "loss": 0.4726, + "step": 5897 + }, + { + "epoch": 1.8871587030716723, + "grad_norm": 0.48372170329093933, + "learning_rate": 3.6369526987054372e-06, + "loss": 0.4968, + "step": 5898 + }, + { + "epoch": 1.8874786689419796, + "grad_norm": 0.5298974514007568, + "learning_rate": 3.635161514276715e-06, + "loss": 0.5113, + "step": 5899 + }, + { + "epoch": 1.8877986348122868, + "grad_norm": 0.5127506852149963, + "learning_rate": 3.6333705190844917e-06, + "loss": 0.461, + "step": 5900 + }, + { + "epoch": 1.8881186006825939, + "grad_norm": 0.5458589196205139, + "learning_rate": 3.631579713377088e-06, + "loss": 0.468, + "step": 5901 + }, + { + "epoch": 1.8884385665529009, + "grad_norm": 0.5374716520309448, + "learning_rate": 3.6297890974028037e-06, + "loss": 0.5122, + "step": 5902 + }, + { + "epoch": 1.8887585324232083, + "grad_norm": 0.45911529660224915, + "learning_rate": 3.627998671409911e-06, + "loss": 0.4274, + "step": 5903 + }, + { + "epoch": 1.8890784982935154, + "grad_norm": 0.5461366772651672, + "learning_rate": 3.626208435646652e-06, + "loss": 0.5157, + "step": 5904 + }, + { + "epoch": 1.8893984641638224, + "grad_norm": 0.5829647779464722, + "learning_rate": 3.6244183903612473e-06, + "loss": 0.5233, + "step": 5905 + }, + { + "epoch": 1.8897184300341296, + "grad_norm": 0.4681430757045746, + "learning_rate": 3.6226285358018854e-06, + "loss": 0.4382, + "step": 5906 + }, + { + "epoch": 1.890038395904437, + "grad_norm": 0.5075001120567322, + "learning_rate": 3.6208388722167355e-06, + "loss": 0.5129, + "step": 5907 + }, + { + "epoch": 1.890358361774744, + "grad_norm": 0.515108585357666, + "learning_rate": 3.6190493998539356e-06, + "loss": 0.5281, + "step": 5908 + }, + { + "epoch": 1.8906783276450512, + "grad_norm": 0.49774298071861267, + "learning_rate": 3.6172601189615963e-06, + "loss": 0.4798, + "step": 5909 + }, + { + "epoch": 1.8909982935153584, + "grad_norm": 0.5024615526199341, + "learning_rate": 3.615471029787807e-06, + "loss": 0.4733, + "step": 5910 + }, + { + "epoch": 1.8913182593856654, + "grad_norm": 0.47436338663101196, + "learning_rate": 3.6136821325806227e-06, + "loss": 0.4588, + "step": 5911 + }, + { + "epoch": 1.8916382252559727, + "grad_norm": 0.5262418985366821, + "learning_rate": 3.611893427588079e-06, + "loss": 0.4737, + "step": 5912 + }, + { + "epoch": 1.89195819112628, + "grad_norm": 0.5761246681213379, + "learning_rate": 3.6101049150581836e-06, + "loss": 0.5106, + "step": 5913 + }, + { + "epoch": 1.892278156996587, + "grad_norm": 0.4981935918331146, + "learning_rate": 3.608316595238912e-06, + "loss": 0.4793, + "step": 5914 + }, + { + "epoch": 1.8925981228668942, + "grad_norm": 0.5088338255882263, + "learning_rate": 3.6065284683782197e-06, + "loss": 0.522, + "step": 5915 + }, + { + "epoch": 1.8929180887372015, + "grad_norm": 0.49246954917907715, + "learning_rate": 3.604740534724029e-06, + "loss": 0.4482, + "step": 5916 + }, + { + "epoch": 1.8932380546075085, + "grad_norm": 0.5981255173683167, + "learning_rate": 3.6029527945242436e-06, + "loss": 0.5036, + "step": 5917 + }, + { + "epoch": 1.8935580204778157, + "grad_norm": 0.6465089917182922, + "learning_rate": 3.601165248026736e-06, + "loss": 0.5968, + "step": 5918 + }, + { + "epoch": 1.893877986348123, + "grad_norm": 0.5026783347129822, + "learning_rate": 3.5993778954793467e-06, + "loss": 0.409, + "step": 5919 + }, + { + "epoch": 1.89419795221843, + "grad_norm": 0.501559317111969, + "learning_rate": 3.597590737129897e-06, + "loss": 0.4895, + "step": 5920 + }, + { + "epoch": 1.894517918088737, + "grad_norm": 0.5835351347923279, + "learning_rate": 3.5958037732261806e-06, + "loss": 0.5273, + "step": 5921 + }, + { + "epoch": 1.8948378839590445, + "grad_norm": 0.5803717970848083, + "learning_rate": 3.5940170040159595e-06, + "loss": 0.4824, + "step": 5922 + }, + { + "epoch": 1.8951578498293515, + "grad_norm": 0.4998316764831543, + "learning_rate": 3.5922304297469728e-06, + "loss": 0.4551, + "step": 5923 + }, + { + "epoch": 1.8954778156996586, + "grad_norm": 0.5349484086036682, + "learning_rate": 3.5904440506669303e-06, + "loss": 0.4978, + "step": 5924 + }, + { + "epoch": 1.895797781569966, + "grad_norm": 0.5529834628105164, + "learning_rate": 3.588657867023514e-06, + "loss": 0.5322, + "step": 5925 + }, + { + "epoch": 1.896117747440273, + "grad_norm": 0.49046170711517334, + "learning_rate": 3.5868718790643847e-06, + "loss": 0.4508, + "step": 5926 + }, + { + "epoch": 1.89643771331058, + "grad_norm": 0.5163431763648987, + "learning_rate": 3.5850860870371685e-06, + "loss": 0.4882, + "step": 5927 + }, + { + "epoch": 1.8967576791808873, + "grad_norm": 0.5227925777435303, + "learning_rate": 3.5833004911894697e-06, + "loss": 0.5048, + "step": 5928 + }, + { + "epoch": 1.8970776450511946, + "grad_norm": 0.5189322829246521, + "learning_rate": 3.5815150917688613e-06, + "loss": 0.4615, + "step": 5929 + }, + { + "epoch": 1.8973976109215016, + "grad_norm": 0.5135989189147949, + "learning_rate": 3.5797298890228903e-06, + "loss": 0.4782, + "step": 5930 + }, + { + "epoch": 1.8977175767918089, + "grad_norm": 0.5108239650726318, + "learning_rate": 3.577944883199081e-06, + "loss": 0.4838, + "step": 5931 + }, + { + "epoch": 1.898037542662116, + "grad_norm": 0.5215096473693848, + "learning_rate": 3.5761600745449234e-06, + "loss": 0.4429, + "step": 5932 + }, + { + "epoch": 1.8983575085324231, + "grad_norm": 0.5620769262313843, + "learning_rate": 3.574375463307885e-06, + "loss": 0.4588, + "step": 5933 + }, + { + "epoch": 1.8986774744027304, + "grad_norm": 0.5866304039955139, + "learning_rate": 3.5725910497354037e-06, + "loss": 0.5205, + "step": 5934 + }, + { + "epoch": 1.8989974402730376, + "grad_norm": 0.5532835125923157, + "learning_rate": 3.5708068340748893e-06, + "loss": 0.5163, + "step": 5935 + }, + { + "epoch": 1.8993174061433447, + "grad_norm": 0.44951871037483215, + "learning_rate": 3.5690228165737283e-06, + "loss": 0.4246, + "step": 5936 + }, + { + "epoch": 1.899637372013652, + "grad_norm": 0.5440343618392944, + "learning_rate": 3.5672389974792743e-06, + "loss": 0.5024, + "step": 5937 + }, + { + "epoch": 1.8999573378839592, + "grad_norm": 0.5705434679985046, + "learning_rate": 3.565455377038858e-06, + "loss": 0.5296, + "step": 5938 + }, + { + "epoch": 1.9002773037542662, + "grad_norm": 0.556301474571228, + "learning_rate": 3.563671955499779e-06, + "loss": 0.4251, + "step": 5939 + }, + { + "epoch": 1.9005972696245734, + "grad_norm": 0.4998564124107361, + "learning_rate": 3.56188873310931e-06, + "loss": 0.4588, + "step": 5940 + }, + { + "epoch": 1.9009172354948807, + "grad_norm": 0.5576817393302917, + "learning_rate": 3.5601057101146995e-06, + "loss": 0.507, + "step": 5941 + }, + { + "epoch": 1.9012372013651877, + "grad_norm": 0.4908350110054016, + "learning_rate": 3.558322886763165e-06, + "loss": 0.4606, + "step": 5942 + }, + { + "epoch": 1.9015571672354947, + "grad_norm": 0.5293522477149963, + "learning_rate": 3.5565402633018963e-06, + "loss": 0.4604, + "step": 5943 + }, + { + "epoch": 1.9018771331058022, + "grad_norm": 0.48738279938697815, + "learning_rate": 3.5547578399780557e-06, + "loss": 0.4269, + "step": 5944 + }, + { + "epoch": 1.9021970989761092, + "grad_norm": 0.5112420916557312, + "learning_rate": 3.5529756170387785e-06, + "loss": 0.4986, + "step": 5945 + }, + { + "epoch": 1.9025170648464163, + "grad_norm": 0.5296844244003296, + "learning_rate": 3.5511935947311756e-06, + "loss": 0.5062, + "step": 5946 + }, + { + "epoch": 1.9028370307167235, + "grad_norm": 0.5048282742500305, + "learning_rate": 3.549411773302322e-06, + "loss": 0.4243, + "step": 5947 + }, + { + "epoch": 1.9031569965870307, + "grad_norm": 0.5419057011604309, + "learning_rate": 3.5476301529992717e-06, + "loss": 0.5494, + "step": 5948 + }, + { + "epoch": 1.9034769624573378, + "grad_norm": 0.4937518835067749, + "learning_rate": 3.545848734069048e-06, + "loss": 0.4584, + "step": 5949 + }, + { + "epoch": 1.903796928327645, + "grad_norm": 0.5428810715675354, + "learning_rate": 3.5440675167586446e-06, + "loss": 0.4979, + "step": 5950 + }, + { + "epoch": 1.9041168941979523, + "grad_norm": 0.5214746594429016, + "learning_rate": 3.5422865013150344e-06, + "loss": 0.4961, + "step": 5951 + }, + { + "epoch": 1.9044368600682593, + "grad_norm": 0.5318081378936768, + "learning_rate": 3.5405056879851533e-06, + "loss": 0.5296, + "step": 5952 + }, + { + "epoch": 1.9047568259385665, + "grad_norm": 0.488450825214386, + "learning_rate": 3.5387250770159152e-06, + "loss": 0.4416, + "step": 5953 + }, + { + "epoch": 1.9050767918088738, + "grad_norm": 0.540209174156189, + "learning_rate": 3.536944668654202e-06, + "loss": 0.4512, + "step": 5954 + }, + { + "epoch": 1.9053967576791808, + "grad_norm": 0.5259593725204468, + "learning_rate": 3.53516446314687e-06, + "loss": 0.518, + "step": 5955 + }, + { + "epoch": 1.905716723549488, + "grad_norm": 0.5429878830909729, + "learning_rate": 3.5333844607407497e-06, + "loss": 0.5454, + "step": 5956 + }, + { + "epoch": 1.9060366894197953, + "grad_norm": 0.42499151825904846, + "learning_rate": 3.531604661682636e-06, + "loss": 0.4153, + "step": 5957 + }, + { + "epoch": 1.9063566552901023, + "grad_norm": 0.5061125755310059, + "learning_rate": 3.5298250662193046e-06, + "loss": 0.5393, + "step": 5958 + }, + { + "epoch": 1.9066766211604096, + "grad_norm": 0.5056055188179016, + "learning_rate": 3.5280456745974943e-06, + "loss": 0.469, + "step": 5959 + }, + { + "epoch": 1.9069965870307168, + "grad_norm": 0.5259008407592773, + "learning_rate": 3.526266487063923e-06, + "loss": 0.5018, + "step": 5960 + }, + { + "epoch": 1.9073165529010239, + "grad_norm": 0.4866785407066345, + "learning_rate": 3.524487503865277e-06, + "loss": 0.4832, + "step": 5961 + }, + { + "epoch": 1.9076365187713311, + "grad_norm": 0.44651931524276733, + "learning_rate": 3.522708725248213e-06, + "loss": 0.4442, + "step": 5962 + }, + { + "epoch": 1.9079564846416384, + "grad_norm": 0.5012054443359375, + "learning_rate": 3.520930151459362e-06, + "loss": 0.4776, + "step": 5963 + }, + { + "epoch": 1.9082764505119454, + "grad_norm": 0.4618786573410034, + "learning_rate": 3.5191517827453224e-06, + "loss": 0.5152, + "step": 5964 + }, + { + "epoch": 1.9085964163822524, + "grad_norm": 0.5219776630401611, + "learning_rate": 3.517373619352672e-06, + "loss": 0.4974, + "step": 5965 + }, + { + "epoch": 1.9089163822525599, + "grad_norm": 0.5021116733551025, + "learning_rate": 3.5155956615279528e-06, + "loss": 0.4273, + "step": 5966 + }, + { + "epoch": 1.909236348122867, + "grad_norm": 0.5333878397941589, + "learning_rate": 3.5138179095176805e-06, + "loss": 0.5371, + "step": 5967 + }, + { + "epoch": 1.909556313993174, + "grad_norm": 0.5196536779403687, + "learning_rate": 3.512040363568344e-06, + "loss": 0.4882, + "step": 5968 + }, + { + "epoch": 1.9098762798634812, + "grad_norm": 0.5455049872398376, + "learning_rate": 3.5102630239263986e-06, + "loss": 0.4518, + "step": 5969 + }, + { + "epoch": 1.9101962457337884, + "grad_norm": 0.5605724453926086, + "learning_rate": 3.508485890838278e-06, + "loss": 0.4898, + "step": 5970 + }, + { + "epoch": 1.9105162116040955, + "grad_norm": 0.5000489354133606, + "learning_rate": 3.506708964550385e-06, + "loss": 0.5037, + "step": 5971 + }, + { + "epoch": 1.9108361774744027, + "grad_norm": 0.515364408493042, + "learning_rate": 3.5049322453090883e-06, + "loss": 0.4956, + "step": 5972 + }, + { + "epoch": 1.91115614334471, + "grad_norm": 0.5556088089942932, + "learning_rate": 3.503155733360737e-06, + "loss": 0.4633, + "step": 5973 + }, + { + "epoch": 1.911476109215017, + "grad_norm": 0.5696067810058594, + "learning_rate": 3.5013794289516406e-06, + "loss": 0.4881, + "step": 5974 + }, + { + "epoch": 1.9117960750853242, + "grad_norm": 0.48840129375457764, + "learning_rate": 3.4996033323280907e-06, + "loss": 0.536, + "step": 5975 + }, + { + "epoch": 1.9121160409556315, + "grad_norm": 0.4710192084312439, + "learning_rate": 3.4978274437363447e-06, + "loss": 0.472, + "step": 5976 + }, + { + "epoch": 1.9124360068259385, + "grad_norm": 0.47531548142433167, + "learning_rate": 3.49605176342263e-06, + "loss": 0.4946, + "step": 5977 + }, + { + "epoch": 1.9127559726962458, + "grad_norm": 0.5320174694061279, + "learning_rate": 3.4942762916331485e-06, + "loss": 0.489, + "step": 5978 + }, + { + "epoch": 1.913075938566553, + "grad_norm": 0.5178653001785278, + "learning_rate": 3.4925010286140688e-06, + "loss": 0.4951, + "step": 5979 + }, + { + "epoch": 1.91339590443686, + "grad_norm": 0.45877155661582947, + "learning_rate": 3.4907259746115356e-06, + "loss": 0.4442, + "step": 5980 + }, + { + "epoch": 1.9137158703071673, + "grad_norm": 0.5815134644508362, + "learning_rate": 3.4889511298716634e-06, + "loss": 0.5452, + "step": 5981 + }, + { + "epoch": 1.9140358361774745, + "grad_norm": 0.5238719582557678, + "learning_rate": 3.487176494640533e-06, + "loss": 0.4666, + "step": 5982 + }, + { + "epoch": 1.9143558020477816, + "grad_norm": 0.5438646078109741, + "learning_rate": 3.485402069164203e-06, + "loss": 0.5087, + "step": 5983 + }, + { + "epoch": 1.9146757679180886, + "grad_norm": 0.5344850420951843, + "learning_rate": 3.4836278536886964e-06, + "loss": 0.4674, + "step": 5984 + }, + { + "epoch": 1.914995733788396, + "grad_norm": 0.5671039819717407, + "learning_rate": 3.481853848460013e-06, + "loss": 0.5366, + "step": 5985 + }, + { + "epoch": 1.915315699658703, + "grad_norm": 0.4894713759422302, + "learning_rate": 3.4800800537241207e-06, + "loss": 0.4418, + "step": 5986 + }, + { + "epoch": 1.91563566552901, + "grad_norm": 0.5043044090270996, + "learning_rate": 3.478306469726957e-06, + "loss": 0.5091, + "step": 5987 + }, + { + "epoch": 1.9159556313993176, + "grad_norm": 0.5146811008453369, + "learning_rate": 3.4765330967144328e-06, + "loss": 0.4145, + "step": 5988 + }, + { + "epoch": 1.9162755972696246, + "grad_norm": 0.5499746203422546, + "learning_rate": 3.4747599349324248e-06, + "loss": 0.5641, + "step": 5989 + }, + { + "epoch": 1.9165955631399316, + "grad_norm": 0.536591112613678, + "learning_rate": 3.472986984626788e-06, + "loss": 0.4706, + "step": 5990 + }, + { + "epoch": 1.9169155290102389, + "grad_norm": 0.566261351108551, + "learning_rate": 3.471214246043344e-06, + "loss": 0.5222, + "step": 5991 + }, + { + "epoch": 1.9172354948805461, + "grad_norm": 0.5134425759315491, + "learning_rate": 3.4694417194278834e-06, + "loss": 0.506, + "step": 5992 + }, + { + "epoch": 1.9175554607508531, + "grad_norm": 0.4592892825603485, + "learning_rate": 3.4676694050261684e-06, + "loss": 0.4737, + "step": 5993 + }, + { + "epoch": 1.9178754266211604, + "grad_norm": 0.5180602073669434, + "learning_rate": 3.4658973030839357e-06, + "loss": 0.4904, + "step": 5994 + }, + { + "epoch": 1.9181953924914676, + "grad_norm": 0.5318039059638977, + "learning_rate": 3.464125413846886e-06, + "loss": 0.4827, + "step": 5995 + }, + { + "epoch": 1.9185153583617747, + "grad_norm": 0.5263141393661499, + "learning_rate": 3.4623537375606976e-06, + "loss": 0.4475, + "step": 5996 + }, + { + "epoch": 1.918835324232082, + "grad_norm": 0.539147138595581, + "learning_rate": 3.4605822744710115e-06, + "loss": 0.548, + "step": 5997 + }, + { + "epoch": 1.9191552901023892, + "grad_norm": 0.48374414443969727, + "learning_rate": 3.458811024823444e-06, + "loss": 0.4516, + "step": 5998 + }, + { + "epoch": 1.9194752559726962, + "grad_norm": 0.48777681589126587, + "learning_rate": 3.4570399888635843e-06, + "loss": 0.5028, + "step": 5999 + }, + { + "epoch": 1.9197952218430034, + "grad_norm": 0.5044283270835876, + "learning_rate": 3.455269166836985e-06, + "loss": 0.4647, + "step": 6000 + }, + { + "epoch": 1.9201151877133107, + "grad_norm": 0.4887102544307709, + "learning_rate": 3.453498558989175e-06, + "loss": 0.4405, + "step": 6001 + }, + { + "epoch": 1.9204351535836177, + "grad_norm": 0.494515985250473, + "learning_rate": 3.4517281655656498e-06, + "loss": 0.5235, + "step": 6002 + }, + { + "epoch": 1.920755119453925, + "grad_norm": 0.5164315104484558, + "learning_rate": 3.449957986811876e-06, + "loss": 0.479, + "step": 6003 + }, + { + "epoch": 1.9210750853242322, + "grad_norm": 0.4746878147125244, + "learning_rate": 3.448188022973293e-06, + "loss": 0.4803, + "step": 6004 + }, + { + "epoch": 1.9213950511945392, + "grad_norm": 0.5262110233306885, + "learning_rate": 3.446418274295308e-06, + "loss": 0.4984, + "step": 6005 + }, + { + "epoch": 1.9217150170648463, + "grad_norm": 0.5051013231277466, + "learning_rate": 3.4446487410232986e-06, + "loss": 0.4641, + "step": 6006 + }, + { + "epoch": 1.9220349829351537, + "grad_norm": 0.48198986053466797, + "learning_rate": 3.4428794234026114e-06, + "loss": 0.5024, + "step": 6007 + }, + { + "epoch": 1.9223549488054608, + "grad_norm": 0.4754115343093872, + "learning_rate": 3.4411103216785645e-06, + "loss": 0.4071, + "step": 6008 + }, + { + "epoch": 1.9226749146757678, + "grad_norm": 0.5091566443443298, + "learning_rate": 3.439341436096449e-06, + "loss": 0.5263, + "step": 6009 + }, + { + "epoch": 1.922994880546075, + "grad_norm": 0.45499107241630554, + "learning_rate": 3.4375727669015196e-06, + "loss": 0.4544, + "step": 6010 + }, + { + "epoch": 1.9233148464163823, + "grad_norm": 0.5030335187911987, + "learning_rate": 3.4358043143390067e-06, + "loss": 0.5287, + "step": 6011 + }, + { + "epoch": 1.9236348122866893, + "grad_norm": 0.4837437570095062, + "learning_rate": 3.4340360786541067e-06, + "loss": 0.4994, + "step": 6012 + }, + { + "epoch": 1.9239547781569966, + "grad_norm": 0.4857082664966583, + "learning_rate": 3.432268060091987e-06, + "loss": 0.4865, + "step": 6013 + }, + { + "epoch": 1.9242747440273038, + "grad_norm": 0.4972582161426544, + "learning_rate": 3.4305002588977888e-06, + "loss": 0.4919, + "step": 6014 + }, + { + "epoch": 1.9245947098976108, + "grad_norm": 0.5630078911781311, + "learning_rate": 3.428732675316617e-06, + "loss": 0.5211, + "step": 6015 + }, + { + "epoch": 1.924914675767918, + "grad_norm": 0.48093336820602417, + "learning_rate": 3.4269653095935514e-06, + "loss": 0.4372, + "step": 6016 + }, + { + "epoch": 1.9252346416382253, + "grad_norm": 0.5052854418754578, + "learning_rate": 3.4251981619736363e-06, + "loss": 0.4873, + "step": 6017 + }, + { + "epoch": 1.9255546075085324, + "grad_norm": 0.5380025506019592, + "learning_rate": 3.4234312327018904e-06, + "loss": 0.5225, + "step": 6018 + }, + { + "epoch": 1.9258745733788396, + "grad_norm": 0.49216553568840027, + "learning_rate": 3.4216645220233023e-06, + "loss": 0.4527, + "step": 6019 + }, + { + "epoch": 1.9261945392491469, + "grad_norm": 0.5159412026405334, + "learning_rate": 3.4198980301828256e-06, + "loss": 0.4326, + "step": 6020 + }, + { + "epoch": 1.9265145051194539, + "grad_norm": 0.5136810541152954, + "learning_rate": 3.418131757425389e-06, + "loss": 0.5023, + "step": 6021 + }, + { + "epoch": 1.9268344709897611, + "grad_norm": 0.5028185248374939, + "learning_rate": 3.4163657039958874e-06, + "loss": 0.4665, + "step": 6022 + }, + { + "epoch": 1.9271544368600684, + "grad_norm": 0.5626875162124634, + "learning_rate": 3.4145998701391837e-06, + "loss": 0.5032, + "step": 6023 + }, + { + "epoch": 1.9274744027303754, + "grad_norm": 0.5647227168083191, + "learning_rate": 3.4128342561001184e-06, + "loss": 0.4871, + "step": 6024 + }, + { + "epoch": 1.9277943686006824, + "grad_norm": 0.498357892036438, + "learning_rate": 3.4110688621234912e-06, + "loss": 0.4873, + "step": 6025 + }, + { + "epoch": 1.92811433447099, + "grad_norm": 0.5906538367271423, + "learning_rate": 3.4093036884540786e-06, + "loss": 0.5337, + "step": 6026 + }, + { + "epoch": 1.928434300341297, + "grad_norm": 0.4693874716758728, + "learning_rate": 3.4075387353366206e-06, + "loss": 0.4774, + "step": 6027 + }, + { + "epoch": 1.928754266211604, + "grad_norm": 0.4883122742176056, + "learning_rate": 3.405774003015836e-06, + "loss": 0.4812, + "step": 6028 + }, + { + "epoch": 1.9290742320819114, + "grad_norm": 0.5262741446495056, + "learning_rate": 3.4040094917364026e-06, + "loss": 0.4612, + "step": 6029 + }, + { + "epoch": 1.9293941979522184, + "grad_norm": 0.5191846489906311, + "learning_rate": 3.4022452017429743e-06, + "loss": 0.4578, + "step": 6030 + }, + { + "epoch": 1.9297141638225255, + "grad_norm": 0.5533474683761597, + "learning_rate": 3.400481133280171e-06, + "loss": 0.4828, + "step": 6031 + }, + { + "epoch": 1.9300341296928327, + "grad_norm": 0.513857364654541, + "learning_rate": 3.3987172865925787e-06, + "loss": 0.4628, + "step": 6032 + }, + { + "epoch": 1.93035409556314, + "grad_norm": 0.5177400708198547, + "learning_rate": 3.3969536619247658e-06, + "loss": 0.4707, + "step": 6033 + }, + { + "epoch": 1.930674061433447, + "grad_norm": 0.5255177617073059, + "learning_rate": 3.395190259521254e-06, + "loss": 0.5129, + "step": 6034 + }, + { + "epoch": 1.9309940273037542, + "grad_norm": 0.44974416494369507, + "learning_rate": 3.3934270796265444e-06, + "loss": 0.4809, + "step": 6035 + }, + { + "epoch": 1.9313139931740615, + "grad_norm": 0.48602941632270813, + "learning_rate": 3.391664122485103e-06, + "loss": 0.5041, + "step": 6036 + }, + { + "epoch": 1.9316339590443685, + "grad_norm": 0.4708314836025238, + "learning_rate": 3.3899013883413635e-06, + "loss": 0.4475, + "step": 6037 + }, + { + "epoch": 1.9319539249146758, + "grad_norm": 0.49465641379356384, + "learning_rate": 3.388138877439736e-06, + "loss": 0.5045, + "step": 6038 + }, + { + "epoch": 1.932273890784983, + "grad_norm": 0.525989830493927, + "learning_rate": 3.3863765900245914e-06, + "loss": 0.5234, + "step": 6039 + }, + { + "epoch": 1.93259385665529, + "grad_norm": 0.5063377022743225, + "learning_rate": 3.3846145263402744e-06, + "loss": 0.4521, + "step": 6040 + }, + { + "epoch": 1.9329138225255973, + "grad_norm": 0.4842839241027832, + "learning_rate": 3.3828526866310962e-06, + "loss": 0.4912, + "step": 6041 + }, + { + "epoch": 1.9332337883959045, + "grad_norm": 0.49538734555244446, + "learning_rate": 3.3810910711413376e-06, + "loss": 0.5095, + "step": 6042 + }, + { + "epoch": 1.9335537542662116, + "grad_norm": 0.5384979844093323, + "learning_rate": 3.3793296801152508e-06, + "loss": 0.4735, + "step": 6043 + }, + { + "epoch": 1.9338737201365188, + "grad_norm": 0.5058707594871521, + "learning_rate": 3.377568513797054e-06, + "loss": 0.4972, + "step": 6044 + }, + { + "epoch": 1.934193686006826, + "grad_norm": 0.5415966510772705, + "learning_rate": 3.3758075724309342e-06, + "loss": 0.4945, + "step": 6045 + }, + { + "epoch": 1.934513651877133, + "grad_norm": 0.5162762403488159, + "learning_rate": 3.374046856261048e-06, + "loss": 0.4969, + "step": 6046 + }, + { + "epoch": 1.9348336177474401, + "grad_norm": 0.516521692276001, + "learning_rate": 3.37228636553152e-06, + "loss": 0.4676, + "step": 6047 + }, + { + "epoch": 1.9351535836177476, + "grad_norm": 0.4812181293964386, + "learning_rate": 3.3705261004864476e-06, + "loss": 0.4643, + "step": 6048 + }, + { + "epoch": 1.9354735494880546, + "grad_norm": 0.5185404419898987, + "learning_rate": 3.3687660613698907e-06, + "loss": 0.4941, + "step": 6049 + }, + { + "epoch": 1.9357935153583616, + "grad_norm": 0.558539092540741, + "learning_rate": 3.367006248425883e-06, + "loss": 0.5026, + "step": 6050 + }, + { + "epoch": 1.9361134812286689, + "grad_norm": 0.47996214032173157, + "learning_rate": 3.3652466618984226e-06, + "loss": 0.4369, + "step": 6051 + }, + { + "epoch": 1.9364334470989761, + "grad_norm": 0.5430442690849304, + "learning_rate": 3.363487302031478e-06, + "loss": 0.5251, + "step": 6052 + }, + { + "epoch": 1.9367534129692832, + "grad_norm": 0.5165271162986755, + "learning_rate": 3.3617281690689895e-06, + "loss": 0.4804, + "step": 6053 + }, + { + "epoch": 1.9370733788395904, + "grad_norm": 0.5100358128547668, + "learning_rate": 3.359969263254861e-06, + "loss": 0.4563, + "step": 6054 + }, + { + "epoch": 1.9373933447098977, + "grad_norm": 0.5728858709335327, + "learning_rate": 3.3582105848329677e-06, + "loss": 0.5041, + "step": 6055 + }, + { + "epoch": 1.9377133105802047, + "grad_norm": 0.5265145301818848, + "learning_rate": 3.3564521340471507e-06, + "loss": 0.459, + "step": 6056 + }, + { + "epoch": 1.938033276450512, + "grad_norm": 0.5253676772117615, + "learning_rate": 3.354693911141222e-06, + "loss": 0.5262, + "step": 6057 + }, + { + "epoch": 1.9383532423208192, + "grad_norm": 0.49260178208351135, + "learning_rate": 3.3529359163589636e-06, + "loss": 0.4921, + "step": 6058 + }, + { + "epoch": 1.9386732081911262, + "grad_norm": 0.47888457775115967, + "learning_rate": 3.3511781499441207e-06, + "loss": 0.5452, + "step": 6059 + }, + { + "epoch": 1.9389931740614335, + "grad_norm": 0.5283399224281311, + "learning_rate": 3.349420612140412e-06, + "loss": 0.4636, + "step": 6060 + }, + { + "epoch": 1.9393131399317407, + "grad_norm": 0.5486240386962891, + "learning_rate": 3.3476633031915194e-06, + "loss": 0.5057, + "step": 6061 + }, + { + "epoch": 1.9396331058020477, + "grad_norm": 0.4765787422657013, + "learning_rate": 3.3459062233410983e-06, + "loss": 0.4171, + "step": 6062 + }, + { + "epoch": 1.939953071672355, + "grad_norm": 0.5192815661430359, + "learning_rate": 3.34414937283277e-06, + "loss": 0.4798, + "step": 6063 + }, + { + "epoch": 1.9402730375426622, + "grad_norm": 0.5335144996643066, + "learning_rate": 3.342392751910123e-06, + "loss": 0.4964, + "step": 6064 + }, + { + "epoch": 1.9405930034129693, + "grad_norm": 0.49786290526390076, + "learning_rate": 3.340636360816716e-06, + "loss": 0.4673, + "step": 6065 + }, + { + "epoch": 1.9409129692832765, + "grad_norm": 0.4714604318141937, + "learning_rate": 3.3388801997960714e-06, + "loss": 0.4626, + "step": 6066 + }, + { + "epoch": 1.9412329351535837, + "grad_norm": 0.4573826789855957, + "learning_rate": 3.3371242690916863e-06, + "loss": 0.4782, + "step": 6067 + }, + { + "epoch": 1.9415529010238908, + "grad_norm": 0.4418972432613373, + "learning_rate": 3.335368568947023e-06, + "loss": 0.4485, + "step": 6068 + }, + { + "epoch": 1.9418728668941978, + "grad_norm": 0.5344722270965576, + "learning_rate": 3.3336130996055095e-06, + "loss": 0.4972, + "step": 6069 + }, + { + "epoch": 1.9421928327645053, + "grad_norm": 0.49242472648620605, + "learning_rate": 3.3318578613105453e-06, + "loss": 0.4598, + "step": 6070 + }, + { + "epoch": 1.9425127986348123, + "grad_norm": 0.49834853410720825, + "learning_rate": 3.3301028543054935e-06, + "loss": 0.4652, + "step": 6071 + }, + { + "epoch": 1.9428327645051193, + "grad_norm": 0.4919757843017578, + "learning_rate": 3.328348078833691e-06, + "loss": 0.4946, + "step": 6072 + }, + { + "epoch": 1.9431527303754266, + "grad_norm": 0.5249433517456055, + "learning_rate": 3.3265935351384386e-06, + "loss": 0.4773, + "step": 6073 + }, + { + "epoch": 1.9434726962457338, + "grad_norm": 0.45385199785232544, + "learning_rate": 3.3248392234630056e-06, + "loss": 0.4416, + "step": 6074 + }, + { + "epoch": 1.9437926621160408, + "grad_norm": 0.5273494720458984, + "learning_rate": 3.32308514405063e-06, + "loss": 0.5478, + "step": 6075 + }, + { + "epoch": 1.944112627986348, + "grad_norm": 0.4976264238357544, + "learning_rate": 3.321331297144515e-06, + "loss": 0.4453, + "step": 6076 + }, + { + "epoch": 1.9444325938566553, + "grad_norm": 0.5137234926223755, + "learning_rate": 3.3195776829878357e-06, + "loss": 0.483, + "step": 6077 + }, + { + "epoch": 1.9447525597269624, + "grad_norm": 0.5299399495124817, + "learning_rate": 3.3178243018237334e-06, + "loss": 0.5094, + "step": 6078 + }, + { + "epoch": 1.9450725255972696, + "grad_norm": 0.4810827970504761, + "learning_rate": 3.316071153895315e-06, + "loss": 0.4915, + "step": 6079 + }, + { + "epoch": 1.9453924914675769, + "grad_norm": 0.49671903252601624, + "learning_rate": 3.314318239445657e-06, + "loss": 0.4534, + "step": 6080 + }, + { + "epoch": 1.945712457337884, + "grad_norm": 0.4928573966026306, + "learning_rate": 3.312565558717801e-06, + "loss": 0.5293, + "step": 6081 + }, + { + "epoch": 1.9460324232081911, + "grad_norm": 0.49744442105293274, + "learning_rate": 3.310813111954762e-06, + "loss": 0.5188, + "step": 6082 + }, + { + "epoch": 1.9463523890784984, + "grad_norm": 0.4996615946292877, + "learning_rate": 3.309060899399518e-06, + "loss": 0.4572, + "step": 6083 + }, + { + "epoch": 1.9466723549488054, + "grad_norm": 0.48610150814056396, + "learning_rate": 3.3073089212950137e-06, + "loss": 0.46, + "step": 6084 + }, + { + "epoch": 1.9469923208191127, + "grad_norm": 0.47651350498199463, + "learning_rate": 3.3055571778841645e-06, + "loss": 0.4671, + "step": 6085 + }, + { + "epoch": 1.94731228668942, + "grad_norm": 0.5219941735267639, + "learning_rate": 3.3038056694098485e-06, + "loss": 0.4608, + "step": 6086 + }, + { + "epoch": 1.947632252559727, + "grad_norm": 0.5291367769241333, + "learning_rate": 3.302054396114919e-06, + "loss": 0.468, + "step": 6087 + }, + { + "epoch": 1.947952218430034, + "grad_norm": 0.5114938616752625, + "learning_rate": 3.3003033582421905e-06, + "loss": 0.4465, + "step": 6088 + }, + { + "epoch": 1.9482721843003414, + "grad_norm": 0.5421054363250732, + "learning_rate": 3.2985525560344458e-06, + "loss": 0.5237, + "step": 6089 + }, + { + "epoch": 1.9485921501706485, + "grad_norm": 0.5172793865203857, + "learning_rate": 3.2968019897344367e-06, + "loss": 0.4778, + "step": 6090 + }, + { + "epoch": 1.9489121160409555, + "grad_norm": 0.523431122303009, + "learning_rate": 3.2950516595848785e-06, + "loss": 0.4951, + "step": 6091 + }, + { + "epoch": 1.949232081911263, + "grad_norm": 0.5439819097518921, + "learning_rate": 3.29330156582846e-06, + "loss": 0.528, + "step": 6092 + }, + { + "epoch": 1.94955204778157, + "grad_norm": 0.4954051673412323, + "learning_rate": 3.291551708707834e-06, + "loss": 0.5023, + "step": 6093 + }, + { + "epoch": 1.949872013651877, + "grad_norm": 0.5080917477607727, + "learning_rate": 3.289802088465617e-06, + "loss": 0.4355, + "step": 6094 + }, + { + "epoch": 1.9501919795221843, + "grad_norm": 0.6108106374740601, + "learning_rate": 3.2880527053444e-06, + "loss": 0.5668, + "step": 6095 + }, + { + "epoch": 1.9505119453924915, + "grad_norm": 0.6188663840293884, + "learning_rate": 3.286303559586732e-06, + "loss": 0.5193, + "step": 6096 + }, + { + "epoch": 1.9508319112627985, + "grad_norm": 0.5643511414527893, + "learning_rate": 3.2845546514351384e-06, + "loss": 0.4728, + "step": 6097 + }, + { + "epoch": 1.9511518771331058, + "grad_norm": 0.5128049254417419, + "learning_rate": 3.282805981132107e-06, + "loss": 0.441, + "step": 6098 + }, + { + "epoch": 1.951471843003413, + "grad_norm": 0.52104651927948, + "learning_rate": 3.281057548920091e-06, + "loss": 0.5237, + "step": 6099 + }, + { + "epoch": 1.95179180887372, + "grad_norm": 0.565772533416748, + "learning_rate": 3.2793093550415133e-06, + "loss": 0.5355, + "step": 6100 + }, + { + "epoch": 1.9521117747440273, + "grad_norm": 0.5251613259315491, + "learning_rate": 3.2775613997387656e-06, + "loss": 0.4681, + "step": 6101 + }, + { + "epoch": 1.9524317406143346, + "grad_norm": 0.5024871826171875, + "learning_rate": 3.2758136832542008e-06, + "loss": 0.4825, + "step": 6102 + }, + { + "epoch": 1.9527517064846416, + "grad_norm": 0.5466561317443848, + "learning_rate": 3.274066205830144e-06, + "loss": 0.5212, + "step": 6103 + }, + { + "epoch": 1.9530716723549488, + "grad_norm": 0.5681954622268677, + "learning_rate": 3.272318967708883e-06, + "loss": 0.5492, + "step": 6104 + }, + { + "epoch": 1.953391638225256, + "grad_norm": 0.5645036101341248, + "learning_rate": 3.270571969132674e-06, + "loss": 0.533, + "step": 6105 + }, + { + "epoch": 1.953711604095563, + "grad_norm": 0.535334587097168, + "learning_rate": 3.2688252103437456e-06, + "loss": 0.4735, + "step": 6106 + }, + { + "epoch": 1.9540315699658704, + "grad_norm": 0.4753856360912323, + "learning_rate": 3.2670786915842824e-06, + "loss": 0.4832, + "step": 6107 + }, + { + "epoch": 1.9543515358361776, + "grad_norm": 0.4911159873008728, + "learning_rate": 3.265332413096444e-06, + "loss": 0.4495, + "step": 6108 + }, + { + "epoch": 1.9546715017064846, + "grad_norm": 0.5127443075180054, + "learning_rate": 3.263586375122353e-06, + "loss": 0.5102, + "step": 6109 + }, + { + "epoch": 1.9549914675767917, + "grad_norm": 0.4534604251384735, + "learning_rate": 3.2618405779040992e-06, + "loss": 0.4494, + "step": 6110 + }, + { + "epoch": 1.9553114334470991, + "grad_norm": 0.4901772737503052, + "learning_rate": 3.2600950216837423e-06, + "loss": 0.5097, + "step": 6111 + }, + { + "epoch": 1.9556313993174061, + "grad_norm": 0.48854494094848633, + "learning_rate": 3.258349706703302e-06, + "loss": 0.414, + "step": 6112 + }, + { + "epoch": 1.9559513651877132, + "grad_norm": 0.5062629580497742, + "learning_rate": 3.256604633204772e-06, + "loss": 0.4779, + "step": 6113 + }, + { + "epoch": 1.9562713310580204, + "grad_norm": 0.5243249535560608, + "learning_rate": 3.2548598014301057e-06, + "loss": 0.5221, + "step": 6114 + }, + { + "epoch": 1.9565912969283277, + "grad_norm": 0.5110526084899902, + "learning_rate": 3.2531152116212263e-06, + "loss": 0.4385, + "step": 6115 + }, + { + "epoch": 1.9569112627986347, + "grad_norm": 0.47971436381340027, + "learning_rate": 3.251370864020026e-06, + "loss": 0.477, + "step": 6116 + }, + { + "epoch": 1.957231228668942, + "grad_norm": 0.4537786543369293, + "learning_rate": 3.2496267588683583e-06, + "loss": 0.4254, + "step": 6117 + }, + { + "epoch": 1.9575511945392492, + "grad_norm": 0.5468919277191162, + "learning_rate": 3.2478828964080476e-06, + "loss": 0.4988, + "step": 6118 + }, + { + "epoch": 1.9578711604095562, + "grad_norm": 0.5255555510520935, + "learning_rate": 3.24613927688088e-06, + "loss": 0.4974, + "step": 6119 + }, + { + "epoch": 1.9581911262798635, + "grad_norm": 0.5240233540534973, + "learning_rate": 3.2443959005286106e-06, + "loss": 0.5151, + "step": 6120 + }, + { + "epoch": 1.9585110921501707, + "grad_norm": 0.4944935142993927, + "learning_rate": 3.2426527675929635e-06, + "loss": 0.4733, + "step": 6121 + }, + { + "epoch": 1.9588310580204777, + "grad_norm": 0.5280991792678833, + "learning_rate": 3.2409098783156233e-06, + "loss": 0.5099, + "step": 6122 + }, + { + "epoch": 1.959151023890785, + "grad_norm": 0.5345900058746338, + "learning_rate": 3.239167232938246e-06, + "loss": 0.5017, + "step": 6123 + }, + { + "epoch": 1.9594709897610922, + "grad_norm": 0.49158307909965515, + "learning_rate": 3.237424831702449e-06, + "loss": 0.4864, + "step": 6124 + }, + { + "epoch": 1.9597909556313993, + "grad_norm": 0.5014450550079346, + "learning_rate": 3.2356826748498182e-06, + "loss": 0.4501, + "step": 6125 + }, + { + "epoch": 1.9601109215017065, + "grad_norm": 0.513562023639679, + "learning_rate": 3.2339407626219097e-06, + "loss": 0.5151, + "step": 6126 + }, + { + "epoch": 1.9604308873720138, + "grad_norm": 0.5434235334396362, + "learning_rate": 3.2321990952602377e-06, + "loss": 0.4818, + "step": 6127 + }, + { + "epoch": 1.9607508532423208, + "grad_norm": 0.5348495841026306, + "learning_rate": 3.2304576730062887e-06, + "loss": 0.5835, + "step": 6128 + }, + { + "epoch": 1.961070819112628, + "grad_norm": 0.5259645581245422, + "learning_rate": 3.2287164961015116e-06, + "loss": 0.4939, + "step": 6129 + }, + { + "epoch": 1.9613907849829353, + "grad_norm": 0.5705592632293701, + "learning_rate": 3.226975564787322e-06, + "loss": 0.5378, + "step": 6130 + }, + { + "epoch": 1.9617107508532423, + "grad_norm": 0.4692423343658447, + "learning_rate": 3.225234879305105e-06, + "loss": 0.4094, + "step": 6131 + }, + { + "epoch": 1.9620307167235493, + "grad_norm": 0.5225099325180054, + "learning_rate": 3.2234944398962066e-06, + "loss": 0.481, + "step": 6132 + }, + { + "epoch": 1.9623506825938568, + "grad_norm": 0.49818119406700134, + "learning_rate": 3.2217542468019425e-06, + "loss": 0.5051, + "step": 6133 + }, + { + "epoch": 1.9626706484641638, + "grad_norm": 0.5146864652633667, + "learning_rate": 3.2200143002635885e-06, + "loss": 0.5107, + "step": 6134 + }, + { + "epoch": 1.9629906143344709, + "grad_norm": 0.46393853425979614, + "learning_rate": 3.218274600522395e-06, + "loss": 0.4544, + "step": 6135 + }, + { + "epoch": 1.963310580204778, + "grad_norm": 0.5225887298583984, + "learning_rate": 3.2165351478195715e-06, + "loss": 0.5056, + "step": 6136 + }, + { + "epoch": 1.9636305460750854, + "grad_norm": 0.4513947665691376, + "learning_rate": 3.2147959423962954e-06, + "loss": 0.4366, + "step": 6137 + }, + { + "epoch": 1.9639505119453924, + "grad_norm": 0.4780377149581909, + "learning_rate": 3.2130569844937097e-06, + "loss": 0.5216, + "step": 6138 + }, + { + "epoch": 1.9642704778156996, + "grad_norm": 0.5713197588920593, + "learning_rate": 3.2113182743529214e-06, + "loss": 0.4782, + "step": 6139 + }, + { + "epoch": 1.9645904436860069, + "grad_norm": 0.4937642216682434, + "learning_rate": 3.2095798122150072e-06, + "loss": 0.4634, + "step": 6140 + }, + { + "epoch": 1.964910409556314, + "grad_norm": 0.497175931930542, + "learning_rate": 3.207841598321007e-06, + "loss": 0.5058, + "step": 6141 + }, + { + "epoch": 1.9652303754266212, + "grad_norm": 0.5064600110054016, + "learning_rate": 3.2061036329119243e-06, + "loss": 0.5039, + "step": 6142 + }, + { + "epoch": 1.9655503412969284, + "grad_norm": 0.47617995738983154, + "learning_rate": 3.2043659162287317e-06, + "loss": 0.4448, + "step": 6143 + }, + { + "epoch": 1.9658703071672354, + "grad_norm": 0.4566115438938141, + "learning_rate": 3.2026284485123625e-06, + "loss": 0.4608, + "step": 6144 + }, + { + "epoch": 1.9661902730375427, + "grad_norm": 0.5149037837982178, + "learning_rate": 3.200891230003723e-06, + "loss": 0.4618, + "step": 6145 + }, + { + "epoch": 1.96651023890785, + "grad_norm": 0.5159351825714111, + "learning_rate": 3.1991542609436797e-06, + "loss": 0.4791, + "step": 6146 + }, + { + "epoch": 1.966830204778157, + "grad_norm": 0.5528331995010376, + "learning_rate": 3.1974175415730636e-06, + "loss": 0.5118, + "step": 6147 + }, + { + "epoch": 1.9671501706484642, + "grad_norm": 0.5175594687461853, + "learning_rate": 3.1956810721326747e-06, + "loss": 0.4688, + "step": 6148 + }, + { + "epoch": 1.9674701365187715, + "grad_norm": 0.47869059443473816, + "learning_rate": 3.1939448528632745e-06, + "loss": 0.4518, + "step": 6149 + }, + { + "epoch": 1.9677901023890785, + "grad_norm": 0.48490244150161743, + "learning_rate": 3.192208884005593e-06, + "loss": 0.4598, + "step": 6150 + }, + { + "epoch": 1.9681100682593855, + "grad_norm": 0.5457737445831299, + "learning_rate": 3.1904731658003264e-06, + "loss": 0.5035, + "step": 6151 + }, + { + "epoch": 1.968430034129693, + "grad_norm": 0.4673367738723755, + "learning_rate": 3.1887376984881313e-06, + "loss": 0.4317, + "step": 6152 + }, + { + "epoch": 1.96875, + "grad_norm": 0.5494493246078491, + "learning_rate": 3.187002482309634e-06, + "loss": 0.5019, + "step": 6153 + }, + { + "epoch": 1.969069965870307, + "grad_norm": 0.5415604710578918, + "learning_rate": 3.1852675175054216e-06, + "loss": 0.464, + "step": 6154 + }, + { + "epoch": 1.9693899317406145, + "grad_norm": 0.5042742490768433, + "learning_rate": 3.1835328043160522e-06, + "loss": 0.5252, + "step": 6155 + }, + { + "epoch": 1.9697098976109215, + "grad_norm": 0.4678550362586975, + "learning_rate": 3.181798342982046e-06, + "loss": 0.4642, + "step": 6156 + }, + { + "epoch": 1.9700298634812285, + "grad_norm": 0.5744708180427551, + "learning_rate": 3.180064133743885e-06, + "loss": 0.5409, + "step": 6157 + }, + { + "epoch": 1.9703498293515358, + "grad_norm": 0.46229907870292664, + "learning_rate": 3.178330176842023e-06, + "loss": 0.4953, + "step": 6158 + }, + { + "epoch": 1.970669795221843, + "grad_norm": 0.5091956257820129, + "learning_rate": 3.176596472516872e-06, + "loss": 0.5184, + "step": 6159 + }, + { + "epoch": 1.97098976109215, + "grad_norm": 0.5130820870399475, + "learning_rate": 3.1748630210088137e-06, + "loss": 0.4752, + "step": 6160 + }, + { + "epoch": 1.9713097269624573, + "grad_norm": 0.5606593489646912, + "learning_rate": 3.173129822558195e-06, + "loss": 0.4496, + "step": 6161 + }, + { + "epoch": 1.9716296928327646, + "grad_norm": 0.4711438715457916, + "learning_rate": 3.171396877405323e-06, + "loss": 0.4357, + "step": 6162 + }, + { + "epoch": 1.9719496587030716, + "grad_norm": 0.5281607508659363, + "learning_rate": 3.1696641857904743e-06, + "loss": 0.5107, + "step": 6163 + }, + { + "epoch": 1.9722696245733788, + "grad_norm": 0.494859904050827, + "learning_rate": 3.1679317479538864e-06, + "loss": 0.4721, + "step": 6164 + }, + { + "epoch": 1.972589590443686, + "grad_norm": 0.47621551156044006, + "learning_rate": 3.1661995641357666e-06, + "loss": 0.5002, + "step": 6165 + }, + { + "epoch": 1.9729095563139931, + "grad_norm": 0.5177987217903137, + "learning_rate": 3.164467634576285e-06, + "loss": 0.4769, + "step": 6166 + }, + { + "epoch": 1.9732295221843004, + "grad_norm": 0.5293196439743042, + "learning_rate": 3.1627359595155728e-06, + "loss": 0.5277, + "step": 6167 + }, + { + "epoch": 1.9735494880546076, + "grad_norm": 0.5151974558830261, + "learning_rate": 3.1610045391937282e-06, + "loss": 0.5385, + "step": 6168 + }, + { + "epoch": 1.9738694539249146, + "grad_norm": 0.5350662469863892, + "learning_rate": 3.1592733738508186e-06, + "loss": 0.4679, + "step": 6169 + }, + { + "epoch": 1.974189419795222, + "grad_norm": 0.5367306470870972, + "learning_rate": 3.157542463726871e-06, + "loss": 0.5163, + "step": 6170 + }, + { + "epoch": 1.9745093856655291, + "grad_norm": 0.5237488746643066, + "learning_rate": 3.1558118090618773e-06, + "loss": 0.511, + "step": 6171 + }, + { + "epoch": 1.9748293515358362, + "grad_norm": 0.47949159145355225, + "learning_rate": 3.154081410095793e-06, + "loss": 0.435, + "step": 6172 + }, + { + "epoch": 1.9751493174061432, + "grad_norm": 0.5170743465423584, + "learning_rate": 3.152351267068542e-06, + "loss": 0.515, + "step": 6173 + }, + { + "epoch": 1.9754692832764507, + "grad_norm": 0.5387949347496033, + "learning_rate": 3.1506213802200113e-06, + "loss": 0.4572, + "step": 6174 + }, + { + "epoch": 1.9757892491467577, + "grad_norm": 0.5387278199195862, + "learning_rate": 3.1488917497900523e-06, + "loss": 0.562, + "step": 6175 + }, + { + "epoch": 1.9761092150170647, + "grad_norm": 0.5308703184127808, + "learning_rate": 3.147162376018479e-06, + "loss": 0.5108, + "step": 6176 + }, + { + "epoch": 1.976429180887372, + "grad_norm": 0.4508434534072876, + "learning_rate": 3.1454332591450697e-06, + "loss": 0.4572, + "step": 6177 + }, + { + "epoch": 1.9767491467576792, + "grad_norm": 0.4709256887435913, + "learning_rate": 3.1437043994095697e-06, + "loss": 0.4828, + "step": 6178 + }, + { + "epoch": 1.9770691126279862, + "grad_norm": 0.541800856590271, + "learning_rate": 3.141975797051689e-06, + "loss": 0.533, + "step": 6179 + }, + { + "epoch": 1.9773890784982935, + "grad_norm": 0.4995775818824768, + "learning_rate": 3.1402474523111006e-06, + "loss": 0.5028, + "step": 6180 + }, + { + "epoch": 1.9777090443686007, + "grad_norm": 0.4512108564376831, + "learning_rate": 3.138519365427438e-06, + "loss": 0.4653, + "step": 6181 + }, + { + "epoch": 1.9780290102389078, + "grad_norm": 0.541901171207428, + "learning_rate": 3.136791536640307e-06, + "loss": 0.5351, + "step": 6182 + }, + { + "epoch": 1.978348976109215, + "grad_norm": 0.47517821192741394, + "learning_rate": 3.135063966189268e-06, + "loss": 0.4567, + "step": 6183 + }, + { + "epoch": 1.9786689419795223, + "grad_norm": 0.5447357296943665, + "learning_rate": 3.1333366543138557e-06, + "loss": 0.525, + "step": 6184 + }, + { + "epoch": 1.9789889078498293, + "grad_norm": 0.5360710024833679, + "learning_rate": 3.1316096012535625e-06, + "loss": 0.4772, + "step": 6185 + }, + { + "epoch": 1.9793088737201365, + "grad_norm": 0.5403434634208679, + "learning_rate": 3.1298828072478443e-06, + "loss": 0.5056, + "step": 6186 + }, + { + "epoch": 1.9796288395904438, + "grad_norm": 0.5185534358024597, + "learning_rate": 3.128156272536126e-06, + "loss": 0.4904, + "step": 6187 + }, + { + "epoch": 1.9799488054607508, + "grad_norm": 0.4958355128765106, + "learning_rate": 3.1264299973577906e-06, + "loss": 0.4628, + "step": 6188 + }, + { + "epoch": 1.980268771331058, + "grad_norm": 0.4718395173549652, + "learning_rate": 3.1247039819521907e-06, + "loss": 0.508, + "step": 6189 + }, + { + "epoch": 1.9805887372013653, + "grad_norm": 0.4683765769004822, + "learning_rate": 3.122978226558641e-06, + "loss": 0.4784, + "step": 6190 + }, + { + "epoch": 1.9809087030716723, + "grad_norm": 0.4669650197029114, + "learning_rate": 3.1212527314164177e-06, + "loss": 0.4133, + "step": 6191 + }, + { + "epoch": 1.9812286689419796, + "grad_norm": 0.5345683097839355, + "learning_rate": 3.1195274967647638e-06, + "loss": 0.5364, + "step": 6192 + }, + { + "epoch": 1.9815486348122868, + "grad_norm": 0.4708634912967682, + "learning_rate": 3.1178025228428826e-06, + "loss": 0.4594, + "step": 6193 + }, + { + "epoch": 1.9818686006825939, + "grad_norm": 0.5484588742256165, + "learning_rate": 3.116077809889947e-06, + "loss": 0.5032, + "step": 6194 + }, + { + "epoch": 1.9821885665529009, + "grad_norm": 0.5436263084411621, + "learning_rate": 3.1143533581450906e-06, + "loss": 0.5252, + "step": 6195 + }, + { + "epoch": 1.9825085324232083, + "grad_norm": 0.5184581875801086, + "learning_rate": 3.112629167847409e-06, + "loss": 0.4339, + "step": 6196 + }, + { + "epoch": 1.9828284982935154, + "grad_norm": 0.5556897521018982, + "learning_rate": 3.1109052392359645e-06, + "loss": 0.5362, + "step": 6197 + }, + { + "epoch": 1.9831484641638224, + "grad_norm": 0.44339731335639954, + "learning_rate": 3.1091815725497793e-06, + "loss": 0.4365, + "step": 6198 + }, + { + "epoch": 1.9834684300341296, + "grad_norm": 0.5135619640350342, + "learning_rate": 3.1074581680278447e-06, + "loss": 0.5121, + "step": 6199 + }, + { + "epoch": 1.983788395904437, + "grad_norm": 0.5377285480499268, + "learning_rate": 3.1057350259091125e-06, + "loss": 0.4856, + "step": 6200 + }, + { + "epoch": 1.984108361774744, + "grad_norm": 0.43569591641426086, + "learning_rate": 3.104012146432497e-06, + "loss": 0.4141, + "step": 6201 + }, + { + "epoch": 1.9844283276450512, + "grad_norm": 0.5048389434814453, + "learning_rate": 3.10228952983688e-06, + "loss": 0.4697, + "step": 6202 + }, + { + "epoch": 1.9847482935153584, + "grad_norm": 0.506389856338501, + "learning_rate": 3.1005671763611003e-06, + "loss": 0.4795, + "step": 6203 + }, + { + "epoch": 1.9850682593856654, + "grad_norm": 0.5272321105003357, + "learning_rate": 3.0988450862439677e-06, + "loss": 0.5171, + "step": 6204 + }, + { + "epoch": 1.9853882252559727, + "grad_norm": 0.505204439163208, + "learning_rate": 3.097123259724252e-06, + "loss": 0.4919, + "step": 6205 + }, + { + "epoch": 1.98570819112628, + "grad_norm": 0.5086266398429871, + "learning_rate": 3.095401697040684e-06, + "loss": 0.5065, + "step": 6206 + }, + { + "epoch": 1.986028156996587, + "grad_norm": 0.525647759437561, + "learning_rate": 3.0936803984319618e-06, + "loss": 0.4548, + "step": 6207 + }, + { + "epoch": 1.9863481228668942, + "grad_norm": 0.5006814002990723, + "learning_rate": 3.0919593641367473e-06, + "loss": 0.486, + "step": 6208 + }, + { + "epoch": 1.9866680887372015, + "grad_norm": 0.47532498836517334, + "learning_rate": 3.090238594393661e-06, + "loss": 0.4429, + "step": 6209 + }, + { + "epoch": 1.9869880546075085, + "grad_norm": 0.5055264830589294, + "learning_rate": 3.088518089441293e-06, + "loss": 0.4567, + "step": 6210 + }, + { + "epoch": 1.9873080204778157, + "grad_norm": 0.5233423709869385, + "learning_rate": 3.08679784951819e-06, + "loss": 0.5191, + "step": 6211 + }, + { + "epoch": 1.987627986348123, + "grad_norm": 0.4947076737880707, + "learning_rate": 3.0850778748628656e-06, + "loss": 0.442, + "step": 6212 + }, + { + "epoch": 1.98794795221843, + "grad_norm": 0.5296012759208679, + "learning_rate": 3.0833581657138e-06, + "loss": 0.5251, + "step": 6213 + }, + { + "epoch": 1.988267918088737, + "grad_norm": 0.4480663239955902, + "learning_rate": 3.081638722309429e-06, + "loss": 0.4544, + "step": 6214 + }, + { + "epoch": 1.9885878839590445, + "grad_norm": 0.5297701358795166, + "learning_rate": 3.0799195448881586e-06, + "loss": 0.5031, + "step": 6215 + }, + { + "epoch": 1.9889078498293515, + "grad_norm": 0.5122623443603516, + "learning_rate": 3.078200633688352e-06, + "loss": 0.4682, + "step": 6216 + }, + { + "epoch": 1.9892278156996586, + "grad_norm": 0.5205539464950562, + "learning_rate": 3.076481988948338e-06, + "loss": 0.5313, + "step": 6217 + }, + { + "epoch": 1.989547781569966, + "grad_norm": 0.5184388160705566, + "learning_rate": 3.0747636109064126e-06, + "loss": 0.5621, + "step": 6218 + }, + { + "epoch": 1.989867747440273, + "grad_norm": 0.48881974816322327, + "learning_rate": 3.073045499800828e-06, + "loss": 0.4314, + "step": 6219 + }, + { + "epoch": 1.99018771331058, + "grad_norm": 0.5957432389259338, + "learning_rate": 3.071327655869804e-06, + "loss": 0.5567, + "step": 6220 + }, + { + "epoch": 1.9905076791808873, + "grad_norm": 0.5555700659751892, + "learning_rate": 3.069610079351519e-06, + "loss": 0.4873, + "step": 6221 + }, + { + "epoch": 1.9908276450511946, + "grad_norm": 0.4987030327320099, + "learning_rate": 3.067892770484119e-06, + "loss": 0.4762, + "step": 6222 + }, + { + "epoch": 1.9911476109215016, + "grad_norm": 0.5237892270088196, + "learning_rate": 3.066175729505712e-06, + "loss": 0.491, + "step": 6223 + }, + { + "epoch": 1.9914675767918089, + "grad_norm": 0.46969321370124817, + "learning_rate": 3.064458956654366e-06, + "loss": 0.4579, + "step": 6224 + }, + { + "epoch": 1.991787542662116, + "grad_norm": 0.5169256329536438, + "learning_rate": 3.062742452168115e-06, + "loss": 0.5175, + "step": 6225 + }, + { + "epoch": 1.9921075085324231, + "grad_norm": 0.5002074837684631, + "learning_rate": 3.061026216284952e-06, + "loss": 0.4797, + "step": 6226 + }, + { + "epoch": 1.9924274744027304, + "grad_norm": 0.49317920207977295, + "learning_rate": 3.059310249242836e-06, + "loss": 0.5041, + "step": 6227 + }, + { + "epoch": 1.9927474402730376, + "grad_norm": 0.4308440387248993, + "learning_rate": 3.057594551279691e-06, + "loss": 0.4721, + "step": 6228 + }, + { + "epoch": 1.9930674061433447, + "grad_norm": 0.49936333298683167, + "learning_rate": 3.0558791226333974e-06, + "loss": 0.527, + "step": 6229 + }, + { + "epoch": 1.993387372013652, + "grad_norm": 0.534095048904419, + "learning_rate": 3.054163963541803e-06, + "loss": 0.4522, + "step": 6230 + }, + { + "epoch": 1.9937073378839592, + "grad_norm": 0.5647724270820618, + "learning_rate": 3.052449074242715e-06, + "loss": 0.5229, + "step": 6231 + }, + { + "epoch": 1.9940273037542662, + "grad_norm": 0.4887125492095947, + "learning_rate": 3.050734454973904e-06, + "loss": 0.4598, + "step": 6232 + }, + { + "epoch": 1.9943472696245734, + "grad_norm": 0.5863474607467651, + "learning_rate": 3.049020105973109e-06, + "loss": 0.5209, + "step": 6233 + }, + { + "epoch": 1.9946672354948807, + "grad_norm": 0.5121579170227051, + "learning_rate": 3.0473060274780217e-06, + "loss": 0.4353, + "step": 6234 + }, + { + "epoch": 1.9949872013651877, + "grad_norm": 0.5075271129608154, + "learning_rate": 3.0455922197263044e-06, + "loss": 0.5176, + "step": 6235 + }, + { + "epoch": 1.9953071672354947, + "grad_norm": 0.5416887402534485, + "learning_rate": 3.0438786829555757e-06, + "loss": 0.5651, + "step": 6236 + }, + { + "epoch": 1.9956271331058022, + "grad_norm": 0.4627353847026825, + "learning_rate": 3.04216541740342e-06, + "loss": 0.4136, + "step": 6237 + }, + { + "epoch": 1.9959470989761092, + "grad_norm": 0.5104613304138184, + "learning_rate": 3.0404524233073864e-06, + "loss": 0.4476, + "step": 6238 + }, + { + "epoch": 1.9962670648464163, + "grad_norm": 0.5518894195556641, + "learning_rate": 3.038739700904981e-06, + "loss": 0.5122, + "step": 6239 + }, + { + "epoch": 1.9965870307167235, + "grad_norm": 0.5062071681022644, + "learning_rate": 3.037027250433676e-06, + "loss": 0.4557, + "step": 6240 + }, + { + "epoch": 1.9969069965870307, + "grad_norm": 0.5367639064788818, + "learning_rate": 3.035315072130902e-06, + "loss": 0.5672, + "step": 6241 + }, + { + "epoch": 1.9972269624573378, + "grad_norm": 0.4900611937046051, + "learning_rate": 3.033603166234059e-06, + "loss": 0.3972, + "step": 6242 + }, + { + "epoch": 1.997546928327645, + "grad_norm": 0.5525842905044556, + "learning_rate": 3.031891532980503e-06, + "loss": 0.5552, + "step": 6243 + }, + { + "epoch": 1.9978668941979523, + "grad_norm": 0.5071797966957092, + "learning_rate": 3.030180172607552e-06, + "loss": 0.4945, + "step": 6244 + }, + { + "epoch": 1.9981868600682593, + "grad_norm": 0.490200936794281, + "learning_rate": 3.0284690853524918e-06, + "loss": 0.481, + "step": 6245 + }, + { + "epoch": 1.9985068259385665, + "grad_norm": 0.5068796277046204, + "learning_rate": 3.0267582714525624e-06, + "loss": 0.4779, + "step": 6246 + }, + { + "epoch": 1.9988267918088738, + "grad_norm": 0.5391877889633179, + "learning_rate": 3.0250477311449734e-06, + "loss": 0.5154, + "step": 6247 + }, + { + "epoch": 1.9991467576791808, + "grad_norm": 0.5023802518844604, + "learning_rate": 3.0233374646668935e-06, + "loss": 0.4056, + "step": 6248 + }, + { + "epoch": 1.999466723549488, + "grad_norm": 0.49331384897232056, + "learning_rate": 3.0216274722554513e-06, + "loss": 0.5042, + "step": 6249 + }, + { + "epoch": 1.9997866894197953, + "grad_norm": 0.4648846387863159, + "learning_rate": 3.019917754147741e-06, + "loss": 0.485, + "step": 6250 + }, + { + "epoch": 2.0001066552901023, + "grad_norm": 1.401410460472107, + "learning_rate": 3.0182083105808135e-06, + "loss": 0.8008, + "step": 6251 + }, + { + "epoch": 2.0004266211604094, + "grad_norm": 0.5558966398239136, + "learning_rate": 3.01649914179169e-06, + "loss": 0.5132, + "step": 6252 + }, + { + "epoch": 2.000746587030717, + "grad_norm": 0.5161656737327576, + "learning_rate": 3.014790248017347e-06, + "loss": 0.4847, + "step": 6253 + }, + { + "epoch": 2.001066552901024, + "grad_norm": 0.45154380798339844, + "learning_rate": 3.0130816294947233e-06, + "loss": 0.3873, + "step": 6254 + }, + { + "epoch": 2.001386518771331, + "grad_norm": 0.4930065870285034, + "learning_rate": 3.0113732864607236e-06, + "loss": 0.4263, + "step": 6255 + }, + { + "epoch": 2.0017064846416384, + "grad_norm": 0.5245649218559265, + "learning_rate": 3.009665219152208e-06, + "loss": 0.4314, + "step": 6256 + }, + { + "epoch": 2.0020264505119454, + "grad_norm": 0.5030341744422913, + "learning_rate": 3.007957427806004e-06, + "loss": 0.4453, + "step": 6257 + }, + { + "epoch": 2.0023464163822524, + "grad_norm": 0.5460236668586731, + "learning_rate": 3.0062499126589018e-06, + "loss": 0.4641, + "step": 6258 + }, + { + "epoch": 2.00266638225256, + "grad_norm": 0.5252211689949036, + "learning_rate": 3.004542673947646e-06, + "loss": 0.4158, + "step": 6259 + }, + { + "epoch": 2.002986348122867, + "grad_norm": 0.544998824596405, + "learning_rate": 3.00283571190895e-06, + "loss": 0.4475, + "step": 6260 + }, + { + "epoch": 2.003306313993174, + "grad_norm": 0.5802964568138123, + "learning_rate": 3.0011290267794833e-06, + "loss": 0.4916, + "step": 6261 + }, + { + "epoch": 2.0036262798634814, + "grad_norm": 0.5228466987609863, + "learning_rate": 2.9994226187958824e-06, + "loss": 0.4407, + "step": 6262 + }, + { + "epoch": 2.0039462457337884, + "grad_norm": 0.5055839419364929, + "learning_rate": 2.997716488194744e-06, + "loss": 0.3993, + "step": 6263 + }, + { + "epoch": 2.0042662116040955, + "grad_norm": 0.5381371378898621, + "learning_rate": 2.9960106352126217e-06, + "loss": 0.4764, + "step": 6264 + }, + { + "epoch": 2.004586177474403, + "grad_norm": 0.4837213456630707, + "learning_rate": 2.994305060086037e-06, + "loss": 0.4731, + "step": 6265 + }, + { + "epoch": 2.00490614334471, + "grad_norm": 0.5261613130569458, + "learning_rate": 2.9925997630514658e-06, + "loss": 0.4017, + "step": 6266 + }, + { + "epoch": 2.005226109215017, + "grad_norm": 0.5161585211753845, + "learning_rate": 2.990894744345354e-06, + "loss": 0.4647, + "step": 6267 + }, + { + "epoch": 2.0055460750853245, + "grad_norm": 0.5065637230873108, + "learning_rate": 2.9891900042041043e-06, + "loss": 0.4401, + "step": 6268 + }, + { + "epoch": 2.0058660409556315, + "grad_norm": 0.5628411173820496, + "learning_rate": 2.9874855428640783e-06, + "loss": 0.4596, + "step": 6269 + }, + { + "epoch": 2.0061860068259385, + "grad_norm": 0.5215651392936707, + "learning_rate": 2.9857813605616036e-06, + "loss": 0.4688, + "step": 6270 + }, + { + "epoch": 2.0065059726962455, + "grad_norm": 0.5510452389717102, + "learning_rate": 2.9840774575329644e-06, + "loss": 0.4361, + "step": 6271 + }, + { + "epoch": 2.006825938566553, + "grad_norm": 0.5135353207588196, + "learning_rate": 2.982373834014412e-06, + "loss": 0.398, + "step": 6272 + }, + { + "epoch": 2.00714590443686, + "grad_norm": 0.5171473622322083, + "learning_rate": 2.9806704902421557e-06, + "loss": 0.4695, + "step": 6273 + }, + { + "epoch": 2.007465870307167, + "grad_norm": 0.49266937375068665, + "learning_rate": 2.978967426452365e-06, + "loss": 0.4735, + "step": 6274 + }, + { + "epoch": 2.0077858361774745, + "grad_norm": 0.5164386630058289, + "learning_rate": 2.97726464288117e-06, + "loss": 0.4332, + "step": 6275 + }, + { + "epoch": 2.0081058020477816, + "grad_norm": 0.510276198387146, + "learning_rate": 2.9755621397646682e-06, + "loss": 0.4564, + "step": 6276 + }, + { + "epoch": 2.0084257679180886, + "grad_norm": 0.5183811187744141, + "learning_rate": 2.9738599173389104e-06, + "loss": 0.4232, + "step": 6277 + }, + { + "epoch": 2.008745733788396, + "grad_norm": 0.48706677556037903, + "learning_rate": 2.9721579758399134e-06, + "loss": 0.4091, + "step": 6278 + }, + { + "epoch": 2.009065699658703, + "grad_norm": 0.465788871049881, + "learning_rate": 2.9704563155036515e-06, + "loss": 0.4329, + "step": 6279 + }, + { + "epoch": 2.00938566552901, + "grad_norm": 0.5820158123970032, + "learning_rate": 2.968754936566062e-06, + "loss": 0.4563, + "step": 6280 + }, + { + "epoch": 2.0097056313993176, + "grad_norm": 0.5367518067359924, + "learning_rate": 2.967053839263046e-06, + "loss": 0.4055, + "step": 6281 + }, + { + "epoch": 2.0100255972696246, + "grad_norm": 0.5321242213249207, + "learning_rate": 2.9653530238304603e-06, + "loss": 0.4539, + "step": 6282 + }, + { + "epoch": 2.0103455631399316, + "grad_norm": 0.5166159272193909, + "learning_rate": 2.9636524905041264e-06, + "loss": 0.4784, + "step": 6283 + }, + { + "epoch": 2.010665529010239, + "grad_norm": 0.4449666738510132, + "learning_rate": 2.961952239519823e-06, + "loss": 0.4355, + "step": 6284 + }, + { + "epoch": 2.010985494880546, + "grad_norm": 0.5288645029067993, + "learning_rate": 2.9602522711132932e-06, + "loss": 0.4802, + "step": 6285 + }, + { + "epoch": 2.011305460750853, + "grad_norm": 0.5091286301612854, + "learning_rate": 2.9585525855202415e-06, + "loss": 0.43, + "step": 6286 + }, + { + "epoch": 2.0116254266211606, + "grad_norm": 0.5545825958251953, + "learning_rate": 2.9568531829763294e-06, + "loss": 0.413, + "step": 6287 + }, + { + "epoch": 2.0119453924914676, + "grad_norm": 0.5017845630645752, + "learning_rate": 2.9551540637171815e-06, + "loss": 0.4257, + "step": 6288 + }, + { + "epoch": 2.0122653583617747, + "grad_norm": 0.49946045875549316, + "learning_rate": 2.953455227978382e-06, + "loss": 0.412, + "step": 6289 + }, + { + "epoch": 2.0125853242320817, + "grad_norm": 0.5081928968429565, + "learning_rate": 2.9517566759954762e-06, + "loss": 0.458, + "step": 6290 + }, + { + "epoch": 2.012905290102389, + "grad_norm": 0.5049950480461121, + "learning_rate": 2.950058408003973e-06, + "loss": 0.4547, + "step": 6291 + }, + { + "epoch": 2.013225255972696, + "grad_norm": 0.5251578092575073, + "learning_rate": 2.9483604242393372e-06, + "loss": 0.4132, + "step": 6292 + }, + { + "epoch": 2.013545221843003, + "grad_norm": 0.5312917828559875, + "learning_rate": 2.946662724936998e-06, + "loss": 0.4707, + "step": 6293 + }, + { + "epoch": 2.0138651877133107, + "grad_norm": 0.5079889297485352, + "learning_rate": 2.9449653103323405e-06, + "loss": 0.468, + "step": 6294 + }, + { + "epoch": 2.0141851535836177, + "grad_norm": 0.5116049647331238, + "learning_rate": 2.9432681806607145e-06, + "loss": 0.4347, + "step": 6295 + }, + { + "epoch": 2.0145051194539247, + "grad_norm": 0.4871806800365448, + "learning_rate": 2.941571336157432e-06, + "loss": 0.4103, + "step": 6296 + }, + { + "epoch": 2.014825085324232, + "grad_norm": 0.5868577361106873, + "learning_rate": 2.939874777057758e-06, + "loss": 0.4955, + "step": 6297 + }, + { + "epoch": 2.0151450511945392, + "grad_norm": 0.5232118368148804, + "learning_rate": 2.938178503596926e-06, + "loss": 0.4416, + "step": 6298 + }, + { + "epoch": 2.0154650170648463, + "grad_norm": 0.5466344356536865, + "learning_rate": 2.936482516010124e-06, + "loss": 0.4367, + "step": 6299 + }, + { + "epoch": 2.0157849829351537, + "grad_norm": 0.5040996670722961, + "learning_rate": 2.934786814532502e-06, + "loss": 0.5026, + "step": 6300 + }, + { + "epoch": 2.0161049488054608, + "grad_norm": 0.4805614650249481, + "learning_rate": 2.9330913993991755e-06, + "loss": 0.4239, + "step": 6301 + }, + { + "epoch": 2.016424914675768, + "grad_norm": 0.5113465189933777, + "learning_rate": 2.9313962708452116e-06, + "loss": 0.4594, + "step": 6302 + }, + { + "epoch": 2.0167448805460753, + "grad_norm": 0.5428706407546997, + "learning_rate": 2.9297014291056446e-06, + "loss": 0.4309, + "step": 6303 + }, + { + "epoch": 2.0170648464163823, + "grad_norm": 0.5894550085067749, + "learning_rate": 2.9280068744154645e-06, + "loss": 0.5033, + "step": 6304 + }, + { + "epoch": 2.0173848122866893, + "grad_norm": 0.5202749967575073, + "learning_rate": 2.9263126070096222e-06, + "loss": 0.45, + "step": 6305 + }, + { + "epoch": 2.017704778156997, + "grad_norm": 0.47046372294425964, + "learning_rate": 2.9246186271230335e-06, + "loss": 0.4595, + "step": 6306 + }, + { + "epoch": 2.018024744027304, + "grad_norm": 0.4563703238964081, + "learning_rate": 2.9229249349905686e-06, + "loss": 0.4495, + "step": 6307 + }, + { + "epoch": 2.018344709897611, + "grad_norm": 0.47942081093788147, + "learning_rate": 2.921231530847061e-06, + "loss": 0.4429, + "step": 6308 + }, + { + "epoch": 2.0186646757679183, + "grad_norm": 0.47403234243392944, + "learning_rate": 2.9195384149273e-06, + "loss": 0.3801, + "step": 6309 + }, + { + "epoch": 2.0189846416382253, + "grad_norm": 0.4782538115978241, + "learning_rate": 2.9178455874660423e-06, + "loss": 0.4318, + "step": 6310 + }, + { + "epoch": 2.0193046075085324, + "grad_norm": 0.517080545425415, + "learning_rate": 2.9161530486979993e-06, + "loss": 0.4635, + "step": 6311 + }, + { + "epoch": 2.0196245733788394, + "grad_norm": 0.504300057888031, + "learning_rate": 2.9144607988578433e-06, + "loss": 0.453, + "step": 6312 + }, + { + "epoch": 2.019944539249147, + "grad_norm": 0.45974022150039673, + "learning_rate": 2.9127688381802043e-06, + "loss": 0.415, + "step": 6313 + }, + { + "epoch": 2.020264505119454, + "grad_norm": 0.5267173647880554, + "learning_rate": 2.9110771668996772e-06, + "loss": 0.4255, + "step": 6314 + }, + { + "epoch": 2.020584470989761, + "grad_norm": 0.5032655000686646, + "learning_rate": 2.909385785250816e-06, + "loss": 0.4436, + "step": 6315 + }, + { + "epoch": 2.0209044368600684, + "grad_norm": 0.4646185040473938, + "learning_rate": 2.9076946934681304e-06, + "loss": 0.4413, + "step": 6316 + }, + { + "epoch": 2.0212244027303754, + "grad_norm": 0.4995832145214081, + "learning_rate": 2.9060038917860928e-06, + "loss": 0.4442, + "step": 6317 + }, + { + "epoch": 2.0215443686006824, + "grad_norm": 0.5385740995407104, + "learning_rate": 2.904313380439132e-06, + "loss": 0.4534, + "step": 6318 + }, + { + "epoch": 2.02186433447099, + "grad_norm": 0.5125446319580078, + "learning_rate": 2.9026231596616435e-06, + "loss": 0.428, + "step": 6319 + }, + { + "epoch": 2.022184300341297, + "grad_norm": 0.5198769569396973, + "learning_rate": 2.900933229687978e-06, + "loss": 0.4173, + "step": 6320 + }, + { + "epoch": 2.022504266211604, + "grad_norm": 0.5717306733131409, + "learning_rate": 2.899243590752446e-06, + "loss": 0.4562, + "step": 6321 + }, + { + "epoch": 2.0228242320819114, + "grad_norm": 0.4866477847099304, + "learning_rate": 2.8975542430893177e-06, + "loss": 0.4519, + "step": 6322 + }, + { + "epoch": 2.0231441979522184, + "grad_norm": 0.5071144700050354, + "learning_rate": 2.8958651869328202e-06, + "loss": 0.42, + "step": 6323 + }, + { + "epoch": 2.0234641638225255, + "grad_norm": 0.45792296528816223, + "learning_rate": 2.8941764225171466e-06, + "loss": 0.4365, + "step": 6324 + }, + { + "epoch": 2.023784129692833, + "grad_norm": 0.501045823097229, + "learning_rate": 2.892487950076447e-06, + "loss": 0.4774, + "step": 6325 + }, + { + "epoch": 2.02410409556314, + "grad_norm": 0.4918345808982849, + "learning_rate": 2.890799769844829e-06, + "loss": 0.463, + "step": 6326 + }, + { + "epoch": 2.024424061433447, + "grad_norm": 0.512128472328186, + "learning_rate": 2.8891118820563575e-06, + "loss": 0.4066, + "step": 6327 + }, + { + "epoch": 2.0247440273037545, + "grad_norm": 0.5132451057434082, + "learning_rate": 2.8874242869450655e-06, + "loss": 0.5126, + "step": 6328 + }, + { + "epoch": 2.0250639931740615, + "grad_norm": 0.4674147069454193, + "learning_rate": 2.885736984744935e-06, + "loss": 0.4349, + "step": 6329 + }, + { + "epoch": 2.0253839590443685, + "grad_norm": 0.4867783784866333, + "learning_rate": 2.8840499756899176e-06, + "loss": 0.4214, + "step": 6330 + }, + { + "epoch": 2.0257039249146755, + "grad_norm": 0.6133652925491333, + "learning_rate": 2.8823632600139166e-06, + "loss": 0.4576, + "step": 6331 + }, + { + "epoch": 2.026023890784983, + "grad_norm": 0.5312769412994385, + "learning_rate": 2.8806768379507953e-06, + "loss": 0.4597, + "step": 6332 + }, + { + "epoch": 2.02634385665529, + "grad_norm": 0.49776095151901245, + "learning_rate": 2.8789907097343818e-06, + "loss": 0.4206, + "step": 6333 + }, + { + "epoch": 2.026663822525597, + "grad_norm": 0.49100378155708313, + "learning_rate": 2.8773048755984567e-06, + "loss": 0.4526, + "step": 6334 + }, + { + "epoch": 2.0269837883959045, + "grad_norm": 0.5101442933082581, + "learning_rate": 2.875619335776765e-06, + "loss": 0.4395, + "step": 6335 + }, + { + "epoch": 2.0273037542662116, + "grad_norm": 0.48366159200668335, + "learning_rate": 2.8739340905030087e-06, + "loss": 0.4498, + "step": 6336 + }, + { + "epoch": 2.0276237201365186, + "grad_norm": 0.5114291906356812, + "learning_rate": 2.8722491400108464e-06, + "loss": 0.444, + "step": 6337 + }, + { + "epoch": 2.027943686006826, + "grad_norm": 0.5394448637962341, + "learning_rate": 2.870564484533903e-06, + "loss": 0.4696, + "step": 6338 + }, + { + "epoch": 2.028263651877133, + "grad_norm": 0.5102099776268005, + "learning_rate": 2.8688801243057536e-06, + "loss": 0.3991, + "step": 6339 + }, + { + "epoch": 2.02858361774744, + "grad_norm": 0.5134938955307007, + "learning_rate": 2.86719605955994e-06, + "loss": 0.4225, + "step": 6340 + }, + { + "epoch": 2.0289035836177476, + "grad_norm": 0.6032628417015076, + "learning_rate": 2.865512290529959e-06, + "loss": 0.4376, + "step": 6341 + }, + { + "epoch": 2.0292235494880546, + "grad_norm": 0.5756434798240662, + "learning_rate": 2.8638288174492657e-06, + "loss": 0.4826, + "step": 6342 + }, + { + "epoch": 2.0295435153583616, + "grad_norm": 0.516460657119751, + "learning_rate": 2.8621456405512775e-06, + "loss": 0.4089, + "step": 6343 + }, + { + "epoch": 2.029863481228669, + "grad_norm": 0.5271695852279663, + "learning_rate": 2.860462760069367e-06, + "loss": 0.4177, + "step": 6344 + }, + { + "epoch": 2.030183447098976, + "grad_norm": 0.5140413045883179, + "learning_rate": 2.8587801762368695e-06, + "loss": 0.4658, + "step": 6345 + }, + { + "epoch": 2.030503412969283, + "grad_norm": 0.48984235525131226, + "learning_rate": 2.8570978892870777e-06, + "loss": 0.4598, + "step": 6346 + }, + { + "epoch": 2.0308233788395906, + "grad_norm": 0.4675515294075012, + "learning_rate": 2.855415899453239e-06, + "loss": 0.4044, + "step": 6347 + }, + { + "epoch": 2.0311433447098977, + "grad_norm": 0.4961652457714081, + "learning_rate": 2.8537342069685657e-06, + "loss": 0.4533, + "step": 6348 + }, + { + "epoch": 2.0314633105802047, + "grad_norm": 0.5445696115493774, + "learning_rate": 2.8520528120662296e-06, + "loss": 0.4365, + "step": 6349 + }, + { + "epoch": 2.031783276450512, + "grad_norm": 0.5242259502410889, + "learning_rate": 2.8503717149793543e-06, + "loss": 0.4994, + "step": 6350 + }, + { + "epoch": 2.032103242320819, + "grad_norm": 0.5333598256111145, + "learning_rate": 2.8486909159410266e-06, + "loss": 0.4286, + "step": 6351 + }, + { + "epoch": 2.032423208191126, + "grad_norm": 0.48089703917503357, + "learning_rate": 2.847010415184289e-06, + "loss": 0.4337, + "step": 6352 + }, + { + "epoch": 2.0327431740614332, + "grad_norm": 0.5590294003486633, + "learning_rate": 2.845330212942149e-06, + "loss": 0.4655, + "step": 6353 + }, + { + "epoch": 2.0330631399317407, + "grad_norm": 0.5742276906967163, + "learning_rate": 2.843650309447568e-06, + "loss": 0.4868, + "step": 6354 + }, + { + "epoch": 2.0333831058020477, + "grad_norm": 0.504341185092926, + "learning_rate": 2.8419707049334664e-06, + "loss": 0.4236, + "step": 6355 + }, + { + "epoch": 2.0337030716723548, + "grad_norm": 0.5049486756324768, + "learning_rate": 2.8402913996327217e-06, + "loss": 0.406, + "step": 6356 + }, + { + "epoch": 2.0340230375426622, + "grad_norm": 0.5707356333732605, + "learning_rate": 2.838612393778172e-06, + "loss": 0.4955, + "step": 6357 + }, + { + "epoch": 2.0343430034129693, + "grad_norm": 0.49165648221969604, + "learning_rate": 2.8369336876026133e-06, + "loss": 0.426, + "step": 6358 + }, + { + "epoch": 2.0346629692832763, + "grad_norm": 0.5239745378494263, + "learning_rate": 2.8352552813388035e-06, + "loss": 0.4673, + "step": 6359 + }, + { + "epoch": 2.0349829351535837, + "grad_norm": 0.5204604268074036, + "learning_rate": 2.833577175219453e-06, + "loss": 0.4028, + "step": 6360 + }, + { + "epoch": 2.0353029010238908, + "grad_norm": 0.5317163467407227, + "learning_rate": 2.831899369477233e-06, + "loss": 0.4746, + "step": 6361 + }, + { + "epoch": 2.035622866894198, + "grad_norm": 0.46956390142440796, + "learning_rate": 2.830221864344772e-06, + "loss": 0.4173, + "step": 6362 + }, + { + "epoch": 2.0359428327645053, + "grad_norm": 0.5055503845214844, + "learning_rate": 2.8285446600546594e-06, + "loss": 0.4398, + "step": 6363 + }, + { + "epoch": 2.0362627986348123, + "grad_norm": 0.5239434242248535, + "learning_rate": 2.8268677568394456e-06, + "loss": 0.447, + "step": 6364 + }, + { + "epoch": 2.0365827645051193, + "grad_norm": 0.5025309324264526, + "learning_rate": 2.8251911549316303e-06, + "loss": 0.4443, + "step": 6365 + }, + { + "epoch": 2.036902730375427, + "grad_norm": 0.4989284873008728, + "learning_rate": 2.8235148545636776e-06, + "loss": 0.4252, + "step": 6366 + }, + { + "epoch": 2.037222696245734, + "grad_norm": 0.49771973490715027, + "learning_rate": 2.8218388559680065e-06, + "loss": 0.5184, + "step": 6367 + }, + { + "epoch": 2.037542662116041, + "grad_norm": 0.47321969270706177, + "learning_rate": 2.8201631593769995e-06, + "loss": 0.3978, + "step": 6368 + }, + { + "epoch": 2.0378626279863483, + "grad_norm": 0.5439449548721313, + "learning_rate": 2.818487765022994e-06, + "loss": 0.4636, + "step": 6369 + }, + { + "epoch": 2.0381825938566553, + "grad_norm": 0.5097255706787109, + "learning_rate": 2.816812673138285e-06, + "loss": 0.4238, + "step": 6370 + }, + { + "epoch": 2.0385025597269624, + "grad_norm": 0.5556920766830444, + "learning_rate": 2.8151378839551248e-06, + "loss": 0.4705, + "step": 6371 + }, + { + "epoch": 2.03882252559727, + "grad_norm": 0.5774846076965332, + "learning_rate": 2.8134633977057236e-06, + "loss": 0.4399, + "step": 6372 + }, + { + "epoch": 2.039142491467577, + "grad_norm": 0.5274468064308167, + "learning_rate": 2.8117892146222524e-06, + "loss": 0.4709, + "step": 6373 + }, + { + "epoch": 2.039462457337884, + "grad_norm": 0.4775089621543884, + "learning_rate": 2.8101153349368417e-06, + "loss": 0.4058, + "step": 6374 + }, + { + "epoch": 2.039782423208191, + "grad_norm": 0.530306875705719, + "learning_rate": 2.808441758881574e-06, + "loss": 0.4554, + "step": 6375 + }, + { + "epoch": 2.0401023890784984, + "grad_norm": 0.4880123436450958, + "learning_rate": 2.8067684866884927e-06, + "loss": 0.4235, + "step": 6376 + }, + { + "epoch": 2.0404223549488054, + "grad_norm": 0.5186628699302673, + "learning_rate": 2.8050955185895977e-06, + "loss": 0.4407, + "step": 6377 + }, + { + "epoch": 2.0407423208191124, + "grad_norm": 0.5505031943321228, + "learning_rate": 2.8034228548168496e-06, + "loss": 0.4116, + "step": 6378 + }, + { + "epoch": 2.04106228668942, + "grad_norm": 0.5448146462440491, + "learning_rate": 2.801750495602167e-06, + "loss": 0.4322, + "step": 6379 + }, + { + "epoch": 2.041382252559727, + "grad_norm": 0.5421686768531799, + "learning_rate": 2.800078441177423e-06, + "loss": 0.4143, + "step": 6380 + }, + { + "epoch": 2.041702218430034, + "grad_norm": 0.489848792552948, + "learning_rate": 2.798406691774449e-06, + "loss": 0.4667, + "step": 6381 + }, + { + "epoch": 2.0420221843003414, + "grad_norm": 0.5021038055419922, + "learning_rate": 2.7967352476250344e-06, + "loss": 0.4193, + "step": 6382 + }, + { + "epoch": 2.0423421501706485, + "grad_norm": 0.503178060054779, + "learning_rate": 2.7950641089609275e-06, + "loss": 0.452, + "step": 6383 + }, + { + "epoch": 2.0426621160409555, + "grad_norm": 0.49223408102989197, + "learning_rate": 2.7933932760138375e-06, + "loss": 0.4555, + "step": 6384 + }, + { + "epoch": 2.042982081911263, + "grad_norm": 0.4962237775325775, + "learning_rate": 2.791722749015424e-06, + "loss": 0.4129, + "step": 6385 + }, + { + "epoch": 2.04330204778157, + "grad_norm": 0.5070708394050598, + "learning_rate": 2.7900525281973078e-06, + "loss": 0.4746, + "step": 6386 + }, + { + "epoch": 2.043622013651877, + "grad_norm": 0.4765138328075409, + "learning_rate": 2.788382613791066e-06, + "loss": 0.4022, + "step": 6387 + }, + { + "epoch": 2.0439419795221845, + "grad_norm": 0.5285912156105042, + "learning_rate": 2.7867130060282345e-06, + "loss": 0.471, + "step": 6388 + }, + { + "epoch": 2.0442619453924915, + "grad_norm": 0.506458580493927, + "learning_rate": 2.7850437051403102e-06, + "loss": 0.4715, + "step": 6389 + }, + { + "epoch": 2.0445819112627985, + "grad_norm": 0.5261322855949402, + "learning_rate": 2.7833747113587394e-06, + "loss": 0.4661, + "step": 6390 + }, + { + "epoch": 2.044901877133106, + "grad_norm": 0.520087718963623, + "learning_rate": 2.781706024914933e-06, + "loss": 0.4166, + "step": 6391 + }, + { + "epoch": 2.045221843003413, + "grad_norm": 0.5601601600646973, + "learning_rate": 2.7800376460402523e-06, + "loss": 0.4705, + "step": 6392 + }, + { + "epoch": 2.04554180887372, + "grad_norm": 0.484840989112854, + "learning_rate": 2.7783695749660228e-06, + "loss": 0.4287, + "step": 6393 + }, + { + "epoch": 2.0458617747440275, + "grad_norm": 0.5149590373039246, + "learning_rate": 2.7767018119235263e-06, + "loss": 0.4412, + "step": 6394 + }, + { + "epoch": 2.0461817406143346, + "grad_norm": 0.4719947278499603, + "learning_rate": 2.7750343571439987e-06, + "loss": 0.4471, + "step": 6395 + }, + { + "epoch": 2.0465017064846416, + "grad_norm": 0.476144015789032, + "learning_rate": 2.7733672108586333e-06, + "loss": 0.458, + "step": 6396 + }, + { + "epoch": 2.0468216723549486, + "grad_norm": 0.5308468341827393, + "learning_rate": 2.7717003732985813e-06, + "loss": 0.4377, + "step": 6397 + }, + { + "epoch": 2.047141638225256, + "grad_norm": 0.5079355835914612, + "learning_rate": 2.770033844694954e-06, + "loss": 0.4276, + "step": 6398 + }, + { + "epoch": 2.047461604095563, + "grad_norm": 0.4922444224357605, + "learning_rate": 2.7683676252788172e-06, + "loss": 0.4352, + "step": 6399 + }, + { + "epoch": 2.04778156996587, + "grad_norm": 0.5174881219863892, + "learning_rate": 2.7667017152811947e-06, + "loss": 0.4665, + "step": 6400 + }, + { + "epoch": 2.0481015358361776, + "grad_norm": 0.5103611350059509, + "learning_rate": 2.7650361149330662e-06, + "loss": 0.4197, + "step": 6401 + }, + { + "epoch": 2.0484215017064846, + "grad_norm": 0.5468028783798218, + "learning_rate": 2.7633708244653677e-06, + "loss": 0.4401, + "step": 6402 + }, + { + "epoch": 2.0487414675767917, + "grad_norm": 0.5299778580665588, + "learning_rate": 2.7617058441089943e-06, + "loss": 0.454, + "step": 6403 + }, + { + "epoch": 2.049061433447099, + "grad_norm": 0.5115585327148438, + "learning_rate": 2.7600411740948007e-06, + "loss": 0.4152, + "step": 6404 + }, + { + "epoch": 2.049381399317406, + "grad_norm": 0.4773969054222107, + "learning_rate": 2.7583768146535927e-06, + "loss": 0.4427, + "step": 6405 + }, + { + "epoch": 2.049701365187713, + "grad_norm": 0.4479951560497284, + "learning_rate": 2.7567127660161366e-06, + "loss": 0.3929, + "step": 6406 + }, + { + "epoch": 2.0500213310580206, + "grad_norm": 0.4836898446083069, + "learning_rate": 2.7550490284131516e-06, + "loss": 0.4789, + "step": 6407 + }, + { + "epoch": 2.0503412969283277, + "grad_norm": 0.5370376110076904, + "learning_rate": 2.75338560207532e-06, + "loss": 0.4708, + "step": 6408 + }, + { + "epoch": 2.0506612627986347, + "grad_norm": 0.4753120541572571, + "learning_rate": 2.7517224872332794e-06, + "loss": 0.3785, + "step": 6409 + }, + { + "epoch": 2.050981228668942, + "grad_norm": 0.5000243782997131, + "learning_rate": 2.7500596841176207e-06, + "loss": 0.4492, + "step": 6410 + }, + { + "epoch": 2.051301194539249, + "grad_norm": 0.5311595797538757, + "learning_rate": 2.748397192958893e-06, + "loss": 0.4689, + "step": 6411 + }, + { + "epoch": 2.051621160409556, + "grad_norm": 0.5121088624000549, + "learning_rate": 2.746735013987602e-06, + "loss": 0.4071, + "step": 6412 + }, + { + "epoch": 2.0519411262798637, + "grad_norm": 0.5043737292289734, + "learning_rate": 2.745073147434212e-06, + "loss": 0.4012, + "step": 6413 + }, + { + "epoch": 2.0522610921501707, + "grad_norm": 0.5759963393211365, + "learning_rate": 2.743411593529144e-06, + "loss": 0.5193, + "step": 6414 + }, + { + "epoch": 2.0525810580204777, + "grad_norm": 0.48577389121055603, + "learning_rate": 2.7417503525027742e-06, + "loss": 0.4423, + "step": 6415 + }, + { + "epoch": 2.0529010238907848, + "grad_norm": 0.5283511281013489, + "learning_rate": 2.7400894245854327e-06, + "loss": 0.455, + "step": 6416 + }, + { + "epoch": 2.0532209897610922, + "grad_norm": 0.5407797694206238, + "learning_rate": 2.7384288100074124e-06, + "loss": 0.4829, + "step": 6417 + }, + { + "epoch": 2.0535409556313993, + "grad_norm": 0.4855120778083801, + "learning_rate": 2.736768508998957e-06, + "loss": 0.3815, + "step": 6418 + }, + { + "epoch": 2.0538609215017063, + "grad_norm": 0.45374947786331177, + "learning_rate": 2.735108521790273e-06, + "loss": 0.4237, + "step": 6419 + }, + { + "epoch": 2.0541808873720138, + "grad_norm": 0.5150781869888306, + "learning_rate": 2.7334488486115164e-06, + "loss": 0.447, + "step": 6420 + }, + { + "epoch": 2.054500853242321, + "grad_norm": 0.5354177355766296, + "learning_rate": 2.7317894896928015e-06, + "loss": 0.4674, + "step": 6421 + }, + { + "epoch": 2.054820819112628, + "grad_norm": 0.528226912021637, + "learning_rate": 2.7301304452642054e-06, + "loss": 0.3772, + "step": 6422 + }, + { + "epoch": 2.0551407849829353, + "grad_norm": 0.563994824886322, + "learning_rate": 2.7284717155557512e-06, + "loss": 0.4606, + "step": 6423 + }, + { + "epoch": 2.0554607508532423, + "grad_norm": 0.5252886414527893, + "learning_rate": 2.7268133007974284e-06, + "loss": 0.3824, + "step": 6424 + }, + { + "epoch": 2.0557807167235493, + "grad_norm": 0.5421047806739807, + "learning_rate": 2.7251552012191763e-06, + "loss": 0.4509, + "step": 6425 + }, + { + "epoch": 2.056100682593857, + "grad_norm": 0.5202421545982361, + "learning_rate": 2.72349741705089e-06, + "loss": 0.4586, + "step": 6426 + }, + { + "epoch": 2.056420648464164, + "grad_norm": 0.5115158557891846, + "learning_rate": 2.721839948522428e-06, + "loss": 0.4792, + "step": 6427 + }, + { + "epoch": 2.056740614334471, + "grad_norm": 0.5127246379852295, + "learning_rate": 2.720182795863595e-06, + "loss": 0.4698, + "step": 6428 + }, + { + "epoch": 2.0570605802047783, + "grad_norm": 0.5429564714431763, + "learning_rate": 2.7185259593041633e-06, + "loss": 0.4015, + "step": 6429 + }, + { + "epoch": 2.0573805460750854, + "grad_norm": 0.5757423639297485, + "learning_rate": 2.7168694390738515e-06, + "loss": 0.4856, + "step": 6430 + }, + { + "epoch": 2.0577005119453924, + "grad_norm": 0.5339115262031555, + "learning_rate": 2.7152132354023365e-06, + "loss": 0.4565, + "step": 6431 + }, + { + "epoch": 2.0580204778157, + "grad_norm": 0.49360546469688416, + "learning_rate": 2.7135573485192566e-06, + "loss": 0.4268, + "step": 6432 + }, + { + "epoch": 2.058340443686007, + "grad_norm": 0.5282648205757141, + "learning_rate": 2.7119017786541997e-06, + "loss": 0.4612, + "step": 6433 + }, + { + "epoch": 2.058660409556314, + "grad_norm": 0.4971468448638916, + "learning_rate": 2.7102465260367154e-06, + "loss": 0.3992, + "step": 6434 + }, + { + "epoch": 2.0589803754266214, + "grad_norm": 0.5643179416656494, + "learning_rate": 2.7085915908963053e-06, + "loss": 0.4679, + "step": 6435 + }, + { + "epoch": 2.0593003412969284, + "grad_norm": 0.5144991874694824, + "learning_rate": 2.706936973462425e-06, + "loss": 0.3816, + "step": 6436 + }, + { + "epoch": 2.0596203071672354, + "grad_norm": 0.5589467883110046, + "learning_rate": 2.705282673964495e-06, + "loss": 0.4387, + "step": 6437 + }, + { + "epoch": 2.0599402730375425, + "grad_norm": 0.4803876578807831, + "learning_rate": 2.70362869263188e-06, + "loss": 0.4571, + "step": 6438 + }, + { + "epoch": 2.06026023890785, + "grad_norm": 0.5414488315582275, + "learning_rate": 2.701975029693912e-06, + "loss": 0.4852, + "step": 6439 + }, + { + "epoch": 2.060580204778157, + "grad_norm": 0.5587663054466248, + "learning_rate": 2.700321685379871e-06, + "loss": 0.479, + "step": 6440 + }, + { + "epoch": 2.060900170648464, + "grad_norm": 0.4906553030014038, + "learning_rate": 2.6986686599189914e-06, + "loss": 0.4323, + "step": 6441 + }, + { + "epoch": 2.0612201365187715, + "grad_norm": 0.4768359661102295, + "learning_rate": 2.697015953540474e-06, + "loss": 0.3904, + "step": 6442 + }, + { + "epoch": 2.0615401023890785, + "grad_norm": 0.4859939217567444, + "learning_rate": 2.695363566473463e-06, + "loss": 0.4133, + "step": 6443 + }, + { + "epoch": 2.0618600682593855, + "grad_norm": 0.49089834094047546, + "learning_rate": 2.693711498947068e-06, + "loss": 0.4171, + "step": 6444 + }, + { + "epoch": 2.062180034129693, + "grad_norm": 0.5339997410774231, + "learning_rate": 2.692059751190348e-06, + "loss": 0.4976, + "step": 6445 + }, + { + "epoch": 2.0625, + "grad_norm": 0.5293613076210022, + "learning_rate": 2.690408323432318e-06, + "loss": 0.4639, + "step": 6446 + }, + { + "epoch": 2.062819965870307, + "grad_norm": 0.5344449877738953, + "learning_rate": 2.688757215901955e-06, + "loss": 0.4026, + "step": 6447 + }, + { + "epoch": 2.0631399317406145, + "grad_norm": 0.511799693107605, + "learning_rate": 2.687106428828182e-06, + "loss": 0.439, + "step": 6448 + }, + { + "epoch": 2.0634598976109215, + "grad_norm": 0.5103006362915039, + "learning_rate": 2.6854559624398867e-06, + "loss": 0.4567, + "step": 6449 + }, + { + "epoch": 2.0637798634812285, + "grad_norm": 0.4594227969646454, + "learning_rate": 2.6838058169659076e-06, + "loss": 0.4354, + "step": 6450 + }, + { + "epoch": 2.064099829351536, + "grad_norm": 0.512006402015686, + "learning_rate": 2.6821559926350362e-06, + "loss": 0.4884, + "step": 6451 + }, + { + "epoch": 2.064419795221843, + "grad_norm": 0.49301502108573914, + "learning_rate": 2.6805064896760265e-06, + "loss": 0.3972, + "step": 6452 + }, + { + "epoch": 2.06473976109215, + "grad_norm": 0.460877925157547, + "learning_rate": 2.6788573083175806e-06, + "loss": 0.4664, + "step": 6453 + }, + { + "epoch": 2.0650597269624575, + "grad_norm": 0.4712378978729248, + "learning_rate": 2.6772084487883633e-06, + "loss": 0.4449, + "step": 6454 + }, + { + "epoch": 2.0653796928327646, + "grad_norm": 0.5092011094093323, + "learning_rate": 2.6755599113169866e-06, + "loss": 0.4403, + "step": 6455 + }, + { + "epoch": 2.0656996587030716, + "grad_norm": 0.5052369236946106, + "learning_rate": 2.6739116961320266e-06, + "loss": 0.4583, + "step": 6456 + }, + { + "epoch": 2.0660196245733786, + "grad_norm": 0.5215187072753906, + "learning_rate": 2.672263803462009e-06, + "loss": 0.4734, + "step": 6457 + }, + { + "epoch": 2.066339590443686, + "grad_norm": 0.47432374954223633, + "learning_rate": 2.670616233535413e-06, + "loss": 0.3821, + "step": 6458 + }, + { + "epoch": 2.066659556313993, + "grad_norm": 0.5366644859313965, + "learning_rate": 2.6689689865806806e-06, + "loss": 0.4653, + "step": 6459 + }, + { + "epoch": 2.0669795221843, + "grad_norm": 0.5265777707099915, + "learning_rate": 2.6673220628262002e-06, + "loss": 0.4624, + "step": 6460 + }, + { + "epoch": 2.0672994880546076, + "grad_norm": 0.48621946573257446, + "learning_rate": 2.6656754625003244e-06, + "loss": 0.4698, + "step": 6461 + }, + { + "epoch": 2.0676194539249146, + "grad_norm": 0.47744280099868774, + "learning_rate": 2.6640291858313536e-06, + "loss": 0.4359, + "step": 6462 + }, + { + "epoch": 2.0679394197952217, + "grad_norm": 0.48860299587249756, + "learning_rate": 2.6623832330475454e-06, + "loss": 0.4267, + "step": 6463 + }, + { + "epoch": 2.068259385665529, + "grad_norm": 0.4812034070491791, + "learning_rate": 2.660737604377115e-06, + "loss": 0.3962, + "step": 6464 + }, + { + "epoch": 2.068579351535836, + "grad_norm": 0.5040679574012756, + "learning_rate": 2.6590923000482285e-06, + "loss": 0.4513, + "step": 6465 + }, + { + "epoch": 2.068899317406143, + "grad_norm": 0.4805178642272949, + "learning_rate": 2.6574473202890126e-06, + "loss": 0.392, + "step": 6466 + }, + { + "epoch": 2.0692192832764507, + "grad_norm": 0.5261695384979248, + "learning_rate": 2.655802665327542e-06, + "loss": 0.4297, + "step": 6467 + }, + { + "epoch": 2.0695392491467577, + "grad_norm": 0.5518275499343872, + "learning_rate": 2.6541583353918545e-06, + "loss": 0.4682, + "step": 6468 + }, + { + "epoch": 2.0698592150170647, + "grad_norm": 0.5347665548324585, + "learning_rate": 2.6525143307099353e-06, + "loss": 0.4474, + "step": 6469 + }, + { + "epoch": 2.070179180887372, + "grad_norm": 0.5325558185577393, + "learning_rate": 2.650870651509726e-06, + "loss": 0.4325, + "step": 6470 + }, + { + "epoch": 2.070499146757679, + "grad_norm": 0.5175374746322632, + "learning_rate": 2.649227298019129e-06, + "loss": 0.4919, + "step": 6471 + }, + { + "epoch": 2.0708191126279862, + "grad_norm": 0.5054546594619751, + "learning_rate": 2.6475842704659937e-06, + "loss": 0.4048, + "step": 6472 + }, + { + "epoch": 2.0711390784982937, + "grad_norm": 0.4748407304286957, + "learning_rate": 2.645941569078131e-06, + "loss": 0.4462, + "step": 6473 + }, + { + "epoch": 2.0714590443686007, + "grad_norm": 0.5580139756202698, + "learning_rate": 2.644299194083302e-06, + "loss": 0.4826, + "step": 6474 + }, + { + "epoch": 2.0717790102389078, + "grad_norm": 0.5135101079940796, + "learning_rate": 2.642657145709222e-06, + "loss": 0.4232, + "step": 6475 + }, + { + "epoch": 2.0720989761092152, + "grad_norm": 0.5293991565704346, + "learning_rate": 2.6410154241835663e-06, + "loss": 0.4532, + "step": 6476 + }, + { + "epoch": 2.0724189419795223, + "grad_norm": 0.49501514434814453, + "learning_rate": 2.6393740297339577e-06, + "loss": 0.4468, + "step": 6477 + }, + { + "epoch": 2.0727389078498293, + "grad_norm": 0.45320573449134827, + "learning_rate": 2.637732962587982e-06, + "loss": 0.4311, + "step": 6478 + }, + { + "epoch": 2.0730588737201363, + "grad_norm": 0.5398769378662109, + "learning_rate": 2.6360922229731727e-06, + "loss": 0.4698, + "step": 6479 + }, + { + "epoch": 2.073378839590444, + "grad_norm": 0.49020859599113464, + "learning_rate": 2.634451811117019e-06, + "loss": 0.4481, + "step": 6480 + }, + { + "epoch": 2.073698805460751, + "grad_norm": 0.5167035460472107, + "learning_rate": 2.632811727246969e-06, + "loss": 0.4564, + "step": 6481 + }, + { + "epoch": 2.074018771331058, + "grad_norm": 0.5227987170219421, + "learning_rate": 2.631171971590419e-06, + "loss": 0.4182, + "step": 6482 + }, + { + "epoch": 2.0743387372013653, + "grad_norm": 0.5614644289016724, + "learning_rate": 2.6295325443747272e-06, + "loss": 0.4358, + "step": 6483 + }, + { + "epoch": 2.0746587030716723, + "grad_norm": 0.5467914938926697, + "learning_rate": 2.6278934458271998e-06, + "loss": 0.4996, + "step": 6484 + }, + { + "epoch": 2.0749786689419794, + "grad_norm": 0.4813799262046814, + "learning_rate": 2.6262546761750975e-06, + "loss": 0.4124, + "step": 6485 + }, + { + "epoch": 2.075298634812287, + "grad_norm": 0.5109665393829346, + "learning_rate": 2.6246162356456423e-06, + "loss": 0.4715, + "step": 6486 + }, + { + "epoch": 2.075618600682594, + "grad_norm": 0.48429253697395325, + "learning_rate": 2.6229781244660015e-06, + "loss": 0.4657, + "step": 6487 + }, + { + "epoch": 2.075938566552901, + "grad_norm": 0.4949950873851776, + "learning_rate": 2.6213403428633055e-06, + "loss": 0.4565, + "step": 6488 + }, + { + "epoch": 2.0762585324232083, + "grad_norm": 0.47727730870246887, + "learning_rate": 2.6197028910646304e-06, + "loss": 0.3965, + "step": 6489 + }, + { + "epoch": 2.0765784982935154, + "grad_norm": 0.5116509795188904, + "learning_rate": 2.6180657692970157e-06, + "loss": 0.4626, + "step": 6490 + }, + { + "epoch": 2.0768984641638224, + "grad_norm": 0.5054216980934143, + "learning_rate": 2.6164289777874468e-06, + "loss": 0.4123, + "step": 6491 + }, + { + "epoch": 2.07721843003413, + "grad_norm": 0.4977065324783325, + "learning_rate": 2.614792516762866e-06, + "loss": 0.4229, + "step": 6492 + }, + { + "epoch": 2.077538395904437, + "grad_norm": 0.46970316767692566, + "learning_rate": 2.613156386450174e-06, + "loss": 0.4102, + "step": 6493 + }, + { + "epoch": 2.077858361774744, + "grad_norm": 0.5014659762382507, + "learning_rate": 2.6115205870762187e-06, + "loss": 0.4661, + "step": 6494 + }, + { + "epoch": 2.0781783276450514, + "grad_norm": 0.5311126112937927, + "learning_rate": 2.6098851188678096e-06, + "loss": 0.4616, + "step": 6495 + }, + { + "epoch": 2.0784982935153584, + "grad_norm": 0.5128471255302429, + "learning_rate": 2.608249982051704e-06, + "loss": 0.4308, + "step": 6496 + }, + { + "epoch": 2.0788182593856654, + "grad_norm": 0.5046076774597168, + "learning_rate": 2.606615176854613e-06, + "loss": 0.4322, + "step": 6497 + }, + { + "epoch": 2.0791382252559725, + "grad_norm": 0.5225712060928345, + "learning_rate": 2.6049807035032104e-06, + "loss": 0.481, + "step": 6498 + }, + { + "epoch": 2.07945819112628, + "grad_norm": 0.4911491870880127, + "learning_rate": 2.6033465622241116e-06, + "loss": 0.4292, + "step": 6499 + }, + { + "epoch": 2.079778156996587, + "grad_norm": 0.49582064151763916, + "learning_rate": 2.6017127532438968e-06, + "loss": 0.4471, + "step": 6500 + }, + { + "epoch": 2.080098122866894, + "grad_norm": 0.5208566784858704, + "learning_rate": 2.6000792767890947e-06, + "loss": 0.3907, + "step": 6501 + }, + { + "epoch": 2.0804180887372015, + "grad_norm": 0.5576671361923218, + "learning_rate": 2.5984461330861864e-06, + "loss": 0.4631, + "step": 6502 + }, + { + "epoch": 2.0807380546075085, + "grad_norm": 0.5146280527114868, + "learning_rate": 2.596813322361612e-06, + "loss": 0.4297, + "step": 6503 + }, + { + "epoch": 2.0810580204778155, + "grad_norm": 0.4958689510822296, + "learning_rate": 2.5951808448417603e-06, + "loss": 0.432, + "step": 6504 + }, + { + "epoch": 2.081377986348123, + "grad_norm": 0.4615541696548462, + "learning_rate": 2.593548700752979e-06, + "loss": 0.4, + "step": 6505 + }, + { + "epoch": 2.08169795221843, + "grad_norm": 0.5136170983314514, + "learning_rate": 2.5919168903215652e-06, + "loss": 0.4242, + "step": 6506 + }, + { + "epoch": 2.082017918088737, + "grad_norm": 0.5297406911849976, + "learning_rate": 2.5902854137737704e-06, + "loss": 0.4591, + "step": 6507 + }, + { + "epoch": 2.0823378839590445, + "grad_norm": 0.5089228749275208, + "learning_rate": 2.5886542713358034e-06, + "loss": 0.4652, + "step": 6508 + }, + { + "epoch": 2.0826578498293515, + "grad_norm": 0.5159381628036499, + "learning_rate": 2.587023463233821e-06, + "loss": 0.4216, + "step": 6509 + }, + { + "epoch": 2.0829778156996586, + "grad_norm": 0.5108634829521179, + "learning_rate": 2.5853929896939395e-06, + "loss": 0.4428, + "step": 6510 + }, + { + "epoch": 2.083297781569966, + "grad_norm": 0.5295661091804504, + "learning_rate": 2.5837628509422253e-06, + "loss": 0.5, + "step": 6511 + }, + { + "epoch": 2.083617747440273, + "grad_norm": 0.5104172825813293, + "learning_rate": 2.582133047204697e-06, + "loss": 0.4035, + "step": 6512 + }, + { + "epoch": 2.08393771331058, + "grad_norm": 0.5169571042060852, + "learning_rate": 2.580503578707332e-06, + "loss": 0.4415, + "step": 6513 + }, + { + "epoch": 2.0842576791808876, + "grad_norm": 0.5187350511550903, + "learning_rate": 2.5788744456760538e-06, + "loss": 0.4642, + "step": 6514 + }, + { + "epoch": 2.0845776450511946, + "grad_norm": 0.5074020624160767, + "learning_rate": 2.57724564833675e-06, + "loss": 0.4274, + "step": 6515 + }, + { + "epoch": 2.0848976109215016, + "grad_norm": 0.4734642505645752, + "learning_rate": 2.5756171869152503e-06, + "loss": 0.4001, + "step": 6516 + }, + { + "epoch": 2.085217576791809, + "grad_norm": 0.5345687866210938, + "learning_rate": 2.573989061637343e-06, + "loss": 0.4515, + "step": 6517 + }, + { + "epoch": 2.085537542662116, + "grad_norm": 0.5220152139663696, + "learning_rate": 2.5723612727287726e-06, + "loss": 0.4412, + "step": 6518 + }, + { + "epoch": 2.085857508532423, + "grad_norm": 0.48031431436538696, + "learning_rate": 2.570733820415231e-06, + "loss": 0.4634, + "step": 6519 + }, + { + "epoch": 2.0861774744027306, + "grad_norm": 0.484093576669693, + "learning_rate": 2.5691067049223683e-06, + "loss": 0.3988, + "step": 6520 + }, + { + "epoch": 2.0864974402730376, + "grad_norm": 0.5259190797805786, + "learning_rate": 2.5674799264757867e-06, + "loss": 0.4581, + "step": 6521 + }, + { + "epoch": 2.0868174061433447, + "grad_norm": 0.5825842618942261, + "learning_rate": 2.5658534853010368e-06, + "loss": 0.4392, + "step": 6522 + }, + { + "epoch": 2.0871373720136517, + "grad_norm": 0.585890531539917, + "learning_rate": 2.5642273816236298e-06, + "loss": 0.4769, + "step": 6523 + }, + { + "epoch": 2.087457337883959, + "grad_norm": 0.5051429271697998, + "learning_rate": 2.5626016156690293e-06, + "loss": 0.4291, + "step": 6524 + }, + { + "epoch": 2.087777303754266, + "grad_norm": 0.5297066569328308, + "learning_rate": 2.560976187662646e-06, + "loss": 0.4798, + "step": 6525 + }, + { + "epoch": 2.088097269624573, + "grad_norm": 0.488235205411911, + "learning_rate": 2.5593510978298487e-06, + "loss": 0.4274, + "step": 6526 + }, + { + "epoch": 2.0884172354948807, + "grad_norm": 0.46988943219184875, + "learning_rate": 2.5577263463959563e-06, + "loss": 0.4248, + "step": 6527 + }, + { + "epoch": 2.0887372013651877, + "grad_norm": 0.5096257925033569, + "learning_rate": 2.5561019335862435e-06, + "loss": 0.4614, + "step": 6528 + }, + { + "epoch": 2.0890571672354947, + "grad_norm": 0.48441529273986816, + "learning_rate": 2.55447785962594e-06, + "loss": 0.4261, + "step": 6529 + }, + { + "epoch": 2.089377133105802, + "grad_norm": 0.5083130598068237, + "learning_rate": 2.552854124740224e-06, + "loss": 0.4786, + "step": 6530 + }, + { + "epoch": 2.0896970989761092, + "grad_norm": 0.5102408528327942, + "learning_rate": 2.551230729154227e-06, + "loss": 0.4603, + "step": 6531 + }, + { + "epoch": 2.0900170648464163, + "grad_norm": 0.48783355951309204, + "learning_rate": 2.549607673093033e-06, + "loss": 0.4378, + "step": 6532 + }, + { + "epoch": 2.0903370307167237, + "grad_norm": 0.5020625591278076, + "learning_rate": 2.547984956781683e-06, + "loss": 0.4495, + "step": 6533 + }, + { + "epoch": 2.0906569965870307, + "grad_norm": 0.548353910446167, + "learning_rate": 2.5463625804451714e-06, + "loss": 0.4467, + "step": 6534 + }, + { + "epoch": 2.0909769624573378, + "grad_norm": 0.49052295088768005, + "learning_rate": 2.5447405443084395e-06, + "loss": 0.4383, + "step": 6535 + }, + { + "epoch": 2.0912969283276452, + "grad_norm": 0.5260667204856873, + "learning_rate": 2.543118848596384e-06, + "loss": 0.4672, + "step": 6536 + }, + { + "epoch": 2.0916168941979523, + "grad_norm": 0.41019099950790405, + "learning_rate": 2.5414974935338555e-06, + "loss": 0.3995, + "step": 6537 + }, + { + "epoch": 2.0919368600682593, + "grad_norm": 0.5305437445640564, + "learning_rate": 2.5398764793456554e-06, + "loss": 0.4685, + "step": 6538 + }, + { + "epoch": 2.0922568259385668, + "grad_norm": 0.5135195851325989, + "learning_rate": 2.538255806256544e-06, + "loss": 0.3883, + "step": 6539 + }, + { + "epoch": 2.092576791808874, + "grad_norm": 0.5410434603691101, + "learning_rate": 2.5366354744912267e-06, + "loss": 0.445, + "step": 6540 + }, + { + "epoch": 2.092896757679181, + "grad_norm": 0.5121901631355286, + "learning_rate": 2.5350154842743643e-06, + "loss": 0.4751, + "step": 6541 + }, + { + "epoch": 2.093216723549488, + "grad_norm": 0.44693824648857117, + "learning_rate": 2.5333958358305683e-06, + "loss": 0.3867, + "step": 6542 + }, + { + "epoch": 2.0935366894197953, + "grad_norm": 0.49896085262298584, + "learning_rate": 2.531776529384407e-06, + "loss": 0.4473, + "step": 6543 + }, + { + "epoch": 2.0938566552901023, + "grad_norm": 0.5412548184394836, + "learning_rate": 2.5301575651604014e-06, + "loss": 0.4437, + "step": 6544 + }, + { + "epoch": 2.0941766211604094, + "grad_norm": 0.5537082552909851, + "learning_rate": 2.5285389433830206e-06, + "loss": 0.4355, + "step": 6545 + }, + { + "epoch": 2.094496587030717, + "grad_norm": 0.4909462630748749, + "learning_rate": 2.5269206642766887e-06, + "loss": 0.4478, + "step": 6546 + }, + { + "epoch": 2.094816552901024, + "grad_norm": 0.514731764793396, + "learning_rate": 2.5253027280657792e-06, + "loss": 0.4082, + "step": 6547 + }, + { + "epoch": 2.095136518771331, + "grad_norm": 0.5522803664207458, + "learning_rate": 2.5236851349746242e-06, + "loss": 0.5042, + "step": 6548 + }, + { + "epoch": 2.0954564846416384, + "grad_norm": 0.49755656719207764, + "learning_rate": 2.522067885227506e-06, + "loss": 0.4522, + "step": 6549 + }, + { + "epoch": 2.0957764505119454, + "grad_norm": 0.4478907883167267, + "learning_rate": 2.520450979048657e-06, + "loss": 0.3895, + "step": 6550 + }, + { + "epoch": 2.0960964163822524, + "grad_norm": 0.5032786726951599, + "learning_rate": 2.5188344166622623e-06, + "loss": 0.4628, + "step": 6551 + }, + { + "epoch": 2.09641638225256, + "grad_norm": 0.5694442391395569, + "learning_rate": 2.5172181982924593e-06, + "loss": 0.4592, + "step": 6552 + }, + { + "epoch": 2.096736348122867, + "grad_norm": 0.5020767450332642, + "learning_rate": 2.5156023241633394e-06, + "loss": 0.4322, + "step": 6553 + }, + { + "epoch": 2.097056313993174, + "grad_norm": 0.5417909622192383, + "learning_rate": 2.5139867944989483e-06, + "loss": 0.4904, + "step": 6554 + }, + { + "epoch": 2.0973762798634814, + "grad_norm": 0.49538323283195496, + "learning_rate": 2.5123716095232784e-06, + "loss": 0.4396, + "step": 6555 + }, + { + "epoch": 2.0976962457337884, + "grad_norm": 0.5189866423606873, + "learning_rate": 2.5107567694602776e-06, + "loss": 0.4718, + "step": 6556 + }, + { + "epoch": 2.0980162116040955, + "grad_norm": 0.5295016765594482, + "learning_rate": 2.509142274533844e-06, + "loss": 0.3919, + "step": 6557 + }, + { + "epoch": 2.098336177474403, + "grad_norm": 0.5462619662284851, + "learning_rate": 2.50752812496783e-06, + "loss": 0.5043, + "step": 6558 + }, + { + "epoch": 2.09865614334471, + "grad_norm": 0.5347604751586914, + "learning_rate": 2.5059143209860428e-06, + "loss": 0.4307, + "step": 6559 + }, + { + "epoch": 2.098976109215017, + "grad_norm": 0.49428123235702515, + "learning_rate": 2.5043008628122346e-06, + "loss": 0.432, + "step": 6560 + }, + { + "epoch": 2.0992960750853245, + "grad_norm": 0.5311827659606934, + "learning_rate": 2.502687750670114e-06, + "loss": 0.4273, + "step": 6561 + }, + { + "epoch": 2.0996160409556315, + "grad_norm": 0.6081832647323608, + "learning_rate": 2.50107498478334e-06, + "loss": 0.4983, + "step": 6562 + }, + { + "epoch": 2.0999360068259385, + "grad_norm": 0.4512195885181427, + "learning_rate": 2.499462565375525e-06, + "loss": 0.4293, + "step": 6563 + }, + { + "epoch": 2.1002559726962455, + "grad_norm": 0.493958443403244, + "learning_rate": 2.497850492670236e-06, + "loss": 0.4315, + "step": 6564 + }, + { + "epoch": 2.100575938566553, + "grad_norm": 0.48001089692115784, + "learning_rate": 2.496238766890986e-06, + "loss": 0.42, + "step": 6565 + }, + { + "epoch": 2.10089590443686, + "grad_norm": 0.5479942560195923, + "learning_rate": 2.4946273882612432e-06, + "loss": 0.4792, + "step": 6566 + }, + { + "epoch": 2.101215870307167, + "grad_norm": 0.5242119431495667, + "learning_rate": 2.4930163570044245e-06, + "loss": 0.4022, + "step": 6567 + }, + { + "epoch": 2.1015358361774745, + "grad_norm": 0.58072429895401, + "learning_rate": 2.4914056733439044e-06, + "loss": 0.521, + "step": 6568 + }, + { + "epoch": 2.1018558020477816, + "grad_norm": 0.49397119879722595, + "learning_rate": 2.4897953375030078e-06, + "loss": 0.3568, + "step": 6569 + }, + { + "epoch": 2.1021757679180886, + "grad_norm": 0.5828869342803955, + "learning_rate": 2.4881853497050074e-06, + "loss": 0.4521, + "step": 6570 + }, + { + "epoch": 2.102495733788396, + "grad_norm": 0.5099934935569763, + "learning_rate": 2.4865757101731295e-06, + "loss": 0.4379, + "step": 6571 + }, + { + "epoch": 2.102815699658703, + "grad_norm": 0.5592410564422607, + "learning_rate": 2.4849664191305517e-06, + "loss": 0.4023, + "step": 6572 + }, + { + "epoch": 2.10313566552901, + "grad_norm": 0.5595393776893616, + "learning_rate": 2.483357476800405e-06, + "loss": 0.4601, + "step": 6573 + }, + { + "epoch": 2.1034556313993176, + "grad_norm": 0.5343798398971558, + "learning_rate": 2.4817488834057744e-06, + "loss": 0.4454, + "step": 6574 + }, + { + "epoch": 2.1037755972696246, + "grad_norm": 0.4935324490070343, + "learning_rate": 2.4801406391696907e-06, + "loss": 0.3637, + "step": 6575 + }, + { + "epoch": 2.1040955631399316, + "grad_norm": 0.49617499113082886, + "learning_rate": 2.4785327443151385e-06, + "loss": 0.4685, + "step": 6576 + }, + { + "epoch": 2.104415529010239, + "grad_norm": 0.4559192359447479, + "learning_rate": 2.476925199065053e-06, + "loss": 0.3727, + "step": 6577 + }, + { + "epoch": 2.104735494880546, + "grad_norm": 0.5002991557121277, + "learning_rate": 2.4753180036423237e-06, + "loss": 0.4793, + "step": 6578 + }, + { + "epoch": 2.105055460750853, + "grad_norm": 0.4963497817516327, + "learning_rate": 2.473711158269792e-06, + "loss": 0.4534, + "step": 6579 + }, + { + "epoch": 2.1053754266211606, + "grad_norm": 0.4909219443798065, + "learning_rate": 2.4721046631702478e-06, + "loss": 0.4649, + "step": 6580 + }, + { + "epoch": 2.1056953924914676, + "grad_norm": 0.5330654978752136, + "learning_rate": 2.470498518566433e-06, + "loss": 0.4395, + "step": 6581 + }, + { + "epoch": 2.1060153583617747, + "grad_norm": 0.5223993062973022, + "learning_rate": 2.46889272468104e-06, + "loss": 0.4382, + "step": 6582 + }, + { + "epoch": 2.1063353242320817, + "grad_norm": 0.49935439229011536, + "learning_rate": 2.4672872817367162e-06, + "loss": 0.4072, + "step": 6583 + }, + { + "epoch": 2.106655290102389, + "grad_norm": 0.5430231094360352, + "learning_rate": 2.465682189956059e-06, + "loss": 0.4407, + "step": 6584 + }, + { + "epoch": 2.106975255972696, + "grad_norm": 0.4854907691478729, + "learning_rate": 2.4640774495616147e-06, + "loss": 0.3854, + "step": 6585 + }, + { + "epoch": 2.107295221843003, + "grad_norm": 0.4910756051540375, + "learning_rate": 2.4624730607758833e-06, + "loss": 0.467, + "step": 6586 + }, + { + "epoch": 2.1076151877133107, + "grad_norm": 0.5427428483963013, + "learning_rate": 2.4608690238213128e-06, + "loss": 0.4444, + "step": 6587 + }, + { + "epoch": 2.1079351535836177, + "grad_norm": 0.49487268924713135, + "learning_rate": 2.459265338920307e-06, + "loss": 0.3803, + "step": 6588 + }, + { + "epoch": 2.1082551194539247, + "grad_norm": 0.512978196144104, + "learning_rate": 2.45766200629522e-06, + "loss": 0.4333, + "step": 6589 + }, + { + "epoch": 2.108575085324232, + "grad_norm": 0.5519300699234009, + "learning_rate": 2.456059026168355e-06, + "loss": 0.4219, + "step": 6590 + }, + { + "epoch": 2.1088950511945392, + "grad_norm": 0.5403918623924255, + "learning_rate": 2.454456398761966e-06, + "loss": 0.4367, + "step": 6591 + }, + { + "epoch": 2.1092150170648463, + "grad_norm": 0.49616748094558716, + "learning_rate": 2.452854124298257e-06, + "loss": 0.4434, + "step": 6592 + }, + { + "epoch": 2.1095349829351537, + "grad_norm": 0.512844443321228, + "learning_rate": 2.451252202999389e-06, + "loss": 0.4193, + "step": 6593 + }, + { + "epoch": 2.1098549488054608, + "grad_norm": 0.5188925862312317, + "learning_rate": 2.449650635087471e-06, + "loss": 0.4732, + "step": 6594 + }, + { + "epoch": 2.110174914675768, + "grad_norm": 0.507415235042572, + "learning_rate": 2.4480494207845607e-06, + "loss": 0.4661, + "step": 6595 + }, + { + "epoch": 2.1104948805460753, + "grad_norm": 0.4596986770629883, + "learning_rate": 2.4464485603126665e-06, + "loss": 0.3773, + "step": 6596 + }, + { + "epoch": 2.1108148464163823, + "grad_norm": 0.5360031127929688, + "learning_rate": 2.4448480538937523e-06, + "loss": 0.4952, + "step": 6597 + }, + { + "epoch": 2.1111348122866893, + "grad_norm": 0.5112121105194092, + "learning_rate": 2.4432479017497285e-06, + "loss": 0.4226, + "step": 6598 + }, + { + "epoch": 2.111454778156997, + "grad_norm": 0.4723941385746002, + "learning_rate": 2.44164810410246e-06, + "loss": 0.4349, + "step": 6599 + }, + { + "epoch": 2.111774744027304, + "grad_norm": 0.5009363293647766, + "learning_rate": 2.4400486611737605e-06, + "loss": 0.486, + "step": 6600 + }, + { + "epoch": 2.112094709897611, + "grad_norm": 0.5131105184555054, + "learning_rate": 2.4384495731853918e-06, + "loss": 0.4278, + "step": 6601 + }, + { + "epoch": 2.1124146757679183, + "grad_norm": 0.5704697966575623, + "learning_rate": 2.436850840359073e-06, + "loss": 0.502, + "step": 6602 + }, + { + "epoch": 2.1127346416382253, + "grad_norm": 0.48064613342285156, + "learning_rate": 2.435252462916467e-06, + "loss": 0.3816, + "step": 6603 + }, + { + "epoch": 2.1130546075085324, + "grad_norm": 0.5179963707923889, + "learning_rate": 2.433654441079194e-06, + "loss": 0.4441, + "step": 6604 + }, + { + "epoch": 2.1133745733788394, + "grad_norm": 0.5865185856819153, + "learning_rate": 2.4320567750688206e-06, + "loss": 0.4794, + "step": 6605 + }, + { + "epoch": 2.113694539249147, + "grad_norm": 0.4456147253513336, + "learning_rate": 2.4304594651068626e-06, + "loss": 0.3302, + "step": 6606 + }, + { + "epoch": 2.114014505119454, + "grad_norm": 0.5374706387519836, + "learning_rate": 2.4288625114147914e-06, + "loss": 0.4636, + "step": 6607 + }, + { + "epoch": 2.114334470989761, + "grad_norm": 0.49862635135650635, + "learning_rate": 2.427265914214027e-06, + "loss": 0.4367, + "step": 6608 + }, + { + "epoch": 2.1146544368600684, + "grad_norm": 0.43899568915367126, + "learning_rate": 2.4256696737259393e-06, + "loss": 0.4391, + "step": 6609 + }, + { + "epoch": 2.1149744027303754, + "grad_norm": 0.5214098691940308, + "learning_rate": 2.424073790171849e-06, + "loss": 0.4664, + "step": 6610 + }, + { + "epoch": 2.1152943686006824, + "grad_norm": 0.4941417872905731, + "learning_rate": 2.422478263773023e-06, + "loss": 0.4319, + "step": 6611 + }, + { + "epoch": 2.11561433447099, + "grad_norm": 0.4795205593109131, + "learning_rate": 2.420883094750688e-06, + "loss": 0.431, + "step": 6612 + }, + { + "epoch": 2.115934300341297, + "grad_norm": 0.5306460857391357, + "learning_rate": 2.419288283326016e-06, + "loss": 0.4316, + "step": 6613 + }, + { + "epoch": 2.116254266211604, + "grad_norm": 0.48654627799987793, + "learning_rate": 2.4176938297201285e-06, + "loss": 0.3964, + "step": 6614 + }, + { + "epoch": 2.1165742320819114, + "grad_norm": 0.529757022857666, + "learning_rate": 2.4160997341540983e-06, + "loss": 0.4257, + "step": 6615 + }, + { + "epoch": 2.1168941979522184, + "grad_norm": 0.5562413930892944, + "learning_rate": 2.414505996848946e-06, + "loss": 0.4464, + "step": 6616 + }, + { + "epoch": 2.1172141638225255, + "grad_norm": 0.4944826066493988, + "learning_rate": 2.4129126180256478e-06, + "loss": 0.4361, + "step": 6617 + }, + { + "epoch": 2.117534129692833, + "grad_norm": 0.5483657121658325, + "learning_rate": 2.4113195979051296e-06, + "loss": 0.4597, + "step": 6618 + }, + { + "epoch": 2.11785409556314, + "grad_norm": 0.47472381591796875, + "learning_rate": 2.409726936708263e-06, + "loss": 0.431, + "step": 6619 + }, + { + "epoch": 2.118174061433447, + "grad_norm": 0.5364698171615601, + "learning_rate": 2.408134634655873e-06, + "loss": 0.4373, + "step": 6620 + }, + { + "epoch": 2.1184940273037545, + "grad_norm": 0.531273365020752, + "learning_rate": 2.4065426919687322e-06, + "loss": 0.4414, + "step": 6621 + }, + { + "epoch": 2.1188139931740615, + "grad_norm": 0.48729974031448364, + "learning_rate": 2.404951108867567e-06, + "loss": 0.4352, + "step": 6622 + }, + { + "epoch": 2.1191339590443685, + "grad_norm": 0.5032392740249634, + "learning_rate": 2.4033598855730544e-06, + "loss": 0.421, + "step": 6623 + }, + { + "epoch": 2.1194539249146755, + "grad_norm": 0.5721447467803955, + "learning_rate": 2.4017690223058175e-06, + "loss": 0.4736, + "step": 6624 + }, + { + "epoch": 2.119773890784983, + "grad_norm": 0.4929790794849396, + "learning_rate": 2.4001785192864314e-06, + "loss": 0.4165, + "step": 6625 + }, + { + "epoch": 2.12009385665529, + "grad_norm": 0.48628777265548706, + "learning_rate": 2.3985883767354194e-06, + "loss": 0.4266, + "step": 6626 + }, + { + "epoch": 2.120413822525597, + "grad_norm": 0.48934853076934814, + "learning_rate": 2.3969985948732586e-06, + "loss": 0.443, + "step": 6627 + }, + { + "epoch": 2.1207337883959045, + "grad_norm": 0.4963590204715729, + "learning_rate": 2.395409173920376e-06, + "loss": 0.4678, + "step": 6628 + }, + { + "epoch": 2.1210537542662116, + "grad_norm": 0.5006743669509888, + "learning_rate": 2.393820114097145e-06, + "loss": 0.4004, + "step": 6629 + }, + { + "epoch": 2.1213737201365186, + "grad_norm": 0.5066134929656982, + "learning_rate": 2.3922314156238883e-06, + "loss": 0.4578, + "step": 6630 + }, + { + "epoch": 2.121693686006826, + "grad_norm": 0.5675395131111145, + "learning_rate": 2.390643078720885e-06, + "loss": 0.4805, + "step": 6631 + }, + { + "epoch": 2.122013651877133, + "grad_norm": 0.5577268004417419, + "learning_rate": 2.3890551036083564e-06, + "loss": 0.5025, + "step": 6632 + }, + { + "epoch": 2.12233361774744, + "grad_norm": 0.5691620707511902, + "learning_rate": 2.3874674905064804e-06, + "loss": 0.4362, + "step": 6633 + }, + { + "epoch": 2.1226535836177476, + "grad_norm": 0.497223436832428, + "learning_rate": 2.3858802396353796e-06, + "loss": 0.4201, + "step": 6634 + }, + { + "epoch": 2.1229735494880546, + "grad_norm": 0.5128043293952942, + "learning_rate": 2.3842933512151268e-06, + "loss": 0.4268, + "step": 6635 + }, + { + "epoch": 2.1232935153583616, + "grad_norm": 0.5296140909194946, + "learning_rate": 2.3827068254657493e-06, + "loss": 0.4759, + "step": 6636 + }, + { + "epoch": 2.123613481228669, + "grad_norm": 0.5045025944709778, + "learning_rate": 2.381120662607217e-06, + "loss": 0.4376, + "step": 6637 + }, + { + "epoch": 2.123933447098976, + "grad_norm": 0.48082995414733887, + "learning_rate": 2.379534862859458e-06, + "loss": 0.4393, + "step": 6638 + }, + { + "epoch": 2.124253412969283, + "grad_norm": 0.5115391612052917, + "learning_rate": 2.3779494264423427e-06, + "loss": 0.4435, + "step": 6639 + }, + { + "epoch": 2.1245733788395906, + "grad_norm": 0.5318616628646851, + "learning_rate": 2.3763643535756914e-06, + "loss": 0.4589, + "step": 6640 + }, + { + "epoch": 2.1248933447098977, + "grad_norm": 0.48749759793281555, + "learning_rate": 2.374779644479281e-06, + "loss": 0.4599, + "step": 6641 + }, + { + "epoch": 2.1252133105802047, + "grad_norm": 0.5193473696708679, + "learning_rate": 2.373195299372829e-06, + "loss": 0.4457, + "step": 6642 + }, + { + "epoch": 2.125533276450512, + "grad_norm": 0.4633994400501251, + "learning_rate": 2.371611318476011e-06, + "loss": 0.4168, + "step": 6643 + }, + { + "epoch": 2.125853242320819, + "grad_norm": 0.5086625814437866, + "learning_rate": 2.3700277020084454e-06, + "loss": 0.4255, + "step": 6644 + }, + { + "epoch": 2.126173208191126, + "grad_norm": 0.4725306034088135, + "learning_rate": 2.3684444501897012e-06, + "loss": 0.463, + "step": 6645 + }, + { + "epoch": 2.1264931740614337, + "grad_norm": 0.49431467056274414, + "learning_rate": 2.366861563239302e-06, + "loss": 0.454, + "step": 6646 + }, + { + "epoch": 2.1268131399317407, + "grad_norm": 0.5001236796379089, + "learning_rate": 2.3652790413767125e-06, + "loss": 0.4689, + "step": 6647 + }, + { + "epoch": 2.1271331058020477, + "grad_norm": 0.49575915932655334, + "learning_rate": 2.363696884821355e-06, + "loss": 0.4317, + "step": 6648 + }, + { + "epoch": 2.1274530716723548, + "grad_norm": 0.5135654807090759, + "learning_rate": 2.362115093792597e-06, + "loss": 0.4244, + "step": 6649 + }, + { + "epoch": 2.1277730375426622, + "grad_norm": 0.5357350707054138, + "learning_rate": 2.3605336685097523e-06, + "loss": 0.4559, + "step": 6650 + }, + { + "epoch": 2.1280930034129693, + "grad_norm": 0.4895274341106415, + "learning_rate": 2.3589526091920915e-06, + "loss": 0.3771, + "step": 6651 + }, + { + "epoch": 2.1284129692832763, + "grad_norm": 0.5519123673439026, + "learning_rate": 2.357371916058827e-06, + "loss": 0.4447, + "step": 6652 + }, + { + "epoch": 2.1287329351535837, + "grad_norm": 0.5529068112373352, + "learning_rate": 2.3557915893291276e-06, + "loss": 0.4291, + "step": 6653 + }, + { + "epoch": 2.1290529010238908, + "grad_norm": 0.5326127409934998, + "learning_rate": 2.354211629222105e-06, + "loss": 0.4858, + "step": 6654 + }, + { + "epoch": 2.129372866894198, + "grad_norm": 0.5279597640037537, + "learning_rate": 2.3526320359568215e-06, + "loss": 0.4218, + "step": 6655 + }, + { + "epoch": 2.1296928327645053, + "grad_norm": 0.4971260726451874, + "learning_rate": 2.3510528097522928e-06, + "loss": 0.41, + "step": 6656 + }, + { + "epoch": 2.1300127986348123, + "grad_norm": 0.4707668423652649, + "learning_rate": 2.3494739508274765e-06, + "loss": 0.4274, + "step": 6657 + }, + { + "epoch": 2.1303327645051193, + "grad_norm": 0.5016271471977234, + "learning_rate": 2.3478954594012884e-06, + "loss": 0.4666, + "step": 6658 + }, + { + "epoch": 2.130652730375427, + "grad_norm": 0.5279699563980103, + "learning_rate": 2.3463173356925843e-06, + "loss": 0.4901, + "step": 6659 + }, + { + "epoch": 2.130972696245734, + "grad_norm": 0.4883582890033722, + "learning_rate": 2.3447395799201733e-06, + "loss": 0.4051, + "step": 6660 + }, + { + "epoch": 2.131292662116041, + "grad_norm": 0.561225950717926, + "learning_rate": 2.3431621923028146e-06, + "loss": 0.4648, + "step": 6661 + }, + { + "epoch": 2.1316126279863483, + "grad_norm": 0.5723588466644287, + "learning_rate": 2.341585173059213e-06, + "loss": 0.4945, + "step": 6662 + }, + { + "epoch": 2.1319325938566553, + "grad_norm": 0.5918015241622925, + "learning_rate": 2.340008522408027e-06, + "loss": 0.4603, + "step": 6663 + }, + { + "epoch": 2.1322525597269624, + "grad_norm": 0.5477120876312256, + "learning_rate": 2.3384322405678576e-06, + "loss": 0.4417, + "step": 6664 + }, + { + "epoch": 2.1325725255972694, + "grad_norm": 0.5312170386314392, + "learning_rate": 2.3368563277572624e-06, + "loss": 0.4679, + "step": 6665 + }, + { + "epoch": 2.132892491467577, + "grad_norm": 0.46622106432914734, + "learning_rate": 2.3352807841947406e-06, + "loss": 0.4656, + "step": 6666 + }, + { + "epoch": 2.133212457337884, + "grad_norm": 0.4596792459487915, + "learning_rate": 2.333705610098742e-06, + "loss": 0.3897, + "step": 6667 + }, + { + "epoch": 2.133532423208191, + "grad_norm": 0.5101672410964966, + "learning_rate": 2.33213080568767e-06, + "loss": 0.4683, + "step": 6668 + }, + { + "epoch": 2.1338523890784984, + "grad_norm": 0.5518092513084412, + "learning_rate": 2.3305563711798696e-06, + "loss": 0.5045, + "step": 6669 + }, + { + "epoch": 2.1341723549488054, + "grad_norm": 0.6030656695365906, + "learning_rate": 2.328982306793642e-06, + "loss": 0.49, + "step": 6670 + }, + { + "epoch": 2.1344923208191124, + "grad_norm": 0.5064455270767212, + "learning_rate": 2.32740861274723e-06, + "loss": 0.4244, + "step": 6671 + }, + { + "epoch": 2.13481228668942, + "grad_norm": 0.5312418937683105, + "learning_rate": 2.3258352892588277e-06, + "loss": 0.4291, + "step": 6672 + }, + { + "epoch": 2.135132252559727, + "grad_norm": 0.5422471165657043, + "learning_rate": 2.3242623365465816e-06, + "loss": 0.4856, + "step": 6673 + }, + { + "epoch": 2.135452218430034, + "grad_norm": 0.45353472232818604, + "learning_rate": 2.322689754828579e-06, + "loss": 0.4177, + "step": 6674 + }, + { + "epoch": 2.1357721843003414, + "grad_norm": 0.5094054341316223, + "learning_rate": 2.3211175443228655e-06, + "loss": 0.4675, + "step": 6675 + }, + { + "epoch": 2.1360921501706485, + "grad_norm": 0.4616287648677826, + "learning_rate": 2.319545705247427e-06, + "loss": 0.4171, + "step": 6676 + }, + { + "epoch": 2.1364121160409555, + "grad_norm": 0.5320475101470947, + "learning_rate": 2.317974237820199e-06, + "loss": 0.4731, + "step": 6677 + }, + { + "epoch": 2.136732081911263, + "grad_norm": 0.5035117268562317, + "learning_rate": 2.3164031422590715e-06, + "loss": 0.3585, + "step": 6678 + }, + { + "epoch": 2.13705204778157, + "grad_norm": 0.5568263530731201, + "learning_rate": 2.314832418781875e-06, + "loss": 0.4652, + "step": 6679 + }, + { + "epoch": 2.137372013651877, + "grad_norm": 0.5261285901069641, + "learning_rate": 2.313262067606396e-06, + "loss": 0.461, + "step": 6680 + }, + { + "epoch": 2.1376919795221845, + "grad_norm": 0.5170149803161621, + "learning_rate": 2.311692088950363e-06, + "loss": 0.4082, + "step": 6681 + }, + { + "epoch": 2.1380119453924915, + "grad_norm": 0.5495178699493408, + "learning_rate": 2.3101224830314546e-06, + "loss": 0.5008, + "step": 6682 + }, + { + "epoch": 2.1383319112627985, + "grad_norm": 0.5265787243843079, + "learning_rate": 2.308553250067302e-06, + "loss": 0.4261, + "step": 6683 + }, + { + "epoch": 2.138651877133106, + "grad_norm": 0.5479193925857544, + "learning_rate": 2.3069843902754767e-06, + "loss": 0.4252, + "step": 6684 + }, + { + "epoch": 2.138971843003413, + "grad_norm": 0.4920346140861511, + "learning_rate": 2.3054159038735076e-06, + "loss": 0.4807, + "step": 6685 + }, + { + "epoch": 2.13929180887372, + "grad_norm": 0.5257879495620728, + "learning_rate": 2.303847791078865e-06, + "loss": 0.4687, + "step": 6686 + }, + { + "epoch": 2.1396117747440275, + "grad_norm": 0.46145397424697876, + "learning_rate": 2.302280052108968e-06, + "loss": 0.4159, + "step": 6687 + }, + { + "epoch": 2.1399317406143346, + "grad_norm": 0.49745118618011475, + "learning_rate": 2.3007126871811886e-06, + "loss": 0.4237, + "step": 6688 + }, + { + "epoch": 2.1402517064846416, + "grad_norm": 0.5214394927024841, + "learning_rate": 2.299145696512841e-06, + "loss": 0.458, + "step": 6689 + }, + { + "epoch": 2.1405716723549486, + "grad_norm": 0.46344926953315735, + "learning_rate": 2.297579080321194e-06, + "loss": 0.4218, + "step": 6690 + }, + { + "epoch": 2.140891638225256, + "grad_norm": 0.4784850478172302, + "learning_rate": 2.2960128388234582e-06, + "loss": 0.4421, + "step": 6691 + }, + { + "epoch": 2.141211604095563, + "grad_norm": 0.505057692527771, + "learning_rate": 2.294446972236793e-06, + "loss": 0.4595, + "step": 6692 + }, + { + "epoch": 2.14153156996587, + "grad_norm": 0.5089201331138611, + "learning_rate": 2.292881480778312e-06, + "loss": 0.4492, + "step": 6693 + }, + { + "epoch": 2.1418515358361776, + "grad_norm": 0.5136494636535645, + "learning_rate": 2.291316364665069e-06, + "loss": 0.4154, + "step": 6694 + }, + { + "epoch": 2.1421715017064846, + "grad_norm": 0.5064804553985596, + "learning_rate": 2.289751624114073e-06, + "loss": 0.4165, + "step": 6695 + }, + { + "epoch": 2.1424914675767917, + "grad_norm": 0.4527985751628876, + "learning_rate": 2.2881872593422737e-06, + "loss": 0.3568, + "step": 6696 + }, + { + "epoch": 2.142811433447099, + "grad_norm": 0.5171433687210083, + "learning_rate": 2.286623270566572e-06, + "loss": 0.4885, + "step": 6697 + }, + { + "epoch": 2.143131399317406, + "grad_norm": 0.4854653775691986, + "learning_rate": 2.2850596580038196e-06, + "loss": 0.4323, + "step": 6698 + }, + { + "epoch": 2.143451365187713, + "grad_norm": 0.46086299419403076, + "learning_rate": 2.28349642187081e-06, + "loss": 0.4193, + "step": 6699 + }, + { + "epoch": 2.1437713310580206, + "grad_norm": 0.4680379629135132, + "learning_rate": 2.281933562384292e-06, + "loss": 0.4369, + "step": 6700 + }, + { + "epoch": 2.1440912969283277, + "grad_norm": 0.5378003120422363, + "learning_rate": 2.2803710797609545e-06, + "loss": 0.5075, + "step": 6701 + }, + { + "epoch": 2.1444112627986347, + "grad_norm": 0.505118191242218, + "learning_rate": 2.2788089742174374e-06, + "loss": 0.4213, + "step": 6702 + }, + { + "epoch": 2.144731228668942, + "grad_norm": 0.5250274538993835, + "learning_rate": 2.2772472459703298e-06, + "loss": 0.442, + "step": 6703 + }, + { + "epoch": 2.145051194539249, + "grad_norm": 0.5066376328468323, + "learning_rate": 2.2756858952361687e-06, + "loss": 0.4603, + "step": 6704 + }, + { + "epoch": 2.145371160409556, + "grad_norm": 0.49959373474121094, + "learning_rate": 2.274124922231436e-06, + "loss": 0.4215, + "step": 6705 + }, + { + "epoch": 2.1456911262798632, + "grad_norm": 0.48590216040611267, + "learning_rate": 2.2725643271725617e-06, + "loss": 0.4433, + "step": 6706 + }, + { + "epoch": 2.1460110921501707, + "grad_norm": 0.4953617453575134, + "learning_rate": 2.271004110275923e-06, + "loss": 0.4682, + "step": 6707 + }, + { + "epoch": 2.1463310580204777, + "grad_norm": 0.5015279650688171, + "learning_rate": 2.2694442717578476e-06, + "loss": 0.4483, + "step": 6708 + }, + { + "epoch": 2.1466510238907848, + "grad_norm": 0.4883991479873657, + "learning_rate": 2.2678848118346103e-06, + "loss": 0.4773, + "step": 6709 + }, + { + "epoch": 2.1469709897610922, + "grad_norm": 0.46480992436408997, + "learning_rate": 2.2663257307224308e-06, + "loss": 0.4195, + "step": 6710 + }, + { + "epoch": 2.1472909556313993, + "grad_norm": 0.47466591000556946, + "learning_rate": 2.264767028637477e-06, + "loss": 0.4532, + "step": 6711 + }, + { + "epoch": 2.1476109215017063, + "grad_norm": 0.5161939263343811, + "learning_rate": 2.263208705795864e-06, + "loss": 0.4606, + "step": 6712 + }, + { + "epoch": 2.1479308873720138, + "grad_norm": 0.48355168104171753, + "learning_rate": 2.2616507624136564e-06, + "loss": 0.4097, + "step": 6713 + }, + { + "epoch": 2.148250853242321, + "grad_norm": 0.5150226950645447, + "learning_rate": 2.2600931987068662e-06, + "loss": 0.4998, + "step": 6714 + }, + { + "epoch": 2.148570819112628, + "grad_norm": 0.4854436218738556, + "learning_rate": 2.258536014891451e-06, + "loss": 0.4182, + "step": 6715 + }, + { + "epoch": 2.1488907849829353, + "grad_norm": 0.5275769233703613, + "learning_rate": 2.256979211183315e-06, + "loss": 0.4387, + "step": 6716 + }, + { + "epoch": 2.1492107508532423, + "grad_norm": 0.5012706518173218, + "learning_rate": 2.2554227877983093e-06, + "loss": 0.4187, + "step": 6717 + }, + { + "epoch": 2.1495307167235493, + "grad_norm": 0.5256766676902771, + "learning_rate": 2.253866744952236e-06, + "loss": 0.4305, + "step": 6718 + }, + { + "epoch": 2.149850682593857, + "grad_norm": 0.5074648261070251, + "learning_rate": 2.252311082860844e-06, + "loss": 0.4624, + "step": 6719 + }, + { + "epoch": 2.150170648464164, + "grad_norm": 0.5025088787078857, + "learning_rate": 2.2507558017398263e-06, + "loss": 0.4458, + "step": 6720 + }, + { + "epoch": 2.150490614334471, + "grad_norm": 0.47776737809181213, + "learning_rate": 2.2492009018048234e-06, + "loss": 0.4219, + "step": 6721 + }, + { + "epoch": 2.1508105802047783, + "grad_norm": 0.48077675700187683, + "learning_rate": 2.2476463832714233e-06, + "loss": 0.4464, + "step": 6722 + }, + { + "epoch": 2.1511305460750854, + "grad_norm": 0.5319883227348328, + "learning_rate": 2.246092246355163e-06, + "loss": 0.5005, + "step": 6723 + }, + { + "epoch": 2.1514505119453924, + "grad_norm": 0.5291938185691833, + "learning_rate": 2.2445384912715285e-06, + "loss": 0.4437, + "step": 6724 + }, + { + "epoch": 2.1517704778157, + "grad_norm": 0.5148395895957947, + "learning_rate": 2.2429851182359465e-06, + "loss": 0.4866, + "step": 6725 + }, + { + "epoch": 2.152090443686007, + "grad_norm": 0.4877208471298218, + "learning_rate": 2.2414321274637946e-06, + "loss": 0.401, + "step": 6726 + }, + { + "epoch": 2.152410409556314, + "grad_norm": 0.5249781608581543, + "learning_rate": 2.239879519170395e-06, + "loss": 0.4535, + "step": 6727 + }, + { + "epoch": 2.1527303754266214, + "grad_norm": 0.5234302878379822, + "learning_rate": 2.2383272935710205e-06, + "loss": 0.4467, + "step": 6728 + }, + { + "epoch": 2.1530503412969284, + "grad_norm": 0.4273567199707031, + "learning_rate": 2.2367754508808914e-06, + "loss": 0.3639, + "step": 6729 + }, + { + "epoch": 2.1533703071672354, + "grad_norm": 0.4548000395298004, + "learning_rate": 2.2352239913151696e-06, + "loss": 0.4276, + "step": 6730 + }, + { + "epoch": 2.1536902730375425, + "grad_norm": 0.4965096116065979, + "learning_rate": 2.233672915088968e-06, + "loss": 0.4642, + "step": 6731 + }, + { + "epoch": 2.15401023890785, + "grad_norm": 0.5180421471595764, + "learning_rate": 2.2321222224173426e-06, + "loss": 0.4583, + "step": 6732 + }, + { + "epoch": 2.154330204778157, + "grad_norm": 0.5392321944236755, + "learning_rate": 2.2305719135153013e-06, + "loss": 0.4502, + "step": 6733 + }, + { + "epoch": 2.154650170648464, + "grad_norm": 0.45776090025901794, + "learning_rate": 2.2290219885977976e-06, + "loss": 0.4079, + "step": 6734 + }, + { + "epoch": 2.1549701365187715, + "grad_norm": 0.5151669383049011, + "learning_rate": 2.2274724478797284e-06, + "loss": 0.4057, + "step": 6735 + }, + { + "epoch": 2.1552901023890785, + "grad_norm": 0.5341548323631287, + "learning_rate": 2.22592329157594e-06, + "loss": 0.4673, + "step": 6736 + }, + { + "epoch": 2.1556100682593855, + "grad_norm": 0.5643128156661987, + "learning_rate": 2.224374519901222e-06, + "loss": 0.4162, + "step": 6737 + }, + { + "epoch": 2.155930034129693, + "grad_norm": 0.5588071942329407, + "learning_rate": 2.2228261330703172e-06, + "loss": 0.4296, + "step": 6738 + }, + { + "epoch": 2.15625, + "grad_norm": 0.5966269969940186, + "learning_rate": 2.2212781312979114e-06, + "loss": 0.5066, + "step": 6739 + }, + { + "epoch": 2.156569965870307, + "grad_norm": 0.5233896374702454, + "learning_rate": 2.2197305147986352e-06, + "loss": 0.3812, + "step": 6740 + }, + { + "epoch": 2.1568899317406145, + "grad_norm": 0.48629361391067505, + "learning_rate": 2.2181832837870683e-06, + "loss": 0.4396, + "step": 6741 + }, + { + "epoch": 2.1572098976109215, + "grad_norm": 0.5437921285629272, + "learning_rate": 2.2166364384777337e-06, + "loss": 0.4639, + "step": 6742 + }, + { + "epoch": 2.1575298634812285, + "grad_norm": 0.5298428535461426, + "learning_rate": 2.2150899790851056e-06, + "loss": 0.4484, + "step": 6743 + }, + { + "epoch": 2.157849829351536, + "grad_norm": 0.5280128121376038, + "learning_rate": 2.2135439058236036e-06, + "loss": 0.4625, + "step": 6744 + }, + { + "epoch": 2.158169795221843, + "grad_norm": 0.5505377054214478, + "learning_rate": 2.211998218907592e-06, + "loss": 0.444, + "step": 6745 + }, + { + "epoch": 2.15848976109215, + "grad_norm": 0.4741867184638977, + "learning_rate": 2.2104529185513807e-06, + "loss": 0.4072, + "step": 6746 + }, + { + "epoch": 2.1588097269624575, + "grad_norm": 0.4593965411186218, + "learning_rate": 2.2089080049692263e-06, + "loss": 0.4331, + "step": 6747 + }, + { + "epoch": 2.1591296928327646, + "grad_norm": 0.4532676339149475, + "learning_rate": 2.2073634783753352e-06, + "loss": 0.4536, + "step": 6748 + }, + { + "epoch": 2.1594496587030716, + "grad_norm": 0.5205339193344116, + "learning_rate": 2.205819338983859e-06, + "loss": 0.4531, + "step": 6749 + }, + { + "epoch": 2.1597696245733786, + "grad_norm": 0.501815140247345, + "learning_rate": 2.2042755870088933e-06, + "loss": 0.3847, + "step": 6750 + }, + { + "epoch": 2.160089590443686, + "grad_norm": 0.48559510707855225, + "learning_rate": 2.2027322226644803e-06, + "loss": 0.4477, + "step": 6751 + }, + { + "epoch": 2.160409556313993, + "grad_norm": 0.46358832716941833, + "learning_rate": 2.201189246164606e-06, + "loss": 0.4196, + "step": 6752 + }, + { + "epoch": 2.1607295221843, + "grad_norm": 0.5158883333206177, + "learning_rate": 2.199646657723214e-06, + "loss": 0.4434, + "step": 6753 + }, + { + "epoch": 2.1610494880546076, + "grad_norm": 0.46989548206329346, + "learning_rate": 2.198104457554182e-06, + "loss": 0.4095, + "step": 6754 + }, + { + "epoch": 2.1613694539249146, + "grad_norm": 0.5143417119979858, + "learning_rate": 2.1965626458713375e-06, + "loss": 0.4407, + "step": 6755 + }, + { + "epoch": 2.1616894197952217, + "grad_norm": 0.5397100448608398, + "learning_rate": 2.1950212228884544e-06, + "loss": 0.4951, + "step": 6756 + }, + { + "epoch": 2.162009385665529, + "grad_norm": 0.4593190550804138, + "learning_rate": 2.19348018881925e-06, + "loss": 0.3479, + "step": 6757 + }, + { + "epoch": 2.162329351535836, + "grad_norm": 0.5357738733291626, + "learning_rate": 2.191939543877397e-06, + "loss": 0.4345, + "step": 6758 + }, + { + "epoch": 2.162649317406143, + "grad_norm": 0.47553011775016785, + "learning_rate": 2.1903992882765048e-06, + "loss": 0.4692, + "step": 6759 + }, + { + "epoch": 2.1629692832764507, + "grad_norm": 0.49455299973487854, + "learning_rate": 2.1888594222301307e-06, + "loss": 0.3952, + "step": 6760 + }, + { + "epoch": 2.1632892491467577, + "grad_norm": 0.48963063955307007, + "learning_rate": 2.18731994595178e-06, + "loss": 0.4103, + "step": 6761 + }, + { + "epoch": 2.1636092150170647, + "grad_norm": 0.5070008635520935, + "learning_rate": 2.1857808596548992e-06, + "loss": 0.5065, + "step": 6762 + }, + { + "epoch": 2.163929180887372, + "grad_norm": 0.4946386516094208, + "learning_rate": 2.184242163552891e-06, + "loss": 0.4266, + "step": 6763 + }, + { + "epoch": 2.164249146757679, + "grad_norm": 0.5226266980171204, + "learning_rate": 2.182703857859094e-06, + "loss": 0.4362, + "step": 6764 + }, + { + "epoch": 2.1645691126279862, + "grad_norm": 0.5144748091697693, + "learning_rate": 2.181165942786797e-06, + "loss": 0.4748, + "step": 6765 + }, + { + "epoch": 2.1648890784982937, + "grad_norm": 0.45408496260643005, + "learning_rate": 2.1796284185492333e-06, + "loss": 0.388, + "step": 6766 + }, + { + "epoch": 2.1652090443686007, + "grad_norm": 0.5466834306716919, + "learning_rate": 2.178091285359578e-06, + "loss": 0.4747, + "step": 6767 + }, + { + "epoch": 2.1655290102389078, + "grad_norm": 0.4702305495738983, + "learning_rate": 2.176554543430965e-06, + "loss": 0.4363, + "step": 6768 + }, + { + "epoch": 2.1658489761092152, + "grad_norm": 0.5069165825843811, + "learning_rate": 2.175018192976461e-06, + "loss": 0.4225, + "step": 6769 + }, + { + "epoch": 2.1661689419795223, + "grad_norm": 0.5441505312919617, + "learning_rate": 2.1734822342090835e-06, + "loss": 0.4543, + "step": 6770 + }, + { + "epoch": 2.1664889078498293, + "grad_norm": 0.479708731174469, + "learning_rate": 2.171946667341792e-06, + "loss": 0.4277, + "step": 6771 + }, + { + "epoch": 2.1668088737201368, + "grad_norm": 0.47547414898872375, + "learning_rate": 2.170411492587498e-06, + "loss": 0.437, + "step": 6772 + }, + { + "epoch": 2.167128839590444, + "grad_norm": 0.5153126120567322, + "learning_rate": 2.1688767101590564e-06, + "loss": 0.4622, + "step": 6773 + }, + { + "epoch": 2.167448805460751, + "grad_norm": 0.5160109996795654, + "learning_rate": 2.1673423202692643e-06, + "loss": 0.4555, + "step": 6774 + }, + { + "epoch": 2.167768771331058, + "grad_norm": 0.48195773363113403, + "learning_rate": 2.165808323130868e-06, + "loss": 0.472, + "step": 6775 + }, + { + "epoch": 2.1680887372013653, + "grad_norm": 0.45776796340942383, + "learning_rate": 2.164274718956555e-06, + "loss": 0.439, + "step": 6776 + }, + { + "epoch": 2.1684087030716723, + "grad_norm": 0.4817793071269989, + "learning_rate": 2.1627415079589637e-06, + "loss": 0.427, + "step": 6777 + }, + { + "epoch": 2.1687286689419794, + "grad_norm": 0.4975053071975708, + "learning_rate": 2.161208690350678e-06, + "loss": 0.4498, + "step": 6778 + }, + { + "epoch": 2.169048634812287, + "grad_norm": 0.460365891456604, + "learning_rate": 2.159676266344222e-06, + "loss": 0.378, + "step": 6779 + }, + { + "epoch": 2.169368600682594, + "grad_norm": 0.512991189956665, + "learning_rate": 2.1581442361520688e-06, + "loss": 0.4484, + "step": 6780 + }, + { + "epoch": 2.169688566552901, + "grad_norm": 0.5276342034339905, + "learning_rate": 2.156612599986634e-06, + "loss": 0.4575, + "step": 6781 + }, + { + "epoch": 2.1700085324232083, + "grad_norm": 0.505875289440155, + "learning_rate": 2.155081358060283e-06, + "loss": 0.4636, + "step": 6782 + }, + { + "epoch": 2.1703284982935154, + "grad_norm": 0.5176311135292053, + "learning_rate": 2.153550510585326e-06, + "loss": 0.466, + "step": 6783 + }, + { + "epoch": 2.1706484641638224, + "grad_norm": 0.590843677520752, + "learning_rate": 2.1520200577740154e-06, + "loss": 0.4663, + "step": 6784 + }, + { + "epoch": 2.17096843003413, + "grad_norm": 0.4701339900493622, + "learning_rate": 2.1504899998385497e-06, + "loss": 0.3819, + "step": 6785 + }, + { + "epoch": 2.171288395904437, + "grad_norm": 0.47744128108024597, + "learning_rate": 2.1489603369910724e-06, + "loss": 0.4367, + "step": 6786 + }, + { + "epoch": 2.171608361774744, + "grad_norm": 0.5769877433776855, + "learning_rate": 2.1474310694436732e-06, + "loss": 0.5047, + "step": 6787 + }, + { + "epoch": 2.1719283276450514, + "grad_norm": 0.5574608445167542, + "learning_rate": 2.1459021974083905e-06, + "loss": 0.4447, + "step": 6788 + }, + { + "epoch": 2.1722482935153584, + "grad_norm": 0.5587093830108643, + "learning_rate": 2.1443737210972017e-06, + "loss": 0.4411, + "step": 6789 + }, + { + "epoch": 2.1725682593856654, + "grad_norm": 0.5133920907974243, + "learning_rate": 2.1428456407220317e-06, + "loss": 0.4533, + "step": 6790 + }, + { + "epoch": 2.1728882252559725, + "grad_norm": 0.47900161147117615, + "learning_rate": 2.141317956494749e-06, + "loss": 0.4772, + "step": 6791 + }, + { + "epoch": 2.17320819112628, + "grad_norm": 0.5402735471725464, + "learning_rate": 2.1397906686271706e-06, + "loss": 0.4177, + "step": 6792 + }, + { + "epoch": 2.173528156996587, + "grad_norm": 0.5899126529693604, + "learning_rate": 2.1382637773310587e-06, + "loss": 0.4572, + "step": 6793 + }, + { + "epoch": 2.173848122866894, + "grad_norm": 0.5434718728065491, + "learning_rate": 2.136737282818117e-06, + "loss": 0.4933, + "step": 6794 + }, + { + "epoch": 2.1741680887372015, + "grad_norm": 0.4832441210746765, + "learning_rate": 2.135211185299996e-06, + "loss": 0.4025, + "step": 6795 + }, + { + "epoch": 2.1744880546075085, + "grad_norm": 0.5162698030471802, + "learning_rate": 2.1336854849882883e-06, + "loss": 0.3842, + "step": 6796 + }, + { + "epoch": 2.1748080204778155, + "grad_norm": 0.476591020822525, + "learning_rate": 2.132160182094536e-06, + "loss": 0.438, + "step": 6797 + }, + { + "epoch": 2.175127986348123, + "grad_norm": 0.486438512802124, + "learning_rate": 2.130635276830227e-06, + "loss": 0.4443, + "step": 6798 + }, + { + "epoch": 2.17544795221843, + "grad_norm": 0.5168624520301819, + "learning_rate": 2.1291107694067885e-06, + "loss": 0.3907, + "step": 6799 + }, + { + "epoch": 2.175767918088737, + "grad_norm": 0.5580846071243286, + "learning_rate": 2.1275866600355954e-06, + "loss": 0.4305, + "step": 6800 + }, + { + "epoch": 2.1760878839590445, + "grad_norm": 0.5284036993980408, + "learning_rate": 2.1260629489279662e-06, + "loss": 0.443, + "step": 6801 + }, + { + "epoch": 2.1764078498293515, + "grad_norm": 0.4435729682445526, + "learning_rate": 2.1245396362951664e-06, + "loss": 0.4378, + "step": 6802 + }, + { + "epoch": 2.1767278156996586, + "grad_norm": 0.49425411224365234, + "learning_rate": 2.1230167223484072e-06, + "loss": 0.4531, + "step": 6803 + }, + { + "epoch": 2.177047781569966, + "grad_norm": 0.5321007370948792, + "learning_rate": 2.12149420729884e-06, + "loss": 0.4399, + "step": 6804 + }, + { + "epoch": 2.177367747440273, + "grad_norm": 0.578448474407196, + "learning_rate": 2.119972091357563e-06, + "loss": 0.405, + "step": 6805 + }, + { + "epoch": 2.17768771331058, + "grad_norm": 0.5024625062942505, + "learning_rate": 2.118450374735621e-06, + "loss": 0.469, + "step": 6806 + }, + { + "epoch": 2.1780076791808876, + "grad_norm": 0.5228381156921387, + "learning_rate": 2.116929057644e-06, + "loss": 0.463, + "step": 6807 + }, + { + "epoch": 2.1783276450511946, + "grad_norm": 0.4660845696926117, + "learning_rate": 2.115408140293635e-06, + "loss": 0.3963, + "step": 6808 + }, + { + "epoch": 2.1786476109215016, + "grad_norm": 0.4882470369338989, + "learning_rate": 2.113887622895402e-06, + "loss": 0.4072, + "step": 6809 + }, + { + "epoch": 2.178967576791809, + "grad_norm": 0.5116421580314636, + "learning_rate": 2.11236750566012e-06, + "loss": 0.4615, + "step": 6810 + }, + { + "epoch": 2.179287542662116, + "grad_norm": 0.5525357723236084, + "learning_rate": 2.1108477887985597e-06, + "loss": 0.4693, + "step": 6811 + }, + { + "epoch": 2.179607508532423, + "grad_norm": 0.491583913564682, + "learning_rate": 2.109328472521427e-06, + "loss": 0.3816, + "step": 6812 + }, + { + "epoch": 2.1799274744027306, + "grad_norm": 0.5098915100097656, + "learning_rate": 2.107809557039381e-06, + "loss": 0.4588, + "step": 6813 + }, + { + "epoch": 2.1802474402730376, + "grad_norm": 0.49412649869918823, + "learning_rate": 2.106291042563019e-06, + "loss": 0.4442, + "step": 6814 + }, + { + "epoch": 2.1805674061433447, + "grad_norm": 0.49511006474494934, + "learning_rate": 2.1047729293028834e-06, + "loss": 0.4347, + "step": 6815 + }, + { + "epoch": 2.1808873720136517, + "grad_norm": 0.5126012563705444, + "learning_rate": 2.103255217469466e-06, + "loss": 0.4268, + "step": 6816 + }, + { + "epoch": 2.181207337883959, + "grad_norm": 0.46396976709365845, + "learning_rate": 2.101737907273196e-06, + "loss": 0.4187, + "step": 6817 + }, + { + "epoch": 2.181527303754266, + "grad_norm": 0.5442050695419312, + "learning_rate": 2.100220998924454e-06, + "loss": 0.4358, + "step": 6818 + }, + { + "epoch": 2.181847269624573, + "grad_norm": 0.5309011936187744, + "learning_rate": 2.0987044926335583e-06, + "loss": 0.4384, + "step": 6819 + }, + { + "epoch": 2.1821672354948807, + "grad_norm": 0.5154093503952026, + "learning_rate": 2.0971883886107737e-06, + "loss": 0.3808, + "step": 6820 + }, + { + "epoch": 2.1824872013651877, + "grad_norm": 0.45877769589424133, + "learning_rate": 2.0956726870663123e-06, + "loss": 0.3898, + "step": 6821 + }, + { + "epoch": 2.1828071672354947, + "grad_norm": 0.48768162727355957, + "learning_rate": 2.094157388210326e-06, + "loss": 0.4781, + "step": 6822 + }, + { + "epoch": 2.183127133105802, + "grad_norm": 0.4783969521522522, + "learning_rate": 2.092642492252915e-06, + "loss": 0.3945, + "step": 6823 + }, + { + "epoch": 2.1834470989761092, + "grad_norm": 0.5190688967704773, + "learning_rate": 2.09112799940412e-06, + "loss": 0.4313, + "step": 6824 + }, + { + "epoch": 2.1837670648464163, + "grad_norm": 0.5493019819259644, + "learning_rate": 2.089613909873926e-06, + "loss": 0.463, + "step": 6825 + }, + { + "epoch": 2.1840870307167237, + "grad_norm": 0.4760606586933136, + "learning_rate": 2.0881002238722657e-06, + "loss": 0.4331, + "step": 6826 + }, + { + "epoch": 2.1844069965870307, + "grad_norm": 0.49167048931121826, + "learning_rate": 2.086586941609011e-06, + "loss": 0.4203, + "step": 6827 + }, + { + "epoch": 2.1847269624573378, + "grad_norm": 0.5753871202468872, + "learning_rate": 2.0850740632939837e-06, + "loss": 0.438, + "step": 6828 + }, + { + "epoch": 2.1850469283276452, + "grad_norm": 0.5460748076438904, + "learning_rate": 2.083561589136944e-06, + "loss": 0.5011, + "step": 6829 + }, + { + "epoch": 2.1853668941979523, + "grad_norm": 0.47245848178863525, + "learning_rate": 2.0820495193475972e-06, + "loss": 0.3838, + "step": 6830 + }, + { + "epoch": 2.1856868600682593, + "grad_norm": 0.5486859679222107, + "learning_rate": 2.0805378541355965e-06, + "loss": 0.4317, + "step": 6831 + }, + { + "epoch": 2.1860068259385663, + "grad_norm": 0.5436904430389404, + "learning_rate": 2.0790265937105325e-06, + "loss": 0.4212, + "step": 6832 + }, + { + "epoch": 2.186326791808874, + "grad_norm": 0.5365811586380005, + "learning_rate": 2.0775157382819473e-06, + "loss": 0.4391, + "step": 6833 + }, + { + "epoch": 2.186646757679181, + "grad_norm": 0.514493465423584, + "learning_rate": 2.0760052880593213e-06, + "loss": 0.4833, + "step": 6834 + }, + { + "epoch": 2.186966723549488, + "grad_norm": 0.523497462272644, + "learning_rate": 2.074495243252077e-06, + "loss": 0.4956, + "step": 6835 + }, + { + "epoch": 2.1872866894197953, + "grad_norm": 0.45324259996414185, + "learning_rate": 2.072985604069589e-06, + "loss": 0.3956, + "step": 6836 + }, + { + "epoch": 2.1876066552901023, + "grad_norm": 0.48483917117118835, + "learning_rate": 2.0714763707211664e-06, + "loss": 0.4065, + "step": 6837 + }, + { + "epoch": 2.1879266211604094, + "grad_norm": 0.49716633558273315, + "learning_rate": 2.06996754341607e-06, + "loss": 0.4386, + "step": 6838 + }, + { + "epoch": 2.188246587030717, + "grad_norm": 0.510878324508667, + "learning_rate": 2.0684591223634986e-06, + "loss": 0.4697, + "step": 6839 + }, + { + "epoch": 2.188566552901024, + "grad_norm": 0.4646270275115967, + "learning_rate": 2.0669511077725945e-06, + "loss": 0.3907, + "step": 6840 + }, + { + "epoch": 2.188886518771331, + "grad_norm": 0.4669325053691864, + "learning_rate": 2.0654434998524502e-06, + "loss": 0.4183, + "step": 6841 + }, + { + "epoch": 2.1892064846416384, + "grad_norm": 0.4611457884311676, + "learning_rate": 2.0639362988120933e-06, + "loss": 0.4072, + "step": 6842 + }, + { + "epoch": 2.1895264505119454, + "grad_norm": 0.5439834594726562, + "learning_rate": 2.062429504860502e-06, + "loss": 0.4816, + "step": 6843 + }, + { + "epoch": 2.1898464163822524, + "grad_norm": 0.4837367534637451, + "learning_rate": 2.060923118206592e-06, + "loss": 0.443, + "step": 6844 + }, + { + "epoch": 2.19016638225256, + "grad_norm": 0.4659494161605835, + "learning_rate": 2.0594171390592294e-06, + "loss": 0.3955, + "step": 6845 + }, + { + "epoch": 2.190486348122867, + "grad_norm": 0.5452629327774048, + "learning_rate": 2.057911567627218e-06, + "loss": 0.482, + "step": 6846 + }, + { + "epoch": 2.190806313993174, + "grad_norm": 0.4978730082511902, + "learning_rate": 2.0564064041193053e-06, + "loss": 0.3921, + "step": 6847 + }, + { + "epoch": 2.1911262798634814, + "grad_norm": 0.5474013686180115, + "learning_rate": 2.054901648744188e-06, + "loss": 0.5015, + "step": 6848 + }, + { + "epoch": 2.1914462457337884, + "grad_norm": 0.5499725937843323, + "learning_rate": 2.0533973017104984e-06, + "loss": 0.505, + "step": 6849 + }, + { + "epoch": 2.1917662116040955, + "grad_norm": 0.503440797328949, + "learning_rate": 2.0518933632268197e-06, + "loss": 0.3985, + "step": 6850 + }, + { + "epoch": 2.192086177474403, + "grad_norm": 0.5070424675941467, + "learning_rate": 2.0503898335016724e-06, + "loss": 0.4199, + "step": 6851 + }, + { + "epoch": 2.19240614334471, + "grad_norm": 0.5492558479309082, + "learning_rate": 2.0488867127435213e-06, + "loss": 0.4508, + "step": 6852 + }, + { + "epoch": 2.192726109215017, + "grad_norm": 0.523219108581543, + "learning_rate": 2.04738400116078e-06, + "loss": 0.4231, + "step": 6853 + }, + { + "epoch": 2.1930460750853245, + "grad_norm": 0.511542558670044, + "learning_rate": 2.045881698961797e-06, + "loss": 0.4042, + "step": 6854 + }, + { + "epoch": 2.1933660409556315, + "grad_norm": 0.5142143964767456, + "learning_rate": 2.044379806354872e-06, + "loss": 0.4438, + "step": 6855 + }, + { + "epoch": 2.1936860068259385, + "grad_norm": 0.5721973180770874, + "learning_rate": 2.0428783235482423e-06, + "loss": 0.4605, + "step": 6856 + }, + { + "epoch": 2.1940059726962455, + "grad_norm": 0.5211235284805298, + "learning_rate": 2.0413772507500895e-06, + "loss": 0.4822, + "step": 6857 + }, + { + "epoch": 2.194325938566553, + "grad_norm": 0.48469439148902893, + "learning_rate": 2.0398765881685417e-06, + "loss": 0.4353, + "step": 6858 + }, + { + "epoch": 2.19464590443686, + "grad_norm": 0.4858268201351166, + "learning_rate": 2.0383763360116636e-06, + "loss": 0.4404, + "step": 6859 + }, + { + "epoch": 2.194965870307167, + "grad_norm": 0.46958059072494507, + "learning_rate": 2.036876494487472e-06, + "loss": 0.461, + "step": 6860 + }, + { + "epoch": 2.1952858361774745, + "grad_norm": 0.47352731227874756, + "learning_rate": 2.035377063803919e-06, + "loss": 0.4045, + "step": 6861 + }, + { + "epoch": 2.1956058020477816, + "grad_norm": 0.5191588997840881, + "learning_rate": 2.033878044168901e-06, + "loss": 0.4781, + "step": 6862 + }, + { + "epoch": 2.1959257679180886, + "grad_norm": 0.5520008206367493, + "learning_rate": 2.032379435790262e-06, + "loss": 0.4526, + "step": 6863 + }, + { + "epoch": 2.196245733788396, + "grad_norm": 0.4723661541938782, + "learning_rate": 2.0308812388757828e-06, + "loss": 0.4322, + "step": 6864 + }, + { + "epoch": 2.196565699658703, + "grad_norm": 0.5281932353973389, + "learning_rate": 2.029383453633193e-06, + "loss": 0.4013, + "step": 6865 + }, + { + "epoch": 2.19688566552901, + "grad_norm": 0.6034097075462341, + "learning_rate": 2.0278860802701616e-06, + "loss": 0.4635, + "step": 6866 + }, + { + "epoch": 2.1972056313993176, + "grad_norm": 0.5373727679252625, + "learning_rate": 2.0263891189942993e-06, + "loss": 0.4505, + "step": 6867 + }, + { + "epoch": 2.1975255972696246, + "grad_norm": 0.494740754365921, + "learning_rate": 2.024892570013164e-06, + "loss": 0.4075, + "step": 6868 + }, + { + "epoch": 2.1978455631399316, + "grad_norm": 0.4898551404476166, + "learning_rate": 2.0233964335342517e-06, + "loss": 0.4347, + "step": 6869 + }, + { + "epoch": 2.198165529010239, + "grad_norm": 0.5141493678092957, + "learning_rate": 2.0219007097650063e-06, + "loss": 0.4566, + "step": 6870 + }, + { + "epoch": 2.198485494880546, + "grad_norm": 0.515340268611908, + "learning_rate": 2.02040539891281e-06, + "loss": 0.4114, + "step": 6871 + }, + { + "epoch": 2.198805460750853, + "grad_norm": 0.5628042817115784, + "learning_rate": 2.0189105011849887e-06, + "loss": 0.503, + "step": 6872 + }, + { + "epoch": 2.1991254266211606, + "grad_norm": 0.5190582275390625, + "learning_rate": 2.017416016788814e-06, + "loss": 0.4681, + "step": 6873 + }, + { + "epoch": 2.1994453924914676, + "grad_norm": 0.5403550863265991, + "learning_rate": 2.0159219459314948e-06, + "loss": 0.4603, + "step": 6874 + }, + { + "epoch": 2.1997653583617747, + "grad_norm": 0.5147091150283813, + "learning_rate": 2.01442828882019e-06, + "loss": 0.4396, + "step": 6875 + }, + { + "epoch": 2.2000853242320817, + "grad_norm": 0.475997656583786, + "learning_rate": 2.0129350456619946e-06, + "loss": 0.4314, + "step": 6876 + }, + { + "epoch": 2.200405290102389, + "grad_norm": 0.4674367308616638, + "learning_rate": 2.011442216663947e-06, + "loss": 0.4133, + "step": 6877 + }, + { + "epoch": 2.200725255972696, + "grad_norm": 0.4793749153614044, + "learning_rate": 2.0099498020330305e-06, + "loss": 0.4767, + "step": 6878 + }, + { + "epoch": 2.201045221843003, + "grad_norm": 0.5369402170181274, + "learning_rate": 2.0084578019761738e-06, + "loss": 0.4702, + "step": 6879 + }, + { + "epoch": 2.2013651877133107, + "grad_norm": 0.48916861414909363, + "learning_rate": 2.006966216700241e-06, + "loss": 0.3998, + "step": 6880 + }, + { + "epoch": 2.2016851535836177, + "grad_norm": 0.4863379895687103, + "learning_rate": 2.005475046412044e-06, + "loss": 0.4966, + "step": 6881 + }, + { + "epoch": 2.2020051194539247, + "grad_norm": 0.44687557220458984, + "learning_rate": 2.0039842913183316e-06, + "loss": 0.4415, + "step": 6882 + }, + { + "epoch": 2.202325085324232, + "grad_norm": 0.4204229414463043, + "learning_rate": 2.002493951625802e-06, + "loss": 0.4162, + "step": 6883 + }, + { + "epoch": 2.2026450511945392, + "grad_norm": 0.5047687292098999, + "learning_rate": 2.0010040275410934e-06, + "loss": 0.4318, + "step": 6884 + }, + { + "epoch": 2.2029650170648463, + "grad_norm": 0.5496372580528259, + "learning_rate": 1.9995145192707845e-06, + "loss": 0.4658, + "step": 6885 + }, + { + "epoch": 2.2032849829351537, + "grad_norm": 0.5273929238319397, + "learning_rate": 1.998025427021397e-06, + "loss": 0.4643, + "step": 6886 + }, + { + "epoch": 2.2036049488054608, + "grad_norm": 0.5278632640838623, + "learning_rate": 1.9965367509993942e-06, + "loss": 0.3911, + "step": 6887 + }, + { + "epoch": 2.203924914675768, + "grad_norm": 0.47825103998184204, + "learning_rate": 1.995048491411184e-06, + "loss": 0.4559, + "step": 6888 + }, + { + "epoch": 2.2042448805460753, + "grad_norm": 0.4605184495449066, + "learning_rate": 1.993560648463117e-06, + "loss": 0.4291, + "step": 6889 + }, + { + "epoch": 2.2045648464163823, + "grad_norm": 0.5016769170761108, + "learning_rate": 1.992073222361483e-06, + "loss": 0.4723, + "step": 6890 + }, + { + "epoch": 2.2048848122866893, + "grad_norm": 0.4509531855583191, + "learning_rate": 1.990586213312516e-06, + "loss": 0.4172, + "step": 6891 + }, + { + "epoch": 2.205204778156997, + "grad_norm": 0.4440130591392517, + "learning_rate": 1.9890996215223885e-06, + "loss": 0.3638, + "step": 6892 + }, + { + "epoch": 2.205524744027304, + "grad_norm": 0.5383840203285217, + "learning_rate": 1.98761344719722e-06, + "loss": 0.4994, + "step": 6893 + }, + { + "epoch": 2.205844709897611, + "grad_norm": 0.4807364046573639, + "learning_rate": 1.986127690543074e-06, + "loss": 0.4016, + "step": 6894 + }, + { + "epoch": 2.2061646757679183, + "grad_norm": 0.5220727920532227, + "learning_rate": 1.984642351765949e-06, + "loss": 0.4297, + "step": 6895 + }, + { + "epoch": 2.2064846416382253, + "grad_norm": 0.507565975189209, + "learning_rate": 1.98315743107179e-06, + "loss": 0.4154, + "step": 6896 + }, + { + "epoch": 2.2068046075085324, + "grad_norm": 0.48764318227767944, + "learning_rate": 1.98167292866648e-06, + "loss": 0.4298, + "step": 6897 + }, + { + "epoch": 2.2071245733788394, + "grad_norm": 0.495840460062027, + "learning_rate": 1.9801888447558497e-06, + "loss": 0.4432, + "step": 6898 + }, + { + "epoch": 2.207444539249147, + "grad_norm": 0.5050677061080933, + "learning_rate": 1.9787051795456714e-06, + "loss": 0.4162, + "step": 6899 + }, + { + "epoch": 2.207764505119454, + "grad_norm": 0.48621681332588196, + "learning_rate": 1.9772219332416544e-06, + "loss": 0.4683, + "step": 6900 + }, + { + "epoch": 2.208084470989761, + "grad_norm": 0.45279693603515625, + "learning_rate": 1.9757391060494525e-06, + "loss": 0.3982, + "step": 6901 + }, + { + "epoch": 2.2084044368600684, + "grad_norm": 0.5346647500991821, + "learning_rate": 1.9742566981746596e-06, + "loss": 0.4741, + "step": 6902 + }, + { + "epoch": 2.2087244027303754, + "grad_norm": 0.5127875208854675, + "learning_rate": 1.9727747098228167e-06, + "loss": 0.4462, + "step": 6903 + }, + { + "epoch": 2.2090443686006824, + "grad_norm": 0.49378058314323425, + "learning_rate": 1.971293141199403e-06, + "loss": 0.437, + "step": 6904 + }, + { + "epoch": 2.20936433447099, + "grad_norm": 0.49449095129966736, + "learning_rate": 1.9698119925098398e-06, + "loss": 0.4527, + "step": 6905 + }, + { + "epoch": 2.209684300341297, + "grad_norm": 0.49412205815315247, + "learning_rate": 1.9683312639594883e-06, + "loss": 0.4799, + "step": 6906 + }, + { + "epoch": 2.210004266211604, + "grad_norm": 0.4692303240299225, + "learning_rate": 1.966850955753653e-06, + "loss": 0.3993, + "step": 6907 + }, + { + "epoch": 2.2103242320819114, + "grad_norm": 0.48455101251602173, + "learning_rate": 1.965371068097581e-06, + "loss": 0.4132, + "step": 6908 + }, + { + "epoch": 2.2106441979522184, + "grad_norm": 0.531497061252594, + "learning_rate": 1.9638916011964637e-06, + "loss": 0.4759, + "step": 6909 + }, + { + "epoch": 2.2109641638225255, + "grad_norm": 0.5194461345672607, + "learning_rate": 1.9624125552554287e-06, + "loss": 0.4083, + "step": 6910 + }, + { + "epoch": 2.211284129692833, + "grad_norm": 0.49568623304367065, + "learning_rate": 1.960933930479545e-06, + "loss": 0.4299, + "step": 6911 + }, + { + "epoch": 2.21160409556314, + "grad_norm": 0.4672048091888428, + "learning_rate": 1.9594557270738275e-06, + "loss": 0.4346, + "step": 6912 + }, + { + "epoch": 2.211924061433447, + "grad_norm": 0.509306013584137, + "learning_rate": 1.9579779452432336e-06, + "loss": 0.474, + "step": 6913 + }, + { + "epoch": 2.2122440273037545, + "grad_norm": 0.5298650860786438, + "learning_rate": 1.9565005851926574e-06, + "loss": 0.4254, + "step": 6914 + }, + { + "epoch": 2.2125639931740615, + "grad_norm": 0.5057072639465332, + "learning_rate": 1.9550236471269358e-06, + "loss": 0.4584, + "step": 6915 + }, + { + "epoch": 2.2128839590443685, + "grad_norm": 0.5267782211303711, + "learning_rate": 1.953547131250847e-06, + "loss": 0.4478, + "step": 6916 + }, + { + "epoch": 2.2132039249146755, + "grad_norm": 0.5834232568740845, + "learning_rate": 1.9520710377691137e-06, + "loss": 0.4841, + "step": 6917 + }, + { + "epoch": 2.213523890784983, + "grad_norm": 0.5276156663894653, + "learning_rate": 1.9505953668863996e-06, + "loss": 0.457, + "step": 6918 + }, + { + "epoch": 2.21384385665529, + "grad_norm": 0.477408766746521, + "learning_rate": 1.9491201188073067e-06, + "loss": 0.4318, + "step": 6919 + }, + { + "epoch": 2.214163822525597, + "grad_norm": 0.5019200444221497, + "learning_rate": 1.947645293736379e-06, + "loss": 0.4364, + "step": 6920 + }, + { + "epoch": 2.2144837883959045, + "grad_norm": 0.49255263805389404, + "learning_rate": 1.9461708918781018e-06, + "loss": 0.4518, + "step": 6921 + }, + { + "epoch": 2.2148037542662116, + "grad_norm": 0.5077428817749023, + "learning_rate": 1.944696913436905e-06, + "loss": 0.4301, + "step": 6922 + }, + { + "epoch": 2.2151237201365186, + "grad_norm": 0.5375750660896301, + "learning_rate": 1.9432233586171576e-06, + "loss": 0.4121, + "step": 6923 + }, + { + "epoch": 2.215443686006826, + "grad_norm": 0.5089945197105408, + "learning_rate": 1.941750227623169e-06, + "loss": 0.4434, + "step": 6924 + }, + { + "epoch": 2.215763651877133, + "grad_norm": 0.43878012895584106, + "learning_rate": 1.940277520659191e-06, + "loss": 0.4257, + "step": 6925 + }, + { + "epoch": 2.21608361774744, + "grad_norm": 0.4928267300128937, + "learning_rate": 1.9388052379294135e-06, + "loss": 0.4155, + "step": 6926 + }, + { + "epoch": 2.2164035836177476, + "grad_norm": 0.5031419992446899, + "learning_rate": 1.937333379637973e-06, + "loss": 0.459, + "step": 6927 + }, + { + "epoch": 2.2167235494880546, + "grad_norm": 0.4916204810142517, + "learning_rate": 1.935861945988945e-06, + "loss": 0.3791, + "step": 6928 + }, + { + "epoch": 2.2170435153583616, + "grad_norm": 0.5682798027992249, + "learning_rate": 1.9343909371863452e-06, + "loss": 0.427, + "step": 6929 + }, + { + "epoch": 2.217363481228669, + "grad_norm": 0.5600176453590393, + "learning_rate": 1.932920353434129e-06, + "loss": 0.417, + "step": 6930 + }, + { + "epoch": 2.217683447098976, + "grad_norm": 0.5578386187553406, + "learning_rate": 1.9314501949361946e-06, + "loss": 0.4224, + "step": 6931 + }, + { + "epoch": 2.218003412969283, + "grad_norm": 0.5250586867332458, + "learning_rate": 1.929980461896382e-06, + "loss": 0.4548, + "step": 6932 + }, + { + "epoch": 2.2183233788395906, + "grad_norm": 0.5524583458900452, + "learning_rate": 1.928511154518473e-06, + "loss": 0.439, + "step": 6933 + }, + { + "epoch": 2.2186433447098977, + "grad_norm": 0.49936771392822266, + "learning_rate": 1.9270422730061883e-06, + "loss": 0.4048, + "step": 6934 + }, + { + "epoch": 2.2189633105802047, + "grad_norm": 0.5375549793243408, + "learning_rate": 1.925573817563189e-06, + "loss": 0.4703, + "step": 6935 + }, + { + "epoch": 2.219283276450512, + "grad_norm": 0.560430645942688, + "learning_rate": 1.924105788393077e-06, + "loss": 0.4663, + "step": 6936 + }, + { + "epoch": 2.219603242320819, + "grad_norm": 0.5091943740844727, + "learning_rate": 1.922638185699397e-06, + "loss": 0.4395, + "step": 6937 + }, + { + "epoch": 2.219923208191126, + "grad_norm": 0.5026044845581055, + "learning_rate": 1.9211710096856373e-06, + "loss": 0.4632, + "step": 6938 + }, + { + "epoch": 2.2202431740614337, + "grad_norm": 0.5348517894744873, + "learning_rate": 1.9197042605552206e-06, + "loss": 0.4696, + "step": 6939 + }, + { + "epoch": 2.2205631399317407, + "grad_norm": 0.47700411081314087, + "learning_rate": 1.918237938511514e-06, + "loss": 0.4222, + "step": 6940 + }, + { + "epoch": 2.2208831058020477, + "grad_norm": 0.5197616219520569, + "learning_rate": 1.916772043757823e-06, + "loss": 0.4943, + "step": 6941 + }, + { + "epoch": 2.2212030716723548, + "grad_norm": 0.48118993639945984, + "learning_rate": 1.915306576497397e-06, + "loss": 0.3919, + "step": 6942 + }, + { + "epoch": 2.2215230375426622, + "grad_norm": 0.5554553866386414, + "learning_rate": 1.9138415369334275e-06, + "loss": 0.4865, + "step": 6943 + }, + { + "epoch": 2.2218430034129693, + "grad_norm": 0.4820300340652466, + "learning_rate": 1.912376925269041e-06, + "loss": 0.4447, + "step": 6944 + }, + { + "epoch": 2.2221629692832763, + "grad_norm": 0.5619463324546814, + "learning_rate": 1.9109127417073087e-06, + "loss": 0.486, + "step": 6945 + }, + { + "epoch": 2.2224829351535837, + "grad_norm": 0.5620408654212952, + "learning_rate": 1.9094489864512393e-06, + "loss": 0.4141, + "step": 6946 + }, + { + "epoch": 2.2228029010238908, + "grad_norm": 0.5426426529884338, + "learning_rate": 1.907985659703786e-06, + "loss": 0.448, + "step": 6947 + }, + { + "epoch": 2.223122866894198, + "grad_norm": 0.48584941029548645, + "learning_rate": 1.9065227616678422e-06, + "loss": 0.4807, + "step": 6948 + }, + { + "epoch": 2.2234428327645053, + "grad_norm": 0.4817311465740204, + "learning_rate": 1.9050602925462392e-06, + "loss": 0.4679, + "step": 6949 + }, + { + "epoch": 2.2237627986348123, + "grad_norm": 0.4578678607940674, + "learning_rate": 1.90359825254175e-06, + "loss": 0.4147, + "step": 6950 + }, + { + "epoch": 2.2240827645051193, + "grad_norm": 0.48312684893608093, + "learning_rate": 1.9021366418570864e-06, + "loss": 0.402, + "step": 6951 + }, + { + "epoch": 2.224402730375427, + "grad_norm": 0.524979293346405, + "learning_rate": 1.9006754606949041e-06, + "loss": 0.5166, + "step": 6952 + }, + { + "epoch": 2.224722696245734, + "grad_norm": 0.5091363787651062, + "learning_rate": 1.8992147092577994e-06, + "loss": 0.3952, + "step": 6953 + }, + { + "epoch": 2.225042662116041, + "grad_norm": 0.5416716933250427, + "learning_rate": 1.8977543877483056e-06, + "loss": 0.4783, + "step": 6954 + }, + { + "epoch": 2.2253626279863483, + "grad_norm": 0.5534230470657349, + "learning_rate": 1.8962944963688984e-06, + "loss": 0.4238, + "step": 6955 + }, + { + "epoch": 2.2256825938566553, + "grad_norm": 0.4583647549152374, + "learning_rate": 1.8948350353219913e-06, + "loss": 0.42, + "step": 6956 + }, + { + "epoch": 2.2260025597269624, + "grad_norm": 0.46035411953926086, + "learning_rate": 1.893376004809942e-06, + "loss": 0.439, + "step": 6957 + }, + { + "epoch": 2.2263225255972694, + "grad_norm": 0.49934062361717224, + "learning_rate": 1.8919174050350487e-06, + "loss": 0.4576, + "step": 6958 + }, + { + "epoch": 2.226642491467577, + "grad_norm": 0.5063640475273132, + "learning_rate": 1.8904592361995467e-06, + "loss": 0.3689, + "step": 6959 + }, + { + "epoch": 2.226962457337884, + "grad_norm": 0.5481410622596741, + "learning_rate": 1.8890014985056115e-06, + "loss": 0.4838, + "step": 6960 + }, + { + "epoch": 2.227282423208191, + "grad_norm": 0.4754042327404022, + "learning_rate": 1.8875441921553606e-06, + "loss": 0.4294, + "step": 6961 + }, + { + "epoch": 2.2276023890784984, + "grad_norm": 0.48783764243125916, + "learning_rate": 1.8860873173508509e-06, + "loss": 0.4348, + "step": 6962 + }, + { + "epoch": 2.2279223549488054, + "grad_norm": 0.49847304821014404, + "learning_rate": 1.8846308742940834e-06, + "loss": 0.4538, + "step": 6963 + }, + { + "epoch": 2.2282423208191124, + "grad_norm": 0.5597853660583496, + "learning_rate": 1.8831748631869923e-06, + "loss": 0.4608, + "step": 6964 + }, + { + "epoch": 2.22856228668942, + "grad_norm": 0.5159894227981567, + "learning_rate": 1.8817192842314564e-06, + "loss": 0.4063, + "step": 6965 + }, + { + "epoch": 2.228882252559727, + "grad_norm": 0.482262521982193, + "learning_rate": 1.8802641376292913e-06, + "loss": 0.3973, + "step": 6966 + }, + { + "epoch": 2.229202218430034, + "grad_norm": 0.5018547773361206, + "learning_rate": 1.8788094235822563e-06, + "loss": 0.4351, + "step": 6967 + }, + { + "epoch": 2.2295221843003414, + "grad_norm": 0.48930323123931885, + "learning_rate": 1.8773551422920522e-06, + "loss": 0.4676, + "step": 6968 + }, + { + "epoch": 2.2298421501706485, + "grad_norm": 0.4622410237789154, + "learning_rate": 1.8759012939603133e-06, + "loss": 0.3874, + "step": 6969 + }, + { + "epoch": 2.2301621160409555, + "grad_norm": 0.4761062562465668, + "learning_rate": 1.8744478787886188e-06, + "loss": 0.4573, + "step": 6970 + }, + { + "epoch": 2.230482081911263, + "grad_norm": 0.4705923795700073, + "learning_rate": 1.8729948969784844e-06, + "loss": 0.413, + "step": 6971 + }, + { + "epoch": 2.23080204778157, + "grad_norm": 0.5186229348182678, + "learning_rate": 1.8715423487313694e-06, + "loss": 0.4403, + "step": 6972 + }, + { + "epoch": 2.231122013651877, + "grad_norm": 0.5065740346908569, + "learning_rate": 1.8700902342486732e-06, + "loss": 0.4422, + "step": 6973 + }, + { + "epoch": 2.2314419795221845, + "grad_norm": 0.4577891528606415, + "learning_rate": 1.8686385537317314e-06, + "loss": 0.3759, + "step": 6974 + }, + { + "epoch": 2.2317619453924915, + "grad_norm": 0.5119481682777405, + "learning_rate": 1.8671873073818208e-06, + "loss": 0.4536, + "step": 6975 + }, + { + "epoch": 2.2320819112627985, + "grad_norm": 0.5477125644683838, + "learning_rate": 1.8657364954001578e-06, + "loss": 0.4949, + "step": 6976 + }, + { + "epoch": 2.232401877133106, + "grad_norm": 0.48550277948379517, + "learning_rate": 1.8642861179878996e-06, + "loss": 0.4123, + "step": 6977 + }, + { + "epoch": 2.232721843003413, + "grad_norm": 0.5441885590553284, + "learning_rate": 1.8628361753461455e-06, + "loss": 0.4188, + "step": 6978 + }, + { + "epoch": 2.23304180887372, + "grad_norm": 0.5613593459129333, + "learning_rate": 1.86138666767593e-06, + "loss": 0.5034, + "step": 6979 + }, + { + "epoch": 2.2333617747440275, + "grad_norm": 0.4752897024154663, + "learning_rate": 1.859937595178229e-06, + "loss": 0.4031, + "step": 6980 + }, + { + "epoch": 2.2336817406143346, + "grad_norm": 0.5087677836418152, + "learning_rate": 1.858488958053956e-06, + "loss": 0.4517, + "step": 6981 + }, + { + "epoch": 2.2340017064846416, + "grad_norm": 0.4827876091003418, + "learning_rate": 1.8570407565039683e-06, + "loss": 0.4157, + "step": 6982 + }, + { + "epoch": 2.2343216723549486, + "grad_norm": 0.5602332353591919, + "learning_rate": 1.8555929907290627e-06, + "loss": 0.53, + "step": 6983 + }, + { + "epoch": 2.234641638225256, + "grad_norm": 0.5727624893188477, + "learning_rate": 1.8541456609299712e-06, + "loss": 0.4362, + "step": 6984 + }, + { + "epoch": 2.234961604095563, + "grad_norm": 0.5025842189788818, + "learning_rate": 1.852698767307367e-06, + "loss": 0.4298, + "step": 6985 + }, + { + "epoch": 2.23528156996587, + "grad_norm": 0.5044243335723877, + "learning_rate": 1.8512523100618673e-06, + "loss": 0.4565, + "step": 6986 + }, + { + "epoch": 2.2356015358361776, + "grad_norm": 0.4869266152381897, + "learning_rate": 1.8498062893940206e-06, + "loss": 0.4383, + "step": 6987 + }, + { + "epoch": 2.2359215017064846, + "grad_norm": 0.44381487369537354, + "learning_rate": 1.8483607055043234e-06, + "loss": 0.4324, + "step": 6988 + }, + { + "epoch": 2.2362414675767917, + "grad_norm": 0.4683745205402374, + "learning_rate": 1.8469155585932057e-06, + "loss": 0.4171, + "step": 6989 + }, + { + "epoch": 2.236561433447099, + "grad_norm": 0.5158386826515198, + "learning_rate": 1.8454708488610372e-06, + "loss": 0.4609, + "step": 6990 + }, + { + "epoch": 2.236881399317406, + "grad_norm": 0.48240604996681213, + "learning_rate": 1.8440265765081328e-06, + "loss": 0.4259, + "step": 6991 + }, + { + "epoch": 2.237201365187713, + "grad_norm": 0.4903813898563385, + "learning_rate": 1.8425827417347385e-06, + "loss": 0.4293, + "step": 6992 + }, + { + "epoch": 2.2375213310580206, + "grad_norm": 0.5533493757247925, + "learning_rate": 1.841139344741047e-06, + "loss": 0.4596, + "step": 6993 + }, + { + "epoch": 2.2378412969283277, + "grad_norm": 0.5921840071678162, + "learning_rate": 1.8396963857271855e-06, + "loss": 0.4857, + "step": 6994 + }, + { + "epoch": 2.2381612627986347, + "grad_norm": 0.4598381221294403, + "learning_rate": 1.8382538648932201e-06, + "loss": 0.405, + "step": 6995 + }, + { + "epoch": 2.238481228668942, + "grad_norm": 0.4713385999202728, + "learning_rate": 1.8368117824391623e-06, + "loss": 0.4308, + "step": 6996 + }, + { + "epoch": 2.238801194539249, + "grad_norm": 0.47838225960731506, + "learning_rate": 1.8353701385649548e-06, + "loss": 0.3965, + "step": 6997 + }, + { + "epoch": 2.239121160409556, + "grad_norm": 0.46495163440704346, + "learning_rate": 1.8339289334704857e-06, + "loss": 0.4581, + "step": 6998 + }, + { + "epoch": 2.2394411262798632, + "grad_norm": 0.5104029178619385, + "learning_rate": 1.832488167355579e-06, + "loss": 0.4513, + "step": 6999 + }, + { + "epoch": 2.2397610921501707, + "grad_norm": 0.4873708188533783, + "learning_rate": 1.831047840419997e-06, + "loss": 0.4124, + "step": 7000 + }, + { + "epoch": 2.2400810580204777, + "grad_norm": 0.4746980369091034, + "learning_rate": 1.8296079528634458e-06, + "loss": 0.4306, + "step": 7001 + }, + { + "epoch": 2.2404010238907848, + "grad_norm": 0.4791220426559448, + "learning_rate": 1.8281685048855647e-06, + "loss": 0.4564, + "step": 7002 + }, + { + "epoch": 2.2407209897610922, + "grad_norm": 0.4969911575317383, + "learning_rate": 1.8267294966859373e-06, + "loss": 0.4386, + "step": 7003 + }, + { + "epoch": 2.2410409556313993, + "grad_norm": 0.48363468050956726, + "learning_rate": 1.825290928464083e-06, + "loss": 0.4272, + "step": 7004 + }, + { + "epoch": 2.2413609215017063, + "grad_norm": 0.5502254962921143, + "learning_rate": 1.8238528004194589e-06, + "loss": 0.5083, + "step": 7005 + }, + { + "epoch": 2.2416808873720138, + "grad_norm": 0.5414071083068848, + "learning_rate": 1.8224151127514672e-06, + "loss": 0.481, + "step": 7006 + }, + { + "epoch": 2.242000853242321, + "grad_norm": 0.5145581960678101, + "learning_rate": 1.8209778656594406e-06, + "loss": 0.4167, + "step": 7007 + }, + { + "epoch": 2.242320819112628, + "grad_norm": 0.538152813911438, + "learning_rate": 1.8195410593426594e-06, + "loss": 0.4199, + "step": 7008 + }, + { + "epoch": 2.2426407849829353, + "grad_norm": 0.5265048742294312, + "learning_rate": 1.8181046940003366e-06, + "loss": 0.4057, + "step": 7009 + }, + { + "epoch": 2.2429607508532423, + "grad_norm": 0.49859923124313354, + "learning_rate": 1.8166687698316237e-06, + "loss": 0.4462, + "step": 7010 + }, + { + "epoch": 2.2432807167235493, + "grad_norm": 0.4799681305885315, + "learning_rate": 1.8152332870356177e-06, + "loss": 0.3907, + "step": 7011 + }, + { + "epoch": 2.243600682593857, + "grad_norm": 0.5121408700942993, + "learning_rate": 1.8137982458113457e-06, + "loss": 0.4462, + "step": 7012 + }, + { + "epoch": 2.243920648464164, + "grad_norm": 0.5352764129638672, + "learning_rate": 1.8123636463577821e-06, + "loss": 0.4413, + "step": 7013 + }, + { + "epoch": 2.244240614334471, + "grad_norm": 0.49394893646240234, + "learning_rate": 1.8109294888738338e-06, + "loss": 0.4361, + "step": 7014 + }, + { + "epoch": 2.2445605802047783, + "grad_norm": 0.4686887562274933, + "learning_rate": 1.8094957735583463e-06, + "loss": 0.4071, + "step": 7015 + }, + { + "epoch": 2.2448805460750854, + "grad_norm": 0.5014979243278503, + "learning_rate": 1.8080625006101098e-06, + "loss": 0.4334, + "step": 7016 + }, + { + "epoch": 2.2452005119453924, + "grad_norm": 0.5246603488922119, + "learning_rate": 1.8066296702278464e-06, + "loss": 0.4769, + "step": 7017 + }, + { + "epoch": 2.2455204778157, + "grad_norm": 0.5133163332939148, + "learning_rate": 1.805197282610222e-06, + "loss": 0.3907, + "step": 7018 + }, + { + "epoch": 2.245840443686007, + "grad_norm": 0.48508474230766296, + "learning_rate": 1.8037653379558368e-06, + "loss": 0.4269, + "step": 7019 + }, + { + "epoch": 2.246160409556314, + "grad_norm": 0.48519837856292725, + "learning_rate": 1.8023338364632336e-06, + "loss": 0.4751, + "step": 7020 + }, + { + "epoch": 2.2464803754266214, + "grad_norm": 0.4832985997200012, + "learning_rate": 1.8009027783308914e-06, + "loss": 0.4336, + "step": 7021 + }, + { + "epoch": 2.2468003412969284, + "grad_norm": 0.5109445452690125, + "learning_rate": 1.799472163757226e-06, + "loss": 0.4196, + "step": 7022 + }, + { + "epoch": 2.2471203071672354, + "grad_norm": 0.5440354347229004, + "learning_rate": 1.7980419929405967e-06, + "loss": 0.4751, + "step": 7023 + }, + { + "epoch": 2.2474402730375425, + "grad_norm": 0.7519711852073669, + "learning_rate": 1.7966122660792955e-06, + "loss": 0.4177, + "step": 7024 + }, + { + "epoch": 2.24776023890785, + "grad_norm": 0.5572145581245422, + "learning_rate": 1.795182983371559e-06, + "loss": 0.4815, + "step": 7025 + }, + { + "epoch": 2.248080204778157, + "grad_norm": 0.515664279460907, + "learning_rate": 1.7937541450155578e-06, + "loss": 0.4056, + "step": 7026 + }, + { + "epoch": 2.248400170648464, + "grad_norm": 0.5235795378684998, + "learning_rate": 1.7923257512093994e-06, + "loss": 0.4593, + "step": 7027 + }, + { + "epoch": 2.2487201365187715, + "grad_norm": 0.4442862570285797, + "learning_rate": 1.7908978021511365e-06, + "loss": 0.4151, + "step": 7028 + }, + { + "epoch": 2.2490401023890785, + "grad_norm": 0.5126864910125732, + "learning_rate": 1.7894702980387524e-06, + "loss": 0.4579, + "step": 7029 + }, + { + "epoch": 2.2493600682593855, + "grad_norm": 0.4956403970718384, + "learning_rate": 1.788043239070175e-06, + "loss": 0.4212, + "step": 7030 + }, + { + "epoch": 2.249680034129693, + "grad_norm": 0.5071899890899658, + "learning_rate": 1.786616625443267e-06, + "loss": 0.5038, + "step": 7031 + }, + { + "epoch": 2.25, + "grad_norm": 0.4895451068878174, + "learning_rate": 1.7851904573558276e-06, + "loss": 0.4153, + "step": 7032 + }, + { + "epoch": 2.250319965870307, + "grad_norm": 0.5462383031845093, + "learning_rate": 1.7837647350056008e-06, + "loss": 0.4266, + "step": 7033 + }, + { + "epoch": 2.2506399317406145, + "grad_norm": 0.5014021992683411, + "learning_rate": 1.7823394585902615e-06, + "loss": 0.4568, + "step": 7034 + }, + { + "epoch": 2.2509598976109215, + "grad_norm": 0.47650790214538574, + "learning_rate": 1.780914628307428e-06, + "loss": 0.4126, + "step": 7035 + }, + { + "epoch": 2.2512798634812285, + "grad_norm": 0.5031761527061462, + "learning_rate": 1.7794902443546546e-06, + "loss": 0.4157, + "step": 7036 + }, + { + "epoch": 2.251599829351536, + "grad_norm": 0.5211197137832642, + "learning_rate": 1.7780663069294312e-06, + "loss": 0.4403, + "step": 7037 + }, + { + "epoch": 2.251919795221843, + "grad_norm": 0.49045392870903015, + "learning_rate": 1.776642816229192e-06, + "loss": 0.4109, + "step": 7038 + }, + { + "epoch": 2.25223976109215, + "grad_norm": 0.5195350646972656, + "learning_rate": 1.7752197724513027e-06, + "loss": 0.4583, + "step": 7039 + }, + { + "epoch": 2.252559726962457, + "grad_norm": 0.468694269657135, + "learning_rate": 1.7737971757930732e-06, + "loss": 0.4034, + "step": 7040 + }, + { + "epoch": 2.2528796928327646, + "grad_norm": 0.5023847818374634, + "learning_rate": 1.7723750264517464e-06, + "loss": 0.4527, + "step": 7041 + }, + { + "epoch": 2.2531996587030716, + "grad_norm": 0.4959229528903961, + "learning_rate": 1.7709533246245032e-06, + "loss": 0.4218, + "step": 7042 + }, + { + "epoch": 2.2535196245733786, + "grad_norm": 0.5296538472175598, + "learning_rate": 1.7695320705084678e-06, + "loss": 0.523, + "step": 7043 + }, + { + "epoch": 2.253839590443686, + "grad_norm": 0.4721936583518982, + "learning_rate": 1.768111264300696e-06, + "loss": 0.4433, + "step": 7044 + }, + { + "epoch": 2.254159556313993, + "grad_norm": 0.5049819350242615, + "learning_rate": 1.7666909061981869e-06, + "loss": 0.4704, + "step": 7045 + }, + { + "epoch": 2.2544795221843, + "grad_norm": 0.5241411924362183, + "learning_rate": 1.765270996397873e-06, + "loss": 0.4218, + "step": 7046 + }, + { + "epoch": 2.2547994880546076, + "grad_norm": 0.5348628163337708, + "learning_rate": 1.7638515350966257e-06, + "loss": 0.4897, + "step": 7047 + }, + { + "epoch": 2.2551194539249146, + "grad_norm": 0.549798846244812, + "learning_rate": 1.762432522491258e-06, + "loss": 0.432, + "step": 7048 + }, + { + "epoch": 2.2554394197952217, + "grad_norm": 0.5057252049446106, + "learning_rate": 1.761013958778514e-06, + "loss": 0.4452, + "step": 7049 + }, + { + "epoch": 2.255759385665529, + "grad_norm": 0.47426411509513855, + "learning_rate": 1.759595844155083e-06, + "loss": 0.4149, + "step": 7050 + }, + { + "epoch": 2.256079351535836, + "grad_norm": 0.45348483324050903, + "learning_rate": 1.758178178817585e-06, + "loss": 0.4352, + "step": 7051 + }, + { + "epoch": 2.256399317406143, + "grad_norm": 0.4907292127609253, + "learning_rate": 1.7567609629625836e-06, + "loss": 0.4177, + "step": 7052 + }, + { + "epoch": 2.2567192832764507, + "grad_norm": 0.5103530287742615, + "learning_rate": 1.755344196786577e-06, + "loss": 0.4333, + "step": 7053 + }, + { + "epoch": 2.2570392491467577, + "grad_norm": 0.5013543367385864, + "learning_rate": 1.7539278804859993e-06, + "loss": 0.4393, + "step": 7054 + }, + { + "epoch": 2.2573592150170647, + "grad_norm": 0.4777149260044098, + "learning_rate": 1.752512014257227e-06, + "loss": 0.4428, + "step": 7055 + }, + { + "epoch": 2.257679180887372, + "grad_norm": 0.4968894124031067, + "learning_rate": 1.75109659829657e-06, + "loss": 0.4251, + "step": 7056 + }, + { + "epoch": 2.257999146757679, + "grad_norm": 0.46732962131500244, + "learning_rate": 1.7496816328002797e-06, + "loss": 0.4255, + "step": 7057 + }, + { + "epoch": 2.2583191126279862, + "grad_norm": 0.5094853639602661, + "learning_rate": 1.7482671179645388e-06, + "loss": 0.4488, + "step": 7058 + }, + { + "epoch": 2.2586390784982937, + "grad_norm": 0.5172582864761353, + "learning_rate": 1.746853053985476e-06, + "loss": 0.4294, + "step": 7059 + }, + { + "epoch": 2.2589590443686007, + "grad_norm": 0.534986138343811, + "learning_rate": 1.7454394410591514e-06, + "loss": 0.4278, + "step": 7060 + }, + { + "epoch": 2.2592790102389078, + "grad_norm": 0.46675431728363037, + "learning_rate": 1.7440262793815615e-06, + "loss": 0.4329, + "step": 7061 + }, + { + "epoch": 2.2595989761092152, + "grad_norm": 0.4917410612106323, + "learning_rate": 1.7426135691486467e-06, + "loss": 0.4416, + "step": 7062 + }, + { + "epoch": 2.2599189419795223, + "grad_norm": 0.5392761826515198, + "learning_rate": 1.7412013105562775e-06, + "loss": 0.3825, + "step": 7063 + }, + { + "epoch": 2.2602389078498293, + "grad_norm": 0.548638641834259, + "learning_rate": 1.7397895038002689e-06, + "loss": 0.4479, + "step": 7064 + }, + { + "epoch": 2.2605588737201368, + "grad_norm": 0.5172118544578552, + "learning_rate": 1.7383781490763678e-06, + "loss": 0.4856, + "step": 7065 + }, + { + "epoch": 2.260878839590444, + "grad_norm": 0.49654272198677063, + "learning_rate": 1.7369672465802585e-06, + "loss": 0.4263, + "step": 7066 + }, + { + "epoch": 2.261198805460751, + "grad_norm": 0.4977187514305115, + "learning_rate": 1.7355567965075675e-06, + "loss": 0.4491, + "step": 7067 + }, + { + "epoch": 2.261518771331058, + "grad_norm": 0.542453944683075, + "learning_rate": 1.734146799053852e-06, + "loss": 0.4566, + "step": 7068 + }, + { + "epoch": 2.2618387372013653, + "grad_norm": 0.4947512745857239, + "learning_rate": 1.7327372544146137e-06, + "loss": 0.4352, + "step": 7069 + }, + { + "epoch": 2.2621587030716723, + "grad_norm": 0.4490604102611542, + "learning_rate": 1.7313281627852857e-06, + "loss": 0.4436, + "step": 7070 + }, + { + "epoch": 2.2624786689419794, + "grad_norm": 0.5164769291877747, + "learning_rate": 1.7299195243612382e-06, + "loss": 0.4845, + "step": 7071 + }, + { + "epoch": 2.262798634812287, + "grad_norm": 0.5048587322235107, + "learning_rate": 1.7285113393377845e-06, + "loss": 0.4152, + "step": 7072 + }, + { + "epoch": 2.263118600682594, + "grad_norm": 0.5196941494941711, + "learning_rate": 1.7271036079101678e-06, + "loss": 0.4732, + "step": 7073 + }, + { + "epoch": 2.263438566552901, + "grad_norm": 0.5368337035179138, + "learning_rate": 1.7256963302735752e-06, + "loss": 0.483, + "step": 7074 + }, + { + "epoch": 2.2637585324232083, + "grad_norm": 0.4829712510108948, + "learning_rate": 1.7242895066231247e-06, + "loss": 0.432, + "step": 7075 + }, + { + "epoch": 2.2640784982935154, + "grad_norm": 0.5370932817459106, + "learning_rate": 1.722883137153874e-06, + "loss": 0.4103, + "step": 7076 + }, + { + "epoch": 2.2643984641638224, + "grad_norm": 0.5939807295799255, + "learning_rate": 1.72147722206082e-06, + "loss": 0.4691, + "step": 7077 + }, + { + "epoch": 2.26471843003413, + "grad_norm": 0.45269444584846497, + "learning_rate": 1.7200717615388917e-06, + "loss": 0.4499, + "step": 7078 + }, + { + "epoch": 2.265038395904437, + "grad_norm": 0.43377363681793213, + "learning_rate": 1.7186667557829612e-06, + "loss": 0.4164, + "step": 7079 + }, + { + "epoch": 2.265358361774744, + "grad_norm": 0.49725332856178284, + "learning_rate": 1.7172622049878329e-06, + "loss": 0.4945, + "step": 7080 + }, + { + "epoch": 2.265678327645051, + "grad_norm": 0.5128596425056458, + "learning_rate": 1.7158581093482467e-06, + "loss": 0.4484, + "step": 7081 + }, + { + "epoch": 2.2659982935153584, + "grad_norm": 0.47245702147483826, + "learning_rate": 1.7144544690588872e-06, + "loss": 0.4514, + "step": 7082 + }, + { + "epoch": 2.2663182593856654, + "grad_norm": 0.427586168050766, + "learning_rate": 1.713051284314366e-06, + "loss": 0.3753, + "step": 7083 + }, + { + "epoch": 2.2666382252559725, + "grad_norm": 0.5714744329452515, + "learning_rate": 1.71164855530924e-06, + "loss": 0.5082, + "step": 7084 + }, + { + "epoch": 2.26695819112628, + "grad_norm": 0.5184192061424255, + "learning_rate": 1.7102462822379977e-06, + "loss": 0.424, + "step": 7085 + }, + { + "epoch": 2.267278156996587, + "grad_norm": 0.5154229402542114, + "learning_rate": 1.7088444652950647e-06, + "loss": 0.4218, + "step": 7086 + }, + { + "epoch": 2.267598122866894, + "grad_norm": 0.4990445077419281, + "learning_rate": 1.7074431046748075e-06, + "loss": 0.4072, + "step": 7087 + }, + { + "epoch": 2.2679180887372015, + "grad_norm": 0.48973727226257324, + "learning_rate": 1.7060422005715232e-06, + "loss": 0.4353, + "step": 7088 + }, + { + "epoch": 2.2682380546075085, + "grad_norm": 0.479113906621933, + "learning_rate": 1.7046417531794523e-06, + "loss": 0.4334, + "step": 7089 + }, + { + "epoch": 2.2685580204778155, + "grad_norm": 0.5312931537628174, + "learning_rate": 1.7032417626927666e-06, + "loss": 0.4662, + "step": 7090 + }, + { + "epoch": 2.268877986348123, + "grad_norm": 0.4754203259944916, + "learning_rate": 1.7018422293055753e-06, + "loss": 0.4176, + "step": 7091 + }, + { + "epoch": 2.26919795221843, + "grad_norm": 0.5042561888694763, + "learning_rate": 1.7004431532119264e-06, + "loss": 0.4286, + "step": 7092 + }, + { + "epoch": 2.269517918088737, + "grad_norm": 0.4862496554851532, + "learning_rate": 1.6990445346058055e-06, + "loss": 0.3956, + "step": 7093 + }, + { + "epoch": 2.2698378839590445, + "grad_norm": 0.5385366678237915, + "learning_rate": 1.6976463736811306e-06, + "loss": 0.4688, + "step": 7094 + }, + { + "epoch": 2.2701578498293515, + "grad_norm": 0.5797232389450073, + "learning_rate": 1.6962486706317594e-06, + "loss": 0.4677, + "step": 7095 + }, + { + "epoch": 2.2704778156996586, + "grad_norm": 0.5136889219284058, + "learning_rate": 1.6948514256514826e-06, + "loss": 0.413, + "step": 7096 + }, + { + "epoch": 2.270797781569966, + "grad_norm": 0.5192843675613403, + "learning_rate": 1.6934546389340317e-06, + "loss": 0.5074, + "step": 7097 + }, + { + "epoch": 2.271117747440273, + "grad_norm": 0.4660189747810364, + "learning_rate": 1.6920583106730749e-06, + "loss": 0.3723, + "step": 7098 + }, + { + "epoch": 2.27143771331058, + "grad_norm": 0.5324483513832092, + "learning_rate": 1.6906624410622124e-06, + "loss": 0.4553, + "step": 7099 + }, + { + "epoch": 2.2717576791808876, + "grad_norm": 0.4562549293041229, + "learning_rate": 1.6892670302949842e-06, + "loss": 0.4392, + "step": 7100 + }, + { + "epoch": 2.2720776450511946, + "grad_norm": 0.5018472671508789, + "learning_rate": 1.6878720785648633e-06, + "loss": 0.4063, + "step": 7101 + }, + { + "epoch": 2.2723976109215016, + "grad_norm": 0.4813690483570099, + "learning_rate": 1.6864775860652638e-06, + "loss": 0.4224, + "step": 7102 + }, + { + "epoch": 2.272717576791809, + "grad_norm": 0.5093215703964233, + "learning_rate": 1.6850835529895344e-06, + "loss": 0.466, + "step": 7103 + }, + { + "epoch": 2.273037542662116, + "grad_norm": 0.5317978262901306, + "learning_rate": 1.683689979530958e-06, + "loss": 0.443, + "step": 7104 + }, + { + "epoch": 2.273357508532423, + "grad_norm": 0.5211302638053894, + "learning_rate": 1.682296865882756e-06, + "loss": 0.4428, + "step": 7105 + }, + { + "epoch": 2.2736774744027306, + "grad_norm": 0.45339956879615784, + "learning_rate": 1.680904212238083e-06, + "loss": 0.4145, + "step": 7106 + }, + { + "epoch": 2.2739974402730376, + "grad_norm": 0.5466711521148682, + "learning_rate": 1.6795120187900344e-06, + "loss": 0.4522, + "step": 7107 + }, + { + "epoch": 2.2743174061433447, + "grad_norm": 0.5057246088981628, + "learning_rate": 1.6781202857316402e-06, + "loss": 0.463, + "step": 7108 + }, + { + "epoch": 2.274637372013652, + "grad_norm": 0.47965845465660095, + "learning_rate": 1.676729013255865e-06, + "loss": 0.4392, + "step": 7109 + }, + { + "epoch": 2.274957337883959, + "grad_norm": 0.4975675940513611, + "learning_rate": 1.6753382015556096e-06, + "loss": 0.4446, + "step": 7110 + }, + { + "epoch": 2.275277303754266, + "grad_norm": 0.50555020570755, + "learning_rate": 1.673947850823711e-06, + "loss": 0.4527, + "step": 7111 + }, + { + "epoch": 2.275597269624573, + "grad_norm": 0.49229076504707336, + "learning_rate": 1.6725579612529441e-06, + "loss": 0.4666, + "step": 7112 + }, + { + "epoch": 2.2759172354948807, + "grad_norm": 0.4973360598087311, + "learning_rate": 1.6711685330360212e-06, + "loss": 0.4604, + "step": 7113 + }, + { + "epoch": 2.2762372013651877, + "grad_norm": 0.4849456548690796, + "learning_rate": 1.6697795663655848e-06, + "loss": 0.4274, + "step": 7114 + }, + { + "epoch": 2.2765571672354947, + "grad_norm": 0.47988003492355347, + "learning_rate": 1.6683910614342186e-06, + "loss": 0.4254, + "step": 7115 + }, + { + "epoch": 2.276877133105802, + "grad_norm": 0.5450612306594849, + "learning_rate": 1.6670030184344377e-06, + "loss": 0.4667, + "step": 7116 + }, + { + "epoch": 2.2771970989761092, + "grad_norm": 0.502415657043457, + "learning_rate": 1.665615437558698e-06, + "loss": 0.4497, + "step": 7117 + }, + { + "epoch": 2.2775170648464163, + "grad_norm": 0.456184983253479, + "learning_rate": 1.6642283189993912e-06, + "loss": 0.407, + "step": 7118 + }, + { + "epoch": 2.2778370307167237, + "grad_norm": 0.5014293789863586, + "learning_rate": 1.662841662948841e-06, + "loss": 0.4767, + "step": 7119 + }, + { + "epoch": 2.2781569965870307, + "grad_norm": 0.5107823014259338, + "learning_rate": 1.6614554695993085e-06, + "loss": 0.3857, + "step": 7120 + }, + { + "epoch": 2.2784769624573378, + "grad_norm": 0.5226872563362122, + "learning_rate": 1.6600697391429904e-06, + "loss": 0.4456, + "step": 7121 + }, + { + "epoch": 2.2787969283276452, + "grad_norm": 0.5985218286514282, + "learning_rate": 1.6586844717720208e-06, + "loss": 0.442, + "step": 7122 + }, + { + "epoch": 2.2791168941979523, + "grad_norm": 0.5533532500267029, + "learning_rate": 1.6572996676784704e-06, + "loss": 0.4335, + "step": 7123 + }, + { + "epoch": 2.2794368600682593, + "grad_norm": 0.49049973487854004, + "learning_rate": 1.6559153270543421e-06, + "loss": 0.4206, + "step": 7124 + }, + { + "epoch": 2.2797568259385663, + "grad_norm": 0.4918212592601776, + "learning_rate": 1.6545314500915771e-06, + "loss": 0.3945, + "step": 7125 + }, + { + "epoch": 2.280076791808874, + "grad_norm": 0.48216524720191956, + "learning_rate": 1.6531480369820497e-06, + "loss": 0.4967, + "step": 7126 + }, + { + "epoch": 2.280396757679181, + "grad_norm": 0.5223669409751892, + "learning_rate": 1.6517650879175733e-06, + "loss": 0.4274, + "step": 7127 + }, + { + "epoch": 2.280716723549488, + "grad_norm": 0.5535790920257568, + "learning_rate": 1.6503826030898973e-06, + "loss": 0.5179, + "step": 7128 + }, + { + "epoch": 2.2810366894197953, + "grad_norm": 0.5251159071922302, + "learning_rate": 1.6490005826907029e-06, + "loss": 0.368, + "step": 7129 + }, + { + "epoch": 2.2813566552901023, + "grad_norm": 0.5006959438323975, + "learning_rate": 1.6476190269116087e-06, + "loss": 0.401, + "step": 7130 + }, + { + "epoch": 2.2816766211604094, + "grad_norm": 0.5298047661781311, + "learning_rate": 1.6462379359441682e-06, + "loss": 0.4692, + "step": 7131 + }, + { + "epoch": 2.281996587030717, + "grad_norm": 0.50528484582901, + "learning_rate": 1.6448573099798727e-06, + "loss": 0.4637, + "step": 7132 + }, + { + "epoch": 2.282316552901024, + "grad_norm": 0.47289609909057617, + "learning_rate": 1.6434771492101487e-06, + "loss": 0.4288, + "step": 7133 + }, + { + "epoch": 2.282636518771331, + "grad_norm": 0.5515410900115967, + "learning_rate": 1.6420974538263561e-06, + "loss": 0.5173, + "step": 7134 + }, + { + "epoch": 2.2829564846416384, + "grad_norm": 0.570210874080658, + "learning_rate": 1.6407182240197916e-06, + "loss": 0.4378, + "step": 7135 + }, + { + "epoch": 2.2832764505119454, + "grad_norm": 0.4833798110485077, + "learning_rate": 1.6393394599816847e-06, + "loss": 0.4007, + "step": 7136 + }, + { + "epoch": 2.2835964163822524, + "grad_norm": 0.47280073165893555, + "learning_rate": 1.6379611619032042e-06, + "loss": 0.3989, + "step": 7137 + }, + { + "epoch": 2.28391638225256, + "grad_norm": 0.48720160126686096, + "learning_rate": 1.6365833299754558e-06, + "loss": 0.4593, + "step": 7138 + }, + { + "epoch": 2.284236348122867, + "grad_norm": 0.49480339884757996, + "learning_rate": 1.635205964389474e-06, + "loss": 0.4236, + "step": 7139 + }, + { + "epoch": 2.284556313993174, + "grad_norm": 0.4934588074684143, + "learning_rate": 1.6338290653362338e-06, + "loss": 0.4237, + "step": 7140 + }, + { + "epoch": 2.2848762798634814, + "grad_norm": 0.49687129259109497, + "learning_rate": 1.6324526330066414e-06, + "loss": 0.4424, + "step": 7141 + }, + { + "epoch": 2.2851962457337884, + "grad_norm": 0.5061244964599609, + "learning_rate": 1.631076667591543e-06, + "loss": 0.44, + "step": 7142 + }, + { + "epoch": 2.2855162116040955, + "grad_norm": 0.42954832315444946, + "learning_rate": 1.6297011692817189e-06, + "loss": 0.3868, + "step": 7143 + }, + { + "epoch": 2.285836177474403, + "grad_norm": 0.4803065061569214, + "learning_rate": 1.6283261382678827e-06, + "loss": 0.4496, + "step": 7144 + }, + { + "epoch": 2.28615614334471, + "grad_norm": 0.5327868461608887, + "learning_rate": 1.6269515747406832e-06, + "loss": 0.4348, + "step": 7145 + }, + { + "epoch": 2.286476109215017, + "grad_norm": 0.5120851993560791, + "learning_rate": 1.6255774788907043e-06, + "loss": 0.4487, + "step": 7146 + }, + { + "epoch": 2.2867960750853245, + "grad_norm": 0.535893440246582, + "learning_rate": 1.6242038509084672e-06, + "loss": 0.4904, + "step": 7147 + }, + { + "epoch": 2.2871160409556315, + "grad_norm": 0.49842455983161926, + "learning_rate": 1.6228306909844294e-06, + "loss": 0.4106, + "step": 7148 + }, + { + "epoch": 2.2874360068259385, + "grad_norm": 0.5016635060310364, + "learning_rate": 1.6214579993089784e-06, + "loss": 0.4354, + "step": 7149 + }, + { + "epoch": 2.287755972696246, + "grad_norm": 0.4777538478374481, + "learning_rate": 1.6200857760724408e-06, + "loss": 0.4368, + "step": 7150 + }, + { + "epoch": 2.288075938566553, + "grad_norm": 0.5343720316886902, + "learning_rate": 1.6187140214650753e-06, + "loss": 0.4194, + "step": 7151 + }, + { + "epoch": 2.28839590443686, + "grad_norm": 0.5467803478240967, + "learning_rate": 1.6173427356770778e-06, + "loss": 0.4866, + "step": 7152 + }, + { + "epoch": 2.288715870307167, + "grad_norm": 0.5415265560150146, + "learning_rate": 1.615971918898581e-06, + "loss": 0.4156, + "step": 7153 + }, + { + "epoch": 2.2890358361774745, + "grad_norm": 0.45446741580963135, + "learning_rate": 1.6146015713196495e-06, + "loss": 0.4037, + "step": 7154 + }, + { + "epoch": 2.2893558020477816, + "grad_norm": 0.5302549004554749, + "learning_rate": 1.613231693130282e-06, + "loss": 0.4718, + "step": 7155 + }, + { + "epoch": 2.2896757679180886, + "grad_norm": 0.5353044271469116, + "learning_rate": 1.611862284520414e-06, + "loss": 0.4428, + "step": 7156 + }, + { + "epoch": 2.289995733788396, + "grad_norm": 0.5557809472084045, + "learning_rate": 1.6104933456799154e-06, + "loss": 0.4133, + "step": 7157 + }, + { + "epoch": 2.290315699658703, + "grad_norm": 0.5576965808868408, + "learning_rate": 1.6091248767985939e-06, + "loss": 0.4685, + "step": 7158 + }, + { + "epoch": 2.29063566552901, + "grad_norm": 0.4604426920413971, + "learning_rate": 1.607756878066188e-06, + "loss": 0.4025, + "step": 7159 + }, + { + "epoch": 2.2909556313993176, + "grad_norm": 0.4758125841617584, + "learning_rate": 1.6063893496723699e-06, + "loss": 0.43, + "step": 7160 + }, + { + "epoch": 2.2912755972696246, + "grad_norm": 0.47438958287239075, + "learning_rate": 1.6050222918067527e-06, + "loss": 0.4646, + "step": 7161 + }, + { + "epoch": 2.2915955631399316, + "grad_norm": 0.48651546239852905, + "learning_rate": 1.6036557046588775e-06, + "loss": 0.4537, + "step": 7162 + }, + { + "epoch": 2.291915529010239, + "grad_norm": 0.47285035252571106, + "learning_rate": 1.602289588418226e-06, + "loss": 0.4457, + "step": 7163 + }, + { + "epoch": 2.292235494880546, + "grad_norm": 0.5200011730194092, + "learning_rate": 1.6009239432742114e-06, + "loss": 0.4138, + "step": 7164 + }, + { + "epoch": 2.292555460750853, + "grad_norm": 0.5149220824241638, + "learning_rate": 1.599558769416179e-06, + "loss": 0.4855, + "step": 7165 + }, + { + "epoch": 2.29287542662116, + "grad_norm": 0.5541476607322693, + "learning_rate": 1.5981940670334156e-06, + "loss": 0.4703, + "step": 7166 + }, + { + "epoch": 2.2931953924914676, + "grad_norm": 0.5186576843261719, + "learning_rate": 1.5968298363151364e-06, + "loss": 0.4403, + "step": 7167 + }, + { + "epoch": 2.2935153583617747, + "grad_norm": 0.5135532021522522, + "learning_rate": 1.5954660774504954e-06, + "loss": 0.4437, + "step": 7168 + }, + { + "epoch": 2.2938353242320817, + "grad_norm": 0.5300794839859009, + "learning_rate": 1.5941027906285788e-06, + "loss": 0.4179, + "step": 7169 + }, + { + "epoch": 2.294155290102389, + "grad_norm": 0.5552196502685547, + "learning_rate": 1.592739976038406e-06, + "loss": 0.4478, + "step": 7170 + }, + { + "epoch": 2.294475255972696, + "grad_norm": 0.5434558987617493, + "learning_rate": 1.5913776338689364e-06, + "loss": 0.3948, + "step": 7171 + }, + { + "epoch": 2.294795221843003, + "grad_norm": 0.5269661545753479, + "learning_rate": 1.5900157643090563e-06, + "loss": 0.4469, + "step": 7172 + }, + { + "epoch": 2.2951151877133107, + "grad_norm": 0.5204504132270813, + "learning_rate": 1.5886543675475946e-06, + "loss": 0.4475, + "step": 7173 + }, + { + "epoch": 2.2954351535836177, + "grad_norm": 0.5484360456466675, + "learning_rate": 1.5872934437733096e-06, + "loss": 0.4304, + "step": 7174 + }, + { + "epoch": 2.2957551194539247, + "grad_norm": 0.5295978784561157, + "learning_rate": 1.5859329931748918e-06, + "loss": 0.4746, + "step": 7175 + }, + { + "epoch": 2.296075085324232, + "grad_norm": 0.5483356714248657, + "learning_rate": 1.5845730159409734e-06, + "loss": 0.4469, + "step": 7176 + }, + { + "epoch": 2.2963950511945392, + "grad_norm": 0.5087558031082153, + "learning_rate": 1.583213512260114e-06, + "loss": 0.4556, + "step": 7177 + }, + { + "epoch": 2.2967150170648463, + "grad_norm": 0.5023752450942993, + "learning_rate": 1.5818544823208126e-06, + "loss": 0.4411, + "step": 7178 + }, + { + "epoch": 2.2970349829351537, + "grad_norm": 0.5368869304656982, + "learning_rate": 1.5804959263115e-06, + "loss": 0.4111, + "step": 7179 + }, + { + "epoch": 2.2973549488054608, + "grad_norm": 0.5652300715446472, + "learning_rate": 1.5791378444205386e-06, + "loss": 0.4498, + "step": 7180 + }, + { + "epoch": 2.297674914675768, + "grad_norm": 0.5881531834602356, + "learning_rate": 1.5777802368362327e-06, + "loss": 0.4156, + "step": 7181 + }, + { + "epoch": 2.2979948805460753, + "grad_norm": 0.495271235704422, + "learning_rate": 1.576423103746812e-06, + "loss": 0.4475, + "step": 7182 + }, + { + "epoch": 2.2983148464163823, + "grad_norm": 0.4825277030467987, + "learning_rate": 1.5750664453404485e-06, + "loss": 0.4638, + "step": 7183 + }, + { + "epoch": 2.2986348122866893, + "grad_norm": 0.47132453322410583, + "learning_rate": 1.573710261805243e-06, + "loss": 0.4329, + "step": 7184 + }, + { + "epoch": 2.298954778156997, + "grad_norm": 0.46688440442085266, + "learning_rate": 1.5723545533292294e-06, + "loss": 0.3808, + "step": 7185 + }, + { + "epoch": 2.299274744027304, + "grad_norm": 0.5292215347290039, + "learning_rate": 1.5709993201003827e-06, + "loss": 0.4736, + "step": 7186 + }, + { + "epoch": 2.299594709897611, + "grad_norm": 0.5536664128303528, + "learning_rate": 1.5696445623066031e-06, + "loss": 0.4321, + "step": 7187 + }, + { + "epoch": 2.2999146757679183, + "grad_norm": 0.5120112299919128, + "learning_rate": 1.568290280135734e-06, + "loss": 0.4162, + "step": 7188 + }, + { + "epoch": 2.3002346416382253, + "grad_norm": 0.49979880452156067, + "learning_rate": 1.5669364737755455e-06, + "loss": 0.4444, + "step": 7189 + }, + { + "epoch": 2.3005546075085324, + "grad_norm": 0.5113063454627991, + "learning_rate": 1.5655831434137437e-06, + "loss": 0.4755, + "step": 7190 + }, + { + "epoch": 2.30087457337884, + "grad_norm": 0.4913010001182556, + "learning_rate": 1.5642302892379708e-06, + "loss": 0.4138, + "step": 7191 + }, + { + "epoch": 2.301194539249147, + "grad_norm": 0.48320209980010986, + "learning_rate": 1.5628779114358034e-06, + "loss": 0.4397, + "step": 7192 + }, + { + "epoch": 2.301514505119454, + "grad_norm": 0.5086919069290161, + "learning_rate": 1.5615260101947482e-06, + "loss": 0.4466, + "step": 7193 + }, + { + "epoch": 2.301834470989761, + "grad_norm": 0.5563963055610657, + "learning_rate": 1.5601745857022483e-06, + "loss": 0.4394, + "step": 7194 + }, + { + "epoch": 2.3021544368600684, + "grad_norm": 0.5221656560897827, + "learning_rate": 1.5588236381456789e-06, + "loss": 0.4436, + "step": 7195 + }, + { + "epoch": 2.3024744027303754, + "grad_norm": 0.5022855997085571, + "learning_rate": 1.5574731677123516e-06, + "loss": 0.4231, + "step": 7196 + }, + { + "epoch": 2.3027943686006824, + "grad_norm": 0.5329191088676453, + "learning_rate": 1.5561231745895127e-06, + "loss": 0.4163, + "step": 7197 + }, + { + "epoch": 2.30311433447099, + "grad_norm": 0.5382068753242493, + "learning_rate": 1.5547736589643393e-06, + "loss": 0.415, + "step": 7198 + }, + { + "epoch": 2.303434300341297, + "grad_norm": 0.4856245219707489, + "learning_rate": 1.5534246210239406e-06, + "loss": 0.4048, + "step": 7199 + }, + { + "epoch": 2.303754266211604, + "grad_norm": 0.6377033591270447, + "learning_rate": 1.5520760609553665e-06, + "loss": 0.4741, + "step": 7200 + }, + { + "epoch": 2.3040742320819114, + "grad_norm": 0.5466446280479431, + "learning_rate": 1.5507279789455925e-06, + "loss": 0.4119, + "step": 7201 + }, + { + "epoch": 2.3043941979522184, + "grad_norm": 0.5324555039405823, + "learning_rate": 1.5493803751815357e-06, + "loss": 0.4644, + "step": 7202 + }, + { + "epoch": 2.3047141638225255, + "grad_norm": 0.5565164685249329, + "learning_rate": 1.548033249850041e-06, + "loss": 0.4468, + "step": 7203 + }, + { + "epoch": 2.305034129692833, + "grad_norm": 0.505616307258606, + "learning_rate": 1.5466866031378874e-06, + "loss": 0.445, + "step": 7204 + }, + { + "epoch": 2.30535409556314, + "grad_norm": 0.5002743005752563, + "learning_rate": 1.5453404352317923e-06, + "loss": 0.4665, + "step": 7205 + }, + { + "epoch": 2.305674061433447, + "grad_norm": 0.49529916048049927, + "learning_rate": 1.5439947463184003e-06, + "loss": 0.4552, + "step": 7206 + }, + { + "epoch": 2.305994027303754, + "grad_norm": 0.4952756464481354, + "learning_rate": 1.5426495365842958e-06, + "loss": 0.3944, + "step": 7207 + }, + { + "epoch": 2.3063139931740615, + "grad_norm": 0.5098868608474731, + "learning_rate": 1.541304806215993e-06, + "loss": 0.4404, + "step": 7208 + }, + { + "epoch": 2.3066339590443685, + "grad_norm": 0.5375380516052246, + "learning_rate": 1.5399605553999374e-06, + "loss": 0.4544, + "step": 7209 + }, + { + "epoch": 2.3069539249146755, + "grad_norm": 0.5279937982559204, + "learning_rate": 1.5386167843225153e-06, + "loss": 0.4714, + "step": 7210 + }, + { + "epoch": 2.307273890784983, + "grad_norm": 0.4866258203983307, + "learning_rate": 1.5372734931700384e-06, + "loss": 0.4295, + "step": 7211 + }, + { + "epoch": 2.30759385665529, + "grad_norm": 0.4863169491291046, + "learning_rate": 1.5359306821287595e-06, + "loss": 0.4156, + "step": 7212 + }, + { + "epoch": 2.307913822525597, + "grad_norm": 0.530881404876709, + "learning_rate": 1.5345883513848586e-06, + "loss": 0.5004, + "step": 7213 + }, + { + "epoch": 2.3082337883959045, + "grad_norm": 0.5486757755279541, + "learning_rate": 1.5332465011244497e-06, + "loss": 0.4835, + "step": 7214 + }, + { + "epoch": 2.3085537542662116, + "grad_norm": 0.4945541322231293, + "learning_rate": 1.5319051315335865e-06, + "loss": 0.4081, + "step": 7215 + }, + { + "epoch": 2.3088737201365186, + "grad_norm": 0.511341392993927, + "learning_rate": 1.530564242798247e-06, + "loss": 0.478, + "step": 7216 + }, + { + "epoch": 2.309193686006826, + "grad_norm": 0.555908739566803, + "learning_rate": 1.5292238351043503e-06, + "loss": 0.4588, + "step": 7217 + }, + { + "epoch": 2.309513651877133, + "grad_norm": 0.47775009274482727, + "learning_rate": 1.5278839086377445e-06, + "loss": 0.3828, + "step": 7218 + }, + { + "epoch": 2.30983361774744, + "grad_norm": 0.5278932452201843, + "learning_rate": 1.5265444635842109e-06, + "loss": 0.4886, + "step": 7219 + }, + { + "epoch": 2.3101535836177476, + "grad_norm": 0.509454607963562, + "learning_rate": 1.525205500129467e-06, + "loss": 0.4453, + "step": 7220 + }, + { + "epoch": 2.3104735494880546, + "grad_norm": 0.4423026442527771, + "learning_rate": 1.5238670184591591e-06, + "loss": 0.4265, + "step": 7221 + }, + { + "epoch": 2.3107935153583616, + "grad_norm": 0.48173072934150696, + "learning_rate": 1.5225290187588732e-06, + "loss": 0.4352, + "step": 7222 + }, + { + "epoch": 2.311113481228669, + "grad_norm": 0.5268172025680542, + "learning_rate": 1.521191501214122e-06, + "loss": 0.3967, + "step": 7223 + }, + { + "epoch": 2.311433447098976, + "grad_norm": 0.5228315591812134, + "learning_rate": 1.5198544660103525e-06, + "loss": 0.448, + "step": 7224 + }, + { + "epoch": 2.311753412969283, + "grad_norm": 0.48992231488227844, + "learning_rate": 1.5185179133329498e-06, + "loss": 0.4338, + "step": 7225 + }, + { + "epoch": 2.3120733788395906, + "grad_norm": 0.51529461145401, + "learning_rate": 1.5171818433672254e-06, + "loss": 0.427, + "step": 7226 + }, + { + "epoch": 2.3123933447098977, + "grad_norm": 0.5545088052749634, + "learning_rate": 1.5158462562984294e-06, + "loss": 0.475, + "step": 7227 + }, + { + "epoch": 2.3127133105802047, + "grad_norm": 0.49719321727752686, + "learning_rate": 1.5145111523117417e-06, + "loss": 0.3874, + "step": 7228 + }, + { + "epoch": 2.313033276450512, + "grad_norm": 0.5087726712226868, + "learning_rate": 1.513176531592274e-06, + "loss": 0.4186, + "step": 7229 + }, + { + "epoch": 2.313353242320819, + "grad_norm": 0.4915749430656433, + "learning_rate": 1.511842394325077e-06, + "loss": 0.4615, + "step": 7230 + }, + { + "epoch": 2.313673208191126, + "grad_norm": 0.48190948367118835, + "learning_rate": 1.5105087406951268e-06, + "loss": 0.4323, + "step": 7231 + }, + { + "epoch": 2.3139931740614337, + "grad_norm": 0.5035275816917419, + "learning_rate": 1.5091755708873396e-06, + "loss": 0.4938, + "step": 7232 + }, + { + "epoch": 2.3143131399317407, + "grad_norm": 0.4648093283176422, + "learning_rate": 1.5078428850865568e-06, + "loss": 0.438, + "step": 7233 + }, + { + "epoch": 2.3146331058020477, + "grad_norm": 0.5146045088768005, + "learning_rate": 1.5065106834775606e-06, + "loss": 0.441, + "step": 7234 + }, + { + "epoch": 2.3149530716723548, + "grad_norm": 0.5253714919090271, + "learning_rate": 1.5051789662450616e-06, + "loss": 0.4765, + "step": 7235 + }, + { + "epoch": 2.3152730375426622, + "grad_norm": 0.4963928461074829, + "learning_rate": 1.503847733573701e-06, + "loss": 0.4049, + "step": 7236 + }, + { + "epoch": 2.3155930034129693, + "grad_norm": 0.5003252029418945, + "learning_rate": 1.5025169856480604e-06, + "loss": 0.488, + "step": 7237 + }, + { + "epoch": 2.3159129692832763, + "grad_norm": 0.4468475878238678, + "learning_rate": 1.5011867226526455e-06, + "loss": 0.3802, + "step": 7238 + }, + { + "epoch": 2.3162329351535837, + "grad_norm": 0.4494239389896393, + "learning_rate": 1.4998569447719019e-06, + "loss": 0.4524, + "step": 7239 + }, + { + "epoch": 2.3165529010238908, + "grad_norm": 0.5028368234634399, + "learning_rate": 1.4985276521902043e-06, + "loss": 0.4187, + "step": 7240 + }, + { + "epoch": 2.316872866894198, + "grad_norm": 0.5029499530792236, + "learning_rate": 1.497198845091858e-06, + "loss": 0.4275, + "step": 7241 + }, + { + "epoch": 2.3171928327645053, + "grad_norm": 0.5458824634552002, + "learning_rate": 1.4958705236611076e-06, + "loss": 0.4246, + "step": 7242 + }, + { + "epoch": 2.3175127986348123, + "grad_norm": 0.472836971282959, + "learning_rate": 1.494542688082123e-06, + "loss": 0.4972, + "step": 7243 + }, + { + "epoch": 2.3178327645051193, + "grad_norm": 0.45238152146339417, + "learning_rate": 1.4932153385390135e-06, + "loss": 0.4245, + "step": 7244 + }, + { + "epoch": 2.318152730375427, + "grad_norm": 0.5061943531036377, + "learning_rate": 1.4918884752158158e-06, + "loss": 0.4201, + "step": 7245 + }, + { + "epoch": 2.318472696245734, + "grad_norm": 0.4750024974346161, + "learning_rate": 1.4905620982965002e-06, + "loss": 0.431, + "step": 7246 + }, + { + "epoch": 2.318792662116041, + "grad_norm": 0.5249196290969849, + "learning_rate": 1.489236207964973e-06, + "loss": 0.4492, + "step": 7247 + }, + { + "epoch": 2.319112627986348, + "grad_norm": 0.5278351306915283, + "learning_rate": 1.4879108044050677e-06, + "loss": 0.4652, + "step": 7248 + }, + { + "epoch": 2.3194325938566553, + "grad_norm": 0.4881724417209625, + "learning_rate": 1.4865858878005556e-06, + "loss": 0.4085, + "step": 7249 + }, + { + "epoch": 2.3197525597269624, + "grad_norm": 0.43812426924705505, + "learning_rate": 1.4852614583351376e-06, + "loss": 0.4386, + "step": 7250 + }, + { + "epoch": 2.3200725255972694, + "grad_norm": 0.5007078647613525, + "learning_rate": 1.4839375161924446e-06, + "loss": 0.471, + "step": 7251 + }, + { + "epoch": 2.320392491467577, + "grad_norm": 0.5291122794151306, + "learning_rate": 1.482614061556047e-06, + "loss": 0.4495, + "step": 7252 + }, + { + "epoch": 2.320712457337884, + "grad_norm": 0.5274209380149841, + "learning_rate": 1.4812910946094394e-06, + "loss": 0.4757, + "step": 7253 + }, + { + "epoch": 2.321032423208191, + "grad_norm": 0.4935450851917267, + "learning_rate": 1.479968615536056e-06, + "loss": 0.4118, + "step": 7254 + }, + { + "epoch": 2.3213523890784984, + "grad_norm": 0.514780580997467, + "learning_rate": 1.4786466245192594e-06, + "loss": 0.4844, + "step": 7255 + }, + { + "epoch": 2.3216723549488054, + "grad_norm": 0.4826414883136749, + "learning_rate": 1.4773251217423424e-06, + "loss": 0.4283, + "step": 7256 + }, + { + "epoch": 2.3219923208191124, + "grad_norm": 0.4890921115875244, + "learning_rate": 1.4760041073885367e-06, + "loss": 0.3864, + "step": 7257 + }, + { + "epoch": 2.32231228668942, + "grad_norm": 0.4811008870601654, + "learning_rate": 1.474683581641e-06, + "loss": 0.4213, + "step": 7258 + }, + { + "epoch": 2.322632252559727, + "grad_norm": 0.4668603837490082, + "learning_rate": 1.473363544682827e-06, + "loss": 0.4649, + "step": 7259 + }, + { + "epoch": 2.322952218430034, + "grad_norm": 0.4900484085083008, + "learning_rate": 1.4720439966970412e-06, + "loss": 0.4389, + "step": 7260 + }, + { + "epoch": 2.3232721843003414, + "grad_norm": 0.5131734609603882, + "learning_rate": 1.4707249378665978e-06, + "loss": 0.4487, + "step": 7261 + }, + { + "epoch": 2.3235921501706485, + "grad_norm": 0.4661282002925873, + "learning_rate": 1.4694063683743893e-06, + "loss": 0.3979, + "step": 7262 + }, + { + "epoch": 2.3239121160409555, + "grad_norm": 0.5389195680618286, + "learning_rate": 1.4680882884032333e-06, + "loss": 0.4789, + "step": 7263 + }, + { + "epoch": 2.324232081911263, + "grad_norm": 0.5405575037002563, + "learning_rate": 1.466770698135887e-06, + "loss": 0.5275, + "step": 7264 + }, + { + "epoch": 2.32455204778157, + "grad_norm": 0.5626694560050964, + "learning_rate": 1.4654535977550343e-06, + "loss": 0.47, + "step": 7265 + }, + { + "epoch": 2.324872013651877, + "grad_norm": 0.575118362903595, + "learning_rate": 1.4641369874432904e-06, + "loss": 0.4373, + "step": 7266 + }, + { + "epoch": 2.3251919795221845, + "grad_norm": 0.4923705756664276, + "learning_rate": 1.4628208673832072e-06, + "loss": 0.4485, + "step": 7267 + }, + { + "epoch": 2.3255119453924915, + "grad_norm": 0.4365101754665375, + "learning_rate": 1.4615052377572675e-06, + "loss": 0.418, + "step": 7268 + }, + { + "epoch": 2.3258319112627985, + "grad_norm": 0.4855596721172333, + "learning_rate": 1.4601900987478834e-06, + "loss": 0.4342, + "step": 7269 + }, + { + "epoch": 2.326151877133106, + "grad_norm": 0.5019798278808594, + "learning_rate": 1.458875450537401e-06, + "loss": 0.4214, + "step": 7270 + }, + { + "epoch": 2.326471843003413, + "grad_norm": 0.5033852458000183, + "learning_rate": 1.4575612933080952e-06, + "loss": 0.4651, + "step": 7271 + }, + { + "epoch": 2.32679180887372, + "grad_norm": 0.46395087242126465, + "learning_rate": 1.4562476272421784e-06, + "loss": 0.4352, + "step": 7272 + }, + { + "epoch": 2.3271117747440275, + "grad_norm": 0.46923959255218506, + "learning_rate": 1.4549344525217923e-06, + "loss": 0.4099, + "step": 7273 + }, + { + "epoch": 2.3274317406143346, + "grad_norm": 0.4986625909805298, + "learning_rate": 1.4536217693290094e-06, + "loss": 0.3984, + "step": 7274 + }, + { + "epoch": 2.3277517064846416, + "grad_norm": 0.549985408782959, + "learning_rate": 1.4523095778458352e-06, + "loss": 0.4783, + "step": 7275 + }, + { + "epoch": 2.328071672354949, + "grad_norm": 0.4645216763019562, + "learning_rate": 1.450997878254204e-06, + "loss": 0.4007, + "step": 7276 + }, + { + "epoch": 2.328391638225256, + "grad_norm": 0.49397537112236023, + "learning_rate": 1.4496866707359863e-06, + "loss": 0.4575, + "step": 7277 + }, + { + "epoch": 2.328711604095563, + "grad_norm": 0.5235719084739685, + "learning_rate": 1.4483759554729844e-06, + "loss": 0.4712, + "step": 7278 + }, + { + "epoch": 2.32903156996587, + "grad_norm": 0.4658564329147339, + "learning_rate": 1.447065732646929e-06, + "loss": 0.414, + "step": 7279 + }, + { + "epoch": 2.3293515358361776, + "grad_norm": 0.5144032835960388, + "learning_rate": 1.4457560024394846e-06, + "loss": 0.4855, + "step": 7280 + }, + { + "epoch": 2.3296715017064846, + "grad_norm": 0.491800993680954, + "learning_rate": 1.4444467650322442e-06, + "loss": 0.3843, + "step": 7281 + }, + { + "epoch": 2.3299914675767917, + "grad_norm": 0.5158237814903259, + "learning_rate": 1.4431380206067374e-06, + "loss": 0.4379, + "step": 7282 + }, + { + "epoch": 2.330311433447099, + "grad_norm": 0.5110154747962952, + "learning_rate": 1.4418297693444243e-06, + "loss": 0.4045, + "step": 7283 + }, + { + "epoch": 2.330631399317406, + "grad_norm": 0.49886924028396606, + "learning_rate": 1.4405220114266949e-06, + "loss": 0.4197, + "step": 7284 + }, + { + "epoch": 2.330951365187713, + "grad_norm": 0.48179560899734497, + "learning_rate": 1.4392147470348706e-06, + "loss": 0.4358, + "step": 7285 + }, + { + "epoch": 2.3312713310580206, + "grad_norm": 0.5632516741752625, + "learning_rate": 1.4379079763502036e-06, + "loss": 0.4751, + "step": 7286 + }, + { + "epoch": 2.3315912969283277, + "grad_norm": 0.5645520687103271, + "learning_rate": 1.4366016995538813e-06, + "loss": 0.4382, + "step": 7287 + }, + { + "epoch": 2.3319112627986347, + "grad_norm": 0.5333660840988159, + "learning_rate": 1.4352959168270214e-06, + "loss": 0.4607, + "step": 7288 + }, + { + "epoch": 2.332231228668942, + "grad_norm": 0.48292359709739685, + "learning_rate": 1.433990628350671e-06, + "loss": 0.4023, + "step": 7289 + }, + { + "epoch": 2.332551194539249, + "grad_norm": 0.5508813858032227, + "learning_rate": 1.4326858343058104e-06, + "loss": 0.5212, + "step": 7290 + }, + { + "epoch": 2.332871160409556, + "grad_norm": 0.4942038059234619, + "learning_rate": 1.4313815348733485e-06, + "loss": 0.3886, + "step": 7291 + }, + { + "epoch": 2.3331911262798632, + "grad_norm": 0.5608505606651306, + "learning_rate": 1.43007773023413e-06, + "loss": 0.4302, + "step": 7292 + }, + { + "epoch": 2.3335110921501707, + "grad_norm": 0.5095614194869995, + "learning_rate": 1.4287744205689308e-06, + "loss": 0.4134, + "step": 7293 + }, + { + "epoch": 2.3338310580204777, + "grad_norm": 0.535039484500885, + "learning_rate": 1.4274716060584542e-06, + "loss": 0.4951, + "step": 7294 + }, + { + "epoch": 2.3341510238907848, + "grad_norm": 0.5037428736686707, + "learning_rate": 1.4261692868833376e-06, + "loss": 0.4663, + "step": 7295 + }, + { + "epoch": 2.3344709897610922, + "grad_norm": 0.5374264717102051, + "learning_rate": 1.424867463224147e-06, + "loss": 0.4799, + "step": 7296 + }, + { + "epoch": 2.3347909556313993, + "grad_norm": 0.4959246814250946, + "learning_rate": 1.4235661352613832e-06, + "loss": 0.4277, + "step": 7297 + }, + { + "epoch": 2.3351109215017063, + "grad_norm": 0.5037208199501038, + "learning_rate": 1.4222653031754795e-06, + "loss": 0.4383, + "step": 7298 + }, + { + "epoch": 2.3354308873720138, + "grad_norm": 0.46232280135154724, + "learning_rate": 1.4209649671467956e-06, + "loss": 0.3771, + "step": 7299 + }, + { + "epoch": 2.335750853242321, + "grad_norm": 0.5119737386703491, + "learning_rate": 1.4196651273556244e-06, + "loss": 0.4818, + "step": 7300 + }, + { + "epoch": 2.336070819112628, + "grad_norm": 0.49208730459213257, + "learning_rate": 1.418365783982189e-06, + "loss": 0.4633, + "step": 7301 + }, + { + "epoch": 2.3363907849829353, + "grad_norm": 0.4549662172794342, + "learning_rate": 1.417066937206647e-06, + "loss": 0.4074, + "step": 7302 + }, + { + "epoch": 2.3367107508532423, + "grad_norm": 0.46344906091690063, + "learning_rate": 1.4157685872090854e-06, + "loss": 0.4241, + "step": 7303 + }, + { + "epoch": 2.3370307167235493, + "grad_norm": 0.5186562538146973, + "learning_rate": 1.4144707341695213e-06, + "loss": 0.4605, + "step": 7304 + }, + { + "epoch": 2.337350682593857, + "grad_norm": 0.47448334097862244, + "learning_rate": 1.4131733782679037e-06, + "loss": 0.3952, + "step": 7305 + }, + { + "epoch": 2.337670648464164, + "grad_norm": 0.568246066570282, + "learning_rate": 1.4118765196841106e-06, + "loss": 0.4662, + "step": 7306 + }, + { + "epoch": 2.337990614334471, + "grad_norm": 0.49980390071868896, + "learning_rate": 1.4105801585979546e-06, + "loss": 0.4151, + "step": 7307 + }, + { + "epoch": 2.3383105802047783, + "grad_norm": 0.49078819155693054, + "learning_rate": 1.4092842951891788e-06, + "loss": 0.4645, + "step": 7308 + }, + { + "epoch": 2.3386305460750854, + "grad_norm": 0.4779077172279358, + "learning_rate": 1.4079889296374561e-06, + "loss": 0.4436, + "step": 7309 + }, + { + "epoch": 2.3389505119453924, + "grad_norm": 0.5242723822593689, + "learning_rate": 1.406694062122389e-06, + "loss": 0.4812, + "step": 7310 + }, + { + "epoch": 2.3392704778157, + "grad_norm": 0.48143717646598816, + "learning_rate": 1.4053996928235113e-06, + "loss": 0.4092, + "step": 7311 + }, + { + "epoch": 2.339590443686007, + "grad_norm": 0.5340338349342346, + "learning_rate": 1.4041058219202914e-06, + "loss": 0.4572, + "step": 7312 + }, + { + "epoch": 2.339910409556314, + "grad_norm": 0.47235822677612305, + "learning_rate": 1.4028124495921263e-06, + "loss": 0.4072, + "step": 7313 + }, + { + "epoch": 2.3402303754266214, + "grad_norm": 0.5129916667938232, + "learning_rate": 1.4015195760183426e-06, + "loss": 0.4698, + "step": 7314 + }, + { + "epoch": 2.3405503412969284, + "grad_norm": 0.5445955991744995, + "learning_rate": 1.4002272013781988e-06, + "loss": 0.4632, + "step": 7315 + }, + { + "epoch": 2.3408703071672354, + "grad_norm": 0.5130749940872192, + "learning_rate": 1.3989353258508825e-06, + "loss": 0.4364, + "step": 7316 + }, + { + "epoch": 2.341190273037543, + "grad_norm": 0.48371532559394836, + "learning_rate": 1.3976439496155158e-06, + "loss": 0.4288, + "step": 7317 + }, + { + "epoch": 2.34151023890785, + "grad_norm": 0.5707089304924011, + "learning_rate": 1.396353072851151e-06, + "loss": 0.4909, + "step": 7318 + }, + { + "epoch": 2.341830204778157, + "grad_norm": 0.476028710603714, + "learning_rate": 1.3950626957367673e-06, + "loss": 0.389, + "step": 7319 + }, + { + "epoch": 2.342150170648464, + "grad_norm": 0.512357771396637, + "learning_rate": 1.3937728184512777e-06, + "loss": 0.4261, + "step": 7320 + }, + { + "epoch": 2.3424701365187715, + "grad_norm": 0.586431622505188, + "learning_rate": 1.3924834411735238e-06, + "loss": 0.5006, + "step": 7321 + }, + { + "epoch": 2.3427901023890785, + "grad_norm": 0.5070300698280334, + "learning_rate": 1.391194564082281e-06, + "loss": 0.4903, + "step": 7322 + }, + { + "epoch": 2.3431100682593855, + "grad_norm": 0.486686110496521, + "learning_rate": 1.3899061873562548e-06, + "loss": 0.3947, + "step": 7323 + }, + { + "epoch": 2.343430034129693, + "grad_norm": 0.4885390102863312, + "learning_rate": 1.3886183111740786e-06, + "loss": 0.4241, + "step": 7324 + }, + { + "epoch": 2.34375, + "grad_norm": 0.5397626161575317, + "learning_rate": 1.3873309357143183e-06, + "loss": 0.4963, + "step": 7325 + }, + { + "epoch": 2.344069965870307, + "grad_norm": 0.562702476978302, + "learning_rate": 1.3860440611554682e-06, + "loss": 0.4929, + "step": 7326 + }, + { + "epoch": 2.3443899317406145, + "grad_norm": 0.5303139090538025, + "learning_rate": 1.3847576876759573e-06, + "loss": 0.4108, + "step": 7327 + }, + { + "epoch": 2.3447098976109215, + "grad_norm": 0.5334408283233643, + "learning_rate": 1.3834718154541432e-06, + "loss": 0.4835, + "step": 7328 + }, + { + "epoch": 2.3450298634812285, + "grad_norm": 0.440531849861145, + "learning_rate": 1.3821864446683126e-06, + "loss": 0.4238, + "step": 7329 + }, + { + "epoch": 2.345349829351536, + "grad_norm": 0.5001742243766785, + "learning_rate": 1.3809015754966843e-06, + "loss": 0.4815, + "step": 7330 + }, + { + "epoch": 2.345669795221843, + "grad_norm": 0.4859180748462677, + "learning_rate": 1.3796172081174041e-06, + "loss": 0.4336, + "step": 7331 + }, + { + "epoch": 2.34598976109215, + "grad_norm": 0.5429733395576477, + "learning_rate": 1.3783333427085543e-06, + "loss": 0.4077, + "step": 7332 + }, + { + "epoch": 2.346309726962457, + "grad_norm": 0.510171115398407, + "learning_rate": 1.3770499794481446e-06, + "loss": 0.4306, + "step": 7333 + }, + { + "epoch": 2.3466296928327646, + "grad_norm": 0.48114025592803955, + "learning_rate": 1.3757671185141136e-06, + "loss": 0.4434, + "step": 7334 + }, + { + "epoch": 2.3469496587030716, + "grad_norm": 0.5100584030151367, + "learning_rate": 1.374484760084332e-06, + "loss": 0.4625, + "step": 7335 + }, + { + "epoch": 2.3472696245733786, + "grad_norm": 0.4963550269603729, + "learning_rate": 1.373202904336597e-06, + "loss": 0.4786, + "step": 7336 + }, + { + "epoch": 2.347589590443686, + "grad_norm": 0.5266077518463135, + "learning_rate": 1.3719215514486455e-06, + "loss": 0.4208, + "step": 7337 + }, + { + "epoch": 2.347909556313993, + "grad_norm": 0.5163435935974121, + "learning_rate": 1.3706407015981355e-06, + "loss": 0.4432, + "step": 7338 + }, + { + "epoch": 2.3482295221843, + "grad_norm": 0.5281748175621033, + "learning_rate": 1.3693603549626584e-06, + "loss": 0.4732, + "step": 7339 + }, + { + "epoch": 2.3485494880546076, + "grad_norm": 0.5066831707954407, + "learning_rate": 1.3680805117197343e-06, + "loss": 0.4656, + "step": 7340 + }, + { + "epoch": 2.3488694539249146, + "grad_norm": 0.47515058517456055, + "learning_rate": 1.3668011720468166e-06, + "loss": 0.3988, + "step": 7341 + }, + { + "epoch": 2.3491894197952217, + "grad_norm": 0.535394012928009, + "learning_rate": 1.3655223361212884e-06, + "loss": 0.453, + "step": 7342 + }, + { + "epoch": 2.349509385665529, + "grad_norm": 0.48910632729530334, + "learning_rate": 1.3642440041204614e-06, + "loss": 0.4184, + "step": 7343 + }, + { + "epoch": 2.349829351535836, + "grad_norm": 0.4723266065120697, + "learning_rate": 1.3629661762215762e-06, + "loss": 0.4477, + "step": 7344 + }, + { + "epoch": 2.350149317406143, + "grad_norm": 0.48808911442756653, + "learning_rate": 1.3616888526018057e-06, + "loss": 0.5345, + "step": 7345 + }, + { + "epoch": 2.3504692832764507, + "grad_norm": 0.4407173693180084, + "learning_rate": 1.360412033438252e-06, + "loss": 0.3538, + "step": 7346 + }, + { + "epoch": 2.3507892491467577, + "grad_norm": 0.4907718300819397, + "learning_rate": 1.35913571890795e-06, + "loss": 0.4158, + "step": 7347 + }, + { + "epoch": 2.3511092150170647, + "grad_norm": 0.5179418325424194, + "learning_rate": 1.3578599091878608e-06, + "loss": 0.4353, + "step": 7348 + }, + { + "epoch": 2.351429180887372, + "grad_norm": 0.4885033071041107, + "learning_rate": 1.356584604454877e-06, + "loss": 0.4483, + "step": 7349 + }, + { + "epoch": 2.351749146757679, + "grad_norm": 0.49781283736228943, + "learning_rate": 1.3553098048858194e-06, + "loss": 0.4388, + "step": 7350 + }, + { + "epoch": 2.3520691126279862, + "grad_norm": 0.5235708355903625, + "learning_rate": 1.3540355106574415e-06, + "loss": 0.4452, + "step": 7351 + }, + { + "epoch": 2.3523890784982937, + "grad_norm": 0.4559749364852905, + "learning_rate": 1.3527617219464284e-06, + "loss": 0.4062, + "step": 7352 + }, + { + "epoch": 2.3527090443686007, + "grad_norm": 0.5198562741279602, + "learning_rate": 1.3514884389293898e-06, + "loss": 0.4991, + "step": 7353 + }, + { + "epoch": 2.3530290102389078, + "grad_norm": 0.5171627402305603, + "learning_rate": 1.3502156617828687e-06, + "loss": 0.4422, + "step": 7354 + }, + { + "epoch": 2.3533489761092152, + "grad_norm": 0.48058363795280457, + "learning_rate": 1.3489433906833355e-06, + "loss": 0.3699, + "step": 7355 + }, + { + "epoch": 2.3536689419795223, + "grad_norm": 0.4736981689929962, + "learning_rate": 1.347671625807193e-06, + "loss": 0.4322, + "step": 7356 + }, + { + "epoch": 2.3539889078498293, + "grad_norm": 0.5326634645462036, + "learning_rate": 1.3464003673307751e-06, + "loss": 0.4543, + "step": 7357 + }, + { + "epoch": 2.3543088737201368, + "grad_norm": 0.48081734776496887, + "learning_rate": 1.3451296154303423e-06, + "loss": 0.4068, + "step": 7358 + }, + { + "epoch": 2.354628839590444, + "grad_norm": 0.5038999319076538, + "learning_rate": 1.3438593702820846e-06, + "loss": 0.427, + "step": 7359 + }, + { + "epoch": 2.354948805460751, + "grad_norm": 0.4914281368255615, + "learning_rate": 1.3425896320621224e-06, + "loss": 0.4472, + "step": 7360 + }, + { + "epoch": 2.355268771331058, + "grad_norm": 0.4974600076675415, + "learning_rate": 1.3413204009465075e-06, + "loss": 0.4807, + "step": 7361 + }, + { + "epoch": 2.3555887372013653, + "grad_norm": 0.46198078989982605, + "learning_rate": 1.340051677111222e-06, + "loss": 0.3904, + "step": 7362 + }, + { + "epoch": 2.3559087030716723, + "grad_norm": 0.5218572616577148, + "learning_rate": 1.3387834607321741e-06, + "loss": 0.479, + "step": 7363 + }, + { + "epoch": 2.3562286689419794, + "grad_norm": 0.5179147124290466, + "learning_rate": 1.337515751985205e-06, + "loss": 0.4263, + "step": 7364 + }, + { + "epoch": 2.356548634812287, + "grad_norm": 0.5002532005310059, + "learning_rate": 1.3362485510460803e-06, + "loss": 0.4497, + "step": 7365 + }, + { + "epoch": 2.356868600682594, + "grad_norm": 0.5245099663734436, + "learning_rate": 1.3349818580905027e-06, + "loss": 0.4351, + "step": 7366 + }, + { + "epoch": 2.357188566552901, + "grad_norm": 0.5275083780288696, + "learning_rate": 1.3337156732941002e-06, + "loss": 0.4113, + "step": 7367 + }, + { + "epoch": 2.3575085324232083, + "grad_norm": 0.5387864708900452, + "learning_rate": 1.3324499968324306e-06, + "loss": 0.4592, + "step": 7368 + }, + { + "epoch": 2.3578284982935154, + "grad_norm": 0.50733482837677, + "learning_rate": 1.3311848288809815e-06, + "loss": 0.434, + "step": 7369 + }, + { + "epoch": 2.3581484641638224, + "grad_norm": 0.4945586919784546, + "learning_rate": 1.3299201696151676e-06, + "loss": 0.4244, + "step": 7370 + }, + { + "epoch": 2.35846843003413, + "grad_norm": 0.5445435643196106, + "learning_rate": 1.3286560192103377e-06, + "loss": 0.4904, + "step": 7371 + }, + { + "epoch": 2.358788395904437, + "grad_norm": 0.46514472365379333, + "learning_rate": 1.3273923778417686e-06, + "loss": 0.4432, + "step": 7372 + }, + { + "epoch": 2.359108361774744, + "grad_norm": 0.4684540629386902, + "learning_rate": 1.3261292456846648e-06, + "loss": 0.409, + "step": 7373 + }, + { + "epoch": 2.359428327645051, + "grad_norm": 0.5185834169387817, + "learning_rate": 1.3248666229141589e-06, + "loss": 0.465, + "step": 7374 + }, + { + "epoch": 2.3597482935153584, + "grad_norm": 0.5179154872894287, + "learning_rate": 1.3236045097053191e-06, + "loss": 0.4905, + "step": 7375 + }, + { + "epoch": 2.3600682593856654, + "grad_norm": 0.4815397560596466, + "learning_rate": 1.322342906233135e-06, + "loss": 0.4296, + "step": 7376 + }, + { + "epoch": 2.3603882252559725, + "grad_norm": 0.4623855650424957, + "learning_rate": 1.3210818126725328e-06, + "loss": 0.4024, + "step": 7377 + }, + { + "epoch": 2.36070819112628, + "grad_norm": 0.5654313564300537, + "learning_rate": 1.3198212291983636e-06, + "loss": 0.5295, + "step": 7378 + }, + { + "epoch": 2.361028156996587, + "grad_norm": 0.501433253288269, + "learning_rate": 1.3185611559854061e-06, + "loss": 0.3838, + "step": 7379 + }, + { + "epoch": 2.361348122866894, + "grad_norm": 0.48429253697395325, + "learning_rate": 1.3173015932083754e-06, + "loss": 0.3983, + "step": 7380 + }, + { + "epoch": 2.3616680887372015, + "grad_norm": 0.5181745886802673, + "learning_rate": 1.316042541041907e-06, + "loss": 0.4655, + "step": 7381 + }, + { + "epoch": 2.3619880546075085, + "grad_norm": 0.5786999464035034, + "learning_rate": 1.3147839996605743e-06, + "loss": 0.4845, + "step": 7382 + }, + { + "epoch": 2.3623080204778155, + "grad_norm": 0.49806299805641174, + "learning_rate": 1.313525969238873e-06, + "loss": 0.3984, + "step": 7383 + }, + { + "epoch": 2.362627986348123, + "grad_norm": 0.5032016634941101, + "learning_rate": 1.31226844995123e-06, + "loss": 0.4439, + "step": 7384 + }, + { + "epoch": 2.36294795221843, + "grad_norm": 0.48412343859672546, + "learning_rate": 1.3110114419720043e-06, + "loss": 0.3864, + "step": 7385 + }, + { + "epoch": 2.363267918088737, + "grad_norm": 0.5202109217643738, + "learning_rate": 1.3097549454754782e-06, + "loss": 0.4555, + "step": 7386 + }, + { + "epoch": 2.3635878839590445, + "grad_norm": 0.5044285655021667, + "learning_rate": 1.30849896063587e-06, + "loss": 0.4295, + "step": 7387 + }, + { + "epoch": 2.3639078498293515, + "grad_norm": 0.5070471167564392, + "learning_rate": 1.3072434876273222e-06, + "loss": 0.4818, + "step": 7388 + }, + { + "epoch": 2.3642278156996586, + "grad_norm": 0.5185414552688599, + "learning_rate": 1.3059885266239058e-06, + "loss": 0.4449, + "step": 7389 + }, + { + "epoch": 2.364547781569966, + "grad_norm": 0.49383845925331116, + "learning_rate": 1.3047340777996254e-06, + "loss": 0.477, + "step": 7390 + }, + { + "epoch": 2.364867747440273, + "grad_norm": 0.49830183386802673, + "learning_rate": 1.3034801413284092e-06, + "loss": 0.4316, + "step": 7391 + }, + { + "epoch": 2.36518771331058, + "grad_norm": 0.479976087808609, + "learning_rate": 1.3022267173841202e-06, + "loss": 0.4081, + "step": 7392 + }, + { + "epoch": 2.3655076791808876, + "grad_norm": 0.48571282625198364, + "learning_rate": 1.3009738061405452e-06, + "loss": 0.4148, + "step": 7393 + }, + { + "epoch": 2.3658276450511946, + "grad_norm": 0.5023415684700012, + "learning_rate": 1.2997214077714004e-06, + "loss": 0.4553, + "step": 7394 + }, + { + "epoch": 2.3661476109215016, + "grad_norm": 0.4909842908382416, + "learning_rate": 1.2984695224503351e-06, + "loss": 0.433, + "step": 7395 + }, + { + "epoch": 2.366467576791809, + "grad_norm": 0.4887163043022156, + "learning_rate": 1.2972181503509228e-06, + "loss": 0.4577, + "step": 7396 + }, + { + "epoch": 2.366787542662116, + "grad_norm": 0.46021318435668945, + "learning_rate": 1.2959672916466698e-06, + "loss": 0.4268, + "step": 7397 + }, + { + "epoch": 2.367107508532423, + "grad_norm": 0.4698842167854309, + "learning_rate": 1.2947169465110077e-06, + "loss": 0.3931, + "step": 7398 + }, + { + "epoch": 2.3674274744027306, + "grad_norm": 0.48359882831573486, + "learning_rate": 1.2934671151172974e-06, + "loss": 0.4401, + "step": 7399 + }, + { + "epoch": 2.3677474402730376, + "grad_norm": 0.49636223912239075, + "learning_rate": 1.2922177976388318e-06, + "loss": 0.4625, + "step": 7400 + }, + { + "epoch": 2.3680674061433447, + "grad_norm": 0.5095014572143555, + "learning_rate": 1.2909689942488273e-06, + "loss": 0.4227, + "step": 7401 + }, + { + "epoch": 2.368387372013652, + "grad_norm": 0.5275278091430664, + "learning_rate": 1.2897207051204358e-06, + "loss": 0.4322, + "step": 7402 + }, + { + "epoch": 2.368707337883959, + "grad_norm": 0.4835013449192047, + "learning_rate": 1.288472930426732e-06, + "loss": 0.4291, + "step": 7403 + }, + { + "epoch": 2.369027303754266, + "grad_norm": 0.5402299165725708, + "learning_rate": 1.28722567034072e-06, + "loss": 0.4736, + "step": 7404 + }, + { + "epoch": 2.369347269624573, + "grad_norm": 0.45481786131858826, + "learning_rate": 1.2859789250353367e-06, + "loss": 0.411, + "step": 7405 + }, + { + "epoch": 2.3696672354948807, + "grad_norm": 0.470196008682251, + "learning_rate": 1.2847326946834427e-06, + "loss": 0.393, + "step": 7406 + }, + { + "epoch": 2.3699872013651877, + "grad_norm": 0.5229184031486511, + "learning_rate": 1.2834869794578308e-06, + "loss": 0.5201, + "step": 7407 + }, + { + "epoch": 2.3703071672354947, + "grad_norm": 0.5163376927375793, + "learning_rate": 1.2822417795312193e-06, + "loss": 0.4365, + "step": 7408 + }, + { + "epoch": 2.370627133105802, + "grad_norm": 0.4873303472995758, + "learning_rate": 1.280997095076259e-06, + "loss": 0.4496, + "step": 7409 + }, + { + "epoch": 2.3709470989761092, + "grad_norm": 0.505264401435852, + "learning_rate": 1.279752926265525e-06, + "loss": 0.4387, + "step": 7410 + }, + { + "epoch": 2.3712670648464163, + "grad_norm": 0.49931851029396057, + "learning_rate": 1.2785092732715225e-06, + "loss": 0.4161, + "step": 7411 + }, + { + "epoch": 2.3715870307167237, + "grad_norm": 0.4455562233924866, + "learning_rate": 1.2772661362666877e-06, + "loss": 0.4451, + "step": 7412 + }, + { + "epoch": 2.3719069965870307, + "grad_norm": 0.5132851600646973, + "learning_rate": 1.2760235154233801e-06, + "loss": 0.4357, + "step": 7413 + }, + { + "epoch": 2.3722269624573378, + "grad_norm": 0.5052698254585266, + "learning_rate": 1.2747814109138939e-06, + "loss": 0.4303, + "step": 7414 + }, + { + "epoch": 2.3725469283276452, + "grad_norm": 0.5079882144927979, + "learning_rate": 1.2735398229104463e-06, + "loss": 0.4499, + "step": 7415 + }, + { + "epoch": 2.3728668941979523, + "grad_norm": 0.5103276371955872, + "learning_rate": 1.2722987515851836e-06, + "loss": 0.4663, + "step": 7416 + }, + { + "epoch": 2.3731868600682593, + "grad_norm": 0.5091108679771423, + "learning_rate": 1.2710581971101855e-06, + "loss": 0.3743, + "step": 7417 + }, + { + "epoch": 2.3735068259385663, + "grad_norm": 0.5644721984863281, + "learning_rate": 1.2698181596574528e-06, + "loss": 0.4721, + "step": 7418 + }, + { + "epoch": 2.373826791808874, + "grad_norm": 0.5130641460418701, + "learning_rate": 1.2685786393989213e-06, + "loss": 0.443, + "step": 7419 + }, + { + "epoch": 2.374146757679181, + "grad_norm": 0.4706560969352722, + "learning_rate": 1.2673396365064504e-06, + "loss": 0.3924, + "step": 7420 + }, + { + "epoch": 2.374466723549488, + "grad_norm": 0.5112258195877075, + "learning_rate": 1.2661011511518278e-06, + "loss": 0.4407, + "step": 7421 + }, + { + "epoch": 2.3747866894197953, + "grad_norm": 0.5168745517730713, + "learning_rate": 1.2648631835067742e-06, + "loss": 0.4241, + "step": 7422 + }, + { + "epoch": 2.3751066552901023, + "grad_norm": 0.49925437569618225, + "learning_rate": 1.2636257337429319e-06, + "loss": 0.4137, + "step": 7423 + }, + { + "epoch": 2.3754266211604094, + "grad_norm": 0.5110605359077454, + "learning_rate": 1.2623888020318787e-06, + "loss": 0.4459, + "step": 7424 + }, + { + "epoch": 2.375746587030717, + "grad_norm": 0.5304185152053833, + "learning_rate": 1.2611523885451137e-06, + "loss": 0.4655, + "step": 7425 + }, + { + "epoch": 2.376066552901024, + "grad_norm": 0.5524001717567444, + "learning_rate": 1.259916493454067e-06, + "loss": 0.4396, + "step": 7426 + }, + { + "epoch": 2.376386518771331, + "grad_norm": 0.5668018460273743, + "learning_rate": 1.2586811169300994e-06, + "loss": 0.4552, + "step": 7427 + }, + { + "epoch": 2.3767064846416384, + "grad_norm": 0.4877117872238159, + "learning_rate": 1.257446259144494e-06, + "loss": 0.4348, + "step": 7428 + }, + { + "epoch": 2.3770264505119454, + "grad_norm": 0.47735869884490967, + "learning_rate": 1.2562119202684692e-06, + "loss": 0.4409, + "step": 7429 + }, + { + "epoch": 2.3773464163822524, + "grad_norm": 0.5047822594642639, + "learning_rate": 1.2549781004731653e-06, + "loss": 0.434, + "step": 7430 + }, + { + "epoch": 2.37766638225256, + "grad_norm": 0.4749005436897278, + "learning_rate": 1.253744799929652e-06, + "loss": 0.3931, + "step": 7431 + }, + { + "epoch": 2.377986348122867, + "grad_norm": 0.5150494575500488, + "learning_rate": 1.25251201880893e-06, + "loss": 0.4553, + "step": 7432 + }, + { + "epoch": 2.378306313993174, + "grad_norm": 0.4489477276802063, + "learning_rate": 1.2512797572819236e-06, + "loss": 0.4603, + "step": 7433 + }, + { + "epoch": 2.3786262798634814, + "grad_norm": 0.5023684501647949, + "learning_rate": 1.2500480155194899e-06, + "loss": 0.4793, + "step": 7434 + }, + { + "epoch": 2.3789462457337884, + "grad_norm": 0.4785566031932831, + "learning_rate": 1.2488167936924105e-06, + "loss": 0.3931, + "step": 7435 + }, + { + "epoch": 2.3792662116040955, + "grad_norm": 0.5234454870223999, + "learning_rate": 1.2475860919713939e-06, + "loss": 0.4481, + "step": 7436 + }, + { + "epoch": 2.379586177474403, + "grad_norm": 0.5080173015594482, + "learning_rate": 1.2463559105270817e-06, + "loss": 0.4192, + "step": 7437 + }, + { + "epoch": 2.37990614334471, + "grad_norm": 0.5243739485740662, + "learning_rate": 1.2451262495300366e-06, + "loss": 0.4371, + "step": 7438 + }, + { + "epoch": 2.380226109215017, + "grad_norm": 0.5148787498474121, + "learning_rate": 1.2438971091507552e-06, + "loss": 0.4154, + "step": 7439 + }, + { + "epoch": 2.3805460750853245, + "grad_norm": 0.4937624931335449, + "learning_rate": 1.2426684895596591e-06, + "loss": 0.4816, + "step": 7440 + }, + { + "epoch": 2.3808660409556315, + "grad_norm": 0.4702324867248535, + "learning_rate": 1.241440390927095e-06, + "loss": 0.3737, + "step": 7441 + }, + { + "epoch": 2.3811860068259385, + "grad_norm": 0.5017414093017578, + "learning_rate": 1.2402128134233437e-06, + "loss": 0.4706, + "step": 7442 + }, + { + "epoch": 2.381505972696246, + "grad_norm": 0.49768519401550293, + "learning_rate": 1.2389857572186082e-06, + "loss": 0.425, + "step": 7443 + }, + { + "epoch": 2.381825938566553, + "grad_norm": 0.47565585374832153, + "learning_rate": 1.2377592224830232e-06, + "loss": 0.4117, + "step": 7444 + }, + { + "epoch": 2.38214590443686, + "grad_norm": 0.46606653928756714, + "learning_rate": 1.2365332093866477e-06, + "loss": 0.4822, + "step": 7445 + }, + { + "epoch": 2.382465870307167, + "grad_norm": 0.46190446615219116, + "learning_rate": 1.2353077180994693e-06, + "loss": 0.4384, + "step": 7446 + }, + { + "epoch": 2.3827858361774745, + "grad_norm": 0.5038273930549622, + "learning_rate": 1.2340827487914048e-06, + "loss": 0.4312, + "step": 7447 + }, + { + "epoch": 2.3831058020477816, + "grad_norm": 0.5104163289070129, + "learning_rate": 1.2328583016322992e-06, + "loss": 0.4671, + "step": 7448 + }, + { + "epoch": 2.3834257679180886, + "grad_norm": 0.5152691006660461, + "learning_rate": 1.231634376791922e-06, + "loss": 0.4744, + "step": 7449 + }, + { + "epoch": 2.383745733788396, + "grad_norm": 0.5327239632606506, + "learning_rate": 1.2304109744399717e-06, + "loss": 0.436, + "step": 7450 + }, + { + "epoch": 2.384065699658703, + "grad_norm": 0.5346876978874207, + "learning_rate": 1.2291880947460732e-06, + "loss": 0.4468, + "step": 7451 + }, + { + "epoch": 2.38438566552901, + "grad_norm": 0.4823782444000244, + "learning_rate": 1.227965737879782e-06, + "loss": 0.414, + "step": 7452 + }, + { + "epoch": 2.3847056313993176, + "grad_norm": 0.4793708920478821, + "learning_rate": 1.2267439040105806e-06, + "loss": 0.4748, + "step": 7453 + }, + { + "epoch": 2.3850255972696246, + "grad_norm": 0.5064877867698669, + "learning_rate": 1.2255225933078762e-06, + "loss": 0.4756, + "step": 7454 + }, + { + "epoch": 2.3853455631399316, + "grad_norm": 0.44026434421539307, + "learning_rate": 1.224301805941005e-06, + "loss": 0.3681, + "step": 7455 + }, + { + "epoch": 2.385665529010239, + "grad_norm": 0.519895613193512, + "learning_rate": 1.223081542079229e-06, + "loss": 0.4246, + "step": 7456 + }, + { + "epoch": 2.385985494880546, + "grad_norm": 0.554022490978241, + "learning_rate": 1.2218618018917412e-06, + "loss": 0.48, + "step": 7457 + }, + { + "epoch": 2.386305460750853, + "grad_norm": 0.5087018013000488, + "learning_rate": 1.2206425855476612e-06, + "loss": 0.4163, + "step": 7458 + }, + { + "epoch": 2.38662542662116, + "grad_norm": 0.4920301139354706, + "learning_rate": 1.2194238932160335e-06, + "loss": 0.4094, + "step": 7459 + }, + { + "epoch": 2.3869453924914676, + "grad_norm": 0.5391239523887634, + "learning_rate": 1.2182057250658307e-06, + "loss": 0.4407, + "step": 7460 + }, + { + "epoch": 2.3872653583617747, + "grad_norm": 0.5132863521575928, + "learning_rate": 1.216988081265953e-06, + "loss": 0.4178, + "step": 7461 + }, + { + "epoch": 2.3875853242320817, + "grad_norm": 0.5124856233596802, + "learning_rate": 1.2157709619852282e-06, + "loss": 0.4094, + "step": 7462 + }, + { + "epoch": 2.387905290102389, + "grad_norm": 0.5079914331436157, + "learning_rate": 1.2145543673924142e-06, + "loss": 0.4194, + "step": 7463 + }, + { + "epoch": 2.388225255972696, + "grad_norm": 0.4989244043827057, + "learning_rate": 1.213338297656191e-06, + "loss": 0.4285, + "step": 7464 + }, + { + "epoch": 2.388545221843003, + "grad_norm": 0.46511325240135193, + "learning_rate": 1.2121227529451678e-06, + "loss": 0.4163, + "step": 7465 + }, + { + "epoch": 2.3888651877133107, + "grad_norm": 0.5314695239067078, + "learning_rate": 1.2109077334278807e-06, + "loss": 0.4577, + "step": 7466 + }, + { + "epoch": 2.3891851535836177, + "grad_norm": 0.4955805838108063, + "learning_rate": 1.2096932392727946e-06, + "loss": 0.3961, + "step": 7467 + }, + { + "epoch": 2.3895051194539247, + "grad_norm": 0.5038869380950928, + "learning_rate": 1.2084792706483022e-06, + "loss": 0.5004, + "step": 7468 + }, + { + "epoch": 2.389825085324232, + "grad_norm": 0.5208796858787537, + "learning_rate": 1.2072658277227206e-06, + "loss": 0.4661, + "step": 7469 + }, + { + "epoch": 2.3901450511945392, + "grad_norm": 0.5192559361457825, + "learning_rate": 1.2060529106642943e-06, + "loss": 0.4651, + "step": 7470 + }, + { + "epoch": 2.3904650170648463, + "grad_norm": 0.482944130897522, + "learning_rate": 1.204840519641195e-06, + "loss": 0.4322, + "step": 7471 + }, + { + "epoch": 2.3907849829351537, + "grad_norm": 0.4873718321323395, + "learning_rate": 1.203628654821523e-06, + "loss": 0.4704, + "step": 7472 + }, + { + "epoch": 2.3911049488054608, + "grad_norm": 0.4696694016456604, + "learning_rate": 1.2024173163733072e-06, + "loss": 0.3924, + "step": 7473 + }, + { + "epoch": 2.391424914675768, + "grad_norm": 0.5030756592750549, + "learning_rate": 1.2012065044644988e-06, + "loss": 0.4325, + "step": 7474 + }, + { + "epoch": 2.3917448805460753, + "grad_norm": 0.5451432466506958, + "learning_rate": 1.1999962192629782e-06, + "loss": 0.4564, + "step": 7475 + }, + { + "epoch": 2.3920648464163823, + "grad_norm": 0.5294658541679382, + "learning_rate": 1.1987864609365523e-06, + "loss": 0.4487, + "step": 7476 + }, + { + "epoch": 2.3923848122866893, + "grad_norm": 0.4775328040122986, + "learning_rate": 1.1975772296529564e-06, + "loss": 0.4161, + "step": 7477 + }, + { + "epoch": 2.392704778156997, + "grad_norm": 0.4947337210178375, + "learning_rate": 1.1963685255798535e-06, + "loss": 0.3921, + "step": 7478 + }, + { + "epoch": 2.393024744027304, + "grad_norm": 0.5113786458969116, + "learning_rate": 1.1951603488848307e-06, + "loss": 0.4679, + "step": 7479 + }, + { + "epoch": 2.393344709897611, + "grad_norm": 0.4797898828983307, + "learning_rate": 1.1939526997354029e-06, + "loss": 0.4468, + "step": 7480 + }, + { + "epoch": 2.3936646757679183, + "grad_norm": 0.49606427550315857, + "learning_rate": 1.1927455782990088e-06, + "loss": 0.397, + "step": 7481 + }, + { + "epoch": 2.3939846416382253, + "grad_norm": 0.5278741717338562, + "learning_rate": 1.1915389847430231e-06, + "loss": 0.4815, + "step": 7482 + }, + { + "epoch": 2.3943046075085324, + "grad_norm": 0.505317747592926, + "learning_rate": 1.1903329192347397e-06, + "loss": 0.431, + "step": 7483 + }, + { + "epoch": 2.39462457337884, + "grad_norm": 0.5160964131355286, + "learning_rate": 1.1891273819413796e-06, + "loss": 0.4359, + "step": 7484 + }, + { + "epoch": 2.394944539249147, + "grad_norm": 0.4927464723587036, + "learning_rate": 1.1879223730300926e-06, + "loss": 0.4245, + "step": 7485 + }, + { + "epoch": 2.395264505119454, + "grad_norm": 0.5302851796150208, + "learning_rate": 1.1867178926679519e-06, + "loss": 0.3868, + "step": 7486 + }, + { + "epoch": 2.395584470989761, + "grad_norm": 0.5058532953262329, + "learning_rate": 1.1855139410219657e-06, + "loss": 0.4229, + "step": 7487 + }, + { + "epoch": 2.3959044368600684, + "grad_norm": 0.5502296686172485, + "learning_rate": 1.1843105182590609e-06, + "loss": 0.4453, + "step": 7488 + }, + { + "epoch": 2.3962244027303754, + "grad_norm": 0.5319601893424988, + "learning_rate": 1.1831076245460926e-06, + "loss": 0.4374, + "step": 7489 + }, + { + "epoch": 2.3965443686006824, + "grad_norm": 0.5216536521911621, + "learning_rate": 1.1819052600498444e-06, + "loss": 0.4219, + "step": 7490 + }, + { + "epoch": 2.39686433447099, + "grad_norm": 0.5190451741218567, + "learning_rate": 1.1807034249370218e-06, + "loss": 0.4338, + "step": 7491 + }, + { + "epoch": 2.397184300341297, + "grad_norm": 0.4999808073043823, + "learning_rate": 1.1795021193742673e-06, + "loss": 0.4069, + "step": 7492 + }, + { + "epoch": 2.397504266211604, + "grad_norm": 0.47495537996292114, + "learning_rate": 1.1783013435281398e-06, + "loss": 0.426, + "step": 7493 + }, + { + "epoch": 2.3978242320819114, + "grad_norm": 0.5205498933792114, + "learning_rate": 1.1771010975651287e-06, + "loss": 0.4721, + "step": 7494 + }, + { + "epoch": 2.3981441979522184, + "grad_norm": 0.5490537881851196, + "learning_rate": 1.1759013816516486e-06, + "loss": 0.4514, + "step": 7495 + }, + { + "epoch": 2.3984641638225255, + "grad_norm": 0.5139085054397583, + "learning_rate": 1.1747021959540418e-06, + "loss": 0.4604, + "step": 7496 + }, + { + "epoch": 2.398784129692833, + "grad_norm": 0.5006527304649353, + "learning_rate": 1.1735035406385792e-06, + "loss": 0.4862, + "step": 7497 + }, + { + "epoch": 2.39910409556314, + "grad_norm": 0.5009004473686218, + "learning_rate": 1.1723054158714542e-06, + "loss": 0.4324, + "step": 7498 + }, + { + "epoch": 2.399424061433447, + "grad_norm": 0.48826852440834045, + "learning_rate": 1.1711078218187878e-06, + "loss": 0.4846, + "step": 7499 + }, + { + "epoch": 2.399744027303754, + "grad_norm": 0.4620589315891266, + "learning_rate": 1.169910758646627e-06, + "loss": 0.4054, + "step": 7500 + }, + { + "epoch": 2.4000639931740615, + "grad_norm": 0.5198194980621338, + "learning_rate": 1.1687142265209471e-06, + "loss": 0.4111, + "step": 7501 + }, + { + "epoch": 2.4003839590443685, + "grad_norm": 0.5378292202949524, + "learning_rate": 1.16751822560765e-06, + "loss": 0.4924, + "step": 7502 + }, + { + "epoch": 2.4007039249146755, + "grad_norm": 0.4919576644897461, + "learning_rate": 1.166322756072562e-06, + "loss": 0.4569, + "step": 7503 + }, + { + "epoch": 2.401023890784983, + "grad_norm": 0.4672337472438812, + "learning_rate": 1.1651278180814356e-06, + "loss": 0.3991, + "step": 7504 + }, + { + "epoch": 2.40134385665529, + "grad_norm": 0.509486973285675, + "learning_rate": 1.1639334117999495e-06, + "loss": 0.4398, + "step": 7505 + }, + { + "epoch": 2.401663822525597, + "grad_norm": 0.5020511150360107, + "learning_rate": 1.1627395373937105e-06, + "loss": 0.4081, + "step": 7506 + }, + { + "epoch": 2.4019837883959045, + "grad_norm": 0.534724771976471, + "learning_rate": 1.1615461950282526e-06, + "loss": 0.4332, + "step": 7507 + }, + { + "epoch": 2.4023037542662116, + "grad_norm": 0.4937615990638733, + "learning_rate": 1.1603533848690323e-06, + "loss": 0.4093, + "step": 7508 + }, + { + "epoch": 2.4026237201365186, + "grad_norm": 0.5317237377166748, + "learning_rate": 1.1591611070814345e-06, + "loss": 0.4902, + "step": 7509 + }, + { + "epoch": 2.402943686006826, + "grad_norm": 0.5365759134292603, + "learning_rate": 1.1579693618307685e-06, + "loss": 0.4281, + "step": 7510 + }, + { + "epoch": 2.403263651877133, + "grad_norm": 0.4887164831161499, + "learning_rate": 1.1567781492822728e-06, + "loss": 0.4479, + "step": 7511 + }, + { + "epoch": 2.40358361774744, + "grad_norm": 0.4960786700248718, + "learning_rate": 1.155587469601111e-06, + "loss": 0.4094, + "step": 7512 + }, + { + "epoch": 2.4039035836177476, + "grad_norm": 0.4631226658821106, + "learning_rate": 1.154397322952372e-06, + "loss": 0.4193, + "step": 7513 + }, + { + "epoch": 2.4042235494880546, + "grad_norm": 0.4892306923866272, + "learning_rate": 1.1532077095010701e-06, + "loss": 0.4171, + "step": 7514 + }, + { + "epoch": 2.4045435153583616, + "grad_norm": 0.45591482520103455, + "learning_rate": 1.152018629412146e-06, + "loss": 0.4065, + "step": 7515 + }, + { + "epoch": 2.404863481228669, + "grad_norm": 0.5342705249786377, + "learning_rate": 1.1508300828504682e-06, + "loss": 0.5449, + "step": 7516 + }, + { + "epoch": 2.405183447098976, + "grad_norm": 0.4758126437664032, + "learning_rate": 1.1496420699808313e-06, + "loss": 0.4127, + "step": 7517 + }, + { + "epoch": 2.405503412969283, + "grad_norm": 0.5583046078681946, + "learning_rate": 1.148454590967954e-06, + "loss": 0.4841, + "step": 7518 + }, + { + "epoch": 2.4058233788395906, + "grad_norm": 0.5164111852645874, + "learning_rate": 1.1472676459764804e-06, + "loss": 0.4191, + "step": 7519 + }, + { + "epoch": 2.4061433447098977, + "grad_norm": 0.46658846735954285, + "learning_rate": 1.1460812351709816e-06, + "loss": 0.4027, + "step": 7520 + }, + { + "epoch": 2.4064633105802047, + "grad_norm": 0.5523401498794556, + "learning_rate": 1.1448953587159562e-06, + "loss": 0.4675, + "step": 7521 + }, + { + "epoch": 2.406783276450512, + "grad_norm": 0.4960285723209381, + "learning_rate": 1.143710016775828e-06, + "loss": 0.4117, + "step": 7522 + }, + { + "epoch": 2.407103242320819, + "grad_norm": 0.5251019597053528, + "learning_rate": 1.1425252095149454e-06, + "loss": 0.4264, + "step": 7523 + }, + { + "epoch": 2.407423208191126, + "grad_norm": 0.5679874420166016, + "learning_rate": 1.141340937097583e-06, + "loss": 0.5073, + "step": 7524 + }, + { + "epoch": 2.4077431740614337, + "grad_norm": 0.5471564531326294, + "learning_rate": 1.1401571996879406e-06, + "loss": 0.4609, + "step": 7525 + }, + { + "epoch": 2.4080631399317407, + "grad_norm": 0.5527142286300659, + "learning_rate": 1.1389739974501462e-06, + "loss": 0.4086, + "step": 7526 + }, + { + "epoch": 2.4083831058020477, + "grad_norm": 0.6052654981613159, + "learning_rate": 1.137791330548253e-06, + "loss": 0.4123, + "step": 7527 + }, + { + "epoch": 2.4087030716723548, + "grad_norm": 0.4928595721721649, + "learning_rate": 1.1366091991462386e-06, + "loss": 0.5074, + "step": 7528 + }, + { + "epoch": 2.4090230375426622, + "grad_norm": 0.5060864090919495, + "learning_rate": 1.1354276034080059e-06, + "loss": 0.4108, + "step": 7529 + }, + { + "epoch": 2.4093430034129693, + "grad_norm": 0.5305314064025879, + "learning_rate": 1.1342465434973844e-06, + "loss": 0.4396, + "step": 7530 + }, + { + "epoch": 2.4096629692832763, + "grad_norm": 0.5134252905845642, + "learning_rate": 1.1330660195781295e-06, + "loss": 0.4404, + "step": 7531 + }, + { + "epoch": 2.4099829351535837, + "grad_norm": 0.551588237285614, + "learning_rate": 1.1318860318139251e-06, + "loss": 0.4645, + "step": 7532 + }, + { + "epoch": 2.4103029010238908, + "grad_norm": 0.5015571117401123, + "learning_rate": 1.1307065803683752e-06, + "loss": 0.4487, + "step": 7533 + }, + { + "epoch": 2.410622866894198, + "grad_norm": 0.4951358139514923, + "learning_rate": 1.1295276654050135e-06, + "loss": 0.4366, + "step": 7534 + }, + { + "epoch": 2.4109428327645053, + "grad_norm": 0.5068965554237366, + "learning_rate": 1.1283492870872952e-06, + "loss": 0.4491, + "step": 7535 + }, + { + "epoch": 2.4112627986348123, + "grad_norm": 0.5098022818565369, + "learning_rate": 1.1271714455786066e-06, + "loss": 0.4491, + "step": 7536 + }, + { + "epoch": 2.4115827645051193, + "grad_norm": 0.5193520188331604, + "learning_rate": 1.1259941410422575e-06, + "loss": 0.4263, + "step": 7537 + }, + { + "epoch": 2.411902730375427, + "grad_norm": 0.5024391412734985, + "learning_rate": 1.1248173736414807e-06, + "loss": 0.4529, + "step": 7538 + }, + { + "epoch": 2.412222696245734, + "grad_norm": 0.5296918749809265, + "learning_rate": 1.1236411435394378e-06, + "loss": 0.4679, + "step": 7539 + }, + { + "epoch": 2.412542662116041, + "grad_norm": 0.5214098691940308, + "learning_rate": 1.1224654508992117e-06, + "loss": 0.4602, + "step": 7540 + }, + { + "epoch": 2.412862627986348, + "grad_norm": 0.510133683681488, + "learning_rate": 1.1212902958838162e-06, + "loss": 0.3656, + "step": 7541 + }, + { + "epoch": 2.4131825938566553, + "grad_norm": 0.5108845233917236, + "learning_rate": 1.1201156786561884e-06, + "loss": 0.4555, + "step": 7542 + }, + { + "epoch": 2.4135025597269624, + "grad_norm": 0.5690526962280273, + "learning_rate": 1.1189415993791891e-06, + "loss": 0.4609, + "step": 7543 + }, + { + "epoch": 2.4138225255972694, + "grad_norm": 0.6002010703086853, + "learning_rate": 1.1177680582156064e-06, + "loss": 0.5271, + "step": 7544 + }, + { + "epoch": 2.414142491467577, + "grad_norm": 0.4729880690574646, + "learning_rate": 1.1165950553281518e-06, + "loss": 0.3858, + "step": 7545 + }, + { + "epoch": 2.414462457337884, + "grad_norm": 0.521458089351654, + "learning_rate": 1.115422590879464e-06, + "loss": 0.4827, + "step": 7546 + }, + { + "epoch": 2.414782423208191, + "grad_norm": 0.5038489103317261, + "learning_rate": 1.1142506650321088e-06, + "loss": 0.4493, + "step": 7547 + }, + { + "epoch": 2.4151023890784984, + "grad_norm": 0.47998735308647156, + "learning_rate": 1.1130792779485739e-06, + "loss": 0.4509, + "step": 7548 + }, + { + "epoch": 2.4154223549488054, + "grad_norm": 0.47233617305755615, + "learning_rate": 1.111908429791273e-06, + "loss": 0.4703, + "step": 7549 + }, + { + "epoch": 2.4157423208191124, + "grad_norm": 0.514860987663269, + "learning_rate": 1.1107381207225443e-06, + "loss": 0.4379, + "step": 7550 + }, + { + "epoch": 2.41606228668942, + "grad_norm": 0.5110282897949219, + "learning_rate": 1.109568350904654e-06, + "loss": 0.4381, + "step": 7551 + }, + { + "epoch": 2.416382252559727, + "grad_norm": 0.4801933765411377, + "learning_rate": 1.108399120499794e-06, + "loss": 0.4222, + "step": 7552 + }, + { + "epoch": 2.416702218430034, + "grad_norm": 0.45939740538597107, + "learning_rate": 1.107230429670077e-06, + "loss": 0.4076, + "step": 7553 + }, + { + "epoch": 2.4170221843003414, + "grad_norm": 0.5356783866882324, + "learning_rate": 1.1060622785775433e-06, + "loss": 0.4688, + "step": 7554 + }, + { + "epoch": 2.4173421501706485, + "grad_norm": 0.5647779703140259, + "learning_rate": 1.1048946673841598e-06, + "loss": 0.4787, + "step": 7555 + }, + { + "epoch": 2.4176621160409555, + "grad_norm": 0.5190406441688538, + "learning_rate": 1.1037275962518162e-06, + "loss": 0.4312, + "step": 7556 + }, + { + "epoch": 2.417982081911263, + "grad_norm": 0.4770490825176239, + "learning_rate": 1.1025610653423292e-06, + "loss": 0.4366, + "step": 7557 + }, + { + "epoch": 2.41830204778157, + "grad_norm": 0.4899611473083496, + "learning_rate": 1.1013950748174396e-06, + "loss": 0.4556, + "step": 7558 + }, + { + "epoch": 2.418622013651877, + "grad_norm": 0.559202253818512, + "learning_rate": 1.1002296248388116e-06, + "loss": 0.4417, + "step": 7559 + }, + { + "epoch": 2.4189419795221845, + "grad_norm": 0.5386381149291992, + "learning_rate": 1.099064715568039e-06, + "loss": 0.4158, + "step": 7560 + }, + { + "epoch": 2.4192619453924915, + "grad_norm": 0.4839184582233429, + "learning_rate": 1.0979003471666355e-06, + "loss": 0.4706, + "step": 7561 + }, + { + "epoch": 2.4195819112627985, + "grad_norm": 0.5128084421157837, + "learning_rate": 1.0967365197960444e-06, + "loss": 0.4509, + "step": 7562 + }, + { + "epoch": 2.419901877133106, + "grad_norm": 0.4908483326435089, + "learning_rate": 1.09557323361763e-06, + "loss": 0.4002, + "step": 7563 + }, + { + "epoch": 2.420221843003413, + "grad_norm": 0.48702266812324524, + "learning_rate": 1.0944104887926833e-06, + "loss": 0.4484, + "step": 7564 + }, + { + "epoch": 2.42054180887372, + "grad_norm": 0.5335679054260254, + "learning_rate": 1.0932482854824216e-06, + "loss": 0.455, + "step": 7565 + }, + { + "epoch": 2.4208617747440275, + "grad_norm": 0.5421773791313171, + "learning_rate": 1.0920866238479843e-06, + "loss": 0.4386, + "step": 7566 + }, + { + "epoch": 2.4211817406143346, + "grad_norm": 0.5170466899871826, + "learning_rate": 1.090925504050439e-06, + "loss": 0.4574, + "step": 7567 + }, + { + "epoch": 2.4215017064846416, + "grad_norm": 0.537121057510376, + "learning_rate": 1.0897649262507753e-06, + "loss": 0.4154, + "step": 7568 + }, + { + "epoch": 2.421821672354949, + "grad_norm": 0.4950486123561859, + "learning_rate": 1.0886048906099073e-06, + "loss": 0.4175, + "step": 7569 + }, + { + "epoch": 2.422141638225256, + "grad_norm": 0.46320489048957825, + "learning_rate": 1.0874453972886783e-06, + "loss": 0.4659, + "step": 7570 + }, + { + "epoch": 2.422461604095563, + "grad_norm": 0.4615775942802429, + "learning_rate": 1.08628644644785e-06, + "loss": 0.4238, + "step": 7571 + }, + { + "epoch": 2.42278156996587, + "grad_norm": 0.49678483605384827, + "learning_rate": 1.0851280382481161e-06, + "loss": 0.416, + "step": 7572 + }, + { + "epoch": 2.4231015358361776, + "grad_norm": 0.5429278016090393, + "learning_rate": 1.0839701728500896e-06, + "loss": 0.4555, + "step": 7573 + }, + { + "epoch": 2.4234215017064846, + "grad_norm": 0.5306439399719238, + "learning_rate": 1.0828128504143076e-06, + "loss": 0.4428, + "step": 7574 + }, + { + "epoch": 2.4237414675767917, + "grad_norm": 0.5775001645088196, + "learning_rate": 1.0816560711012386e-06, + "loss": 0.4523, + "step": 7575 + }, + { + "epoch": 2.424061433447099, + "grad_norm": 0.5291292071342468, + "learning_rate": 1.080499835071267e-06, + "loss": 0.4678, + "step": 7576 + }, + { + "epoch": 2.424381399317406, + "grad_norm": 0.4802502989768982, + "learning_rate": 1.0793441424847106e-06, + "loss": 0.4286, + "step": 7577 + }, + { + "epoch": 2.424701365187713, + "grad_norm": 0.48766323924064636, + "learning_rate": 1.0781889935018053e-06, + "loss": 0.4102, + "step": 7578 + }, + { + "epoch": 2.4250213310580206, + "grad_norm": 0.5157595872879028, + "learning_rate": 1.0770343882827127e-06, + "loss": 0.4421, + "step": 7579 + }, + { + "epoch": 2.4253412969283277, + "grad_norm": 0.5109654664993286, + "learning_rate": 1.0758803269875228e-06, + "loss": 0.4911, + "step": 7580 + }, + { + "epoch": 2.4256612627986347, + "grad_norm": 0.462458997964859, + "learning_rate": 1.0747268097762454e-06, + "loss": 0.4184, + "step": 7581 + }, + { + "epoch": 2.425981228668942, + "grad_norm": 0.5116490125656128, + "learning_rate": 1.0735738368088188e-06, + "loss": 0.4274, + "step": 7582 + }, + { + "epoch": 2.426301194539249, + "grad_norm": 0.5581761002540588, + "learning_rate": 1.072421408245104e-06, + "loss": 0.4847, + "step": 7583 + }, + { + "epoch": 2.426621160409556, + "grad_norm": 0.539322018623352, + "learning_rate": 1.071269524244884e-06, + "loss": 0.4175, + "step": 7584 + }, + { + "epoch": 2.4269411262798632, + "grad_norm": 0.4879395365715027, + "learning_rate": 1.0701181849678726e-06, + "loss": 0.4472, + "step": 7585 + }, + { + "epoch": 2.4272610921501707, + "grad_norm": 0.4611521363258362, + "learning_rate": 1.0689673905737013e-06, + "loss": 0.4399, + "step": 7586 + }, + { + "epoch": 2.4275810580204777, + "grad_norm": 0.4945104420185089, + "learning_rate": 1.0678171412219317e-06, + "loss": 0.4799, + "step": 7587 + }, + { + "epoch": 2.4279010238907848, + "grad_norm": 0.4701174199581146, + "learning_rate": 1.0666674370720442e-06, + "loss": 0.4129, + "step": 7588 + }, + { + "epoch": 2.4282209897610922, + "grad_norm": 0.5425107479095459, + "learning_rate": 1.0655182782834505e-06, + "loss": 0.4462, + "step": 7589 + }, + { + "epoch": 2.4285409556313993, + "grad_norm": 0.5583622455596924, + "learning_rate": 1.0643696650154805e-06, + "loss": 0.4759, + "step": 7590 + }, + { + "epoch": 2.4288609215017063, + "grad_norm": 0.5604155659675598, + "learning_rate": 1.0632215974273897e-06, + "loss": 0.4477, + "step": 7591 + }, + { + "epoch": 2.4291808873720138, + "grad_norm": 0.47683799266815186, + "learning_rate": 1.062074075678362e-06, + "loss": 0.395, + "step": 7592 + }, + { + "epoch": 2.429500853242321, + "grad_norm": 0.5691552758216858, + "learning_rate": 1.0609270999275e-06, + "loss": 0.473, + "step": 7593 + }, + { + "epoch": 2.429820819112628, + "grad_norm": 0.591694712638855, + "learning_rate": 1.0597806703338354e-06, + "loss": 0.4721, + "step": 7594 + }, + { + "epoch": 2.4301407849829353, + "grad_norm": 0.5700772404670715, + "learning_rate": 1.0586347870563213e-06, + "loss": 0.439, + "step": 7595 + }, + { + "epoch": 2.4304607508532423, + "grad_norm": 0.5037002563476562, + "learning_rate": 1.057489450253834e-06, + "loss": 0.3811, + "step": 7596 + }, + { + "epoch": 2.4307807167235493, + "grad_norm": 0.5461633801460266, + "learning_rate": 1.0563446600851784e-06, + "loss": 0.4682, + "step": 7597 + }, + { + "epoch": 2.431100682593857, + "grad_norm": 0.4951798915863037, + "learning_rate": 1.0552004167090795e-06, + "loss": 0.4089, + "step": 7598 + }, + { + "epoch": 2.431420648464164, + "grad_norm": 0.5015349984169006, + "learning_rate": 1.0540567202841894e-06, + "loss": 0.4096, + "step": 7599 + }, + { + "epoch": 2.431740614334471, + "grad_norm": 0.522261917591095, + "learning_rate": 1.052913570969082e-06, + "loss": 0.5391, + "step": 7600 + }, + { + "epoch": 2.4320605802047783, + "grad_norm": 0.4680025577545166, + "learning_rate": 1.0517709689222556e-06, + "loss": 0.3876, + "step": 7601 + }, + { + "epoch": 2.4323805460750854, + "grad_norm": 0.5641904473304749, + "learning_rate": 1.0506289143021348e-06, + "loss": 0.4187, + "step": 7602 + }, + { + "epoch": 2.4327005119453924, + "grad_norm": 0.4958467185497284, + "learning_rate": 1.0494874072670653e-06, + "loss": 0.4498, + "step": 7603 + }, + { + "epoch": 2.4330204778157, + "grad_norm": 0.48778653144836426, + "learning_rate": 1.0483464479753208e-06, + "loss": 0.4209, + "step": 7604 + }, + { + "epoch": 2.433340443686007, + "grad_norm": 0.5240517854690552, + "learning_rate": 1.047206036585095e-06, + "loss": 0.4644, + "step": 7605 + }, + { + "epoch": 2.433660409556314, + "grad_norm": 0.4938124120235443, + "learning_rate": 1.0460661732545063e-06, + "loss": 0.4193, + "step": 7606 + }, + { + "epoch": 2.4339803754266214, + "grad_norm": 0.4557558000087738, + "learning_rate": 1.0449268581416012e-06, + "loss": 0.4508, + "step": 7607 + }, + { + "epoch": 2.4343003412969284, + "grad_norm": 0.4963361620903015, + "learning_rate": 1.043788091404343e-06, + "loss": 0.4048, + "step": 7608 + }, + { + "epoch": 2.4346203071672354, + "grad_norm": 0.5271517038345337, + "learning_rate": 1.0426498732006273e-06, + "loss": 0.479, + "step": 7609 + }, + { + "epoch": 2.434940273037543, + "grad_norm": 0.4944269359111786, + "learning_rate": 1.0415122036882674e-06, + "loss": 0.4145, + "step": 7610 + }, + { + "epoch": 2.43526023890785, + "grad_norm": 0.48690417408943176, + "learning_rate": 1.0403750830250014e-06, + "loss": 0.4536, + "step": 7611 + }, + { + "epoch": 2.435580204778157, + "grad_norm": 0.5064367055892944, + "learning_rate": 1.0392385113684949e-06, + "loss": 0.4342, + "step": 7612 + }, + { + "epoch": 2.435900170648464, + "grad_norm": 0.5235652923583984, + "learning_rate": 1.0381024888763324e-06, + "loss": 0.4116, + "step": 7613 + }, + { + "epoch": 2.4362201365187715, + "grad_norm": 0.57554030418396, + "learning_rate": 1.0369670157060275e-06, + "loss": 0.4318, + "step": 7614 + }, + { + "epoch": 2.4365401023890785, + "grad_norm": 0.6025678515434265, + "learning_rate": 1.0358320920150133e-06, + "loss": 0.4823, + "step": 7615 + }, + { + "epoch": 2.4368600682593855, + "grad_norm": 0.4879830777645111, + "learning_rate": 1.0346977179606477e-06, + "loss": 0.3813, + "step": 7616 + }, + { + "epoch": 2.437180034129693, + "grad_norm": 0.5200275778770447, + "learning_rate": 1.0335638937002145e-06, + "loss": 0.4524, + "step": 7617 + }, + { + "epoch": 2.4375, + "grad_norm": 0.5162732601165771, + "learning_rate": 1.0324306193909185e-06, + "loss": 0.4334, + "step": 7618 + }, + { + "epoch": 2.437819965870307, + "grad_norm": 0.5059532523155212, + "learning_rate": 1.031297895189891e-06, + "loss": 0.4214, + "step": 7619 + }, + { + "epoch": 2.4381399317406145, + "grad_norm": 0.5162801742553711, + "learning_rate": 1.0301657212541854e-06, + "loss": 0.4186, + "step": 7620 + }, + { + "epoch": 2.4384598976109215, + "grad_norm": 0.5356256365776062, + "learning_rate": 1.029034097740777e-06, + "loss": 0.4177, + "step": 7621 + }, + { + "epoch": 2.4387798634812285, + "grad_norm": 0.50387042760849, + "learning_rate": 1.0279030248065675e-06, + "loss": 0.4223, + "step": 7622 + }, + { + "epoch": 2.439099829351536, + "grad_norm": 0.5306486487388611, + "learning_rate": 1.0267725026083842e-06, + "loss": 0.4674, + "step": 7623 + }, + { + "epoch": 2.439419795221843, + "grad_norm": 0.5405206680297852, + "learning_rate": 1.0256425313029727e-06, + "loss": 0.5039, + "step": 7624 + }, + { + "epoch": 2.43973976109215, + "grad_norm": 0.5215193629264832, + "learning_rate": 1.0245131110470058e-06, + "loss": 0.4404, + "step": 7625 + }, + { + "epoch": 2.440059726962457, + "grad_norm": 0.477377325296402, + "learning_rate": 1.0233842419970773e-06, + "loss": 0.4434, + "step": 7626 + }, + { + "epoch": 2.4403796928327646, + "grad_norm": 0.455512672662735, + "learning_rate": 1.022255924309708e-06, + "loss": 0.3599, + "step": 7627 + }, + { + "epoch": 2.4406996587030716, + "grad_norm": 0.4892837107181549, + "learning_rate": 1.0211281581413407e-06, + "loss": 0.4431, + "step": 7628 + }, + { + "epoch": 2.4410196245733786, + "grad_norm": 0.492992639541626, + "learning_rate": 1.0200009436483416e-06, + "loss": 0.4433, + "step": 7629 + }, + { + "epoch": 2.441339590443686, + "grad_norm": 0.5320823192596436, + "learning_rate": 1.0188742809869994e-06, + "loss": 0.4859, + "step": 7630 + }, + { + "epoch": 2.441659556313993, + "grad_norm": 0.5062793493270874, + "learning_rate": 1.0177481703135268e-06, + "loss": 0.4247, + "step": 7631 + }, + { + "epoch": 2.4419795221843, + "grad_norm": 0.48646122217178345, + "learning_rate": 1.0166226117840606e-06, + "loss": 0.4214, + "step": 7632 + }, + { + "epoch": 2.4422994880546076, + "grad_norm": 0.5569018721580505, + "learning_rate": 1.0154976055546627e-06, + "loss": 0.452, + "step": 7633 + }, + { + "epoch": 2.4426194539249146, + "grad_norm": 0.5125252604484558, + "learning_rate": 1.0143731517813154e-06, + "loss": 0.444, + "step": 7634 + }, + { + "epoch": 2.4429394197952217, + "grad_norm": 0.5162290930747986, + "learning_rate": 1.0132492506199244e-06, + "loss": 0.4762, + "step": 7635 + }, + { + "epoch": 2.443259385665529, + "grad_norm": 0.5181790590286255, + "learning_rate": 1.0121259022263219e-06, + "loss": 0.4177, + "step": 7636 + }, + { + "epoch": 2.443579351535836, + "grad_norm": 0.4434118866920471, + "learning_rate": 1.0110031067562592e-06, + "loss": 0.4134, + "step": 7637 + }, + { + "epoch": 2.443899317406143, + "grad_norm": 0.5199148654937744, + "learning_rate": 1.0098808643654163e-06, + "loss": 0.4569, + "step": 7638 + }, + { + "epoch": 2.4442192832764507, + "grad_norm": 0.5765395760536194, + "learning_rate": 1.0087591752093912e-06, + "loss": 0.4519, + "step": 7639 + }, + { + "epoch": 2.4445392491467577, + "grad_norm": 0.483059823513031, + "learning_rate": 1.0076380394437073e-06, + "loss": 0.438, + "step": 7640 + }, + { + "epoch": 2.4448592150170647, + "grad_norm": 0.4759012758731842, + "learning_rate": 1.0065174572238133e-06, + "loss": 0.4232, + "step": 7641 + }, + { + "epoch": 2.445179180887372, + "grad_norm": 0.5737249255180359, + "learning_rate": 1.0053974287050767e-06, + "loss": 0.4544, + "step": 7642 + }, + { + "epoch": 2.445499146757679, + "grad_norm": 0.5525760054588318, + "learning_rate": 1.0042779540427937e-06, + "loss": 0.4319, + "step": 7643 + }, + { + "epoch": 2.4458191126279862, + "grad_norm": 0.5611857771873474, + "learning_rate": 1.0031590333921791e-06, + "loss": 0.4473, + "step": 7644 + }, + { + "epoch": 2.4461390784982937, + "grad_norm": 0.5297382473945618, + "learning_rate": 1.0020406669083721e-06, + "loss": 0.4774, + "step": 7645 + }, + { + "epoch": 2.4464590443686007, + "grad_norm": 0.49018895626068115, + "learning_rate": 1.0009228547464373e-06, + "loss": 0.4144, + "step": 7646 + }, + { + "epoch": 2.4467790102389078, + "grad_norm": 0.48025065660476685, + "learning_rate": 9.99805597061358e-07, + "loss": 0.4625, + "step": 7647 + }, + { + "epoch": 2.4470989761092152, + "grad_norm": 0.540985643863678, + "learning_rate": 9.986888940080468e-07, + "loss": 0.4483, + "step": 7648 + }, + { + "epoch": 2.4474189419795223, + "grad_norm": 0.4940546751022339, + "learning_rate": 9.975727457413342e-07, + "loss": 0.4043, + "step": 7649 + }, + { + "epoch": 2.4477389078498293, + "grad_norm": 0.47084319591522217, + "learning_rate": 9.964571524159738e-07, + "loss": 0.4393, + "step": 7650 + }, + { + "epoch": 2.4480588737201368, + "grad_norm": 0.4575006067752838, + "learning_rate": 9.95342114186647e-07, + "loss": 0.4384, + "step": 7651 + }, + { + "epoch": 2.448378839590444, + "grad_norm": 0.5515487790107727, + "learning_rate": 9.942276312079524e-07, + "loss": 0.5155, + "step": 7652 + }, + { + "epoch": 2.448698805460751, + "grad_norm": 0.5114273428916931, + "learning_rate": 9.931137036344174e-07, + "loss": 0.489, + "step": 7653 + }, + { + "epoch": 2.449018771331058, + "grad_norm": 0.5337884426116943, + "learning_rate": 9.920003316204874e-07, + "loss": 0.4619, + "step": 7654 + }, + { + "epoch": 2.4493387372013653, + "grad_norm": 0.4916205108165741, + "learning_rate": 9.908875153205321e-07, + "loss": 0.3839, + "step": 7655 + }, + { + "epoch": 2.4496587030716723, + "grad_norm": 0.4589358866214752, + "learning_rate": 9.897752548888461e-07, + "loss": 0.4375, + "step": 7656 + }, + { + "epoch": 2.4499786689419794, + "grad_norm": 0.45303237438201904, + "learning_rate": 9.886635504796467e-07, + "loss": 0.4195, + "step": 7657 + }, + { + "epoch": 2.450298634812287, + "grad_norm": 0.4342546761035919, + "learning_rate": 9.875524022470713e-07, + "loss": 0.4156, + "step": 7658 + }, + { + "epoch": 2.450618600682594, + "grad_norm": 0.480857253074646, + "learning_rate": 9.86441810345183e-07, + "loss": 0.4451, + "step": 7659 + }, + { + "epoch": 2.450938566552901, + "grad_norm": 0.5475195050239563, + "learning_rate": 9.85331774927964e-07, + "loss": 0.5084, + "step": 7660 + }, + { + "epoch": 2.4512585324232083, + "grad_norm": 0.48804011940956116, + "learning_rate": 9.842222961493247e-07, + "loss": 0.4208, + "step": 7661 + }, + { + "epoch": 2.4515784982935154, + "grad_norm": 0.4797302186489105, + "learning_rate": 9.831133741630966e-07, + "loss": 0.4538, + "step": 7662 + }, + { + "epoch": 2.4518984641638224, + "grad_norm": 0.48906010389328003, + "learning_rate": 9.82005009123031e-07, + "loss": 0.4558, + "step": 7663 + }, + { + "epoch": 2.45221843003413, + "grad_norm": 0.522978663444519, + "learning_rate": 9.808972011828055e-07, + "loss": 0.4595, + "step": 7664 + }, + { + "epoch": 2.452538395904437, + "grad_norm": 0.5094500780105591, + "learning_rate": 9.797899504960162e-07, + "loss": 0.4489, + "step": 7665 + }, + { + "epoch": 2.452858361774744, + "grad_norm": 0.5380294322967529, + "learning_rate": 9.78683257216187e-07, + "loss": 0.4189, + "step": 7666 + }, + { + "epoch": 2.453178327645051, + "grad_norm": 0.5117015242576599, + "learning_rate": 9.775771214967633e-07, + "loss": 0.4287, + "step": 7667 + }, + { + "epoch": 2.4534982935153584, + "grad_norm": 0.4672139585018158, + "learning_rate": 9.764715434911104e-07, + "loss": 0.4106, + "step": 7668 + }, + { + "epoch": 2.4538182593856654, + "grad_norm": 0.5067694187164307, + "learning_rate": 9.753665233525188e-07, + "loss": 0.469, + "step": 7669 + }, + { + "epoch": 2.4541382252559725, + "grad_norm": 0.48697370290756226, + "learning_rate": 9.742620612341992e-07, + "loss": 0.4012, + "step": 7670 + }, + { + "epoch": 2.45445819112628, + "grad_norm": 0.5106503963470459, + "learning_rate": 9.731581572892878e-07, + "loss": 0.4187, + "step": 7671 + }, + { + "epoch": 2.454778156996587, + "grad_norm": 0.4614931046962738, + "learning_rate": 9.720548116708434e-07, + "loss": 0.429, + "step": 7672 + }, + { + "epoch": 2.455098122866894, + "grad_norm": 0.5304867625236511, + "learning_rate": 9.709520245318453e-07, + "loss": 0.4584, + "step": 7673 + }, + { + "epoch": 2.4554180887372015, + "grad_norm": 0.5220844149589539, + "learning_rate": 9.698497960251957e-07, + "loss": 0.4016, + "step": 7674 + }, + { + "epoch": 2.4557380546075085, + "grad_norm": 0.4315398633480072, + "learning_rate": 9.687481263037196e-07, + "loss": 0.4386, + "step": 7675 + }, + { + "epoch": 2.4560580204778155, + "grad_norm": 0.44316765666007996, + "learning_rate": 9.67647015520165e-07, + "loss": 0.4077, + "step": 7676 + }, + { + "epoch": 2.456377986348123, + "grad_norm": 0.5713621973991394, + "learning_rate": 9.665464638272044e-07, + "loss": 0.5292, + "step": 7677 + }, + { + "epoch": 2.45669795221843, + "grad_norm": 0.5227526426315308, + "learning_rate": 9.654464713774286e-07, + "loss": 0.4227, + "step": 7678 + }, + { + "epoch": 2.457017918088737, + "grad_norm": 0.5182657837867737, + "learning_rate": 9.64347038323354e-07, + "loss": 0.4476, + "step": 7679 + }, + { + "epoch": 2.4573378839590445, + "grad_norm": 0.4926142990589142, + "learning_rate": 9.632481648174151e-07, + "loss": 0.4527, + "step": 7680 + }, + { + "epoch": 2.4576578498293515, + "grad_norm": 0.48976460099220276, + "learning_rate": 9.621498510119753e-07, + "loss": 0.4618, + "step": 7681 + }, + { + "epoch": 2.4579778156996586, + "grad_norm": 0.4745182394981384, + "learning_rate": 9.610520970593173e-07, + "loss": 0.4175, + "step": 7682 + }, + { + "epoch": 2.458297781569966, + "grad_norm": 0.5499017238616943, + "learning_rate": 9.59954903111645e-07, + "loss": 0.4633, + "step": 7683 + }, + { + "epoch": 2.458617747440273, + "grad_norm": 0.5084142684936523, + "learning_rate": 9.588582693210858e-07, + "loss": 0.4562, + "step": 7684 + }, + { + "epoch": 2.45893771331058, + "grad_norm": 0.4486640691757202, + "learning_rate": 9.577621958396876e-07, + "loss": 0.3917, + "step": 7685 + }, + { + "epoch": 2.4592576791808876, + "grad_norm": 0.49820923805236816, + "learning_rate": 9.566666828194243e-07, + "loss": 0.448, + "step": 7686 + }, + { + "epoch": 2.4595776450511946, + "grad_norm": 0.5000406503677368, + "learning_rate": 9.55571730412191e-07, + "loss": 0.4524, + "step": 7687 + }, + { + "epoch": 2.4598976109215016, + "grad_norm": 0.48786410689353943, + "learning_rate": 9.544773387698025e-07, + "loss": 0.4435, + "step": 7688 + }, + { + "epoch": 2.460217576791809, + "grad_norm": 0.5494711995124817, + "learning_rate": 9.533835080439979e-07, + "loss": 0.4356, + "step": 7689 + }, + { + "epoch": 2.460537542662116, + "grad_norm": 0.5086516737937927, + "learning_rate": 9.522902383864369e-07, + "loss": 0.4204, + "step": 7690 + }, + { + "epoch": 2.460857508532423, + "grad_norm": 0.4816153645515442, + "learning_rate": 9.511975299487036e-07, + "loss": 0.4525, + "step": 7691 + }, + { + "epoch": 2.4611774744027306, + "grad_norm": 0.45684894919395447, + "learning_rate": 9.501053828823054e-07, + "loss": 0.4117, + "step": 7692 + }, + { + "epoch": 2.4614974402730376, + "grad_norm": 0.45524606108665466, + "learning_rate": 9.490137973386676e-07, + "loss": 0.4343, + "step": 7693 + }, + { + "epoch": 2.4618174061433447, + "grad_norm": 0.4852887690067291, + "learning_rate": 9.479227734691404e-07, + "loss": 0.4329, + "step": 7694 + }, + { + "epoch": 2.462137372013652, + "grad_norm": 0.5166252851486206, + "learning_rate": 9.468323114249944e-07, + "loss": 0.4293, + "step": 7695 + }, + { + "epoch": 2.462457337883959, + "grad_norm": 0.5214166641235352, + "learning_rate": 9.457424113574238e-07, + "loss": 0.4486, + "step": 7696 + }, + { + "epoch": 2.462777303754266, + "grad_norm": 0.5261161923408508, + "learning_rate": 9.446530734175474e-07, + "loss": 0.432, + "step": 7697 + }, + { + "epoch": 2.463097269624573, + "grad_norm": 0.4776790142059326, + "learning_rate": 9.435642977564002e-07, + "loss": 0.423, + "step": 7698 + }, + { + "epoch": 2.4634172354948807, + "grad_norm": 0.5481368899345398, + "learning_rate": 9.424760845249436e-07, + "loss": 0.4894, + "step": 7699 + }, + { + "epoch": 2.4637372013651877, + "grad_norm": 0.5501801371574402, + "learning_rate": 9.413884338740581e-07, + "loss": 0.4141, + "step": 7700 + }, + { + "epoch": 2.4640571672354947, + "grad_norm": 0.5201511383056641, + "learning_rate": 9.403013459545485e-07, + "loss": 0.4394, + "step": 7701 + }, + { + "epoch": 2.464377133105802, + "grad_norm": 0.5060268044471741, + "learning_rate": 9.392148209171432e-07, + "loss": 0.418, + "step": 7702 + }, + { + "epoch": 2.4646970989761092, + "grad_norm": 0.5173326134681702, + "learning_rate": 9.381288589124877e-07, + "loss": 0.4517, + "step": 7703 + }, + { + "epoch": 2.4650170648464163, + "grad_norm": 0.4509728252887726, + "learning_rate": 9.370434600911521e-07, + "loss": 0.4435, + "step": 7704 + }, + { + "epoch": 2.4653370307167237, + "grad_norm": 0.5011699199676514, + "learning_rate": 9.359586246036278e-07, + "loss": 0.4386, + "step": 7705 + }, + { + "epoch": 2.4656569965870307, + "grad_norm": 0.5216577053070068, + "learning_rate": 9.34874352600329e-07, + "loss": 0.4481, + "step": 7706 + }, + { + "epoch": 2.4659769624573378, + "grad_norm": 0.5623379349708557, + "learning_rate": 9.337906442315936e-07, + "loss": 0.4597, + "step": 7707 + }, + { + "epoch": 2.4662969283276452, + "grad_norm": 0.5247960090637207, + "learning_rate": 9.327074996476765e-07, + "loss": 0.4602, + "step": 7708 + }, + { + "epoch": 2.4666168941979523, + "grad_norm": 0.49052828550338745, + "learning_rate": 9.31624918998758e-07, + "loss": 0.3796, + "step": 7709 + }, + { + "epoch": 2.4669368600682593, + "grad_norm": 0.44440314173698425, + "learning_rate": 9.305429024349371e-07, + "loss": 0.404, + "step": 7710 + }, + { + "epoch": 2.4672568259385663, + "grad_norm": 0.47280484437942505, + "learning_rate": 9.294614501062393e-07, + "loss": 0.4225, + "step": 7711 + }, + { + "epoch": 2.467576791808874, + "grad_norm": 0.47758734226226807, + "learning_rate": 9.283805621626091e-07, + "loss": 0.436, + "step": 7712 + }, + { + "epoch": 2.467896757679181, + "grad_norm": 0.5277053117752075, + "learning_rate": 9.273002387539126e-07, + "loss": 0.4753, + "step": 7713 + }, + { + "epoch": 2.468216723549488, + "grad_norm": 0.4714113473892212, + "learning_rate": 9.262204800299374e-07, + "loss": 0.3981, + "step": 7714 + }, + { + "epoch": 2.4685366894197953, + "grad_norm": 0.5600582957267761, + "learning_rate": 9.251412861403919e-07, + "loss": 0.5198, + "step": 7715 + }, + { + "epoch": 2.4688566552901023, + "grad_norm": 0.48871341347694397, + "learning_rate": 9.240626572349093e-07, + "loss": 0.4157, + "step": 7716 + }, + { + "epoch": 2.4691766211604094, + "grad_norm": 0.4396373927593231, + "learning_rate": 9.229845934630443e-07, + "loss": 0.414, + "step": 7717 + }, + { + "epoch": 2.469496587030717, + "grad_norm": 0.5115680694580078, + "learning_rate": 9.219070949742703e-07, + "loss": 0.4254, + "step": 7718 + }, + { + "epoch": 2.469816552901024, + "grad_norm": 0.5043185949325562, + "learning_rate": 9.208301619179833e-07, + "loss": 0.4315, + "step": 7719 + }, + { + "epoch": 2.470136518771331, + "grad_norm": 0.5135460495948792, + "learning_rate": 9.197537944435009e-07, + "loss": 0.4285, + "step": 7720 + }, + { + "epoch": 2.4704564846416384, + "grad_norm": 0.5104293823242188, + "learning_rate": 9.186779927000633e-07, + "loss": 0.4469, + "step": 7721 + }, + { + "epoch": 2.4707764505119454, + "grad_norm": 0.5346876382827759, + "learning_rate": 9.176027568368328e-07, + "loss": 0.4353, + "step": 7722 + }, + { + "epoch": 2.4710964163822524, + "grad_norm": 0.4901260435581207, + "learning_rate": 9.165280870028919e-07, + "loss": 0.4265, + "step": 7723 + }, + { + "epoch": 2.47141638225256, + "grad_norm": 0.49625712633132935, + "learning_rate": 9.154539833472442e-07, + "loss": 0.4654, + "step": 7724 + }, + { + "epoch": 2.471736348122867, + "grad_norm": 0.4737713932991028, + "learning_rate": 9.143804460188144e-07, + "loss": 0.4264, + "step": 7725 + }, + { + "epoch": 2.472056313993174, + "grad_norm": 0.4819645285606384, + "learning_rate": 9.133074751664506e-07, + "loss": 0.4697, + "step": 7726 + }, + { + "epoch": 2.4723762798634814, + "grad_norm": 0.4653226137161255, + "learning_rate": 9.122350709389228e-07, + "loss": 0.4025, + "step": 7727 + }, + { + "epoch": 2.4726962457337884, + "grad_norm": 0.5325989723205566, + "learning_rate": 9.111632334849207e-07, + "loss": 0.5155, + "step": 7728 + }, + { + "epoch": 2.4730162116040955, + "grad_norm": 0.4841369390487671, + "learning_rate": 9.100919629530536e-07, + "loss": 0.4066, + "step": 7729 + }, + { + "epoch": 2.473336177474403, + "grad_norm": 0.546658992767334, + "learning_rate": 9.090212594918574e-07, + "loss": 0.4723, + "step": 7730 + }, + { + "epoch": 2.47365614334471, + "grad_norm": 0.48808419704437256, + "learning_rate": 9.07951123249784e-07, + "loss": 0.4369, + "step": 7731 + }, + { + "epoch": 2.473976109215017, + "grad_norm": 0.49957290291786194, + "learning_rate": 9.068815543752113e-07, + "loss": 0.4063, + "step": 7732 + }, + { + "epoch": 2.4742960750853245, + "grad_norm": 0.5041957497596741, + "learning_rate": 9.058125530164358e-07, + "loss": 0.4257, + "step": 7733 + }, + { + "epoch": 2.4746160409556315, + "grad_norm": 0.4725160300731659, + "learning_rate": 9.047441193216733e-07, + "loss": 0.4227, + "step": 7734 + }, + { + "epoch": 2.4749360068259385, + "grad_norm": 0.4849986732006073, + "learning_rate": 9.036762534390669e-07, + "loss": 0.4623, + "step": 7735 + }, + { + "epoch": 2.475255972696246, + "grad_norm": 0.4827047288417816, + "learning_rate": 9.026089555166745e-07, + "loss": 0.3774, + "step": 7736 + }, + { + "epoch": 2.475575938566553, + "grad_norm": 0.5167376399040222, + "learning_rate": 9.015422257024814e-07, + "loss": 0.4669, + "step": 7737 + }, + { + "epoch": 2.47589590443686, + "grad_norm": 0.4632802903652191, + "learning_rate": 9.004760641443888e-07, + "loss": 0.4574, + "step": 7738 + }, + { + "epoch": 2.476215870307167, + "grad_norm": 0.4840168356895447, + "learning_rate": 8.994104709902207e-07, + "loss": 0.4308, + "step": 7739 + }, + { + "epoch": 2.4765358361774745, + "grad_norm": 0.5287582278251648, + "learning_rate": 8.983454463877256e-07, + "loss": 0.4503, + "step": 7740 + }, + { + "epoch": 2.4768558020477816, + "grad_norm": 0.4669416844844818, + "learning_rate": 8.972809904845669e-07, + "loss": 0.4093, + "step": 7741 + }, + { + "epoch": 2.4771757679180886, + "grad_norm": 0.5467846393585205, + "learning_rate": 8.962171034283362e-07, + "loss": 0.4409, + "step": 7742 + }, + { + "epoch": 2.477495733788396, + "grad_norm": 0.5437021851539612, + "learning_rate": 8.95153785366541e-07, + "loss": 0.491, + "step": 7743 + }, + { + "epoch": 2.477815699658703, + "grad_norm": 0.44450342655181885, + "learning_rate": 8.940910364466104e-07, + "loss": 0.414, + "step": 7744 + }, + { + "epoch": 2.47813566552901, + "grad_norm": 0.5120328664779663, + "learning_rate": 8.930288568158985e-07, + "loss": 0.4784, + "step": 7745 + }, + { + "epoch": 2.4784556313993176, + "grad_norm": 0.48859596252441406, + "learning_rate": 8.919672466216744e-07, + "loss": 0.4486, + "step": 7746 + }, + { + "epoch": 2.4787755972696246, + "grad_norm": 0.49027806520462036, + "learning_rate": 8.909062060111356e-07, + "loss": 0.4157, + "step": 7747 + }, + { + "epoch": 2.4790955631399316, + "grad_norm": 0.4924556314945221, + "learning_rate": 8.89845735131395e-07, + "loss": 0.4658, + "step": 7748 + }, + { + "epoch": 2.479415529010239, + "grad_norm": 0.5062253475189209, + "learning_rate": 8.887858341294859e-07, + "loss": 0.4544, + "step": 7749 + }, + { + "epoch": 2.479735494880546, + "grad_norm": 0.4840138852596283, + "learning_rate": 8.877265031523685e-07, + "loss": 0.4084, + "step": 7750 + }, + { + "epoch": 2.480055460750853, + "grad_norm": 0.49299880862236023, + "learning_rate": 8.866677423469172e-07, + "loss": 0.3985, + "step": 7751 + }, + { + "epoch": 2.48037542662116, + "grad_norm": 0.5082073211669922, + "learning_rate": 8.856095518599333e-07, + "loss": 0.4692, + "step": 7752 + }, + { + "epoch": 2.4806953924914676, + "grad_norm": 0.49071094393730164, + "learning_rate": 8.845519318381346e-07, + "loss": 0.475, + "step": 7753 + }, + { + "epoch": 2.4810153583617747, + "grad_norm": 0.4831361174583435, + "learning_rate": 8.834948824281603e-07, + "loss": 0.3992, + "step": 7754 + }, + { + "epoch": 2.4813353242320817, + "grad_norm": 0.49088916182518005, + "learning_rate": 8.824384037765743e-07, + "loss": 0.4297, + "step": 7755 + }, + { + "epoch": 2.481655290102389, + "grad_norm": 0.4936610162258148, + "learning_rate": 8.813824960298556e-07, + "loss": 0.4683, + "step": 7756 + }, + { + "epoch": 2.481975255972696, + "grad_norm": 0.5130487084388733, + "learning_rate": 8.803271593344098e-07, + "loss": 0.4086, + "step": 7757 + }, + { + "epoch": 2.482295221843003, + "grad_norm": 0.5862554907798767, + "learning_rate": 8.792723938365599e-07, + "loss": 0.4342, + "step": 7758 + }, + { + "epoch": 2.4826151877133107, + "grad_norm": 0.5383480191230774, + "learning_rate": 8.782181996825484e-07, + "loss": 0.4829, + "step": 7759 + }, + { + "epoch": 2.4829351535836177, + "grad_norm": 0.4855513870716095, + "learning_rate": 8.771645770185433e-07, + "loss": 0.4232, + "step": 7760 + }, + { + "epoch": 2.4832551194539247, + "grad_norm": 0.5087752342224121, + "learning_rate": 8.761115259906289e-07, + "loss": 0.4893, + "step": 7761 + }, + { + "epoch": 2.483575085324232, + "grad_norm": 0.5325577855110168, + "learning_rate": 8.750590467448128e-07, + "loss": 0.4296, + "step": 7762 + }, + { + "epoch": 2.4838950511945392, + "grad_norm": 0.5027719140052795, + "learning_rate": 8.740071394270217e-07, + "loss": 0.4864, + "step": 7763 + }, + { + "epoch": 2.4842150170648463, + "grad_norm": 0.48960229754447937, + "learning_rate": 8.729558041831055e-07, + "loss": 0.4272, + "step": 7764 + }, + { + "epoch": 2.4845349829351537, + "grad_norm": 0.473838210105896, + "learning_rate": 8.719050411588321e-07, + "loss": 0.436, + "step": 7765 + }, + { + "epoch": 2.4848549488054608, + "grad_norm": 0.4734203517436981, + "learning_rate": 8.708548504998898e-07, + "loss": 0.4474, + "step": 7766 + }, + { + "epoch": 2.485174914675768, + "grad_norm": 0.5132763981819153, + "learning_rate": 8.698052323518913e-07, + "loss": 0.4424, + "step": 7767 + }, + { + "epoch": 2.4854948805460753, + "grad_norm": 0.5405433177947998, + "learning_rate": 8.687561868603644e-07, + "loss": 0.4259, + "step": 7768 + }, + { + "epoch": 2.4858148464163823, + "grad_norm": 0.5086086988449097, + "learning_rate": 8.677077141707635e-07, + "loss": 0.4623, + "step": 7769 + }, + { + "epoch": 2.4861348122866893, + "grad_norm": 0.437712162733078, + "learning_rate": 8.666598144284594e-07, + "loss": 0.3828, + "step": 7770 + }, + { + "epoch": 2.486454778156997, + "grad_norm": 0.4647720456123352, + "learning_rate": 8.656124877787437e-07, + "loss": 0.3754, + "step": 7771 + }, + { + "epoch": 2.486774744027304, + "grad_norm": 0.4738596975803375, + "learning_rate": 8.645657343668312e-07, + "loss": 0.4491, + "step": 7772 + }, + { + "epoch": 2.487094709897611, + "grad_norm": 0.49070629477500916, + "learning_rate": 8.635195543378532e-07, + "loss": 0.4705, + "step": 7773 + }, + { + "epoch": 2.4874146757679183, + "grad_norm": 0.5047940611839294, + "learning_rate": 8.624739478368671e-07, + "loss": 0.4486, + "step": 7774 + }, + { + "epoch": 2.4877346416382253, + "grad_norm": 0.4897439479827881, + "learning_rate": 8.61428915008844e-07, + "loss": 0.4371, + "step": 7775 + }, + { + "epoch": 2.4880546075085324, + "grad_norm": 0.48489508032798767, + "learning_rate": 8.603844559986823e-07, + "loss": 0.4098, + "step": 7776 + }, + { + "epoch": 2.48837457337884, + "grad_norm": 0.44628262519836426, + "learning_rate": 8.593405709511954e-07, + "loss": 0.4284, + "step": 7777 + }, + { + "epoch": 2.488694539249147, + "grad_norm": 0.5165276527404785, + "learning_rate": 8.582972600111189e-07, + "loss": 0.4968, + "step": 7778 + }, + { + "epoch": 2.489014505119454, + "grad_norm": 0.5198566913604736, + "learning_rate": 8.572545233231111e-07, + "loss": 0.4104, + "step": 7779 + }, + { + "epoch": 2.489334470989761, + "grad_norm": 0.5088862180709839, + "learning_rate": 8.562123610317457e-07, + "loss": 0.4453, + "step": 7780 + }, + { + "epoch": 2.4896544368600684, + "grad_norm": 0.49031898379325867, + "learning_rate": 8.551707732815234e-07, + "loss": 0.4559, + "step": 7781 + }, + { + "epoch": 2.4899744027303754, + "grad_norm": 0.5151739120483398, + "learning_rate": 8.541297602168591e-07, + "loss": 0.4375, + "step": 7782 + }, + { + "epoch": 2.4902943686006824, + "grad_norm": 0.5170591473579407, + "learning_rate": 8.530893219820907e-07, + "loss": 0.4197, + "step": 7783 + }, + { + "epoch": 2.49061433447099, + "grad_norm": 0.5250839591026306, + "learning_rate": 8.520494587214773e-07, + "loss": 0.46, + "step": 7784 + }, + { + "epoch": 2.490934300341297, + "grad_norm": 0.47464457154273987, + "learning_rate": 8.510101705791957e-07, + "loss": 0.4299, + "step": 7785 + }, + { + "epoch": 2.491254266211604, + "grad_norm": 0.4447919726371765, + "learning_rate": 8.499714576993462e-07, + "loss": 0.4009, + "step": 7786 + }, + { + "epoch": 2.4915742320819114, + "grad_norm": 0.5237582325935364, + "learning_rate": 8.48933320225947e-07, + "loss": 0.4642, + "step": 7787 + }, + { + "epoch": 2.4918941979522184, + "grad_norm": 0.5427510738372803, + "learning_rate": 8.478957583029351e-07, + "loss": 0.5008, + "step": 7788 + }, + { + "epoch": 2.4922141638225255, + "grad_norm": 0.47621169686317444, + "learning_rate": 8.468587720741728e-07, + "loss": 0.38, + "step": 7789 + }, + { + "epoch": 2.492534129692833, + "grad_norm": 0.517081618309021, + "learning_rate": 8.458223616834371e-07, + "loss": 0.4361, + "step": 7790 + }, + { + "epoch": 2.49285409556314, + "grad_norm": 0.537865400314331, + "learning_rate": 8.447865272744299e-07, + "loss": 0.4542, + "step": 7791 + }, + { + "epoch": 2.493174061433447, + "grad_norm": 0.5118927955627441, + "learning_rate": 8.437512689907696e-07, + "loss": 0.4356, + "step": 7792 + }, + { + "epoch": 2.493494027303754, + "grad_norm": 0.4743349552154541, + "learning_rate": 8.42716586975994e-07, + "loss": 0.4347, + "step": 7793 + }, + { + "epoch": 2.4938139931740615, + "grad_norm": 0.5667382478713989, + "learning_rate": 8.416824813735669e-07, + "loss": 0.511, + "step": 7794 + }, + { + "epoch": 2.4941339590443685, + "grad_norm": 0.532633364200592, + "learning_rate": 8.406489523268646e-07, + "loss": 0.4498, + "step": 7795 + }, + { + "epoch": 2.4944539249146755, + "grad_norm": 0.5290338397026062, + "learning_rate": 8.396159999791903e-07, + "loss": 0.4787, + "step": 7796 + }, + { + "epoch": 2.494773890784983, + "grad_norm": 0.495960533618927, + "learning_rate": 8.385836244737628e-07, + "loss": 0.3912, + "step": 7797 + }, + { + "epoch": 2.49509385665529, + "grad_norm": 0.5184521675109863, + "learning_rate": 8.375518259537208e-07, + "loss": 0.4055, + "step": 7798 + }, + { + "epoch": 2.495413822525597, + "grad_norm": 0.5542027950286865, + "learning_rate": 8.365206045621271e-07, + "loss": 0.4543, + "step": 7799 + }, + { + "epoch": 2.4957337883959045, + "grad_norm": 0.5616625547409058, + "learning_rate": 8.354899604419586e-07, + "loss": 0.4486, + "step": 7800 + }, + { + "epoch": 2.4960537542662116, + "grad_norm": 0.4697994291782379, + "learning_rate": 8.344598937361186e-07, + "loss": 0.3706, + "step": 7801 + }, + { + "epoch": 2.4963737201365186, + "grad_norm": 0.505582869052887, + "learning_rate": 8.334304045874248e-07, + "loss": 0.4675, + "step": 7802 + }, + { + "epoch": 2.496693686006826, + "grad_norm": 0.5572977066040039, + "learning_rate": 8.324014931386187e-07, + "loss": 0.4193, + "step": 7803 + }, + { + "epoch": 2.497013651877133, + "grad_norm": 0.4983901381492615, + "learning_rate": 8.313731595323593e-07, + "loss": 0.4471, + "step": 7804 + }, + { + "epoch": 2.49733361774744, + "grad_norm": 0.4947405159473419, + "learning_rate": 8.303454039112252e-07, + "loss": 0.4262, + "step": 7805 + }, + { + "epoch": 2.4976535836177476, + "grad_norm": 0.5503841042518616, + "learning_rate": 8.293182264177185e-07, + "loss": 0.4817, + "step": 7806 + }, + { + "epoch": 2.4979735494880546, + "grad_norm": 0.47878727316856384, + "learning_rate": 8.282916271942553e-07, + "loss": 0.4825, + "step": 7807 + }, + { + "epoch": 2.4982935153583616, + "grad_norm": 0.4644586145877838, + "learning_rate": 8.272656063831785e-07, + "loss": 0.4158, + "step": 7808 + }, + { + "epoch": 2.498613481228669, + "grad_norm": 0.5075016021728516, + "learning_rate": 8.262401641267454e-07, + "loss": 0.4104, + "step": 7809 + }, + { + "epoch": 2.498933447098976, + "grad_norm": 0.5721554756164551, + "learning_rate": 8.252153005671332e-07, + "loss": 0.4476, + "step": 7810 + }, + { + "epoch": 2.499253412969283, + "grad_norm": 0.5362179279327393, + "learning_rate": 8.241910158464428e-07, + "loss": 0.4312, + "step": 7811 + }, + { + "epoch": 2.4995733788395906, + "grad_norm": 0.4949038624763489, + "learning_rate": 8.231673101066911e-07, + "loss": 0.4742, + "step": 7812 + }, + { + "epoch": 2.4998933447098977, + "grad_norm": 0.45464375615119934, + "learning_rate": 8.221441834898175e-07, + "loss": 0.4002, + "step": 7813 + }, + { + "epoch": 2.5002133105802047, + "grad_norm": 0.4683190882205963, + "learning_rate": 8.21121636137679e-07, + "loss": 0.4823, + "step": 7814 + }, + { + "epoch": 2.500533276450512, + "grad_norm": 0.4764581322669983, + "learning_rate": 8.20099668192052e-07, + "loss": 0.4656, + "step": 7815 + }, + { + "epoch": 2.500853242320819, + "grad_norm": 0.4716465175151825, + "learning_rate": 8.190782797946356e-07, + "loss": 0.4191, + "step": 7816 + }, + { + "epoch": 2.501173208191126, + "grad_norm": 0.5018923878669739, + "learning_rate": 8.180574710870442e-07, + "loss": 0.4141, + "step": 7817 + }, + { + "epoch": 2.5014931740614337, + "grad_norm": 0.5134195685386658, + "learning_rate": 8.170372422108164e-07, + "loss": 0.4714, + "step": 7818 + }, + { + "epoch": 2.5018131399317407, + "grad_norm": 0.4809771478176117, + "learning_rate": 8.160175933074077e-07, + "loss": 0.404, + "step": 7819 + }, + { + "epoch": 2.5021331058020477, + "grad_norm": 0.5265135765075684, + "learning_rate": 8.149985245181919e-07, + "loss": 0.4516, + "step": 7820 + }, + { + "epoch": 2.502453071672355, + "grad_norm": 0.5177709460258484, + "learning_rate": 8.139800359844663e-07, + "loss": 0.4256, + "step": 7821 + }, + { + "epoch": 2.5027730375426622, + "grad_norm": 0.5242189764976501, + "learning_rate": 8.129621278474437e-07, + "loss": 0.4616, + "step": 7822 + }, + { + "epoch": 2.5030930034129693, + "grad_norm": 0.4369068145751953, + "learning_rate": 8.1194480024826e-07, + "loss": 0.4181, + "step": 7823 + }, + { + "epoch": 2.5034129692832767, + "grad_norm": 0.5571264624595642, + "learning_rate": 8.109280533279685e-07, + "loss": 0.4737, + "step": 7824 + }, + { + "epoch": 2.5037329351535837, + "grad_norm": 0.5230574607849121, + "learning_rate": 8.099118872275403e-07, + "loss": 0.4011, + "step": 7825 + }, + { + "epoch": 2.5040529010238908, + "grad_norm": 0.4745146632194519, + "learning_rate": 8.088963020878715e-07, + "loss": 0.4128, + "step": 7826 + }, + { + "epoch": 2.504372866894198, + "grad_norm": 0.5156494379043579, + "learning_rate": 8.078812980497708e-07, + "loss": 0.4402, + "step": 7827 + }, + { + "epoch": 2.5046928327645053, + "grad_norm": 0.5434547662734985, + "learning_rate": 8.068668752539726e-07, + "loss": 0.4597, + "step": 7828 + }, + { + "epoch": 2.5050127986348123, + "grad_norm": 0.5473108291625977, + "learning_rate": 8.058530338411269e-07, + "loss": 0.4405, + "step": 7829 + }, + { + "epoch": 2.5053327645051193, + "grad_norm": 0.5330970287322998, + "learning_rate": 8.048397739518021e-07, + "loss": 0.5092, + "step": 7830 + }, + { + "epoch": 2.5056527303754264, + "grad_norm": 0.5310120582580566, + "learning_rate": 8.03827095726491e-07, + "loss": 0.4714, + "step": 7831 + }, + { + "epoch": 2.505972696245734, + "grad_norm": 0.5227869749069214, + "learning_rate": 8.028149993056e-07, + "loss": 0.3721, + "step": 7832 + }, + { + "epoch": 2.506292662116041, + "grad_norm": 0.5100311040878296, + "learning_rate": 8.018034848294598e-07, + "loss": 0.4337, + "step": 7833 + }, + { + "epoch": 2.506612627986348, + "grad_norm": 0.5334123373031616, + "learning_rate": 8.007925524383175e-07, + "loss": 0.4494, + "step": 7834 + }, + { + "epoch": 2.5069325938566553, + "grad_norm": 0.5083934664726257, + "learning_rate": 7.997822022723378e-07, + "loss": 0.4058, + "step": 7835 + }, + { + "epoch": 2.5072525597269624, + "grad_norm": 0.47573235630989075, + "learning_rate": 7.987724344716091e-07, + "loss": 0.4076, + "step": 7836 + }, + { + "epoch": 2.5075725255972694, + "grad_norm": 0.5759677886962891, + "learning_rate": 7.977632491761372e-07, + "loss": 0.438, + "step": 7837 + }, + { + "epoch": 2.507892491467577, + "grad_norm": 1.110116958618164, + "learning_rate": 7.967546465258469e-07, + "loss": 0.431, + "step": 7838 + }, + { + "epoch": 2.508212457337884, + "grad_norm": 0.5045497417449951, + "learning_rate": 7.957466266605812e-07, + "loss": 0.4415, + "step": 7839 + }, + { + "epoch": 2.508532423208191, + "grad_norm": 0.48604580760002136, + "learning_rate": 7.947391897201024e-07, + "loss": 0.3933, + "step": 7840 + }, + { + "epoch": 2.5088523890784984, + "grad_norm": 0.5094166994094849, + "learning_rate": 7.937323358440935e-07, + "loss": 0.4343, + "step": 7841 + }, + { + "epoch": 2.5091723549488054, + "grad_norm": 0.538398265838623, + "learning_rate": 7.92726065172158e-07, + "loss": 0.4715, + "step": 7842 + }, + { + "epoch": 2.5094923208191124, + "grad_norm": 0.5149754881858826, + "learning_rate": 7.917203778438154e-07, + "loss": 0.4317, + "step": 7843 + }, + { + "epoch": 2.50981228668942, + "grad_norm": 0.48621267080307007, + "learning_rate": 7.907152739985047e-07, + "loss": 0.4721, + "step": 7844 + }, + { + "epoch": 2.510132252559727, + "grad_norm": 0.4904319643974304, + "learning_rate": 7.897107537755844e-07, + "loss": 0.4281, + "step": 7845 + }, + { + "epoch": 2.510452218430034, + "grad_norm": 0.5282472372055054, + "learning_rate": 7.887068173143325e-07, + "loss": 0.4568, + "step": 7846 + }, + { + "epoch": 2.5107721843003414, + "grad_norm": 0.49507468938827515, + "learning_rate": 7.877034647539483e-07, + "loss": 0.4193, + "step": 7847 + }, + { + "epoch": 2.5110921501706485, + "grad_norm": 0.4763088524341583, + "learning_rate": 7.867006962335466e-07, + "loss": 0.4008, + "step": 7848 + }, + { + "epoch": 2.5114121160409555, + "grad_norm": 0.47316059470176697, + "learning_rate": 7.856985118921618e-07, + "loss": 0.4714, + "step": 7849 + }, + { + "epoch": 2.511732081911263, + "grad_norm": 0.46475642919540405, + "learning_rate": 7.846969118687464e-07, + "loss": 0.4173, + "step": 7850 + }, + { + "epoch": 2.51205204778157, + "grad_norm": 0.5356907248497009, + "learning_rate": 7.836958963021762e-07, + "loss": 0.4541, + "step": 7851 + }, + { + "epoch": 2.512372013651877, + "grad_norm": 0.5031693577766418, + "learning_rate": 7.826954653312429e-07, + "loss": 0.4761, + "step": 7852 + }, + { + "epoch": 2.5126919795221845, + "grad_norm": 0.49371716380119324, + "learning_rate": 7.816956190946573e-07, + "loss": 0.4227, + "step": 7853 + }, + { + "epoch": 2.5130119453924915, + "grad_norm": 0.527224600315094, + "learning_rate": 7.80696357731049e-07, + "loss": 0.4879, + "step": 7854 + }, + { + "epoch": 2.5133319112627985, + "grad_norm": 0.46800824999809265, + "learning_rate": 7.796976813789647e-07, + "loss": 0.4489, + "step": 7855 + }, + { + "epoch": 2.513651877133106, + "grad_norm": 0.4291780889034271, + "learning_rate": 7.786995901768746e-07, + "loss": 0.399, + "step": 7856 + }, + { + "epoch": 2.513971843003413, + "grad_norm": 0.5125072002410889, + "learning_rate": 7.777020842631655e-07, + "loss": 0.5024, + "step": 7857 + }, + { + "epoch": 2.51429180887372, + "grad_norm": 0.5313035249710083, + "learning_rate": 7.767051637761424e-07, + "loss": 0.4516, + "step": 7858 + }, + { + "epoch": 2.5146117747440275, + "grad_norm": 0.4871886968612671, + "learning_rate": 7.757088288540288e-07, + "loss": 0.4269, + "step": 7859 + }, + { + "epoch": 2.5149317406143346, + "grad_norm": 0.5050264000892639, + "learning_rate": 7.747130796349672e-07, + "loss": 0.3796, + "step": 7860 + }, + { + "epoch": 2.5152517064846416, + "grad_norm": 0.4975477457046509, + "learning_rate": 7.737179162570196e-07, + "loss": 0.4757, + "step": 7861 + }, + { + "epoch": 2.515571672354949, + "grad_norm": 0.561652660369873, + "learning_rate": 7.72723338858169e-07, + "loss": 0.4734, + "step": 7862 + }, + { + "epoch": 2.515891638225256, + "grad_norm": 0.5012123584747314, + "learning_rate": 7.717293475763126e-07, + "loss": 0.4151, + "step": 7863 + }, + { + "epoch": 2.516211604095563, + "grad_norm": 0.5105333924293518, + "learning_rate": 7.707359425492694e-07, + "loss": 0.4373, + "step": 7864 + }, + { + "epoch": 2.5165315699658706, + "grad_norm": 0.48380863666534424, + "learning_rate": 7.69743123914774e-07, + "loss": 0.4431, + "step": 7865 + }, + { + "epoch": 2.5168515358361776, + "grad_norm": 0.46601566672325134, + "learning_rate": 7.687508918104841e-07, + "loss": 0.4391, + "step": 7866 + }, + { + "epoch": 2.5171715017064846, + "grad_norm": 0.5079662203788757, + "learning_rate": 7.677592463739741e-07, + "loss": 0.4003, + "step": 7867 + }, + { + "epoch": 2.5174914675767917, + "grad_norm": 0.4948737323284149, + "learning_rate": 7.667681877427363e-07, + "loss": 0.4707, + "step": 7868 + }, + { + "epoch": 2.517811433447099, + "grad_norm": 0.4404037296772003, + "learning_rate": 7.657777160541818e-07, + "loss": 0.4013, + "step": 7869 + }, + { + "epoch": 2.518131399317406, + "grad_norm": 0.47359374165534973, + "learning_rate": 7.647878314456403e-07, + "loss": 0.4264, + "step": 7870 + }, + { + "epoch": 2.518451365187713, + "grad_norm": 0.5073278546333313, + "learning_rate": 7.637985340543602e-07, + "loss": 0.4524, + "step": 7871 + }, + { + "epoch": 2.51877133105802, + "grad_norm": 0.44690725207328796, + "learning_rate": 7.628098240175113e-07, + "loss": 0.3978, + "step": 7872 + }, + { + "epoch": 2.5190912969283277, + "grad_norm": 0.4936218559741974, + "learning_rate": 7.618217014721779e-07, + "loss": 0.4774, + "step": 7873 + }, + { + "epoch": 2.5194112627986347, + "grad_norm": 0.5053908228874207, + "learning_rate": 7.608341665553637e-07, + "loss": 0.4424, + "step": 7874 + }, + { + "epoch": 2.5197312286689417, + "grad_norm": 0.4724701941013336, + "learning_rate": 7.598472194039913e-07, + "loss": 0.3928, + "step": 7875 + }, + { + "epoch": 2.520051194539249, + "grad_norm": 0.5237969160079956, + "learning_rate": 7.58860860154903e-07, + "loss": 0.4649, + "step": 7876 + }, + { + "epoch": 2.520371160409556, + "grad_norm": 0.5442495346069336, + "learning_rate": 7.5787508894486e-07, + "loss": 0.4657, + "step": 7877 + }, + { + "epoch": 2.5206911262798632, + "grad_norm": 0.4638236165046692, + "learning_rate": 7.56889905910539e-07, + "loss": 0.4268, + "step": 7878 + }, + { + "epoch": 2.5210110921501707, + "grad_norm": 0.5017217397689819, + "learning_rate": 7.559053111885373e-07, + "loss": 0.4371, + "step": 7879 + }, + { + "epoch": 2.5213310580204777, + "grad_norm": 0.4847548305988312, + "learning_rate": 7.549213049153687e-07, + "loss": 0.4173, + "step": 7880 + }, + { + "epoch": 2.5216510238907848, + "grad_norm": 0.4981611669063568, + "learning_rate": 7.539378872274677e-07, + "loss": 0.519, + "step": 7881 + }, + { + "epoch": 2.5219709897610922, + "grad_norm": 0.41899481415748596, + "learning_rate": 7.529550582611883e-07, + "loss": 0.4235, + "step": 7882 + }, + { + "epoch": 2.5222909556313993, + "grad_norm": 0.47802966833114624, + "learning_rate": 7.519728181527997e-07, + "loss": 0.4436, + "step": 7883 + }, + { + "epoch": 2.5226109215017063, + "grad_norm": 0.5048873424530029, + "learning_rate": 7.5099116703849e-07, + "loss": 0.4518, + "step": 7884 + }, + { + "epoch": 2.5229308873720138, + "grad_norm": 0.43372148275375366, + "learning_rate": 7.500101050543657e-07, + "loss": 0.3671, + "step": 7885 + }, + { + "epoch": 2.523250853242321, + "grad_norm": 0.49529823660850525, + "learning_rate": 7.490296323364532e-07, + "loss": 0.4827, + "step": 7886 + }, + { + "epoch": 2.523570819112628, + "grad_norm": 0.49398353695869446, + "learning_rate": 7.480497490206973e-07, + "loss": 0.4172, + "step": 7887 + }, + { + "epoch": 2.5238907849829353, + "grad_norm": 0.4616610109806061, + "learning_rate": 7.470704552429592e-07, + "loss": 0.4183, + "step": 7888 + }, + { + "epoch": 2.5242107508532423, + "grad_norm": 0.532124936580658, + "learning_rate": 7.460917511390186e-07, + "loss": 0.4552, + "step": 7889 + }, + { + "epoch": 2.5245307167235493, + "grad_norm": 0.4939459562301636, + "learning_rate": 7.451136368445727e-07, + "loss": 0.4066, + "step": 7890 + }, + { + "epoch": 2.524850682593857, + "grad_norm": 0.49357783794403076, + "learning_rate": 7.441361124952406e-07, + "loss": 0.4387, + "step": 7891 + }, + { + "epoch": 2.525170648464164, + "grad_norm": 0.48628488183021545, + "learning_rate": 7.431591782265574e-07, + "loss": 0.4396, + "step": 7892 + }, + { + "epoch": 2.525490614334471, + "grad_norm": 0.48584863543510437, + "learning_rate": 7.421828341739751e-07, + "loss": 0.3749, + "step": 7893 + }, + { + "epoch": 2.5258105802047783, + "grad_norm": 0.4655524492263794, + "learning_rate": 7.41207080472865e-07, + "loss": 0.4627, + "step": 7894 + }, + { + "epoch": 2.5261305460750854, + "grad_norm": 0.540604829788208, + "learning_rate": 7.402319172585154e-07, + "loss": 0.4555, + "step": 7895 + }, + { + "epoch": 2.5264505119453924, + "grad_norm": 0.4794887900352478, + "learning_rate": 7.392573446661355e-07, + "loss": 0.3782, + "step": 7896 + }, + { + "epoch": 2.5267704778157, + "grad_norm": 0.4764775335788727, + "learning_rate": 7.382833628308511e-07, + "loss": 0.4542, + "step": 7897 + }, + { + "epoch": 2.527090443686007, + "grad_norm": 0.4774472415447235, + "learning_rate": 7.373099718877059e-07, + "loss": 0.4158, + "step": 7898 + }, + { + "epoch": 2.527410409556314, + "grad_norm": 0.4650486409664154, + "learning_rate": 7.363371719716611e-07, + "loss": 0.455, + "step": 7899 + }, + { + "epoch": 2.5277303754266214, + "grad_norm": 0.5249002575874329, + "learning_rate": 7.353649632175957e-07, + "loss": 0.4391, + "step": 7900 + }, + { + "epoch": 2.5280503412969284, + "grad_norm": 0.4526771605014801, + "learning_rate": 7.343933457603086e-07, + "loss": 0.4209, + "step": 7901 + }, + { + "epoch": 2.5283703071672354, + "grad_norm": 0.4585656225681305, + "learning_rate": 7.334223197345169e-07, + "loss": 0.466, + "step": 7902 + }, + { + "epoch": 2.528690273037543, + "grad_norm": 0.5290406942367554, + "learning_rate": 7.324518852748535e-07, + "loss": 0.4633, + "step": 7903 + }, + { + "epoch": 2.52901023890785, + "grad_norm": 0.5417910218238831, + "learning_rate": 7.314820425158687e-07, + "loss": 0.466, + "step": 7904 + }, + { + "epoch": 2.529330204778157, + "grad_norm": 0.5115374326705933, + "learning_rate": 7.305127915920351e-07, + "loss": 0.471, + "step": 7905 + }, + { + "epoch": 2.5296501706484644, + "grad_norm": 0.46070459485054016, + "learning_rate": 7.295441326377384e-07, + "loss": 0.3598, + "step": 7906 + }, + { + "epoch": 2.5299701365187715, + "grad_norm": 0.4890165328979492, + "learning_rate": 7.285760657872859e-07, + "loss": 0.5009, + "step": 7907 + }, + { + "epoch": 2.5302901023890785, + "grad_norm": 0.4732837975025177, + "learning_rate": 7.276085911749009e-07, + "loss": 0.459, + "step": 7908 + }, + { + "epoch": 2.5306100682593855, + "grad_norm": 0.47583815455436707, + "learning_rate": 7.266417089347233e-07, + "loss": 0.4212, + "step": 7909 + }, + { + "epoch": 2.530930034129693, + "grad_norm": 0.48661455512046814, + "learning_rate": 7.256754192008148e-07, + "loss": 0.4319, + "step": 7910 + }, + { + "epoch": 2.53125, + "grad_norm": 0.5234128832817078, + "learning_rate": 7.247097221071509e-07, + "loss": 0.4463, + "step": 7911 + }, + { + "epoch": 2.531569965870307, + "grad_norm": 0.4903889000415802, + "learning_rate": 7.237446177876278e-07, + "loss": 0.375, + "step": 7912 + }, + { + "epoch": 2.5318899317406145, + "grad_norm": 0.5294584631919861, + "learning_rate": 7.227801063760581e-07, + "loss": 0.4596, + "step": 7913 + }, + { + "epoch": 2.5322098976109215, + "grad_norm": 0.4898788034915924, + "learning_rate": 7.218161880061714e-07, + "loss": 0.4073, + "step": 7914 + }, + { + "epoch": 2.5325298634812285, + "grad_norm": 0.4349392056465149, + "learning_rate": 7.208528628116179e-07, + "loss": 0.4212, + "step": 7915 + }, + { + "epoch": 2.5328498293515356, + "grad_norm": 0.4884548783302307, + "learning_rate": 7.198901309259615e-07, + "loss": 0.4469, + "step": 7916 + }, + { + "epoch": 2.533169795221843, + "grad_norm": 0.5282070636749268, + "learning_rate": 7.189279924826892e-07, + "loss": 0.4284, + "step": 7917 + }, + { + "epoch": 2.53348976109215, + "grad_norm": 0.5929446816444397, + "learning_rate": 7.17966447615201e-07, + "loss": 0.4968, + "step": 7918 + }, + { + "epoch": 2.533809726962457, + "grad_norm": 0.5412628054618835, + "learning_rate": 7.170054964568146e-07, + "loss": 0.4404, + "step": 7919 + }, + { + "epoch": 2.5341296928327646, + "grad_norm": 0.5096564292907715, + "learning_rate": 7.160451391407691e-07, + "loss": 0.5102, + "step": 7920 + }, + { + "epoch": 2.5344496587030716, + "grad_norm": 0.46906861662864685, + "learning_rate": 7.150853758002196e-07, + "loss": 0.4025, + "step": 7921 + }, + { + "epoch": 2.5347696245733786, + "grad_norm": 0.509339451789856, + "learning_rate": 7.14126206568238e-07, + "loss": 0.4543, + "step": 7922 + }, + { + "epoch": 2.535089590443686, + "grad_norm": 0.44657838344573975, + "learning_rate": 7.131676315778136e-07, + "loss": 0.4469, + "step": 7923 + }, + { + "epoch": 2.535409556313993, + "grad_norm": 0.4389392137527466, + "learning_rate": 7.122096509618537e-07, + "loss": 0.3871, + "step": 7924 + }, + { + "epoch": 2.5357295221843, + "grad_norm": 0.48517391085624695, + "learning_rate": 7.112522648531833e-07, + "loss": 0.449, + "step": 7925 + }, + { + "epoch": 2.5360494880546076, + "grad_norm": 0.49374955892562866, + "learning_rate": 7.102954733845474e-07, + "loss": 0.359, + "step": 7926 + }, + { + "epoch": 2.5363694539249146, + "grad_norm": 0.5568493604660034, + "learning_rate": 7.093392766886048e-07, + "loss": 0.4509, + "step": 7927 + }, + { + "epoch": 2.5366894197952217, + "grad_norm": 0.473632276058197, + "learning_rate": 7.083836748979328e-07, + "loss": 0.4456, + "step": 7928 + }, + { + "epoch": 2.537009385665529, + "grad_norm": 0.4755510091781616, + "learning_rate": 7.074286681450266e-07, + "loss": 0.4893, + "step": 7929 + }, + { + "epoch": 2.537329351535836, + "grad_norm": 0.4428599774837494, + "learning_rate": 7.064742565622995e-07, + "loss": 0.401, + "step": 7930 + }, + { + "epoch": 2.537649317406143, + "grad_norm": 0.4869743287563324, + "learning_rate": 7.055204402820831e-07, + "loss": 0.4499, + "step": 7931 + }, + { + "epoch": 2.5379692832764507, + "grad_norm": 0.44322407245635986, + "learning_rate": 7.045672194366238e-07, + "loss": 0.4087, + "step": 7932 + }, + { + "epoch": 2.5382892491467577, + "grad_norm": 0.5037073493003845, + "learning_rate": 7.036145941580874e-07, + "loss": 0.4724, + "step": 7933 + }, + { + "epoch": 2.5386092150170647, + "grad_norm": 0.5240101218223572, + "learning_rate": 7.026625645785551e-07, + "loss": 0.4887, + "step": 7934 + }, + { + "epoch": 2.538929180887372, + "grad_norm": 0.46680012345314026, + "learning_rate": 7.017111308300278e-07, + "loss": 0.4437, + "step": 7935 + }, + { + "epoch": 2.539249146757679, + "grad_norm": 0.45819762349128723, + "learning_rate": 7.00760293044424e-07, + "loss": 0.4034, + "step": 7936 + }, + { + "epoch": 2.5395691126279862, + "grad_norm": 0.47192758321762085, + "learning_rate": 6.998100513535782e-07, + "loss": 0.4662, + "step": 7937 + }, + { + "epoch": 2.5398890784982937, + "grad_norm": 0.44931477308273315, + "learning_rate": 6.988604058892418e-07, + "loss": 0.4233, + "step": 7938 + }, + { + "epoch": 2.5402090443686007, + "grad_norm": 0.4655084013938904, + "learning_rate": 6.979113567830837e-07, + "loss": 0.3816, + "step": 7939 + }, + { + "epoch": 2.5405290102389078, + "grad_norm": 0.5024960041046143, + "learning_rate": 6.969629041666914e-07, + "loss": 0.4849, + "step": 7940 + }, + { + "epoch": 2.5408489761092152, + "grad_norm": 0.5373795032501221, + "learning_rate": 6.9601504817157e-07, + "loss": 0.427, + "step": 7941 + }, + { + "epoch": 2.5411689419795223, + "grad_norm": 0.49928146600723267, + "learning_rate": 6.950677889291402e-07, + "loss": 0.4607, + "step": 7942 + }, + { + "epoch": 2.5414889078498293, + "grad_norm": 0.5113884210586548, + "learning_rate": 6.941211265707393e-07, + "loss": 0.4202, + "step": 7943 + }, + { + "epoch": 2.5418088737201368, + "grad_norm": 0.5413162112236023, + "learning_rate": 6.93175061227625e-07, + "loss": 0.4412, + "step": 7944 + }, + { + "epoch": 2.542128839590444, + "grad_norm": 0.5174480080604553, + "learning_rate": 6.922295930309691e-07, + "loss": 0.4355, + "step": 7945 + }, + { + "epoch": 2.542448805460751, + "grad_norm": 0.5377246141433716, + "learning_rate": 6.912847221118635e-07, + "loss": 0.4573, + "step": 7946 + }, + { + "epoch": 2.5427687713310583, + "grad_norm": 0.5317425727844238, + "learning_rate": 6.903404486013149e-07, + "loss": 0.4932, + "step": 7947 + }, + { + "epoch": 2.5430887372013653, + "grad_norm": 0.48629119992256165, + "learning_rate": 6.893967726302464e-07, + "loss": 0.3659, + "step": 7948 + }, + { + "epoch": 2.5434087030716723, + "grad_norm": 0.4948200583457947, + "learning_rate": 6.884536943295023e-07, + "loss": 0.4538, + "step": 7949 + }, + { + "epoch": 2.54372866894198, + "grad_norm": 0.47697150707244873, + "learning_rate": 6.87511213829839e-07, + "loss": 0.4669, + "step": 7950 + }, + { + "epoch": 2.544048634812287, + "grad_norm": 0.5171055197715759, + "learning_rate": 6.865693312619354e-07, + "loss": 0.4483, + "step": 7951 + }, + { + "epoch": 2.544368600682594, + "grad_norm": 0.46479135751724243, + "learning_rate": 6.856280467563831e-07, + "loss": 0.3814, + "step": 7952 + }, + { + "epoch": 2.544688566552901, + "grad_norm": 0.5061951279640198, + "learning_rate": 6.84687360443691e-07, + "loss": 0.4685, + "step": 7953 + }, + { + "epoch": 2.5450085324232083, + "grad_norm": 0.45566174387931824, + "learning_rate": 6.837472724542893e-07, + "loss": 0.4357, + "step": 7954 + }, + { + "epoch": 2.5453284982935154, + "grad_norm": 0.44831565022468567, + "learning_rate": 6.828077829185198e-07, + "loss": 0.4058, + "step": 7955 + }, + { + "epoch": 2.5456484641638224, + "grad_norm": 0.507569432258606, + "learning_rate": 6.818688919666461e-07, + "loss": 0.4659, + "step": 7956 + }, + { + "epoch": 2.5459684300341294, + "grad_norm": 0.4822544455528259, + "learning_rate": 6.809305997288451e-07, + "loss": 0.4137, + "step": 7957 + }, + { + "epoch": 2.546288395904437, + "grad_norm": 0.4946458339691162, + "learning_rate": 6.799929063352112e-07, + "loss": 0.419, + "step": 7958 + }, + { + "epoch": 2.546608361774744, + "grad_norm": 0.5048835277557373, + "learning_rate": 6.790558119157597e-07, + "loss": 0.4896, + "step": 7959 + }, + { + "epoch": 2.546928327645051, + "grad_norm": 0.4941258132457733, + "learning_rate": 6.781193166004163e-07, + "loss": 0.4414, + "step": 7960 + }, + { + "epoch": 2.5472482935153584, + "grad_norm": 0.4280059039592743, + "learning_rate": 6.771834205190303e-07, + "loss": 0.4074, + "step": 7961 + }, + { + "epoch": 2.5475682593856654, + "grad_norm": 0.5590117573738098, + "learning_rate": 6.762481238013635e-07, + "loss": 0.5315, + "step": 7962 + }, + { + "epoch": 2.5478882252559725, + "grad_norm": 0.46806591749191284, + "learning_rate": 6.75313426577095e-07, + "loss": 0.3526, + "step": 7963 + }, + { + "epoch": 2.54820819112628, + "grad_norm": 0.5423508882522583, + "learning_rate": 6.743793289758238e-07, + "loss": 0.4518, + "step": 7964 + }, + { + "epoch": 2.548528156996587, + "grad_norm": 0.5363116264343262, + "learning_rate": 6.734458311270609e-07, + "loss": 0.4359, + "step": 7965 + }, + { + "epoch": 2.548848122866894, + "grad_norm": 0.4925404489040375, + "learning_rate": 6.725129331602398e-07, + "loss": 0.4212, + "step": 7966 + }, + { + "epoch": 2.5491680887372015, + "grad_norm": 0.5390374660491943, + "learning_rate": 6.715806352047072e-07, + "loss": 0.4659, + "step": 7967 + }, + { + "epoch": 2.5494880546075085, + "grad_norm": 0.5575571060180664, + "learning_rate": 6.706489373897251e-07, + "loss": 0.4608, + "step": 7968 + }, + { + "epoch": 2.5498080204778155, + "grad_norm": 0.45139390230178833, + "learning_rate": 6.697178398444771e-07, + "loss": 0.4012, + "step": 7969 + }, + { + "epoch": 2.550127986348123, + "grad_norm": 0.49906036257743835, + "learning_rate": 6.68787342698059e-07, + "loss": 0.4613, + "step": 7970 + }, + { + "epoch": 2.55044795221843, + "grad_norm": 0.5017614960670471, + "learning_rate": 6.678574460794879e-07, + "loss": 0.4129, + "step": 7971 + }, + { + "epoch": 2.550767918088737, + "grad_norm": 0.48051193356513977, + "learning_rate": 6.669281501176933e-07, + "loss": 0.3836, + "step": 7972 + }, + { + "epoch": 2.5510878839590445, + "grad_norm": 0.5058014392852783, + "learning_rate": 6.659994549415227e-07, + "loss": 0.5042, + "step": 7973 + }, + { + "epoch": 2.5514078498293515, + "grad_norm": 0.4778083264827728, + "learning_rate": 6.650713606797426e-07, + "loss": 0.4889, + "step": 7974 + }, + { + "epoch": 2.5517278156996586, + "grad_norm": 0.42727649211883545, + "learning_rate": 6.641438674610318e-07, + "loss": 0.3505, + "step": 7975 + }, + { + "epoch": 2.552047781569966, + "grad_norm": 0.5494928359985352, + "learning_rate": 6.632169754139917e-07, + "loss": 0.5076, + "step": 7976 + }, + { + "epoch": 2.552367747440273, + "grad_norm": 0.4783819913864136, + "learning_rate": 6.622906846671339e-07, + "loss": 0.4863, + "step": 7977 + }, + { + "epoch": 2.55268771331058, + "grad_norm": 0.5169709324836731, + "learning_rate": 6.613649953488921e-07, + "loss": 0.4364, + "step": 7978 + }, + { + "epoch": 2.5530076791808876, + "grad_norm": 0.475924551486969, + "learning_rate": 6.604399075876133e-07, + "loss": 0.406, + "step": 7979 + }, + { + "epoch": 2.5533276450511946, + "grad_norm": 0.5216423869132996, + "learning_rate": 6.595154215115607e-07, + "loss": 0.4441, + "step": 7980 + }, + { + "epoch": 2.5536476109215016, + "grad_norm": 0.516577959060669, + "learning_rate": 6.585915372489177e-07, + "loss": 0.4712, + "step": 7981 + }, + { + "epoch": 2.553967576791809, + "grad_norm": 0.508666455745697, + "learning_rate": 6.576682549277797e-07, + "loss": 0.4248, + "step": 7982 + }, + { + "epoch": 2.554287542662116, + "grad_norm": 0.5177512168884277, + "learning_rate": 6.567455746761636e-07, + "loss": 0.4926, + "step": 7983 + }, + { + "epoch": 2.554607508532423, + "grad_norm": 0.4615013003349304, + "learning_rate": 6.558234966219984e-07, + "loss": 0.4214, + "step": 7984 + }, + { + "epoch": 2.5549274744027306, + "grad_norm": 0.47998058795928955, + "learning_rate": 6.549020208931306e-07, + "loss": 0.4224, + "step": 7985 + }, + { + "epoch": 2.5552474402730376, + "grad_norm": 0.4983801245689392, + "learning_rate": 6.539811476173258e-07, + "loss": 0.4008, + "step": 7986 + }, + { + "epoch": 2.5555674061433447, + "grad_norm": 0.5446333885192871, + "learning_rate": 6.530608769222624e-07, + "loss": 0.4695, + "step": 7987 + }, + { + "epoch": 2.555887372013652, + "grad_norm": 0.487994909286499, + "learning_rate": 6.521412089355389e-07, + "loss": 0.4271, + "step": 7988 + }, + { + "epoch": 2.556207337883959, + "grad_norm": 0.4926753044128418, + "learning_rate": 6.512221437846672e-07, + "loss": 0.4037, + "step": 7989 + }, + { + "epoch": 2.556527303754266, + "grad_norm": 0.4782487154006958, + "learning_rate": 6.503036815970759e-07, + "loss": 0.4544, + "step": 7990 + }, + { + "epoch": 2.5568472696245736, + "grad_norm": 0.47307702898979187, + "learning_rate": 6.493858225001131e-07, + "loss": 0.456, + "step": 7991 + }, + { + "epoch": 2.5571672354948807, + "grad_norm": 0.4964895248413086, + "learning_rate": 6.484685666210389e-07, + "loss": 0.4531, + "step": 7992 + }, + { + "epoch": 2.5574872013651877, + "grad_norm": 0.4947316646575928, + "learning_rate": 6.47551914087034e-07, + "loss": 0.4128, + "step": 7993 + }, + { + "epoch": 2.5578071672354947, + "grad_norm": 0.48979294300079346, + "learning_rate": 6.466358650251925e-07, + "loss": 0.4305, + "step": 7994 + }, + { + "epoch": 2.558127133105802, + "grad_norm": 0.4775591790676117, + "learning_rate": 6.457204195625244e-07, + "loss": 0.4633, + "step": 7995 + }, + { + "epoch": 2.5584470989761092, + "grad_norm": 0.45607683062553406, + "learning_rate": 6.448055778259594e-07, + "loss": 0.4346, + "step": 7996 + }, + { + "epoch": 2.5587670648464163, + "grad_norm": 0.45880717039108276, + "learning_rate": 6.438913399423396e-07, + "loss": 0.461, + "step": 7997 + }, + { + "epoch": 2.5590870307167233, + "grad_norm": 0.4997277855873108, + "learning_rate": 6.429777060384268e-07, + "loss": 0.3966, + "step": 7998 + }, + { + "epoch": 2.5594069965870307, + "grad_norm": 0.527934193611145, + "learning_rate": 6.420646762408972e-07, + "loss": 0.4403, + "step": 7999 + }, + { + "epoch": 2.5597269624573378, + "grad_norm": 0.5341289639472961, + "learning_rate": 6.411522506763412e-07, + "loss": 0.4561, + "step": 8000 + }, + { + "epoch": 2.560046928327645, + "grad_norm": 0.47350558638572693, + "learning_rate": 6.402404294712711e-07, + "loss": 0.4201, + "step": 8001 + }, + { + "epoch": 2.5603668941979523, + "grad_norm": 0.4907589554786682, + "learning_rate": 6.393292127521089e-07, + "loss": 0.4745, + "step": 8002 + }, + { + "epoch": 2.5606868600682593, + "grad_norm": 0.4462029039859772, + "learning_rate": 6.384186006451987e-07, + "loss": 0.4005, + "step": 8003 + }, + { + "epoch": 2.5610068259385663, + "grad_norm": 0.5270899534225464, + "learning_rate": 6.375085932767966e-07, + "loss": 0.4886, + "step": 8004 + }, + { + "epoch": 2.561326791808874, + "grad_norm": 0.5296903252601624, + "learning_rate": 6.365991907730751e-07, + "loss": 0.4888, + "step": 8005 + }, + { + "epoch": 2.561646757679181, + "grad_norm": 0.4909511208534241, + "learning_rate": 6.35690393260126e-07, + "loss": 0.4011, + "step": 8006 + }, + { + "epoch": 2.561966723549488, + "grad_norm": 0.4494198262691498, + "learning_rate": 6.347822008639537e-07, + "loss": 0.4451, + "step": 8007 + }, + { + "epoch": 2.5622866894197953, + "grad_norm": 0.5234202742576599, + "learning_rate": 6.338746137104817e-07, + "loss": 0.4932, + "step": 8008 + }, + { + "epoch": 2.5626066552901023, + "grad_norm": 0.5346943736076355, + "learning_rate": 6.329676319255468e-07, + "loss": 0.4351, + "step": 8009 + }, + { + "epoch": 2.5629266211604094, + "grad_norm": 0.46330562233924866, + "learning_rate": 6.320612556349027e-07, + "loss": 0.4454, + "step": 8010 + }, + { + "epoch": 2.563246587030717, + "grad_norm": 0.4751325845718384, + "learning_rate": 6.311554849642198e-07, + "loss": 0.4381, + "step": 8011 + }, + { + "epoch": 2.563566552901024, + "grad_norm": 0.5119494795799255, + "learning_rate": 6.302503200390858e-07, + "loss": 0.4228, + "step": 8012 + }, + { + "epoch": 2.563886518771331, + "grad_norm": 0.5533508658409119, + "learning_rate": 6.293457609850024e-07, + "loss": 0.4703, + "step": 8013 + }, + { + "epoch": 2.5642064846416384, + "grad_norm": 0.5752602219581604, + "learning_rate": 6.284418079273869e-07, + "loss": 0.441, + "step": 8014 + }, + { + "epoch": 2.5645264505119454, + "grad_norm": 0.556894063949585, + "learning_rate": 6.275384609915724e-07, + "loss": 0.4381, + "step": 8015 + }, + { + "epoch": 2.5648464163822524, + "grad_norm": 0.5390867590904236, + "learning_rate": 6.266357203028111e-07, + "loss": 0.4704, + "step": 8016 + }, + { + "epoch": 2.56516638225256, + "grad_norm": 0.5242125988006592, + "learning_rate": 6.257335859862696e-07, + "loss": 0.4442, + "step": 8017 + }, + { + "epoch": 2.565486348122867, + "grad_norm": 0.5321682691574097, + "learning_rate": 6.248320581670281e-07, + "loss": 0.4569, + "step": 8018 + }, + { + "epoch": 2.565806313993174, + "grad_norm": 0.5042716264724731, + "learning_rate": 6.239311369700857e-07, + "loss": 0.4921, + "step": 8019 + }, + { + "epoch": 2.5661262798634814, + "grad_norm": 0.5050718188285828, + "learning_rate": 6.230308225203546e-07, + "loss": 0.4014, + "step": 8020 + }, + { + "epoch": 2.5664462457337884, + "grad_norm": 0.47061100602149963, + "learning_rate": 6.221311149426657e-07, + "loss": 0.3965, + "step": 8021 + }, + { + "epoch": 2.5667662116040955, + "grad_norm": 0.4673781991004944, + "learning_rate": 6.212320143617651e-07, + "loss": 0.4455, + "step": 8022 + }, + { + "epoch": 2.567086177474403, + "grad_norm": 0.45127198100090027, + "learning_rate": 6.203335209023137e-07, + "loss": 0.3997, + "step": 8023 + }, + { + "epoch": 2.56740614334471, + "grad_norm": 0.463192880153656, + "learning_rate": 6.194356346888885e-07, + "loss": 0.4131, + "step": 8024 + }, + { + "epoch": 2.567726109215017, + "grad_norm": 0.4891464114189148, + "learning_rate": 6.185383558459818e-07, + "loss": 0.4509, + "step": 8025 + }, + { + "epoch": 2.5680460750853245, + "grad_norm": 0.5318314433097839, + "learning_rate": 6.176416844980027e-07, + "loss": 0.4168, + "step": 8026 + }, + { + "epoch": 2.5683660409556315, + "grad_norm": 0.5327971577644348, + "learning_rate": 6.167456207692768e-07, + "loss": 0.4366, + "step": 8027 + }, + { + "epoch": 2.5686860068259385, + "grad_norm": 0.4995075762271881, + "learning_rate": 6.158501647840443e-07, + "loss": 0.4467, + "step": 8028 + }, + { + "epoch": 2.569005972696246, + "grad_norm": 0.49919721484184265, + "learning_rate": 6.149553166664607e-07, + "loss": 0.4718, + "step": 8029 + }, + { + "epoch": 2.569325938566553, + "grad_norm": 0.48580458760261536, + "learning_rate": 6.140610765405969e-07, + "loss": 0.4002, + "step": 8030 + }, + { + "epoch": 2.56964590443686, + "grad_norm": 0.5121810436248779, + "learning_rate": 6.131674445304409e-07, + "loss": 0.4802, + "step": 8031 + }, + { + "epoch": 2.5699658703071675, + "grad_norm": 0.4804602563381195, + "learning_rate": 6.122744207598974e-07, + "loss": 0.4165, + "step": 8032 + }, + { + "epoch": 2.5702858361774745, + "grad_norm": 0.5108646154403687, + "learning_rate": 6.113820053527835e-07, + "loss": 0.4041, + "step": 8033 + }, + { + "epoch": 2.5706058020477816, + "grad_norm": 0.5154693126678467, + "learning_rate": 6.104901984328348e-07, + "loss": 0.4377, + "step": 8034 + }, + { + "epoch": 2.5709257679180886, + "grad_norm": 0.5470184087753296, + "learning_rate": 6.095990001236996e-07, + "loss": 0.4881, + "step": 8035 + }, + { + "epoch": 2.571245733788396, + "grad_norm": 0.47853484749794006, + "learning_rate": 6.087084105489449e-07, + "loss": 0.3808, + "step": 8036 + }, + { + "epoch": 2.571565699658703, + "grad_norm": 0.43800899386405945, + "learning_rate": 6.078184298320527e-07, + "loss": 0.3837, + "step": 8037 + }, + { + "epoch": 2.57188566552901, + "grad_norm": 0.47548708319664, + "learning_rate": 6.069290580964194e-07, + "loss": 0.4467, + "step": 8038 + }, + { + "epoch": 2.5722056313993176, + "grad_norm": 0.49081090092658997, + "learning_rate": 6.060402954653571e-07, + "loss": 0.3955, + "step": 8039 + }, + { + "epoch": 2.5725255972696246, + "grad_norm": 0.5398766398429871, + "learning_rate": 6.051521420620932e-07, + "loss": 0.4942, + "step": 8040 + }, + { + "epoch": 2.5728455631399316, + "grad_norm": 0.47375577688217163, + "learning_rate": 6.042645980097717e-07, + "loss": 0.4444, + "step": 8041 + }, + { + "epoch": 2.5731655290102387, + "grad_norm": 0.4815525710582733, + "learning_rate": 6.033776634314525e-07, + "loss": 0.3972, + "step": 8042 + }, + { + "epoch": 2.573485494880546, + "grad_norm": 0.5203426480293274, + "learning_rate": 6.024913384501097e-07, + "loss": 0.4016, + "step": 8043 + }, + { + "epoch": 2.573805460750853, + "grad_norm": 0.5383763909339905, + "learning_rate": 6.016056231886336e-07, + "loss": 0.4223, + "step": 8044 + }, + { + "epoch": 2.57412542662116, + "grad_norm": 0.5123094916343689, + "learning_rate": 6.007205177698277e-07, + "loss": 0.4369, + "step": 8045 + }, + { + "epoch": 2.5744453924914676, + "grad_norm": 0.4762115180492401, + "learning_rate": 5.998360223164145e-07, + "loss": 0.4577, + "step": 8046 + }, + { + "epoch": 2.5747653583617747, + "grad_norm": 0.483458936214447, + "learning_rate": 5.989521369510321e-07, + "loss": 0.4668, + "step": 8047 + }, + { + "epoch": 2.5750853242320817, + "grad_norm": 0.5269235372543335, + "learning_rate": 5.980688617962299e-07, + "loss": 0.4695, + "step": 8048 + }, + { + "epoch": 2.575405290102389, + "grad_norm": 0.5068848729133606, + "learning_rate": 5.971861969744758e-07, + "loss": 0.4666, + "step": 8049 + }, + { + "epoch": 2.575725255972696, + "grad_norm": 0.4762909710407257, + "learning_rate": 5.963041426081511e-07, + "loss": 0.3967, + "step": 8050 + }, + { + "epoch": 2.576045221843003, + "grad_norm": 0.5501658320426941, + "learning_rate": 5.954226988195544e-07, + "loss": 0.502, + "step": 8051 + }, + { + "epoch": 2.5763651877133107, + "grad_norm": 0.4227347671985626, + "learning_rate": 5.94541865730901e-07, + "loss": 0.3864, + "step": 8052 + }, + { + "epoch": 2.5766851535836177, + "grad_norm": 0.48304030299186707, + "learning_rate": 5.936616434643177e-07, + "loss": 0.4318, + "step": 8053 + }, + { + "epoch": 2.5770051194539247, + "grad_norm": 0.5073933601379395, + "learning_rate": 5.927820321418482e-07, + "loss": 0.4967, + "step": 8054 + }, + { + "epoch": 2.577325085324232, + "grad_norm": 0.48393183946609497, + "learning_rate": 5.919030318854513e-07, + "loss": 0.4358, + "step": 8055 + }, + { + "epoch": 2.5776450511945392, + "grad_norm": 0.48558008670806885, + "learning_rate": 5.910246428170019e-07, + "loss": 0.3956, + "step": 8056 + }, + { + "epoch": 2.5779650170648463, + "grad_norm": 0.5119196176528931, + "learning_rate": 5.901468650582908e-07, + "loss": 0.4767, + "step": 8057 + }, + { + "epoch": 2.5782849829351537, + "grad_norm": 0.5263835191726685, + "learning_rate": 5.89269698731022e-07, + "loss": 0.4206, + "step": 8058 + }, + { + "epoch": 2.5786049488054608, + "grad_norm": 0.48206663131713867, + "learning_rate": 5.883931439568163e-07, + "loss": 0.4976, + "step": 8059 + }, + { + "epoch": 2.578924914675768, + "grad_norm": 0.48202216625213623, + "learning_rate": 5.875172008572055e-07, + "loss": 0.4053, + "step": 8060 + }, + { + "epoch": 2.5792448805460753, + "grad_norm": 0.5368192195892334, + "learning_rate": 5.866418695536458e-07, + "loss": 0.4666, + "step": 8061 + }, + { + "epoch": 2.5795648464163823, + "grad_norm": 0.5510113835334778, + "learning_rate": 5.857671501675005e-07, + "loss": 0.3894, + "step": 8062 + }, + { + "epoch": 2.5798848122866893, + "grad_norm": 0.594622790813446, + "learning_rate": 5.8489304282005e-07, + "loss": 0.5042, + "step": 8063 + }, + { + "epoch": 2.580204778156997, + "grad_norm": 0.48127028346061707, + "learning_rate": 5.840195476324906e-07, + "loss": 0.3677, + "step": 8064 + }, + { + "epoch": 2.580524744027304, + "grad_norm": 0.4667394459247589, + "learning_rate": 5.831466647259315e-07, + "loss": 0.4419, + "step": 8065 + }, + { + "epoch": 2.580844709897611, + "grad_norm": 0.5215911865234375, + "learning_rate": 5.822743942214026e-07, + "loss": 0.4373, + "step": 8066 + }, + { + "epoch": 2.5811646757679183, + "grad_norm": 0.4830494821071625, + "learning_rate": 5.814027362398439e-07, + "loss": 0.4105, + "step": 8067 + }, + { + "epoch": 2.5814846416382253, + "grad_norm": 0.486611008644104, + "learning_rate": 5.80531690902112e-07, + "loss": 0.4652, + "step": 8068 + }, + { + "epoch": 2.5818046075085324, + "grad_norm": 0.4940856397151947, + "learning_rate": 5.796612583289773e-07, + "loss": 0.4603, + "step": 8069 + }, + { + "epoch": 2.58212457337884, + "grad_norm": 0.49336355924606323, + "learning_rate": 5.787914386411248e-07, + "loss": 0.4013, + "step": 8070 + }, + { + "epoch": 2.582444539249147, + "grad_norm": 0.4824593663215637, + "learning_rate": 5.779222319591604e-07, + "loss": 0.4137, + "step": 8071 + }, + { + "epoch": 2.582764505119454, + "grad_norm": 0.4997718930244446, + "learning_rate": 5.770536384035985e-07, + "loss": 0.4918, + "step": 8072 + }, + { + "epoch": 2.5830844709897613, + "grad_norm": 0.46724408864974976, + "learning_rate": 5.761856580948705e-07, + "loss": 0.4103, + "step": 8073 + }, + { + "epoch": 2.5834044368600684, + "grad_norm": 0.4812827408313751, + "learning_rate": 5.753182911533228e-07, + "loss": 0.4547, + "step": 8074 + }, + { + "epoch": 2.5837244027303754, + "grad_norm": 0.5146414041519165, + "learning_rate": 5.744515376992155e-07, + "loss": 0.4209, + "step": 8075 + }, + { + "epoch": 2.5840443686006824, + "grad_norm": 0.5635401010513306, + "learning_rate": 5.735853978527284e-07, + "loss": 0.4415, + "step": 8076 + }, + { + "epoch": 2.58436433447099, + "grad_norm": 0.5143701434135437, + "learning_rate": 5.727198717339511e-07, + "loss": 0.4301, + "step": 8077 + }, + { + "epoch": 2.584684300341297, + "grad_norm": 0.4612678587436676, + "learning_rate": 5.718549594628892e-07, + "loss": 0.4266, + "step": 8078 + }, + { + "epoch": 2.585004266211604, + "grad_norm": 0.5282837748527527, + "learning_rate": 5.70990661159464e-07, + "loss": 0.4799, + "step": 8079 + }, + { + "epoch": 2.5853242320819114, + "grad_norm": 0.4639507234096527, + "learning_rate": 5.701269769435113e-07, + "loss": 0.4183, + "step": 8080 + }, + { + "epoch": 2.5856441979522184, + "grad_norm": 0.5629203915596008, + "learning_rate": 5.692639069347838e-07, + "loss": 0.3848, + "step": 8081 + }, + { + "epoch": 2.5859641638225255, + "grad_norm": 0.5728508830070496, + "learning_rate": 5.684014512529462e-07, + "loss": 0.4767, + "step": 8082 + }, + { + "epoch": 2.5862841296928325, + "grad_norm": 0.5335341691970825, + "learning_rate": 5.675396100175784e-07, + "loss": 0.4757, + "step": 8083 + }, + { + "epoch": 2.58660409556314, + "grad_norm": 0.5024179220199585, + "learning_rate": 5.666783833481749e-07, + "loss": 0.4289, + "step": 8084 + }, + { + "epoch": 2.586924061433447, + "grad_norm": 0.49825990200042725, + "learning_rate": 5.658177713641471e-07, + "loss": 0.4189, + "step": 8085 + }, + { + "epoch": 2.587244027303754, + "grad_norm": 0.45117056369781494, + "learning_rate": 5.649577741848211e-07, + "loss": 0.3969, + "step": 8086 + }, + { + "epoch": 2.5875639931740615, + "grad_norm": 0.4884481132030487, + "learning_rate": 5.640983919294347e-07, + "loss": 0.44, + "step": 8087 + }, + { + "epoch": 2.5878839590443685, + "grad_norm": 0.5033963918685913, + "learning_rate": 5.632396247171429e-07, + "loss": 0.3851, + "step": 8088 + }, + { + "epoch": 2.5882039249146755, + "grad_norm": 0.5034910440444946, + "learning_rate": 5.62381472667014e-07, + "loss": 0.4661, + "step": 8089 + }, + { + "epoch": 2.588523890784983, + "grad_norm": 0.5116603970527649, + "learning_rate": 5.615239358980318e-07, + "loss": 0.4745, + "step": 8090 + }, + { + "epoch": 2.58884385665529, + "grad_norm": 0.507566511631012, + "learning_rate": 5.606670145290971e-07, + "loss": 0.4319, + "step": 8091 + }, + { + "epoch": 2.589163822525597, + "grad_norm": 0.4920852482318878, + "learning_rate": 5.598107086790216e-07, + "loss": 0.4434, + "step": 8092 + }, + { + "epoch": 2.5894837883959045, + "grad_norm": 0.4991411864757538, + "learning_rate": 5.589550184665327e-07, + "loss": 0.4522, + "step": 8093 + }, + { + "epoch": 2.5898037542662116, + "grad_norm": 0.47811561822891235, + "learning_rate": 5.580999440102719e-07, + "loss": 0.4392, + "step": 8094 + }, + { + "epoch": 2.5901237201365186, + "grad_norm": 0.48993173241615295, + "learning_rate": 5.572454854287984e-07, + "loss": 0.4458, + "step": 8095 + }, + { + "epoch": 2.590443686006826, + "grad_norm": 0.4690512716770172, + "learning_rate": 5.563916428405836e-07, + "loss": 0.3775, + "step": 8096 + }, + { + "epoch": 2.590763651877133, + "grad_norm": 0.5116325616836548, + "learning_rate": 5.555384163640132e-07, + "loss": 0.4414, + "step": 8097 + }, + { + "epoch": 2.59108361774744, + "grad_norm": 0.534573495388031, + "learning_rate": 5.546858061173882e-07, + "loss": 0.486, + "step": 8098 + }, + { + "epoch": 2.5914035836177476, + "grad_norm": 0.5088597536087036, + "learning_rate": 5.538338122189235e-07, + "loss": 0.4007, + "step": 8099 + }, + { + "epoch": 2.5917235494880546, + "grad_norm": 0.49535033106803894, + "learning_rate": 5.52982434786749e-07, + "loss": 0.4555, + "step": 8100 + }, + { + "epoch": 2.5920435153583616, + "grad_norm": 0.48237720131874084, + "learning_rate": 5.521316739389116e-07, + "loss": 0.4199, + "step": 8101 + }, + { + "epoch": 2.592363481228669, + "grad_norm": 0.4965192675590515, + "learning_rate": 5.512815297933677e-07, + "loss": 0.5094, + "step": 8102 + }, + { + "epoch": 2.592683447098976, + "grad_norm": 0.4722084403038025, + "learning_rate": 5.504320024679922e-07, + "loss": 0.379, + "step": 8103 + }, + { + "epoch": 2.593003412969283, + "grad_norm": 0.5180158615112305, + "learning_rate": 5.495830920805711e-07, + "loss": 0.4783, + "step": 8104 + }, + { + "epoch": 2.5933233788395906, + "grad_norm": 0.4634142220020294, + "learning_rate": 5.487347987488084e-07, + "loss": 0.394, + "step": 8105 + }, + { + "epoch": 2.5936433447098977, + "grad_norm": 0.5356070399284363, + "learning_rate": 5.478871225903221e-07, + "loss": 0.4783, + "step": 8106 + }, + { + "epoch": 2.5939633105802047, + "grad_norm": 0.5005761981010437, + "learning_rate": 5.47040063722642e-07, + "loss": 0.4703, + "step": 8107 + }, + { + "epoch": 2.594283276450512, + "grad_norm": 0.48422369360923767, + "learning_rate": 5.461936222632141e-07, + "loss": 0.4207, + "step": 8108 + }, + { + "epoch": 2.594603242320819, + "grad_norm": 0.5311996936798096, + "learning_rate": 5.453477983293975e-07, + "loss": 0.4514, + "step": 8109 + }, + { + "epoch": 2.594923208191126, + "grad_norm": 0.5696889758110046, + "learning_rate": 5.445025920384678e-07, + "loss": 0.4826, + "step": 8110 + }, + { + "epoch": 2.5952431740614337, + "grad_norm": 0.5132286548614502, + "learning_rate": 5.43658003507615e-07, + "loss": 0.4557, + "step": 8111 + }, + { + "epoch": 2.5955631399317407, + "grad_norm": 0.46289798617362976, + "learning_rate": 5.428140328539411e-07, + "loss": 0.3898, + "step": 8112 + }, + { + "epoch": 2.5958831058020477, + "grad_norm": 0.5173072218894958, + "learning_rate": 5.419706801944635e-07, + "loss": 0.4478, + "step": 8113 + }, + { + "epoch": 2.596203071672355, + "grad_norm": 0.5169364809989929, + "learning_rate": 5.411279456461133e-07, + "loss": 0.4411, + "step": 8114 + }, + { + "epoch": 2.5965230375426622, + "grad_norm": 0.5261008143424988, + "learning_rate": 5.402858293257379e-07, + "loss": 0.4669, + "step": 8115 + }, + { + "epoch": 2.5968430034129693, + "grad_norm": 0.4407719671726227, + "learning_rate": 5.39444331350098e-07, + "loss": 0.4142, + "step": 8116 + }, + { + "epoch": 2.5971629692832767, + "grad_norm": 0.4699857831001282, + "learning_rate": 5.386034518358679e-07, + "loss": 0.4539, + "step": 8117 + }, + { + "epoch": 2.5974829351535837, + "grad_norm": 0.4793130159378052, + "learning_rate": 5.37763190899635e-07, + "loss": 0.4232, + "step": 8118 + }, + { + "epoch": 2.5978029010238908, + "grad_norm": 0.48637887835502625, + "learning_rate": 5.369235486579055e-07, + "loss": 0.469, + "step": 8119 + }, + { + "epoch": 2.598122866894198, + "grad_norm": 0.4321134090423584, + "learning_rate": 5.360845252270941e-07, + "loss": 0.4343, + "step": 8120 + }, + { + "epoch": 2.5984428327645053, + "grad_norm": 0.4490710198879242, + "learning_rate": 5.352461207235344e-07, + "loss": 0.4117, + "step": 8121 + }, + { + "epoch": 2.5987627986348123, + "grad_norm": 0.5368947386741638, + "learning_rate": 5.344083352634711e-07, + "loss": 0.4582, + "step": 8122 + }, + { + "epoch": 2.5990827645051193, + "grad_norm": 0.49436429142951965, + "learning_rate": 5.335711689630635e-07, + "loss": 0.4019, + "step": 8123 + }, + { + "epoch": 2.5994027303754264, + "grad_norm": 0.505104124546051, + "learning_rate": 5.327346219383872e-07, + "loss": 0.4713, + "step": 8124 + }, + { + "epoch": 2.599722696245734, + "grad_norm": 0.43805694580078125, + "learning_rate": 5.318986943054294e-07, + "loss": 0.4003, + "step": 8125 + }, + { + "epoch": 2.600042662116041, + "grad_norm": 0.48276111483573914, + "learning_rate": 5.310633861800934e-07, + "loss": 0.4441, + "step": 8126 + }, + { + "epoch": 2.600362627986348, + "grad_norm": 0.4786250591278076, + "learning_rate": 5.302286976781956e-07, + "loss": 0.4285, + "step": 8127 + }, + { + "epoch": 2.6006825938566553, + "grad_norm": 0.49598589539527893, + "learning_rate": 5.29394628915465e-07, + "loss": 0.4733, + "step": 8128 + }, + { + "epoch": 2.6010025597269624, + "grad_norm": 0.5020879507064819, + "learning_rate": 5.285611800075479e-07, + "loss": 0.4614, + "step": 8129 + }, + { + "epoch": 2.6013225255972694, + "grad_norm": 0.4940076172351837, + "learning_rate": 5.277283510700021e-07, + "loss": 0.4899, + "step": 8130 + }, + { + "epoch": 2.601642491467577, + "grad_norm": 0.49958762526512146, + "learning_rate": 5.268961422183016e-07, + "loss": 0.4297, + "step": 8131 + }, + { + "epoch": 2.601962457337884, + "grad_norm": 0.5203652381896973, + "learning_rate": 5.260645535678321e-07, + "loss": 0.4593, + "step": 8132 + }, + { + "epoch": 2.602282423208191, + "grad_norm": 0.47786155343055725, + "learning_rate": 5.252335852338936e-07, + "loss": 0.4475, + "step": 8133 + }, + { + "epoch": 2.6026023890784984, + "grad_norm": 0.49558225274086, + "learning_rate": 5.244032373317027e-07, + "loss": 0.4031, + "step": 8134 + }, + { + "epoch": 2.6029223549488054, + "grad_norm": 0.562053382396698, + "learning_rate": 5.235735099763861e-07, + "loss": 0.4867, + "step": 8135 + }, + { + "epoch": 2.6032423208191124, + "grad_norm": 0.46849918365478516, + "learning_rate": 5.227444032829887e-07, + "loss": 0.4359, + "step": 8136 + }, + { + "epoch": 2.60356228668942, + "grad_norm": 0.5348246693611145, + "learning_rate": 5.219159173664662e-07, + "loss": 0.4731, + "step": 8137 + }, + { + "epoch": 2.603882252559727, + "grad_norm": 0.48375827074050903, + "learning_rate": 5.210880523416884e-07, + "loss": 0.4579, + "step": 8138 + }, + { + "epoch": 2.604202218430034, + "grad_norm": 0.4847160577774048, + "learning_rate": 5.202608083234406e-07, + "loss": 0.4329, + "step": 8139 + }, + { + "epoch": 2.6045221843003414, + "grad_norm": 0.4918648898601532, + "learning_rate": 5.194341854264206e-07, + "loss": 0.3665, + "step": 8140 + }, + { + "epoch": 2.6048421501706485, + "grad_norm": 0.5149650573730469, + "learning_rate": 5.186081837652418e-07, + "loss": 0.5078, + "step": 8141 + }, + { + "epoch": 2.6051621160409555, + "grad_norm": 0.45677924156188965, + "learning_rate": 5.177828034544291e-07, + "loss": 0.3819, + "step": 8142 + }, + { + "epoch": 2.605482081911263, + "grad_norm": 0.475281298160553, + "learning_rate": 5.169580446084226e-07, + "loss": 0.3766, + "step": 8143 + }, + { + "epoch": 2.60580204778157, + "grad_norm": 0.5269748568534851, + "learning_rate": 5.161339073415772e-07, + "loss": 0.5267, + "step": 8144 + }, + { + "epoch": 2.606122013651877, + "grad_norm": 0.459114670753479, + "learning_rate": 5.15310391768159e-07, + "loss": 0.3896, + "step": 8145 + }, + { + "epoch": 2.6064419795221845, + "grad_norm": 0.5193907022476196, + "learning_rate": 5.144874980023512e-07, + "loss": 0.4628, + "step": 8146 + }, + { + "epoch": 2.6067619453924915, + "grad_norm": 0.5411141514778137, + "learning_rate": 5.136652261582476e-07, + "loss": 0.4008, + "step": 8147 + }, + { + "epoch": 2.6070819112627985, + "grad_norm": 0.5180327296257019, + "learning_rate": 5.128435763498568e-07, + "loss": 0.4461, + "step": 8148 + }, + { + "epoch": 2.607401877133106, + "grad_norm": 0.5204225778579712, + "learning_rate": 5.120225486911035e-07, + "loss": 0.4305, + "step": 8149 + }, + { + "epoch": 2.607721843003413, + "grad_norm": 0.526789665222168, + "learning_rate": 5.112021432958214e-07, + "loss": 0.4612, + "step": 8150 + }, + { + "epoch": 2.60804180887372, + "grad_norm": 0.47868508100509644, + "learning_rate": 5.103823602777641e-07, + "loss": 0.3933, + "step": 8151 + }, + { + "epoch": 2.6083617747440275, + "grad_norm": 0.5638915300369263, + "learning_rate": 5.095631997505923e-07, + "loss": 0.4622, + "step": 8152 + }, + { + "epoch": 2.6086817406143346, + "grad_norm": 0.48375749588012695, + "learning_rate": 5.087446618278858e-07, + "loss": 0.4052, + "step": 8153 + }, + { + "epoch": 2.6090017064846416, + "grad_norm": 0.4716733396053314, + "learning_rate": 5.079267466231352e-07, + "loss": 0.4573, + "step": 8154 + }, + { + "epoch": 2.609321672354949, + "grad_norm": 0.4670661687850952, + "learning_rate": 5.07109454249744e-07, + "loss": 0.4325, + "step": 8155 + }, + { + "epoch": 2.609641638225256, + "grad_norm": 0.47152283787727356, + "learning_rate": 5.062927848210336e-07, + "loss": 0.3957, + "step": 8156 + }, + { + "epoch": 2.609961604095563, + "grad_norm": 0.47465530037879944, + "learning_rate": 5.054767384502335e-07, + "loss": 0.4826, + "step": 8157 + }, + { + "epoch": 2.6102815699658706, + "grad_norm": 0.5024846792221069, + "learning_rate": 5.046613152504914e-07, + "loss": 0.4282, + "step": 8158 + }, + { + "epoch": 2.6106015358361776, + "grad_norm": 0.4924025237560272, + "learning_rate": 5.038465153348659e-07, + "loss": 0.4253, + "step": 8159 + }, + { + "epoch": 2.6109215017064846, + "grad_norm": 0.49057185649871826, + "learning_rate": 5.030323388163294e-07, + "loss": 0.4388, + "step": 8160 + }, + { + "epoch": 2.6112414675767917, + "grad_norm": 0.5156389474868774, + "learning_rate": 5.022187858077698e-07, + "loss": 0.4316, + "step": 8161 + }, + { + "epoch": 2.611561433447099, + "grad_norm": 0.49020588397979736, + "learning_rate": 5.014058564219859e-07, + "loss": 0.4459, + "step": 8162 + }, + { + "epoch": 2.611881399317406, + "grad_norm": 0.529273509979248, + "learning_rate": 5.00593550771693e-07, + "loss": 0.4647, + "step": 8163 + }, + { + "epoch": 2.612201365187713, + "grad_norm": 0.5657989382743835, + "learning_rate": 4.997818689695166e-07, + "loss": 0.4385, + "step": 8164 + }, + { + "epoch": 2.61252133105802, + "grad_norm": 0.4963492155075073, + "learning_rate": 4.989708111279973e-07, + "loss": 0.4139, + "step": 8165 + }, + { + "epoch": 2.6128412969283277, + "grad_norm": 0.4489327073097229, + "learning_rate": 4.981603773595911e-07, + "loss": 0.3985, + "step": 8166 + }, + { + "epoch": 2.6131612627986347, + "grad_norm": 0.49600255489349365, + "learning_rate": 4.973505677766633e-07, + "loss": 0.4416, + "step": 8167 + }, + { + "epoch": 2.6134812286689417, + "grad_norm": 0.5063232779502869, + "learning_rate": 4.965413824914966e-07, + "loss": 0.4359, + "step": 8168 + }, + { + "epoch": 2.613801194539249, + "grad_norm": 0.5594180822372437, + "learning_rate": 4.95732821616286e-07, + "loss": 0.3855, + "step": 8169 + }, + { + "epoch": 2.614121160409556, + "grad_norm": 0.5148386359214783, + "learning_rate": 4.949248852631367e-07, + "loss": 0.4629, + "step": 8170 + }, + { + "epoch": 2.6144411262798632, + "grad_norm": 0.4605209529399872, + "learning_rate": 4.941175735440728e-07, + "loss": 0.3764, + "step": 8171 + }, + { + "epoch": 2.6147610921501707, + "grad_norm": 0.5135962963104248, + "learning_rate": 4.933108865710274e-07, + "loss": 0.4638, + "step": 8172 + }, + { + "epoch": 2.6150810580204777, + "grad_norm": 0.5130895376205444, + "learning_rate": 4.9250482445585e-07, + "loss": 0.4997, + "step": 8173 + }, + { + "epoch": 2.6154010238907848, + "grad_norm": 0.48560479283332825, + "learning_rate": 4.916993873103015e-07, + "loss": 0.3539, + "step": 8174 + }, + { + "epoch": 2.6157209897610922, + "grad_norm": 0.55850750207901, + "learning_rate": 4.908945752460553e-07, + "loss": 0.5305, + "step": 8175 + }, + { + "epoch": 2.6160409556313993, + "grad_norm": 0.5068272352218628, + "learning_rate": 4.900903883747021e-07, + "loss": 0.4184, + "step": 8176 + }, + { + "epoch": 2.6163609215017063, + "grad_norm": 0.4663023352622986, + "learning_rate": 4.892868268077411e-07, + "loss": 0.3963, + "step": 8177 + }, + { + "epoch": 2.6166808873720138, + "grad_norm": 0.5439704060554504, + "learning_rate": 4.884838906565892e-07, + "loss": 0.4643, + "step": 8178 + }, + { + "epoch": 2.617000853242321, + "grad_norm": 0.5312791466712952, + "learning_rate": 4.87681580032573e-07, + "loss": 0.4374, + "step": 8179 + }, + { + "epoch": 2.617320819112628, + "grad_norm": 0.5001805424690247, + "learning_rate": 4.868798950469333e-07, + "loss": 0.4276, + "step": 8180 + }, + { + "epoch": 2.6176407849829353, + "grad_norm": 0.48276445269584656, + "learning_rate": 4.860788358108265e-07, + "loss": 0.4505, + "step": 8181 + }, + { + "epoch": 2.6179607508532423, + "grad_norm": 0.5378854870796204, + "learning_rate": 4.852784024353185e-07, + "loss": 0.4857, + "step": 8182 + }, + { + "epoch": 2.6182807167235493, + "grad_norm": 0.47105562686920166, + "learning_rate": 4.844785950313924e-07, + "loss": 0.4303, + "step": 8183 + }, + { + "epoch": 2.618600682593857, + "grad_norm": 0.5040062069892883, + "learning_rate": 4.836794137099415e-07, + "loss": 0.4188, + "step": 8184 + }, + { + "epoch": 2.618920648464164, + "grad_norm": 0.49857619404792786, + "learning_rate": 4.828808585817724e-07, + "loss": 0.4326, + "step": 8185 + }, + { + "epoch": 2.619240614334471, + "grad_norm": 0.4615936875343323, + "learning_rate": 4.820829297576068e-07, + "loss": 0.4422, + "step": 8186 + }, + { + "epoch": 2.6195605802047783, + "grad_norm": 0.48037397861480713, + "learning_rate": 4.812856273480776e-07, + "loss": 0.3965, + "step": 8187 + }, + { + "epoch": 2.6198805460750854, + "grad_norm": 0.4728635549545288, + "learning_rate": 4.804889514637334e-07, + "loss": 0.4579, + "step": 8188 + }, + { + "epoch": 2.6202005119453924, + "grad_norm": 0.47411414980888367, + "learning_rate": 4.796929022150332e-07, + "loss": 0.4767, + "step": 8189 + }, + { + "epoch": 2.6205204778157, + "grad_norm": 0.4786132276058197, + "learning_rate": 4.788974797123491e-07, + "loss": 0.4402, + "step": 8190 + }, + { + "epoch": 2.620840443686007, + "grad_norm": 0.5201287269592285, + "learning_rate": 4.78102684065968e-07, + "loss": 0.4799, + "step": 8191 + }, + { + "epoch": 2.621160409556314, + "grad_norm": 0.44401898980140686, + "learning_rate": 4.773085153860912e-07, + "loss": 0.3678, + "step": 8192 + }, + { + "epoch": 2.6214803754266214, + "grad_norm": 0.49735119938850403, + "learning_rate": 4.765149737828295e-07, + "loss": 0.4713, + "step": 8193 + }, + { + "epoch": 2.6218003412969284, + "grad_norm": 0.5207501649856567, + "learning_rate": 4.7572205936620874e-07, + "loss": 0.4354, + "step": 8194 + }, + { + "epoch": 2.6221203071672354, + "grad_norm": 0.5102244019508362, + "learning_rate": 4.74929772246166e-07, + "loss": 0.4234, + "step": 8195 + }, + { + "epoch": 2.622440273037543, + "grad_norm": 0.49887150526046753, + "learning_rate": 4.7413811253255446e-07, + "loss": 0.4551, + "step": 8196 + }, + { + "epoch": 2.62276023890785, + "grad_norm": 0.5083227157592773, + "learning_rate": 4.733470803351392e-07, + "loss": 0.4755, + "step": 8197 + }, + { + "epoch": 2.623080204778157, + "grad_norm": 0.49517136812210083, + "learning_rate": 4.7255667576359687e-07, + "loss": 0.3996, + "step": 8198 + }, + { + "epoch": 2.6234001706484644, + "grad_norm": 0.4763628840446472, + "learning_rate": 4.717668989275187e-07, + "loss": 0.4546, + "step": 8199 + }, + { + "epoch": 2.6237201365187715, + "grad_norm": 0.4698621332645416, + "learning_rate": 4.70977749936406e-07, + "loss": 0.4112, + "step": 8200 + }, + { + "epoch": 2.6240401023890785, + "grad_norm": 0.5033742189407349, + "learning_rate": 4.701892288996768e-07, + "loss": 0.4685, + "step": 8201 + }, + { + "epoch": 2.6243600682593855, + "grad_norm": 0.4517287015914917, + "learning_rate": 4.694013359266614e-07, + "loss": 0.3929, + "step": 8202 + }, + { + "epoch": 2.624680034129693, + "grad_norm": 0.5576598048210144, + "learning_rate": 4.6861407112660184e-07, + "loss": 0.5224, + "step": 8203 + }, + { + "epoch": 2.625, + "grad_norm": 0.4753642976284027, + "learning_rate": 4.6782743460865197e-07, + "loss": 0.381, + "step": 8204 + }, + { + "epoch": 2.625319965870307, + "grad_norm": 0.49122047424316406, + "learning_rate": 4.6704142648188013e-07, + "loss": 0.4727, + "step": 8205 + }, + { + "epoch": 2.6256399317406145, + "grad_norm": 0.4982339143753052, + "learning_rate": 4.662560468552674e-07, + "loss": 0.4416, + "step": 8206 + }, + { + "epoch": 2.6259598976109215, + "grad_norm": 0.5182000398635864, + "learning_rate": 4.6547129583770943e-07, + "loss": 0.4386, + "step": 8207 + }, + { + "epoch": 2.6262798634812285, + "grad_norm": 0.5248851180076599, + "learning_rate": 4.6468717353801095e-07, + "loss": 0.3901, + "step": 8208 + }, + { + "epoch": 2.6265998293515356, + "grad_norm": 0.5372112393379211, + "learning_rate": 4.639036800648927e-07, + "loss": 0.4414, + "step": 8209 + }, + { + "epoch": 2.626919795221843, + "grad_norm": 0.5512729287147522, + "learning_rate": 4.6312081552698454e-07, + "loss": 0.4765, + "step": 8210 + }, + { + "epoch": 2.62723976109215, + "grad_norm": 0.4698147177696228, + "learning_rate": 4.6233858003283394e-07, + "loss": 0.3375, + "step": 8211 + }, + { + "epoch": 2.627559726962457, + "grad_norm": 0.520575761795044, + "learning_rate": 4.6155697369089867e-07, + "loss": 0.5129, + "step": 8212 + }, + { + "epoch": 2.6278796928327646, + "grad_norm": 0.46919360756874084, + "learning_rate": 4.607759966095493e-07, + "loss": 0.4307, + "step": 8213 + }, + { + "epoch": 2.6281996587030716, + "grad_norm": 0.5309707522392273, + "learning_rate": 4.5999564889706806e-07, + "loss": 0.5125, + "step": 8214 + }, + { + "epoch": 2.6285196245733786, + "grad_norm": 0.4674423038959503, + "learning_rate": 4.592159306616517e-07, + "loss": 0.4261, + "step": 8215 + }, + { + "epoch": 2.628839590443686, + "grad_norm": 0.5125669240951538, + "learning_rate": 4.584368420114088e-07, + "loss": 0.4309, + "step": 8216 + }, + { + "epoch": 2.629159556313993, + "grad_norm": 0.553999662399292, + "learning_rate": 4.576583830543624e-07, + "loss": 0.48, + "step": 8217 + }, + { + "epoch": 2.6294795221843, + "grad_norm": 0.5225322842597961, + "learning_rate": 4.568805538984461e-07, + "loss": 0.4011, + "step": 8218 + }, + { + "epoch": 2.6297994880546076, + "grad_norm": 0.5288572907447815, + "learning_rate": 4.5610335465150535e-07, + "loss": 0.4341, + "step": 8219 + }, + { + "epoch": 2.6301194539249146, + "grad_norm": 0.4576762914657593, + "learning_rate": 4.553267854213017e-07, + "loss": 0.3901, + "step": 8220 + }, + { + "epoch": 2.6304394197952217, + "grad_norm": 0.46657612919807434, + "learning_rate": 4.5455084631550575e-07, + "loss": 0.4513, + "step": 8221 + }, + { + "epoch": 2.630759385665529, + "grad_norm": 0.4978737235069275, + "learning_rate": 4.537755374417041e-07, + "loss": 0.3943, + "step": 8222 + }, + { + "epoch": 2.631079351535836, + "grad_norm": 0.4882107973098755, + "learning_rate": 4.5300085890739366e-07, + "loss": 0.4666, + "step": 8223 + }, + { + "epoch": 2.631399317406143, + "grad_norm": 0.5072929859161377, + "learning_rate": 4.522268108199834e-07, + "loss": 0.4409, + "step": 8224 + }, + { + "epoch": 2.6317192832764507, + "grad_norm": 0.5407156944274902, + "learning_rate": 4.5145339328679705e-07, + "loss": 0.4459, + "step": 8225 + }, + { + "epoch": 2.6320392491467577, + "grad_norm": 0.44135379791259766, + "learning_rate": 4.506806064150704e-07, + "loss": 0.4064, + "step": 8226 + }, + { + "epoch": 2.6323592150170647, + "grad_norm": 0.4765286147594452, + "learning_rate": 4.499084503119505e-07, + "loss": 0.4493, + "step": 8227 + }, + { + "epoch": 2.632679180887372, + "grad_norm": 0.5037778615951538, + "learning_rate": 4.491369250844979e-07, + "loss": 0.4995, + "step": 8228 + }, + { + "epoch": 2.632999146757679, + "grad_norm": 0.45145121216773987, + "learning_rate": 4.4836603083968476e-07, + "loss": 0.4353, + "step": 8229 + }, + { + "epoch": 2.6333191126279862, + "grad_norm": 0.48700815439224243, + "learning_rate": 4.475957676843973e-07, + "loss": 0.3971, + "step": 8230 + }, + { + "epoch": 2.6336390784982937, + "grad_norm": 0.552137017250061, + "learning_rate": 4.468261357254339e-07, + "loss": 0.4473, + "step": 8231 + }, + { + "epoch": 2.6339590443686007, + "grad_norm": 0.4934519827365875, + "learning_rate": 4.4605713506950477e-07, + "loss": 0.4481, + "step": 8232 + }, + { + "epoch": 2.6342790102389078, + "grad_norm": 0.46769335865974426, + "learning_rate": 4.452887658232319e-07, + "loss": 0.3953, + "step": 8233 + }, + { + "epoch": 2.6345989761092152, + "grad_norm": 0.49531757831573486, + "learning_rate": 4.445210280931506e-07, + "loss": 0.3934, + "step": 8234 + }, + { + "epoch": 2.6349189419795223, + "grad_norm": 0.5289334654808044, + "learning_rate": 4.43753921985709e-07, + "loss": 0.4995, + "step": 8235 + }, + { + "epoch": 2.6352389078498293, + "grad_norm": 0.4507151246070862, + "learning_rate": 4.429874476072682e-07, + "loss": 0.352, + "step": 8236 + }, + { + "epoch": 2.6355588737201368, + "grad_norm": 0.4960966408252716, + "learning_rate": 4.422216050641004e-07, + "loss": 0.4657, + "step": 8237 + }, + { + "epoch": 2.635878839590444, + "grad_norm": 0.5034002661705017, + "learning_rate": 4.4145639446238955e-07, + "loss": 0.4416, + "step": 8238 + }, + { + "epoch": 2.636198805460751, + "grad_norm": 0.48766377568244934, + "learning_rate": 4.406918159082335e-07, + "loss": 0.4567, + "step": 8239 + }, + { + "epoch": 2.6365187713310583, + "grad_norm": 0.44892147183418274, + "learning_rate": 4.3992786950764145e-07, + "loss": 0.3961, + "step": 8240 + }, + { + "epoch": 2.6368387372013653, + "grad_norm": 0.48189565539360046, + "learning_rate": 4.3916455536653744e-07, + "loss": 0.4709, + "step": 8241 + }, + { + "epoch": 2.6371587030716723, + "grad_norm": 0.4697919189929962, + "learning_rate": 4.384018735907547e-07, + "loss": 0.4415, + "step": 8242 + }, + { + "epoch": 2.63747866894198, + "grad_norm": 0.48356616497039795, + "learning_rate": 4.376398242860397e-07, + "loss": 0.4241, + "step": 8243 + }, + { + "epoch": 2.637798634812287, + "grad_norm": 0.5298052430152893, + "learning_rate": 4.368784075580512e-07, + "loss": 0.4637, + "step": 8244 + }, + { + "epoch": 2.638118600682594, + "grad_norm": 0.4812069535255432, + "learning_rate": 4.36117623512361e-07, + "loss": 0.3976, + "step": 8245 + }, + { + "epoch": 2.638438566552901, + "grad_norm": 0.5246396064758301, + "learning_rate": 4.3535747225445345e-07, + "loss": 0.4488, + "step": 8246 + }, + { + "epoch": 2.6387585324232083, + "grad_norm": 0.50551438331604, + "learning_rate": 4.345979538897238e-07, + "loss": 0.4682, + "step": 8247 + }, + { + "epoch": 2.6390784982935154, + "grad_norm": 0.527849018573761, + "learning_rate": 4.3383906852348e-07, + "loss": 0.4303, + "step": 8248 + }, + { + "epoch": 2.6393984641638224, + "grad_norm": 0.5653607845306396, + "learning_rate": 4.330808162609418e-07, + "loss": 0.4698, + "step": 8249 + }, + { + "epoch": 2.6397184300341294, + "grad_norm": 0.4829493761062622, + "learning_rate": 4.323231972072428e-07, + "loss": 0.4029, + "step": 8250 + }, + { + "epoch": 2.640038395904437, + "grad_norm": 0.49808141589164734, + "learning_rate": 4.315662114674285e-07, + "loss": 0.3898, + "step": 8251 + }, + { + "epoch": 2.640358361774744, + "grad_norm": 0.5228855013847351, + "learning_rate": 4.308098591464549e-07, + "loss": 0.4504, + "step": 8252 + }, + { + "epoch": 2.640678327645051, + "grad_norm": 0.4826861321926117, + "learning_rate": 4.300541403491909e-07, + "loss": 0.3995, + "step": 8253 + }, + { + "epoch": 2.6409982935153584, + "grad_norm": 0.4797203242778778, + "learning_rate": 4.2929905518041714e-07, + "loss": 0.4093, + "step": 8254 + }, + { + "epoch": 2.6413182593856654, + "grad_norm": 0.494663268327713, + "learning_rate": 4.285446037448282e-07, + "loss": 0.4543, + "step": 8255 + }, + { + "epoch": 2.6416382252559725, + "grad_norm": 0.4919332265853882, + "learning_rate": 4.2779078614703093e-07, + "loss": 0.499, + "step": 8256 + }, + { + "epoch": 2.64195819112628, + "grad_norm": 0.4947248697280884, + "learning_rate": 4.2703760249154124e-07, + "loss": 0.4133, + "step": 8257 + }, + { + "epoch": 2.642278156996587, + "grad_norm": 0.5169200301170349, + "learning_rate": 4.262850528827894e-07, + "loss": 0.4587, + "step": 8258 + }, + { + "epoch": 2.642598122866894, + "grad_norm": 0.5032166838645935, + "learning_rate": 4.255331374251159e-07, + "loss": 0.4039, + "step": 8259 + }, + { + "epoch": 2.6429180887372015, + "grad_norm": 0.555990993976593, + "learning_rate": 4.247818562227768e-07, + "loss": 0.4784, + "step": 8260 + }, + { + "epoch": 2.6432380546075085, + "grad_norm": 0.5182940363883972, + "learning_rate": 4.240312093799381e-07, + "loss": 0.3817, + "step": 8261 + }, + { + "epoch": 2.6435580204778155, + "grad_norm": 0.5078566670417786, + "learning_rate": 4.2328119700067727e-07, + "loss": 0.473, + "step": 8262 + }, + { + "epoch": 2.643877986348123, + "grad_norm": 0.5581321716308594, + "learning_rate": 4.225318191889843e-07, + "loss": 0.4822, + "step": 8263 + }, + { + "epoch": 2.64419795221843, + "grad_norm": 0.4526835083961487, + "learning_rate": 4.2178307604876067e-07, + "loss": 0.4142, + "step": 8264 + }, + { + "epoch": 2.644517918088737, + "grad_norm": 0.47004154324531555, + "learning_rate": 4.210349676838216e-07, + "loss": 0.4693, + "step": 8265 + }, + { + "epoch": 2.6448378839590445, + "grad_norm": 0.4589392840862274, + "learning_rate": 4.202874941978935e-07, + "loss": 0.4177, + "step": 8266 + }, + { + "epoch": 2.6451578498293515, + "grad_norm": 0.4764752984046936, + "learning_rate": 4.1954065569461355e-07, + "loss": 0.4506, + "step": 8267 + }, + { + "epoch": 2.6454778156996586, + "grad_norm": 0.5085862874984741, + "learning_rate": 4.187944522775328e-07, + "loss": 0.4215, + "step": 8268 + }, + { + "epoch": 2.645797781569966, + "grad_norm": 0.5105277299880981, + "learning_rate": 4.1804888405011125e-07, + "loss": 0.4466, + "step": 8269 + }, + { + "epoch": 2.646117747440273, + "grad_norm": 0.5165952444076538, + "learning_rate": 4.1730395111572397e-07, + "loss": 0.4488, + "step": 8270 + }, + { + "epoch": 2.64643771331058, + "grad_norm": 0.5091129541397095, + "learning_rate": 4.165596535776578e-07, + "loss": 0.4375, + "step": 8271 + }, + { + "epoch": 2.6467576791808876, + "grad_norm": 0.5199570655822754, + "learning_rate": 4.1581599153911026e-07, + "loss": 0.4288, + "step": 8272 + }, + { + "epoch": 2.6470776450511946, + "grad_norm": 0.48100724816322327, + "learning_rate": 4.150729651031898e-07, + "loss": 0.4851, + "step": 8273 + }, + { + "epoch": 2.6473976109215016, + "grad_norm": 0.44028621912002563, + "learning_rate": 4.143305743729181e-07, + "loss": 0.3605, + "step": 8274 + }, + { + "epoch": 2.647717576791809, + "grad_norm": 0.5184507966041565, + "learning_rate": 4.1358881945122884e-07, + "loss": 0.462, + "step": 8275 + }, + { + "epoch": 2.648037542662116, + "grad_norm": 0.5367111563682556, + "learning_rate": 4.128477004409681e-07, + "loss": 0.425, + "step": 8276 + }, + { + "epoch": 2.648357508532423, + "grad_norm": 0.51981121301651, + "learning_rate": 4.12107217444892e-07, + "loss": 0.4634, + "step": 8277 + }, + { + "epoch": 2.6486774744027306, + "grad_norm": 0.5907763242721558, + "learning_rate": 4.113673705656701e-07, + "loss": 0.4485, + "step": 8278 + }, + { + "epoch": 2.6489974402730376, + "grad_norm": 0.47015947103500366, + "learning_rate": 4.1062815990588147e-07, + "loss": 0.4189, + "step": 8279 + }, + { + "epoch": 2.6493174061433447, + "grad_norm": 0.5920960307121277, + "learning_rate": 4.098895855680191e-07, + "loss": 0.4666, + "step": 8280 + }, + { + "epoch": 2.649637372013652, + "grad_norm": 0.5242739319801331, + "learning_rate": 4.0915164765448936e-07, + "loss": 0.4608, + "step": 8281 + }, + { + "epoch": 2.649957337883959, + "grad_norm": 0.5620046854019165, + "learning_rate": 4.08414346267606e-07, + "loss": 0.458, + "step": 8282 + }, + { + "epoch": 2.650277303754266, + "grad_norm": 0.4502480626106262, + "learning_rate": 4.0767768150959785e-07, + "loss": 0.4147, + "step": 8283 + }, + { + "epoch": 2.6505972696245736, + "grad_norm": 0.4997509717941284, + "learning_rate": 4.069416534826026e-07, + "loss": 0.4684, + "step": 8284 + }, + { + "epoch": 2.6509172354948807, + "grad_norm": 0.5171517729759216, + "learning_rate": 4.0620626228867256e-07, + "loss": 0.4625, + "step": 8285 + }, + { + "epoch": 2.6512372013651877, + "grad_norm": 0.5204228162765503, + "learning_rate": 4.054715080297722e-07, + "loss": 0.424, + "step": 8286 + }, + { + "epoch": 2.6515571672354947, + "grad_norm": 0.5608391165733337, + "learning_rate": 4.0473739080777394e-07, + "loss": 0.5094, + "step": 8287 + }, + { + "epoch": 2.651877133105802, + "grad_norm": 0.4928218722343445, + "learning_rate": 4.0400391072446533e-07, + "loss": 0.401, + "step": 8288 + }, + { + "epoch": 2.6521970989761092, + "grad_norm": 0.4749138653278351, + "learning_rate": 4.032710678815421e-07, + "loss": 0.4399, + "step": 8289 + }, + { + "epoch": 2.6525170648464163, + "grad_norm": 0.4711044132709503, + "learning_rate": 4.025388623806159e-07, + "loss": 0.4032, + "step": 8290 + }, + { + "epoch": 2.6528370307167233, + "grad_norm": 0.4894063472747803, + "learning_rate": 4.0180729432320766e-07, + "loss": 0.4009, + "step": 8291 + }, + { + "epoch": 2.6531569965870307, + "grad_norm": 0.5121898651123047, + "learning_rate": 4.010763638107501e-07, + "loss": 0.4611, + "step": 8292 + }, + { + "epoch": 2.6534769624573378, + "grad_norm": 0.5382512211799622, + "learning_rate": 4.003460709445872e-07, + "loss": 0.4387, + "step": 8293 + }, + { + "epoch": 2.653796928327645, + "grad_norm": 0.5028988122940063, + "learning_rate": 3.9961641582597464e-07, + "loss": 0.4295, + "step": 8294 + }, + { + "epoch": 2.6541168941979523, + "grad_norm": 0.5069299340248108, + "learning_rate": 3.988873985560798e-07, + "loss": 0.4483, + "step": 8295 + }, + { + "epoch": 2.6544368600682593, + "grad_norm": 0.43606746196746826, + "learning_rate": 3.9815901923598354e-07, + "loss": 0.4117, + "step": 8296 + }, + { + "epoch": 2.6547568259385663, + "grad_norm": 0.4850480556488037, + "learning_rate": 3.9743127796667556e-07, + "loss": 0.4718, + "step": 8297 + }, + { + "epoch": 2.655076791808874, + "grad_norm": 0.529316246509552, + "learning_rate": 3.967041748490569e-07, + "loss": 0.4571, + "step": 8298 + }, + { + "epoch": 2.655396757679181, + "grad_norm": 0.49332305788993835, + "learning_rate": 3.9597770998394347e-07, + "loss": 0.4133, + "step": 8299 + }, + { + "epoch": 2.655716723549488, + "grad_norm": 0.523765504360199, + "learning_rate": 3.9525188347205813e-07, + "loss": 0.4775, + "step": 8300 + }, + { + "epoch": 2.6560366894197953, + "grad_norm": 0.4612776041030884, + "learning_rate": 3.9452669541403976e-07, + "loss": 0.4597, + "step": 8301 + }, + { + "epoch": 2.6563566552901023, + "grad_norm": 0.4600866436958313, + "learning_rate": 3.9380214591043574e-07, + "loss": 0.4735, + "step": 8302 + }, + { + "epoch": 2.6566766211604094, + "grad_norm": 0.5209136605262756, + "learning_rate": 3.9307823506170405e-07, + "loss": 0.4345, + "step": 8303 + }, + { + "epoch": 2.656996587030717, + "grad_norm": 0.5463297963142395, + "learning_rate": 3.9235496296821884e-07, + "loss": 0.435, + "step": 8304 + }, + { + "epoch": 2.657316552901024, + "grad_norm": 0.57563316822052, + "learning_rate": 3.9163232973026045e-07, + "loss": 0.4648, + "step": 8305 + }, + { + "epoch": 2.657636518771331, + "grad_norm": 0.5193503499031067, + "learning_rate": 3.909103354480237e-07, + "loss": 0.4834, + "step": 8306 + }, + { + "epoch": 2.6579564846416384, + "grad_norm": 0.4927099049091339, + "learning_rate": 3.9018898022161413e-07, + "loss": 0.3804, + "step": 8307 + }, + { + "epoch": 2.6582764505119454, + "grad_norm": 0.5175009965896606, + "learning_rate": 3.894682641510478e-07, + "loss": 0.4284, + "step": 8308 + }, + { + "epoch": 2.6585964163822524, + "grad_norm": 0.4690502882003784, + "learning_rate": 3.8874818733625363e-07, + "loss": 0.4287, + "step": 8309 + }, + { + "epoch": 2.65891638225256, + "grad_norm": 0.5278282165527344, + "learning_rate": 3.880287498770696e-07, + "loss": 0.4618, + "step": 8310 + }, + { + "epoch": 2.659236348122867, + "grad_norm": 0.5248937010765076, + "learning_rate": 3.873099518732493e-07, + "loss": 0.4448, + "step": 8311 + }, + { + "epoch": 2.659556313993174, + "grad_norm": 0.4708818793296814, + "learning_rate": 3.8659179342445283e-07, + "loss": 0.4113, + "step": 8312 + }, + { + "epoch": 2.6598762798634814, + "grad_norm": 0.45589479804039, + "learning_rate": 3.858742746302535e-07, + "loss": 0.4362, + "step": 8313 + }, + { + "epoch": 2.6601962457337884, + "grad_norm": 0.4845450222492218, + "learning_rate": 3.851573955901383e-07, + "loss": 0.3885, + "step": 8314 + }, + { + "epoch": 2.6605162116040955, + "grad_norm": 0.511802077293396, + "learning_rate": 3.8444115640350053e-07, + "loss": 0.4661, + "step": 8315 + }, + { + "epoch": 2.660836177474403, + "grad_norm": 0.5026915669441223, + "learning_rate": 3.837255571696502e-07, + "loss": 0.4425, + "step": 8316 + }, + { + "epoch": 2.66115614334471, + "grad_norm": 0.4273529052734375, + "learning_rate": 3.830105979878046e-07, + "loss": 0.3899, + "step": 8317 + }, + { + "epoch": 2.661476109215017, + "grad_norm": 0.5140239000320435, + "learning_rate": 3.822962789570933e-07, + "loss": 0.4782, + "step": 8318 + }, + { + "epoch": 2.6617960750853245, + "grad_norm": 0.5597017407417297, + "learning_rate": 3.815826001765594e-07, + "loss": 0.4586, + "step": 8319 + }, + { + "epoch": 2.6621160409556315, + "grad_norm": 0.540404200553894, + "learning_rate": 3.8086956174515245e-07, + "loss": 0.5083, + "step": 8320 + }, + { + "epoch": 2.6624360068259385, + "grad_norm": 0.4860124886035919, + "learning_rate": 3.801571637617391e-07, + "loss": 0.4345, + "step": 8321 + }, + { + "epoch": 2.662755972696246, + "grad_norm": 0.5035784244537354, + "learning_rate": 3.79445406325093e-07, + "loss": 0.416, + "step": 8322 + }, + { + "epoch": 2.663075938566553, + "grad_norm": 0.5504601001739502, + "learning_rate": 3.787342895338986e-07, + "loss": 0.4488, + "step": 8323 + }, + { + "epoch": 2.66339590443686, + "grad_norm": 0.5597847104072571, + "learning_rate": 3.780238134867553e-07, + "loss": 0.4273, + "step": 8324 + }, + { + "epoch": 2.6637158703071675, + "grad_norm": 0.4814067482948303, + "learning_rate": 3.773139782821705e-07, + "loss": 0.4353, + "step": 8325 + }, + { + "epoch": 2.6640358361774745, + "grad_norm": 0.520801842212677, + "learning_rate": 3.766047840185638e-07, + "loss": 0.4601, + "step": 8326 + }, + { + "epoch": 2.6643558020477816, + "grad_norm": 0.5612490177154541, + "learning_rate": 3.758962307942665e-07, + "loss": 0.4002, + "step": 8327 + }, + { + "epoch": 2.6646757679180886, + "grad_norm": 0.5105308890342712, + "learning_rate": 3.7518831870751836e-07, + "loss": 0.4121, + "step": 8328 + }, + { + "epoch": 2.664995733788396, + "grad_norm": 0.5145578384399414, + "learning_rate": 3.7448104785647476e-07, + "loss": 0.4577, + "step": 8329 + }, + { + "epoch": 2.665315699658703, + "grad_norm": 0.4928121864795685, + "learning_rate": 3.737744183391978e-07, + "loss": 0.4292, + "step": 8330 + }, + { + "epoch": 2.66563566552901, + "grad_norm": 0.503603994846344, + "learning_rate": 3.7306843025366415e-07, + "loss": 0.4537, + "step": 8331 + }, + { + "epoch": 2.6659556313993176, + "grad_norm": 0.5278449654579163, + "learning_rate": 3.723630836977576e-07, + "loss": 0.4532, + "step": 8332 + }, + { + "epoch": 2.6662755972696246, + "grad_norm": 0.468374639749527, + "learning_rate": 3.716583787692779e-07, + "loss": 0.4234, + "step": 8333 + }, + { + "epoch": 2.6665955631399316, + "grad_norm": 0.47699859738349915, + "learning_rate": 3.709543155659323e-07, + "loss": 0.4883, + "step": 8334 + }, + { + "epoch": 2.6669155290102387, + "grad_norm": 0.456051230430603, + "learning_rate": 3.702508941853383e-07, + "loss": 0.3865, + "step": 8335 + }, + { + "epoch": 2.667235494880546, + "grad_norm": 0.4542686343193054, + "learning_rate": 3.695481147250285e-07, + "loss": 0.3937, + "step": 8336 + }, + { + "epoch": 2.667555460750853, + "grad_norm": 0.48175179958343506, + "learning_rate": 3.6884597728244265e-07, + "loss": 0.4983, + "step": 8337 + }, + { + "epoch": 2.66787542662116, + "grad_norm": 0.4818973243236542, + "learning_rate": 3.68144481954934e-07, + "loss": 0.4271, + "step": 8338 + }, + { + "epoch": 2.6681953924914676, + "grad_norm": 0.47783929109573364, + "learning_rate": 3.6744362883976534e-07, + "loss": 0.4392, + "step": 8339 + }, + { + "epoch": 2.6685153583617747, + "grad_norm": 0.4501132667064667, + "learning_rate": 3.6674341803411004e-07, + "loss": 0.421, + "step": 8340 + }, + { + "epoch": 2.6688353242320817, + "grad_norm": 0.5383400917053223, + "learning_rate": 3.660438496350543e-07, + "loss": 0.4726, + "step": 8341 + }, + { + "epoch": 2.669155290102389, + "grad_norm": 0.5299282670021057, + "learning_rate": 3.653449237395923e-07, + "loss": 0.494, + "step": 8342 + }, + { + "epoch": 2.669475255972696, + "grad_norm": 0.4743891656398773, + "learning_rate": 3.646466404446336e-07, + "loss": 0.4127, + "step": 8343 + }, + { + "epoch": 2.669795221843003, + "grad_norm": 0.4851328432559967, + "learning_rate": 3.639489998469947e-07, + "loss": 0.4157, + "step": 8344 + }, + { + "epoch": 2.6701151877133107, + "grad_norm": 0.49590522050857544, + "learning_rate": 3.632520020434033e-07, + "loss": 0.4207, + "step": 8345 + }, + { + "epoch": 2.6704351535836177, + "grad_norm": 0.5184808969497681, + "learning_rate": 3.6255564713050085e-07, + "loss": 0.4305, + "step": 8346 + }, + { + "epoch": 2.6707551194539247, + "grad_norm": 0.4785478711128235, + "learning_rate": 3.618599352048363e-07, + "loss": 0.477, + "step": 8347 + }, + { + "epoch": 2.671075085324232, + "grad_norm": 0.47505947947502136, + "learning_rate": 3.611648663628725e-07, + "loss": 0.3898, + "step": 8348 + }, + { + "epoch": 2.6713950511945392, + "grad_norm": 0.5582399964332581, + "learning_rate": 3.604704407009801e-07, + "loss": 0.4998, + "step": 8349 + }, + { + "epoch": 2.6717150170648463, + "grad_norm": 0.5541906952857971, + "learning_rate": 3.5977665831544217e-07, + "loss": 0.4653, + "step": 8350 + }, + { + "epoch": 2.6720349829351537, + "grad_norm": 0.512864351272583, + "learning_rate": 3.5908351930245335e-07, + "loss": 0.4074, + "step": 8351 + }, + { + "epoch": 2.6723549488054608, + "grad_norm": 0.4545230269432068, + "learning_rate": 3.5839102375811675e-07, + "loss": 0.4042, + "step": 8352 + }, + { + "epoch": 2.672674914675768, + "grad_norm": 0.4630187153816223, + "learning_rate": 3.5769917177845006e-07, + "loss": 0.4368, + "step": 8353 + }, + { + "epoch": 2.6729948805460753, + "grad_norm": 0.550320029258728, + "learning_rate": 3.570079634593776e-07, + "loss": 0.4591, + "step": 8354 + }, + { + "epoch": 2.6733148464163823, + "grad_norm": 0.4963036775588989, + "learning_rate": 3.563173988967361e-07, + "loss": 0.4149, + "step": 8355 + }, + { + "epoch": 2.6736348122866893, + "grad_norm": 0.49244219064712524, + "learning_rate": 3.5562747818627453e-07, + "loss": 0.4326, + "step": 8356 + }, + { + "epoch": 2.673954778156997, + "grad_norm": 0.4770994484424591, + "learning_rate": 3.5493820142364963e-07, + "loss": 0.3859, + "step": 8357 + }, + { + "epoch": 2.674274744027304, + "grad_norm": 0.46881797909736633, + "learning_rate": 3.5424956870443226e-07, + "loss": 0.4083, + "step": 8358 + }, + { + "epoch": 2.674594709897611, + "grad_norm": 0.5601597428321838, + "learning_rate": 3.535615801240999e-07, + "loss": 0.5077, + "step": 8359 + }, + { + "epoch": 2.6749146757679183, + "grad_norm": 0.5159189105033875, + "learning_rate": 3.528742357780457e-07, + "loss": 0.4357, + "step": 8360 + }, + { + "epoch": 2.6752346416382253, + "grad_norm": 0.4607735574245453, + "learning_rate": 3.5218753576156837e-07, + "loss": 0.3934, + "step": 8361 + }, + { + "epoch": 2.6755546075085324, + "grad_norm": 0.5549046397209167, + "learning_rate": 3.5150148016988073e-07, + "loss": 0.4223, + "step": 8362 + }, + { + "epoch": 2.67587457337884, + "grad_norm": 0.49229738116264343, + "learning_rate": 3.508160690981055e-07, + "loss": 0.4252, + "step": 8363 + }, + { + "epoch": 2.676194539249147, + "grad_norm": 0.4866524636745453, + "learning_rate": 3.50131302641275e-07, + "loss": 0.4251, + "step": 8364 + }, + { + "epoch": 2.676514505119454, + "grad_norm": 0.529787003993988, + "learning_rate": 3.4944718089433337e-07, + "loss": 0.435, + "step": 8365 + }, + { + "epoch": 2.6768344709897613, + "grad_norm": 0.5101364850997925, + "learning_rate": 3.4876370395213464e-07, + "loss": 0.4214, + "step": 8366 + }, + { + "epoch": 2.6771544368600684, + "grad_norm": 0.5263444185256958, + "learning_rate": 3.4808087190944473e-07, + "loss": 0.4555, + "step": 8367 + }, + { + "epoch": 2.6774744027303754, + "grad_norm": 0.47625741362571716, + "learning_rate": 3.473986848609384e-07, + "loss": 0.4217, + "step": 8368 + }, + { + "epoch": 2.6777943686006824, + "grad_norm": 0.5141160488128662, + "learning_rate": 3.4671714290120063e-07, + "loss": 0.4497, + "step": 8369 + }, + { + "epoch": 2.67811433447099, + "grad_norm": 0.4759281873703003, + "learning_rate": 3.4603624612472966e-07, + "loss": 0.423, + "step": 8370 + }, + { + "epoch": 2.678434300341297, + "grad_norm": 0.5235477685928345, + "learning_rate": 3.453559946259316e-07, + "loss": 0.457, + "step": 8371 + }, + { + "epoch": 2.678754266211604, + "grad_norm": 0.45080694556236267, + "learning_rate": 3.44676388499125e-07, + "loss": 0.3974, + "step": 8372 + }, + { + "epoch": 2.6790742320819114, + "grad_norm": 0.48992204666137695, + "learning_rate": 3.4399742783853773e-07, + "loss": 0.477, + "step": 8373 + }, + { + "epoch": 2.6793941979522184, + "grad_norm": 0.4760851263999939, + "learning_rate": 3.433191127383079e-07, + "loss": 0.4041, + "step": 8374 + }, + { + "epoch": 2.6797141638225255, + "grad_norm": 0.5457757711410522, + "learning_rate": 3.426414432924857e-07, + "loss": 0.487, + "step": 8375 + }, + { + "epoch": 2.6800341296928325, + "grad_norm": 0.46977418661117554, + "learning_rate": 3.4196441959502944e-07, + "loss": 0.4664, + "step": 8376 + }, + { + "epoch": 2.68035409556314, + "grad_norm": 0.43681231141090393, + "learning_rate": 3.412880417398112e-07, + "loss": 0.3664, + "step": 8377 + }, + { + "epoch": 2.680674061433447, + "grad_norm": 0.498920202255249, + "learning_rate": 3.4061230982061043e-07, + "loss": 0.4783, + "step": 8378 + }, + { + "epoch": 2.680994027303754, + "grad_norm": 0.49872443079948425, + "learning_rate": 3.3993722393111775e-07, + "loss": 0.4546, + "step": 8379 + }, + { + "epoch": 2.6813139931740615, + "grad_norm": 0.4891470968723297, + "learning_rate": 3.392627841649354e-07, + "loss": 0.4265, + "step": 8380 + }, + { + "epoch": 2.6816339590443685, + "grad_norm": 0.5192527770996094, + "learning_rate": 3.385889906155748e-07, + "loss": 0.4872, + "step": 8381 + }, + { + "epoch": 2.6819539249146755, + "grad_norm": 0.488872766494751, + "learning_rate": 3.37915843376459e-07, + "loss": 0.38, + "step": 8382 + }, + { + "epoch": 2.682273890784983, + "grad_norm": 0.47903430461883545, + "learning_rate": 3.3724334254091997e-07, + "loss": 0.4155, + "step": 8383 + }, + { + "epoch": 2.68259385665529, + "grad_norm": 0.48480987548828125, + "learning_rate": 3.365714882022003e-07, + "loss": 0.4463, + "step": 8384 + }, + { + "epoch": 2.682913822525597, + "grad_norm": 0.44706690311431885, + "learning_rate": 3.35900280453455e-07, + "loss": 0.4247, + "step": 8385 + }, + { + "epoch": 2.6832337883959045, + "grad_norm": 0.4883236289024353, + "learning_rate": 3.3522971938774575e-07, + "loss": 0.4457, + "step": 8386 + }, + { + "epoch": 2.6835537542662116, + "grad_norm": 0.5534349679946899, + "learning_rate": 3.3455980509804865e-07, + "loss": 0.4596, + "step": 8387 + }, + { + "epoch": 2.6838737201365186, + "grad_norm": 0.4790193736553192, + "learning_rate": 3.3389053767724723e-07, + "loss": 0.3802, + "step": 8388 + }, + { + "epoch": 2.684193686006826, + "grad_norm": 0.5198260545730591, + "learning_rate": 3.332219172181356e-07, + "loss": 0.4388, + "step": 8389 + }, + { + "epoch": 2.684513651877133, + "grad_norm": 0.49348023533821106, + "learning_rate": 3.325539438134201e-07, + "loss": 0.3952, + "step": 8390 + }, + { + "epoch": 2.68483361774744, + "grad_norm": 0.5152326226234436, + "learning_rate": 3.318866175557145e-07, + "loss": 0.4842, + "step": 8391 + }, + { + "epoch": 2.6851535836177476, + "grad_norm": 0.49283209443092346, + "learning_rate": 3.312199385375464e-07, + "loss": 0.4116, + "step": 8392 + }, + { + "epoch": 2.6854735494880546, + "grad_norm": 0.5174846649169922, + "learning_rate": 3.3055390685135026e-07, + "loss": 0.3847, + "step": 8393 + }, + { + "epoch": 2.6857935153583616, + "grad_norm": 0.565240740776062, + "learning_rate": 3.2988852258947155e-07, + "loss": 0.4801, + "step": 8394 + }, + { + "epoch": 2.686113481228669, + "grad_norm": 0.5280541777610779, + "learning_rate": 3.2922378584416823e-07, + "loss": 0.4187, + "step": 8395 + }, + { + "epoch": 2.686433447098976, + "grad_norm": 0.5336593985557556, + "learning_rate": 3.285596967076055e-07, + "loss": 0.4802, + "step": 8396 + }, + { + "epoch": 2.686753412969283, + "grad_norm": 0.5000171065330505, + "learning_rate": 3.278962552718612e-07, + "loss": 0.3496, + "step": 8397 + }, + { + "epoch": 2.6870733788395906, + "grad_norm": 0.5063610672950745, + "learning_rate": 3.2723346162892186e-07, + "loss": 0.4361, + "step": 8398 + }, + { + "epoch": 2.6873933447098977, + "grad_norm": 0.49247491359710693, + "learning_rate": 3.26571315870684e-07, + "loss": 0.4618, + "step": 8399 + }, + { + "epoch": 2.6877133105802047, + "grad_norm": 0.5019463300704956, + "learning_rate": 3.2590981808895637e-07, + "loss": 0.4603, + "step": 8400 + }, + { + "epoch": 2.688033276450512, + "grad_norm": 0.5265095829963684, + "learning_rate": 3.252489683754545e-07, + "loss": 0.4829, + "step": 8401 + }, + { + "epoch": 2.688353242320819, + "grad_norm": 0.4545130431652069, + "learning_rate": 3.245887668218078e-07, + "loss": 0.3999, + "step": 8402 + }, + { + "epoch": 2.688673208191126, + "grad_norm": 0.4803597927093506, + "learning_rate": 3.239292135195532e-07, + "loss": 0.4609, + "step": 8403 + }, + { + "epoch": 2.6889931740614337, + "grad_norm": 0.49703601002693176, + "learning_rate": 3.232703085601374e-07, + "loss": 0.4472, + "step": 8404 + }, + { + "epoch": 2.6893131399317407, + "grad_norm": 0.5202013254165649, + "learning_rate": 3.226120520349202e-07, + "loss": 0.435, + "step": 8405 + }, + { + "epoch": 2.6896331058020477, + "grad_norm": 0.529909610748291, + "learning_rate": 3.219544440351696e-07, + "loss": 0.4801, + "step": 8406 + }, + { + "epoch": 2.689953071672355, + "grad_norm": 0.511505663394928, + "learning_rate": 3.2129748465206336e-07, + "loss": 0.4322, + "step": 8407 + }, + { + "epoch": 2.6902730375426622, + "grad_norm": 0.5222049355506897, + "learning_rate": 3.2064117397668903e-07, + "loss": 0.4835, + "step": 8408 + }, + { + "epoch": 2.6905930034129693, + "grad_norm": 0.44696512818336487, + "learning_rate": 3.1998551210004504e-07, + "loss": 0.4163, + "step": 8409 + }, + { + "epoch": 2.6909129692832767, + "grad_norm": 0.4778428077697754, + "learning_rate": 3.193304991130397e-07, + "loss": 0.4261, + "step": 8410 + }, + { + "epoch": 2.6912329351535837, + "grad_norm": 0.46592646837234497, + "learning_rate": 3.1867613510649264e-07, + "loss": 0.4639, + "step": 8411 + }, + { + "epoch": 2.6915529010238908, + "grad_norm": 0.41280320286750793, + "learning_rate": 3.180224201711307e-07, + "loss": 0.3912, + "step": 8412 + }, + { + "epoch": 2.691872866894198, + "grad_norm": 0.48914268612861633, + "learning_rate": 3.173693543975931e-07, + "loss": 0.4419, + "step": 8413 + }, + { + "epoch": 2.6921928327645053, + "grad_norm": 0.5670159459114075, + "learning_rate": 3.167169378764273e-07, + "loss": 0.4238, + "step": 8414 + }, + { + "epoch": 2.6925127986348123, + "grad_norm": 0.5115007162094116, + "learning_rate": 3.1606517069809216e-07, + "loss": 0.4451, + "step": 8415 + }, + { + "epoch": 2.6928327645051193, + "grad_norm": 0.5018688440322876, + "learning_rate": 3.1541405295295645e-07, + "loss": 0.4524, + "step": 8416 + }, + { + "epoch": 2.6931527303754264, + "grad_norm": 0.5401268601417542, + "learning_rate": 3.14763584731298e-07, + "loss": 0.4099, + "step": 8417 + }, + { + "epoch": 2.693472696245734, + "grad_norm": 0.5169517993927002, + "learning_rate": 3.1411376612330513e-07, + "loss": 0.4834, + "step": 8418 + }, + { + "epoch": 2.693792662116041, + "grad_norm": 0.4915733337402344, + "learning_rate": 3.134645972190753e-07, + "loss": 0.4227, + "step": 8419 + }, + { + "epoch": 2.694112627986348, + "grad_norm": 0.5028654336929321, + "learning_rate": 3.12816078108617e-07, + "loss": 0.4139, + "step": 8420 + }, + { + "epoch": 2.6944325938566553, + "grad_norm": 0.5588289499282837, + "learning_rate": 3.1216820888184895e-07, + "loss": 0.5212, + "step": 8421 + }, + { + "epoch": 2.6947525597269624, + "grad_norm": 0.5030496716499329, + "learning_rate": 3.115209896285981e-07, + "loss": 0.4438, + "step": 8422 + }, + { + "epoch": 2.6950725255972694, + "grad_norm": 0.4874051809310913, + "learning_rate": 3.1087442043860274e-07, + "loss": 0.459, + "step": 8423 + }, + { + "epoch": 2.695392491467577, + "grad_norm": 0.4535805583000183, + "learning_rate": 3.102285014015094e-07, + "loss": 0.3931, + "step": 8424 + }, + { + "epoch": 2.695712457337884, + "grad_norm": 0.4591400921344757, + "learning_rate": 3.095832326068765e-07, + "loss": 0.415, + "step": 8425 + }, + { + "epoch": 2.696032423208191, + "grad_norm": 0.4513499140739441, + "learning_rate": 3.089386141441714e-07, + "loss": 0.3891, + "step": 8426 + }, + { + "epoch": 2.6963523890784984, + "grad_norm": 0.46307820081710815, + "learning_rate": 3.082946461027708e-07, + "loss": 0.4436, + "step": 8427 + }, + { + "epoch": 2.6966723549488054, + "grad_norm": 0.45751821994781494, + "learning_rate": 3.076513285719623e-07, + "loss": 0.4138, + "step": 8428 + }, + { + "epoch": 2.6969923208191124, + "grad_norm": 0.5338154435157776, + "learning_rate": 3.070086616409412e-07, + "loss": 0.4908, + "step": 8429 + }, + { + "epoch": 2.69731228668942, + "grad_norm": 0.4924832880496979, + "learning_rate": 3.063666453988151e-07, + "loss": 0.4392, + "step": 8430 + }, + { + "epoch": 2.697632252559727, + "grad_norm": 0.5246785879135132, + "learning_rate": 3.0572527993460054e-07, + "loss": 0.4646, + "step": 8431 + }, + { + "epoch": 2.697952218430034, + "grad_norm": 0.4785500764846802, + "learning_rate": 3.0508456533722363e-07, + "loss": 0.3924, + "step": 8432 + }, + { + "epoch": 2.6982721843003414, + "grad_norm": 0.4865003824234009, + "learning_rate": 3.044445016955194e-07, + "loss": 0.4422, + "step": 8433 + }, + { + "epoch": 2.6985921501706485, + "grad_norm": 0.4638444781303406, + "learning_rate": 3.038050890982336e-07, + "loss": 0.4232, + "step": 8434 + }, + { + "epoch": 2.6989121160409555, + "grad_norm": 0.48224329948425293, + "learning_rate": 3.031663276340219e-07, + "loss": 0.482, + "step": 8435 + }, + { + "epoch": 2.699232081911263, + "grad_norm": 0.4687730669975281, + "learning_rate": 3.0252821739144954e-07, + "loss": 0.407, + "step": 8436 + }, + { + "epoch": 2.69955204778157, + "grad_norm": 0.4968908131122589, + "learning_rate": 3.018907584589914e-07, + "loss": 0.4475, + "step": 8437 + }, + { + "epoch": 2.699872013651877, + "grad_norm": 0.5187657475471497, + "learning_rate": 3.012539509250317e-07, + "loss": 0.4177, + "step": 8438 + }, + { + "epoch": 2.7001919795221845, + "grad_norm": 0.5040234327316284, + "learning_rate": 3.0061779487786325e-07, + "loss": 0.4708, + "step": 8439 + }, + { + "epoch": 2.7005119453924915, + "grad_norm": 0.4529518485069275, + "learning_rate": 2.9998229040569147e-07, + "loss": 0.4633, + "step": 8440 + }, + { + "epoch": 2.7008319112627985, + "grad_norm": 0.4898926615715027, + "learning_rate": 2.9934743759662985e-07, + "loss": 0.4232, + "step": 8441 + }, + { + "epoch": 2.701151877133106, + "grad_norm": 0.4939773976802826, + "learning_rate": 2.987132365387013e-07, + "loss": 0.5227, + "step": 8442 + }, + { + "epoch": 2.701471843003413, + "grad_norm": 0.5070110559463501, + "learning_rate": 2.980796873198383e-07, + "loss": 0.4193, + "step": 8443 + }, + { + "epoch": 2.70179180887372, + "grad_norm": 0.5593889355659485, + "learning_rate": 2.974467900278821e-07, + "loss": 0.4221, + "step": 8444 + }, + { + "epoch": 2.7021117747440275, + "grad_norm": 0.511501669883728, + "learning_rate": 2.96814544750586e-07, + "loss": 0.4591, + "step": 8445 + }, + { + "epoch": 2.7024317406143346, + "grad_norm": 0.4870849847793579, + "learning_rate": 2.9618295157561205e-07, + "loss": 0.4261, + "step": 8446 + }, + { + "epoch": 2.7027517064846416, + "grad_norm": 0.5267578363418579, + "learning_rate": 2.9555201059053073e-07, + "loss": 0.4783, + "step": 8447 + }, + { + "epoch": 2.703071672354949, + "grad_norm": 0.4592973589897156, + "learning_rate": 2.949217218828226e-07, + "loss": 0.4231, + "step": 8448 + }, + { + "epoch": 2.703391638225256, + "grad_norm": 0.5345248579978943, + "learning_rate": 2.942920855398773e-07, + "loss": 0.4243, + "step": 8449 + }, + { + "epoch": 2.703711604095563, + "grad_norm": 0.5086983442306519, + "learning_rate": 2.9366310164899603e-07, + "loss": 0.4987, + "step": 8450 + }, + { + "epoch": 2.7040315699658706, + "grad_norm": 0.4644456207752228, + "learning_rate": 2.9303477029738794e-07, + "loss": 0.4356, + "step": 8451 + }, + { + "epoch": 2.7043515358361776, + "grad_norm": 0.4368300139904022, + "learning_rate": 2.9240709157217107e-07, + "loss": 0.3591, + "step": 8452 + }, + { + "epoch": 2.7046715017064846, + "grad_norm": 0.5334351658821106, + "learning_rate": 2.9178006556037477e-07, + "loss": 0.4868, + "step": 8453 + }, + { + "epoch": 2.7049914675767917, + "grad_norm": 0.4681623876094818, + "learning_rate": 2.911536923489361e-07, + "loss": 0.3648, + "step": 8454 + }, + { + "epoch": 2.705311433447099, + "grad_norm": 0.507057249546051, + "learning_rate": 2.905279720247023e-07, + "loss": 0.4935, + "step": 8455 + }, + { + "epoch": 2.705631399317406, + "grad_norm": 0.4899614751338959, + "learning_rate": 2.8990290467443174e-07, + "loss": 0.4516, + "step": 8456 + }, + { + "epoch": 2.705951365187713, + "grad_norm": 0.4832085967063904, + "learning_rate": 2.8927849038479006e-07, + "loss": 0.4697, + "step": 8457 + }, + { + "epoch": 2.70627133105802, + "grad_norm": 0.5052345991134644, + "learning_rate": 2.886547292423525e-07, + "loss": 0.4208, + "step": 8458 + }, + { + "epoch": 2.7065912969283277, + "grad_norm": 0.49080780148506165, + "learning_rate": 2.8803162133360374e-07, + "loss": 0.4259, + "step": 8459 + }, + { + "epoch": 2.7069112627986347, + "grad_norm": 0.4580221474170685, + "learning_rate": 2.874091667449397e-07, + "loss": 0.4164, + "step": 8460 + }, + { + "epoch": 2.7072312286689417, + "grad_norm": 0.44764164090156555, + "learning_rate": 2.867873655626646e-07, + "loss": 0.464, + "step": 8461 + }, + { + "epoch": 2.707551194539249, + "grad_norm": 0.5174615383148193, + "learning_rate": 2.861662178729918e-07, + "loss": 0.465, + "step": 8462 + }, + { + "epoch": 2.707871160409556, + "grad_norm": 0.5332177877426147, + "learning_rate": 2.8554572376204347e-07, + "loss": 0.4716, + "step": 8463 + }, + { + "epoch": 2.7081911262798632, + "grad_norm": 0.4738543927669525, + "learning_rate": 2.849258833158514e-07, + "loss": 0.3845, + "step": 8464 + }, + { + "epoch": 2.7085110921501707, + "grad_norm": 0.5458797812461853, + "learning_rate": 2.8430669662035784e-07, + "loss": 0.4715, + "step": 8465 + }, + { + "epoch": 2.7088310580204777, + "grad_norm": 0.47262293100357056, + "learning_rate": 2.836881637614147e-07, + "loss": 0.4233, + "step": 8466 + }, + { + "epoch": 2.7091510238907848, + "grad_norm": 0.4871342182159424, + "learning_rate": 2.8307028482478185e-07, + "loss": 0.4052, + "step": 8467 + }, + { + "epoch": 2.7094709897610922, + "grad_norm": 0.5085920095443726, + "learning_rate": 2.824530598961284e-07, + "loss": 0.4503, + "step": 8468 + }, + { + "epoch": 2.7097909556313993, + "grad_norm": 0.5077427625656128, + "learning_rate": 2.8183648906103255e-07, + "loss": 0.4202, + "step": 8469 + }, + { + "epoch": 2.7101109215017063, + "grad_norm": 0.47658786177635193, + "learning_rate": 2.812205724049838e-07, + "loss": 0.4362, + "step": 8470 + }, + { + "epoch": 2.7104308873720138, + "grad_norm": 0.4591486155986786, + "learning_rate": 2.806053100133804e-07, + "loss": 0.4146, + "step": 8471 + }, + { + "epoch": 2.710750853242321, + "grad_norm": 0.5390037298202515, + "learning_rate": 2.799907019715281e-07, + "loss": 0.4764, + "step": 8472 + }, + { + "epoch": 2.711070819112628, + "grad_norm": 0.5763192772865295, + "learning_rate": 2.7937674836464256e-07, + "loss": 0.5183, + "step": 8473 + }, + { + "epoch": 2.7113907849829353, + "grad_norm": 0.44307756423950195, + "learning_rate": 2.787634492778507e-07, + "loss": 0.4003, + "step": 8474 + }, + { + "epoch": 2.7117107508532423, + "grad_norm": 0.46029427647590637, + "learning_rate": 2.781508047961862e-07, + "loss": 0.4284, + "step": 8475 + }, + { + "epoch": 2.7120307167235493, + "grad_norm": 0.5089544057846069, + "learning_rate": 2.775388150045932e-07, + "loss": 0.4204, + "step": 8476 + }, + { + "epoch": 2.712350682593857, + "grad_norm": 0.5391183495521545, + "learning_rate": 2.769274799879246e-07, + "loss": 0.4842, + "step": 8477 + }, + { + "epoch": 2.712670648464164, + "grad_norm": 0.4696575105190277, + "learning_rate": 2.7631679983094293e-07, + "loss": 0.4197, + "step": 8478 + }, + { + "epoch": 2.712990614334471, + "grad_norm": 0.5044712424278259, + "learning_rate": 2.757067746183195e-07, + "loss": 0.4379, + "step": 8479 + }, + { + "epoch": 2.7133105802047783, + "grad_norm": 0.5446345806121826, + "learning_rate": 2.750974044346355e-07, + "loss": 0.4504, + "step": 8480 + }, + { + "epoch": 2.7136305460750854, + "grad_norm": 0.5092166662216187, + "learning_rate": 2.744886893643806e-07, + "loss": 0.4543, + "step": 8481 + }, + { + "epoch": 2.7139505119453924, + "grad_norm": 0.48626455664634705, + "learning_rate": 2.7388062949195384e-07, + "loss": 0.4434, + "step": 8482 + }, + { + "epoch": 2.7142704778157, + "grad_norm": 0.4884015619754791, + "learning_rate": 2.7327322490166287e-07, + "loss": 0.4649, + "step": 8483 + }, + { + "epoch": 2.714590443686007, + "grad_norm": 0.5095328688621521, + "learning_rate": 2.726664756777264e-07, + "loss": 0.4626, + "step": 8484 + }, + { + "epoch": 2.714910409556314, + "grad_norm": 0.4839845299720764, + "learning_rate": 2.7206038190427e-07, + "loss": 0.4162, + "step": 8485 + }, + { + "epoch": 2.7152303754266214, + "grad_norm": 0.5074237585067749, + "learning_rate": 2.7145494366532966e-07, + "loss": 0.4135, + "step": 8486 + }, + { + "epoch": 2.7155503412969284, + "grad_norm": 0.5321500897407532, + "learning_rate": 2.7085016104484994e-07, + "loss": 0.4739, + "step": 8487 + }, + { + "epoch": 2.7158703071672354, + "grad_norm": 0.5675606727600098, + "learning_rate": 2.702460341266838e-07, + "loss": 0.4598, + "step": 8488 + }, + { + "epoch": 2.716190273037543, + "grad_norm": 0.45474451780319214, + "learning_rate": 2.6964256299459646e-07, + "loss": 0.3731, + "step": 8489 + }, + { + "epoch": 2.71651023890785, + "grad_norm": 0.5283628702163696, + "learning_rate": 2.6903974773225703e-07, + "loss": 0.5131, + "step": 8490 + }, + { + "epoch": 2.716830204778157, + "grad_norm": 0.4831763207912445, + "learning_rate": 2.684375884232493e-07, + "loss": 0.3843, + "step": 8491 + }, + { + "epoch": 2.7171501706484644, + "grad_norm": 0.48812925815582275, + "learning_rate": 2.6783608515106195e-07, + "loss": 0.4493, + "step": 8492 + }, + { + "epoch": 2.7174701365187715, + "grad_norm": 0.47067147493362427, + "learning_rate": 2.6723523799909387e-07, + "loss": 0.4413, + "step": 8493 + }, + { + "epoch": 2.7177901023890785, + "grad_norm": 0.4778099060058594, + "learning_rate": 2.6663504705065404e-07, + "loss": 0.4564, + "step": 8494 + }, + { + "epoch": 2.7181100682593855, + "grad_norm": 0.4619276523590088, + "learning_rate": 2.660355123889585e-07, + "loss": 0.3582, + "step": 8495 + }, + { + "epoch": 2.718430034129693, + "grad_norm": 0.5539404153823853, + "learning_rate": 2.654366340971354e-07, + "loss": 0.4361, + "step": 8496 + }, + { + "epoch": 2.71875, + "grad_norm": 0.5206783413887024, + "learning_rate": 2.648384122582187e-07, + "loss": 0.4379, + "step": 8497 + }, + { + "epoch": 2.719069965870307, + "grad_norm": 0.5146101117134094, + "learning_rate": 2.642408469551516e-07, + "loss": 0.4665, + "step": 8498 + }, + { + "epoch": 2.7193899317406145, + "grad_norm": 0.4704565703868866, + "learning_rate": 2.636439382707895e-07, + "loss": 0.448, + "step": 8499 + }, + { + "epoch": 2.7197098976109215, + "grad_norm": 0.4845392405986786, + "learning_rate": 2.630476862878928e-07, + "loss": 0.3829, + "step": 8500 + }, + { + "epoch": 2.7200298634812285, + "grad_norm": 0.5160871148109436, + "learning_rate": 2.6245209108913316e-07, + "loss": 0.4631, + "step": 8501 + }, + { + "epoch": 2.7203498293515356, + "grad_norm": 0.4980718791484833, + "learning_rate": 2.618571527570907e-07, + "loss": 0.491, + "step": 8502 + }, + { + "epoch": 2.720669795221843, + "grad_norm": 0.4564629793167114, + "learning_rate": 2.6126287137425376e-07, + "loss": 0.3945, + "step": 8503 + }, + { + "epoch": 2.72098976109215, + "grad_norm": 0.523655891418457, + "learning_rate": 2.6066924702302044e-07, + "loss": 0.4404, + "step": 8504 + }, + { + "epoch": 2.721309726962457, + "grad_norm": 0.46495863795280457, + "learning_rate": 2.6007627978569817e-07, + "loss": 0.44, + "step": 8505 + }, + { + "epoch": 2.7216296928327646, + "grad_norm": 0.5155764222145081, + "learning_rate": 2.594839697445017e-07, + "loss": 0.4447, + "step": 8506 + }, + { + "epoch": 2.7219496587030716, + "grad_norm": 0.5131596326828003, + "learning_rate": 2.588923169815555e-07, + "loss": 0.448, + "step": 8507 + }, + { + "epoch": 2.7222696245733786, + "grad_norm": 0.5179362893104553, + "learning_rate": 2.5830132157889377e-07, + "loss": 0.4763, + "step": 8508 + }, + { + "epoch": 2.722589590443686, + "grad_norm": 0.40810224413871765, + "learning_rate": 2.577109836184577e-07, + "loss": 0.3922, + "step": 8509 + }, + { + "epoch": 2.722909556313993, + "grad_norm": 0.48155176639556885, + "learning_rate": 2.5712130318209904e-07, + "loss": 0.4474, + "step": 8510 + }, + { + "epoch": 2.7232295221843, + "grad_norm": 0.4342872202396393, + "learning_rate": 2.5653228035157796e-07, + "loss": 0.3829, + "step": 8511 + }, + { + "epoch": 2.7235494880546076, + "grad_norm": 0.4778122007846832, + "learning_rate": 2.5594391520856187e-07, + "loss": 0.4624, + "step": 8512 + }, + { + "epoch": 2.7238694539249146, + "grad_norm": 0.4993056654930115, + "learning_rate": 2.5535620783463e-07, + "loss": 0.446, + "step": 8513 + }, + { + "epoch": 2.7241894197952217, + "grad_norm": 0.426123708486557, + "learning_rate": 2.5476915831126715e-07, + "loss": 0.4275, + "step": 8514 + }, + { + "epoch": 2.724509385665529, + "grad_norm": 0.4558579623699188, + "learning_rate": 2.5418276671986984e-07, + "loss": 0.4425, + "step": 8515 + }, + { + "epoch": 2.724829351535836, + "grad_norm": 0.4622271955013275, + "learning_rate": 2.5359703314174147e-07, + "loss": 0.4783, + "step": 8516 + }, + { + "epoch": 2.725149317406143, + "grad_norm": 0.4740881025791168, + "learning_rate": 2.530119576580936e-07, + "loss": 0.4173, + "step": 8517 + }, + { + "epoch": 2.7254692832764507, + "grad_norm": 0.4327777326107025, + "learning_rate": 2.5242754035004924e-07, + "loss": 0.3782, + "step": 8518 + }, + { + "epoch": 2.7257892491467577, + "grad_norm": 0.470712810754776, + "learning_rate": 2.518437812986374e-07, + "loss": 0.4403, + "step": 8519 + }, + { + "epoch": 2.7261092150170647, + "grad_norm": 0.5009084939956665, + "learning_rate": 2.512606805847978e-07, + "loss": 0.4344, + "step": 8520 + }, + { + "epoch": 2.726429180887372, + "grad_norm": 0.5207602381706238, + "learning_rate": 2.50678238289378e-07, + "loss": 0.4582, + "step": 8521 + }, + { + "epoch": 2.726749146757679, + "grad_norm": 0.5586817264556885, + "learning_rate": 2.500964544931328e-07, + "loss": 0.4734, + "step": 8522 + }, + { + "epoch": 2.7270691126279862, + "grad_norm": 0.5082388520240784, + "learning_rate": 2.495153292767299e-07, + "loss": 0.3952, + "step": 8523 + }, + { + "epoch": 2.7273890784982937, + "grad_norm": 0.46565231680870056, + "learning_rate": 2.4893486272074055e-07, + "loss": 0.4361, + "step": 8524 + }, + { + "epoch": 2.7277090443686007, + "grad_norm": 0.46218451857566833, + "learning_rate": 2.4835505490564846e-07, + "loss": 0.4188, + "step": 8525 + }, + { + "epoch": 2.7280290102389078, + "grad_norm": 0.483097106218338, + "learning_rate": 2.477759059118451e-07, + "loss": 0.3994, + "step": 8526 + }, + { + "epoch": 2.7283489761092152, + "grad_norm": 0.5206490159034729, + "learning_rate": 2.471974158196283e-07, + "loss": 0.4603, + "step": 8527 + }, + { + "epoch": 2.7286689419795223, + "grad_norm": 0.4444100558757782, + "learning_rate": 2.4661958470920845e-07, + "loss": 0.4267, + "step": 8528 + }, + { + "epoch": 2.7289889078498293, + "grad_norm": 0.4927453398704529, + "learning_rate": 2.4604241266070085e-07, + "loss": 0.4268, + "step": 8529 + }, + { + "epoch": 2.7293088737201368, + "grad_norm": 0.46823206543922424, + "learning_rate": 2.454658997541326e-07, + "loss": 0.389, + "step": 8530 + }, + { + "epoch": 2.729628839590444, + "grad_norm": 0.5520978569984436, + "learning_rate": 2.44890046069437e-07, + "loss": 0.5481, + "step": 8531 + }, + { + "epoch": 2.729948805460751, + "grad_norm": 0.4668036699295044, + "learning_rate": 2.4431485168645685e-07, + "loss": 0.4141, + "step": 8532 + }, + { + "epoch": 2.7302687713310583, + "grad_norm": 0.49614331126213074, + "learning_rate": 2.437403166849439e-07, + "loss": 0.4528, + "step": 8533 + }, + { + "epoch": 2.7305887372013653, + "grad_norm": 0.501336932182312, + "learning_rate": 2.431664411445578e-07, + "loss": 0.4259, + "step": 8534 + }, + { + "epoch": 2.7309087030716723, + "grad_norm": 0.5080056190490723, + "learning_rate": 2.425932251448676e-07, + "loss": 0.4172, + "step": 8535 + }, + { + "epoch": 2.73122866894198, + "grad_norm": 0.5129799842834473, + "learning_rate": 2.420206687653498e-07, + "loss": 0.4751, + "step": 8536 + }, + { + "epoch": 2.731548634812287, + "grad_norm": 0.5177556872367859, + "learning_rate": 2.4144877208538974e-07, + "loss": 0.4168, + "step": 8537 + }, + { + "epoch": 2.731868600682594, + "grad_norm": 0.4787764549255371, + "learning_rate": 2.408775351842829e-07, + "loss": 0.456, + "step": 8538 + }, + { + "epoch": 2.732188566552901, + "grad_norm": 0.5124397277832031, + "learning_rate": 2.4030695814123093e-07, + "loss": 0.4248, + "step": 8539 + }, + { + "epoch": 2.7325085324232083, + "grad_norm": 0.5004435181617737, + "learning_rate": 2.397370410353456e-07, + "loss": 0.4468, + "step": 8540 + }, + { + "epoch": 2.7328284982935154, + "grad_norm": 0.4988056421279907, + "learning_rate": 2.391677839456458e-07, + "loss": 0.4366, + "step": 8541 + }, + { + "epoch": 2.7331484641638224, + "grad_norm": 0.5729777812957764, + "learning_rate": 2.385991869510601e-07, + "loss": 0.4458, + "step": 8542 + }, + { + "epoch": 2.7334684300341294, + "grad_norm": 0.4886725842952728, + "learning_rate": 2.380312501304255e-07, + "loss": 0.45, + "step": 8543 + }, + { + "epoch": 2.733788395904437, + "grad_norm": 0.4933490753173828, + "learning_rate": 2.3746397356248617e-07, + "loss": 0.4277, + "step": 8544 + }, + { + "epoch": 2.734108361774744, + "grad_norm": 0.5624293684959412, + "learning_rate": 2.3689735732589757e-07, + "loss": 0.4675, + "step": 8545 + }, + { + "epoch": 2.734428327645051, + "grad_norm": 0.48448383808135986, + "learning_rate": 2.3633140149921962e-07, + "loss": 0.4117, + "step": 8546 + }, + { + "epoch": 2.7347482935153584, + "grad_norm": 0.47338375449180603, + "learning_rate": 2.35766106160924e-07, + "loss": 0.4596, + "step": 8547 + }, + { + "epoch": 2.7350682593856654, + "grad_norm": 0.43746209144592285, + "learning_rate": 2.3520147138938965e-07, + "loss": 0.4113, + "step": 8548 + }, + { + "epoch": 2.7353882252559725, + "grad_norm": 0.44164198637008667, + "learning_rate": 2.3463749726290287e-07, + "loss": 0.4306, + "step": 8549 + }, + { + "epoch": 2.73570819112628, + "grad_norm": 0.46448519825935364, + "learning_rate": 2.3407418385966053e-07, + "loss": 0.4599, + "step": 8550 + }, + { + "epoch": 2.736028156996587, + "grad_norm": 0.46389082074165344, + "learning_rate": 2.3351153125776515e-07, + "loss": 0.4802, + "step": 8551 + }, + { + "epoch": 2.736348122866894, + "grad_norm": 0.4623454511165619, + "learning_rate": 2.3294953953523104e-07, + "loss": 0.3895, + "step": 8552 + }, + { + "epoch": 2.7366680887372015, + "grad_norm": 0.5046341419219971, + "learning_rate": 2.323882087699786e-07, + "loss": 0.4447, + "step": 8553 + }, + { + "epoch": 2.7369880546075085, + "grad_norm": 0.47571954131126404, + "learning_rate": 2.3182753903983556e-07, + "loss": 0.4287, + "step": 8554 + }, + { + "epoch": 2.7373080204778155, + "grad_norm": 0.5265892744064331, + "learning_rate": 2.312675304225409e-07, + "loss": 0.4532, + "step": 8555 + }, + { + "epoch": 2.737627986348123, + "grad_norm": 0.5027657747268677, + "learning_rate": 2.3070818299573972e-07, + "loss": 0.4218, + "step": 8556 + }, + { + "epoch": 2.73794795221843, + "grad_norm": 0.5723645091056824, + "learning_rate": 2.3014949683698717e-07, + "loss": 0.5078, + "step": 8557 + }, + { + "epoch": 2.738267918088737, + "grad_norm": 0.519535481929779, + "learning_rate": 2.2959147202374522e-07, + "loss": 0.4089, + "step": 8558 + }, + { + "epoch": 2.7385878839590445, + "grad_norm": 0.5125702619552612, + "learning_rate": 2.290341086333836e-07, + "loss": 0.4165, + "step": 8559 + }, + { + "epoch": 2.7389078498293515, + "grad_norm": 0.5632970929145813, + "learning_rate": 2.2847740674318385e-07, + "loss": 0.5036, + "step": 8560 + }, + { + "epoch": 2.7392278156996586, + "grad_norm": 0.48991167545318604, + "learning_rate": 2.2792136643033035e-07, + "loss": 0.4415, + "step": 8561 + }, + { + "epoch": 2.739547781569966, + "grad_norm": 0.4932671785354614, + "learning_rate": 2.27365987771922e-07, + "loss": 0.413, + "step": 8562 + }, + { + "epoch": 2.739867747440273, + "grad_norm": 0.5287392735481262, + "learning_rate": 2.2681127084496047e-07, + "loss": 0.4796, + "step": 8563 + }, + { + "epoch": 2.74018771331058, + "grad_norm": 0.5054778456687927, + "learning_rate": 2.2625721572635762e-07, + "loss": 0.382, + "step": 8564 + }, + { + "epoch": 2.7405076791808876, + "grad_norm": 0.4676485061645508, + "learning_rate": 2.2570382249293586e-07, + "loss": 0.3968, + "step": 8565 + }, + { + "epoch": 2.7408276450511946, + "grad_norm": 0.47252145409584045, + "learning_rate": 2.251510912214222e-07, + "loss": 0.4577, + "step": 8566 + }, + { + "epoch": 2.7411476109215016, + "grad_norm": 0.4993182122707367, + "learning_rate": 2.2459902198845419e-07, + "loss": 0.4753, + "step": 8567 + }, + { + "epoch": 2.741467576791809, + "grad_norm": 0.48656511306762695, + "learning_rate": 2.2404761487057726e-07, + "loss": 0.4018, + "step": 8568 + }, + { + "epoch": 2.741787542662116, + "grad_norm": 0.573748767375946, + "learning_rate": 2.2349686994424303e-07, + "loss": 0.4718, + "step": 8569 + }, + { + "epoch": 2.742107508532423, + "grad_norm": 0.5051923990249634, + "learning_rate": 2.2294678728581487e-07, + "loss": 0.4768, + "step": 8570 + }, + { + "epoch": 2.7424274744027306, + "grad_norm": 0.4523454010486603, + "learning_rate": 2.2239736697156067e-07, + "loss": 0.4197, + "step": 8571 + }, + { + "epoch": 2.7427474402730376, + "grad_norm": 0.4958074986934662, + "learning_rate": 2.2184860907766003e-07, + "loss": 0.3901, + "step": 8572 + }, + { + "epoch": 2.7430674061433447, + "grad_norm": 0.5585159659385681, + "learning_rate": 2.2130051368019823e-07, + "loss": 0.5015, + "step": 8573 + }, + { + "epoch": 2.743387372013652, + "grad_norm": 0.4593925476074219, + "learning_rate": 2.2075308085516835e-07, + "loss": 0.4388, + "step": 8574 + }, + { + "epoch": 2.743707337883959, + "grad_norm": 0.5206596255302429, + "learning_rate": 2.202063106784741e-07, + "loss": 0.4959, + "step": 8575 + }, + { + "epoch": 2.744027303754266, + "grad_norm": 0.5175601840019226, + "learning_rate": 2.196602032259243e-07, + "loss": 0.4169, + "step": 8576 + }, + { + "epoch": 2.7443472696245736, + "grad_norm": 0.4904636740684509, + "learning_rate": 2.1911475857323893e-07, + "loss": 0.4093, + "step": 8577 + }, + { + "epoch": 2.7446672354948807, + "grad_norm": 0.5445221662521362, + "learning_rate": 2.1856997679604363e-07, + "loss": 0.4526, + "step": 8578 + }, + { + "epoch": 2.7449872013651877, + "grad_norm": 0.48061031103134155, + "learning_rate": 2.1802585796987243e-07, + "loss": 0.4406, + "step": 8579 + }, + { + "epoch": 2.7453071672354947, + "grad_norm": 0.5459372401237488, + "learning_rate": 2.1748240217016936e-07, + "loss": 0.4488, + "step": 8580 + }, + { + "epoch": 2.745627133105802, + "grad_norm": 0.49022161960601807, + "learning_rate": 2.1693960947228532e-07, + "loss": 0.4235, + "step": 8581 + }, + { + "epoch": 2.7459470989761092, + "grad_norm": 0.5033886432647705, + "learning_rate": 2.1639747995147843e-07, + "loss": 0.4521, + "step": 8582 + }, + { + "epoch": 2.7462670648464163, + "grad_norm": 0.4898601472377777, + "learning_rate": 2.1585601368291575e-07, + "loss": 0.468, + "step": 8583 + }, + { + "epoch": 2.7465870307167233, + "grad_norm": 0.4970014691352844, + "learning_rate": 2.1531521074167172e-07, + "loss": 0.5213, + "step": 8584 + }, + { + "epoch": 2.7469069965870307, + "grad_norm": 0.44684627652168274, + "learning_rate": 2.1477507120272967e-07, + "loss": 0.3995, + "step": 8585 + }, + { + "epoch": 2.7472269624573378, + "grad_norm": 0.5002879500389099, + "learning_rate": 2.1423559514098135e-07, + "loss": 0.438, + "step": 8586 + }, + { + "epoch": 2.747546928327645, + "grad_norm": 0.5029903054237366, + "learning_rate": 2.1369678263122583e-07, + "loss": 0.405, + "step": 8587 + }, + { + "epoch": 2.7478668941979523, + "grad_norm": 0.49236345291137695, + "learning_rate": 2.1315863374816893e-07, + "loss": 0.4375, + "step": 8588 + }, + { + "epoch": 2.7481868600682593, + "grad_norm": 0.44991016387939453, + "learning_rate": 2.1262114856642535e-07, + "loss": 0.3779, + "step": 8589 + }, + { + "epoch": 2.7485068259385663, + "grad_norm": 0.5005273818969727, + "learning_rate": 2.1208432716051886e-07, + "loss": 0.4834, + "step": 8590 + }, + { + "epoch": 2.748826791808874, + "grad_norm": 0.4878215491771698, + "learning_rate": 2.1154816960488157e-07, + "loss": 0.449, + "step": 8591 + }, + { + "epoch": 2.749146757679181, + "grad_norm": 0.45098960399627686, + "learning_rate": 2.1101267597385066e-07, + "loss": 0.3973, + "step": 8592 + }, + { + "epoch": 2.749466723549488, + "grad_norm": 0.47468888759613037, + "learning_rate": 2.1047784634167402e-07, + "loss": 0.4342, + "step": 8593 + }, + { + "epoch": 2.7497866894197953, + "grad_norm": 0.474040150642395, + "learning_rate": 2.0994368078250505e-07, + "loss": 0.461, + "step": 8594 + }, + { + "epoch": 2.7501066552901023, + "grad_norm": 0.4642796814441681, + "learning_rate": 2.094101793704073e-07, + "loss": 0.4474, + "step": 8595 + }, + { + "epoch": 2.7504266211604094, + "grad_norm": 0.4851500988006592, + "learning_rate": 2.0887734217935218e-07, + "loss": 0.4678, + "step": 8596 + }, + { + "epoch": 2.750746587030717, + "grad_norm": 0.5419781804084778, + "learning_rate": 2.0834516928321725e-07, + "loss": 0.476, + "step": 8597 + }, + { + "epoch": 2.751066552901024, + "grad_norm": 0.49836188554763794, + "learning_rate": 2.0781366075578902e-07, + "loss": 0.4385, + "step": 8598 + }, + { + "epoch": 2.751386518771331, + "grad_norm": 0.49485135078430176, + "learning_rate": 2.0728281667076133e-07, + "loss": 0.4694, + "step": 8599 + }, + { + "epoch": 2.7517064846416384, + "grad_norm": 0.5159654021263123, + "learning_rate": 2.0675263710173697e-07, + "loss": 0.452, + "step": 8600 + }, + { + "epoch": 2.7520264505119454, + "grad_norm": 0.4692765474319458, + "learning_rate": 2.0622312212222662e-07, + "loss": 0.4189, + "step": 8601 + }, + { + "epoch": 2.7523464163822524, + "grad_norm": 0.44056499004364014, + "learning_rate": 2.0569427180564704e-07, + "loss": 0.441, + "step": 8602 + }, + { + "epoch": 2.75266638225256, + "grad_norm": 0.51723313331604, + "learning_rate": 2.0516608622532518e-07, + "loss": 0.4848, + "step": 8603 + }, + { + "epoch": 2.752986348122867, + "grad_norm": 0.5240875482559204, + "learning_rate": 2.0463856545449246e-07, + "loss": 0.457, + "step": 8604 + }, + { + "epoch": 2.753306313993174, + "grad_norm": 0.4982944130897522, + "learning_rate": 2.0411170956629145e-07, + "loss": 0.4306, + "step": 8605 + }, + { + "epoch": 2.7536262798634814, + "grad_norm": 0.5124058723449707, + "learning_rate": 2.0358551863377264e-07, + "loss": 0.4723, + "step": 8606 + }, + { + "epoch": 2.7539462457337884, + "grad_norm": 0.44902369379997253, + "learning_rate": 2.030599927298915e-07, + "loss": 0.4156, + "step": 8607 + }, + { + "epoch": 2.7542662116040955, + "grad_norm": 0.4867861270904541, + "learning_rate": 2.0253513192751374e-07, + "loss": 0.4051, + "step": 8608 + }, + { + "epoch": 2.754586177474403, + "grad_norm": 0.48332637548446655, + "learning_rate": 2.0201093629941048e-07, + "loss": 0.4343, + "step": 8609 + }, + { + "epoch": 2.75490614334471, + "grad_norm": 0.5181475281715393, + "learning_rate": 2.014874059182631e-07, + "loss": 0.4967, + "step": 8610 + }, + { + "epoch": 2.755226109215017, + "grad_norm": 0.4403003454208374, + "learning_rate": 2.0096454085666016e-07, + "loss": 0.321, + "step": 8611 + }, + { + "epoch": 2.7555460750853245, + "grad_norm": 0.44011250138282776, + "learning_rate": 2.0044234118709704e-07, + "loss": 0.4645, + "step": 8612 + }, + { + "epoch": 2.7558660409556315, + "grad_norm": 0.4728182554244995, + "learning_rate": 1.9992080698197746e-07, + "loss": 0.4957, + "step": 8613 + }, + { + "epoch": 2.7561860068259385, + "grad_norm": 0.47703278064727783, + "learning_rate": 1.9939993831361137e-07, + "loss": 0.4213, + "step": 8614 + }, + { + "epoch": 2.756505972696246, + "grad_norm": 0.4957233667373657, + "learning_rate": 1.9887973525421988e-07, + "loss": 0.4404, + "step": 8615 + }, + { + "epoch": 2.756825938566553, + "grad_norm": 0.4639700651168823, + "learning_rate": 1.983601978759292e-07, + "loss": 0.4406, + "step": 8616 + }, + { + "epoch": 2.75714590443686, + "grad_norm": 0.4967518448829651, + "learning_rate": 1.978413262507739e-07, + "loss": 0.4768, + "step": 8617 + }, + { + "epoch": 2.7574658703071675, + "grad_norm": 0.5306515097618103, + "learning_rate": 1.9732312045069534e-07, + "loss": 0.4878, + "step": 8618 + }, + { + "epoch": 2.7577858361774745, + "grad_norm": 0.4871678352355957, + "learning_rate": 1.9680558054754384e-07, + "loss": 0.4272, + "step": 8619 + }, + { + "epoch": 2.7581058020477816, + "grad_norm": 0.448634535074234, + "learning_rate": 1.9628870661307698e-07, + "loss": 0.3849, + "step": 8620 + }, + { + "epoch": 2.7584257679180886, + "grad_norm": 0.47452399134635925, + "learning_rate": 1.957724987189602e-07, + "loss": 0.4708, + "step": 8621 + }, + { + "epoch": 2.758745733788396, + "grad_norm": 0.46345850825309753, + "learning_rate": 1.9525695693676628e-07, + "loss": 0.4261, + "step": 8622 + }, + { + "epoch": 2.759065699658703, + "grad_norm": 0.5059054493904114, + "learning_rate": 1.9474208133797524e-07, + "loss": 0.4908, + "step": 8623 + }, + { + "epoch": 2.75938566552901, + "grad_norm": 0.47566303610801697, + "learning_rate": 1.9422787199397553e-07, + "loss": 0.4469, + "step": 8624 + }, + { + "epoch": 2.7597056313993176, + "grad_norm": 0.4674135446548462, + "learning_rate": 1.9371432897606236e-07, + "loss": 0.3585, + "step": 8625 + }, + { + "epoch": 2.7600255972696246, + "grad_norm": 0.48827576637268066, + "learning_rate": 1.9320145235544042e-07, + "loss": 0.4537, + "step": 8626 + }, + { + "epoch": 2.7603455631399316, + "grad_norm": 0.5700124502182007, + "learning_rate": 1.926892422032195e-07, + "loss": 0.4124, + "step": 8627 + }, + { + "epoch": 2.7606655290102387, + "grad_norm": 0.5084264278411865, + "learning_rate": 1.9217769859041834e-07, + "loss": 0.4028, + "step": 8628 + }, + { + "epoch": 2.760985494880546, + "grad_norm": 0.5398024320602417, + "learning_rate": 1.9166682158796303e-07, + "loss": 0.4654, + "step": 8629 + }, + { + "epoch": 2.761305460750853, + "grad_norm": 0.5033600926399231, + "learning_rate": 1.9115661126668738e-07, + "loss": 0.4298, + "step": 8630 + }, + { + "epoch": 2.76162542662116, + "grad_norm": 0.491950124502182, + "learning_rate": 1.9064706769733378e-07, + "loss": 0.4715, + "step": 8631 + }, + { + "epoch": 2.7619453924914676, + "grad_norm": 0.5091946125030518, + "learning_rate": 1.9013819095054954e-07, + "loss": 0.4647, + "step": 8632 + }, + { + "epoch": 2.7622653583617747, + "grad_norm": 0.48028603196144104, + "learning_rate": 1.8962998109689156e-07, + "loss": 0.4132, + "step": 8633 + }, + { + "epoch": 2.7625853242320817, + "grad_norm": 0.5029857158660889, + "learning_rate": 1.8912243820682296e-07, + "loss": 0.486, + "step": 8634 + }, + { + "epoch": 2.762905290102389, + "grad_norm": 0.4869309067726135, + "learning_rate": 1.8861556235071632e-07, + "loss": 0.4707, + "step": 8635 + }, + { + "epoch": 2.763225255972696, + "grad_norm": 0.4437768757343292, + "learning_rate": 1.881093535988504e-07, + "loss": 0.3903, + "step": 8636 + }, + { + "epoch": 2.763545221843003, + "grad_norm": 0.5005133748054504, + "learning_rate": 1.8760381202141188e-07, + "loss": 0.4381, + "step": 8637 + }, + { + "epoch": 2.7638651877133107, + "grad_norm": 0.4766649901866913, + "learning_rate": 1.8709893768849406e-07, + "loss": 0.4338, + "step": 8638 + }, + { + "epoch": 2.7641851535836177, + "grad_norm": 0.4825212061405182, + "learning_rate": 1.8659473067009826e-07, + "loss": 0.4023, + "step": 8639 + }, + { + "epoch": 2.7645051194539247, + "grad_norm": 0.47575074434280396, + "learning_rate": 1.860911910361335e-07, + "loss": 0.4149, + "step": 8640 + }, + { + "epoch": 2.764825085324232, + "grad_norm": 0.5040088891983032, + "learning_rate": 1.855883188564167e-07, + "loss": 0.4926, + "step": 8641 + }, + { + "epoch": 2.7651450511945392, + "grad_norm": 0.4763101637363434, + "learning_rate": 1.8508611420067157e-07, + "loss": 0.406, + "step": 8642 + }, + { + "epoch": 2.7654650170648463, + "grad_norm": 0.5055087804794312, + "learning_rate": 1.8458457713852907e-07, + "loss": 0.485, + "step": 8643 + }, + { + "epoch": 2.7657849829351537, + "grad_norm": 0.527959406375885, + "learning_rate": 1.8408370773952744e-07, + "loss": 0.4299, + "step": 8644 + }, + { + "epoch": 2.7661049488054608, + "grad_norm": 0.49691763520240784, + "learning_rate": 1.8358350607311392e-07, + "loss": 0.4119, + "step": 8645 + }, + { + "epoch": 2.766424914675768, + "grad_norm": 0.4913119673728943, + "learning_rate": 1.8308397220864193e-07, + "loss": 0.406, + "step": 8646 + }, + { + "epoch": 2.7667448805460753, + "grad_norm": 0.4808558225631714, + "learning_rate": 1.8258510621537219e-07, + "loss": 0.4341, + "step": 8647 + }, + { + "epoch": 2.7670648464163823, + "grad_norm": 0.5470263361930847, + "learning_rate": 1.8208690816247266e-07, + "loss": 0.4325, + "step": 8648 + }, + { + "epoch": 2.7673848122866893, + "grad_norm": 0.5051500201225281, + "learning_rate": 1.815893781190181e-07, + "loss": 0.4475, + "step": 8649 + }, + { + "epoch": 2.767704778156997, + "grad_norm": 0.5044153928756714, + "learning_rate": 1.810925161539945e-07, + "loss": 0.4295, + "step": 8650 + }, + { + "epoch": 2.768024744027304, + "grad_norm": 0.5254626870155334, + "learning_rate": 1.8059632233629054e-07, + "loss": 0.4999, + "step": 8651 + }, + { + "epoch": 2.768344709897611, + "grad_norm": 0.479126900434494, + "learning_rate": 1.801007967347046e-07, + "loss": 0.4403, + "step": 8652 + }, + { + "epoch": 2.7686646757679183, + "grad_norm": 0.4540811777114868, + "learning_rate": 1.7960593941794113e-07, + "loss": 0.4404, + "step": 8653 + }, + { + "epoch": 2.7689846416382253, + "grad_norm": 0.46963462233543396, + "learning_rate": 1.7911175045461305e-07, + "loss": 0.4471, + "step": 8654 + }, + { + "epoch": 2.7693046075085324, + "grad_norm": 0.48203012347221375, + "learning_rate": 1.7861822991324107e-07, + "loss": 0.4271, + "step": 8655 + }, + { + "epoch": 2.76962457337884, + "grad_norm": 0.48530858755111694, + "learning_rate": 1.781253778622516e-07, + "loss": 0.4257, + "step": 8656 + }, + { + "epoch": 2.769944539249147, + "grad_norm": 0.47368013858795166, + "learning_rate": 1.776331943699794e-07, + "loss": 0.4403, + "step": 8657 + }, + { + "epoch": 2.770264505119454, + "grad_norm": 0.5373647809028625, + "learning_rate": 1.7714167950466597e-07, + "loss": 0.4392, + "step": 8658 + }, + { + "epoch": 2.7705844709897613, + "grad_norm": 0.47049710154533386, + "learning_rate": 1.7665083333446077e-07, + "loss": 0.3905, + "step": 8659 + }, + { + "epoch": 2.7709044368600684, + "grad_norm": 0.47844138741493225, + "learning_rate": 1.7616065592742038e-07, + "loss": 0.4395, + "step": 8660 + }, + { + "epoch": 2.7712244027303754, + "grad_norm": 0.47040674090385437, + "learning_rate": 1.7567114735150826e-07, + "loss": 0.4368, + "step": 8661 + }, + { + "epoch": 2.7715443686006824, + "grad_norm": 0.5076830983161926, + "learning_rate": 1.7518230767459566e-07, + "loss": 0.4793, + "step": 8662 + }, + { + "epoch": 2.77186433447099, + "grad_norm": 0.4674657881259918, + "learning_rate": 1.7469413696445946e-07, + "loss": 0.4004, + "step": 8663 + }, + { + "epoch": 2.772184300341297, + "grad_norm": 0.48944738507270813, + "learning_rate": 1.7420663528878666e-07, + "loss": 0.4275, + "step": 8664 + }, + { + "epoch": 2.772504266211604, + "grad_norm": 0.5075875520706177, + "learning_rate": 1.7371980271516987e-07, + "loss": 0.4179, + "step": 8665 + }, + { + "epoch": 2.7728242320819114, + "grad_norm": 0.46065860986709595, + "learning_rate": 1.7323363931110836e-07, + "loss": 0.458, + "step": 8666 + }, + { + "epoch": 2.7731441979522184, + "grad_norm": 0.5107364654541016, + "learning_rate": 1.7274814514400995e-07, + "loss": 0.4784, + "step": 8667 + }, + { + "epoch": 2.7734641638225255, + "grad_norm": 0.46682408452033997, + "learning_rate": 1.7226332028118742e-07, + "loss": 0.4453, + "step": 8668 + }, + { + "epoch": 2.7737841296928325, + "grad_norm": 0.46138402819633484, + "learning_rate": 1.7177916478986368e-07, + "loss": 0.4547, + "step": 8669 + }, + { + "epoch": 2.77410409556314, + "grad_norm": 0.47206464409828186, + "learning_rate": 1.712956787371678e-07, + "loss": 0.4108, + "step": 8670 + }, + { + "epoch": 2.774424061433447, + "grad_norm": 0.4193723499774933, + "learning_rate": 1.708128621901356e-07, + "loss": 0.4219, + "step": 8671 + }, + { + "epoch": 2.774744027303754, + "grad_norm": 0.5017774701118469, + "learning_rate": 1.703307152157091e-07, + "loss": 0.408, + "step": 8672 + }, + { + "epoch": 2.7750639931740615, + "grad_norm": 0.5490996241569519, + "learning_rate": 1.698492378807387e-07, + "loss": 0.4885, + "step": 8673 + }, + { + "epoch": 2.7753839590443685, + "grad_norm": 0.459357351064682, + "learning_rate": 1.6936843025198268e-07, + "loss": 0.3526, + "step": 8674 + }, + { + "epoch": 2.7757039249146755, + "grad_norm": 0.5254474878311157, + "learning_rate": 1.6888829239610548e-07, + "loss": 0.4909, + "step": 8675 + }, + { + "epoch": 2.776023890784983, + "grad_norm": 0.5045756101608276, + "learning_rate": 1.6840882437967832e-07, + "loss": 0.3988, + "step": 8676 + }, + { + "epoch": 2.77634385665529, + "grad_norm": 0.47853589057922363, + "learning_rate": 1.679300262691802e-07, + "loss": 0.4152, + "step": 8677 + }, + { + "epoch": 2.776663822525597, + "grad_norm": 0.4418056607246399, + "learning_rate": 1.6745189813099694e-07, + "loss": 0.4152, + "step": 8678 + }, + { + "epoch": 2.7769837883959045, + "grad_norm": 0.5065884590148926, + "learning_rate": 1.669744400314216e-07, + "loss": 0.4627, + "step": 8679 + }, + { + "epoch": 2.7773037542662116, + "grad_norm": 0.5764408707618713, + "learning_rate": 1.6649765203665458e-07, + "loss": 0.4382, + "step": 8680 + }, + { + "epoch": 2.7776237201365186, + "grad_norm": 0.48257893323898315, + "learning_rate": 1.6602153421280298e-07, + "loss": 0.4424, + "step": 8681 + }, + { + "epoch": 2.777943686006826, + "grad_norm": 0.5135520696640015, + "learning_rate": 1.655460866258818e-07, + "loss": 0.4749, + "step": 8682 + }, + { + "epoch": 2.778263651877133, + "grad_norm": 0.4605827033519745, + "learning_rate": 1.6507130934181048e-07, + "loss": 0.4067, + "step": 8683 + }, + { + "epoch": 2.77858361774744, + "grad_norm": 0.49601656198501587, + "learning_rate": 1.6459720242641864e-07, + "loss": 0.4145, + "step": 8684 + }, + { + "epoch": 2.7789035836177476, + "grad_norm": 0.5408211946487427, + "learning_rate": 1.6412376594544255e-07, + "loss": 0.4534, + "step": 8685 + }, + { + "epoch": 2.7792235494880546, + "grad_norm": 0.5210228562355042, + "learning_rate": 1.6365099996452416e-07, + "loss": 0.4529, + "step": 8686 + }, + { + "epoch": 2.7795435153583616, + "grad_norm": 0.5135594606399536, + "learning_rate": 1.6317890454921214e-07, + "loss": 0.4649, + "step": 8687 + }, + { + "epoch": 2.779863481228669, + "grad_norm": 0.5040867924690247, + "learning_rate": 1.627074797649647e-07, + "loss": 0.4136, + "step": 8688 + }, + { + "epoch": 2.780183447098976, + "grad_norm": 0.5140244364738464, + "learning_rate": 1.6223672567714342e-07, + "loss": 0.41, + "step": 8689 + }, + { + "epoch": 2.780503412969283, + "grad_norm": 0.5380143523216248, + "learning_rate": 1.6176664235102114e-07, + "loss": 0.4999, + "step": 8690 + }, + { + "epoch": 2.7808233788395906, + "grad_norm": 0.46018654108047485, + "learning_rate": 1.6129722985177453e-07, + "loss": 0.4032, + "step": 8691 + }, + { + "epoch": 2.7811433447098977, + "grad_norm": 0.4694654047489166, + "learning_rate": 1.6082848824448715e-07, + "loss": 0.4299, + "step": 8692 + }, + { + "epoch": 2.7814633105802047, + "grad_norm": 0.47146186232566833, + "learning_rate": 1.6036041759415255e-07, + "loss": 0.4119, + "step": 8693 + }, + { + "epoch": 2.781783276450512, + "grad_norm": 0.48588356375694275, + "learning_rate": 1.5989301796566713e-07, + "loss": 0.4093, + "step": 8694 + }, + { + "epoch": 2.782103242320819, + "grad_norm": 0.4830414950847626, + "learning_rate": 1.5942628942383854e-07, + "loss": 0.4287, + "step": 8695 + }, + { + "epoch": 2.782423208191126, + "grad_norm": 0.5008724927902222, + "learning_rate": 1.589602320333783e-07, + "loss": 0.422, + "step": 8696 + }, + { + "epoch": 2.7827431740614337, + "grad_norm": 0.5461366772651672, + "learning_rate": 1.584948458589053e-07, + "loss": 0.4599, + "step": 8697 + }, + { + "epoch": 2.7830631399317407, + "grad_norm": 0.557327926158905, + "learning_rate": 1.580301309649468e-07, + "loss": 0.4525, + "step": 8698 + }, + { + "epoch": 2.7833831058020477, + "grad_norm": 0.549267590045929, + "learning_rate": 1.575660874159346e-07, + "loss": 0.43, + "step": 8699 + }, + { + "epoch": 2.783703071672355, + "grad_norm": 0.5106770992279053, + "learning_rate": 1.5710271527621112e-07, + "loss": 0.4263, + "step": 8700 + }, + { + "epoch": 2.7840230375426622, + "grad_norm": 0.5354407429695129, + "learning_rate": 1.566400146100222e-07, + "loss": 0.4589, + "step": 8701 + }, + { + "epoch": 2.7843430034129693, + "grad_norm": 0.45193251967430115, + "learning_rate": 1.561779854815215e-07, + "loss": 0.4203, + "step": 8702 + }, + { + "epoch": 2.7846629692832767, + "grad_norm": 0.4626031219959259, + "learning_rate": 1.5571662795477116e-07, + "loss": 0.4436, + "step": 8703 + }, + { + "epoch": 2.7849829351535837, + "grad_norm": 0.5113726854324341, + "learning_rate": 1.5525594209373716e-07, + "loss": 0.4268, + "step": 8704 + }, + { + "epoch": 2.7853029010238908, + "grad_norm": 0.5348974466323853, + "learning_rate": 1.547959279622957e-07, + "loss": 0.47, + "step": 8705 + }, + { + "epoch": 2.785622866894198, + "grad_norm": 0.5126128792762756, + "learning_rate": 1.5433658562422794e-07, + "loss": 0.4555, + "step": 8706 + }, + { + "epoch": 2.7859428327645053, + "grad_norm": 0.48683688044548035, + "learning_rate": 1.5387791514322182e-07, + "loss": 0.3778, + "step": 8707 + }, + { + "epoch": 2.7862627986348123, + "grad_norm": 0.5526103973388672, + "learning_rate": 1.5341991658287259e-07, + "loss": 0.499, + "step": 8708 + }, + { + "epoch": 2.7865827645051193, + "grad_norm": 0.49776914715766907, + "learning_rate": 1.5296259000668279e-07, + "loss": 0.4261, + "step": 8709 + }, + { + "epoch": 2.7869027303754264, + "grad_norm": 0.5074841380119324, + "learning_rate": 1.5250593547806114e-07, + "loss": 0.4391, + "step": 8710 + }, + { + "epoch": 2.787222696245734, + "grad_norm": 0.5573603510856628, + "learning_rate": 1.520499530603231e-07, + "loss": 0.4533, + "step": 8711 + }, + { + "epoch": 2.787542662116041, + "grad_norm": 0.5001792311668396, + "learning_rate": 1.515946428166909e-07, + "loss": 0.398, + "step": 8712 + }, + { + "epoch": 2.787862627986348, + "grad_norm": 0.49975937604904175, + "learning_rate": 1.5114000481029456e-07, + "loss": 0.4646, + "step": 8713 + }, + { + "epoch": 2.7881825938566553, + "grad_norm": 0.482695609331131, + "learning_rate": 1.506860391041698e-07, + "loss": 0.4291, + "step": 8714 + }, + { + "epoch": 2.7885025597269624, + "grad_norm": 0.49805188179016113, + "learning_rate": 1.5023274576125958e-07, + "loss": 0.4629, + "step": 8715 + }, + { + "epoch": 2.7888225255972694, + "grad_norm": 0.5283095836639404, + "learning_rate": 1.497801248444136e-07, + "loss": 0.4183, + "step": 8716 + }, + { + "epoch": 2.789142491467577, + "grad_norm": 0.5180644989013672, + "learning_rate": 1.493281764163884e-07, + "loss": 0.4163, + "step": 8717 + }, + { + "epoch": 2.789462457337884, + "grad_norm": 0.48102542757987976, + "learning_rate": 1.4887690053984716e-07, + "loss": 0.4574, + "step": 8718 + }, + { + "epoch": 2.789782423208191, + "grad_norm": 0.5150607228279114, + "learning_rate": 1.4842629727735924e-07, + "loss": 0.4088, + "step": 8719 + }, + { + "epoch": 2.7901023890784984, + "grad_norm": 0.5212371945381165, + "learning_rate": 1.4797636669140192e-07, + "loss": 0.4726, + "step": 8720 + }, + { + "epoch": 2.7904223549488054, + "grad_norm": 0.5373845100402832, + "learning_rate": 1.4752710884435806e-07, + "loss": 0.461, + "step": 8721 + }, + { + "epoch": 2.7907423208191124, + "grad_norm": 0.4621168076992035, + "learning_rate": 1.470785237985195e-07, + "loss": 0.4207, + "step": 8722 + }, + { + "epoch": 2.79106228668942, + "grad_norm": 0.48803746700286865, + "learning_rate": 1.4663061161608095e-07, + "loss": 0.4637, + "step": 8723 + }, + { + "epoch": 2.791382252559727, + "grad_norm": 0.5152938961982727, + "learning_rate": 1.4618337235914658e-07, + "loss": 0.4399, + "step": 8724 + }, + { + "epoch": 2.791702218430034, + "grad_norm": 0.47996997833251953, + "learning_rate": 1.4573680608972796e-07, + "loss": 0.4034, + "step": 8725 + }, + { + "epoch": 2.7920221843003414, + "grad_norm": 0.5782660841941833, + "learning_rate": 1.4529091286973994e-07, + "loss": 0.5179, + "step": 8726 + }, + { + "epoch": 2.7923421501706485, + "grad_norm": 0.5318395495414734, + "learning_rate": 1.4484569276100756e-07, + "loss": 0.4363, + "step": 8727 + }, + { + "epoch": 2.7926621160409555, + "grad_norm": 0.5264575481414795, + "learning_rate": 1.4440114582526089e-07, + "loss": 0.5013, + "step": 8728 + }, + { + "epoch": 2.792982081911263, + "grad_norm": 0.49003055691719055, + "learning_rate": 1.4395727212413668e-07, + "loss": 0.4107, + "step": 8729 + }, + { + "epoch": 2.79330204778157, + "grad_norm": 0.4770995080471039, + "learning_rate": 1.4351407171917908e-07, + "loss": 0.4214, + "step": 8730 + }, + { + "epoch": 2.793622013651877, + "grad_norm": 0.4444619119167328, + "learning_rate": 1.4307154467183727e-07, + "loss": 0.4126, + "step": 8731 + }, + { + "epoch": 2.7939419795221845, + "grad_norm": 0.4984275996685028, + "learning_rate": 1.426296910434688e-07, + "loss": 0.4686, + "step": 8732 + }, + { + "epoch": 2.7942619453924915, + "grad_norm": 0.4909839928150177, + "learning_rate": 1.4218851089533802e-07, + "loss": 0.4467, + "step": 8733 + }, + { + "epoch": 2.7945819112627985, + "grad_norm": 0.4700436592102051, + "learning_rate": 1.4174800428861324e-07, + "loss": 0.4161, + "step": 8734 + }, + { + "epoch": 2.794901877133106, + "grad_norm": 0.47440141439437866, + "learning_rate": 1.4130817128437336e-07, + "loss": 0.4766, + "step": 8735 + }, + { + "epoch": 2.795221843003413, + "grad_norm": 0.49831926822662354, + "learning_rate": 1.4086901194359958e-07, + "loss": 0.4411, + "step": 8736 + }, + { + "epoch": 2.79554180887372, + "grad_norm": 0.4541718065738678, + "learning_rate": 1.4043052632718378e-07, + "loss": 0.4208, + "step": 8737 + }, + { + "epoch": 2.7958617747440275, + "grad_norm": 0.4604977071285248, + "learning_rate": 1.3999271449592177e-07, + "loss": 0.4676, + "step": 8738 + }, + { + "epoch": 2.7961817406143346, + "grad_norm": 0.4659779965877533, + "learning_rate": 1.395555765105161e-07, + "loss": 0.387, + "step": 8739 + }, + { + "epoch": 2.7965017064846416, + "grad_norm": 0.5293183326721191, + "learning_rate": 1.3911911243157716e-07, + "loss": 0.4838, + "step": 8740 + }, + { + "epoch": 2.796821672354949, + "grad_norm": 0.5436280369758606, + "learning_rate": 1.38683322319621e-07, + "loss": 0.4305, + "step": 8741 + }, + { + "epoch": 2.797141638225256, + "grad_norm": 0.5356222987174988, + "learning_rate": 1.3824820623507095e-07, + "loss": 0.4546, + "step": 8742 + }, + { + "epoch": 2.797461604095563, + "grad_norm": 0.4895310699939728, + "learning_rate": 1.3781376423825598e-07, + "loss": 0.4064, + "step": 8743 + }, + { + "epoch": 2.7977815699658706, + "grad_norm": 0.5118739008903503, + "learning_rate": 1.3737999638941124e-07, + "loss": 0.4892, + "step": 8744 + }, + { + "epoch": 2.7981015358361776, + "grad_norm": 0.5162504315376282, + "learning_rate": 1.3694690274868028e-07, + "loss": 0.4285, + "step": 8745 + }, + { + "epoch": 2.7984215017064846, + "grad_norm": 0.542949914932251, + "learning_rate": 1.3651448337611117e-07, + "loss": 0.4784, + "step": 8746 + }, + { + "epoch": 2.7987414675767917, + "grad_norm": 0.5507196187973022, + "learning_rate": 1.3608273833166096e-07, + "loss": 0.4858, + "step": 8747 + }, + { + "epoch": 2.799061433447099, + "grad_norm": 0.5387033224105835, + "learning_rate": 1.3565166767519012e-07, + "loss": 0.4182, + "step": 8748 + }, + { + "epoch": 2.799381399317406, + "grad_norm": 0.5092856287956238, + "learning_rate": 1.3522127146646691e-07, + "loss": 0.4973, + "step": 8749 + }, + { + "epoch": 2.799701365187713, + "grad_norm": 0.5302609801292419, + "learning_rate": 1.3479154976516696e-07, + "loss": 0.3835, + "step": 8750 + }, + { + "epoch": 2.80002133105802, + "grad_norm": 0.4946901798248291, + "learning_rate": 1.3436250263087204e-07, + "loss": 0.4143, + "step": 8751 + }, + { + "epoch": 2.8003412969283277, + "grad_norm": 0.46833574771881104, + "learning_rate": 1.3393413012306955e-07, + "loss": 0.4492, + "step": 8752 + }, + { + "epoch": 2.8006612627986347, + "grad_norm": 0.42688828706741333, + "learning_rate": 1.3350643230115424e-07, + "loss": 0.4154, + "step": 8753 + }, + { + "epoch": 2.8009812286689417, + "grad_norm": 0.5493572950363159, + "learning_rate": 1.3307940922442585e-07, + "loss": 0.5317, + "step": 8754 + }, + { + "epoch": 2.801301194539249, + "grad_norm": 0.45775240659713745, + "learning_rate": 1.3265306095209208e-07, + "loss": 0.3769, + "step": 8755 + }, + { + "epoch": 2.801621160409556, + "grad_norm": 0.5258458852767944, + "learning_rate": 1.3222738754326836e-07, + "loss": 0.4656, + "step": 8756 + }, + { + "epoch": 2.8019411262798632, + "grad_norm": 0.5466823577880859, + "learning_rate": 1.3180238905697306e-07, + "loss": 0.4292, + "step": 8757 + }, + { + "epoch": 2.8022610921501707, + "grad_norm": 0.5115330815315247, + "learning_rate": 1.3137806555213295e-07, + "loss": 0.4582, + "step": 8758 + }, + { + "epoch": 2.8025810580204777, + "grad_norm": 0.44748345017433167, + "learning_rate": 1.309544170875804e-07, + "loss": 0.4004, + "step": 8759 + }, + { + "epoch": 2.8029010238907848, + "grad_norm": 0.5248578190803528, + "learning_rate": 1.3053144372205619e-07, + "loss": 0.4329, + "step": 8760 + }, + { + "epoch": 2.8032209897610922, + "grad_norm": 0.5367554426193237, + "learning_rate": 1.301091455142056e-07, + "loss": 0.4666, + "step": 8761 + }, + { + "epoch": 2.8035409556313993, + "grad_norm": 0.5284185409545898, + "learning_rate": 1.296875225225802e-07, + "loss": 0.4363, + "step": 8762 + }, + { + "epoch": 2.8038609215017063, + "grad_norm": 0.4763930141925812, + "learning_rate": 1.292665748056393e-07, + "loss": 0.3859, + "step": 8763 + }, + { + "epoch": 2.8041808873720138, + "grad_norm": 0.49560821056365967, + "learning_rate": 1.2884630242174734e-07, + "loss": 0.431, + "step": 8764 + }, + { + "epoch": 2.804500853242321, + "grad_norm": 0.4569384753704071, + "learning_rate": 1.284267054291749e-07, + "loss": 0.4236, + "step": 8765 + }, + { + "epoch": 2.804820819112628, + "grad_norm": 0.4791736900806427, + "learning_rate": 1.2800778388610157e-07, + "loss": 0.4238, + "step": 8766 + }, + { + "epoch": 2.8051407849829353, + "grad_norm": 0.4712306261062622, + "learning_rate": 1.275895378506098e-07, + "loss": 0.4277, + "step": 8767 + }, + { + "epoch": 2.8054607508532423, + "grad_norm": 0.43769967555999756, + "learning_rate": 1.2717196738069036e-07, + "loss": 0.4321, + "step": 8768 + }, + { + "epoch": 2.8057807167235493, + "grad_norm": 0.49380260705947876, + "learning_rate": 1.2675507253423924e-07, + "loss": 0.4275, + "step": 8769 + }, + { + "epoch": 2.806100682593857, + "grad_norm": 0.5135136246681213, + "learning_rate": 1.2633885336906014e-07, + "loss": 0.4889, + "step": 8770 + }, + { + "epoch": 2.806420648464164, + "grad_norm": 0.5384788513183594, + "learning_rate": 1.2592330994286251e-07, + "loss": 0.4352, + "step": 8771 + }, + { + "epoch": 2.806740614334471, + "grad_norm": 0.4903396666049957, + "learning_rate": 1.2550844231326188e-07, + "loss": 0.4284, + "step": 8772 + }, + { + "epoch": 2.8070605802047783, + "grad_norm": 0.5049971342086792, + "learning_rate": 1.2509425053777947e-07, + "loss": 0.4427, + "step": 8773 + }, + { + "epoch": 2.8073805460750854, + "grad_norm": 0.4411507546901703, + "learning_rate": 1.246807346738438e-07, + "loss": 0.3914, + "step": 8774 + }, + { + "epoch": 2.8077005119453924, + "grad_norm": 0.5614057779312134, + "learning_rate": 1.2426789477878898e-07, + "loss": 0.4433, + "step": 8775 + }, + { + "epoch": 2.8080204778157, + "grad_norm": 0.46131646633148193, + "learning_rate": 1.2385573090985703e-07, + "loss": 0.4479, + "step": 8776 + }, + { + "epoch": 2.808340443686007, + "grad_norm": 0.5110061168670654, + "learning_rate": 1.234442431241939e-07, + "loss": 0.4529, + "step": 8777 + }, + { + "epoch": 2.808660409556314, + "grad_norm": 0.548893928527832, + "learning_rate": 1.2303343147885283e-07, + "loss": 0.5141, + "step": 8778 + }, + { + "epoch": 2.8089803754266214, + "grad_norm": 0.5144462585449219, + "learning_rate": 1.2262329603079382e-07, + "loss": 0.3636, + "step": 8779 + }, + { + "epoch": 2.8093003412969284, + "grad_norm": 0.5226725339889526, + "learning_rate": 1.2221383683688192e-07, + "loss": 0.4317, + "step": 8780 + }, + { + "epoch": 2.8096203071672354, + "grad_norm": 0.4646565318107605, + "learning_rate": 1.2180505395389063e-07, + "loss": 0.4245, + "step": 8781 + }, + { + "epoch": 2.809940273037543, + "grad_norm": 0.49531203508377075, + "learning_rate": 1.2139694743849684e-07, + "loss": 0.4439, + "step": 8782 + }, + { + "epoch": 2.81026023890785, + "grad_norm": 0.5245881080627441, + "learning_rate": 1.2098951734728525e-07, + "loss": 0.4501, + "step": 8783 + }, + { + "epoch": 2.810580204778157, + "grad_norm": 0.5053141117095947, + "learning_rate": 1.2058276373674682e-07, + "loss": 0.4572, + "step": 8784 + }, + { + "epoch": 2.8109001706484644, + "grad_norm": 0.4849434792995453, + "learning_rate": 1.2017668666327752e-07, + "loss": 0.4485, + "step": 8785 + }, + { + "epoch": 2.8112201365187715, + "grad_norm": 0.4660979211330414, + "learning_rate": 1.1977128618318234e-07, + "loss": 0.3886, + "step": 8786 + }, + { + "epoch": 2.8115401023890785, + "grad_norm": 0.47356531023979187, + "learning_rate": 1.193665623526691e-07, + "loss": 0.4243, + "step": 8787 + }, + { + "epoch": 2.8118600682593855, + "grad_norm": 0.46527403593063354, + "learning_rate": 1.189625152278534e-07, + "loss": 0.4342, + "step": 8788 + }, + { + "epoch": 2.812180034129693, + "grad_norm": 0.5244778990745544, + "learning_rate": 1.1855914486475661e-07, + "loss": 0.4673, + "step": 8789 + }, + { + "epoch": 2.8125, + "grad_norm": 0.4573207497596741, + "learning_rate": 1.181564513193073e-07, + "loss": 0.3831, + "step": 8790 + }, + { + "epoch": 2.812819965870307, + "grad_norm": 0.448699951171875, + "learning_rate": 1.177544346473386e-07, + "loss": 0.3922, + "step": 8791 + }, + { + "epoch": 2.8131399317406145, + "grad_norm": 0.5018029808998108, + "learning_rate": 1.1735309490459201e-07, + "loss": 0.4931, + "step": 8792 + }, + { + "epoch": 2.8134598976109215, + "grad_norm": 0.4705091416835785, + "learning_rate": 1.1695243214671192e-07, + "loss": 0.3911, + "step": 8793 + }, + { + "epoch": 2.8137798634812285, + "grad_norm": 0.4831168055534363, + "learning_rate": 1.1655244642925113e-07, + "loss": 0.4585, + "step": 8794 + }, + { + "epoch": 2.8140998293515356, + "grad_norm": 0.46551117300987244, + "learning_rate": 1.1615313780766967e-07, + "loss": 0.4352, + "step": 8795 + }, + { + "epoch": 2.814419795221843, + "grad_norm": 0.5043496489524841, + "learning_rate": 1.1575450633733054e-07, + "loss": 0.4108, + "step": 8796 + }, + { + "epoch": 2.81473976109215, + "grad_norm": 0.5622084736824036, + "learning_rate": 1.153565520735056e-07, + "loss": 0.5166, + "step": 8797 + }, + { + "epoch": 2.815059726962457, + "grad_norm": 0.45480015873908997, + "learning_rate": 1.1495927507137073e-07, + "loss": 0.3995, + "step": 8798 + }, + { + "epoch": 2.8153796928327646, + "grad_norm": 0.5136435627937317, + "learning_rate": 1.1456267538600797e-07, + "loss": 0.4597, + "step": 8799 + }, + { + "epoch": 2.8156996587030716, + "grad_norm": 0.5320642590522766, + "learning_rate": 1.1416675307240887e-07, + "loss": 0.4379, + "step": 8800 + }, + { + "epoch": 2.8160196245733786, + "grad_norm": 0.4744340181350708, + "learning_rate": 1.1377150818546734e-07, + "loss": 0.4363, + "step": 8801 + }, + { + "epoch": 2.816339590443686, + "grad_norm": 0.4676035940647125, + "learning_rate": 1.1337694077998451e-07, + "loss": 0.4385, + "step": 8802 + }, + { + "epoch": 2.816659556313993, + "grad_norm": 0.4505445063114166, + "learning_rate": 1.1298305091066664e-07, + "loss": 0.4517, + "step": 8803 + }, + { + "epoch": 2.8169795221843, + "grad_norm": 0.43211179971694946, + "learning_rate": 1.1258983863212835e-07, + "loss": 0.3816, + "step": 8804 + }, + { + "epoch": 2.8172994880546076, + "grad_norm": 0.4516248404979706, + "learning_rate": 1.121973039988894e-07, + "loss": 0.4566, + "step": 8805 + }, + { + "epoch": 2.8176194539249146, + "grad_norm": 0.48179879784584045, + "learning_rate": 1.11805447065374e-07, + "loss": 0.4273, + "step": 8806 + }, + { + "epoch": 2.8179394197952217, + "grad_norm": 0.4546855390071869, + "learning_rate": 1.1141426788591426e-07, + "loss": 0.4601, + "step": 8807 + }, + { + "epoch": 2.818259385665529, + "grad_norm": 0.5028133392333984, + "learning_rate": 1.1102376651474678e-07, + "loss": 0.4323, + "step": 8808 + }, + { + "epoch": 2.818579351535836, + "grad_norm": 0.5004110932350159, + "learning_rate": 1.1063394300601548e-07, + "loss": 0.4164, + "step": 8809 + }, + { + "epoch": 2.818899317406143, + "grad_norm": 0.5113457441329956, + "learning_rate": 1.1024479741377048e-07, + "loss": 0.4388, + "step": 8810 + }, + { + "epoch": 2.8192192832764507, + "grad_norm": 0.5095594525337219, + "learning_rate": 1.0985632979196692e-07, + "loss": 0.4749, + "step": 8811 + }, + { + "epoch": 2.8195392491467577, + "grad_norm": 0.46509113907814026, + "learning_rate": 1.0946854019446673e-07, + "loss": 0.429, + "step": 8812 + }, + { + "epoch": 2.8198592150170647, + "grad_norm": 0.5171486735343933, + "learning_rate": 1.090814286750358e-07, + "loss": 0.4134, + "step": 8813 + }, + { + "epoch": 2.820179180887372, + "grad_norm": 0.5246515274047852, + "learning_rate": 1.0869499528734894e-07, + "loss": 0.4441, + "step": 8814 + }, + { + "epoch": 2.820499146757679, + "grad_norm": 0.49848657846450806, + "learning_rate": 1.0830924008498612e-07, + "loss": 0.4639, + "step": 8815 + }, + { + "epoch": 2.8208191126279862, + "grad_norm": 0.5230051279067993, + "learning_rate": 1.0792416312143172e-07, + "loss": 0.3844, + "step": 8816 + }, + { + "epoch": 2.8211390784982937, + "grad_norm": 0.5380415916442871, + "learning_rate": 1.0753976445007696e-07, + "loss": 0.4908, + "step": 8817 + }, + { + "epoch": 2.8214590443686007, + "grad_norm": 0.4953198730945587, + "learning_rate": 1.0715604412421976e-07, + "loss": 0.4454, + "step": 8818 + }, + { + "epoch": 2.8217790102389078, + "grad_norm": 0.5087031126022339, + "learning_rate": 1.0677300219706311e-07, + "loss": 0.3919, + "step": 8819 + }, + { + "epoch": 2.8220989761092152, + "grad_norm": 0.45869845151901245, + "learning_rate": 1.0639063872171674e-07, + "loss": 0.4644, + "step": 8820 + }, + { + "epoch": 2.8224189419795223, + "grad_norm": 0.4913416802883148, + "learning_rate": 1.0600895375119602e-07, + "loss": 0.4402, + "step": 8821 + }, + { + "epoch": 2.8227389078498293, + "grad_norm": 0.5042554140090942, + "learning_rate": 1.0562794733842086e-07, + "loss": 0.4958, + "step": 8822 + }, + { + "epoch": 2.8230588737201368, + "grad_norm": 0.5037972927093506, + "learning_rate": 1.0524761953621843e-07, + "loss": 0.414, + "step": 8823 + }, + { + "epoch": 2.823378839590444, + "grad_norm": 0.4720671474933624, + "learning_rate": 1.0486797039732266e-07, + "loss": 0.3998, + "step": 8824 + }, + { + "epoch": 2.823698805460751, + "grad_norm": 0.4891723692417145, + "learning_rate": 1.0448899997437145e-07, + "loss": 0.409, + "step": 8825 + }, + { + "epoch": 2.8240187713310583, + "grad_norm": 0.49462926387786865, + "learning_rate": 1.0411070831991055e-07, + "loss": 0.4409, + "step": 8826 + }, + { + "epoch": 2.8243387372013653, + "grad_norm": 0.4781778156757355, + "learning_rate": 1.0373309548638966e-07, + "loss": 0.4708, + "step": 8827 + }, + { + "epoch": 2.8246587030716723, + "grad_norm": 0.4797002375125885, + "learning_rate": 1.0335616152616468e-07, + "loss": 0.4244, + "step": 8828 + }, + { + "epoch": 2.82497866894198, + "grad_norm": 0.42282843589782715, + "learning_rate": 1.029799064914988e-07, + "loss": 0.4112, + "step": 8829 + }, + { + "epoch": 2.825298634812287, + "grad_norm": 0.5273585319519043, + "learning_rate": 1.0260433043456086e-07, + "loss": 0.4617, + "step": 8830 + }, + { + "epoch": 2.825618600682594, + "grad_norm": 0.47264957427978516, + "learning_rate": 1.0222943340742365e-07, + "loss": 0.4212, + "step": 8831 + }, + { + "epoch": 2.825938566552901, + "grad_norm": 0.4755886197090149, + "learning_rate": 1.0185521546206834e-07, + "loss": 0.4078, + "step": 8832 + }, + { + "epoch": 2.8262585324232083, + "grad_norm": 0.47338294982910156, + "learning_rate": 1.01481676650379e-07, + "loss": 0.4228, + "step": 8833 + }, + { + "epoch": 2.8265784982935154, + "grad_norm": 0.4957878291606903, + "learning_rate": 1.0110881702414865e-07, + "loss": 0.4529, + "step": 8834 + }, + { + "epoch": 2.8268984641638224, + "grad_norm": 0.5062832236289978, + "learning_rate": 1.0073663663507482e-07, + "loss": 0.4709, + "step": 8835 + }, + { + "epoch": 2.8272184300341294, + "grad_norm": 0.483052134513855, + "learning_rate": 1.0036513553476013e-07, + "loss": 0.4047, + "step": 8836 + }, + { + "epoch": 2.827538395904437, + "grad_norm": 0.5118340253829956, + "learning_rate": 9.999431377471391e-08, + "loss": 0.4398, + "step": 8837 + }, + { + "epoch": 2.827858361774744, + "grad_norm": 0.5285395383834839, + "learning_rate": 9.962417140635006e-08, + "loss": 0.4436, + "step": 8838 + }, + { + "epoch": 2.828178327645051, + "grad_norm": 0.5113656520843506, + "learning_rate": 9.925470848099083e-08, + "loss": 0.4113, + "step": 8839 + }, + { + "epoch": 2.8284982935153584, + "grad_norm": 0.5032393336296082, + "learning_rate": 9.888592504986194e-08, + "loss": 0.4706, + "step": 8840 + }, + { + "epoch": 2.8288182593856654, + "grad_norm": 0.5037457346916199, + "learning_rate": 9.851782116409581e-08, + "loss": 0.5053, + "step": 8841 + }, + { + "epoch": 2.8291382252559725, + "grad_norm": 0.46051284670829773, + "learning_rate": 9.81503968747305e-08, + "loss": 0.4571, + "step": 8842 + }, + { + "epoch": 2.82945819112628, + "grad_norm": 0.4622531533241272, + "learning_rate": 9.778365223270914e-08, + "loss": 0.431, + "step": 8843 + }, + { + "epoch": 2.829778156996587, + "grad_norm": 0.476308673620224, + "learning_rate": 9.741758728888218e-08, + "loss": 0.4512, + "step": 8844 + }, + { + "epoch": 2.830098122866894, + "grad_norm": 0.45065200328826904, + "learning_rate": 9.705220209400457e-08, + "loss": 0.4513, + "step": 8845 + }, + { + "epoch": 2.8304180887372015, + "grad_norm": 0.4600454866886139, + "learning_rate": 9.6687496698738e-08, + "loss": 0.4176, + "step": 8846 + }, + { + "epoch": 2.8307380546075085, + "grad_norm": 0.48312658071517944, + "learning_rate": 9.63234711536487e-08, + "loss": 0.5039, + "step": 8847 + }, + { + "epoch": 2.8310580204778155, + "grad_norm": 0.5284901857376099, + "learning_rate": 9.596012550920907e-08, + "loss": 0.417, + "step": 8848 + }, + { + "epoch": 2.831377986348123, + "grad_norm": 0.4941372275352478, + "learning_rate": 9.559745981579771e-08, + "loss": 0.482, + "step": 8849 + }, + { + "epoch": 2.83169795221843, + "grad_norm": 0.5106128454208374, + "learning_rate": 9.523547412369937e-08, + "loss": 0.4064, + "step": 8850 + }, + { + "epoch": 2.832017918088737, + "grad_norm": 0.5166605114936829, + "learning_rate": 9.487416848310283e-08, + "loss": 0.415, + "step": 8851 + }, + { + "epoch": 2.8323378839590445, + "grad_norm": 0.5388782024383545, + "learning_rate": 9.451354294410353e-08, + "loss": 0.4282, + "step": 8852 + }, + { + "epoch": 2.8326578498293515, + "grad_norm": 0.49470970034599304, + "learning_rate": 9.415359755670317e-08, + "loss": 0.4512, + "step": 8853 + }, + { + "epoch": 2.8329778156996586, + "grad_norm": 0.4966568648815155, + "learning_rate": 9.379433237080793e-08, + "loss": 0.4595, + "step": 8854 + }, + { + "epoch": 2.833297781569966, + "grad_norm": 0.4812455177307129, + "learning_rate": 9.34357474362313e-08, + "loss": 0.4621, + "step": 8855 + }, + { + "epoch": 2.833617747440273, + "grad_norm": 0.5319587588310242, + "learning_rate": 9.307784280269128e-08, + "loss": 0.4628, + "step": 8856 + }, + { + "epoch": 2.83393771331058, + "grad_norm": 0.4560822546482086, + "learning_rate": 9.272061851981151e-08, + "loss": 0.3885, + "step": 8857 + }, + { + "epoch": 2.8342576791808876, + "grad_norm": 0.49199801683425903, + "learning_rate": 9.236407463712071e-08, + "loss": 0.4398, + "step": 8858 + }, + { + "epoch": 2.8345776450511946, + "grad_norm": 0.4829440414905548, + "learning_rate": 9.200821120405545e-08, + "loss": 0.3572, + "step": 8859 + }, + { + "epoch": 2.8348976109215016, + "grad_norm": 0.5515627861022949, + "learning_rate": 9.165302826995626e-08, + "loss": 0.5003, + "step": 8860 + }, + { + "epoch": 2.835217576791809, + "grad_norm": 0.6842747330665588, + "learning_rate": 9.129852588406985e-08, + "loss": 0.4109, + "step": 8861 + }, + { + "epoch": 2.835537542662116, + "grad_norm": 0.5266549587249756, + "learning_rate": 9.094470409554746e-08, + "loss": 0.4544, + "step": 8862 + }, + { + "epoch": 2.835857508532423, + "grad_norm": 0.45223504304885864, + "learning_rate": 9.059156295344818e-08, + "loss": 0.4239, + "step": 8863 + }, + { + "epoch": 2.8361774744027306, + "grad_norm": 0.5521233677864075, + "learning_rate": 9.023910250673451e-08, + "loss": 0.4705, + "step": 8864 + }, + { + "epoch": 2.8364974402730376, + "grad_norm": 0.5143077969551086, + "learning_rate": 8.988732280427625e-08, + "loss": 0.4167, + "step": 8865 + }, + { + "epoch": 2.8368174061433447, + "grad_norm": 0.4757063090801239, + "learning_rate": 8.953622389484774e-08, + "loss": 0.4431, + "step": 8866 + }, + { + "epoch": 2.837137372013652, + "grad_norm": 0.47248420119285583, + "learning_rate": 8.918580582712833e-08, + "loss": 0.4305, + "step": 8867 + }, + { + "epoch": 2.837457337883959, + "grad_norm": 0.5008695721626282, + "learning_rate": 8.883606864970585e-08, + "loss": 0.4174, + "step": 8868 + }, + { + "epoch": 2.837777303754266, + "grad_norm": 0.5308582782745361, + "learning_rate": 8.848701241106982e-08, + "loss": 0.4573, + "step": 8869 + }, + { + "epoch": 2.8380972696245736, + "grad_norm": 0.4937674403190613, + "learning_rate": 8.813863715961879e-08, + "loss": 0.4102, + "step": 8870 + }, + { + "epoch": 2.8384172354948807, + "grad_norm": 0.50037682056427, + "learning_rate": 8.77909429436552e-08, + "loss": 0.4047, + "step": 8871 + }, + { + "epoch": 2.8387372013651877, + "grad_norm": 0.577347457408905, + "learning_rate": 8.744392981138605e-08, + "loss": 0.5141, + "step": 8872 + }, + { + "epoch": 2.8390571672354947, + "grad_norm": 0.5387556552886963, + "learning_rate": 8.709759781092675e-08, + "loss": 0.4247, + "step": 8873 + }, + { + "epoch": 2.839377133105802, + "grad_norm": 0.485126256942749, + "learning_rate": 8.675194699029499e-08, + "loss": 0.4278, + "step": 8874 + }, + { + "epoch": 2.8396970989761092, + "grad_norm": 0.5208066701889038, + "learning_rate": 8.640697739741744e-08, + "loss": 0.4033, + "step": 8875 + }, + { + "epoch": 2.8400170648464163, + "grad_norm": 0.5068288445472717, + "learning_rate": 8.606268908012305e-08, + "loss": 0.4318, + "step": 8876 + }, + { + "epoch": 2.8403370307167233, + "grad_norm": 0.470518559217453, + "learning_rate": 8.571908208614866e-08, + "loss": 0.458, + "step": 8877 + }, + { + "epoch": 2.8406569965870307, + "grad_norm": 0.44177472591400146, + "learning_rate": 8.537615646313557e-08, + "loss": 0.4154, + "step": 8878 + }, + { + "epoch": 2.8409769624573378, + "grad_norm": 0.4870431423187256, + "learning_rate": 8.503391225863078e-08, + "loss": 0.4035, + "step": 8879 + }, + { + "epoch": 2.841296928327645, + "grad_norm": 0.4561026394367218, + "learning_rate": 8.469234952008687e-08, + "loss": 0.4108, + "step": 8880 + }, + { + "epoch": 2.8416168941979523, + "grad_norm": 0.4789717495441437, + "learning_rate": 8.435146829486263e-08, + "loss": 0.4655, + "step": 8881 + }, + { + "epoch": 2.8419368600682593, + "grad_norm": 0.48523077368736267, + "learning_rate": 8.401126863021969e-08, + "loss": 0.4506, + "step": 8882 + }, + { + "epoch": 2.8422568259385663, + "grad_norm": 0.46199101209640503, + "learning_rate": 8.367175057332976e-08, + "loss": 0.4377, + "step": 8883 + }, + { + "epoch": 2.842576791808874, + "grad_norm": 0.48472705483436584, + "learning_rate": 8.33329141712652e-08, + "loss": 0.4305, + "step": 8884 + }, + { + "epoch": 2.842896757679181, + "grad_norm": 0.5093603134155273, + "learning_rate": 8.299475947100733e-08, + "loss": 0.4637, + "step": 8885 + }, + { + "epoch": 2.843216723549488, + "grad_norm": 0.4923092722892761, + "learning_rate": 8.265728651944138e-08, + "loss": 0.4038, + "step": 8886 + }, + { + "epoch": 2.8435366894197953, + "grad_norm": 0.48508721590042114, + "learning_rate": 8.232049536335774e-08, + "loss": 0.4471, + "step": 8887 + }, + { + "epoch": 2.8438566552901023, + "grad_norm": 0.4965299963951111, + "learning_rate": 8.198438604945457e-08, + "loss": 0.4184, + "step": 8888 + }, + { + "epoch": 2.8441766211604094, + "grad_norm": 0.5033362507820129, + "learning_rate": 8.164895862433186e-08, + "loss": 0.4325, + "step": 8889 + }, + { + "epoch": 2.844496587030717, + "grad_norm": 0.5296280980110168, + "learning_rate": 8.13142131344985e-08, + "loss": 0.4991, + "step": 8890 + }, + { + "epoch": 2.844816552901024, + "grad_norm": 0.549327552318573, + "learning_rate": 8.09801496263668e-08, + "loss": 0.4385, + "step": 8891 + }, + { + "epoch": 2.845136518771331, + "grad_norm": 0.503280520439148, + "learning_rate": 8.064676814625472e-08, + "loss": 0.4513, + "step": 8892 + }, + { + "epoch": 2.8454564846416384, + "grad_norm": 0.4724968373775482, + "learning_rate": 8.031406874038694e-08, + "loss": 0.3853, + "step": 8893 + }, + { + "epoch": 2.8457764505119454, + "grad_norm": 0.5086998343467712, + "learning_rate": 7.998205145489157e-08, + "loss": 0.467, + "step": 8894 + }, + { + "epoch": 2.8460964163822524, + "grad_norm": 0.5187538266181946, + "learning_rate": 7.965071633580401e-08, + "loss": 0.4731, + "step": 8895 + }, + { + "epoch": 2.84641638225256, + "grad_norm": 0.511500358581543, + "learning_rate": 7.932006342906361e-08, + "loss": 0.4319, + "step": 8896 + }, + { + "epoch": 2.846736348122867, + "grad_norm": 0.4847926199436188, + "learning_rate": 7.89900927805165e-08, + "loss": 0.4282, + "step": 8897 + }, + { + "epoch": 2.847056313993174, + "grad_norm": 0.4554595947265625, + "learning_rate": 7.866080443591273e-08, + "loss": 0.4007, + "step": 8898 + }, + { + "epoch": 2.8473762798634814, + "grad_norm": 0.5097051858901978, + "learning_rate": 7.833219844090911e-08, + "loss": 0.4787, + "step": 8899 + }, + { + "epoch": 2.8476962457337884, + "grad_norm": 0.5208517909049988, + "learning_rate": 7.800427484106754e-08, + "loss": 0.457, + "step": 8900 + }, + { + "epoch": 2.8480162116040955, + "grad_norm": 0.5112928748130798, + "learning_rate": 7.767703368185386e-08, + "loss": 0.4765, + "step": 8901 + }, + { + "epoch": 2.848336177474403, + "grad_norm": 0.4776992201805115, + "learning_rate": 7.735047500864179e-08, + "loss": 0.4128, + "step": 8902 + }, + { + "epoch": 2.84865614334471, + "grad_norm": 0.49549901485443115, + "learning_rate": 7.702459886670788e-08, + "loss": 0.5057, + "step": 8903 + }, + { + "epoch": 2.848976109215017, + "grad_norm": 0.48889100551605225, + "learning_rate": 7.669940530123598e-08, + "loss": 0.4079, + "step": 8904 + }, + { + "epoch": 2.8492960750853245, + "grad_norm": 0.5252945423126221, + "learning_rate": 7.637489435731505e-08, + "loss": 0.4028, + "step": 8905 + }, + { + "epoch": 2.8496160409556315, + "grad_norm": 0.5165454745292664, + "learning_rate": 7.60510660799374e-08, + "loss": 0.4237, + "step": 8906 + }, + { + "epoch": 2.8499360068259385, + "grad_norm": 0.5192331075668335, + "learning_rate": 7.572792051400325e-08, + "loss": 0.4722, + "step": 8907 + }, + { + "epoch": 2.850255972696246, + "grad_norm": 0.4691384732723236, + "learning_rate": 7.540545770431729e-08, + "loss": 0.3964, + "step": 8908 + }, + { + "epoch": 2.850575938566553, + "grad_norm": 0.5160424709320068, + "learning_rate": 7.508367769558877e-08, + "loss": 0.4559, + "step": 8909 + }, + { + "epoch": 2.85089590443686, + "grad_norm": 0.4484829902648926, + "learning_rate": 7.476258053243368e-08, + "loss": 0.3618, + "step": 8910 + }, + { + "epoch": 2.8512158703071675, + "grad_norm": 0.48174214363098145, + "learning_rate": 7.44421662593714e-08, + "loss": 0.4129, + "step": 8911 + }, + { + "epoch": 2.8515358361774745, + "grad_norm": 0.4934282600879669, + "learning_rate": 7.41224349208286e-08, + "loss": 0.4777, + "step": 8912 + }, + { + "epoch": 2.8518558020477816, + "grad_norm": 0.49322712421417236, + "learning_rate": 7.38033865611365e-08, + "loss": 0.415, + "step": 8913 + }, + { + "epoch": 2.8521757679180886, + "grad_norm": 0.48833104968070984, + "learning_rate": 7.348502122453083e-08, + "loss": 0.4436, + "step": 8914 + }, + { + "epoch": 2.852495733788396, + "grad_norm": 0.46358126401901245, + "learning_rate": 7.316733895515349e-08, + "loss": 0.465, + "step": 8915 + }, + { + "epoch": 2.852815699658703, + "grad_norm": 0.46667855978012085, + "learning_rate": 7.285033979705203e-08, + "loss": 0.4088, + "step": 8916 + }, + { + "epoch": 2.85313566552901, + "grad_norm": 0.5348223447799683, + "learning_rate": 7.253402379417851e-08, + "loss": 0.4822, + "step": 8917 + }, + { + "epoch": 2.8534556313993176, + "grad_norm": 0.47619935870170593, + "learning_rate": 7.221839099039063e-08, + "loss": 0.4191, + "step": 8918 + }, + { + "epoch": 2.8537755972696246, + "grad_norm": 0.4943060576915741, + "learning_rate": 7.19034414294506e-08, + "loss": 0.4274, + "step": 8919 + }, + { + "epoch": 2.8540955631399316, + "grad_norm": 0.47631967067718506, + "learning_rate": 7.158917515502739e-08, + "loss": 0.4444, + "step": 8920 + }, + { + "epoch": 2.8544155290102387, + "grad_norm": 0.5007773637771606, + "learning_rate": 7.127559221069391e-08, + "loss": 0.4482, + "step": 8921 + }, + { + "epoch": 2.854735494880546, + "grad_norm": 0.511054515838623, + "learning_rate": 7.096269263992928e-08, + "loss": 0.4479, + "step": 8922 + }, + { + "epoch": 2.855055460750853, + "grad_norm": 0.510521411895752, + "learning_rate": 7.065047648611711e-08, + "loss": 0.4566, + "step": 8923 + }, + { + "epoch": 2.85537542662116, + "grad_norm": 0.5748690366744995, + "learning_rate": 7.033894379254558e-08, + "loss": 0.4625, + "step": 8924 + }, + { + "epoch": 2.8556953924914676, + "grad_norm": 0.4568046033382416, + "learning_rate": 7.002809460241123e-08, + "loss": 0.3904, + "step": 8925 + }, + { + "epoch": 2.8560153583617747, + "grad_norm": 0.4974222481250763, + "learning_rate": 6.97179289588118e-08, + "loss": 0.4638, + "step": 8926 + }, + { + "epoch": 2.8563353242320817, + "grad_norm": 0.5003941655158997, + "learning_rate": 6.940844690475291e-08, + "loss": 0.4169, + "step": 8927 + }, + { + "epoch": 2.856655290102389, + "grad_norm": 0.49940621852874756, + "learning_rate": 6.909964848314465e-08, + "loss": 0.4534, + "step": 8928 + }, + { + "epoch": 2.856975255972696, + "grad_norm": 0.493630588054657, + "learning_rate": 6.879153373680169e-08, + "loss": 0.4351, + "step": 8929 + }, + { + "epoch": 2.857295221843003, + "grad_norm": 0.43891242146492004, + "learning_rate": 6.848410270844596e-08, + "loss": 0.411, + "step": 8930 + }, + { + "epoch": 2.8576151877133107, + "grad_norm": 0.4711717367172241, + "learning_rate": 6.817735544070114e-08, + "loss": 0.4678, + "step": 8931 + }, + { + "epoch": 2.8579351535836177, + "grad_norm": 0.483100950717926, + "learning_rate": 6.787129197609988e-08, + "loss": 0.4362, + "step": 8932 + }, + { + "epoch": 2.8582551194539247, + "grad_norm": 0.5003790855407715, + "learning_rate": 6.75659123570771e-08, + "loss": 0.4405, + "step": 8933 + }, + { + "epoch": 2.858575085324232, + "grad_norm": 0.5048513412475586, + "learning_rate": 6.726121662597396e-08, + "loss": 0.4123, + "step": 8934 + }, + { + "epoch": 2.8588950511945392, + "grad_norm": 0.5274044275283813, + "learning_rate": 6.695720482503776e-08, + "loss": 0.5795, + "step": 8935 + }, + { + "epoch": 2.8592150170648463, + "grad_norm": 0.46992045640945435, + "learning_rate": 6.665387699641979e-08, + "loss": 0.3656, + "step": 8936 + }, + { + "epoch": 2.8595349829351537, + "grad_norm": 0.5334693789482117, + "learning_rate": 6.635123318217696e-08, + "loss": 0.5043, + "step": 8937 + }, + { + "epoch": 2.8598549488054608, + "grad_norm": 0.4710192084312439, + "learning_rate": 6.60492734242707e-08, + "loss": 0.4081, + "step": 8938 + }, + { + "epoch": 2.860174914675768, + "grad_norm": 0.4978450834751129, + "learning_rate": 6.574799776456864e-08, + "loss": 0.4646, + "step": 8939 + }, + { + "epoch": 2.8604948805460753, + "grad_norm": 0.4480099380016327, + "learning_rate": 6.544740624484236e-08, + "loss": 0.4159, + "step": 8940 + }, + { + "epoch": 2.8608148464163823, + "grad_norm": 0.4901712238788605, + "learning_rate": 6.514749890676963e-08, + "loss": 0.4074, + "step": 8941 + }, + { + "epoch": 2.8611348122866893, + "grad_norm": 0.4576753079891205, + "learning_rate": 6.484827579193331e-08, + "loss": 0.4081, + "step": 8942 + }, + { + "epoch": 2.861454778156997, + "grad_norm": 0.48580124974250793, + "learning_rate": 6.454973694182021e-08, + "loss": 0.4985, + "step": 8943 + }, + { + "epoch": 2.861774744027304, + "grad_norm": 0.4531460702419281, + "learning_rate": 6.425188239782387e-08, + "loss": 0.3939, + "step": 8944 + }, + { + "epoch": 2.862094709897611, + "grad_norm": 0.507771372795105, + "learning_rate": 6.395471220124183e-08, + "loss": 0.4617, + "step": 8945 + }, + { + "epoch": 2.8624146757679183, + "grad_norm": 0.4530040919780731, + "learning_rate": 6.365822639327724e-08, + "loss": 0.3676, + "step": 8946 + }, + { + "epoch": 2.8627346416382253, + "grad_norm": 0.5020298361778259, + "learning_rate": 6.33624250150383e-08, + "loss": 0.4218, + "step": 8947 + }, + { + "epoch": 2.8630546075085324, + "grad_norm": 0.5119328498840332, + "learning_rate": 6.306730810753725e-08, + "loss": 0.4719, + "step": 8948 + }, + { + "epoch": 2.86337457337884, + "grad_norm": 0.49358975887298584, + "learning_rate": 6.27728757116941e-08, + "loss": 0.4098, + "step": 8949 + }, + { + "epoch": 2.863694539249147, + "grad_norm": 0.47846266627311707, + "learning_rate": 6.247912786833066e-08, + "loss": 0.4491, + "step": 8950 + }, + { + "epoch": 2.864014505119454, + "grad_norm": 0.46984216570854187, + "learning_rate": 6.218606461817656e-08, + "loss": 0.4107, + "step": 8951 + }, + { + "epoch": 2.8643344709897613, + "grad_norm": 0.5021507143974304, + "learning_rate": 6.18936860018654e-08, + "loss": 0.4461, + "step": 8952 + }, + { + "epoch": 2.8646544368600684, + "grad_norm": 0.5089204907417297, + "learning_rate": 6.160199205993422e-08, + "loss": 0.481, + "step": 8953 + }, + { + "epoch": 2.8649744027303754, + "grad_norm": 0.551600992679596, + "learning_rate": 6.131098283282899e-08, + "loss": 0.4607, + "step": 8954 + }, + { + "epoch": 2.8652943686006824, + "grad_norm": 0.5164521932601929, + "learning_rate": 6.102065836089632e-08, + "loss": 0.438, + "step": 8955 + }, + { + "epoch": 2.86561433447099, + "grad_norm": 0.45446866750717163, + "learning_rate": 6.07310186843918e-08, + "loss": 0.412, + "step": 8956 + }, + { + "epoch": 2.865934300341297, + "grad_norm": 0.495629221200943, + "learning_rate": 6.044206384347384e-08, + "loss": 0.4731, + "step": 8957 + }, + { + "epoch": 2.866254266211604, + "grad_norm": 0.4896862804889679, + "learning_rate": 6.015379387820541e-08, + "loss": 0.4561, + "step": 8958 + }, + { + "epoch": 2.8665742320819114, + "grad_norm": 0.5254265666007996, + "learning_rate": 5.986620882855676e-08, + "loss": 0.4918, + "step": 8959 + }, + { + "epoch": 2.8668941979522184, + "grad_norm": 0.46429502964019775, + "learning_rate": 5.9579308734400984e-08, + "loss": 0.4501, + "step": 8960 + }, + { + "epoch": 2.8672141638225255, + "grad_norm": 0.4449886083602905, + "learning_rate": 5.929309363551794e-08, + "loss": 0.4275, + "step": 8961 + }, + { + "epoch": 2.8675341296928325, + "grad_norm": 0.5188478231430054, + "learning_rate": 5.900756357159143e-08, + "loss": 0.4413, + "step": 8962 + }, + { + "epoch": 2.86785409556314, + "grad_norm": 0.5296725034713745, + "learning_rate": 5.8722718582209796e-08, + "loss": 0.4131, + "step": 8963 + }, + { + "epoch": 2.868174061433447, + "grad_norm": 0.5184751152992249, + "learning_rate": 5.8438558706868095e-08, + "loss": 0.493, + "step": 8964 + }, + { + "epoch": 2.868494027303754, + "grad_norm": 0.502002477645874, + "learning_rate": 5.8155083984964835e-08, + "loss": 0.4376, + "step": 8965 + }, + { + "epoch": 2.8688139931740615, + "grad_norm": 0.5247063636779785, + "learning_rate": 5.7872294455804666e-08, + "loss": 0.4131, + "step": 8966 + }, + { + "epoch": 2.8691339590443685, + "grad_norm": 0.4924825429916382, + "learning_rate": 5.7590190158596235e-08, + "loss": 0.486, + "step": 8967 + }, + { + "epoch": 2.8694539249146755, + "grad_norm": 0.506610631942749, + "learning_rate": 5.730877113245381e-08, + "loss": 0.4687, + "step": 8968 + }, + { + "epoch": 2.869773890784983, + "grad_norm": 0.45550239086151123, + "learning_rate": 5.702803741639618e-08, + "loss": 0.3963, + "step": 8969 + }, + { + "epoch": 2.87009385665529, + "grad_norm": 0.4927195608615875, + "learning_rate": 5.6747989049348304e-08, + "loss": 0.415, + "step": 8970 + }, + { + "epoch": 2.870413822525597, + "grad_norm": 0.47503048181533813, + "learning_rate": 5.646862607013859e-08, + "loss": 0.457, + "step": 8971 + }, + { + "epoch": 2.8707337883959045, + "grad_norm": 0.5261685252189636, + "learning_rate": 5.618994851750104e-08, + "loss": 0.4459, + "step": 8972 + }, + { + "epoch": 2.8710537542662116, + "grad_norm": 0.47380179166793823, + "learning_rate": 5.5911956430074187e-08, + "loss": 0.4289, + "step": 8973 + }, + { + "epoch": 2.8713737201365186, + "grad_norm": 0.4902719557285309, + "learning_rate": 5.563464984640221e-08, + "loss": 0.4489, + "step": 8974 + }, + { + "epoch": 2.871693686006826, + "grad_norm": 0.48171642422676086, + "learning_rate": 5.535802880493491e-08, + "loss": 0.4179, + "step": 8975 + }, + { + "epoch": 2.872013651877133, + "grad_norm": 0.5338441729545593, + "learning_rate": 5.508209334402548e-08, + "loss": 0.4554, + "step": 8976 + }, + { + "epoch": 2.87233361774744, + "grad_norm": 0.46447697281837463, + "learning_rate": 5.4806843501932774e-08, + "loss": 0.438, + "step": 8977 + }, + { + "epoch": 2.8726535836177476, + "grad_norm": 0.5013840794563293, + "learning_rate": 5.453227931682015e-08, + "loss": 0.4158, + "step": 8978 + }, + { + "epoch": 2.8729735494880546, + "grad_norm": 0.5281605124473572, + "learning_rate": 5.42584008267566e-08, + "loss": 0.4786, + "step": 8979 + }, + { + "epoch": 2.8732935153583616, + "grad_norm": 0.4782954156398773, + "learning_rate": 5.398520806971563e-08, + "loss": 0.3993, + "step": 8980 + }, + { + "epoch": 2.873613481228669, + "grad_norm": 0.5424149036407471, + "learning_rate": 5.371270108357585e-08, + "loss": 0.4876, + "step": 8981 + }, + { + "epoch": 2.873933447098976, + "grad_norm": 0.4415556788444519, + "learning_rate": 5.3440879906120344e-08, + "loss": 0.4059, + "step": 8982 + }, + { + "epoch": 2.874253412969283, + "grad_norm": 0.4531678259372711, + "learning_rate": 5.316974457503787e-08, + "loss": 0.4149, + "step": 8983 + }, + { + "epoch": 2.8745733788395906, + "grad_norm": 0.543174147605896, + "learning_rate": 5.2899295127921133e-08, + "loss": 0.4273, + "step": 8984 + }, + { + "epoch": 2.8748933447098977, + "grad_norm": 0.549419105052948, + "learning_rate": 5.262953160226958e-08, + "loss": 0.4826, + "step": 8985 + }, + { + "epoch": 2.8752133105802047, + "grad_norm": 0.48486778140068054, + "learning_rate": 5.23604540354844e-08, + "loss": 0.4248, + "step": 8986 + }, + { + "epoch": 2.875533276450512, + "grad_norm": 0.6450443863868713, + "learning_rate": 5.2092062464875194e-08, + "loss": 0.4711, + "step": 8987 + }, + { + "epoch": 2.875853242320819, + "grad_norm": 0.492783784866333, + "learning_rate": 5.182435692765331e-08, + "loss": 0.4773, + "step": 8988 + }, + { + "epoch": 2.876173208191126, + "grad_norm": 0.47193995118141174, + "learning_rate": 5.155733746093683e-08, + "loss": 0.4228, + "step": 8989 + }, + { + "epoch": 2.8764931740614337, + "grad_norm": 0.4988788962364197, + "learning_rate": 5.1291004101749476e-08, + "loss": 0.5133, + "step": 8990 + }, + { + "epoch": 2.8768131399317407, + "grad_norm": 0.5246462821960449, + "learning_rate": 5.1025356887017815e-08, + "loss": 0.4284, + "step": 8991 + }, + { + "epoch": 2.8771331058020477, + "grad_norm": 0.47200262546539307, + "learning_rate": 5.076039585357406e-08, + "loss": 0.3902, + "step": 8992 + }, + { + "epoch": 2.877453071672355, + "grad_norm": 0.5065959692001343, + "learning_rate": 5.049612103815493e-08, + "loss": 0.4616, + "step": 8993 + }, + { + "epoch": 2.8777730375426622, + "grad_norm": 0.5456526279449463, + "learning_rate": 5.023253247740334e-08, + "loss": 0.4638, + "step": 8994 + }, + { + "epoch": 2.8780930034129693, + "grad_norm": 0.4624581038951874, + "learning_rate": 4.996963020786616e-08, + "loss": 0.4235, + "step": 8995 + }, + { + "epoch": 2.8784129692832767, + "grad_norm": 0.5095905065536499, + "learning_rate": 4.970741426599479e-08, + "loss": 0.5033, + "step": 8996 + }, + { + "epoch": 2.8787329351535837, + "grad_norm": 0.47514355182647705, + "learning_rate": 4.944588468814571e-08, + "loss": 0.4508, + "step": 8997 + }, + { + "epoch": 2.8790529010238908, + "grad_norm": 0.4647156596183777, + "learning_rate": 4.91850415105799e-08, + "loss": 0.4389, + "step": 8998 + }, + { + "epoch": 2.879372866894198, + "grad_norm": 0.5182019472122192, + "learning_rate": 4.8924884769464e-08, + "loss": 0.5105, + "step": 8999 + }, + { + "epoch": 2.8796928327645053, + "grad_norm": 0.4541059732437134, + "learning_rate": 4.86654145008697e-08, + "loss": 0.3809, + "step": 9000 + }, + { + "epoch": 2.8800127986348123, + "grad_norm": 0.5203688144683838, + "learning_rate": 4.840663074077212e-08, + "loss": 0.4513, + "step": 9001 + }, + { + "epoch": 2.8803327645051193, + "grad_norm": 0.5316109657287598, + "learning_rate": 4.814853352505255e-08, + "loss": 0.4178, + "step": 9002 + }, + { + "epoch": 2.8806527303754264, + "grad_norm": 0.5072743892669678, + "learning_rate": 4.789112288949515e-08, + "loss": 0.4966, + "step": 9003 + }, + { + "epoch": 2.880972696245734, + "grad_norm": 0.45236936211586, + "learning_rate": 4.7634398869791375e-08, + "loss": 0.4092, + "step": 9004 + }, + { + "epoch": 2.881292662116041, + "grad_norm": 0.5064358711242676, + "learning_rate": 4.737836150153607e-08, + "loss": 0.4787, + "step": 9005 + }, + { + "epoch": 2.881612627986348, + "grad_norm": 0.48052626848220825, + "learning_rate": 4.7123010820229186e-08, + "loss": 0.4048, + "step": 9006 + }, + { + "epoch": 2.8819325938566553, + "grad_norm": 0.50408536195755, + "learning_rate": 4.6868346861275725e-08, + "loss": 0.4808, + "step": 9007 + }, + { + "epoch": 2.8822525597269624, + "grad_norm": 0.4835711121559143, + "learning_rate": 4.6614369659984116e-08, + "loss": 0.4072, + "step": 9008 + }, + { + "epoch": 2.8825725255972694, + "grad_norm": 0.5072152614593506, + "learning_rate": 4.636107925156896e-08, + "loss": 0.506, + "step": 9009 + }, + { + "epoch": 2.882892491467577, + "grad_norm": 0.5044692158699036, + "learning_rate": 4.610847567115051e-08, + "loss": 0.464, + "step": 9010 + }, + { + "epoch": 2.883212457337884, + "grad_norm": 0.4609028100967407, + "learning_rate": 4.5856558953750744e-08, + "loss": 0.3974, + "step": 9011 + }, + { + "epoch": 2.883532423208191, + "grad_norm": 0.5189031362533569, + "learning_rate": 4.5605329134299495e-08, + "loss": 0.499, + "step": 9012 + }, + { + "epoch": 2.8838523890784984, + "grad_norm": 0.4667864739894867, + "learning_rate": 4.5354786247629456e-08, + "loss": 0.3638, + "step": 9013 + }, + { + "epoch": 2.8841723549488054, + "grad_norm": 0.45510849356651306, + "learning_rate": 4.510493032847896e-08, + "loss": 0.4153, + "step": 9014 + }, + { + "epoch": 2.8844923208191124, + "grad_norm": 0.461737722158432, + "learning_rate": 4.485576141149084e-08, + "loss": 0.4357, + "step": 9015 + }, + { + "epoch": 2.88481228668942, + "grad_norm": 0.48075640201568604, + "learning_rate": 4.460727953121247e-08, + "loss": 0.4205, + "step": 9016 + }, + { + "epoch": 2.885132252559727, + "grad_norm": 0.39439621567726135, + "learning_rate": 4.435948472209628e-08, + "loss": 0.3687, + "step": 9017 + }, + { + "epoch": 2.885452218430034, + "grad_norm": 0.5058373212814331, + "learning_rate": 4.411237701849924e-08, + "loss": 0.5451, + "step": 9018 + }, + { + "epoch": 2.8857721843003414, + "grad_norm": 0.4888904392719269, + "learning_rate": 4.386595645468339e-08, + "loss": 0.4556, + "step": 9019 + }, + { + "epoch": 2.8860921501706485, + "grad_norm": 0.5107361674308777, + "learning_rate": 4.362022306481473e-08, + "loss": 0.394, + "step": 9020 + }, + { + "epoch": 2.8864121160409555, + "grad_norm": 0.5496082305908203, + "learning_rate": 4.337517688296544e-08, + "loss": 0.4273, + "step": 9021 + }, + { + "epoch": 2.886732081911263, + "grad_norm": 0.5090062022209167, + "learning_rate": 4.313081794311058e-08, + "loss": 0.4966, + "step": 9022 + }, + { + "epoch": 2.88705204778157, + "grad_norm": 0.5143817663192749, + "learning_rate": 4.288714627913082e-08, + "loss": 0.4048, + "step": 9023 + }, + { + "epoch": 2.887372013651877, + "grad_norm": 0.5198468565940857, + "learning_rate": 4.2644161924811353e-08, + "loss": 0.4356, + "step": 9024 + }, + { + "epoch": 2.8876919795221845, + "grad_norm": 0.5902714133262634, + "learning_rate": 4.240186491384357e-08, + "loss": 0.5307, + "step": 9025 + }, + { + "epoch": 2.8880119453924915, + "grad_norm": 0.4560842216014862, + "learning_rate": 4.216025527982115e-08, + "loss": 0.387, + "step": 9026 + }, + { + "epoch": 2.8883319112627985, + "grad_norm": 0.4865311086177826, + "learning_rate": 4.1919333056243964e-08, + "loss": 0.4501, + "step": 9027 + }, + { + "epoch": 2.888651877133106, + "grad_norm": 0.5078125596046448, + "learning_rate": 4.1679098276515286e-08, + "loss": 0.476, + "step": 9028 + }, + { + "epoch": 2.888971843003413, + "grad_norm": 0.5110453963279724, + "learning_rate": 4.143955097394459e-08, + "loss": 0.4446, + "step": 9029 + }, + { + "epoch": 2.88929180887372, + "grad_norm": 0.503882884979248, + "learning_rate": 4.120069118174586e-08, + "loss": 0.4279, + "step": 9030 + }, + { + "epoch": 2.8896117747440275, + "grad_norm": 0.534324049949646, + "learning_rate": 4.0962518933037046e-08, + "loss": 0.4493, + "step": 9031 + }, + { + "epoch": 2.8899317406143346, + "grad_norm": 0.49789708852767944, + "learning_rate": 4.0725034260840626e-08, + "loss": 0.3742, + "step": 9032 + }, + { + "epoch": 2.8902517064846416, + "grad_norm": 0.5121422410011292, + "learning_rate": 4.04882371980847e-08, + "loss": 0.4662, + "step": 9033 + }, + { + "epoch": 2.890571672354949, + "grad_norm": 0.4967876970767975, + "learning_rate": 4.0252127777600233e-08, + "loss": 0.4169, + "step": 9034 + }, + { + "epoch": 2.890891638225256, + "grad_norm": 0.5471508502960205, + "learning_rate": 4.001670603212604e-08, + "loss": 0.4557, + "step": 9035 + }, + { + "epoch": 2.891211604095563, + "grad_norm": 0.46366459131240845, + "learning_rate": 3.978197199430212e-08, + "loss": 0.4757, + "step": 9036 + }, + { + "epoch": 2.8915315699658706, + "grad_norm": 0.5181254148483276, + "learning_rate": 3.9547925696675206e-08, + "loss": 0.4907, + "step": 9037 + }, + { + "epoch": 2.8918515358361776, + "grad_norm": 0.5185445547103882, + "learning_rate": 3.931456717169546e-08, + "loss": 0.4245, + "step": 9038 + }, + { + "epoch": 2.8921715017064846, + "grad_norm": 0.4751928448677063, + "learning_rate": 3.908189645171867e-08, + "loss": 0.3828, + "step": 9039 + }, + { + "epoch": 2.8924914675767917, + "grad_norm": 0.5134016871452332, + "learning_rate": 3.8849913569005694e-08, + "loss": 0.4076, + "step": 9040 + }, + { + "epoch": 2.892811433447099, + "grad_norm": 0.4868691563606262, + "learning_rate": 3.86186185557208e-08, + "loss": 0.4647, + "step": 9041 + }, + { + "epoch": 2.893131399317406, + "grad_norm": 0.48324650526046753, + "learning_rate": 3.838801144393278e-08, + "loss": 0.4142, + "step": 9042 + }, + { + "epoch": 2.893451365187713, + "grad_norm": 0.46404677629470825, + "learning_rate": 3.8158092265616065e-08, + "loss": 0.4704, + "step": 9043 + }, + { + "epoch": 2.89377133105802, + "grad_norm": 0.46939048171043396, + "learning_rate": 3.792886105264848e-08, + "loss": 0.3937, + "step": 9044 + }, + { + "epoch": 2.8940912969283277, + "grad_norm": 0.4731648564338684, + "learning_rate": 3.7700317836814605e-08, + "loss": 0.4177, + "step": 9045 + }, + { + "epoch": 2.8944112627986347, + "grad_norm": 0.4886084496974945, + "learning_rate": 3.7472462649801314e-08, + "loss": 0.474, + "step": 9046 + }, + { + "epoch": 2.8947312286689417, + "grad_norm": 0.5067484378814697, + "learning_rate": 3.7245295523201106e-08, + "loss": 0.434, + "step": 9047 + }, + { + "epoch": 2.895051194539249, + "grad_norm": 0.4630763530731201, + "learning_rate": 3.701881648851102e-08, + "loss": 0.4451, + "step": 9048 + }, + { + "epoch": 2.895371160409556, + "grad_norm": 0.509672999382019, + "learning_rate": 3.679302557713205e-08, + "loss": 0.4372, + "step": 9049 + }, + { + "epoch": 2.8956911262798632, + "grad_norm": 0.4773665964603424, + "learning_rate": 3.656792282037136e-08, + "loss": 0.4372, + "step": 9050 + }, + { + "epoch": 2.8960110921501707, + "grad_norm": 0.5030218958854675, + "learning_rate": 3.634350824943955e-08, + "loss": 0.389, + "step": 9051 + }, + { + "epoch": 2.8963310580204777, + "grad_norm": 0.48134681582450867, + "learning_rate": 3.6119781895451734e-08, + "loss": 0.4991, + "step": 9052 + }, + { + "epoch": 2.8966510238907848, + "grad_norm": 0.43066349625587463, + "learning_rate": 3.5896743789427535e-08, + "loss": 0.3744, + "step": 9053 + }, + { + "epoch": 2.8969709897610922, + "grad_norm": 0.4765835702419281, + "learning_rate": 3.567439396229222e-08, + "loss": 0.4312, + "step": 9054 + }, + { + "epoch": 2.8972909556313993, + "grad_norm": 0.476105272769928, + "learning_rate": 3.545273244487391e-08, + "loss": 0.4339, + "step": 9055 + }, + { + "epoch": 2.8976109215017063, + "grad_norm": 0.5043574571609497, + "learning_rate": 3.523175926790745e-08, + "loss": 0.4558, + "step": 9056 + }, + { + "epoch": 2.8979308873720138, + "grad_norm": 0.5222077369689941, + "learning_rate": 3.501147446202946e-08, + "loss": 0.4618, + "step": 9057 + }, + { + "epoch": 2.898250853242321, + "grad_norm": 0.4806877076625824, + "learning_rate": 3.479187805778439e-08, + "loss": 0.3512, + "step": 9058 + }, + { + "epoch": 2.898570819112628, + "grad_norm": 0.4741007685661316, + "learning_rate": 3.4572970085617883e-08, + "loss": 0.4531, + "step": 9059 + }, + { + "epoch": 2.8988907849829353, + "grad_norm": 0.5301935076713562, + "learning_rate": 3.4354750575883446e-08, + "loss": 0.4383, + "step": 9060 + }, + { + "epoch": 2.8992107508532423, + "grad_norm": 0.5026822090148926, + "learning_rate": 3.413721955883631e-08, + "loss": 0.4406, + "step": 9061 + }, + { + "epoch": 2.8995307167235493, + "grad_norm": 0.4861147701740265, + "learning_rate": 3.3920377064637355e-08, + "loss": 0.4568, + "step": 9062 + }, + { + "epoch": 2.899850682593857, + "grad_norm": 0.5339757800102234, + "learning_rate": 3.370422312335309e-08, + "loss": 0.4428, + "step": 9063 + }, + { + "epoch": 2.900170648464164, + "grad_norm": 0.5031761527061462, + "learning_rate": 3.3488757764952306e-08, + "loss": 0.3909, + "step": 9064 + }, + { + "epoch": 2.900490614334471, + "grad_norm": 0.5004369616508484, + "learning_rate": 3.3273981019310566e-08, + "loss": 0.4481, + "step": 9065 + }, + { + "epoch": 2.9008105802047783, + "grad_norm": 0.46735748648643494, + "learning_rate": 3.305989291620626e-08, + "loss": 0.406, + "step": 9066 + }, + { + "epoch": 2.9011305460750854, + "grad_norm": 0.550030529499054, + "learning_rate": 3.2846493485323426e-08, + "loss": 0.4686, + "step": 9067 + }, + { + "epoch": 2.9014505119453924, + "grad_norm": 0.4990123510360718, + "learning_rate": 3.263378275624951e-08, + "loss": 0.4335, + "step": 9068 + }, + { + "epoch": 2.9017704778157, + "grad_norm": 0.43858641386032104, + "learning_rate": 3.2421760758477585e-08, + "loss": 0.4257, + "step": 9069 + }, + { + "epoch": 2.902090443686007, + "grad_norm": 0.4837779104709625, + "learning_rate": 3.221042752140524e-08, + "loss": 0.4663, + "step": 9070 + }, + { + "epoch": 2.902410409556314, + "grad_norm": 0.4876711368560791, + "learning_rate": 3.1999783074332934e-08, + "loss": 0.4159, + "step": 9071 + }, + { + "epoch": 2.9027303754266214, + "grad_norm": 0.5029987692832947, + "learning_rate": 3.1789827446467304e-08, + "loss": 0.4138, + "step": 9072 + }, + { + "epoch": 2.9030503412969284, + "grad_norm": 0.4941578507423401, + "learning_rate": 3.15805606669195e-08, + "loss": 0.449, + "step": 9073 + }, + { + "epoch": 2.9033703071672354, + "grad_norm": 0.4757187068462372, + "learning_rate": 3.137198276470355e-08, + "loss": 0.4441, + "step": 9074 + }, + { + "epoch": 2.903690273037543, + "grad_norm": 0.509201169013977, + "learning_rate": 3.116409376874019e-08, + "loss": 0.4959, + "step": 9075 + }, + { + "epoch": 2.90401023890785, + "grad_norm": 0.4921058714389801, + "learning_rate": 3.095689370785249e-08, + "loss": 0.4105, + "step": 9076 + }, + { + "epoch": 2.904330204778157, + "grad_norm": 0.48168763518333435, + "learning_rate": 3.0750382610769124e-08, + "loss": 0.4493, + "step": 9077 + }, + { + "epoch": 2.9046501706484644, + "grad_norm": 0.49827325344085693, + "learning_rate": 3.0544560506123863e-08, + "loss": 0.3925, + "step": 9078 + }, + { + "epoch": 2.9049701365187715, + "grad_norm": 0.48839378356933594, + "learning_rate": 3.0339427422453327e-08, + "loss": 0.3944, + "step": 9079 + }, + { + "epoch": 2.9052901023890785, + "grad_norm": 0.4565756916999817, + "learning_rate": 3.013498338820031e-08, + "loss": 0.4225, + "step": 9080 + }, + { + "epoch": 2.9056100682593855, + "grad_norm": 0.5143803954124451, + "learning_rate": 2.993122843171048e-08, + "loss": 0.4555, + "step": 9081 + }, + { + "epoch": 2.905930034129693, + "grad_norm": 0.5379270911216736, + "learning_rate": 2.9728162581234566e-08, + "loss": 0.4531, + "step": 9082 + }, + { + "epoch": 2.90625, + "grad_norm": 0.5080519914627075, + "learning_rate": 2.9525785864928934e-08, + "loss": 0.4265, + "step": 9083 + }, + { + "epoch": 2.906569965870307, + "grad_norm": 0.4725300669670105, + "learning_rate": 2.9324098310852257e-08, + "loss": 0.4317, + "step": 9084 + }, + { + "epoch": 2.9068899317406145, + "grad_norm": 0.4862784743309021, + "learning_rate": 2.912309994696938e-08, + "loss": 0.4418, + "step": 9085 + }, + { + "epoch": 2.9072098976109215, + "grad_norm": 0.4593566060066223, + "learning_rate": 2.8922790801149126e-08, + "loss": 0.4044, + "step": 9086 + }, + { + "epoch": 2.9075298634812285, + "grad_norm": 0.4919830858707428, + "learning_rate": 2.8723170901163722e-08, + "loss": 0.4559, + "step": 9087 + }, + { + "epoch": 2.9078498293515356, + "grad_norm": 0.5015954375267029, + "learning_rate": 2.8524240274691582e-08, + "loss": 0.464, + "step": 9088 + }, + { + "epoch": 2.908169795221843, + "grad_norm": 0.5168560147285461, + "learning_rate": 2.8325998949314536e-08, + "loss": 0.4154, + "step": 9089 + }, + { + "epoch": 2.90848976109215, + "grad_norm": 0.5248748660087585, + "learning_rate": 2.8128446952518373e-08, + "loss": 0.4314, + "step": 9090 + }, + { + "epoch": 2.908809726962457, + "grad_norm": 0.4684077799320221, + "learning_rate": 2.7931584311694514e-08, + "loss": 0.4436, + "step": 9091 + }, + { + "epoch": 2.9091296928327646, + "grad_norm": 0.5102325677871704, + "learning_rate": 2.773541105413835e-08, + "loss": 0.4127, + "step": 9092 + }, + { + "epoch": 2.9094496587030716, + "grad_norm": 0.5072308778762817, + "learning_rate": 2.7539927207048678e-08, + "loss": 0.4392, + "step": 9093 + }, + { + "epoch": 2.9097696245733786, + "grad_norm": 0.4549092650413513, + "learning_rate": 2.7345132797530482e-08, + "loss": 0.4767, + "step": 9094 + }, + { + "epoch": 2.910089590443686, + "grad_norm": 0.4574567377567291, + "learning_rate": 2.7151027852591606e-08, + "loss": 0.411, + "step": 9095 + }, + { + "epoch": 2.910409556313993, + "grad_norm": 0.527850866317749, + "learning_rate": 2.6957612399145517e-08, + "loss": 0.4501, + "step": 9096 + }, + { + "epoch": 2.9107295221843, + "grad_norm": 0.5241904258728027, + "learning_rate": 2.6764886464009097e-08, + "loss": 0.4183, + "step": 9097 + }, + { + "epoch": 2.9110494880546076, + "grad_norm": 0.5417770147323608, + "learning_rate": 2.6572850073903755e-08, + "loss": 0.4734, + "step": 9098 + }, + { + "epoch": 2.9113694539249146, + "grad_norm": 0.48998844623565674, + "learning_rate": 2.638150325545652e-08, + "loss": 0.4398, + "step": 9099 + }, + { + "epoch": 2.9116894197952217, + "grad_norm": 0.5481809973716736, + "learning_rate": 2.6190846035196727e-08, + "loss": 0.4864, + "step": 9100 + }, + { + "epoch": 2.912009385665529, + "grad_norm": 0.5044158101081848, + "learning_rate": 2.6000878439559896e-08, + "loss": 0.4075, + "step": 9101 + }, + { + "epoch": 2.912329351535836, + "grad_norm": 0.5413654446601868, + "learning_rate": 2.5811600494885512e-08, + "loss": 0.4452, + "step": 9102 + }, + { + "epoch": 2.912649317406143, + "grad_norm": 0.4841751456260681, + "learning_rate": 2.5623012227416476e-08, + "loss": 0.4071, + "step": 9103 + }, + { + "epoch": 2.9129692832764507, + "grad_norm": 0.5086470246315002, + "learning_rate": 2.5435113663301315e-08, + "loss": 0.4196, + "step": 9104 + }, + { + "epoch": 2.9132892491467577, + "grad_norm": 0.4594883620738983, + "learning_rate": 2.5247904828591963e-08, + "loss": 0.4427, + "step": 9105 + }, + { + "epoch": 2.9136092150170647, + "grad_norm": 0.4493332803249359, + "learning_rate": 2.5061385749245437e-08, + "loss": 0.4014, + "step": 9106 + }, + { + "epoch": 2.913929180887372, + "grad_norm": 0.46458470821380615, + "learning_rate": 2.4875556451123274e-08, + "loss": 0.4496, + "step": 9107 + }, + { + "epoch": 2.914249146757679, + "grad_norm": 0.5013710856437683, + "learning_rate": 2.469041695998986e-08, + "loss": 0.411, + "step": 9108 + }, + { + "epoch": 2.9145691126279862, + "grad_norm": 0.48893997073173523, + "learning_rate": 2.4505967301515778e-08, + "loss": 0.4391, + "step": 9109 + }, + { + "epoch": 2.9148890784982937, + "grad_norm": 0.49383464455604553, + "learning_rate": 2.4322207501274453e-08, + "loss": 0.4164, + "step": 9110 + }, + { + "epoch": 2.9152090443686007, + "grad_norm": 0.44160425662994385, + "learning_rate": 2.413913758474551e-08, + "loss": 0.4516, + "step": 9111 + }, + { + "epoch": 2.9155290102389078, + "grad_norm": 0.47670525312423706, + "learning_rate": 2.3956757577311418e-08, + "loss": 0.4217, + "step": 9112 + }, + { + "epoch": 2.9158489761092152, + "grad_norm": 0.49880895018577576, + "learning_rate": 2.377506750425862e-08, + "loss": 0.4099, + "step": 9113 + }, + { + "epoch": 2.9161689419795223, + "grad_norm": 0.49023520946502686, + "learning_rate": 2.359406739077974e-08, + "loss": 0.4096, + "step": 9114 + }, + { + "epoch": 2.9164889078498293, + "grad_norm": 0.5317866206169128, + "learning_rate": 2.341375726197026e-08, + "loss": 0.4396, + "step": 9115 + }, + { + "epoch": 2.9168088737201368, + "grad_norm": 0.4866684675216675, + "learning_rate": 2.3234137142830182e-08, + "loss": 0.4448, + "step": 9116 + }, + { + "epoch": 2.917128839590444, + "grad_norm": 0.49172016978263855, + "learning_rate": 2.3055207058264028e-08, + "loss": 0.435, + "step": 9117 + }, + { + "epoch": 2.917448805460751, + "grad_norm": 0.4982343018054962, + "learning_rate": 2.2876967033080845e-08, + "loss": 0.5196, + "step": 9118 + }, + { + "epoch": 2.9177687713310583, + "grad_norm": 0.4825814366340637, + "learning_rate": 2.2699417091993635e-08, + "loss": 0.4358, + "step": 9119 + }, + { + "epoch": 2.9180887372013653, + "grad_norm": 0.5194673538208008, + "learning_rate": 2.2522557259620492e-08, + "loss": 0.4736, + "step": 9120 + }, + { + "epoch": 2.9184087030716723, + "grad_norm": 0.4512815773487091, + "learning_rate": 2.234638756048235e-08, + "loss": 0.4362, + "step": 9121 + }, + { + "epoch": 2.91872866894198, + "grad_norm": 0.46083351969718933, + "learning_rate": 2.217090801900634e-08, + "loss": 0.4274, + "step": 9122 + }, + { + "epoch": 2.919048634812287, + "grad_norm": 0.42566928267478943, + "learning_rate": 2.1996118659521894e-08, + "loss": 0.4126, + "step": 9123 + }, + { + "epoch": 2.919368600682594, + "grad_norm": 0.4711596369743347, + "learning_rate": 2.1822019506264612e-08, + "loss": 0.463, + "step": 9124 + }, + { + "epoch": 2.919688566552901, + "grad_norm": 0.49391430616378784, + "learning_rate": 2.164861058337353e-08, + "loss": 0.4164, + "step": 9125 + }, + { + "epoch": 2.9200085324232083, + "grad_norm": 0.547099232673645, + "learning_rate": 2.1475891914891078e-08, + "loss": 0.4504, + "step": 9126 + }, + { + "epoch": 2.9203284982935154, + "grad_norm": 0.4892997741699219, + "learning_rate": 2.1303863524765878e-08, + "loss": 0.4054, + "step": 9127 + }, + { + "epoch": 2.9206484641638224, + "grad_norm": 0.5020419955253601, + "learning_rate": 2.1132525436849406e-08, + "loss": 0.4643, + "step": 9128 + }, + { + "epoch": 2.9209684300341294, + "grad_norm": 0.5101315975189209, + "learning_rate": 2.0961877674898213e-08, + "loss": 0.4345, + "step": 9129 + }, + { + "epoch": 2.921288395904437, + "grad_norm": 0.44621843099594116, + "learning_rate": 2.0791920262572262e-08, + "loss": 0.4093, + "step": 9130 + }, + { + "epoch": 2.921608361774744, + "grad_norm": 0.457491934299469, + "learning_rate": 2.062265322343715e-08, + "loss": 0.4225, + "step": 9131 + }, + { + "epoch": 2.921928327645051, + "grad_norm": 0.5055102705955505, + "learning_rate": 2.0454076580961323e-08, + "loss": 0.4948, + "step": 9132 + }, + { + "epoch": 2.9222482935153584, + "grad_norm": 0.5012072920799255, + "learning_rate": 2.0286190358517754e-08, + "loss": 0.4174, + "step": 9133 + }, + { + "epoch": 2.9225682593856654, + "grad_norm": 0.4871751368045807, + "learning_rate": 2.0118994579385042e-08, + "loss": 0.3989, + "step": 9134 + }, + { + "epoch": 2.9228882252559725, + "grad_norm": 0.49165430665016174, + "learning_rate": 1.9952489266744647e-08, + "loss": 0.4385, + "step": 9135 + }, + { + "epoch": 2.92320819112628, + "grad_norm": 0.4882758855819702, + "learning_rate": 1.9786674443683096e-08, + "loss": 0.444, + "step": 9136 + }, + { + "epoch": 2.923528156996587, + "grad_norm": 0.45996996760368347, + "learning_rate": 1.9621550133190337e-08, + "loss": 0.4277, + "step": 9137 + }, + { + "epoch": 2.923848122866894, + "grad_norm": 0.46109628677368164, + "learning_rate": 1.945711635816083e-08, + "loss": 0.3773, + "step": 9138 + }, + { + "epoch": 2.9241680887372015, + "grad_norm": 0.5478599667549133, + "learning_rate": 1.9293373141394124e-08, + "loss": 0.4842, + "step": 9139 + }, + { + "epoch": 2.9244880546075085, + "grad_norm": 0.49445387721061707, + "learning_rate": 1.9130320505593158e-08, + "loss": 0.4225, + "step": 9140 + }, + { + "epoch": 2.9248080204778155, + "grad_norm": 0.5079352259635925, + "learning_rate": 1.896795847336541e-08, + "loss": 0.4762, + "step": 9141 + }, + { + "epoch": 2.925127986348123, + "grad_norm": 0.4407789409160614, + "learning_rate": 1.8806287067222873e-08, + "loss": 0.3709, + "step": 9142 + }, + { + "epoch": 2.92544795221843, + "grad_norm": 0.47105005383491516, + "learning_rate": 1.8645306309580948e-08, + "loss": 0.4343, + "step": 9143 + }, + { + "epoch": 2.925767918088737, + "grad_norm": 0.49983924627304077, + "learning_rate": 1.8485016222759555e-08, + "loss": 0.4664, + "step": 9144 + }, + { + "epoch": 2.9260878839590445, + "grad_norm": 0.48586347699165344, + "learning_rate": 1.83254168289837e-08, + "loss": 0.4526, + "step": 9145 + }, + { + "epoch": 2.9264078498293515, + "grad_norm": 0.5456278324127197, + "learning_rate": 1.8166508150382344e-08, + "loss": 0.4782, + "step": 9146 + }, + { + "epoch": 2.9267278156996586, + "grad_norm": 0.4486168622970581, + "learning_rate": 1.8008290208987868e-08, + "loss": 0.417, + "step": 9147 + }, + { + "epoch": 2.927047781569966, + "grad_norm": 0.5022203326225281, + "learning_rate": 1.785076302673716e-08, + "loss": 0.4686, + "step": 9148 + }, + { + "epoch": 2.927367747440273, + "grad_norm": 0.4830387532711029, + "learning_rate": 1.7693926625471646e-08, + "loss": 0.4596, + "step": 9149 + }, + { + "epoch": 2.92768771331058, + "grad_norm": 0.4588986039161682, + "learning_rate": 1.7537781026937816e-08, + "loss": 0.4222, + "step": 9150 + }, + { + "epoch": 2.9280076791808876, + "grad_norm": 0.49632179737091064, + "learning_rate": 1.738232625278391e-08, + "loss": 0.4845, + "step": 9151 + }, + { + "epoch": 2.9283276450511946, + "grad_norm": 0.4504736065864563, + "learning_rate": 1.7227562324565463e-08, + "loss": 0.38, + "step": 9152 + }, + { + "epoch": 2.9286476109215016, + "grad_norm": 0.49249497056007385, + "learning_rate": 1.7073489263739197e-08, + "loss": 0.4617, + "step": 9153 + }, + { + "epoch": 2.928967576791809, + "grad_norm": 0.46786049008369446, + "learning_rate": 1.6920107091668582e-08, + "loss": 0.4591, + "step": 9154 + }, + { + "epoch": 2.929287542662116, + "grad_norm": 0.46979671716690063, + "learning_rate": 1.676741582961994e-08, + "loss": 0.4544, + "step": 9155 + }, + { + "epoch": 2.929607508532423, + "grad_norm": 0.5029367208480835, + "learning_rate": 1.6615415498764108e-08, + "loss": 0.4066, + "step": 9156 + }, + { + "epoch": 2.9299274744027306, + "grad_norm": 0.5062602162361145, + "learning_rate": 1.6464106120175905e-08, + "loss": 0.4635, + "step": 9157 + }, + { + "epoch": 2.9302474402730376, + "grad_norm": 0.491169810295105, + "learning_rate": 1.6313487714834654e-08, + "loss": 0.4324, + "step": 9158 + }, + { + "epoch": 2.9305674061433447, + "grad_norm": 0.5217692255973816, + "learning_rate": 1.616356030362365e-08, + "loss": 0.4255, + "step": 9159 + }, + { + "epoch": 2.930887372013652, + "grad_norm": 0.4856349527835846, + "learning_rate": 1.6014323907331265e-08, + "loss": 0.4283, + "step": 9160 + }, + { + "epoch": 2.931207337883959, + "grad_norm": 0.5454949140548706, + "learning_rate": 1.5865778546648723e-08, + "loss": 0.4642, + "step": 9161 + }, + { + "epoch": 2.931527303754266, + "grad_norm": 0.49683502316474915, + "learning_rate": 1.571792424217178e-08, + "loss": 0.4336, + "step": 9162 + }, + { + "epoch": 2.9318472696245736, + "grad_norm": 0.5096780061721802, + "learning_rate": 1.557076101440125e-08, + "loss": 0.4651, + "step": 9163 + }, + { + "epoch": 2.9321672354948807, + "grad_norm": 0.47015881538391113, + "learning_rate": 1.5424288883740257e-08, + "loss": 0.4176, + "step": 9164 + }, + { + "epoch": 2.9324872013651877, + "grad_norm": 0.5371462106704712, + "learning_rate": 1.5278507870499226e-08, + "loss": 0.4615, + "step": 9165 + }, + { + "epoch": 2.9328071672354947, + "grad_norm": 0.5015416741371155, + "learning_rate": 1.513341799488921e-08, + "loss": 0.4304, + "step": 9166 + }, + { + "epoch": 2.933127133105802, + "grad_norm": 0.49367377161979675, + "learning_rate": 1.4989019277028004e-08, + "loss": 0.3683, + "step": 9167 + }, + { + "epoch": 2.9334470989761092, + "grad_norm": 0.4831542372703552, + "learning_rate": 1.4845311736936263e-08, + "loss": 0.4894, + "step": 9168 + }, + { + "epoch": 2.9337670648464163, + "grad_norm": 0.4677984416484833, + "learning_rate": 1.4702295394539157e-08, + "loss": 0.3889, + "step": 9169 + }, + { + "epoch": 2.9340870307167233, + "grad_norm": 0.5112711787223816, + "learning_rate": 1.4559970269666379e-08, + "loss": 0.4102, + "step": 9170 + }, + { + "epoch": 2.9344069965870307, + "grad_norm": 0.5255852937698364, + "learning_rate": 1.441833638205159e-08, + "loss": 0.4651, + "step": 9171 + }, + { + "epoch": 2.9347269624573378, + "grad_norm": 0.48539504408836365, + "learning_rate": 1.4277393751332414e-08, + "loss": 0.4518, + "step": 9172 + }, + { + "epoch": 2.935046928327645, + "grad_norm": 0.4464510679244995, + "learning_rate": 1.4137142397049885e-08, + "loss": 0.3935, + "step": 9173 + }, + { + "epoch": 2.9353668941979523, + "grad_norm": 0.4925348162651062, + "learning_rate": 1.3997582338651228e-08, + "loss": 0.456, + "step": 9174 + }, + { + "epoch": 2.9356868600682593, + "grad_norm": 0.48311126232147217, + "learning_rate": 1.3858713595485961e-08, + "loss": 0.4418, + "step": 9175 + }, + { + "epoch": 2.9360068259385663, + "grad_norm": 0.46080172061920166, + "learning_rate": 1.3720536186808686e-08, + "loss": 0.4259, + "step": 9176 + }, + { + "epoch": 2.936326791808874, + "grad_norm": 0.43515244126319885, + "learning_rate": 1.3583050131777408e-08, + "loss": 0.3937, + "step": 9177 + }, + { + "epoch": 2.936646757679181, + "grad_norm": 0.5481618046760559, + "learning_rate": 1.3446255449455215e-08, + "loss": 0.5069, + "step": 9178 + }, + { + "epoch": 2.936966723549488, + "grad_norm": 0.4946882426738739, + "learning_rate": 1.3310152158808598e-08, + "loss": 0.4196, + "step": 9179 + }, + { + "epoch": 2.9372866894197953, + "grad_norm": 0.5199326872825623, + "learning_rate": 1.3174740278708575e-08, + "loss": 0.4455, + "step": 9180 + }, + { + "epoch": 2.9376066552901023, + "grad_norm": 0.5129525065422058, + "learning_rate": 1.3040019827930128e-08, + "loss": 0.4764, + "step": 9181 + }, + { + "epoch": 2.9379266211604094, + "grad_norm": 0.506719708442688, + "learning_rate": 1.2905990825152204e-08, + "loss": 0.4782, + "step": 9182 + }, + { + "epoch": 2.938246587030717, + "grad_norm": 0.5135883688926697, + "learning_rate": 1.277265328895827e-08, + "loss": 0.4102, + "step": 9183 + }, + { + "epoch": 2.938566552901024, + "grad_norm": 0.4967620074748993, + "learning_rate": 1.264000723783576e-08, + "loss": 0.4614, + "step": 9184 + }, + { + "epoch": 2.938886518771331, + "grad_norm": 0.4385746419429779, + "learning_rate": 1.250805269017663e-08, + "loss": 0.4591, + "step": 9185 + }, + { + "epoch": 2.9392064846416384, + "grad_norm": 0.5029006004333496, + "learning_rate": 1.237678966427569e-08, + "loss": 0.3995, + "step": 9186 + }, + { + "epoch": 2.9395264505119454, + "grad_norm": 0.5193878412246704, + "learning_rate": 1.224621817833338e-08, + "loss": 0.4779, + "step": 9187 + }, + { + "epoch": 2.9398464163822524, + "grad_norm": 0.4947200417518616, + "learning_rate": 1.2116338250452997e-08, + "loss": 0.4176, + "step": 9188 + }, + { + "epoch": 2.94016638225256, + "grad_norm": 0.4940420091152191, + "learning_rate": 1.1987149898642914e-08, + "loss": 0.4224, + "step": 9189 + }, + { + "epoch": 2.940486348122867, + "grad_norm": 0.5115923881530762, + "learning_rate": 1.1858653140815468e-08, + "loss": 0.4605, + "step": 9190 + }, + { + "epoch": 2.940806313993174, + "grad_norm": 0.5231035351753235, + "learning_rate": 1.173084799478641e-08, + "loss": 0.4506, + "step": 9191 + }, + { + "epoch": 2.9411262798634814, + "grad_norm": 0.4822664260864258, + "learning_rate": 1.1603734478276562e-08, + "loss": 0.3945, + "step": 9192 + }, + { + "epoch": 2.9414462457337884, + "grad_norm": 0.4457946717739105, + "learning_rate": 1.1477312608910162e-08, + "loss": 0.4186, + "step": 9193 + }, + { + "epoch": 2.9417662116040955, + "grad_norm": 0.5013504028320312, + "learning_rate": 1.135158240421541e-08, + "loss": 0.495, + "step": 9194 + }, + { + "epoch": 2.942086177474403, + "grad_norm": 0.5230039954185486, + "learning_rate": 1.1226543881625584e-08, + "loss": 0.4535, + "step": 9195 + }, + { + "epoch": 2.94240614334471, + "grad_norm": 0.5289256572723389, + "learning_rate": 1.1102197058476816e-08, + "loss": 0.438, + "step": 9196 + }, + { + "epoch": 2.942726109215017, + "grad_norm": 0.5156910419464111, + "learning_rate": 1.0978541952010313e-08, + "loss": 0.4271, + "step": 9197 + }, + { + "epoch": 2.9430460750853245, + "grad_norm": 0.4751654863357544, + "learning_rate": 1.0855578579370696e-08, + "loss": 0.414, + "step": 9198 + }, + { + "epoch": 2.9433660409556315, + "grad_norm": 0.5000825524330139, + "learning_rate": 1.0733306957607104e-08, + "loss": 0.4947, + "step": 9199 + }, + { + "epoch": 2.9436860068259385, + "grad_norm": 0.5216525793075562, + "learning_rate": 1.0611727103673196e-08, + "loss": 0.3871, + "step": 9200 + }, + { + "epoch": 2.944005972696246, + "grad_norm": 0.4903882145881653, + "learning_rate": 1.0490839034425493e-08, + "loss": 0.4292, + "step": 9201 + }, + { + "epoch": 2.944325938566553, + "grad_norm": 0.49286291003227234, + "learning_rate": 1.0370642766625583e-08, + "loss": 0.424, + "step": 9202 + }, + { + "epoch": 2.94464590443686, + "grad_norm": 0.46055853366851807, + "learning_rate": 1.025113831693847e-08, + "loss": 0.4237, + "step": 9203 + }, + { + "epoch": 2.9449658703071675, + "grad_norm": 0.5028175115585327, + "learning_rate": 1.0132325701934232e-08, + "loss": 0.4496, + "step": 9204 + }, + { + "epoch": 2.9452858361774745, + "grad_norm": 0.4903533160686493, + "learning_rate": 1.0014204938085803e-08, + "loss": 0.4153, + "step": 9205 + }, + { + "epoch": 2.9456058020477816, + "grad_norm": 0.4870935082435608, + "learning_rate": 9.89677604177064e-09, + "loss": 0.42, + "step": 9206 + }, + { + "epoch": 2.9459257679180886, + "grad_norm": 0.5625002980232239, + "learning_rate": 9.780039029271271e-09, + "loss": 0.4543, + "step": 9207 + }, + { + "epoch": 2.946245733788396, + "grad_norm": 0.46420183777809143, + "learning_rate": 9.663993916772529e-09, + "loss": 0.4816, + "step": 9208 + }, + { + "epoch": 2.946565699658703, + "grad_norm": 0.46914464235305786, + "learning_rate": 9.548640720364321e-09, + "loss": 0.4378, + "step": 9209 + }, + { + "epoch": 2.94688566552901, + "grad_norm": 0.4514414370059967, + "learning_rate": 9.43397945604163e-09, + "loss": 0.3825, + "step": 9210 + }, + { + "epoch": 2.9472056313993176, + "grad_norm": 0.4555622637271881, + "learning_rate": 9.320010139700631e-09, + "loss": 0.4416, + "step": 9211 + }, + { + "epoch": 2.9475255972696246, + "grad_norm": 0.5080385804176331, + "learning_rate": 9.20673278714479e-09, + "loss": 0.4571, + "step": 9212 + }, + { + "epoch": 2.9478455631399316, + "grad_norm": 0.5202873945236206, + "learning_rate": 9.094147414078769e-09, + "loss": 0.4595, + "step": 9213 + }, + { + "epoch": 2.9481655290102387, + "grad_norm": 0.4716443121433258, + "learning_rate": 8.982254036113968e-09, + "loss": 0.399, + "step": 9214 + }, + { + "epoch": 2.948485494880546, + "grad_norm": 0.5085718035697937, + "learning_rate": 8.871052668763535e-09, + "loss": 0.44, + "step": 9215 + }, + { + "epoch": 2.948805460750853, + "grad_norm": 0.47173622250556946, + "learning_rate": 8.760543327446802e-09, + "loss": 0.4069, + "step": 9216 + }, + { + "epoch": 2.94912542662116, + "grad_norm": 0.4833856225013733, + "learning_rate": 8.650726027484846e-09, + "loss": 0.392, + "step": 9217 + }, + { + "epoch": 2.9494453924914676, + "grad_norm": 0.45571327209472656, + "learning_rate": 8.541600784104376e-09, + "loss": 0.4697, + "step": 9218 + }, + { + "epoch": 2.9497653583617747, + "grad_norm": 0.4360343813896179, + "learning_rate": 8.433167612436066e-09, + "loss": 0.3868, + "step": 9219 + }, + { + "epoch": 2.9500853242320817, + "grad_norm": 0.4683946371078491, + "learning_rate": 8.325426527514003e-09, + "loss": 0.4701, + "step": 9220 + }, + { + "epoch": 2.950405290102389, + "grad_norm": 0.5024915933609009, + "learning_rate": 8.21837754427679e-09, + "loss": 0.4405, + "step": 9221 + }, + { + "epoch": 2.950725255972696, + "grad_norm": 0.4808758795261383, + "learning_rate": 8.112020677567001e-09, + "loss": 0.4219, + "step": 9222 + }, + { + "epoch": 2.951045221843003, + "grad_norm": 0.47212451696395874, + "learning_rate": 8.006355942130616e-09, + "loss": 0.4211, + "step": 9223 + }, + { + "epoch": 2.9513651877133107, + "grad_norm": 0.5444512367248535, + "learning_rate": 7.901383352618142e-09, + "loss": 0.5328, + "step": 9224 + }, + { + "epoch": 2.9516851535836177, + "grad_norm": 0.4807533323764801, + "learning_rate": 7.797102923585154e-09, + "loss": 0.3795, + "step": 9225 + }, + { + "epoch": 2.9520051194539247, + "grad_norm": 0.5102534890174866, + "learning_rate": 7.693514669489533e-09, + "loss": 0.4764, + "step": 9226 + }, + { + "epoch": 2.952325085324232, + "grad_norm": 0.4702845811843872, + "learning_rate": 7.590618604693678e-09, + "loss": 0.4287, + "step": 9227 + }, + { + "epoch": 2.9526450511945392, + "grad_norm": 0.5033854842185974, + "learning_rate": 7.488414743464512e-09, + "loss": 0.457, + "step": 9228 + }, + { + "epoch": 2.9529650170648463, + "grad_norm": 0.5184718370437622, + "learning_rate": 7.38690309997292e-09, + "loss": 0.4333, + "step": 9229 + }, + { + "epoch": 2.9532849829351537, + "grad_norm": 0.5364457368850708, + "learning_rate": 7.286083688293755e-09, + "loss": 0.4411, + "step": 9230 + }, + { + "epoch": 2.9536049488054608, + "grad_norm": 0.47790831327438354, + "learning_rate": 7.1859565224052796e-09, + "loss": 0.4661, + "step": 9231 + }, + { + "epoch": 2.953924914675768, + "grad_norm": 0.44603395462036133, + "learning_rate": 7.0865216161902785e-09, + "loss": 0.4012, + "step": 9232 + }, + { + "epoch": 2.9542448805460753, + "grad_norm": 0.5121586918830872, + "learning_rate": 6.9877789834355e-09, + "loss": 0.4345, + "step": 9233 + }, + { + "epoch": 2.9545648464163823, + "grad_norm": 0.4695791006088257, + "learning_rate": 6.889728637832216e-09, + "loss": 0.3884, + "step": 9234 + }, + { + "epoch": 2.9548848122866893, + "grad_norm": 0.5383626818656921, + "learning_rate": 6.792370592975106e-09, + "loss": 0.4676, + "step": 9235 + }, + { + "epoch": 2.955204778156997, + "grad_norm": 0.48278743028640747, + "learning_rate": 6.6957048623628175e-09, + "loss": 0.4897, + "step": 9236 + }, + { + "epoch": 2.955524744027304, + "grad_norm": 0.47618865966796875, + "learning_rate": 6.599731459397962e-09, + "loss": 0.3991, + "step": 9237 + }, + { + "epoch": 2.955844709897611, + "grad_norm": 0.5008996725082397, + "learning_rate": 6.504450397387119e-09, + "loss": 0.4679, + "step": 9238 + }, + { + "epoch": 2.9561646757679183, + "grad_norm": 0.4668422043323517, + "learning_rate": 6.40986168954194e-09, + "loss": 0.3855, + "step": 9239 + }, + { + "epoch": 2.9564846416382253, + "grad_norm": 0.510735034942627, + "learning_rate": 6.315965348976938e-09, + "loss": 0.4149, + "step": 9240 + }, + { + "epoch": 2.9568046075085324, + "grad_norm": 0.5300773978233337, + "learning_rate": 6.222761388711141e-09, + "loss": 0.4485, + "step": 9241 + }, + { + "epoch": 2.95712457337884, + "grad_norm": 0.4763416051864624, + "learning_rate": 6.130249821666434e-09, + "loss": 0.4383, + "step": 9242 + }, + { + "epoch": 2.957444539249147, + "grad_norm": 0.5363229513168335, + "learning_rate": 6.038430660670891e-09, + "loss": 0.4253, + "step": 9243 + }, + { + "epoch": 2.957764505119454, + "grad_norm": 0.5187003016471863, + "learning_rate": 5.947303918454328e-09, + "loss": 0.4287, + "step": 9244 + }, + { + "epoch": 2.9580844709897613, + "grad_norm": 0.49342259764671326, + "learning_rate": 5.856869607652749e-09, + "loss": 0.3999, + "step": 9245 + }, + { + "epoch": 2.9584044368600684, + "grad_norm": 0.5214436650276184, + "learning_rate": 5.767127740803902e-09, + "loss": 0.4429, + "step": 9246 + }, + { + "epoch": 2.9587244027303754, + "grad_norm": 0.4750804007053375, + "learning_rate": 5.678078330351166e-09, + "loss": 0.3914, + "step": 9247 + }, + { + "epoch": 2.9590443686006824, + "grad_norm": 0.4475878179073334, + "learning_rate": 5.589721388640778e-09, + "loss": 0.393, + "step": 9248 + }, + { + "epoch": 2.95936433447099, + "grad_norm": 0.4907412827014923, + "learning_rate": 5.5020569279246035e-09, + "loss": 0.4505, + "step": 9249 + }, + { + "epoch": 2.959684300341297, + "grad_norm": 0.4824308454990387, + "learning_rate": 5.415084960356254e-09, + "loss": 0.4407, + "step": 9250 + }, + { + "epoch": 2.960004266211604, + "grad_norm": 0.4802906811237335, + "learning_rate": 5.3288054979949714e-09, + "loss": 0.4897, + "step": 9251 + }, + { + "epoch": 2.9603242320819114, + "grad_norm": 0.48446446657180786, + "learning_rate": 5.243218552803964e-09, + "loss": 0.4279, + "step": 9252 + }, + { + "epoch": 2.9606441979522184, + "grad_norm": 0.4675389528274536, + "learning_rate": 5.1583241366492955e-09, + "loss": 0.3847, + "step": 9253 + }, + { + "epoch": 2.9609641638225255, + "grad_norm": 0.5105575919151306, + "learning_rate": 5.074122261301551e-09, + "loss": 0.4312, + "step": 9254 + }, + { + "epoch": 2.9612841296928325, + "grad_norm": 0.4578179121017456, + "learning_rate": 4.99061293843639e-09, + "loss": 0.3716, + "step": 9255 + }, + { + "epoch": 2.96160409556314, + "grad_norm": 0.5118416547775269, + "learning_rate": 4.907796179631774e-09, + "loss": 0.484, + "step": 9256 + }, + { + "epoch": 2.961924061433447, + "grad_norm": 0.47899672389030457, + "learning_rate": 4.825671996370184e-09, + "loss": 0.4052, + "step": 9257 + }, + { + "epoch": 2.962244027303754, + "grad_norm": 0.5589344501495361, + "learning_rate": 4.744240400038624e-09, + "loss": 0.4194, + "step": 9258 + }, + { + "epoch": 2.9625639931740615, + "grad_norm": 0.49536803364753723, + "learning_rate": 4.663501401927506e-09, + "loss": 0.4636, + "step": 9259 + }, + { + "epoch": 2.9628839590443685, + "grad_norm": 0.49311965703964233, + "learning_rate": 4.583455013231764e-09, + "loss": 0.4318, + "step": 9260 + }, + { + "epoch": 2.9632039249146755, + "grad_norm": 0.5307648777961731, + "learning_rate": 4.504101245049741e-09, + "loss": 0.3942, + "step": 9261 + }, + { + "epoch": 2.963523890784983, + "grad_norm": 0.5528405904769897, + "learning_rate": 4.425440108383195e-09, + "loss": 0.4615, + "step": 9262 + }, + { + "epoch": 2.96384385665529, + "grad_norm": 0.5247182846069336, + "learning_rate": 4.347471614139509e-09, + "loss": 0.4698, + "step": 9263 + }, + { + "epoch": 2.964163822525597, + "grad_norm": 0.5076631307601929, + "learning_rate": 4.270195773128927e-09, + "loss": 0.4439, + "step": 9264 + }, + { + "epoch": 2.9644837883959045, + "grad_norm": 0.4680469036102295, + "learning_rate": 4.193612596065655e-09, + "loss": 0.4456, + "step": 9265 + }, + { + "epoch": 2.9648037542662116, + "grad_norm": 0.4061358571052551, + "learning_rate": 4.117722093567867e-09, + "loss": 0.3656, + "step": 9266 + }, + { + "epoch": 2.9651237201365186, + "grad_norm": 0.4813680350780487, + "learning_rate": 4.042524276158255e-09, + "loss": 0.4453, + "step": 9267 + }, + { + "epoch": 2.965443686006826, + "grad_norm": 0.5288780927658081, + "learning_rate": 3.968019154262926e-09, + "loss": 0.4411, + "step": 9268 + }, + { + "epoch": 2.965763651877133, + "grad_norm": 0.5130483508110046, + "learning_rate": 3.894206738212502e-09, + "loss": 0.438, + "step": 9269 + }, + { + "epoch": 2.96608361774744, + "grad_norm": 0.5286841988563538, + "learning_rate": 3.8210870382404674e-09, + "loss": 0.4828, + "step": 9270 + }, + { + "epoch": 2.9664035836177476, + "grad_norm": 0.4753878116607666, + "learning_rate": 3.748660064484821e-09, + "loss": 0.4217, + "step": 9271 + }, + { + "epoch": 2.9667235494880546, + "grad_norm": 0.4546852707862854, + "learning_rate": 3.6769258269886422e-09, + "loss": 0.4621, + "step": 9272 + }, + { + "epoch": 2.9670435153583616, + "grad_norm": 0.5022838711738586, + "learning_rate": 3.6058843356973074e-09, + "loss": 0.4346, + "step": 9273 + }, + { + "epoch": 2.967363481228669, + "grad_norm": 0.48522311449050903, + "learning_rate": 3.5355356004607154e-09, + "loss": 0.3815, + "step": 9274 + }, + { + "epoch": 2.967683447098976, + "grad_norm": 0.5357267260551453, + "learning_rate": 3.4658796310332866e-09, + "loss": 0.4844, + "step": 9275 + }, + { + "epoch": 2.968003412969283, + "grad_norm": 0.4758678674697876, + "learning_rate": 3.3969164370722953e-09, + "loss": 0.3583, + "step": 9276 + }, + { + "epoch": 2.9683233788395906, + "grad_norm": 0.46168282628059387, + "learning_rate": 3.3286460281400922e-09, + "loss": 0.4112, + "step": 9277 + }, + { + "epoch": 2.9686433447098977, + "grad_norm": 0.5056720972061157, + "learning_rate": 3.2610684137024397e-09, + "loss": 0.5097, + "step": 9278 + }, + { + "epoch": 2.9689633105802047, + "grad_norm": 0.5224137306213379, + "learning_rate": 3.194183603128509e-09, + "loss": 0.4443, + "step": 9279 + }, + { + "epoch": 2.969283276450512, + "grad_norm": 0.5291537046432495, + "learning_rate": 3.1279916056925484e-09, + "loss": 0.4944, + "step": 9280 + }, + { + "epoch": 2.969603242320819, + "grad_norm": 0.4236122965812683, + "learning_rate": 3.062492430572217e-09, + "loss": 0.3447, + "step": 9281 + }, + { + "epoch": 2.969923208191126, + "grad_norm": 0.4741148054599762, + "learning_rate": 2.9976860868485837e-09, + "loss": 0.4604, + "step": 9282 + }, + { + "epoch": 2.9702431740614337, + "grad_norm": 0.4553922414779663, + "learning_rate": 2.933572583506683e-09, + "loss": 0.4211, + "step": 9283 + }, + { + "epoch": 2.9705631399317407, + "grad_norm": 0.5211372971534729, + "learning_rate": 2.8701519294371815e-09, + "loss": 0.4789, + "step": 9284 + }, + { + "epoch": 2.9708831058020477, + "grad_norm": 0.49699050188064575, + "learning_rate": 2.8074241334330454e-09, + "loss": 0.4508, + "step": 9285 + }, + { + "epoch": 2.971203071672355, + "grad_norm": 0.4980183243751526, + "learning_rate": 2.7453892041906516e-09, + "loss": 0.41, + "step": 9286 + }, + { + "epoch": 2.9715230375426622, + "grad_norm": 0.4982837736606598, + "learning_rate": 2.684047150312563e-09, + "loss": 0.4297, + "step": 9287 + }, + { + "epoch": 2.9718430034129693, + "grad_norm": 0.5232399106025696, + "learning_rate": 2.6233979803025334e-09, + "loss": 0.4717, + "step": 9288 + }, + { + "epoch": 2.9721629692832767, + "grad_norm": 0.48403775691986084, + "learning_rate": 2.5634417025710565e-09, + "loss": 0.4464, + "step": 9289 + }, + { + "epoch": 2.9724829351535837, + "grad_norm": 0.5008255243301392, + "learning_rate": 2.504178325429818e-09, + "loss": 0.3797, + "step": 9290 + }, + { + "epoch": 2.9728029010238908, + "grad_norm": 0.5325677394866943, + "learning_rate": 2.445607857096688e-09, + "loss": 0.4651, + "step": 9291 + }, + { + "epoch": 2.973122866894198, + "grad_norm": 0.48184674978256226, + "learning_rate": 2.3877303056923928e-09, + "loss": 0.4226, + "step": 9292 + }, + { + "epoch": 2.9734428327645053, + "grad_norm": 0.46005287766456604, + "learning_rate": 2.330545679241625e-09, + "loss": 0.4657, + "step": 9293 + }, + { + "epoch": 2.9737627986348123, + "grad_norm": 0.45725423097610474, + "learning_rate": 2.2740539856724866e-09, + "loss": 0.4548, + "step": 9294 + }, + { + "epoch": 2.9740827645051193, + "grad_norm": 0.49043959379196167, + "learning_rate": 2.218255232818711e-09, + "loss": 0.4154, + "step": 9295 + }, + { + "epoch": 2.9744027303754264, + "grad_norm": 0.428393691778183, + "learning_rate": 2.1631494284163334e-09, + "loss": 0.3719, + "step": 9296 + }, + { + "epoch": 2.974722696245734, + "grad_norm": 0.5096278786659241, + "learning_rate": 2.1087365801053526e-09, + "loss": 0.5253, + "step": 9297 + }, + { + "epoch": 2.975042662116041, + "grad_norm": 0.5088695287704468, + "learning_rate": 2.055016695430845e-09, + "loss": 0.4528, + "step": 9298 + }, + { + "epoch": 2.975362627986348, + "grad_norm": 0.4937836825847626, + "learning_rate": 2.001989781841296e-09, + "loss": 0.4428, + "step": 9299 + }, + { + "epoch": 2.9756825938566553, + "grad_norm": 0.4787329435348511, + "learning_rate": 1.949655846688048e-09, + "loss": 0.4407, + "step": 9300 + }, + { + "epoch": 2.9760025597269624, + "grad_norm": 0.47262081503868103, + "learning_rate": 1.898014897228628e-09, + "loss": 0.4247, + "step": 9301 + }, + { + "epoch": 2.9763225255972694, + "grad_norm": 0.501465916633606, + "learning_rate": 1.8470669406217556e-09, + "loss": 0.419, + "step": 9302 + }, + { + "epoch": 2.976642491467577, + "grad_norm": 0.49505382776260376, + "learning_rate": 1.7968119839317787e-09, + "loss": 0.4641, + "step": 9303 + }, + { + "epoch": 2.976962457337884, + "grad_norm": 0.5140721797943115, + "learning_rate": 1.7472500341270126e-09, + "loss": 0.4348, + "step": 9304 + }, + { + "epoch": 2.977282423208191, + "grad_norm": 0.5133354663848877, + "learning_rate": 1.6983810980791826e-09, + "loss": 0.4573, + "step": 9305 + }, + { + "epoch": 2.9776023890784984, + "grad_norm": 0.4683056175708771, + "learning_rate": 1.6502051825639798e-09, + "loss": 0.3955, + "step": 9306 + }, + { + "epoch": 2.9779223549488054, + "grad_norm": 0.5300466418266296, + "learning_rate": 1.602722294261061e-09, + "loss": 0.4995, + "step": 9307 + }, + { + "epoch": 2.9782423208191124, + "grad_norm": 0.5027908086776733, + "learning_rate": 1.5559324397534937e-09, + "loss": 0.4312, + "step": 9308 + }, + { + "epoch": 2.97856228668942, + "grad_norm": 0.5182973146438599, + "learning_rate": 1.5098356255294211e-09, + "loss": 0.4583, + "step": 9309 + }, + { + "epoch": 2.978882252559727, + "grad_norm": 0.49473249912261963, + "learning_rate": 1.4644318579798422e-09, + "loss": 0.3811, + "step": 9310 + }, + { + "epoch": 2.979202218430034, + "grad_norm": 0.5003598928451538, + "learning_rate": 1.4197211434002767e-09, + "loss": 0.4496, + "step": 9311 + }, + { + "epoch": 2.9795221843003414, + "grad_norm": 0.5183418989181519, + "learning_rate": 1.3757034879896547e-09, + "loss": 0.4161, + "step": 9312 + }, + { + "epoch": 2.9798421501706485, + "grad_norm": 0.49360767006874084, + "learning_rate": 1.3323788978514275e-09, + "loss": 0.4789, + "step": 9313 + }, + { + "epoch": 2.9801621160409555, + "grad_norm": 0.5218793153762817, + "learning_rate": 1.2897473789924563e-09, + "loss": 0.4396, + "step": 9314 + }, + { + "epoch": 2.980482081911263, + "grad_norm": 0.5581367611885071, + "learning_rate": 1.2478089373235691e-09, + "loss": 0.4587, + "step": 9315 + }, + { + "epoch": 2.98080204778157, + "grad_norm": 0.5236092805862427, + "learning_rate": 1.2065635786595586e-09, + "loss": 0.4835, + "step": 9316 + }, + { + "epoch": 2.981122013651877, + "grad_norm": 0.5084599256515503, + "learning_rate": 1.1660113087191839e-09, + "loss": 0.4404, + "step": 9317 + }, + { + "epoch": 2.9814419795221845, + "grad_norm": 0.49674293398857117, + "learning_rate": 1.126152133125169e-09, + "loss": 0.4192, + "step": 9318 + }, + { + "epoch": 2.9817619453924915, + "grad_norm": 0.5019124746322632, + "learning_rate": 1.0869860574036494e-09, + "loss": 0.4674, + "step": 9319 + }, + { + "epoch": 2.9820819112627985, + "grad_norm": 0.513853132724762, + "learning_rate": 1.0485130869858362e-09, + "loss": 0.4452, + "step": 9320 + }, + { + "epoch": 2.982401877133106, + "grad_norm": 0.5213266015052795, + "learning_rate": 1.0107332272052406e-09, + "loss": 0.4618, + "step": 9321 + }, + { + "epoch": 2.982721843003413, + "grad_norm": 0.46743345260620117, + "learning_rate": 9.7364648330045e-10, + "loss": 0.4253, + "step": 9322 + }, + { + "epoch": 2.98304180887372, + "grad_norm": 0.5130221247673035, + "learning_rate": 9.372528604134623e-10, + "loss": 0.4654, + "step": 9323 + }, + { + "epoch": 2.9833617747440275, + "grad_norm": 0.5008715391159058, + "learning_rate": 9.015523635907963e-10, + "loss": 0.4023, + "step": 9324 + }, + { + "epoch": 2.9836817406143346, + "grad_norm": 0.44110965728759766, + "learning_rate": 8.665449977818263e-10, + "loss": 0.414, + "step": 9325 + }, + { + "epoch": 2.9840017064846416, + "grad_norm": 0.5047010183334351, + "learning_rate": 8.322307678398922e-10, + "loss": 0.4989, + "step": 9326 + }, + { + "epoch": 2.984321672354949, + "grad_norm": 0.481261283159256, + "learning_rate": 7.98609678523965e-10, + "loss": 0.4233, + "step": 9327 + }, + { + "epoch": 2.984641638225256, + "grad_norm": 0.4948277771472931, + "learning_rate": 7.656817344947609e-10, + "loss": 0.4815, + "step": 9328 + }, + { + "epoch": 2.984961604095563, + "grad_norm": 0.44666704535484314, + "learning_rate": 7.334469403180721e-10, + "loss": 0.3975, + "step": 9329 + }, + { + "epoch": 2.9852815699658706, + "grad_norm": 0.5035291910171509, + "learning_rate": 7.019053004631016e-10, + "loss": 0.4343, + "step": 9330 + }, + { + "epoch": 2.9856015358361776, + "grad_norm": 0.5192816853523254, + "learning_rate": 6.710568193035727e-10, + "loss": 0.4249, + "step": 9331 + }, + { + "epoch": 2.9859215017064846, + "grad_norm": 0.5104855298995972, + "learning_rate": 6.409015011160646e-10, + "loss": 0.4427, + "step": 9332 + }, + { + "epoch": 2.9862414675767917, + "grad_norm": 0.5305399298667908, + "learning_rate": 6.114393500822324e-10, + "loss": 0.4319, + "step": 9333 + }, + { + "epoch": 2.986561433447099, + "grad_norm": 0.5200382471084595, + "learning_rate": 5.826703702865866e-10, + "loss": 0.4428, + "step": 9334 + }, + { + "epoch": 2.986881399317406, + "grad_norm": 0.5008326768875122, + "learning_rate": 5.545945657181584e-10, + "loss": 0.4537, + "step": 9335 + }, + { + "epoch": 2.987201365187713, + "grad_norm": 0.5317647457122803, + "learning_rate": 5.272119402693898e-10, + "loss": 0.4286, + "step": 9336 + }, + { + "epoch": 2.98752133105802, + "grad_norm": 0.5395078659057617, + "learning_rate": 5.005224977372436e-10, + "loss": 0.428, + "step": 9337 + }, + { + "epoch": 2.9878412969283277, + "grad_norm": 0.4737189710140228, + "learning_rate": 4.745262418226481e-10, + "loss": 0.4336, + "step": 9338 + }, + { + "epoch": 2.9881612627986347, + "grad_norm": 0.4658548831939697, + "learning_rate": 4.492231761293875e-10, + "loss": 0.3766, + "step": 9339 + }, + { + "epoch": 2.9884812286689417, + "grad_norm": 0.5189458727836609, + "learning_rate": 4.246133041657663e-10, + "loss": 0.4898, + "step": 9340 + }, + { + "epoch": 2.988801194539249, + "grad_norm": 0.5214295387268066, + "learning_rate": 4.006966293440551e-10, + "loss": 0.4769, + "step": 9341 + }, + { + "epoch": 2.989121160409556, + "grad_norm": 0.5181043744087219, + "learning_rate": 3.7747315498049e-10, + "loss": 0.3411, + "step": 9342 + }, + { + "epoch": 2.9894411262798632, + "grad_norm": 0.5489748120307922, + "learning_rate": 3.5494288429527293e-10, + "loss": 0.4787, + "step": 9343 + }, + { + "epoch": 2.9897610921501707, + "grad_norm": 0.540492594242096, + "learning_rate": 3.3310582041146124e-10, + "loss": 0.4249, + "step": 9344 + }, + { + "epoch": 2.9900810580204777, + "grad_norm": 0.5374926328659058, + "learning_rate": 3.1196196635774335e-10, + "loss": 0.4918, + "step": 9345 + }, + { + "epoch": 2.9904010238907848, + "grad_norm": 0.4693905711174011, + "learning_rate": 2.91511325065108e-10, + "loss": 0.4418, + "step": 9346 + }, + { + "epoch": 2.9907209897610922, + "grad_norm": 0.43871355056762695, + "learning_rate": 2.7175389936906496e-10, + "loss": 0.4497, + "step": 9347 + }, + { + "epoch": 2.9910409556313993, + "grad_norm": 0.5393214821815491, + "learning_rate": 2.526896920090893e-10, + "loss": 0.4206, + "step": 9348 + }, + { + "epoch": 2.9913609215017063, + "grad_norm": 0.5345500111579895, + "learning_rate": 2.3431870562917735e-10, + "loss": 0.4384, + "step": 9349 + }, + { + "epoch": 2.9916808873720138, + "grad_norm": 0.4962906539440155, + "learning_rate": 2.166409427750704e-10, + "loss": 0.4434, + "step": 9350 + }, + { + "epoch": 2.992000853242321, + "grad_norm": 0.48774805665016174, + "learning_rate": 1.9965640589925118e-10, + "loss": 0.4356, + "step": 9351 + }, + { + "epoch": 2.992320819112628, + "grad_norm": 0.5116692781448364, + "learning_rate": 1.8336509735539243e-10, + "loss": 0.4268, + "step": 9352 + }, + { + "epoch": 2.9926407849829353, + "grad_norm": 0.5087159276008606, + "learning_rate": 1.6776701940335317e-10, + "loss": 0.4679, + "step": 9353 + }, + { + "epoch": 2.9929607508532423, + "grad_norm": 0.5321253538131714, + "learning_rate": 1.528621742047376e-10, + "loss": 0.4676, + "step": 9354 + }, + { + "epoch": 2.9932807167235493, + "grad_norm": 0.5010133981704712, + "learning_rate": 1.386505638273361e-10, + "loss": 0.4319, + "step": 9355 + }, + { + "epoch": 2.993600682593857, + "grad_norm": 0.48460671305656433, + "learning_rate": 1.2513219024068435e-10, + "loss": 0.3796, + "step": 9356 + }, + { + "epoch": 2.993920648464164, + "grad_norm": 0.4713209569454193, + "learning_rate": 1.1230705531994901e-10, + "loss": 0.4135, + "step": 9357 + }, + { + "epoch": 2.994240614334471, + "grad_norm": 0.4603041410446167, + "learning_rate": 1.0017516084204204e-10, + "loss": 0.455, + "step": 9358 + }, + { + "epoch": 2.9945605802047783, + "grad_norm": 0.4950495660305023, + "learning_rate": 8.873650849061666e-11, + "loss": 0.4907, + "step": 9359 + }, + { + "epoch": 2.9948805460750854, + "grad_norm": 0.47134140133857727, + "learning_rate": 7.799109985051623e-11, + "loss": 0.4352, + "step": 9360 + }, + { + "epoch": 2.9952005119453924, + "grad_norm": 0.5164924263954163, + "learning_rate": 6.793893641221516e-11, + "loss": 0.4315, + "step": 9361 + }, + { + "epoch": 2.9955204778157, + "grad_norm": 0.5457006096839905, + "learning_rate": 5.858001956904335e-11, + "loss": 0.444, + "step": 9362 + }, + { + "epoch": 2.995840443686007, + "grad_norm": 0.511311411857605, + "learning_rate": 4.9914350618851527e-11, + "loss": 0.4658, + "step": 9363 + }, + { + "epoch": 2.996160409556314, + "grad_norm": 0.49430567026138306, + "learning_rate": 4.1941930763456094e-11, + "loss": 0.4161, + "step": 9364 + }, + { + "epoch": 2.9964803754266214, + "grad_norm": 0.5116695761680603, + "learning_rate": 3.466276110752897e-11, + "loss": 0.4619, + "step": 9365 + }, + { + "epoch": 2.9968003412969284, + "grad_norm": 0.49776217341423035, + "learning_rate": 2.8076842660818003e-11, + "loss": 0.4505, + "step": 9366 + }, + { + "epoch": 2.9971203071672354, + "grad_norm": 0.478668212890625, + "learning_rate": 2.2184176336481623e-11, + "loss": 0.427, + "step": 9367 + }, + { + "epoch": 2.997440273037543, + "grad_norm": 0.4620792269706726, + "learning_rate": 1.698476295108886e-11, + "loss": 0.4259, + "step": 9368 + }, + { + "epoch": 2.99776023890785, + "grad_norm": 0.5169647336006165, + "learning_rate": 1.2478603226284691e-11, + "loss": 0.4261, + "step": 9369 + }, + { + "epoch": 2.998080204778157, + "grad_norm": 0.5284416079521179, + "learning_rate": 8.66569778656956e-12, + "loss": 0.5411, + "step": 9370 + }, + { + "epoch": 2.9984001706484644, + "grad_norm": 0.409438818693161, + "learning_rate": 5.5460471598545174e-12, + "loss": 0.3569, + "step": 9371 + }, + { + "epoch": 2.9987201365187715, + "grad_norm": 0.5102793574333191, + "learning_rate": 3.119651779681654e-12, + "loss": 0.4702, + "step": 9372 + }, + { + "epoch": 2.9990401023890785, + "grad_norm": 0.5210458040237427, + "learning_rate": 1.3865119824485462e-12, + "loss": 0.4577, + "step": 9373 + }, + { + "epoch": 2.9993600682593855, + "grad_norm": 0.508808434009552, + "learning_rate": 3.466280074082562e-13, + "loss": 0.4223, + "step": 9374 + }, + { + "epoch": 2.999680034129693, + "grad_norm": 0.5104720592498779, + "learning_rate": 0.0, + "loss": 0.4301, + "step": 9375 + }, + { + "epoch": 2.999680034129693, + "step": 9375, + "total_flos": 5611685260214272.0, + "train_loss": 0.5075681580543518, + "train_runtime": 141329.7996, + "train_samples_per_second": 6.368, + "train_steps_per_second": 0.066 + } + ], + "logging_steps": 1.0, + "max_steps": 9375, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 5611685260214272.0, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}