diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,21033 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.07790039877585088, + "eval_steps": 500, + "global_step": 3000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 2.596679959195029e-05, + "grad_norm": 10.167712131798243, + "learning_rate": 0.0, + "loss": 2.0279, + "step": 1 + }, + { + "epoch": 5.193359918390058e-05, + "grad_norm": 12.71908591386877, + "learning_rate": 9.82808161164113e-07, + "loss": 2.051, + "step": 2 + }, + { + "epoch": 7.790039877585088e-05, + "grad_norm": 31.52939070180938, + "learning_rate": 1.5577140808478338e-06, + "loss": 2.0822, + "step": 3 + }, + { + "epoch": 0.00010386719836780116, + "grad_norm": 8.52302663566333, + "learning_rate": 1.965616322328226e-06, + "loss": 2.0849, + "step": 4 + }, + { + "epoch": 0.00012983399795975147, + "grad_norm": 12.410036369227404, + "learning_rate": 2.2820098812915408e-06, + "loss": 2.0754, + "step": 5 + }, + { + "epoch": 0.00015580079755170176, + "grad_norm": 8.92003156921548, + "learning_rate": 2.540522242011947e-06, + "loss": 2.0222, + "step": 6 + }, + { + "epoch": 0.00018176759714365204, + "grad_norm": 13.189865984810673, + "learning_rate": 2.759091328682456e-06, + "loss": 2.0348, + "step": 7 + }, + { + "epoch": 0.00020773439673560232, + "grad_norm": 12.068025420458316, + "learning_rate": 2.948424483492339e-06, + "loss": 2.0539, + "step": 8 + }, + { + "epoch": 0.00023370119632755264, + "grad_norm": 8.673342798781551, + "learning_rate": 3.1154281616956676e-06, + "loss": 2.0554, + "step": 9 + }, + { + "epoch": 0.00025966799591950295, + "grad_norm": 58.430923760988925, + "learning_rate": 3.2648180424556543e-06, + "loss": 2.065, + "step": 10 + }, + { + "epoch": 0.0002856347955114532, + "grad_norm": 8.40080362744625, + "learning_rate": 3.3999576277859138e-06, + "loss": 2.039, + "step": 11 + }, + { + "epoch": 0.0003116015951034035, + "grad_norm": 11.594089554495266, + "learning_rate": 3.52333040317606e-06, + "loss": 2.0788, + "step": 12 + }, + { + "epoch": 0.0003375683946953538, + "grad_norm": 11.471317513183969, + "learning_rate": 3.636822354884896e-06, + "loss": 2.0432, + "step": 13 + }, + { + "epoch": 0.0003635351942873041, + "grad_norm": 11.566295433866307, + "learning_rate": 3.7418994898465686e-06, + "loss": 2.0643, + "step": 14 + }, + { + "epoch": 0.0003895019938792544, + "grad_norm": 10.147984542311645, + "learning_rate": 3.839723962139374e-06, + "loss": 2.0618, + "step": 15 + }, + { + "epoch": 0.00041546879347120465, + "grad_norm": 7.1649375722205955, + "learning_rate": 3.931232644656452e-06, + "loss": 2.0061, + "step": 16 + }, + { + "epoch": 0.00044143559306315496, + "grad_norm": 8.161162201638032, + "learning_rate": 4.017191838835887e-06, + "loss": 1.9924, + "step": 17 + }, + { + "epoch": 0.00046740239265510527, + "grad_norm": 14.897103436436844, + "learning_rate": 4.09823632285978e-06, + "loss": 1.9987, + "step": 18 + }, + { + "epoch": 0.0004933691922470555, + "grad_norm": 9.367730335735887, + "learning_rate": 4.174897828245933e-06, + "loss": 1.9955, + "step": 19 + }, + { + "epoch": 0.0005193359918390059, + "grad_norm": 13.833683116734269, + "learning_rate": 4.2476262036197665e-06, + "loss": 1.9203, + "step": 20 + }, + { + "epoch": 0.0005453027914309561, + "grad_norm": 9.656884033720331, + "learning_rate": 4.3168054095302896e-06, + "loss": 1.9405, + "step": 21 + }, + { + "epoch": 0.0005712695910229064, + "grad_norm": 7.4336704685183275, + "learning_rate": 4.382765788950027e-06, + "loss": 1.9728, + "step": 22 + }, + { + "epoch": 0.0005972363906148568, + "grad_norm": 10.20043087217532, + "learning_rate": 4.445793607944332e-06, + "loss": 1.9714, + "step": 23 + }, + { + "epoch": 0.000623203190206807, + "grad_norm": 7.451984822133405, + "learning_rate": 4.506138564340173e-06, + "loss": 1.934, + "step": 24 + }, + { + "epoch": 0.0006491699897987573, + "grad_norm": 9.163112529353, + "learning_rate": 4.5640197625830816e-06, + "loss": 1.8725, + "step": 25 + }, + { + "epoch": 0.0006751367893907076, + "grad_norm": 7.889725254240628, + "learning_rate": 4.619630516049009e-06, + "loss": 1.9659, + "step": 26 + }, + { + "epoch": 0.0007011035889826579, + "grad_norm": 14.977112665951129, + "learning_rate": 4.673142242543501e-06, + "loss": 1.8455, + "step": 27 + }, + { + "epoch": 0.0007270703885746082, + "grad_norm": 11.452971786577518, + "learning_rate": 4.724707651010682e-06, + "loss": 1.8676, + "step": 28 + }, + { + "epoch": 0.0007530371881665584, + "grad_norm": 8.541682198363038, + "learning_rate": 4.774463368791538e-06, + "loss": 1.8726, + "step": 29 + }, + { + "epoch": 0.0007790039877585088, + "grad_norm": 8.082582524093958, + "learning_rate": 4.822532123303487e-06, + "loss": 1.8402, + "step": 30 + }, + { + "epoch": 0.000804970787350459, + "grad_norm": 17.061179979845694, + "learning_rate": 4.869024565857359e-06, + "loss": 1.855, + "step": 31 + }, + { + "epoch": 0.0008309375869424093, + "grad_norm": 10.305135790926464, + "learning_rate": 4.914040805820566e-06, + "loss": 1.8031, + "step": 32 + }, + { + "epoch": 0.0008569043865343597, + "grad_norm": 17.521505166809924, + "learning_rate": 4.957671708633747e-06, + "loss": 1.7963, + "step": 33 + }, + { + "epoch": 0.0008828711861263099, + "grad_norm": 27.768773782133074, + "learning_rate": 5e-06, + "loss": 1.797, + "step": 34 + }, + { + "epoch": 0.0009088379857182602, + "grad_norm": 7.709614772647655, + "learning_rate": 5.041101209973997e-06, + "loss": 1.7693, + "step": 35 + }, + { + "epoch": 0.0009348047853102105, + "grad_norm": 40.43078274482668, + "learning_rate": 5.081044484023894e-06, + "loss": 1.7098, + "step": 36 + }, + { + "epoch": 0.0009607715849021608, + "grad_norm": 13.228036202417233, + "learning_rate": 5.1198932829439884e-06, + "loss": 1.7796, + "step": 37 + }, + { + "epoch": 0.000986738384494111, + "grad_norm": 12.98342308108088, + "learning_rate": 5.157705989410047e-06, + "loss": 1.7911, + "step": 38 + }, + { + "epoch": 0.0010127051840860613, + "grad_norm": 11.634216592772333, + "learning_rate": 5.19453643573273e-06, + "loss": 1.6965, + "step": 39 + }, + { + "epoch": 0.0010386719836780118, + "grad_norm": 11.958818343868286, + "learning_rate": 5.23043436478388e-06, + "loss": 1.755, + "step": 40 + }, + { + "epoch": 0.001064638783269962, + "grad_norm": 6.580766233685262, + "learning_rate": 5.265445833999807e-06, + "loss": 1.74, + "step": 41 + }, + { + "epoch": 0.0010906055828619123, + "grad_norm": 21.555133478985347, + "learning_rate": 5.299613570694403e-06, + "loss": 1.6821, + "step": 42 + }, + { + "epoch": 0.0011165723824538626, + "grad_norm": 12.351837985147316, + "learning_rate": 5.332977285558406e-06, + "loss": 1.6646, + "step": 43 + }, + { + "epoch": 0.0011425391820458128, + "grad_norm": 7.967517952014531, + "learning_rate": 5.365573950114139e-06, + "loss": 1.6549, + "step": 44 + }, + { + "epoch": 0.001168505981637763, + "grad_norm": 43.7458996404413, + "learning_rate": 5.397438042987208e-06, + "loss": 1.641, + "step": 45 + }, + { + "epoch": 0.0011944727812297135, + "grad_norm": 10.5030422344766, + "learning_rate": 5.428601769108445e-06, + "loss": 1.6464, + "step": 46 + }, + { + "epoch": 0.0012204395808216638, + "grad_norm": 10.938006405687435, + "learning_rate": 5.4590952553399815e-06, + "loss": 1.6533, + "step": 47 + }, + { + "epoch": 0.001246406380413614, + "grad_norm": 7.8619150110899545, + "learning_rate": 5.488946725504286e-06, + "loss": 1.5973, + "step": 48 + }, + { + "epoch": 0.0012723731800055643, + "grad_norm": 14.23050306233854, + "learning_rate": 5.518182657364912e-06, + "loss": 1.5932, + "step": 49 + }, + { + "epoch": 0.0012983399795975146, + "grad_norm": 9.252604349903836, + "learning_rate": 5.546827923747195e-06, + "loss": 1.6129, + "step": 50 + }, + { + "epoch": 0.0013243067791894648, + "grad_norm": 13.090581615282309, + "learning_rate": 5.5749059196837205e-06, + "loss": 1.5941, + "step": 51 + }, + { + "epoch": 0.0013502735787814153, + "grad_norm": 7.915832020446602, + "learning_rate": 5.602438677213122e-06, + "loss": 1.6106, + "step": 52 + }, + { + "epoch": 0.0013762403783733656, + "grad_norm": 9.105486764648631, + "learning_rate": 5.62944696924357e-06, + "loss": 1.5117, + "step": 53 + }, + { + "epoch": 0.0014022071779653158, + "grad_norm": 7.126014771057802, + "learning_rate": 5.655950403707615e-06, + "loss": 1.5508, + "step": 54 + }, + { + "epoch": 0.001428173977557266, + "grad_norm": 9.865708241007317, + "learning_rate": 5.681967509077455e-06, + "loss": 1.5662, + "step": 55 + }, + { + "epoch": 0.0014541407771492163, + "grad_norm": 9.65133077507866, + "learning_rate": 5.7075158121747964e-06, + "loss": 1.4955, + "step": 56 + }, + { + "epoch": 0.0014801075767411666, + "grad_norm": 8.443505078765376, + "learning_rate": 5.732611909093767e-06, + "loss": 1.528, + "step": 57 + }, + { + "epoch": 0.0015060743763331168, + "grad_norm": 23.391536962432454, + "learning_rate": 5.75727152995565e-06, + "loss": 1.532, + "step": 58 + }, + { + "epoch": 0.0015320411759250673, + "grad_norm": 10.05259226299638, + "learning_rate": 5.781509598128162e-06, + "loss": 1.5033, + "step": 59 + }, + { + "epoch": 0.0015580079755170176, + "grad_norm": 7.054507704225335, + "learning_rate": 5.805340284467601e-06, + "loss": 1.4956, + "step": 60 + }, + { + "epoch": 0.0015839747751089678, + "grad_norm": 7.676786819867121, + "learning_rate": 5.828777057077528e-06, + "loss": 1.5122, + "step": 61 + }, + { + "epoch": 0.001609941574700918, + "grad_norm": 8.308300821998932, + "learning_rate": 5.851832727021472e-06, + "loss": 1.4783, + "step": 62 + }, + { + "epoch": 0.0016359083742928683, + "grad_norm": 7.94511036723585, + "learning_rate": 5.8745194903781236e-06, + "loss": 1.4798, + "step": 63 + }, + { + "epoch": 0.0016618751738848186, + "grad_norm": 10.036220566856906, + "learning_rate": 5.896848966984678e-06, + "loss": 1.4794, + "step": 64 + }, + { + "epoch": 0.001687841973476769, + "grad_norm": 11.66207388439793, + "learning_rate": 5.918832236176436e-06, + "loss": 1.5142, + "step": 65 + }, + { + "epoch": 0.0017138087730687193, + "grad_norm": 8.391434671275535, + "learning_rate": 5.940479869797859e-06, + "loss": 1.5192, + "step": 66 + }, + { + "epoch": 0.0017397755726606696, + "grad_norm": 6.706762220689152, + "learning_rate": 5.961801962731307e-06, + "loss": 1.4689, + "step": 67 + }, + { + "epoch": 0.0017657423722526198, + "grad_norm": 6.743265253698558, + "learning_rate": 5.982808161164114e-06, + "loss": 1.4538, + "step": 68 + }, + { + "epoch": 0.00179170917184457, + "grad_norm": 8.66399777717597, + "learning_rate": 6.0035076887921665e-06, + "loss": 1.4629, + "step": 69 + }, + { + "epoch": 0.0018176759714365204, + "grad_norm": 11.238826713278984, + "learning_rate": 6.02390937113811e-06, + "loss": 1.4716, + "step": 70 + }, + { + "epoch": 0.0018436427710284708, + "grad_norm": 7.419337302584475, + "learning_rate": 6.0440216581446974e-06, + "loss": 1.473, + "step": 71 + }, + { + "epoch": 0.001869609570620421, + "grad_norm": 7.494123063652215, + "learning_rate": 6.063852645188006e-06, + "loss": 1.4674, + "step": 72 + }, + { + "epoch": 0.0018955763702123713, + "grad_norm": 18.654266507572654, + "learning_rate": 6.083410092641338e-06, + "loss": 1.4646, + "step": 73 + }, + { + "epoch": 0.0019215431698043216, + "grad_norm": 6.675700678821861, + "learning_rate": 6.102701444108102e-06, + "loss": 1.4292, + "step": 74 + }, + { + "epoch": 0.0019475099693962719, + "grad_norm": 16.474908419225734, + "learning_rate": 6.121733843430915e-06, + "loss": 1.4084, + "step": 75 + }, + { + "epoch": 0.001973476768988222, + "grad_norm": 8.395304648237534, + "learning_rate": 6.140514150574159e-06, + "loss": 1.3988, + "step": 76 + }, + { + "epoch": 0.0019994435685801726, + "grad_norm": 17.2232361088312, + "learning_rate": 6.159048956468369e-06, + "loss": 1.496, + "step": 77 + }, + { + "epoch": 0.0020254103681721226, + "grad_norm": 17.05890100550516, + "learning_rate": 6.177344596896842e-06, + "loss": 1.4618, + "step": 78 + }, + { + "epoch": 0.002051377167764073, + "grad_norm": 8.844316336355204, + "learning_rate": 6.1954071654976755e-06, + "loss": 1.4066, + "step": 79 + }, + { + "epoch": 0.0020773439673560236, + "grad_norm": 7.807838941601465, + "learning_rate": 6.2132425259479926e-06, + "loss": 1.4033, + "step": 80 + }, + { + "epoch": 0.0021033107669479736, + "grad_norm": 7.118514472206657, + "learning_rate": 6.230856323391335e-06, + "loss": 1.4293, + "step": 81 + }, + { + "epoch": 0.002129277566539924, + "grad_norm": 7.4210338236309505, + "learning_rate": 6.2482539951639196e-06, + "loss": 1.4293, + "step": 82 + }, + { + "epoch": 0.002155244366131874, + "grad_norm": 8.614845379522835, + "learning_rate": 6.265440780870785e-06, + "loss": 1.4266, + "step": 83 + }, + { + "epoch": 0.0021812111657238246, + "grad_norm": 13.527631949897033, + "learning_rate": 6.282421731858516e-06, + "loss": 1.4275, + "step": 84 + }, + { + "epoch": 0.0022071779653157746, + "grad_norm": 10.933003400461018, + "learning_rate": 6.299201720127428e-06, + "loss": 1.4149, + "step": 85 + }, + { + "epoch": 0.002233144764907725, + "grad_norm": 8.947921545688304, + "learning_rate": 6.315785446722519e-06, + "loss": 1.4006, + "step": 86 + }, + { + "epoch": 0.0022591115644996756, + "grad_norm": 6.766296776243929, + "learning_rate": 6.332177449639372e-06, + "loss": 1.4137, + "step": 87 + }, + { + "epoch": 0.0022850783640916256, + "grad_norm": 23.311286465147226, + "learning_rate": 6.348382111278253e-06, + "loss": 1.3642, + "step": 88 + }, + { + "epoch": 0.002311045163683576, + "grad_norm": 7.53677507397063, + "learning_rate": 6.364403665477058e-06, + "loss": 1.3805, + "step": 89 + }, + { + "epoch": 0.002337011963275526, + "grad_norm": 7.973492576174175, + "learning_rate": 6.380246204151321e-06, + "loss": 1.3802, + "step": 90 + }, + { + "epoch": 0.0023629787628674766, + "grad_norm": 5.992278216494705, + "learning_rate": 6.395913683567352e-06, + "loss": 1.404, + "step": 91 + }, + { + "epoch": 0.002388945562459427, + "grad_norm": 9.849752479192926, + "learning_rate": 6.411409930272558e-06, + "loss": 1.3895, + "step": 92 + }, + { + "epoch": 0.002414912362051377, + "grad_norm": 9.836286543151555, + "learning_rate": 6.426738646705193e-06, + "loss": 1.3441, + "step": 93 + }, + { + "epoch": 0.0024408791616433276, + "grad_norm": 7.4713873256950505, + "learning_rate": 6.441903416504094e-06, + "loss": 1.3697, + "step": 94 + }, + { + "epoch": 0.0024668459612352776, + "grad_norm": 6.267332534043902, + "learning_rate": 6.4569077095374745e-06, + "loss": 1.4325, + "step": 95 + }, + { + "epoch": 0.002492812760827228, + "grad_norm": 8.904193841597902, + "learning_rate": 6.471754886668398e-06, + "loss": 1.4295, + "step": 96 + }, + { + "epoch": 0.002518779560419178, + "grad_norm": 6.633396226338773, + "learning_rate": 6.486448204273347e-06, + "loss": 1.3815, + "step": 97 + }, + { + "epoch": 0.0025447463600111286, + "grad_norm": 7.783640299045919, + "learning_rate": 6.500990818529026e-06, + "loss": 1.386, + "step": 98 + }, + { + "epoch": 0.002570713159603079, + "grad_norm": 8.83256033148245, + "learning_rate": 6.515385789481581e-06, + "loss": 1.3658, + "step": 99 + }, + { + "epoch": 0.002596679959195029, + "grad_norm": 17.54175403902429, + "learning_rate": 6.5296360849113085e-06, + "loss": 1.4168, + "step": 100 + }, + { + "epoch": 0.0026226467587869796, + "grad_norm": 17.514768635677953, + "learning_rate": 6.543744584005075e-06, + "loss": 1.3831, + "step": 101 + }, + { + "epoch": 0.0026486135583789296, + "grad_norm": 11.530537404938231, + "learning_rate": 6.5577140808478336e-06, + "loss": 1.3836, + "step": 102 + }, + { + "epoch": 0.00267458035797088, + "grad_norm": 15.291566825908028, + "learning_rate": 6.571547287743812e-06, + "loss": 1.4025, + "step": 103 + }, + { + "epoch": 0.0027005471575628306, + "grad_norm": 13.018564906569484, + "learning_rate": 6.585246838377235e-06, + "loss": 1.3979, + "step": 104 + }, + { + "epoch": 0.0027265139571547806, + "grad_norm": 10.037675077400776, + "learning_rate": 6.598815290821831e-06, + "loss": 1.3604, + "step": 105 + }, + { + "epoch": 0.002752480756746731, + "grad_norm": 8.77763102069398, + "learning_rate": 6.612255130407681e-06, + "loss": 1.3627, + "step": 106 + }, + { + "epoch": 0.002778447556338681, + "grad_norm": 7.253699838457306, + "learning_rate": 6.625568772453486e-06, + "loss": 1.323, + "step": 107 + }, + { + "epoch": 0.0028044143559306316, + "grad_norm": 6.200582833736466, + "learning_rate": 6.638758564871728e-06, + "loss": 1.4246, + "step": 108 + }, + { + "epoch": 0.0028303811555225817, + "grad_norm": 9.93854330092373, + "learning_rate": 6.6518267906537845e-06, + "loss": 1.3579, + "step": 109 + }, + { + "epoch": 0.002856347955114532, + "grad_norm": 9.143808289848597, + "learning_rate": 6.664775670241568e-06, + "loss": 1.4106, + "step": 110 + }, + { + "epoch": 0.0028823147547064826, + "grad_norm": 8.970037387867794, + "learning_rate": 6.677607363791822e-06, + "loss": 1.3676, + "step": 111 + }, + { + "epoch": 0.0029082815542984326, + "grad_norm": 10.658336804240012, + "learning_rate": 6.690323973338908e-06, + "loss": 1.3303, + "step": 112 + }, + { + "epoch": 0.002934248353890383, + "grad_norm": 6.462273197879613, + "learning_rate": 6.70292754486144e-06, + "loss": 1.3445, + "step": 113 + }, + { + "epoch": 0.002960215153482333, + "grad_norm": 6.922270683879563, + "learning_rate": 6.715420070257881e-06, + "loss": 1.3677, + "step": 114 + }, + { + "epoch": 0.0029861819530742836, + "grad_norm": 12.297885024434843, + "learning_rate": 6.727803489235872e-06, + "loss": 1.3731, + "step": 115 + }, + { + "epoch": 0.0030121487526662337, + "grad_norm": 9.740242795326719, + "learning_rate": 6.740079691119763e-06, + "loss": 1.4205, + "step": 116 + }, + { + "epoch": 0.003038115552258184, + "grad_norm": 9.065516318478803, + "learning_rate": 6.752250516580564e-06, + "loss": 1.3875, + "step": 117 + }, + { + "epoch": 0.0030640823518501346, + "grad_norm": 9.111923253942726, + "learning_rate": 6.764317759292276e-06, + "loss": 1.3462, + "step": 118 + }, + { + "epoch": 0.0030900491514420847, + "grad_norm": 7.087398424715804, + "learning_rate": 6.776283167518344e-06, + "loss": 1.3987, + "step": 119 + }, + { + "epoch": 0.003116015951034035, + "grad_norm": 6.798203893079039, + "learning_rate": 6.7881484456317135e-06, + "loss": 1.3966, + "step": 120 + }, + { + "epoch": 0.003141982750625985, + "grad_norm": 8.597048816428273, + "learning_rate": 6.7999152555718276e-06, + "loss": 1.3295, + "step": 121 + }, + { + "epoch": 0.0031679495502179356, + "grad_norm": 9.601550978303706, + "learning_rate": 6.811585218241642e-06, + "loss": 1.3596, + "step": 122 + }, + { + "epoch": 0.003193916349809886, + "grad_norm": 7.453082006558386, + "learning_rate": 6.8231599148476405e-06, + "loss": 1.3269, + "step": 123 + }, + { + "epoch": 0.003219883149401836, + "grad_norm": 11.568309087610185, + "learning_rate": 6.834640888185586e-06, + "loss": 1.3935, + "step": 124 + }, + { + "epoch": 0.0032458499489937866, + "grad_norm": 7.52277172466284, + "learning_rate": 6.846029643874624e-06, + "loss": 1.3854, + "step": 125 + }, + { + "epoch": 0.0032718167485857367, + "grad_norm": 7.104760049181486, + "learning_rate": 6.8573276515422375e-06, + "loss": 1.4156, + "step": 126 + }, + { + "epoch": 0.003297783548177687, + "grad_norm": 10.389672257189655, + "learning_rate": 6.868536345962348e-06, + "loss": 1.3725, + "step": 127 + }, + { + "epoch": 0.003323750347769637, + "grad_norm": 9.29401873140395, + "learning_rate": 6.879657128148791e-06, + "loss": 1.3771, + "step": 128 + }, + { + "epoch": 0.0033497171473615877, + "grad_norm": 27.153211136070077, + "learning_rate": 6.8906913664062406e-06, + "loss": 1.3813, + "step": 129 + }, + { + "epoch": 0.003375683946953538, + "grad_norm": 7.2659788640305765, + "learning_rate": 6.90164039734055e-06, + "loss": 1.362, + "step": 130 + }, + { + "epoch": 0.003401650746545488, + "grad_norm": 7.806757410935308, + "learning_rate": 6.9125055268303975e-06, + "loss": 1.3908, + "step": 131 + }, + { + "epoch": 0.0034276175461374386, + "grad_norm": 6.417780923910124, + "learning_rate": 6.923288030961973e-06, + "loss": 1.3829, + "step": 132 + }, + { + "epoch": 0.0034535843457293887, + "grad_norm": 19.56526179211872, + "learning_rate": 6.933989156928389e-06, + "loss": 1.3482, + "step": 133 + }, + { + "epoch": 0.003479551145321339, + "grad_norm": 17.573846701579736, + "learning_rate": 6.94461012389542e-06, + "loss": 1.3563, + "step": 134 + }, + { + "epoch": 0.0035055179449132896, + "grad_norm": 16.164281505282915, + "learning_rate": 6.955152123835043e-06, + "loss": 1.3702, + "step": 135 + }, + { + "epoch": 0.0035314847445052397, + "grad_norm": 7.006035246607669, + "learning_rate": 6.9656163223282265e-06, + "loss": 1.3861, + "step": 136 + }, + { + "epoch": 0.00355745154409719, + "grad_norm": 22.67827996837985, + "learning_rate": 6.976003859338315e-06, + "loss": 1.358, + "step": 137 + }, + { + "epoch": 0.00358341834368914, + "grad_norm": 10.845342777475386, + "learning_rate": 6.986315849956279e-06, + "loss": 1.3868, + "step": 138 + }, + { + "epoch": 0.0036093851432810907, + "grad_norm": 11.680227046484237, + "learning_rate": 6.996553385119069e-06, + "loss": 1.3601, + "step": 139 + }, + { + "epoch": 0.0036353519428730407, + "grad_norm": 8.494801446256888, + "learning_rate": 7.006717532302223e-06, + "loss": 1.3897, + "step": 140 + }, + { + "epoch": 0.003661318742464991, + "grad_norm": 6.680234009989669, + "learning_rate": 7.0168093361878155e-06, + "loss": 1.3977, + "step": 141 + }, + { + "epoch": 0.0036872855420569416, + "grad_norm": 7.6763028110811335, + "learning_rate": 7.026829819308811e-06, + "loss": 1.3739, + "step": 142 + }, + { + "epoch": 0.0037132523416488917, + "grad_norm": 14.889423461983792, + "learning_rate": 7.036779982670809e-06, + "loss": 1.3987, + "step": 143 + }, + { + "epoch": 0.003739219141240842, + "grad_norm": 9.744161569492924, + "learning_rate": 7.04666080635212e-06, + "loss": 1.376, + "step": 144 + }, + { + "epoch": 0.003765185940832792, + "grad_norm": 14.935510478942511, + "learning_rate": 7.056473250083078e-06, + "loss": 1.396, + "step": 145 + }, + { + "epoch": 0.0037911527404247427, + "grad_norm": 7.908486272406978, + "learning_rate": 7.066218253805451e-06, + "loss": 1.3675, + "step": 146 + }, + { + "epoch": 0.003817119540016693, + "grad_norm": 6.528226745893618, + "learning_rate": 7.075896738212745e-06, + "loss": 1.3674, + "step": 147 + }, + { + "epoch": 0.003843086339608643, + "grad_norm": 14.668321400554357, + "learning_rate": 7.085509605272214e-06, + "loss": 1.3038, + "step": 148 + }, + { + "epoch": 0.0038690531392005937, + "grad_norm": 7.779697493136657, + "learning_rate": 7.095057738729268e-06, + "loss": 1.3498, + "step": 149 + }, + { + "epoch": 0.0038950199387925437, + "grad_norm": 7.679658938225753, + "learning_rate": 7.104542004595029e-06, + "loss": 1.3304, + "step": 150 + }, + { + "epoch": 0.003920986738384494, + "grad_norm": 7.0238160571298085, + "learning_rate": 7.1139632516176825e-06, + "loss": 1.3435, + "step": 151 + }, + { + "epoch": 0.003946953537976444, + "grad_norm": 11.49497477989533, + "learning_rate": 7.123322311738273e-06, + "loss": 1.3641, + "step": 152 + }, + { + "epoch": 0.003972920337568395, + "grad_norm": 11.703728002030285, + "learning_rate": 7.1326200005315545e-06, + "loss": 1.334, + "step": 153 + }, + { + "epoch": 0.003998887137160345, + "grad_norm": 23.317689865090223, + "learning_rate": 7.141857117632483e-06, + "loss": 1.3823, + "step": 154 + }, + { + "epoch": 0.004024853936752296, + "grad_norm": 8.019241558180987, + "learning_rate": 7.1510344471489e-06, + "loss": 1.298, + "step": 155 + }, + { + "epoch": 0.004050820736344245, + "grad_norm": 9.57004742936486, + "learning_rate": 7.160152758060956e-06, + "loss": 1.401, + "step": 156 + }, + { + "epoch": 0.004076787535936196, + "grad_norm": 12.620595635868442, + "learning_rate": 7.169212804607765e-06, + "loss": 1.3685, + "step": 157 + }, + { + "epoch": 0.004102754335528146, + "grad_norm": 14.83889209159024, + "learning_rate": 7.1782153266617894e-06, + "loss": 1.3579, + "step": 158 + }, + { + "epoch": 0.004128721135120097, + "grad_norm": 9.262507462595046, + "learning_rate": 7.187161050091403e-06, + "loss": 1.3487, + "step": 159 + }, + { + "epoch": 0.004154687934712047, + "grad_norm": 10.32604682547935, + "learning_rate": 7.1960506871121065e-06, + "loss": 1.3609, + "step": 160 + }, + { + "epoch": 0.004180654734303997, + "grad_norm": 10.239856818509438, + "learning_rate": 7.204884936626788e-06, + "loss": 1.351, + "step": 161 + }, + { + "epoch": 0.004206621533895947, + "grad_norm": 10.286713513778382, + "learning_rate": 7.213664484555448e-06, + "loss": 1.3406, + "step": 162 + }, + { + "epoch": 0.004232588333487898, + "grad_norm": 7.978487893923655, + "learning_rate": 7.222390004154792e-06, + "loss": 1.3988, + "step": 163 + }, + { + "epoch": 0.004258555133079848, + "grad_norm": 9.492382810490653, + "learning_rate": 7.2310621563280335e-06, + "loss": 1.3728, + "step": 164 + }, + { + "epoch": 0.004284521932671798, + "grad_norm": 22.24914549140385, + "learning_rate": 7.239681589925288e-06, + "loss": 1.3557, + "step": 165 + }, + { + "epoch": 0.004310488732263748, + "grad_norm": 16.38696317979202, + "learning_rate": 7.248248942034898e-06, + "loss": 1.3206, + "step": 166 + }, + { + "epoch": 0.004336455531855699, + "grad_norm": 7.292030834005163, + "learning_rate": 7.256764838265993e-06, + "loss": 1.3565, + "step": 167 + }, + { + "epoch": 0.004362422331447649, + "grad_norm": 8.101374624900906, + "learning_rate": 7.265229893022629e-06, + "loss": 1.3202, + "step": 168 + }, + { + "epoch": 0.0043883891310396, + "grad_norm": 18.62859376376098, + "learning_rate": 7.273644709769792e-06, + "loss": 1.2802, + "step": 169 + }, + { + "epoch": 0.004414355930631549, + "grad_norm": 9.860671065516012, + "learning_rate": 7.282009881291542e-06, + "loss": 1.3646, + "step": 170 + }, + { + "epoch": 0.0044403227302235, + "grad_norm": 7.165928322069725, + "learning_rate": 7.290325989941602e-06, + "loss": 1.3744, + "step": 171 + }, + { + "epoch": 0.00446628952981545, + "grad_norm": 10.13862992279656, + "learning_rate": 7.298593607886632e-06, + "loss": 1.3551, + "step": 172 + }, + { + "epoch": 0.004492256329407401, + "grad_norm": 23.03111484441347, + "learning_rate": 7.306813297342459e-06, + "loss": 1.4089, + "step": 173 + }, + { + "epoch": 0.004518223128999351, + "grad_norm": 7.336396126837946, + "learning_rate": 7.314985610803485e-06, + "loss": 1.3533, + "step": 174 + }, + { + "epoch": 0.004544189928591301, + "grad_norm": 10.975889311540966, + "learning_rate": 7.323111091265538e-06, + "loss": 1.3359, + "step": 175 + }, + { + "epoch": 0.004570156728183251, + "grad_norm": 18.203166912505395, + "learning_rate": 7.331190272442366e-06, + "loss": 1.3938, + "step": 176 + }, + { + "epoch": 0.004596123527775202, + "grad_norm": 10.626390433492434, + "learning_rate": 7.339223678975995e-06, + "loss": 1.3613, + "step": 177 + }, + { + "epoch": 0.004622090327367152, + "grad_norm": 8.335399778374882, + "learning_rate": 7.347211826641172e-06, + "loss": 1.3547, + "step": 178 + }, + { + "epoch": 0.004648057126959103, + "grad_norm": 7.513147401762808, + "learning_rate": 7.355155222544062e-06, + "loss": 1.3884, + "step": 179 + }, + { + "epoch": 0.004674023926551052, + "grad_norm": 23.357641292268386, + "learning_rate": 7.363054365315435e-06, + "loss": 1.3547, + "step": 180 + }, + { + "epoch": 0.004699990726143003, + "grad_norm": 7.548544878660804, + "learning_rate": 7.370909745298483e-06, + "loss": 1.3339, + "step": 181 + }, + { + "epoch": 0.004725957525734953, + "grad_norm": 7.7003669056855815, + "learning_rate": 7.378721844731465e-06, + "loss": 1.2954, + "step": 182 + }, + { + "epoch": 0.004751924325326904, + "grad_norm": 8.491756256299768, + "learning_rate": 7.386491137925362e-06, + "loss": 1.2771, + "step": 183 + }, + { + "epoch": 0.004777891124918854, + "grad_norm": 11.947803783908952, + "learning_rate": 7.394218091436671e-06, + "loss": 1.3229, + "step": 184 + }, + { + "epoch": 0.004803857924510804, + "grad_norm": 16.63183820244135, + "learning_rate": 7.401903164235529e-06, + "loss": 1.3632, + "step": 185 + }, + { + "epoch": 0.004829824724102754, + "grad_norm": 16.49252890012579, + "learning_rate": 7.409546807869306e-06, + "loss": 1.3492, + "step": 186 + }, + { + "epoch": 0.004855791523694705, + "grad_norm": 12.450384319455702, + "learning_rate": 7.417149466621801e-06, + "loss": 1.3701, + "step": 187 + }, + { + "epoch": 0.004881758323286655, + "grad_norm": 7.413463057794066, + "learning_rate": 7.4247115776682076e-06, + "loss": 1.3011, + "step": 188 + }, + { + "epoch": 0.004907725122878605, + "grad_norm": 8.050823420710756, + "learning_rate": 7.4322335712259576e-06, + "loss": 1.4044, + "step": 189 + }, + { + "epoch": 0.004933691922470555, + "grad_norm": 10.789497383297318, + "learning_rate": 7.439715870701588e-06, + "loss": 1.3529, + "step": 190 + }, + { + "epoch": 0.004959658722062506, + "grad_norm": 29.81939553541715, + "learning_rate": 7.447158892833755e-06, + "loss": 1.3545, + "step": 191 + }, + { + "epoch": 0.004985625521654456, + "grad_norm": 8.831806218799514, + "learning_rate": 7.454563047832512e-06, + "loss": 1.382, + "step": 192 + }, + { + "epoch": 0.005011592321246407, + "grad_norm": 26.37743435152005, + "learning_rate": 7.461928739514972e-06, + "loss": 1.3224, + "step": 193 + }, + { + "epoch": 0.005037559120838356, + "grad_norm": 9.06957521160958, + "learning_rate": 7.469256365437459e-06, + "loss": 1.4097, + "step": 194 + }, + { + "epoch": 0.005063525920430307, + "grad_norm": 8.583746841494522, + "learning_rate": 7.4765463170242706e-06, + "loss": 1.3359, + "step": 195 + }, + { + "epoch": 0.005089492720022257, + "grad_norm": 9.050779142614127, + "learning_rate": 7.483798979693137e-06, + "loss": 1.3384, + "step": 196 + }, + { + "epoch": 0.005115459519614208, + "grad_norm": 7.649394551172773, + "learning_rate": 7.491014732977504e-06, + "loss": 1.3715, + "step": 197 + }, + { + "epoch": 0.005141426319206158, + "grad_norm": 9.36104389270932, + "learning_rate": 7.498193950645695e-06, + "loss": 1.3808, + "step": 198 + }, + { + "epoch": 0.005167393118798108, + "grad_norm": 10.646616051572883, + "learning_rate": 7.505337000817097e-06, + "loss": 1.3675, + "step": 199 + }, + { + "epoch": 0.005193359918390058, + "grad_norm": 10.842199209173675, + "learning_rate": 7.512444246075421e-06, + "loss": 1.3477, + "step": 200 + }, + { + "epoch": 0.005219326717982009, + "grad_norm": 9.077553919887066, + "learning_rate": 7.519516043579141e-06, + "loss": 1.3332, + "step": 201 + }, + { + "epoch": 0.005245293517573959, + "grad_norm": 12.194503571625965, + "learning_rate": 7.526552745169188e-06, + "loss": 1.3095, + "step": 202 + }, + { + "epoch": 0.00527126031716591, + "grad_norm": 6.785703260369773, + "learning_rate": 7.533554697473993e-06, + "loss": 1.3439, + "step": 203 + }, + { + "epoch": 0.005297227116757859, + "grad_norm": 9.0612556392459, + "learning_rate": 7.5405222420119475e-06, + "loss": 1.3357, + "step": 204 + }, + { + "epoch": 0.00532319391634981, + "grad_norm": 10.883029122260794, + "learning_rate": 7.5474557152913486e-06, + "loss": 1.3381, + "step": 205 + }, + { + "epoch": 0.00534916071594176, + "grad_norm": 7.539056175110689, + "learning_rate": 7.5543554489079255e-06, + "loss": 1.3359, + "step": 206 + }, + { + "epoch": 0.005375127515533711, + "grad_norm": 9.94351378998889, + "learning_rate": 7.561221769639999e-06, + "loss": 1.3504, + "step": 207 + }, + { + "epoch": 0.005401094315125661, + "grad_norm": 10.110755979382025, + "learning_rate": 7.568054999541348e-06, + "loss": 1.3391, + "step": 208 + }, + { + "epoch": 0.005427061114717611, + "grad_norm": 8.226666369239807, + "learning_rate": 7.574855456031846e-06, + "loss": 1.3481, + "step": 209 + }, + { + "epoch": 0.005453027914309561, + "grad_norm": 17.50960342664, + "learning_rate": 7.581623451985943e-06, + "loss": 1.3117, + "step": 210 + }, + { + "epoch": 0.005478994713901512, + "grad_norm": 12.655596530197865, + "learning_rate": 7.588359295819035e-06, + "loss": 1.3926, + "step": 211 + }, + { + "epoch": 0.005504961513493462, + "grad_norm": 22.713369537823286, + "learning_rate": 7.595063291571795e-06, + "loss": 1.3602, + "step": 212 + }, + { + "epoch": 0.005530928313085412, + "grad_norm": 18.895159130693205, + "learning_rate": 7.6017357389925314e-06, + "loss": 1.3178, + "step": 213 + }, + { + "epoch": 0.005556895112677362, + "grad_norm": 15.828232971003901, + "learning_rate": 7.6083769336176e-06, + "loss": 1.3411, + "step": 214 + }, + { + "epoch": 0.005582861912269313, + "grad_norm": 7.981957355122498, + "learning_rate": 7.614987166849946e-06, + "loss": 1.3366, + "step": 215 + }, + { + "epoch": 0.005608828711861263, + "grad_norm": 55.26763054743821, + "learning_rate": 7.621566726035842e-06, + "loss": 1.3349, + "step": 216 + }, + { + "epoch": 0.005634795511453214, + "grad_norm": 9.36976619856241, + "learning_rate": 7.628115894539815e-06, + "loss": 1.317, + "step": 217 + }, + { + "epoch": 0.005660762311045163, + "grad_norm": 13.451327142862034, + "learning_rate": 7.634634951817898e-06, + "loss": 1.3679, + "step": 218 + }, + { + "epoch": 0.005686729110637114, + "grad_norm": 123.50038746143673, + "learning_rate": 7.641124173489172e-06, + "loss": 1.3204, + "step": 219 + }, + { + "epoch": 0.005712695910229064, + "grad_norm": 7.227459826046035, + "learning_rate": 7.647583831405682e-06, + "loss": 1.3339, + "step": 220 + }, + { + "epoch": 0.005738662709821015, + "grad_norm": 20.714984810593634, + "learning_rate": 7.654014193720782e-06, + "loss": 1.3468, + "step": 221 + }, + { + "epoch": 0.005764629509412965, + "grad_norm": 8.575404335031518, + "learning_rate": 7.660415524955935e-06, + "loss": 1.3678, + "step": 222 + }, + { + "epoch": 0.005790596309004915, + "grad_norm": 7.160803266678416, + "learning_rate": 7.66678808606599e-06, + "loss": 1.3147, + "step": 223 + }, + { + "epoch": 0.005816563108596865, + "grad_norm": 14.710955651708133, + "learning_rate": 7.67313213450302e-06, + "loss": 1.3462, + "step": 224 + }, + { + "epoch": 0.005842529908188816, + "grad_norm": 9.41572169961027, + "learning_rate": 7.679447924278749e-06, + "loss": 1.31, + "step": 225 + }, + { + "epoch": 0.005868496707780766, + "grad_norm": 7.9713889146903405, + "learning_rate": 7.685735706025554e-06, + "loss": 1.3749, + "step": 226 + }, + { + "epoch": 0.005894463507372717, + "grad_norm": 6.4611887208376215, + "learning_rate": 7.691995727056155e-06, + "loss": 1.3525, + "step": 227 + }, + { + "epoch": 0.005920430306964666, + "grad_norm": 10.622313395401079, + "learning_rate": 7.698228231421993e-06, + "loss": 1.3047, + "step": 228 + }, + { + "epoch": 0.005946397106556617, + "grad_norm": 7.760333647486589, + "learning_rate": 7.704433459970307e-06, + "loss": 1.3254, + "step": 229 + }, + { + "epoch": 0.005972363906148567, + "grad_norm": 5.8583065432084735, + "learning_rate": 7.710611650399987e-06, + "loss": 1.3149, + "step": 230 + }, + { + "epoch": 0.005998330705740518, + "grad_norm": 8.20501605806982, + "learning_rate": 7.716763037316203e-06, + "loss": 1.2932, + "step": 231 + }, + { + "epoch": 0.006024297505332467, + "grad_norm": 6.875382291395296, + "learning_rate": 7.722887852283877e-06, + "loss": 1.3203, + "step": 232 + }, + { + "epoch": 0.006050264304924418, + "grad_norm": 26.869264922229217, + "learning_rate": 7.72898632387997e-06, + "loss": 1.336, + "step": 233 + }, + { + "epoch": 0.006076231104516368, + "grad_norm": 9.363350102088319, + "learning_rate": 7.735058677744676e-06, + "loss": 1.4033, + "step": 234 + }, + { + "epoch": 0.006102197904108319, + "grad_norm": 9.617671802514312, + "learning_rate": 7.741105136631522e-06, + "loss": 1.3055, + "step": 235 + }, + { + "epoch": 0.006128164703700269, + "grad_norm": 8.310921655400623, + "learning_rate": 7.747125920456388e-06, + "loss": 1.3452, + "step": 236 + }, + { + "epoch": 0.006154131503292219, + "grad_norm": 67.017805324173, + "learning_rate": 7.75312124634551e-06, + "loss": 1.3431, + "step": 237 + }, + { + "epoch": 0.006180098302884169, + "grad_norm": 9.122703367743568, + "learning_rate": 7.759091328682457e-06, + "loss": 1.3231, + "step": 238 + }, + { + "epoch": 0.00620606510247612, + "grad_norm": 6.3775782140035275, + "learning_rate": 7.765036379154136e-06, + "loss": 1.3677, + "step": 239 + }, + { + "epoch": 0.00623203190206807, + "grad_norm": 17.411403803883104, + "learning_rate": 7.770956606795828e-06, + "loss": 1.2984, + "step": 240 + }, + { + "epoch": 0.006257998701660021, + "grad_norm": 10.143552007993412, + "learning_rate": 7.776852218035288e-06, + "loss": 1.3445, + "step": 241 + }, + { + "epoch": 0.00628396550125197, + "grad_norm": 10.123678238860833, + "learning_rate": 7.782723416735941e-06, + "loss": 1.375, + "step": 242 + }, + { + "epoch": 0.006309932300843921, + "grad_norm": 6.371850526022763, + "learning_rate": 7.788570404239169e-06, + "loss": 1.3653, + "step": 243 + }, + { + "epoch": 0.006335899100435871, + "grad_norm": 5.89328177767566, + "learning_rate": 7.794393379405754e-06, + "loss": 1.3096, + "step": 244 + }, + { + "epoch": 0.006361865900027822, + "grad_norm": 8.603886511231103, + "learning_rate": 7.800192538656454e-06, + "loss": 1.3078, + "step": 245 + }, + { + "epoch": 0.006387832699619772, + "grad_norm": 8.645650716178183, + "learning_rate": 7.805968076011753e-06, + "loss": 1.2719, + "step": 246 + }, + { + "epoch": 0.006413799499211722, + "grad_norm": 8.08649558385069, + "learning_rate": 7.81172018313083e-06, + "loss": 1.3295, + "step": 247 + }, + { + "epoch": 0.006439766298803672, + "grad_norm": 11.904470395833384, + "learning_rate": 7.817449049349698e-06, + "loss": 1.2976, + "step": 248 + }, + { + "epoch": 0.006465733098395623, + "grad_norm": 13.380287337475172, + "learning_rate": 7.823154861718619e-06, + "loss": 1.3463, + "step": 249 + }, + { + "epoch": 0.006491699897987573, + "grad_norm": 10.258324568925755, + "learning_rate": 7.828837805038734e-06, + "loss": 1.3248, + "step": 250 + }, + { + "epoch": 0.006517666697579524, + "grad_norm": 6.645810245804659, + "learning_rate": 7.834498061897997e-06, + "loss": 1.3229, + "step": 251 + }, + { + "epoch": 0.006543633497171473, + "grad_norm": 12.370636758412791, + "learning_rate": 7.84013581270635e-06, + "loss": 1.3401, + "step": 252 + }, + { + "epoch": 0.006569600296763424, + "grad_norm": 6.798062205493258, + "learning_rate": 7.845751235730246e-06, + "loss": 1.3291, + "step": 253 + }, + { + "epoch": 0.006595567096355374, + "grad_norm": 8.803854739622139, + "learning_rate": 7.851344507126461e-06, + "loss": 1.3397, + "step": 254 + }, + { + "epoch": 0.006621533895947325, + "grad_norm": 13.118091327184736, + "learning_rate": 7.856915800975262e-06, + "loss": 1.3658, + "step": 255 + }, + { + "epoch": 0.006647500695539274, + "grad_norm": 8.944820367193921, + "learning_rate": 7.862465289312904e-06, + "loss": 1.3595, + "step": 256 + }, + { + "epoch": 0.006673467495131225, + "grad_norm": 9.043443730776387, + "learning_rate": 7.867993142163519e-06, + "loss": 1.3599, + "step": 257 + }, + { + "epoch": 0.006699434294723175, + "grad_norm": 9.929264902986116, + "learning_rate": 7.873499527570354e-06, + "loss": 1.3672, + "step": 258 + }, + { + "epoch": 0.006725401094315126, + "grad_norm": 11.242870190789944, + "learning_rate": 7.878984611626443e-06, + "loss": 1.3362, + "step": 259 + }, + { + "epoch": 0.006751367893907076, + "grad_norm": 9.896732415297485, + "learning_rate": 7.884448558504662e-06, + "loss": 1.4016, + "step": 260 + }, + { + "epoch": 0.006777334693499026, + "grad_norm": 8.787737747383957, + "learning_rate": 7.889891530487207e-06, + "loss": 1.3494, + "step": 261 + }, + { + "epoch": 0.006803301493090976, + "grad_norm": 9.270952156819869, + "learning_rate": 7.895313687994511e-06, + "loss": 1.2866, + "step": 262 + }, + { + "epoch": 0.006829268292682927, + "grad_norm": 8.802471108707424, + "learning_rate": 7.90071518961364e-06, + "loss": 1.372, + "step": 263 + }, + { + "epoch": 0.006855235092274877, + "grad_norm": 9.18034073473634, + "learning_rate": 7.906096192126086e-06, + "loss": 1.3346, + "step": 264 + }, + { + "epoch": 0.006881201891866828, + "grad_norm": 6.511979434831468, + "learning_rate": 7.91145685053511e-06, + "loss": 1.3295, + "step": 265 + }, + { + "epoch": 0.006907168691458777, + "grad_norm": 7.1220536414217674, + "learning_rate": 7.916797318092504e-06, + "loss": 1.3226, + "step": 266 + }, + { + "epoch": 0.006933135491050728, + "grad_norm": 14.632575787260487, + "learning_rate": 7.922117746324892e-06, + "loss": 1.2882, + "step": 267 + }, + { + "epoch": 0.006959102290642678, + "grad_norm": 12.03047252966854, + "learning_rate": 7.927418285059532e-06, + "loss": 1.3179, + "step": 268 + }, + { + "epoch": 0.006985069090234629, + "grad_norm": 7.132617559202905, + "learning_rate": 7.932699082449625e-06, + "loss": 1.3161, + "step": 269 + }, + { + "epoch": 0.007011035889826579, + "grad_norm": 8.880372046847457, + "learning_rate": 7.937960284999156e-06, + "loss": 1.3506, + "step": 270 + }, + { + "epoch": 0.007037002689418529, + "grad_norm": 8.621517196923953, + "learning_rate": 7.943202037587298e-06, + "loss": 1.3627, + "step": 271 + }, + { + "epoch": 0.007062969489010479, + "grad_norm": 18.460590347984372, + "learning_rate": 7.948424483492341e-06, + "loss": 1.3132, + "step": 272 + }, + { + "epoch": 0.00708893628860243, + "grad_norm": 21.72006586054306, + "learning_rate": 7.953627764415187e-06, + "loss": 1.2672, + "step": 273 + }, + { + "epoch": 0.00711490308819438, + "grad_norm": 6.496909066863696, + "learning_rate": 7.958812020502428e-06, + "loss": 1.302, + "step": 274 + }, + { + "epoch": 0.00714086988778633, + "grad_norm": 7.720008407746318, + "learning_rate": 7.963977390368995e-06, + "loss": 1.3412, + "step": 275 + }, + { + "epoch": 0.00716683668737828, + "grad_norm": 7.97355596688705, + "learning_rate": 7.969124011120391e-06, + "loss": 1.2785, + "step": 276 + }, + { + "epoch": 0.007192803486970231, + "grad_norm": 7.065621825415309, + "learning_rate": 7.974252018374531e-06, + "loss": 1.312, + "step": 277 + }, + { + "epoch": 0.007218770286562181, + "grad_norm": 6.136903678686395, + "learning_rate": 7.979361546283181e-06, + "loss": 1.3367, + "step": 278 + }, + { + "epoch": 0.007244737086154132, + "grad_norm": 6.6185127021245185, + "learning_rate": 7.984452727553026e-06, + "loss": 1.3175, + "step": 279 + }, + { + "epoch": 0.007270703885746081, + "grad_norm": 18.039243516307025, + "learning_rate": 7.989525693466335e-06, + "loss": 1.3621, + "step": 280 + }, + { + "epoch": 0.007296670685338032, + "grad_norm": 6.549995352474049, + "learning_rate": 7.99458057390129e-06, + "loss": 1.2807, + "step": 281 + }, + { + "epoch": 0.007322637484929982, + "grad_norm": 6.850537422427106, + "learning_rate": 7.99961749735193e-06, + "loss": 1.3491, + "step": 282 + }, + { + "epoch": 0.007348604284521933, + "grad_norm": 35.012267660332284, + "learning_rate": 8.004636590947738e-06, + "loss": 1.3712, + "step": 283 + }, + { + "epoch": 0.007374571084113883, + "grad_norm": 8.012232012142825, + "learning_rate": 8.009637980472924e-06, + "loss": 1.3338, + "step": 284 + }, + { + "epoch": 0.007400537883705833, + "grad_norm": 7.53472668188132, + "learning_rate": 8.014621790385308e-06, + "loss": 1.3488, + "step": 285 + }, + { + "epoch": 0.007426504683297783, + "grad_norm": 6.701185821812214, + "learning_rate": 8.019588143834921e-06, + "loss": 1.3261, + "step": 286 + }, + { + "epoch": 0.007452471482889734, + "grad_norm": 13.09963310834476, + "learning_rate": 8.024537162682263e-06, + "loss": 1.3009, + "step": 287 + }, + { + "epoch": 0.007478438282481684, + "grad_norm": 8.611267214218124, + "learning_rate": 8.029468967516232e-06, + "loss": 1.3634, + "step": 288 + }, + { + "epoch": 0.007504405082073635, + "grad_norm": 10.418050770674501, + "learning_rate": 8.034383677671775e-06, + "loss": 1.3336, + "step": 289 + }, + { + "epoch": 0.007530371881665584, + "grad_norm": 8.163813449401028, + "learning_rate": 8.039281411247192e-06, + "loss": 1.3103, + "step": 290 + }, + { + "epoch": 0.007556338681257535, + "grad_norm": 10.877005556582644, + "learning_rate": 8.044162285121181e-06, + "loss": 1.2917, + "step": 291 + }, + { + "epoch": 0.007582305480849485, + "grad_norm": 11.176676879963095, + "learning_rate": 8.049026414969564e-06, + "loss": 1.337, + "step": 292 + }, + { + "epoch": 0.007608272280441436, + "grad_norm": 9.174352671891471, + "learning_rate": 8.053873915281741e-06, + "loss": 1.2822, + "step": 293 + }, + { + "epoch": 0.007634239080033386, + "grad_norm": 9.914802397405095, + "learning_rate": 8.05870489937686e-06, + "loss": 1.2873, + "step": 294 + }, + { + "epoch": 0.007660205879625336, + "grad_norm": 7.7322134369007856, + "learning_rate": 8.063519479419703e-06, + "loss": 1.3384, + "step": 295 + }, + { + "epoch": 0.007686172679217286, + "grad_norm": 28.19253393589449, + "learning_rate": 8.068317766436328e-06, + "loss": 1.3465, + "step": 296 + }, + { + "epoch": 0.007712139478809237, + "grad_norm": 13.530092960779823, + "learning_rate": 8.073099870329416e-06, + "loss": 1.2604, + "step": 297 + }, + { + "epoch": 0.007738106278401187, + "grad_norm": 48.151400530481716, + "learning_rate": 8.077865899893382e-06, + "loss": 1.3166, + "step": 298 + }, + { + "epoch": 0.007764073077993137, + "grad_norm": 21.928781543007297, + "learning_rate": 8.082615962829228e-06, + "loss": 1.3289, + "step": 299 + }, + { + "epoch": 0.007790039877585087, + "grad_norm": 7.46148813822125, + "learning_rate": 8.087350165759142e-06, + "loss": 1.3655, + "step": 300 + }, + { + "epoch": 0.007816006677177039, + "grad_norm": 9.11497594950522, + "learning_rate": 8.092068614240862e-06, + "loss": 1.3542, + "step": 301 + }, + { + "epoch": 0.007841973476768987, + "grad_norm": 30.764909804726695, + "learning_rate": 8.096771412781796e-06, + "loss": 1.3151, + "step": 302 + }, + { + "epoch": 0.007867940276360938, + "grad_norm": 16.4965650277518, + "learning_rate": 8.101458664852908e-06, + "loss": 1.3254, + "step": 303 + }, + { + "epoch": 0.007893907075952888, + "grad_norm": 7.479352376716694, + "learning_rate": 8.106130472902386e-06, + "loss": 1.2903, + "step": 304 + }, + { + "epoch": 0.007919873875544839, + "grad_norm": 7.482740506462893, + "learning_rate": 8.11078693836907e-06, + "loss": 1.341, + "step": 305 + }, + { + "epoch": 0.00794584067513679, + "grad_norm": 8.554184934581409, + "learning_rate": 8.115428161695668e-06, + "loss": 1.2741, + "step": 306 + }, + { + "epoch": 0.00797180747472874, + "grad_norm": 12.29465302649864, + "learning_rate": 8.120054242341762e-06, + "loss": 1.4047, + "step": 307 + }, + { + "epoch": 0.00799777427432069, + "grad_norm": 8.519258560371146, + "learning_rate": 8.124665278796595e-06, + "loss": 1.3364, + "step": 308 + }, + { + "epoch": 0.00802374107391264, + "grad_norm": 9.562780438468483, + "learning_rate": 8.129261368591647e-06, + "loss": 1.3074, + "step": 309 + }, + { + "epoch": 0.008049707873504591, + "grad_norm": 9.465584908945512, + "learning_rate": 8.133842608313013e-06, + "loss": 1.3066, + "step": 310 + }, + { + "epoch": 0.00807567467309654, + "grad_norm": 11.084808442659224, + "learning_rate": 8.138409093613595e-06, + "loss": 1.3473, + "step": 311 + }, + { + "epoch": 0.00810164147268849, + "grad_norm": 7.739093366253686, + "learning_rate": 8.14296091922507e-06, + "loss": 1.3603, + "step": 312 + }, + { + "epoch": 0.008127608272280441, + "grad_norm": 7.3436679267355265, + "learning_rate": 8.147498178969687e-06, + "loss": 1.3074, + "step": 313 + }, + { + "epoch": 0.008153575071872391, + "grad_norm": 6.684921765893247, + "learning_rate": 8.152020965771878e-06, + "loss": 1.284, + "step": 314 + }, + { + "epoch": 0.008179541871464342, + "grad_norm": 9.769524612027933, + "learning_rate": 8.156529371669665e-06, + "loss": 1.3036, + "step": 315 + }, + { + "epoch": 0.008205508671056292, + "grad_norm": 6.773991088068427, + "learning_rate": 8.161023487825902e-06, + "loss": 1.3137, + "step": 316 + }, + { + "epoch": 0.008231475470648243, + "grad_norm": 12.179017230193878, + "learning_rate": 8.16550340453934e-06, + "loss": 1.3228, + "step": 317 + }, + { + "epoch": 0.008257442270240193, + "grad_norm": 14.197717424068946, + "learning_rate": 8.169969211255516e-06, + "loss": 1.347, + "step": 318 + }, + { + "epoch": 0.008283409069832144, + "grad_norm": 11.358211662403935, + "learning_rate": 8.174420996577452e-06, + "loss": 1.3583, + "step": 319 + }, + { + "epoch": 0.008309375869424094, + "grad_norm": 16.21989586812, + "learning_rate": 8.17885884827622e-06, + "loss": 1.2989, + "step": 320 + }, + { + "epoch": 0.008335342669016043, + "grad_norm": 15.520815746459554, + "learning_rate": 8.18328285330132e-06, + "loss": 1.3354, + "step": 321 + }, + { + "epoch": 0.008361309468607993, + "grad_norm": 10.815957988296267, + "learning_rate": 8.187693097790901e-06, + "loss": 1.368, + "step": 322 + }, + { + "epoch": 0.008387276268199944, + "grad_norm": 10.876482983027103, + "learning_rate": 8.192089667081821e-06, + "loss": 1.3192, + "step": 323 + }, + { + "epoch": 0.008413243067791894, + "grad_norm": 16.671686459219753, + "learning_rate": 8.19647264571956e-06, + "loss": 1.3481, + "step": 324 + }, + { + "epoch": 0.008439209867383845, + "grad_norm": 12.076490126254866, + "learning_rate": 8.200842117467977e-06, + "loss": 1.4066, + "step": 325 + }, + { + "epoch": 0.008465176666975795, + "grad_norm": 8.878343409074304, + "learning_rate": 8.205198165318907e-06, + "loss": 1.3239, + "step": 326 + }, + { + "epoch": 0.008491143466567746, + "grad_norm": 17.014450214840362, + "learning_rate": 8.20954087150162e-06, + "loss": 1.368, + "step": 327 + }, + { + "epoch": 0.008517110266159696, + "grad_norm": 7.500569385531857, + "learning_rate": 8.213870317492146e-06, + "loss": 1.2655, + "step": 328 + }, + { + "epoch": 0.008543077065751647, + "grad_norm": 8.38835513990228, + "learning_rate": 8.218186584022438e-06, + "loss": 1.2958, + "step": 329 + }, + { + "epoch": 0.008569043865343596, + "grad_norm": 11.806358072766674, + "learning_rate": 8.222489751089401e-06, + "loss": 1.3349, + "step": 330 + }, + { + "epoch": 0.008595010664935546, + "grad_norm": 11.669761537862971, + "learning_rate": 8.2267798979638e-06, + "loss": 1.2796, + "step": 331 + }, + { + "epoch": 0.008620977464527496, + "grad_norm": 6.457679305236312, + "learning_rate": 8.23105710319901e-06, + "loss": 1.3004, + "step": 332 + }, + { + "epoch": 0.008646944264119447, + "grad_norm": 7.702922454709103, + "learning_rate": 8.235321444639656e-06, + "loss": 1.3333, + "step": 333 + }, + { + "epoch": 0.008672911063711397, + "grad_norm": 9.694037612593371, + "learning_rate": 8.239572999430105e-06, + "loss": 1.3713, + "step": 334 + }, + { + "epoch": 0.008698877863303348, + "grad_norm": 6.260321067888362, + "learning_rate": 8.243811844022847e-06, + "loss": 1.3499, + "step": 335 + }, + { + "epoch": 0.008724844662895298, + "grad_norm": 6.271144043922839, + "learning_rate": 8.248038054186743e-06, + "loss": 1.3211, + "step": 336 + }, + { + "epoch": 0.008750811462487249, + "grad_norm": 8.871865162133503, + "learning_rate": 8.252251705015143e-06, + "loss": 1.3437, + "step": 337 + }, + { + "epoch": 0.0087767782620792, + "grad_norm": 8.427454867540591, + "learning_rate": 8.256452870933905e-06, + "loss": 1.3057, + "step": 338 + }, + { + "epoch": 0.00880274506167115, + "grad_norm": 5.303857392881011, + "learning_rate": 8.260641625709273e-06, + "loss": 1.3295, + "step": 339 + }, + { + "epoch": 0.008828711861263099, + "grad_norm": 10.905614796151166, + "learning_rate": 8.264818042455655e-06, + "loss": 1.3395, + "step": 340 + }, + { + "epoch": 0.008854678660855049, + "grad_norm": 6.666194453999609, + "learning_rate": 8.268982193643272e-06, + "loss": 1.3235, + "step": 341 + }, + { + "epoch": 0.008880645460447, + "grad_norm": 6.029204501033549, + "learning_rate": 8.273134151105714e-06, + "loss": 1.3478, + "step": 342 + }, + { + "epoch": 0.00890661226003895, + "grad_norm": 7.000323696826566, + "learning_rate": 8.277273986047367e-06, + "loss": 1.3205, + "step": 343 + }, + { + "epoch": 0.0089325790596309, + "grad_norm": 8.103916925996776, + "learning_rate": 8.281401769050745e-06, + "loss": 1.3429, + "step": 344 + }, + { + "epoch": 0.008958545859222851, + "grad_norm": 7.185255147907229, + "learning_rate": 8.285517570083706e-06, + "loss": 1.2911, + "step": 345 + }, + { + "epoch": 0.008984512658814801, + "grad_norm": 7.380007837803098, + "learning_rate": 8.289621458506573e-06, + "loss": 1.3415, + "step": 346 + }, + { + "epoch": 0.009010479458406752, + "grad_norm": 8.760110735607036, + "learning_rate": 8.293713503079141e-06, + "loss": 1.3316, + "step": 347 + }, + { + "epoch": 0.009036446257998702, + "grad_norm": 12.26091295835125, + "learning_rate": 8.297793771967598e-06, + "loss": 1.3194, + "step": 348 + }, + { + "epoch": 0.009062413057590653, + "grad_norm": 7.346531377794021, + "learning_rate": 8.301862332751332e-06, + "loss": 1.3705, + "step": 349 + }, + { + "epoch": 0.009088379857182602, + "grad_norm": 6.2407475458986985, + "learning_rate": 8.30591925242965e-06, + "loss": 1.2924, + "step": 350 + }, + { + "epoch": 0.009114346656774552, + "grad_norm": 11.724512255188712, + "learning_rate": 8.309964597428397e-06, + "loss": 1.3795, + "step": 351 + }, + { + "epoch": 0.009140313456366502, + "grad_norm": 12.390461212268669, + "learning_rate": 8.313998433606479e-06, + "loss": 1.2961, + "step": 352 + }, + { + "epoch": 0.009166280255958453, + "grad_norm": 8.084909449290242, + "learning_rate": 8.3180208262623e-06, + "loss": 1.3388, + "step": 353 + }, + { + "epoch": 0.009192247055550403, + "grad_norm": 8.476629104001004, + "learning_rate": 8.322031840140109e-06, + "loss": 1.3583, + "step": 354 + }, + { + "epoch": 0.009218213855142354, + "grad_norm": 11.72318340617448, + "learning_rate": 8.326031539436239e-06, + "loss": 1.3318, + "step": 355 + }, + { + "epoch": 0.009244180654734304, + "grad_norm": 17.19596471247188, + "learning_rate": 8.330019987805285e-06, + "loss": 1.3348, + "step": 356 + }, + { + "epoch": 0.009270147454326255, + "grad_norm": 9.533497029568581, + "learning_rate": 8.333997248366176e-06, + "loss": 1.2795, + "step": 357 + }, + { + "epoch": 0.009296114253918205, + "grad_norm": 25.8323416570833, + "learning_rate": 8.337963383708176e-06, + "loss": 1.3164, + "step": 358 + }, + { + "epoch": 0.009322081053510154, + "grad_norm": 10.029549617981397, + "learning_rate": 8.341918455896776e-06, + "loss": 1.2734, + "step": 359 + }, + { + "epoch": 0.009348047853102105, + "grad_norm": 8.751897987510427, + "learning_rate": 8.345862526479548e-06, + "loss": 1.3325, + "step": 360 + }, + { + "epoch": 0.009374014652694055, + "grad_norm": 8.871129011103413, + "learning_rate": 8.349795656491867e-06, + "loss": 1.345, + "step": 361 + }, + { + "epoch": 0.009399981452286005, + "grad_norm": 7.675148570458639, + "learning_rate": 8.353717906462596e-06, + "loss": 1.3233, + "step": 362 + }, + { + "epoch": 0.009425948251877956, + "grad_norm": 11.736448224183725, + "learning_rate": 8.35762933641966e-06, + "loss": 1.2794, + "step": 363 + }, + { + "epoch": 0.009451915051469906, + "grad_norm": 6.15664160397889, + "learning_rate": 8.361530005895578e-06, + "loss": 1.2959, + "step": 364 + }, + { + "epoch": 0.009477881851061857, + "grad_norm": 7.896612150756339, + "learning_rate": 8.365419973932877e-06, + "loss": 1.3352, + "step": 365 + }, + { + "epoch": 0.009503848650653807, + "grad_norm": 5.725559870234239, + "learning_rate": 8.369299299089475e-06, + "loss": 1.312, + "step": 366 + }, + { + "epoch": 0.009529815450245758, + "grad_norm": 9.00910489156751, + "learning_rate": 8.373168039443953e-06, + "loss": 1.2983, + "step": 367 + }, + { + "epoch": 0.009555782249837708, + "grad_norm": 6.023136670706045, + "learning_rate": 8.377026252600784e-06, + "loss": 1.3382, + "step": 368 + }, + { + "epoch": 0.009581749049429657, + "grad_norm": 6.440249550290697, + "learning_rate": 8.380873995695474e-06, + "loss": 1.3093, + "step": 369 + }, + { + "epoch": 0.009607715849021608, + "grad_norm": 6.385236114222794, + "learning_rate": 8.384711325399641e-06, + "loss": 1.3444, + "step": 370 + }, + { + "epoch": 0.009633682648613558, + "grad_norm": 9.37516316589781, + "learning_rate": 8.388538297926024e-06, + "loss": 1.2719, + "step": 371 + }, + { + "epoch": 0.009659649448205508, + "grad_norm": 10.452128318632534, + "learning_rate": 8.392354969033419e-06, + "loss": 1.2962, + "step": 372 + }, + { + "epoch": 0.009685616247797459, + "grad_norm": 7.880888901741465, + "learning_rate": 8.396161394031556e-06, + "loss": 1.3398, + "step": 373 + }, + { + "epoch": 0.00971158304738941, + "grad_norm": 6.493042082837747, + "learning_rate": 8.399957627785914e-06, + "loss": 1.3446, + "step": 374 + }, + { + "epoch": 0.00973754984698136, + "grad_norm": 11.569726283891926, + "learning_rate": 8.403743724722457e-06, + "loss": 1.3247, + "step": 375 + }, + { + "epoch": 0.00976351664657331, + "grad_norm": 6.853232029167367, + "learning_rate": 8.40751973883232e-06, + "loss": 1.2906, + "step": 376 + }, + { + "epoch": 0.00978948344616526, + "grad_norm": 13.805150074486752, + "learning_rate": 8.411285723676433e-06, + "loss": 1.2665, + "step": 377 + }, + { + "epoch": 0.00981545024575721, + "grad_norm": 17.736827852123373, + "learning_rate": 8.41504173239007e-06, + "loss": 1.3635, + "step": 378 + }, + { + "epoch": 0.00984141704534916, + "grad_norm": 10.441537772740675, + "learning_rate": 8.418787817687356e-06, + "loss": 1.3385, + "step": 379 + }, + { + "epoch": 0.00986738384494111, + "grad_norm": 7.712042361242046, + "learning_rate": 8.422524031865701e-06, + "loss": 1.3786, + "step": 380 + }, + { + "epoch": 0.009893350644533061, + "grad_norm": 5.769924307976265, + "learning_rate": 8.426250426810182e-06, + "loss": 1.3124, + "step": 381 + }, + { + "epoch": 0.009919317444125011, + "grad_norm": 6.186067049192837, + "learning_rate": 8.429967053997868e-06, + "loss": 1.3281, + "step": 382 + }, + { + "epoch": 0.009945284243716962, + "grad_norm": 8.321126720723807, + "learning_rate": 8.433673964502083e-06, + "loss": 1.3269, + "step": 383 + }, + { + "epoch": 0.009971251043308912, + "grad_norm": 5.898701036641689, + "learning_rate": 8.437371208996625e-06, + "loss": 1.3661, + "step": 384 + }, + { + "epoch": 0.009997217842900863, + "grad_norm": 7.703732108745755, + "learning_rate": 8.441058837759911e-06, + "loss": 1.2866, + "step": 385 + }, + { + "epoch": 0.010023184642492813, + "grad_norm": 6.797432010671182, + "learning_rate": 8.444736900679085e-06, + "loss": 1.2959, + "step": 386 + }, + { + "epoch": 0.010049151442084764, + "grad_norm": 8.701968213769632, + "learning_rate": 8.448405447254075e-06, + "loss": 1.3062, + "step": 387 + }, + { + "epoch": 0.010075118241676713, + "grad_norm": 17.26909078620039, + "learning_rate": 8.452064526601574e-06, + "loss": 1.3395, + "step": 388 + }, + { + "epoch": 0.010101085041268663, + "grad_norm": 11.113619694978292, + "learning_rate": 8.455714187458997e-06, + "loss": 1.2838, + "step": 389 + }, + { + "epoch": 0.010127051840860614, + "grad_norm": 14.090985147489079, + "learning_rate": 8.459354478188384e-06, + "loss": 1.305, + "step": 390 + }, + { + "epoch": 0.010153018640452564, + "grad_norm": 8.184098097525583, + "learning_rate": 8.462985446780219e-06, + "loss": 1.3261, + "step": 391 + }, + { + "epoch": 0.010178985440044514, + "grad_norm": 7.603606181630663, + "learning_rate": 8.466607140857251e-06, + "loss": 1.3311, + "step": 392 + }, + { + "epoch": 0.010204952239636465, + "grad_norm": 8.722347050750185, + "learning_rate": 8.470219607678232e-06, + "loss": 1.3135, + "step": 393 + }, + { + "epoch": 0.010230919039228415, + "grad_norm": 20.83685875119346, + "learning_rate": 8.473822894141617e-06, + "loss": 1.302, + "step": 394 + }, + { + "epoch": 0.010256885838820366, + "grad_norm": 6.976740477065645, + "learning_rate": 8.477417046789217e-06, + "loss": 1.4206, + "step": 395 + }, + { + "epoch": 0.010282852638412316, + "grad_norm": 10.933475473301844, + "learning_rate": 8.481002111809807e-06, + "loss": 1.2899, + "step": 396 + }, + { + "epoch": 0.010308819438004265, + "grad_norm": 8.75623034325952, + "learning_rate": 8.484578135042691e-06, + "loss": 1.3136, + "step": 397 + }, + { + "epoch": 0.010334786237596216, + "grad_norm": 7.781121310353383, + "learning_rate": 8.48814516198121e-06, + "loss": 1.2797, + "step": 398 + }, + { + "epoch": 0.010360753037188166, + "grad_norm": 7.018769305127057, + "learning_rate": 8.491703237776225e-06, + "loss": 1.3303, + "step": 399 + }, + { + "epoch": 0.010386719836780117, + "grad_norm": 8.044161025944726, + "learning_rate": 8.495252407239533e-06, + "loss": 1.318, + "step": 400 + }, + { + "epoch": 0.010412686636372067, + "grad_norm": 6.36698619313673, + "learning_rate": 8.498792714847276e-06, + "loss": 1.2821, + "step": 401 + }, + { + "epoch": 0.010438653435964017, + "grad_norm": 10.383729869767452, + "learning_rate": 8.502324204743254e-06, + "loss": 1.3111, + "step": 402 + }, + { + "epoch": 0.010464620235555968, + "grad_norm": 19.929967719021775, + "learning_rate": 8.505846920742254e-06, + "loss": 1.3135, + "step": 403 + }, + { + "epoch": 0.010490587035147918, + "grad_norm": 20.099977468601335, + "learning_rate": 8.509360906333301e-06, + "loss": 1.328, + "step": 404 + }, + { + "epoch": 0.010516553834739869, + "grad_norm": 9.359546589751165, + "learning_rate": 8.512866204682875e-06, + "loss": 1.292, + "step": 405 + }, + { + "epoch": 0.01054252063433182, + "grad_norm": 21.01337624469642, + "learning_rate": 8.516362858638106e-06, + "loss": 1.3853, + "step": 406 + }, + { + "epoch": 0.010568487433923768, + "grad_norm": 7.672721986961927, + "learning_rate": 8.5198509107299e-06, + "loss": 1.3317, + "step": 407 + }, + { + "epoch": 0.010594454233515719, + "grad_norm": 5.951840399766034, + "learning_rate": 8.52333040317606e-06, + "loss": 1.3225, + "step": 408 + }, + { + "epoch": 0.010620421033107669, + "grad_norm": 7.012137924883387, + "learning_rate": 8.526801377884334e-06, + "loss": 1.3224, + "step": 409 + }, + { + "epoch": 0.01064638783269962, + "grad_norm": 7.990295007103124, + "learning_rate": 8.53026387645546e-06, + "loss": 1.3227, + "step": 410 + }, + { + "epoch": 0.01067235463229157, + "grad_norm": 7.019355716537877, + "learning_rate": 8.533717940186149e-06, + "loss": 1.3206, + "step": 411 + }, + { + "epoch": 0.01069832143188352, + "grad_norm": 13.27508760987304, + "learning_rate": 8.537163610072039e-06, + "loss": 1.3406, + "step": 412 + }, + { + "epoch": 0.010724288231475471, + "grad_norm": 7.280744150296732, + "learning_rate": 8.54060092681062e-06, + "loss": 1.3355, + "step": 413 + }, + { + "epoch": 0.010750255031067421, + "grad_norm": 10.454443057316437, + "learning_rate": 8.544029930804112e-06, + "loss": 1.3307, + "step": 414 + }, + { + "epoch": 0.010776221830659372, + "grad_norm": 7.501350789816125, + "learning_rate": 8.547450662162324e-06, + "loss": 1.3074, + "step": 415 + }, + { + "epoch": 0.010802188630251322, + "grad_norm": 6.930846726954727, + "learning_rate": 8.550863160705462e-06, + "loss": 1.3242, + "step": 416 + }, + { + "epoch": 0.010828155429843271, + "grad_norm": 9.40454288846183, + "learning_rate": 8.554267465966904e-06, + "loss": 1.3099, + "step": 417 + }, + { + "epoch": 0.010854122229435222, + "grad_norm": 9.64250948647588, + "learning_rate": 8.557663617195961e-06, + "loss": 1.3177, + "step": 418 + }, + { + "epoch": 0.010880089029027172, + "grad_norm": 10.216995321923326, + "learning_rate": 8.561051653360584e-06, + "loss": 1.3154, + "step": 419 + }, + { + "epoch": 0.010906055828619123, + "grad_norm": 10.533806646151053, + "learning_rate": 8.564431613150058e-06, + "loss": 1.2982, + "step": 420 + }, + { + "epoch": 0.010932022628211073, + "grad_norm": 8.249024304579242, + "learning_rate": 8.567803534977628e-06, + "loss": 1.3643, + "step": 421 + }, + { + "epoch": 0.010957989427803023, + "grad_norm": 6.821933199834766, + "learning_rate": 8.571167456983147e-06, + "loss": 1.265, + "step": 422 + }, + { + "epoch": 0.010983956227394974, + "grad_norm": 11.64364539778733, + "learning_rate": 8.574523417035649e-06, + "loss": 1.3597, + "step": 423 + }, + { + "epoch": 0.011009923026986924, + "grad_norm": 8.547013360111805, + "learning_rate": 8.577871452735907e-06, + "loss": 1.3389, + "step": 424 + }, + { + "epoch": 0.011035889826578875, + "grad_norm": 12.51575728023884, + "learning_rate": 8.58121160141897e-06, + "loss": 1.2996, + "step": 425 + }, + { + "epoch": 0.011061856626170824, + "grad_norm": 6.503887187941285, + "learning_rate": 8.584543900156646e-06, + "loss": 1.3073, + "step": 426 + }, + { + "epoch": 0.011087823425762774, + "grad_norm": 8.662711145599271, + "learning_rate": 8.587868385759984e-06, + "loss": 1.3097, + "step": 427 + }, + { + "epoch": 0.011113790225354725, + "grad_norm": 6.834107455932946, + "learning_rate": 8.591185094781712e-06, + "loss": 1.2679, + "step": 428 + }, + { + "epoch": 0.011139757024946675, + "grad_norm": 6.4298619261729435, + "learning_rate": 8.594494063518644e-06, + "loss": 1.2943, + "step": 429 + }, + { + "epoch": 0.011165723824538626, + "grad_norm": 8.70481409635774, + "learning_rate": 8.59779532801406e-06, + "loss": 1.326, + "step": 430 + }, + { + "epoch": 0.011191690624130576, + "grad_norm": 5.903954438258581, + "learning_rate": 8.601088924060076e-06, + "loss": 1.3182, + "step": 431 + }, + { + "epoch": 0.011217657423722526, + "grad_norm": 11.8824664284336, + "learning_rate": 8.604374887199955e-06, + "loss": 1.3358, + "step": 432 + }, + { + "epoch": 0.011243624223314477, + "grad_norm": 12.624693611884025, + "learning_rate": 8.60765325273042e-06, + "loss": 1.3101, + "step": 433 + }, + { + "epoch": 0.011269591022906427, + "grad_norm": 7.643104946463594, + "learning_rate": 8.610924055703927e-06, + "loss": 1.2687, + "step": 434 + }, + { + "epoch": 0.011295557822498378, + "grad_norm": 13.344452671962902, + "learning_rate": 8.614187330930911e-06, + "loss": 1.3315, + "step": 435 + }, + { + "epoch": 0.011321524622090327, + "grad_norm": 49.021965241837, + "learning_rate": 8.61744311298201e-06, + "loss": 1.2778, + "step": 436 + }, + { + "epoch": 0.011347491421682277, + "grad_norm": 7.14513216573154, + "learning_rate": 8.620691436190264e-06, + "loss": 1.2894, + "step": 437 + }, + { + "epoch": 0.011373458221274228, + "grad_norm": 8.229940360352616, + "learning_rate": 8.623932334653285e-06, + "loss": 1.3516, + "step": 438 + }, + { + "epoch": 0.011399425020866178, + "grad_norm": 11.419731972349986, + "learning_rate": 8.627165842235402e-06, + "loss": 1.2959, + "step": 439 + }, + { + "epoch": 0.011425391820458129, + "grad_norm": 7.209055811079275, + "learning_rate": 8.630391992569792e-06, + "loss": 1.3563, + "step": 440 + }, + { + "epoch": 0.011451358620050079, + "grad_norm": 6.40050088591324, + "learning_rate": 8.633610819060579e-06, + "loss": 1.3164, + "step": 441 + }, + { + "epoch": 0.01147732541964203, + "grad_norm": 6.273048322299633, + "learning_rate": 8.636822354884896e-06, + "loss": 1.3187, + "step": 442 + }, + { + "epoch": 0.01150329221923398, + "grad_norm": 8.258419942032788, + "learning_rate": 8.640026632994951e-06, + "loss": 1.3302, + "step": 443 + }, + { + "epoch": 0.01152925901882593, + "grad_norm": 16.629842640860005, + "learning_rate": 8.643223686120049e-06, + "loss": 1.3237, + "step": 444 + }, + { + "epoch": 0.01155522581841788, + "grad_norm": 7.6472244313807245, + "learning_rate": 8.6464135467686e-06, + "loss": 1.2647, + "step": 445 + }, + { + "epoch": 0.01158119261800983, + "grad_norm": 8.715713464418599, + "learning_rate": 8.649596247230101e-06, + "loss": 1.3273, + "step": 446 + }, + { + "epoch": 0.01160715941760178, + "grad_norm": 11.869932358250956, + "learning_rate": 8.652771819577103e-06, + "loss": 1.2864, + "step": 447 + }, + { + "epoch": 0.01163312621719373, + "grad_norm": 7.558537982271662, + "learning_rate": 8.655940295667134e-06, + "loss": 1.2688, + "step": 448 + }, + { + "epoch": 0.011659093016785681, + "grad_norm": 6.43557888364574, + "learning_rate": 8.659101707144642e-06, + "loss": 1.3129, + "step": 449 + }, + { + "epoch": 0.011685059816377632, + "grad_norm": 10.57662975838232, + "learning_rate": 8.662256085442863e-06, + "loss": 1.3073, + "step": 450 + }, + { + "epoch": 0.011711026615969582, + "grad_norm": 9.285370705085466, + "learning_rate": 8.66540346178572e-06, + "loss": 1.2818, + "step": 451 + }, + { + "epoch": 0.011736993415561532, + "grad_norm": 7.901559757822838, + "learning_rate": 8.668543867189666e-06, + "loss": 1.3301, + "step": 452 + }, + { + "epoch": 0.011762960215153483, + "grad_norm": 8.429572103734019, + "learning_rate": 8.671677332465516e-06, + "loss": 1.3729, + "step": 453 + }, + { + "epoch": 0.011788927014745433, + "grad_norm": 6.463813409208414, + "learning_rate": 8.674803888220268e-06, + "loss": 1.2708, + "step": 454 + }, + { + "epoch": 0.011814893814337382, + "grad_norm": 16.758537280363814, + "learning_rate": 8.677923564858893e-06, + "loss": 1.3361, + "step": 455 + }, + { + "epoch": 0.011840860613929333, + "grad_norm": 28.670275202302843, + "learning_rate": 8.681036392586107e-06, + "loss": 1.3323, + "step": 456 + }, + { + "epoch": 0.011866827413521283, + "grad_norm": 8.377091331702639, + "learning_rate": 8.684142401408131e-06, + "loss": 1.3586, + "step": 457 + }, + { + "epoch": 0.011892794213113234, + "grad_norm": 12.474321158062757, + "learning_rate": 8.687241621134421e-06, + "loss": 1.3726, + "step": 458 + }, + { + "epoch": 0.011918761012705184, + "grad_norm": 11.219645599915147, + "learning_rate": 8.690334081379388e-06, + "loss": 1.2291, + "step": 459 + }, + { + "epoch": 0.011944727812297135, + "grad_norm": 9.00940806011971, + "learning_rate": 8.693419811564099e-06, + "loss": 1.2881, + "step": 460 + }, + { + "epoch": 0.011970694611889085, + "grad_norm": 12.080513741387797, + "learning_rate": 8.696498840917945e-06, + "loss": 1.2973, + "step": 461 + }, + { + "epoch": 0.011996661411481035, + "grad_norm": 10.166901716006194, + "learning_rate": 8.699571198480317e-06, + "loss": 1.3592, + "step": 462 + }, + { + "epoch": 0.012022628211072986, + "grad_norm": 24.717390676514963, + "learning_rate": 8.702636913102236e-06, + "loss": 1.3226, + "step": 463 + }, + { + "epoch": 0.012048595010664935, + "grad_norm": 29.858104572986647, + "learning_rate": 8.705696013447989e-06, + "loss": 1.2975, + "step": 464 + }, + { + "epoch": 0.012074561810256885, + "grad_norm": 8.549380310088178, + "learning_rate": 8.708748527996734e-06, + "loss": 1.2562, + "step": 465 + }, + { + "epoch": 0.012100528609848836, + "grad_norm": 14.784844572342324, + "learning_rate": 8.711794485044082e-06, + "loss": 1.2887, + "step": 466 + }, + { + "epoch": 0.012126495409440786, + "grad_norm": 8.702597692873313, + "learning_rate": 8.714833912703688e-06, + "loss": 1.324, + "step": 467 + }, + { + "epoch": 0.012152462209032737, + "grad_norm": 8.740970421172273, + "learning_rate": 8.71786683890879e-06, + "loss": 1.2953, + "step": 468 + }, + { + "epoch": 0.012178429008624687, + "grad_norm": 14.943312786858824, + "learning_rate": 8.720893291413763e-06, + "loss": 1.334, + "step": 469 + }, + { + "epoch": 0.012204395808216638, + "grad_norm": 22.592041288689337, + "learning_rate": 8.723913297795636e-06, + "loss": 1.259, + "step": 470 + }, + { + "epoch": 0.012230362607808588, + "grad_norm": 7.939431263423161, + "learning_rate": 8.726926885455599e-06, + "loss": 1.3223, + "step": 471 + }, + { + "epoch": 0.012256329407400538, + "grad_norm": 14.375975467922059, + "learning_rate": 8.729934081620502e-06, + "loss": 1.3075, + "step": 472 + }, + { + "epoch": 0.012282296206992489, + "grad_norm": 7.938068279438846, + "learning_rate": 8.732934913344318e-06, + "loss": 1.2826, + "step": 473 + }, + { + "epoch": 0.012308263006584438, + "grad_norm": 11.73295298162169, + "learning_rate": 8.735929407509624e-06, + "loss": 1.3223, + "step": 474 + }, + { + "epoch": 0.012334229806176388, + "grad_norm": 34.04614716408818, + "learning_rate": 8.738917590829017e-06, + "loss": 1.3326, + "step": 475 + }, + { + "epoch": 0.012360196605768339, + "grad_norm": 7.837383711102742, + "learning_rate": 8.74189948984657e-06, + "loss": 1.289, + "step": 476 + }, + { + "epoch": 0.012386163405360289, + "grad_norm": 97.16839643436028, + "learning_rate": 8.744875130939237e-06, + "loss": 1.293, + "step": 477 + }, + { + "epoch": 0.01241213020495224, + "grad_norm": 7.972486868913614, + "learning_rate": 8.74784454031825e-06, + "loss": 1.3219, + "step": 478 + }, + { + "epoch": 0.01243809700454419, + "grad_norm": 9.091655686672258, + "learning_rate": 8.750807744030504e-06, + "loss": 1.3014, + "step": 479 + }, + { + "epoch": 0.01246406380413614, + "grad_norm": 12.356849383421743, + "learning_rate": 8.75376476795994e-06, + "loss": 1.3343, + "step": 480 + }, + { + "epoch": 0.012490030603728091, + "grad_norm": 11.799624576092189, + "learning_rate": 8.756715637828884e-06, + "loss": 1.3273, + "step": 481 + }, + { + "epoch": 0.012515997403320041, + "grad_norm": 11.711360300281424, + "learning_rate": 8.7596603791994e-06, + "loss": 1.3486, + "step": 482 + }, + { + "epoch": 0.012541964202911992, + "grad_norm": 6.304515963918772, + "learning_rate": 8.76259901747462e-06, + "loss": 1.3623, + "step": 483 + }, + { + "epoch": 0.01256793100250394, + "grad_norm": 10.630146889304193, + "learning_rate": 8.765531577900054e-06, + "loss": 1.3145, + "step": 484 + }, + { + "epoch": 0.012593897802095891, + "grad_norm": 8.559846118720351, + "learning_rate": 8.768458085564887e-06, + "loss": 1.2686, + "step": 485 + }, + { + "epoch": 0.012619864601687842, + "grad_norm": 12.44509425813039, + "learning_rate": 8.771378565403283e-06, + "loss": 1.2983, + "step": 486 + }, + { + "epoch": 0.012645831401279792, + "grad_norm": 7.424097888745131, + "learning_rate": 8.77429304219564e-06, + "loss": 1.3874, + "step": 487 + }, + { + "epoch": 0.012671798200871743, + "grad_norm": 19.253763202577762, + "learning_rate": 8.777201540569868e-06, + "loss": 1.363, + "step": 488 + }, + { + "epoch": 0.012697765000463693, + "grad_norm": 12.012439002796162, + "learning_rate": 8.780104085002626e-06, + "loss": 1.2677, + "step": 489 + }, + { + "epoch": 0.012723731800055644, + "grad_norm": 8.46542371902764, + "learning_rate": 8.783000699820565e-06, + "loss": 1.318, + "step": 490 + }, + { + "epoch": 0.012749698599647594, + "grad_norm": 31.918480033920495, + "learning_rate": 8.78589140920155e-06, + "loss": 1.2861, + "step": 491 + }, + { + "epoch": 0.012775665399239544, + "grad_norm": 9.0983791039578, + "learning_rate": 8.788776237175867e-06, + "loss": 1.3323, + "step": 492 + }, + { + "epoch": 0.012801632198831493, + "grad_norm": 34.21629022266386, + "learning_rate": 8.791655207627424e-06, + "loss": 1.257, + "step": 493 + }, + { + "epoch": 0.012827598998423444, + "grad_norm": 8.234301340544596, + "learning_rate": 8.794528344294943e-06, + "loss": 1.344, + "step": 494 + }, + { + "epoch": 0.012853565798015394, + "grad_norm": 8.278718842245176, + "learning_rate": 8.797395670773122e-06, + "loss": 1.3074, + "step": 495 + }, + { + "epoch": 0.012879532597607345, + "grad_norm": 8.319601825576319, + "learning_rate": 8.800257210513812e-06, + "loss": 1.3175, + "step": 496 + }, + { + "epoch": 0.012905499397199295, + "grad_norm": 7.7405290877214155, + "learning_rate": 8.803112986827153e-06, + "loss": 1.3357, + "step": 497 + }, + { + "epoch": 0.012931466196791246, + "grad_norm": 8.526207855661813, + "learning_rate": 8.805963022882731e-06, + "loss": 1.289, + "step": 498 + }, + { + "epoch": 0.012957432996383196, + "grad_norm": 16.773957840089313, + "learning_rate": 8.808807341710687e-06, + "loss": 1.3642, + "step": 499 + }, + { + "epoch": 0.012983399795975147, + "grad_norm": 21.823192697264833, + "learning_rate": 8.811645966202848e-06, + "loss": 1.3052, + "step": 500 + }, + { + "epoch": 0.013009366595567097, + "grad_norm": 5.908800008589215, + "learning_rate": 8.814478919113826e-06, + "loss": 1.3261, + "step": 501 + }, + { + "epoch": 0.013035333395159047, + "grad_norm": 7.658593142325667, + "learning_rate": 8.81730622306211e-06, + "loss": 1.3202, + "step": 502 + }, + { + "epoch": 0.013061300194750996, + "grad_norm": 9.109407244931138, + "learning_rate": 8.82012790053116e-06, + "loss": 1.3604, + "step": 503 + }, + { + "epoch": 0.013087266994342947, + "grad_norm": 8.423662155288833, + "learning_rate": 8.822943973870464e-06, + "loss": 1.3056, + "step": 504 + }, + { + "epoch": 0.013113233793934897, + "grad_norm": 6.343488716043891, + "learning_rate": 8.825754465296615e-06, + "loss": 1.3373, + "step": 505 + }, + { + "epoch": 0.013139200593526848, + "grad_norm": 6.476887325520665, + "learning_rate": 8.828559396894359e-06, + "loss": 1.2722, + "step": 506 + }, + { + "epoch": 0.013165167393118798, + "grad_norm": 8.869875865871887, + "learning_rate": 8.831358790617626e-06, + "loss": 1.361, + "step": 507 + }, + { + "epoch": 0.013191134192710749, + "grad_norm": 20.233810169481412, + "learning_rate": 8.834152668290574e-06, + "loss": 1.3159, + "step": 508 + }, + { + "epoch": 0.013217100992302699, + "grad_norm": 8.093723135060385, + "learning_rate": 8.836941051608605e-06, + "loss": 1.3335, + "step": 509 + }, + { + "epoch": 0.01324306779189465, + "grad_norm": 9.801160735852644, + "learning_rate": 8.839723962139376e-06, + "loss": 1.3044, + "step": 510 + }, + { + "epoch": 0.0132690345914866, + "grad_norm": 9.961510783135234, + "learning_rate": 8.842501421323793e-06, + "loss": 1.3411, + "step": 511 + }, + { + "epoch": 0.013295001391078549, + "grad_norm": 9.041207391234236, + "learning_rate": 8.845273450477018e-06, + "loss": 1.322, + "step": 512 + }, + { + "epoch": 0.0133209681906705, + "grad_norm": 7.125103063005283, + "learning_rate": 8.848040070789435e-06, + "loss": 1.283, + "step": 513 + }, + { + "epoch": 0.01334693499026245, + "grad_norm": 9.695936110710935, + "learning_rate": 8.850801303327631e-06, + "loss": 1.3237, + "step": 514 + }, + { + "epoch": 0.0133729017898544, + "grad_norm": 7.0853662902004295, + "learning_rate": 8.853557169035352e-06, + "loss": 1.2671, + "step": 515 + }, + { + "epoch": 0.01339886858944635, + "grad_norm": 16.186120876640057, + "learning_rate": 8.856307688734466e-06, + "loss": 1.3301, + "step": 516 + }, + { + "epoch": 0.013424835389038301, + "grad_norm": 21.04460710683975, + "learning_rate": 8.859052883125895e-06, + "loss": 1.3258, + "step": 517 + }, + { + "epoch": 0.013450802188630252, + "grad_norm": 11.375608323148247, + "learning_rate": 8.861792772790557e-06, + "loss": 1.325, + "step": 518 + }, + { + "epoch": 0.013476768988222202, + "grad_norm": 15.634870417551685, + "learning_rate": 8.864527378190292e-06, + "loss": 1.3483, + "step": 519 + }, + { + "epoch": 0.013502735787814153, + "grad_norm": 17.707983012006636, + "learning_rate": 8.867256719668776e-06, + "loss": 1.3044, + "step": 520 + }, + { + "epoch": 0.013528702587406103, + "grad_norm": 17.390560071195054, + "learning_rate": 8.869980817452429e-06, + "loss": 1.3198, + "step": 521 + }, + { + "epoch": 0.013554669386998052, + "grad_norm": 24.013850983599028, + "learning_rate": 8.872699691651319e-06, + "loss": 1.315, + "step": 522 + }, + { + "epoch": 0.013580636186590002, + "grad_norm": 13.636292423871524, + "learning_rate": 8.875413362260044e-06, + "loss": 1.3022, + "step": 523 + }, + { + "epoch": 0.013606602986181953, + "grad_norm": 13.332696772320904, + "learning_rate": 8.878121849158624e-06, + "loss": 1.3065, + "step": 524 + }, + { + "epoch": 0.013632569785773903, + "grad_norm": 12.810938595789073, + "learning_rate": 8.880825172113371e-06, + "loss": 1.3111, + "step": 525 + }, + { + "epoch": 0.013658536585365854, + "grad_norm": 11.527270946059486, + "learning_rate": 8.883523350777752e-06, + "loss": 1.3245, + "step": 526 + }, + { + "epoch": 0.013684503384957804, + "grad_norm": 8.155897344806606, + "learning_rate": 8.886216404693247e-06, + "loss": 1.3301, + "step": 527 + }, + { + "epoch": 0.013710470184549755, + "grad_norm": 8.968821455508033, + "learning_rate": 8.8889043532902e-06, + "loss": 1.3234, + "step": 528 + }, + { + "epoch": 0.013736436984141705, + "grad_norm": 9.720529473094802, + "learning_rate": 8.891587215888663e-06, + "loss": 1.3125, + "step": 529 + }, + { + "epoch": 0.013762403783733656, + "grad_norm": 17.16864199831793, + "learning_rate": 8.894265011699222e-06, + "loss": 1.288, + "step": 530 + }, + { + "epoch": 0.013788370583325604, + "grad_norm": 49.356201920009504, + "learning_rate": 8.89693775982383e-06, + "loss": 1.2833, + "step": 531 + }, + { + "epoch": 0.013814337382917555, + "grad_norm": 11.400028402728118, + "learning_rate": 8.899605479256616e-06, + "loss": 1.3399, + "step": 532 + }, + { + "epoch": 0.013840304182509505, + "grad_norm": 9.460091538097275, + "learning_rate": 8.902268188884702e-06, + "loss": 1.345, + "step": 533 + }, + { + "epoch": 0.013866270982101456, + "grad_norm": 6.814165944077656, + "learning_rate": 8.904925907489006e-06, + "loss": 1.3619, + "step": 534 + }, + { + "epoch": 0.013892237781693406, + "grad_norm": 15.578239601195818, + "learning_rate": 8.907578653745027e-06, + "loss": 1.2911, + "step": 535 + }, + { + "epoch": 0.013918204581285357, + "grad_norm": 10.999393446411709, + "learning_rate": 8.910226446223646e-06, + "loss": 1.3281, + "step": 536 + }, + { + "epoch": 0.013944171380877307, + "grad_norm": 15.10227179205996, + "learning_rate": 8.912869303391895e-06, + "loss": 1.3268, + "step": 537 + }, + { + "epoch": 0.013970138180469258, + "grad_norm": 7.852517626917796, + "learning_rate": 8.915507243613737e-06, + "loss": 1.3016, + "step": 538 + }, + { + "epoch": 0.013996104980061208, + "grad_norm": 12.065995890764157, + "learning_rate": 8.918140285150826e-06, + "loss": 1.3555, + "step": 539 + }, + { + "epoch": 0.014022071779653159, + "grad_norm": 8.603419187527049, + "learning_rate": 8.92076844616327e-06, + "loss": 1.2917, + "step": 540 + }, + { + "epoch": 0.014048038579245107, + "grad_norm": 11.660028658221508, + "learning_rate": 8.923391744710377e-06, + "loss": 1.3089, + "step": 541 + }, + { + "epoch": 0.014074005378837058, + "grad_norm": 18.65551797270926, + "learning_rate": 8.926010198751412e-06, + "loss": 1.336, + "step": 542 + }, + { + "epoch": 0.014099972178429008, + "grad_norm": 7.000661818689587, + "learning_rate": 8.928623826146315e-06, + "loss": 1.3511, + "step": 543 + }, + { + "epoch": 0.014125938978020959, + "grad_norm": 15.827903753543781, + "learning_rate": 8.931232644656452e-06, + "loss": 1.2936, + "step": 544 + }, + { + "epoch": 0.01415190577761291, + "grad_norm": 13.601446388971663, + "learning_rate": 8.933836671945326e-06, + "loss": 1.3131, + "step": 545 + }, + { + "epoch": 0.01417787257720486, + "grad_norm": 83.19046705358583, + "learning_rate": 8.936435925579299e-06, + "loss": 1.2495, + "step": 546 + }, + { + "epoch": 0.01420383937679681, + "grad_norm": 23.934707837318854, + "learning_rate": 8.9390304230283e-06, + "loss": 1.3317, + "step": 547 + }, + { + "epoch": 0.01422980617638876, + "grad_norm": 11.469646699447907, + "learning_rate": 8.941620181666542e-06, + "loss": 1.2713, + "step": 548 + }, + { + "epoch": 0.014255772975980711, + "grad_norm": 9.729578002686013, + "learning_rate": 8.944205218773196e-06, + "loss": 1.311, + "step": 549 + }, + { + "epoch": 0.01428173977557266, + "grad_norm": 12.578228114872365, + "learning_rate": 8.946785551533108e-06, + "loss": 1.286, + "step": 550 + }, + { + "epoch": 0.01430770657516461, + "grad_norm": 38.60564088073788, + "learning_rate": 8.949361197037471e-06, + "loss": 1.3017, + "step": 551 + }, + { + "epoch": 0.01433367337475656, + "grad_norm": 11.08377839153315, + "learning_rate": 8.951932172284505e-06, + "loss": 1.3343, + "step": 552 + }, + { + "epoch": 0.014359640174348511, + "grad_norm": 19.78722957732343, + "learning_rate": 8.954498494180131e-06, + "loss": 1.2972, + "step": 553 + }, + { + "epoch": 0.014385606973940462, + "grad_norm": 16.061551908425017, + "learning_rate": 8.957060179538645e-06, + "loss": 1.2834, + "step": 554 + }, + { + "epoch": 0.014411573773532412, + "grad_norm": 10.370988920727264, + "learning_rate": 8.959617245083364e-06, + "loss": 1.2948, + "step": 555 + }, + { + "epoch": 0.014437540573124363, + "grad_norm": 11.34506344087767, + "learning_rate": 8.962169707447295e-06, + "loss": 1.3255, + "step": 556 + }, + { + "epoch": 0.014463507372716313, + "grad_norm": 9.757952845643032, + "learning_rate": 8.964717583173782e-06, + "loss": 1.2764, + "step": 557 + }, + { + "epoch": 0.014489474172308264, + "grad_norm": 13.268599788841524, + "learning_rate": 8.967260888717138e-06, + "loss": 1.3444, + "step": 558 + }, + { + "epoch": 0.014515440971900214, + "grad_norm": 19.686153692806446, + "learning_rate": 8.969799640443301e-06, + "loss": 1.3064, + "step": 559 + }, + { + "epoch": 0.014541407771492163, + "grad_norm": 22.32061397788173, + "learning_rate": 8.972333854630449e-06, + "loss": 1.3004, + "step": 560 + }, + { + "epoch": 0.014567374571084113, + "grad_norm": 12.073207076065684, + "learning_rate": 8.974863547469635e-06, + "loss": 1.3336, + "step": 561 + }, + { + "epoch": 0.014593341370676064, + "grad_norm": 8.93231477576882, + "learning_rate": 8.977388735065404e-06, + "loss": 1.3105, + "step": 562 + }, + { + "epoch": 0.014619308170268014, + "grad_norm": 9.706787204552938, + "learning_rate": 8.979909433436415e-06, + "loss": 1.337, + "step": 563 + }, + { + "epoch": 0.014645274969859965, + "grad_norm": 8.49702940250251, + "learning_rate": 8.982425658516042e-06, + "loss": 1.3846, + "step": 564 + }, + { + "epoch": 0.014671241769451915, + "grad_norm": 12.46072932819845, + "learning_rate": 8.984937426152981e-06, + "loss": 1.302, + "step": 565 + }, + { + "epoch": 0.014697208569043866, + "grad_norm": 11.22518048045766, + "learning_rate": 8.987444752111852e-06, + "loss": 1.3049, + "step": 566 + }, + { + "epoch": 0.014723175368635816, + "grad_norm": 9.653347134589065, + "learning_rate": 8.98994765207379e-06, + "loss": 1.3146, + "step": 567 + }, + { + "epoch": 0.014749142168227767, + "grad_norm": 12.618630190351858, + "learning_rate": 8.992446141637037e-06, + "loss": 1.3258, + "step": 568 + }, + { + "epoch": 0.014775108967819717, + "grad_norm": 9.694785252081882, + "learning_rate": 8.994940236317517e-06, + "loss": 1.3053, + "step": 569 + }, + { + "epoch": 0.014801075767411666, + "grad_norm": 14.601681630733923, + "learning_rate": 8.997429951549422e-06, + "loss": 1.2976, + "step": 570 + }, + { + "epoch": 0.014827042567003616, + "grad_norm": 8.100937645412614, + "learning_rate": 8.999915302685781e-06, + "loss": 1.2719, + "step": 571 + }, + { + "epoch": 0.014853009366595567, + "grad_norm": 9.764920287073041, + "learning_rate": 9.002396304999035e-06, + "loss": 1.3126, + "step": 572 + }, + { + "epoch": 0.014878976166187517, + "grad_norm": 8.66681871525339, + "learning_rate": 9.004872973681589e-06, + "loss": 1.3199, + "step": 573 + }, + { + "epoch": 0.014904942965779468, + "grad_norm": 12.930710637389216, + "learning_rate": 9.007345323846377e-06, + "loss": 1.2854, + "step": 574 + }, + { + "epoch": 0.014930909765371418, + "grad_norm": 14.495094258691779, + "learning_rate": 9.009813370527414e-06, + "loss": 1.2773, + "step": 575 + }, + { + "epoch": 0.014956876564963369, + "grad_norm": 7.94574363536512, + "learning_rate": 9.012277128680346e-06, + "loss": 1.2841, + "step": 576 + }, + { + "epoch": 0.014982843364555319, + "grad_norm": 7.950306810900519, + "learning_rate": 9.014736613182994e-06, + "loss": 1.274, + "step": 577 + }, + { + "epoch": 0.01500881016414727, + "grad_norm": 10.119029981239827, + "learning_rate": 9.017191838835889e-06, + "loss": 1.3142, + "step": 578 + }, + { + "epoch": 0.015034776963739218, + "grad_norm": 9.057933840910332, + "learning_rate": 9.019642820362806e-06, + "loss": 1.3321, + "step": 579 + }, + { + "epoch": 0.015060743763331169, + "grad_norm": 8.03784099229282, + "learning_rate": 9.022089572411304e-06, + "loss": 1.2983, + "step": 580 + }, + { + "epoch": 0.01508671056292312, + "grad_norm": 10.008949090171678, + "learning_rate": 9.02453210955324e-06, + "loss": 1.3003, + "step": 581 + }, + { + "epoch": 0.01511267736251507, + "grad_norm": 10.371568594944012, + "learning_rate": 9.026970446285293e-06, + "loss": 1.3236, + "step": 582 + }, + { + "epoch": 0.01513864416210702, + "grad_norm": 11.90331674723314, + "learning_rate": 9.029404597029482e-06, + "loss": 1.2916, + "step": 583 + }, + { + "epoch": 0.01516461096169897, + "grad_norm": 11.661498169222565, + "learning_rate": 9.031834576133676e-06, + "loss": 1.2984, + "step": 584 + }, + { + "epoch": 0.015190577761290921, + "grad_norm": 10.018358164149243, + "learning_rate": 9.034260397872104e-06, + "loss": 1.3306, + "step": 585 + }, + { + "epoch": 0.015216544560882872, + "grad_norm": 13.27678693393332, + "learning_rate": 9.036682076445854e-06, + "loss": 1.3203, + "step": 586 + }, + { + "epoch": 0.015242511360474822, + "grad_norm": 15.120002866875916, + "learning_rate": 9.039099625983376e-06, + "loss": 1.3284, + "step": 587 + }, + { + "epoch": 0.015268478160066773, + "grad_norm": 17.973215307948085, + "learning_rate": 9.041513060540972e-06, + "loss": 1.3476, + "step": 588 + }, + { + "epoch": 0.015294444959658721, + "grad_norm": 8.089184036506083, + "learning_rate": 9.043922394103292e-06, + "loss": 1.3123, + "step": 589 + }, + { + "epoch": 0.015320411759250672, + "grad_norm": 9.2648085544856, + "learning_rate": 9.046327640583815e-06, + "loss": 1.2739, + "step": 590 + }, + { + "epoch": 0.015346378558842622, + "grad_norm": 9.459103545927496, + "learning_rate": 9.048728813825338e-06, + "loss": 1.334, + "step": 591 + }, + { + "epoch": 0.015372345358434573, + "grad_norm": 23.601991782755988, + "learning_rate": 9.05112592760044e-06, + "loss": 1.3188, + "step": 592 + }, + { + "epoch": 0.015398312158026523, + "grad_norm": 7.51043991811877, + "learning_rate": 9.053518995611976e-06, + "loss": 1.3281, + "step": 593 + }, + { + "epoch": 0.015424278957618474, + "grad_norm": 22.389406478532152, + "learning_rate": 9.055908031493528e-06, + "loss": 1.3311, + "step": 594 + }, + { + "epoch": 0.015450245757210424, + "grad_norm": 10.556096229314086, + "learning_rate": 9.058293048809884e-06, + "loss": 1.298, + "step": 595 + }, + { + "epoch": 0.015476212556802375, + "grad_norm": 7.762648075184148, + "learning_rate": 9.060674061057494e-06, + "loss": 1.3115, + "step": 596 + }, + { + "epoch": 0.015502179356394325, + "grad_norm": 47.315583102154356, + "learning_rate": 9.06305108166493e-06, + "loss": 1.328, + "step": 597 + }, + { + "epoch": 0.015528146155986274, + "grad_norm": 18.333856933514937, + "learning_rate": 9.06542412399334e-06, + "loss": 1.2809, + "step": 598 + }, + { + "epoch": 0.015554112955578224, + "grad_norm": 7.2518300344724675, + "learning_rate": 9.067793201336898e-06, + "loss": 1.2767, + "step": 599 + }, + { + "epoch": 0.015580079755170175, + "grad_norm": 8.965675172502522, + "learning_rate": 9.070158326923256e-06, + "loss": 1.3126, + "step": 600 + }, + { + "epoch": 0.015606046554762125, + "grad_norm": 15.710259573265331, + "learning_rate": 9.072519513913973e-06, + "loss": 1.2567, + "step": 601 + }, + { + "epoch": 0.015632013354354077, + "grad_norm": 8.087489000457406, + "learning_rate": 9.074876775404974e-06, + "loss": 1.2956, + "step": 602 + }, + { + "epoch": 0.015657980153946024, + "grad_norm": 7.332846220773334, + "learning_rate": 9.077230124426974e-06, + "loss": 1.2562, + "step": 603 + }, + { + "epoch": 0.015683946953537975, + "grad_norm": 8.517217812618046, + "learning_rate": 9.07957957394591e-06, + "loss": 1.3486, + "step": 604 + }, + { + "epoch": 0.015709913753129925, + "grad_norm": 16.28920207524134, + "learning_rate": 9.081925136863369e-06, + "loss": 1.3155, + "step": 605 + }, + { + "epoch": 0.015735880552721876, + "grad_norm": 17.08201204478111, + "learning_rate": 9.08426682601702e-06, + "loss": 1.2697, + "step": 606 + }, + { + "epoch": 0.015761847352313826, + "grad_norm": 7.118860376857067, + "learning_rate": 9.086604654181038e-06, + "loss": 1.3118, + "step": 607 + }, + { + "epoch": 0.015787814151905777, + "grad_norm": 11.266422407523432, + "learning_rate": 9.0889386340665e-06, + "loss": 1.2964, + "step": 608 + }, + { + "epoch": 0.015813780951497727, + "grad_norm": 11.018286676992332, + "learning_rate": 9.091268778321827e-06, + "loss": 1.3387, + "step": 609 + }, + { + "epoch": 0.015839747751089678, + "grad_norm": 7.319764009851145, + "learning_rate": 9.093595099533182e-06, + "loss": 1.3338, + "step": 610 + }, + { + "epoch": 0.01586571455068163, + "grad_norm": 9.811724784017573, + "learning_rate": 9.095917610224879e-06, + "loss": 1.3304, + "step": 611 + }, + { + "epoch": 0.01589168135027358, + "grad_norm": 10.221337413733616, + "learning_rate": 9.098236322859781e-06, + "loss": 1.3093, + "step": 612 + }, + { + "epoch": 0.01591764814986553, + "grad_norm": 9.241624906931149, + "learning_rate": 9.100551249839722e-06, + "loss": 1.3273, + "step": 613 + }, + { + "epoch": 0.01594361494945748, + "grad_norm": 8.14333821073124, + "learning_rate": 9.102862403505876e-06, + "loss": 1.2895, + "step": 614 + }, + { + "epoch": 0.01596958174904943, + "grad_norm": 8.330253744118858, + "learning_rate": 9.105169796139182e-06, + "loss": 1.3284, + "step": 615 + }, + { + "epoch": 0.01599554854864138, + "grad_norm": 14.354473286063564, + "learning_rate": 9.107473439960709e-06, + "loss": 1.2391, + "step": 616 + }, + { + "epoch": 0.01602151534823333, + "grad_norm": 9.313910665492, + "learning_rate": 9.109773347132062e-06, + "loss": 1.3036, + "step": 617 + }, + { + "epoch": 0.01604748214782528, + "grad_norm": 14.74312004370333, + "learning_rate": 9.112069529755758e-06, + "loss": 1.2741, + "step": 618 + }, + { + "epoch": 0.016073448947417232, + "grad_norm": 8.355429486525326, + "learning_rate": 9.114361999875616e-06, + "loss": 1.3031, + "step": 619 + }, + { + "epoch": 0.016099415747009183, + "grad_norm": 80.22214640258325, + "learning_rate": 9.116650769477127e-06, + "loss": 1.324, + "step": 620 + }, + { + "epoch": 0.016125382546601133, + "grad_norm": 11.7324230129952, + "learning_rate": 9.118935850487833e-06, + "loss": 1.3167, + "step": 621 + }, + { + "epoch": 0.01615134934619308, + "grad_norm": 10.691009480672081, + "learning_rate": 9.121217254777707e-06, + "loss": 1.3022, + "step": 622 + }, + { + "epoch": 0.01617731614578503, + "grad_norm": 11.96406579049354, + "learning_rate": 9.123494994159514e-06, + "loss": 1.3093, + "step": 623 + }, + { + "epoch": 0.01620328294537698, + "grad_norm": 8.40092759233295, + "learning_rate": 9.125769080389182e-06, + "loss": 1.2821, + "step": 624 + }, + { + "epoch": 0.01622924974496893, + "grad_norm": 12.362364820892633, + "learning_rate": 9.128039525166163e-06, + "loss": 1.2593, + "step": 625 + }, + { + "epoch": 0.016255216544560882, + "grad_norm": 11.000279509721205, + "learning_rate": 9.130306340133801e-06, + "loss": 1.2966, + "step": 626 + }, + { + "epoch": 0.016281183344152832, + "grad_norm": 10.213677410110082, + "learning_rate": 9.132569536879682e-06, + "loss": 1.2504, + "step": 627 + }, + { + "epoch": 0.016307150143744783, + "grad_norm": 18.052368279211493, + "learning_rate": 9.134829126935992e-06, + "loss": 1.263, + "step": 628 + }, + { + "epoch": 0.016333116943336733, + "grad_norm": 11.542408245697757, + "learning_rate": 9.137085121779877e-06, + "loss": 1.2882, + "step": 629 + }, + { + "epoch": 0.016359083742928684, + "grad_norm": 25.364351791379864, + "learning_rate": 9.139337532833777e-06, + "loss": 1.3332, + "step": 630 + }, + { + "epoch": 0.016385050542520634, + "grad_norm": 11.710337475247231, + "learning_rate": 9.141586371465793e-06, + "loss": 1.2768, + "step": 631 + }, + { + "epoch": 0.016411017342112585, + "grad_norm": 7.828728262371525, + "learning_rate": 9.143831648990016e-06, + "loss": 1.2837, + "step": 632 + }, + { + "epoch": 0.016436984141704535, + "grad_norm": 7.066480207082579, + "learning_rate": 9.146073376666868e-06, + "loss": 1.2865, + "step": 633 + }, + { + "epoch": 0.016462950941296486, + "grad_norm": 8.120703867204377, + "learning_rate": 9.148311565703455e-06, + "loss": 1.2979, + "step": 634 + }, + { + "epoch": 0.016488917740888436, + "grad_norm": 27.393370156218914, + "learning_rate": 9.150546227253889e-06, + "loss": 1.3217, + "step": 635 + }, + { + "epoch": 0.016514884540480387, + "grad_norm": 8.034100334728, + "learning_rate": 9.152777372419628e-06, + "loss": 1.2805, + "step": 636 + }, + { + "epoch": 0.016540851340072337, + "grad_norm": 12.914565245620869, + "learning_rate": 9.155005012249807e-06, + "loss": 1.2925, + "step": 637 + }, + { + "epoch": 0.016566818139664288, + "grad_norm": 10.3796041912254, + "learning_rate": 9.157229157741564e-06, + "loss": 1.2607, + "step": 638 + }, + { + "epoch": 0.016592784939256238, + "grad_norm": 13.795358645791548, + "learning_rate": 9.159449819840365e-06, + "loss": 1.3279, + "step": 639 + }, + { + "epoch": 0.01661875173884819, + "grad_norm": 9.413164911177299, + "learning_rate": 9.161667009440333e-06, + "loss": 1.2971, + "step": 640 + }, + { + "epoch": 0.016644718538440136, + "grad_norm": 9.72507361847537, + "learning_rate": 9.16388073738456e-06, + "loss": 1.3089, + "step": 641 + }, + { + "epoch": 0.016670685338032086, + "grad_norm": 7.724865919017422, + "learning_rate": 9.166091014465433e-06, + "loss": 1.2698, + "step": 642 + }, + { + "epoch": 0.016696652137624036, + "grad_norm": 7.844480928811063, + "learning_rate": 9.168297851424947e-06, + "loss": 1.2959, + "step": 643 + }, + { + "epoch": 0.016722618937215987, + "grad_norm": 10.170541702292283, + "learning_rate": 9.170501258955015e-06, + "loss": 1.2719, + "step": 644 + }, + { + "epoch": 0.016748585736807937, + "grad_norm": 8.59401056068061, + "learning_rate": 9.172701247697781e-06, + "loss": 1.3337, + "step": 645 + }, + { + "epoch": 0.016774552536399888, + "grad_norm": 8.043001950210739, + "learning_rate": 9.174897828245934e-06, + "loss": 1.2882, + "step": 646 + }, + { + "epoch": 0.01680051933599184, + "grad_norm": 8.777179901535902, + "learning_rate": 9.177091011143006e-06, + "loss": 1.3435, + "step": 647 + }, + { + "epoch": 0.01682648613558379, + "grad_norm": 33.96567553071501, + "learning_rate": 9.179280806883674e-06, + "loss": 1.2949, + "step": 648 + }, + { + "epoch": 0.01685245293517574, + "grad_norm": 28.6817187631937, + "learning_rate": 9.181467225914077e-06, + "loss": 1.3144, + "step": 649 + }, + { + "epoch": 0.01687841973476769, + "grad_norm": 9.929303321639692, + "learning_rate": 9.18365027863209e-06, + "loss": 1.3016, + "step": 650 + }, + { + "epoch": 0.01690438653435964, + "grad_norm": 14.84143845077309, + "learning_rate": 9.185829975387648e-06, + "loss": 1.3121, + "step": 651 + }, + { + "epoch": 0.01693035333395159, + "grad_norm": 11.151662700855525, + "learning_rate": 9.188006326483019e-06, + "loss": 1.3064, + "step": 652 + }, + { + "epoch": 0.01695632013354354, + "grad_norm": 11.958299661161398, + "learning_rate": 9.190179342173104e-06, + "loss": 1.3112, + "step": 653 + }, + { + "epoch": 0.01698228693313549, + "grad_norm": 9.915681891604242, + "learning_rate": 9.192349032665733e-06, + "loss": 1.3032, + "step": 654 + }, + { + "epoch": 0.017008253732727442, + "grad_norm": 37.723450476067434, + "learning_rate": 9.194515408121939e-06, + "loss": 1.267, + "step": 655 + }, + { + "epoch": 0.017034220532319393, + "grad_norm": 9.993901109537108, + "learning_rate": 9.19667847865626e-06, + "loss": 1.2801, + "step": 656 + }, + { + "epoch": 0.017060187331911343, + "grad_norm": 9.149354805633838, + "learning_rate": 9.198838254337004e-06, + "loss": 1.2861, + "step": 657 + }, + { + "epoch": 0.017086154131503294, + "grad_norm": 9.15138047440944, + "learning_rate": 9.200994745186552e-06, + "loss": 1.2721, + "step": 658 + }, + { + "epoch": 0.017112120931095244, + "grad_norm": 19.171036108887836, + "learning_rate": 9.20314796118161e-06, + "loss": 1.2741, + "step": 659 + }, + { + "epoch": 0.01713808773068719, + "grad_norm": 27.297866637325686, + "learning_rate": 9.205297912253515e-06, + "loss": 1.2858, + "step": 660 + }, + { + "epoch": 0.01716405453027914, + "grad_norm": 21.454637800585143, + "learning_rate": 9.207444608288486e-06, + "loss": 1.3048, + "step": 661 + }, + { + "epoch": 0.017190021329871092, + "grad_norm": 6.771246074909877, + "learning_rate": 9.209588059127914e-06, + "loss": 1.3081, + "step": 662 + }, + { + "epoch": 0.017215988129463042, + "grad_norm": 12.409302919592031, + "learning_rate": 9.211728274568617e-06, + "loss": 1.289, + "step": 663 + }, + { + "epoch": 0.017241954929054993, + "grad_norm": 14.705951674354077, + "learning_rate": 9.213865264363124e-06, + "loss": 1.2987, + "step": 664 + }, + { + "epoch": 0.017267921728646943, + "grad_norm": 8.296533380002053, + "learning_rate": 9.215999038219931e-06, + "loss": 1.269, + "step": 665 + }, + { + "epoch": 0.017293888528238894, + "grad_norm": 10.64632044882743, + "learning_rate": 9.21812960580377e-06, + "loss": 1.3025, + "step": 666 + }, + { + "epoch": 0.017319855327830844, + "grad_norm": 13.388991408705243, + "learning_rate": 9.220256976735869e-06, + "loss": 1.2994, + "step": 667 + }, + { + "epoch": 0.017345822127422795, + "grad_norm": 10.900458341988639, + "learning_rate": 9.222381160594218e-06, + "loss": 1.2921, + "step": 668 + }, + { + "epoch": 0.017371788927014745, + "grad_norm": 9.746180642147722, + "learning_rate": 9.224502166913824e-06, + "loss": 1.3148, + "step": 669 + }, + { + "epoch": 0.017397755726606696, + "grad_norm": 10.674548058691956, + "learning_rate": 9.226620005186961e-06, + "loss": 1.2596, + "step": 670 + }, + { + "epoch": 0.017423722526198646, + "grad_norm": 40.8701437768911, + "learning_rate": 9.228734684863441e-06, + "loss": 1.3363, + "step": 671 + }, + { + "epoch": 0.017449689325790597, + "grad_norm": 6.747756019871416, + "learning_rate": 9.230846215350855e-06, + "loss": 1.304, + "step": 672 + }, + { + "epoch": 0.017475656125382547, + "grad_norm": 19.778278434979182, + "learning_rate": 9.232954606014825e-06, + "loss": 1.3071, + "step": 673 + }, + { + "epoch": 0.017501622924974498, + "grad_norm": 8.453818132395858, + "learning_rate": 9.235059866179255e-06, + "loss": 1.3355, + "step": 674 + }, + { + "epoch": 0.017527589724566448, + "grad_norm": 6.875251904120836, + "learning_rate": 9.237162005126584e-06, + "loss": 1.3299, + "step": 675 + }, + { + "epoch": 0.0175535565241584, + "grad_norm": 7.727957604396621, + "learning_rate": 9.239261032098019e-06, + "loss": 1.2462, + "step": 676 + }, + { + "epoch": 0.01757952332375035, + "grad_norm": 10.502004522900542, + "learning_rate": 9.24135695629379e-06, + "loss": 1.2725, + "step": 677 + }, + { + "epoch": 0.0176054901233423, + "grad_norm": 9.042296310041701, + "learning_rate": 9.243449786873387e-06, + "loss": 1.3239, + "step": 678 + }, + { + "epoch": 0.01763145692293425, + "grad_norm": 15.823222147554505, + "learning_rate": 9.245539532955803e-06, + "loss": 1.3487, + "step": 679 + }, + { + "epoch": 0.017657423722526197, + "grad_norm": 8.707504403362098, + "learning_rate": 9.247626203619767e-06, + "loss": 1.2611, + "step": 680 + }, + { + "epoch": 0.017683390522118148, + "grad_norm": 20.351384806086063, + "learning_rate": 9.249709807903988e-06, + "loss": 1.2986, + "step": 681 + }, + { + "epoch": 0.017709357321710098, + "grad_norm": 8.237904322517537, + "learning_rate": 9.251790354807385e-06, + "loss": 1.2947, + "step": 682 + }, + { + "epoch": 0.01773532412130205, + "grad_norm": 13.219234232228562, + "learning_rate": 9.253867853289318e-06, + "loss": 1.2485, + "step": 683 + }, + { + "epoch": 0.017761290920894, + "grad_norm": 6.886971056336054, + "learning_rate": 9.255942312269827e-06, + "loss": 1.2678, + "step": 684 + }, + { + "epoch": 0.01778725772048595, + "grad_norm": 22.48223718337504, + "learning_rate": 9.258013740629857e-06, + "loss": 1.3017, + "step": 685 + }, + { + "epoch": 0.0178132245200779, + "grad_norm": 8.939413226027959, + "learning_rate": 9.26008214721148e-06, + "loss": 1.307, + "step": 686 + }, + { + "epoch": 0.01783919131966985, + "grad_norm": 7.990898677240369, + "learning_rate": 9.26214754081814e-06, + "loss": 1.3052, + "step": 687 + }, + { + "epoch": 0.0178651581192618, + "grad_norm": 9.292740923673259, + "learning_rate": 9.264209930214859e-06, + "loss": 1.2895, + "step": 688 + }, + { + "epoch": 0.01789112491885375, + "grad_norm": 44.645794295120204, + "learning_rate": 9.266269324128465e-06, + "loss": 1.2751, + "step": 689 + }, + { + "epoch": 0.017917091718445702, + "grad_norm": 16.521185019342873, + "learning_rate": 9.26832573124782e-06, + "loss": 1.3404, + "step": 690 + }, + { + "epoch": 0.017943058518037652, + "grad_norm": 15.21818751366837, + "learning_rate": 9.270379160224033e-06, + "loss": 1.3005, + "step": 691 + }, + { + "epoch": 0.017969025317629603, + "grad_norm": 15.675672715640966, + "learning_rate": 9.272429619670684e-06, + "loss": 1.325, + "step": 692 + }, + { + "epoch": 0.017994992117221553, + "grad_norm": 14.719263152877996, + "learning_rate": 9.274477118164037e-06, + "loss": 1.2533, + "step": 693 + }, + { + "epoch": 0.018020958916813504, + "grad_norm": 12.439264066951218, + "learning_rate": 9.276521664243253e-06, + "loss": 1.3126, + "step": 694 + }, + { + "epoch": 0.018046925716405454, + "grad_norm": 9.623182248959004, + "learning_rate": 9.27856326641061e-06, + "loss": 1.3116, + "step": 695 + }, + { + "epoch": 0.018072892515997405, + "grad_norm": 9.49080373465731, + "learning_rate": 9.280601933131712e-06, + "loss": 1.2609, + "step": 696 + }, + { + "epoch": 0.018098859315589355, + "grad_norm": 12.38593648980537, + "learning_rate": 9.282637672835695e-06, + "loss": 1.308, + "step": 697 + }, + { + "epoch": 0.018124826115181306, + "grad_norm": 13.301384095185593, + "learning_rate": 9.284670493915446e-06, + "loss": 1.3239, + "step": 698 + }, + { + "epoch": 0.018150792914773253, + "grad_norm": 15.319515193907645, + "learning_rate": 9.286700404727803e-06, + "loss": 1.2822, + "step": 699 + }, + { + "epoch": 0.018176759714365203, + "grad_norm": 33.45554419673965, + "learning_rate": 9.288727413593764e-06, + "loss": 1.29, + "step": 700 + }, + { + "epoch": 0.018202726513957154, + "grad_norm": 8.482947354132778, + "learning_rate": 9.29075152879869e-06, + "loss": 1.2667, + "step": 701 + }, + { + "epoch": 0.018228693313549104, + "grad_norm": 8.608671349988299, + "learning_rate": 9.29277275859251e-06, + "loss": 1.299, + "step": 702 + }, + { + "epoch": 0.018254660113141054, + "grad_norm": 8.489157549394257, + "learning_rate": 9.294791111189922e-06, + "loss": 1.2708, + "step": 703 + }, + { + "epoch": 0.018280626912733005, + "grad_norm": 14.702670958883468, + "learning_rate": 9.296806594770593e-06, + "loss": 1.273, + "step": 704 + }, + { + "epoch": 0.018306593712324955, + "grad_norm": 8.877941235228574, + "learning_rate": 9.298819217479357e-06, + "loss": 1.3226, + "step": 705 + }, + { + "epoch": 0.018332560511916906, + "grad_norm": 13.776428901792105, + "learning_rate": 9.300828987426414e-06, + "loss": 1.2911, + "step": 706 + }, + { + "epoch": 0.018358527311508856, + "grad_norm": 30.326219108713296, + "learning_rate": 9.30283591268753e-06, + "loss": 1.2732, + "step": 707 + }, + { + "epoch": 0.018384494111100807, + "grad_norm": 10.18200797671993, + "learning_rate": 9.304840001304221e-06, + "loss": 1.3039, + "step": 708 + }, + { + "epoch": 0.018410460910692757, + "grad_norm": 14.94280324351539, + "learning_rate": 9.30684126128396e-06, + "loss": 1.3127, + "step": 709 + }, + { + "epoch": 0.018436427710284708, + "grad_norm": 9.642062038358937, + "learning_rate": 9.308839700600353e-06, + "loss": 1.2957, + "step": 710 + }, + { + "epoch": 0.01846239450987666, + "grad_norm": 88.29279238703462, + "learning_rate": 9.310835327193344e-06, + "loss": 1.2355, + "step": 711 + }, + { + "epoch": 0.01848836130946861, + "grad_norm": 111.35942652061387, + "learning_rate": 9.312828148969399e-06, + "loss": 1.2957, + "step": 712 + }, + { + "epoch": 0.01851432810906056, + "grad_norm": 11.701025658669575, + "learning_rate": 9.31481817380169e-06, + "loss": 1.3128, + "step": 713 + }, + { + "epoch": 0.01854029490865251, + "grad_norm": 11.232881534293707, + "learning_rate": 9.31680540953029e-06, + "loss": 1.2767, + "step": 714 + }, + { + "epoch": 0.01856626170824446, + "grad_norm": 10.595722026192819, + "learning_rate": 9.31878986396235e-06, + "loss": 1.36, + "step": 715 + }, + { + "epoch": 0.01859222850783641, + "grad_norm": 8.54474531067582, + "learning_rate": 9.320771544872288e-06, + "loss": 1.2775, + "step": 716 + }, + { + "epoch": 0.01861819530742836, + "grad_norm": 13.221015754031889, + "learning_rate": 9.322750460001969e-06, + "loss": 1.3249, + "step": 717 + }, + { + "epoch": 0.018644162107020308, + "grad_norm": 8.739058711083347, + "learning_rate": 9.32472661706089e-06, + "loss": 1.3474, + "step": 718 + }, + { + "epoch": 0.01867012890661226, + "grad_norm": 6.732240272155417, + "learning_rate": 9.326700023726357e-06, + "loss": 1.2855, + "step": 719 + }, + { + "epoch": 0.01869609570620421, + "grad_norm": 9.009528308507436, + "learning_rate": 9.328670687643661e-06, + "loss": 1.3067, + "step": 720 + }, + { + "epoch": 0.01872206250579616, + "grad_norm": 10.285638029330517, + "learning_rate": 9.330638616426268e-06, + "loss": 1.3101, + "step": 721 + }, + { + "epoch": 0.01874802930538811, + "grad_norm": 11.14549715955254, + "learning_rate": 9.33260381765598e-06, + "loss": 1.3528, + "step": 722 + }, + { + "epoch": 0.01877399610498006, + "grad_norm": 12.534631591544372, + "learning_rate": 9.334566298883122e-06, + "loss": 1.312, + "step": 723 + }, + { + "epoch": 0.01879996290457201, + "grad_norm": 13.551120748305516, + "learning_rate": 9.336526067626708e-06, + "loss": 1.3116, + "step": 724 + }, + { + "epoch": 0.01882592970416396, + "grad_norm": 8.36327532037973, + "learning_rate": 9.33848313137462e-06, + "loss": 1.3082, + "step": 725 + }, + { + "epoch": 0.018851896503755912, + "grad_norm": 9.228729122861095, + "learning_rate": 9.340437497583775e-06, + "loss": 1.252, + "step": 726 + }, + { + "epoch": 0.018877863303347862, + "grad_norm": 39.62267713025901, + "learning_rate": 9.342389173680298e-06, + "loss": 1.2739, + "step": 727 + }, + { + "epoch": 0.018903830102939813, + "grad_norm": 48.02572196961879, + "learning_rate": 9.344338167059692e-06, + "loss": 1.2428, + "step": 728 + }, + { + "epoch": 0.018929796902531763, + "grad_norm": 7.551323821362591, + "learning_rate": 9.346284485087002e-06, + "loss": 1.2825, + "step": 729 + }, + { + "epoch": 0.018955763702123714, + "grad_norm": 10.56927597929785, + "learning_rate": 9.348228135096991e-06, + "loss": 1.3145, + "step": 730 + }, + { + "epoch": 0.018981730501715664, + "grad_norm": 9.27424125755419, + "learning_rate": 9.350169124394292e-06, + "loss": 1.3003, + "step": 731 + }, + { + "epoch": 0.019007697301307615, + "grad_norm": 9.001938928918618, + "learning_rate": 9.352107460253587e-06, + "loss": 1.3159, + "step": 732 + }, + { + "epoch": 0.019033664100899565, + "grad_norm": 13.985923775095209, + "learning_rate": 9.354043149919761e-06, + "loss": 1.315, + "step": 733 + }, + { + "epoch": 0.019059630900491516, + "grad_norm": 17.833782344881563, + "learning_rate": 9.355976200608065e-06, + "loss": 1.3294, + "step": 734 + }, + { + "epoch": 0.019085597700083466, + "grad_norm": 7.499492119441192, + "learning_rate": 9.357906619504287e-06, + "loss": 1.225, + "step": 735 + }, + { + "epoch": 0.019111564499675417, + "grad_norm": 8.104860199293856, + "learning_rate": 9.359834413764898e-06, + "loss": 1.3128, + "step": 736 + }, + { + "epoch": 0.019137531299267364, + "grad_norm": 38.36440267459048, + "learning_rate": 9.36175959051722e-06, + "loss": 1.2852, + "step": 737 + }, + { + "epoch": 0.019163498098859314, + "grad_norm": 8.564558909338148, + "learning_rate": 9.363682156859588e-06, + "loss": 1.2998, + "step": 738 + }, + { + "epoch": 0.019189464898451265, + "grad_norm": 15.0431498098361, + "learning_rate": 9.365602119861494e-06, + "loss": 1.3233, + "step": 739 + }, + { + "epoch": 0.019215431698043215, + "grad_norm": 7.23419861547596, + "learning_rate": 9.367519486563755e-06, + "loss": 1.3532, + "step": 740 + }, + { + "epoch": 0.019241398497635166, + "grad_norm": 20.118620246309995, + "learning_rate": 9.369434263978663e-06, + "loss": 1.279, + "step": 741 + }, + { + "epoch": 0.019267365297227116, + "grad_norm": 22.442236100300715, + "learning_rate": 9.371346459090138e-06, + "loss": 1.2863, + "step": 742 + }, + { + "epoch": 0.019293332096819066, + "grad_norm": 10.269761922689858, + "learning_rate": 9.373256078853882e-06, + "loss": 1.2896, + "step": 743 + }, + { + "epoch": 0.019319298896411017, + "grad_norm": 12.349139077762226, + "learning_rate": 9.375163130197531e-06, + "loss": 1.3074, + "step": 744 + }, + { + "epoch": 0.019345265696002967, + "grad_norm": 12.604691591707839, + "learning_rate": 9.37706762002081e-06, + "loss": 1.261, + "step": 745 + }, + { + "epoch": 0.019371232495594918, + "grad_norm": 6.887276525731672, + "learning_rate": 9.378969555195668e-06, + "loss": 1.2556, + "step": 746 + }, + { + "epoch": 0.01939719929518687, + "grad_norm": 8.542389133572351, + "learning_rate": 9.380868942566452e-06, + "loss": 1.3205, + "step": 747 + }, + { + "epoch": 0.01942316609477882, + "grad_norm": 8.700608278253945, + "learning_rate": 9.382765788950028e-06, + "loss": 1.2898, + "step": 748 + }, + { + "epoch": 0.01944913289437077, + "grad_norm": 8.489947489473856, + "learning_rate": 9.384660101135941e-06, + "loss": 1.2905, + "step": 749 + }, + { + "epoch": 0.01947509969396272, + "grad_norm": 10.502281304443157, + "learning_rate": 9.38655188588657e-06, + "loss": 1.292, + "step": 750 + }, + { + "epoch": 0.01950106649355467, + "grad_norm": 6.105573715508463, + "learning_rate": 9.388441149937248e-06, + "loss": 1.2616, + "step": 751 + }, + { + "epoch": 0.01952703329314662, + "grad_norm": 10.081980548836095, + "learning_rate": 9.390327899996435e-06, + "loss": 1.2787, + "step": 752 + }, + { + "epoch": 0.01955300009273857, + "grad_norm": 13.091469796050344, + "learning_rate": 9.392212142745832e-06, + "loss": 1.3429, + "step": 753 + }, + { + "epoch": 0.01957896689233052, + "grad_norm": 13.511628405667896, + "learning_rate": 9.394093884840547e-06, + "loss": 1.3442, + "step": 754 + }, + { + "epoch": 0.019604933691922472, + "grad_norm": 7.892695330962823, + "learning_rate": 9.395973132909223e-06, + "loss": 1.292, + "step": 755 + }, + { + "epoch": 0.01963090049151442, + "grad_norm": 7.252237698092822, + "learning_rate": 9.397849893554185e-06, + "loss": 1.2646, + "step": 756 + }, + { + "epoch": 0.01965686729110637, + "grad_norm": 6.032199923169485, + "learning_rate": 9.399724173351567e-06, + "loss": 1.3044, + "step": 757 + }, + { + "epoch": 0.01968283409069832, + "grad_norm": 7.917219069097807, + "learning_rate": 9.40159597885147e-06, + "loss": 1.316, + "step": 758 + }, + { + "epoch": 0.01970880089029027, + "grad_norm": 33.126442701483775, + "learning_rate": 9.40346531657808e-06, + "loss": 1.3299, + "step": 759 + }, + { + "epoch": 0.01973476768988222, + "grad_norm": 6.957160927079609, + "learning_rate": 9.405332193029815e-06, + "loss": 1.2675, + "step": 760 + }, + { + "epoch": 0.01976073448947417, + "grad_norm": 9.926834236747744, + "learning_rate": 9.407196614679458e-06, + "loss": 1.3629, + "step": 761 + }, + { + "epoch": 0.019786701289066122, + "grad_norm": 13.584940684571292, + "learning_rate": 9.409058587974296e-06, + "loss": 1.3248, + "step": 762 + }, + { + "epoch": 0.019812668088658072, + "grad_norm": 10.016061853650482, + "learning_rate": 9.41091811933624e-06, + "loss": 1.2911, + "step": 763 + }, + { + "epoch": 0.019838634888250023, + "grad_norm": 8.734101156401493, + "learning_rate": 9.412775215161982e-06, + "loss": 1.2621, + "step": 764 + }, + { + "epoch": 0.019864601687841973, + "grad_norm": 9.316102016127902, + "learning_rate": 9.414629881823095e-06, + "loss": 1.3703, + "step": 765 + }, + { + "epoch": 0.019890568487433924, + "grad_norm": 8.310801397027836, + "learning_rate": 9.416482125666197e-06, + "loss": 1.3004, + "step": 766 + }, + { + "epoch": 0.019916535287025874, + "grad_norm": 9.479428152872343, + "learning_rate": 9.418331953013058e-06, + "loss": 1.3436, + "step": 767 + }, + { + "epoch": 0.019942502086617825, + "grad_norm": 13.372233167016153, + "learning_rate": 9.42017937016074e-06, + "loss": 1.2542, + "step": 768 + }, + { + "epoch": 0.019968468886209775, + "grad_norm": 8.900459414851019, + "learning_rate": 9.42202438338172e-06, + "loss": 1.2621, + "step": 769 + }, + { + "epoch": 0.019994435685801726, + "grad_norm": 9.76160893983355, + "learning_rate": 9.423866998924024e-06, + "loss": 1.2915, + "step": 770 + }, + { + "epoch": 0.020020402485393676, + "grad_norm": 6.760180108193483, + "learning_rate": 9.425707223011352e-06, + "loss": 1.2811, + "step": 771 + }, + { + "epoch": 0.020046369284985627, + "grad_norm": 46.14461441439429, + "learning_rate": 9.427545061843199e-06, + "loss": 1.2612, + "step": 772 + }, + { + "epoch": 0.020072336084577577, + "grad_norm": 6.536094219709947, + "learning_rate": 9.429380521594988e-06, + "loss": 1.2701, + "step": 773 + }, + { + "epoch": 0.020098302884169528, + "grad_norm": 8.885132928561234, + "learning_rate": 9.431213608418187e-06, + "loss": 1.3699, + "step": 774 + }, + { + "epoch": 0.020124269683761475, + "grad_norm": 10.123906205989652, + "learning_rate": 9.43304432844044e-06, + "loss": 1.2437, + "step": 775 + }, + { + "epoch": 0.020150236483353425, + "grad_norm": 7.299895810990617, + "learning_rate": 9.434872687765686e-06, + "loss": 1.2887, + "step": 776 + }, + { + "epoch": 0.020176203282945376, + "grad_norm": 8.515268412832219, + "learning_rate": 9.436698692474278e-06, + "loss": 1.352, + "step": 777 + }, + { + "epoch": 0.020202170082537326, + "grad_norm": 11.817555470974964, + "learning_rate": 9.438522348623111e-06, + "loss": 1.2841, + "step": 778 + }, + { + "epoch": 0.020228136882129277, + "grad_norm": 7.456795674235343, + "learning_rate": 9.44034366224574e-06, + "loss": 1.2854, + "step": 779 + }, + { + "epoch": 0.020254103681721227, + "grad_norm": 12.625504587495392, + "learning_rate": 9.442162639352497e-06, + "loss": 1.2714, + "step": 780 + }, + { + "epoch": 0.020280070481313178, + "grad_norm": 7.162390327612572, + "learning_rate": 9.44397928593061e-06, + "loss": 1.3103, + "step": 781 + }, + { + "epoch": 0.020306037280905128, + "grad_norm": 9.687976850033104, + "learning_rate": 9.445793607944333e-06, + "loss": 1.3012, + "step": 782 + }, + { + "epoch": 0.02033200408049708, + "grad_norm": 11.083621105005188, + "learning_rate": 9.44760561133504e-06, + "loss": 1.284, + "step": 783 + }, + { + "epoch": 0.02035797088008903, + "grad_norm": 12.402071847413175, + "learning_rate": 9.449415302021363e-06, + "loss": 1.3215, + "step": 784 + }, + { + "epoch": 0.02038393767968098, + "grad_norm": 8.763750878920135, + "learning_rate": 9.451222685899305e-06, + "loss": 1.2608, + "step": 785 + }, + { + "epoch": 0.02040990447927293, + "grad_norm": 7.151611690672944, + "learning_rate": 9.453027768842346e-06, + "loss": 1.366, + "step": 786 + }, + { + "epoch": 0.02043587127886488, + "grad_norm": 10.316221363356519, + "learning_rate": 9.454830556701558e-06, + "loss": 1.277, + "step": 787 + }, + { + "epoch": 0.02046183807845683, + "grad_norm": 5.82706979807524, + "learning_rate": 9.45663105530573e-06, + "loss": 1.3201, + "step": 788 + }, + { + "epoch": 0.02048780487804878, + "grad_norm": 9.118958445853728, + "learning_rate": 9.45842927046147e-06, + "loss": 1.3263, + "step": 789 + }, + { + "epoch": 0.020513771677640732, + "grad_norm": 13.75415309260886, + "learning_rate": 9.460225207953329e-06, + "loss": 1.3388, + "step": 790 + }, + { + "epoch": 0.020539738477232682, + "grad_norm": 7.998784926645695, + "learning_rate": 9.462018873543895e-06, + "loss": 1.2874, + "step": 791 + }, + { + "epoch": 0.020565705276824633, + "grad_norm": 11.592095516635329, + "learning_rate": 9.463810272973921e-06, + "loss": 1.3074, + "step": 792 + }, + { + "epoch": 0.020591672076416583, + "grad_norm": 11.969750106102428, + "learning_rate": 9.465599411962424e-06, + "loss": 1.2917, + "step": 793 + }, + { + "epoch": 0.02061763887600853, + "grad_norm": 15.218198041678331, + "learning_rate": 9.467386296206804e-06, + "loss": 1.2511, + "step": 794 + }, + { + "epoch": 0.02064360567560048, + "grad_norm": 6.054787990597522, + "learning_rate": 9.469170931382943e-06, + "loss": 1.2647, + "step": 795 + }, + { + "epoch": 0.02066957247519243, + "grad_norm": 6.2885017097606095, + "learning_rate": 9.470953323145324e-06, + "loss": 1.2985, + "step": 796 + }, + { + "epoch": 0.02069553927478438, + "grad_norm": 12.999428956621218, + "learning_rate": 9.472733477127124e-06, + "loss": 1.3291, + "step": 797 + }, + { + "epoch": 0.020721506074376332, + "grad_norm": 6.862496953800269, + "learning_rate": 9.474511398940337e-06, + "loss": 1.2704, + "step": 798 + }, + { + "epoch": 0.020747472873968283, + "grad_norm": 6.035594581068338, + "learning_rate": 9.47628709417587e-06, + "loss": 1.2695, + "step": 799 + }, + { + "epoch": 0.020773439673560233, + "grad_norm": 13.75671389264251, + "learning_rate": 9.478060568403647e-06, + "loss": 1.3492, + "step": 800 + }, + { + "epoch": 0.020799406473152184, + "grad_norm": 7.501462920882109, + "learning_rate": 9.479831827172727e-06, + "loss": 1.2786, + "step": 801 + }, + { + "epoch": 0.020825373272744134, + "grad_norm": 8.663390151577905, + "learning_rate": 9.481600876011389e-06, + "loss": 1.3202, + "step": 802 + }, + { + "epoch": 0.020851340072336084, + "grad_norm": 8.690349742668722, + "learning_rate": 9.48336772042725e-06, + "loss": 1.3365, + "step": 803 + }, + { + "epoch": 0.020877306871928035, + "grad_norm": 15.078268194753988, + "learning_rate": 9.485132365907367e-06, + "loss": 1.2758, + "step": 804 + }, + { + "epoch": 0.020903273671519985, + "grad_norm": 7.777259876023124, + "learning_rate": 9.486894817918328e-06, + "loss": 1.3495, + "step": 805 + }, + { + "epoch": 0.020929240471111936, + "grad_norm": 7.724683203266029, + "learning_rate": 9.488655081906368e-06, + "loss": 1.3174, + "step": 806 + }, + { + "epoch": 0.020955207270703886, + "grad_norm": 11.227759523758062, + "learning_rate": 9.490413163297458e-06, + "loss": 1.2436, + "step": 807 + }, + { + "epoch": 0.020981174070295837, + "grad_norm": 9.121196633520393, + "learning_rate": 9.492169067497413e-06, + "loss": 1.3519, + "step": 808 + }, + { + "epoch": 0.021007140869887787, + "grad_norm": 6.272152817849625, + "learning_rate": 9.493922799891992e-06, + "loss": 1.3085, + "step": 809 + }, + { + "epoch": 0.021033107669479738, + "grad_norm": 12.326125917211503, + "learning_rate": 9.495674365846988e-06, + "loss": 1.3724, + "step": 810 + }, + { + "epoch": 0.02105907446907169, + "grad_norm": 5.911233081531251, + "learning_rate": 9.497423770708341e-06, + "loss": 1.2948, + "step": 811 + }, + { + "epoch": 0.02108504126866364, + "grad_norm": 7.176595626080198, + "learning_rate": 9.49917101980222e-06, + "loss": 1.2966, + "step": 812 + }, + { + "epoch": 0.02111100806825559, + "grad_norm": 11.241405340559592, + "learning_rate": 9.500916118435131e-06, + "loss": 1.3274, + "step": 813 + }, + { + "epoch": 0.021136974867847536, + "grad_norm": 7.827668136696391, + "learning_rate": 9.502659071894014e-06, + "loss": 1.2669, + "step": 814 + }, + { + "epoch": 0.021162941667439487, + "grad_norm": 7.770025652790313, + "learning_rate": 9.504399885446334e-06, + "loss": 1.3116, + "step": 815 + }, + { + "epoch": 0.021188908467031437, + "grad_norm": 8.230935929431967, + "learning_rate": 9.506138564340174e-06, + "loss": 1.3104, + "step": 816 + }, + { + "epoch": 0.021214875266623388, + "grad_norm": 6.358543847808392, + "learning_rate": 9.50787511380434e-06, + "loss": 1.3019, + "step": 817 + }, + { + "epoch": 0.021240842066215338, + "grad_norm": 5.515899219126165, + "learning_rate": 9.509609539048448e-06, + "loss": 1.2849, + "step": 818 + }, + { + "epoch": 0.02126680886580729, + "grad_norm": 6.932183780494624, + "learning_rate": 9.51134184526302e-06, + "loss": 1.2885, + "step": 819 + }, + { + "epoch": 0.02129277566539924, + "grad_norm": 7.265344101592903, + "learning_rate": 9.513072037619573e-06, + "loss": 1.2536, + "step": 820 + }, + { + "epoch": 0.02131874246499119, + "grad_norm": 7.585652873457997, + "learning_rate": 9.514800121270723e-06, + "loss": 1.2911, + "step": 821 + }, + { + "epoch": 0.02134470926458314, + "grad_norm": 9.816762982226818, + "learning_rate": 9.51652610135026e-06, + "loss": 1.3065, + "step": 822 + }, + { + "epoch": 0.02137067606417509, + "grad_norm": 55.033406048054125, + "learning_rate": 9.518249982973259e-06, + "loss": 1.3019, + "step": 823 + }, + { + "epoch": 0.02139664286376704, + "grad_norm": 7.799871874025926, + "learning_rate": 9.519971771236152e-06, + "loss": 1.3562, + "step": 824 + }, + { + "epoch": 0.02142260966335899, + "grad_norm": 10.481865982034497, + "learning_rate": 9.521691471216828e-06, + "loss": 1.2744, + "step": 825 + }, + { + "epoch": 0.021448576462950942, + "grad_norm": 5.641024673123932, + "learning_rate": 9.523409087974731e-06, + "loss": 1.2588, + "step": 826 + }, + { + "epoch": 0.021474543262542892, + "grad_norm": 5.960853666063499, + "learning_rate": 9.52512462655093e-06, + "loss": 1.2903, + "step": 827 + }, + { + "epoch": 0.021500510062134843, + "grad_norm": 5.084301382140283, + "learning_rate": 9.526838091968224e-06, + "loss": 1.3001, + "step": 828 + }, + { + "epoch": 0.021526476861726793, + "grad_norm": 10.903577766846796, + "learning_rate": 9.528549489231225e-06, + "loss": 1.3012, + "step": 829 + }, + { + "epoch": 0.021552443661318744, + "grad_norm": 6.290636024522288, + "learning_rate": 9.530258823326437e-06, + "loss": 1.2671, + "step": 830 + }, + { + "epoch": 0.021578410460910694, + "grad_norm": 5.72639163232039, + "learning_rate": 9.531966099222364e-06, + "loss": 1.364, + "step": 831 + }, + { + "epoch": 0.021604377260502645, + "grad_norm": 8.199991961219382, + "learning_rate": 9.533671321869575e-06, + "loss": 1.3145, + "step": 832 + }, + { + "epoch": 0.021630344060094592, + "grad_norm": 8.670850868580098, + "learning_rate": 9.5353744962008e-06, + "loss": 1.2914, + "step": 833 + }, + { + "epoch": 0.021656310859686542, + "grad_norm": 7.603732607920527, + "learning_rate": 9.537075627131016e-06, + "loss": 1.2973, + "step": 834 + }, + { + "epoch": 0.021682277659278493, + "grad_norm": 9.550542798020247, + "learning_rate": 9.538774719557533e-06, + "loss": 1.3278, + "step": 835 + }, + { + "epoch": 0.021708244458870443, + "grad_norm": 6.90376478678128, + "learning_rate": 9.540471778360073e-06, + "loss": 1.2987, + "step": 836 + }, + { + "epoch": 0.021734211258462394, + "grad_norm": 6.150753142598995, + "learning_rate": 9.54216680840086e-06, + "loss": 1.3168, + "step": 837 + }, + { + "epoch": 0.021760178058054344, + "grad_norm": 6.09144932280446, + "learning_rate": 9.543859814524698e-06, + "loss": 1.2922, + "step": 838 + }, + { + "epoch": 0.021786144857646295, + "grad_norm": 12.822298796843992, + "learning_rate": 9.54555080155906e-06, + "loss": 1.3162, + "step": 839 + }, + { + "epoch": 0.021812111657238245, + "grad_norm": 8.13494030237128, + "learning_rate": 9.54723977431417e-06, + "loss": 1.2857, + "step": 840 + }, + { + "epoch": 0.021838078456830196, + "grad_norm": 7.570909096718917, + "learning_rate": 9.548926737583076e-06, + "loss": 1.302, + "step": 841 + }, + { + "epoch": 0.021864045256422146, + "grad_norm": 6.438118321621105, + "learning_rate": 9.55061169614174e-06, + "loss": 1.3208, + "step": 842 + }, + { + "epoch": 0.021890012056014096, + "grad_norm": 7.3194385442482774, + "learning_rate": 9.552294654749126e-06, + "loss": 1.3066, + "step": 843 + }, + { + "epoch": 0.021915978855606047, + "grad_norm": 5.81334393024937, + "learning_rate": 9.55397561814726e-06, + "loss": 1.284, + "step": 844 + }, + { + "epoch": 0.021941945655197997, + "grad_norm": 5.669002425557477, + "learning_rate": 9.555654591061334e-06, + "loss": 1.2842, + "step": 845 + }, + { + "epoch": 0.021967912454789948, + "grad_norm": 9.83168981578391, + "learning_rate": 9.557331578199763e-06, + "loss": 1.2914, + "step": 846 + }, + { + "epoch": 0.0219938792543819, + "grad_norm": 5.004403654654639, + "learning_rate": 9.559006584254283e-06, + "loss": 1.2675, + "step": 847 + }, + { + "epoch": 0.02201984605397385, + "grad_norm": 5.598515531787687, + "learning_rate": 9.560679613900021e-06, + "loss": 1.2897, + "step": 848 + }, + { + "epoch": 0.0220458128535658, + "grad_norm": 13.405993336886036, + "learning_rate": 9.562350671795573e-06, + "loss": 1.2598, + "step": 849 + }, + { + "epoch": 0.02207177965315775, + "grad_norm": 8.810016688007392, + "learning_rate": 9.564019762583084e-06, + "loss": 1.2713, + "step": 850 + }, + { + "epoch": 0.0220977464527497, + "grad_norm": 4.822900803651806, + "learning_rate": 9.565686890888321e-06, + "loss": 1.2995, + "step": 851 + }, + { + "epoch": 0.022123713252341647, + "grad_norm": 8.380725782187989, + "learning_rate": 9.567352061320757e-06, + "loss": 1.3437, + "step": 852 + }, + { + "epoch": 0.022149680051933598, + "grad_norm": 15.632187799862061, + "learning_rate": 9.569015278473649e-06, + "loss": 1.3065, + "step": 853 + }, + { + "epoch": 0.022175646851525548, + "grad_norm": 6.554750935516936, + "learning_rate": 9.570676546924098e-06, + "loss": 1.271, + "step": 854 + }, + { + "epoch": 0.0222016136511175, + "grad_norm": 5.467269958763876, + "learning_rate": 9.572335871233142e-06, + "loss": 1.2906, + "step": 855 + }, + { + "epoch": 0.02222758045070945, + "grad_norm": 7.0094679426203985, + "learning_rate": 9.573993255945826e-06, + "loss": 1.3023, + "step": 856 + }, + { + "epoch": 0.0222535472503014, + "grad_norm": 5.637045445693295, + "learning_rate": 9.575648705591272e-06, + "loss": 1.2869, + "step": 857 + }, + { + "epoch": 0.02227951404989335, + "grad_norm": 7.239452713993271, + "learning_rate": 9.577302224682756e-06, + "loss": 1.3024, + "step": 858 + }, + { + "epoch": 0.0223054808494853, + "grad_norm": 10.993978412764509, + "learning_rate": 9.578953817717786e-06, + "loss": 1.2899, + "step": 859 + }, + { + "epoch": 0.02233144764907725, + "grad_norm": 5.587018955354914, + "learning_rate": 9.580603489178172e-06, + "loss": 1.3164, + "step": 860 + }, + { + "epoch": 0.0223574144486692, + "grad_norm": 9.673838574784886, + "learning_rate": 9.582251243530096e-06, + "loss": 1.2987, + "step": 861 + }, + { + "epoch": 0.022383381248261152, + "grad_norm": 12.867992679346035, + "learning_rate": 9.583897085224188e-06, + "loss": 1.2432, + "step": 862 + }, + { + "epoch": 0.022409348047853102, + "grad_norm": 5.948304314571507, + "learning_rate": 9.585541018695598e-06, + "loss": 1.2896, + "step": 863 + }, + { + "epoch": 0.022435314847445053, + "grad_norm": 6.3771274980945085, + "learning_rate": 9.587183048364066e-06, + "loss": 1.318, + "step": 864 + }, + { + "epoch": 0.022461281647037003, + "grad_norm": 4.9450878315252424, + "learning_rate": 9.588823178633999e-06, + "loss": 1.2777, + "step": 865 + }, + { + "epoch": 0.022487248446628954, + "grad_norm": 7.322631338393158, + "learning_rate": 9.590461413894532e-06, + "loss": 1.2895, + "step": 866 + }, + { + "epoch": 0.022513215246220904, + "grad_norm": 5.996004999852148, + "learning_rate": 9.592097758519608e-06, + "loss": 1.2961, + "step": 867 + }, + { + "epoch": 0.022539182045812855, + "grad_norm": 15.872299325058666, + "learning_rate": 9.593732216868041e-06, + "loss": 1.3118, + "step": 868 + }, + { + "epoch": 0.022565148845404805, + "grad_norm": 10.305991228416378, + "learning_rate": 9.595364793283588e-06, + "loss": 1.3102, + "step": 869 + }, + { + "epoch": 0.022591115644996756, + "grad_norm": 7.229728244512511, + "learning_rate": 9.596995492095025e-06, + "loss": 1.2981, + "step": 870 + }, + { + "epoch": 0.022617082444588703, + "grad_norm": 25.113071931484956, + "learning_rate": 9.598624317616204e-06, + "loss": 1.3444, + "step": 871 + }, + { + "epoch": 0.022643049244180653, + "grad_norm": 11.490754100996606, + "learning_rate": 9.600251274146124e-06, + "loss": 1.3223, + "step": 872 + }, + { + "epoch": 0.022669016043772604, + "grad_norm": 4.9926604761761135, + "learning_rate": 9.601876365969014e-06, + "loss": 1.2855, + "step": 873 + }, + { + "epoch": 0.022694982843364554, + "grad_norm": 4.931973418536181, + "learning_rate": 9.603499597354378e-06, + "loss": 1.3321, + "step": 874 + }, + { + "epoch": 0.022720949642956505, + "grad_norm": 5.436327928522666, + "learning_rate": 9.605120972557077e-06, + "loss": 1.3053, + "step": 875 + }, + { + "epoch": 0.022746916442548455, + "grad_norm": 8.78772255410856, + "learning_rate": 9.606740495817399e-06, + "loss": 1.3042, + "step": 876 + }, + { + "epoch": 0.022772883242140406, + "grad_norm": 8.863838161492176, + "learning_rate": 9.608358171361102e-06, + "loss": 1.2963, + "step": 877 + }, + { + "epoch": 0.022798850041732356, + "grad_norm": 5.743450249123793, + "learning_rate": 9.609974003399514e-06, + "loss": 1.3559, + "step": 878 + }, + { + "epoch": 0.022824816841324307, + "grad_norm": 5.4974475486511825, + "learning_rate": 9.611587996129575e-06, + "loss": 1.2636, + "step": 879 + }, + { + "epoch": 0.022850783640916257, + "grad_norm": 7.229712415280847, + "learning_rate": 9.613200153733906e-06, + "loss": 1.3024, + "step": 880 + }, + { + "epoch": 0.022876750440508208, + "grad_norm": 9.665796394197534, + "learning_rate": 9.614810480380884e-06, + "loss": 1.3013, + "step": 881 + }, + { + "epoch": 0.022902717240100158, + "grad_norm": 7.92096291131216, + "learning_rate": 9.616418980224693e-06, + "loss": 1.3147, + "step": 882 + }, + { + "epoch": 0.02292868403969211, + "grad_norm": 6.818307903164356, + "learning_rate": 9.6180256574054e-06, + "loss": 1.29, + "step": 883 + }, + { + "epoch": 0.02295465083928406, + "grad_norm": 5.7847685133399835, + "learning_rate": 9.61963051604901e-06, + "loss": 1.2674, + "step": 884 + }, + { + "epoch": 0.02298061763887601, + "grad_norm": 6.394619837982118, + "learning_rate": 9.621233560267536e-06, + "loss": 1.3388, + "step": 885 + }, + { + "epoch": 0.02300658443846796, + "grad_norm": 5.668685478676017, + "learning_rate": 9.622834794159064e-06, + "loss": 1.2962, + "step": 886 + }, + { + "epoch": 0.02303255123805991, + "grad_norm": 9.472242258994749, + "learning_rate": 9.624434221807802e-06, + "loss": 1.3117, + "step": 887 + }, + { + "epoch": 0.02305851803765186, + "grad_norm": 11.334118161421749, + "learning_rate": 9.626031847284162e-06, + "loss": 1.2593, + "step": 888 + }, + { + "epoch": 0.02308448483724381, + "grad_norm": 6.080792074810655, + "learning_rate": 9.627627674644803e-06, + "loss": 1.3573, + "step": 889 + }, + { + "epoch": 0.02311045163683576, + "grad_norm": 8.550900363920247, + "learning_rate": 9.629221707932712e-06, + "loss": 1.2709, + "step": 890 + }, + { + "epoch": 0.02313641843642771, + "grad_norm": 8.165138049666838, + "learning_rate": 9.630813951177247e-06, + "loss": 1.246, + "step": 891 + }, + { + "epoch": 0.02316238523601966, + "grad_norm": 5.552837423199632, + "learning_rate": 9.632404408394215e-06, + "loss": 1.2925, + "step": 892 + }, + { + "epoch": 0.02318835203561161, + "grad_norm": 18.042003367088892, + "learning_rate": 9.633993083585915e-06, + "loss": 1.2752, + "step": 893 + }, + { + "epoch": 0.02321431883520356, + "grad_norm": 6.71109239534936, + "learning_rate": 9.635579980741213e-06, + "loss": 1.3258, + "step": 894 + }, + { + "epoch": 0.02324028563479551, + "grad_norm": 6.9044891460144715, + "learning_rate": 9.637165103835603e-06, + "loss": 1.2974, + "step": 895 + }, + { + "epoch": 0.02326625243438746, + "grad_norm": 9.853750035925806, + "learning_rate": 9.638748456831248e-06, + "loss": 1.3332, + "step": 896 + }, + { + "epoch": 0.02329221923397941, + "grad_norm": 9.558153947534306, + "learning_rate": 9.640330043677061e-06, + "loss": 1.3096, + "step": 897 + }, + { + "epoch": 0.023318186033571362, + "grad_norm": 8.610965533376161, + "learning_rate": 9.641909868308754e-06, + "loss": 1.3152, + "step": 898 + }, + { + "epoch": 0.023344152833163313, + "grad_norm": 7.142766385183039, + "learning_rate": 9.643487934648895e-06, + "loss": 1.2709, + "step": 899 + }, + { + "epoch": 0.023370119632755263, + "grad_norm": 6.166920046712851, + "learning_rate": 9.645064246606975e-06, + "loss": 1.35, + "step": 900 + }, + { + "epoch": 0.023396086432347214, + "grad_norm": 5.530068481462772, + "learning_rate": 9.646638808079456e-06, + "loss": 1.2657, + "step": 901 + }, + { + "epoch": 0.023422053231939164, + "grad_norm": 10.349556851690181, + "learning_rate": 9.648211622949834e-06, + "loss": 1.3126, + "step": 902 + }, + { + "epoch": 0.023448020031531114, + "grad_norm": 16.52874194555455, + "learning_rate": 9.649782695088697e-06, + "loss": 1.3455, + "step": 903 + }, + { + "epoch": 0.023473986831123065, + "grad_norm": 17.179491209372024, + "learning_rate": 9.651352028353778e-06, + "loss": 1.2803, + "step": 904 + }, + { + "epoch": 0.023499953630715015, + "grad_norm": 15.271347862967362, + "learning_rate": 9.652919626590021e-06, + "loss": 1.2948, + "step": 905 + }, + { + "epoch": 0.023525920430306966, + "grad_norm": 6.534933557267464, + "learning_rate": 9.654485493629629e-06, + "loss": 1.3214, + "step": 906 + }, + { + "epoch": 0.023551887229898916, + "grad_norm": 18.19862974269543, + "learning_rate": 9.656049633292117e-06, + "loss": 1.3455, + "step": 907 + }, + { + "epoch": 0.023577854029490867, + "grad_norm": 8.118425262041375, + "learning_rate": 9.657612049384382e-06, + "loss": 1.2953, + "step": 908 + }, + { + "epoch": 0.023603820829082814, + "grad_norm": 7.907715431677386, + "learning_rate": 9.659172745700743e-06, + "loss": 1.2986, + "step": 909 + }, + { + "epoch": 0.023629787628674764, + "grad_norm": 6.806759716945963, + "learning_rate": 9.660731726023005e-06, + "loss": 1.2975, + "step": 910 + }, + { + "epoch": 0.023655754428266715, + "grad_norm": 11.35526651544202, + "learning_rate": 9.66228899412052e-06, + "loss": 1.3155, + "step": 911 + }, + { + "epoch": 0.023681721227858665, + "grad_norm": 8.15906040691162, + "learning_rate": 9.66384455375022e-06, + "loss": 1.3255, + "step": 912 + }, + { + "epoch": 0.023707688027450616, + "grad_norm": 6.326080217960587, + "learning_rate": 9.665398408656697e-06, + "loss": 1.3095, + "step": 913 + }, + { + "epoch": 0.023733654827042566, + "grad_norm": 6.377296781991996, + "learning_rate": 9.666950562572242e-06, + "loss": 1.3069, + "step": 914 + }, + { + "epoch": 0.023759621626634517, + "grad_norm": 5.0605873921935505, + "learning_rate": 9.668501019216903e-06, + "loss": 1.2717, + "step": 915 + }, + { + "epoch": 0.023785588426226467, + "grad_norm": 7.202311505039545, + "learning_rate": 9.670049782298533e-06, + "loss": 1.2766, + "step": 916 + }, + { + "epoch": 0.023811555225818418, + "grad_norm": 5.985144025428875, + "learning_rate": 9.671596855512855e-06, + "loss": 1.2753, + "step": 917 + }, + { + "epoch": 0.023837522025410368, + "grad_norm": 4.749374348444241, + "learning_rate": 9.673142242543502e-06, + "loss": 1.2848, + "step": 918 + }, + { + "epoch": 0.02386348882500232, + "grad_norm": 6.7693010523288555, + "learning_rate": 9.67468594706208e-06, + "loss": 1.3632, + "step": 919 + }, + { + "epoch": 0.02388945562459427, + "grad_norm": 7.2012825883135605, + "learning_rate": 9.676227972728213e-06, + "loss": 1.2742, + "step": 920 + }, + { + "epoch": 0.02391542242418622, + "grad_norm": 6.82820991795022, + "learning_rate": 9.677768323189597e-06, + "loss": 1.3219, + "step": 921 + }, + { + "epoch": 0.02394138922377817, + "grad_norm": 6.974731068229161, + "learning_rate": 9.67930700208206e-06, + "loss": 1.3178, + "step": 922 + }, + { + "epoch": 0.02396735602337012, + "grad_norm": 15.13070596302278, + "learning_rate": 9.680844013029594e-06, + "loss": 1.284, + "step": 923 + }, + { + "epoch": 0.02399332282296207, + "grad_norm": 6.438319181893629, + "learning_rate": 9.68237935964443e-06, + "loss": 1.2973, + "step": 924 + }, + { + "epoch": 0.02401928962255402, + "grad_norm": 6.625264576361242, + "learning_rate": 9.68391304552707e-06, + "loss": 1.2586, + "step": 925 + }, + { + "epoch": 0.024045256422145972, + "grad_norm": 15.059806614580237, + "learning_rate": 9.68544507426635e-06, + "loss": 1.3258, + "step": 926 + }, + { + "epoch": 0.024071223221737922, + "grad_norm": 8.527297225594568, + "learning_rate": 9.68697544943948e-06, + "loss": 1.2771, + "step": 927 + }, + { + "epoch": 0.02409719002132987, + "grad_norm": 5.311482410584564, + "learning_rate": 9.688504174612103e-06, + "loss": 1.2686, + "step": 928 + }, + { + "epoch": 0.02412315682092182, + "grad_norm": 7.752117729025274, + "learning_rate": 9.69003125333834e-06, + "loss": 1.2727, + "step": 929 + }, + { + "epoch": 0.02414912362051377, + "grad_norm": 4.8332415385682514, + "learning_rate": 9.691556689160846e-06, + "loss": 1.3241, + "step": 930 + }, + { + "epoch": 0.02417509042010572, + "grad_norm": 6.002508274948092, + "learning_rate": 9.693080485610847e-06, + "loss": 1.3002, + "step": 931 + }, + { + "epoch": 0.02420105721969767, + "grad_norm": 6.7418202889208985, + "learning_rate": 9.694602646208196e-06, + "loss": 1.2637, + "step": 932 + }, + { + "epoch": 0.024227024019289622, + "grad_norm": 4.9300360941774635, + "learning_rate": 9.696123174461428e-06, + "loss": 1.3186, + "step": 933 + }, + { + "epoch": 0.024252990818881572, + "grad_norm": 4.788797841063375, + "learning_rate": 9.697642073867802e-06, + "loss": 1.3139, + "step": 934 + }, + { + "epoch": 0.024278957618473523, + "grad_norm": 25.32759725313847, + "learning_rate": 9.699159347913341e-06, + "loss": 1.3188, + "step": 935 + }, + { + "epoch": 0.024304924418065473, + "grad_norm": 6.124397403891343, + "learning_rate": 9.700675000072903e-06, + "loss": 1.2059, + "step": 936 + }, + { + "epoch": 0.024330891217657424, + "grad_norm": 6.15710195465208, + "learning_rate": 9.702189033810203e-06, + "loss": 1.2634, + "step": 937 + }, + { + "epoch": 0.024356858017249374, + "grad_norm": 5.224370431159867, + "learning_rate": 9.703701452577877e-06, + "loss": 1.2649, + "step": 938 + }, + { + "epoch": 0.024382824816841325, + "grad_norm": 9.860583247949691, + "learning_rate": 9.705212259817522e-06, + "loss": 1.3026, + "step": 939 + }, + { + "epoch": 0.024408791616433275, + "grad_norm": 9.254174192775624, + "learning_rate": 9.70672145895975e-06, + "loss": 1.2627, + "step": 940 + }, + { + "epoch": 0.024434758416025226, + "grad_norm": 10.661632621519558, + "learning_rate": 9.708229053424222e-06, + "loss": 1.2983, + "step": 941 + }, + { + "epoch": 0.024460725215617176, + "grad_norm": 5.977909110957374, + "learning_rate": 9.709735046619713e-06, + "loss": 1.3384, + "step": 942 + }, + { + "epoch": 0.024486692015209126, + "grad_norm": 5.354247017144938, + "learning_rate": 9.711239441944137e-06, + "loss": 1.2425, + "step": 943 + }, + { + "epoch": 0.024512658814801077, + "grad_norm": 4.443145813892533, + "learning_rate": 9.712742242784614e-06, + "loss": 1.2567, + "step": 944 + }, + { + "epoch": 0.024538625614393027, + "grad_norm": 16.416654996658025, + "learning_rate": 9.7142434525175e-06, + "loss": 1.3159, + "step": 945 + }, + { + "epoch": 0.024564592413984978, + "grad_norm": 6.161107040111931, + "learning_rate": 9.715743074508432e-06, + "loss": 1.3107, + "step": 946 + }, + { + "epoch": 0.024590559213576925, + "grad_norm": 5.047379506880239, + "learning_rate": 9.717241112112394e-06, + "loss": 1.2834, + "step": 947 + }, + { + "epoch": 0.024616526013168875, + "grad_norm": 7.430957384821238, + "learning_rate": 9.718737568673735e-06, + "loss": 1.2714, + "step": 948 + }, + { + "epoch": 0.024642492812760826, + "grad_norm": 13.865081816362068, + "learning_rate": 9.720232447526234e-06, + "loss": 1.2922, + "step": 949 + }, + { + "epoch": 0.024668459612352776, + "grad_norm": 7.719543416253316, + "learning_rate": 9.721725751993129e-06, + "loss": 1.318, + "step": 950 + }, + { + "epoch": 0.024694426411944727, + "grad_norm": 6.456343247810244, + "learning_rate": 9.723217485387176e-06, + "loss": 1.2868, + "step": 951 + }, + { + "epoch": 0.024720393211536677, + "grad_norm": 6.514205375453248, + "learning_rate": 9.724707651010681e-06, + "loss": 1.2431, + "step": 952 + }, + { + "epoch": 0.024746360011128628, + "grad_norm": 6.89294754296579, + "learning_rate": 9.726196252155556e-06, + "loss": 1.292, + "step": 953 + }, + { + "epoch": 0.024772326810720578, + "grad_norm": 8.456403790955324, + "learning_rate": 9.72768329210335e-06, + "loss": 1.2675, + "step": 954 + }, + { + "epoch": 0.02479829361031253, + "grad_norm": 6.805975072823477, + "learning_rate": 9.729168774125295e-06, + "loss": 1.2808, + "step": 955 + }, + { + "epoch": 0.02482426040990448, + "grad_norm": 4.884198240863559, + "learning_rate": 9.730652701482362e-06, + "loss": 1.305, + "step": 956 + }, + { + "epoch": 0.02485022720949643, + "grad_norm": 10.604731716447711, + "learning_rate": 9.732135077425285e-06, + "loss": 1.2619, + "step": 957 + }, + { + "epoch": 0.02487619400908838, + "grad_norm": 5.267105573561905, + "learning_rate": 9.733615905194618e-06, + "loss": 1.2869, + "step": 958 + }, + { + "epoch": 0.02490216080868033, + "grad_norm": 6.79009095930207, + "learning_rate": 9.735095188020771e-06, + "loss": 1.3256, + "step": 959 + }, + { + "epoch": 0.02492812760827228, + "grad_norm": 5.721454585332429, + "learning_rate": 9.736572929124054e-06, + "loss": 1.3219, + "step": 960 + }, + { + "epoch": 0.02495409440786423, + "grad_norm": 5.676630385875216, + "learning_rate": 9.738049131714717e-06, + "loss": 1.2697, + "step": 961 + }, + { + "epoch": 0.024980061207456182, + "grad_norm": 19.56131748123225, + "learning_rate": 9.739523798992998e-06, + "loss": 1.2768, + "step": 962 + }, + { + "epoch": 0.025006028007048132, + "grad_norm": 6.551889033587738, + "learning_rate": 9.740996934149154e-06, + "loss": 1.2909, + "step": 963 + }, + { + "epoch": 0.025031994806640083, + "grad_norm": 6.254231355736228, + "learning_rate": 9.742468540363514e-06, + "loss": 1.2957, + "step": 964 + }, + { + "epoch": 0.025057961606232033, + "grad_norm": 6.289857044141956, + "learning_rate": 9.743938620806514e-06, + "loss": 1.2626, + "step": 965 + }, + { + "epoch": 0.025083928405823984, + "grad_norm": 7.581360808440901, + "learning_rate": 9.745407178638734e-06, + "loss": 1.316, + "step": 966 + }, + { + "epoch": 0.02510989520541593, + "grad_norm": 9.473038813151181, + "learning_rate": 9.746874217010951e-06, + "loss": 1.2473, + "step": 967 + }, + { + "epoch": 0.02513586200500788, + "grad_norm": 4.546115457863514, + "learning_rate": 9.748339739064166e-06, + "loss": 1.3249, + "step": 968 + }, + { + "epoch": 0.025161828804599832, + "grad_norm": 8.662045298170554, + "learning_rate": 9.749803747929655e-06, + "loss": 1.3347, + "step": 969 + }, + { + "epoch": 0.025187795604191782, + "grad_norm": 7.319306425578429, + "learning_rate": 9.751266246729001e-06, + "loss": 1.2844, + "step": 970 + }, + { + "epoch": 0.025213762403783733, + "grad_norm": 5.191683817060053, + "learning_rate": 9.752727238574139e-06, + "loss": 1.3249, + "step": 971 + }, + { + "epoch": 0.025239729203375683, + "grad_norm": 7.46033568826115, + "learning_rate": 9.754186726567395e-06, + "loss": 1.3356, + "step": 972 + }, + { + "epoch": 0.025265696002967634, + "grad_norm": 4.9832783512072005, + "learning_rate": 9.755644713801525e-06, + "loss": 1.3125, + "step": 973 + }, + { + "epoch": 0.025291662802559584, + "grad_norm": 6.483181542562136, + "learning_rate": 9.757101203359754e-06, + "loss": 1.2658, + "step": 974 + }, + { + "epoch": 0.025317629602151535, + "grad_norm": 5.959667065429219, + "learning_rate": 9.758556198315812e-06, + "loss": 1.2394, + "step": 975 + }, + { + "epoch": 0.025343596401743485, + "grad_norm": 9.80932254734702, + "learning_rate": 9.76000970173398e-06, + "loss": 1.3495, + "step": 976 + }, + { + "epoch": 0.025369563201335436, + "grad_norm": 4.057787465402776, + "learning_rate": 9.761461716669126e-06, + "loss": 1.3405, + "step": 977 + }, + { + "epoch": 0.025395530000927386, + "grad_norm": 10.611116561311272, + "learning_rate": 9.76291224616674e-06, + "loss": 1.2725, + "step": 978 + }, + { + "epoch": 0.025421496800519337, + "grad_norm": 6.494670737292218, + "learning_rate": 9.764361293262972e-06, + "loss": 1.2891, + "step": 979 + }, + { + "epoch": 0.025447463600111287, + "grad_norm": 9.537197171224495, + "learning_rate": 9.765808860984678e-06, + "loss": 1.3093, + "step": 980 + }, + { + "epoch": 0.025473430399703238, + "grad_norm": 5.815006167014108, + "learning_rate": 9.767254952349454e-06, + "loss": 1.3046, + "step": 981 + }, + { + "epoch": 0.025499397199295188, + "grad_norm": 7.86228521930036, + "learning_rate": 9.768699570365663e-06, + "loss": 1.3013, + "step": 982 + }, + { + "epoch": 0.02552536399888714, + "grad_norm": 5.721077867335082, + "learning_rate": 9.770142718032494e-06, + "loss": 1.2978, + "step": 983 + }, + { + "epoch": 0.02555133079847909, + "grad_norm": 5.856146245829168, + "learning_rate": 9.771584398339979e-06, + "loss": 1.2845, + "step": 984 + }, + { + "epoch": 0.02557729759807104, + "grad_norm": 8.408391176395932, + "learning_rate": 9.773024614269044e-06, + "loss": 1.301, + "step": 985 + }, + { + "epoch": 0.025603264397662986, + "grad_norm": 5.087963538645011, + "learning_rate": 9.774463368791538e-06, + "loss": 1.3114, + "step": 986 + }, + { + "epoch": 0.025629231197254937, + "grad_norm": 9.0692643801184, + "learning_rate": 9.775900664870271e-06, + "loss": 1.2596, + "step": 987 + }, + { + "epoch": 0.025655197996846887, + "grad_norm": 12.134828339860874, + "learning_rate": 9.777336505459056e-06, + "loss": 1.2722, + "step": 988 + }, + { + "epoch": 0.025681164796438838, + "grad_norm": 6.718353401721282, + "learning_rate": 9.778770893502738e-06, + "loss": 1.2841, + "step": 989 + }, + { + "epoch": 0.02570713159603079, + "grad_norm": 6.623555151813948, + "learning_rate": 9.780203831937234e-06, + "loss": 1.2548, + "step": 990 + }, + { + "epoch": 0.02573309839562274, + "grad_norm": 10.501194840522352, + "learning_rate": 9.781635323689574e-06, + "loss": 1.3054, + "step": 991 + }, + { + "epoch": 0.02575906519521469, + "grad_norm": 7.389525511172787, + "learning_rate": 9.783065371677924e-06, + "loss": 1.3585, + "step": 992 + }, + { + "epoch": 0.02578503199480664, + "grad_norm": 7.097115008429206, + "learning_rate": 9.784493978811635e-06, + "loss": 1.2735, + "step": 993 + }, + { + "epoch": 0.02581099879439859, + "grad_norm": 5.317230272863703, + "learning_rate": 9.785921147991267e-06, + "loss": 1.2838, + "step": 994 + }, + { + "epoch": 0.02583696559399054, + "grad_norm": 6.095939047421082, + "learning_rate": 9.787346882108637e-06, + "loss": 1.2572, + "step": 995 + }, + { + "epoch": 0.02586293239358249, + "grad_norm": 6.978365025575531, + "learning_rate": 9.788771184046845e-06, + "loss": 1.2687, + "step": 996 + }, + { + "epoch": 0.02588889919317444, + "grad_norm": 13.259350626192834, + "learning_rate": 9.790194056680308e-06, + "loss": 1.2572, + "step": 997 + }, + { + "epoch": 0.025914865992766392, + "grad_norm": 9.119014919686707, + "learning_rate": 9.791615502874801e-06, + "loss": 1.3067, + "step": 998 + }, + { + "epoch": 0.025940832792358343, + "grad_norm": 7.927527463270512, + "learning_rate": 9.79303552548749e-06, + "loss": 1.2221, + "step": 999 + }, + { + "epoch": 0.025966799591950293, + "grad_norm": 6.758283547979663, + "learning_rate": 9.794454127366962e-06, + "loss": 1.2497, + "step": 1000 + }, + { + "epoch": 0.025992766391542244, + "grad_norm": 7.403770836629018, + "learning_rate": 9.795871311353265e-06, + "loss": 1.3469, + "step": 1001 + }, + { + "epoch": 0.026018733191134194, + "grad_norm": 15.645674798249456, + "learning_rate": 9.797287080277939e-06, + "loss": 1.2602, + "step": 1002 + }, + { + "epoch": 0.026044699990726145, + "grad_norm": 7.0923842405137565, + "learning_rate": 9.79870143696405e-06, + "loss": 1.3024, + "step": 1003 + }, + { + "epoch": 0.026070666790318095, + "grad_norm": 15.947687939333136, + "learning_rate": 9.800114384226223e-06, + "loss": 1.2514, + "step": 1004 + }, + { + "epoch": 0.026096633589910042, + "grad_norm": 8.867820737301034, + "learning_rate": 9.801525924870682e-06, + "loss": 1.2825, + "step": 1005 + }, + { + "epoch": 0.026122600389501992, + "grad_norm": 21.461414577651404, + "learning_rate": 9.802936061695272e-06, + "loss": 1.2828, + "step": 1006 + }, + { + "epoch": 0.026148567189093943, + "grad_norm": 8.469150435211214, + "learning_rate": 9.804344797489503e-06, + "loss": 1.2981, + "step": 1007 + }, + { + "epoch": 0.026174533988685893, + "grad_norm": 10.420956595473665, + "learning_rate": 9.805752135034576e-06, + "loss": 1.3322, + "step": 1008 + }, + { + "epoch": 0.026200500788277844, + "grad_norm": 6.847053462131083, + "learning_rate": 9.807158077103422e-06, + "loss": 1.3094, + "step": 1009 + }, + { + "epoch": 0.026226467587869794, + "grad_norm": 10.458620186528178, + "learning_rate": 9.808562626460728e-06, + "loss": 1.363, + "step": 1010 + }, + { + "epoch": 0.026252434387461745, + "grad_norm": 9.791115949578268, + "learning_rate": 9.809965785862977e-06, + "loss": 1.3119, + "step": 1011 + }, + { + "epoch": 0.026278401187053695, + "grad_norm": 7.743233218433294, + "learning_rate": 9.81136755805847e-06, + "loss": 1.2529, + "step": 1012 + }, + { + "epoch": 0.026304367986645646, + "grad_norm": 6.762522631482175, + "learning_rate": 9.812767945787375e-06, + "loss": 1.2703, + "step": 1013 + }, + { + "epoch": 0.026330334786237596, + "grad_norm": 5.997442336129237, + "learning_rate": 9.81416695178174e-06, + "loss": 1.2476, + "step": 1014 + }, + { + "epoch": 0.026356301585829547, + "grad_norm": 6.057686379305182, + "learning_rate": 9.815564578765534e-06, + "loss": 1.2868, + "step": 1015 + }, + { + "epoch": 0.026382268385421497, + "grad_norm": 6.9888953274426155, + "learning_rate": 9.816960829454688e-06, + "loss": 1.2332, + "step": 1016 + }, + { + "epoch": 0.026408235185013448, + "grad_norm": 5.994918260693183, + "learning_rate": 9.818355706557106e-06, + "loss": 1.3004, + "step": 1017 + }, + { + "epoch": 0.026434201984605398, + "grad_norm": 20.47101982285911, + "learning_rate": 9.819749212772719e-06, + "loss": 1.2649, + "step": 1018 + }, + { + "epoch": 0.02646016878419735, + "grad_norm": 7.834661925714113, + "learning_rate": 9.821141350793496e-06, + "loss": 1.269, + "step": 1019 + }, + { + "epoch": 0.0264861355837893, + "grad_norm": 7.016448994187666, + "learning_rate": 9.822532123303488e-06, + "loss": 1.2671, + "step": 1020 + }, + { + "epoch": 0.02651210238338125, + "grad_norm": 10.319168062018734, + "learning_rate": 9.823921532978857e-06, + "loss": 1.2819, + "step": 1021 + }, + { + "epoch": 0.0265380691829732, + "grad_norm": 8.335203901793443, + "learning_rate": 9.825309582487907e-06, + "loss": 1.2271, + "step": 1022 + }, + { + "epoch": 0.02656403598256515, + "grad_norm": 8.86659035709595, + "learning_rate": 9.826696274491106e-06, + "loss": 1.3004, + "step": 1023 + }, + { + "epoch": 0.026590002782157098, + "grad_norm": 5.759697076461371, + "learning_rate": 9.828081611641132e-06, + "loss": 1.2976, + "step": 1024 + }, + { + "epoch": 0.026615969581749048, + "grad_norm": 5.144906438008312, + "learning_rate": 9.829465596582888e-06, + "loss": 1.2771, + "step": 1025 + }, + { + "epoch": 0.026641936381341, + "grad_norm": 8.762602966835512, + "learning_rate": 9.830848231953547e-06, + "loss": 1.2333, + "step": 1026 + }, + { + "epoch": 0.02666790318093295, + "grad_norm": 9.294652849581757, + "learning_rate": 9.832229520382572e-06, + "loss": 1.2944, + "step": 1027 + }, + { + "epoch": 0.0266938699805249, + "grad_norm": 9.372601418423425, + "learning_rate": 9.833609464491743e-06, + "loss": 1.3702, + "step": 1028 + }, + { + "epoch": 0.02671983678011685, + "grad_norm": 12.490429075153717, + "learning_rate": 9.834988066895202e-06, + "loss": 1.3249, + "step": 1029 + }, + { + "epoch": 0.0267458035797088, + "grad_norm": 7.940191338297178, + "learning_rate": 9.836365330199466e-06, + "loss": 1.2973, + "step": 1030 + }, + { + "epoch": 0.02677177037930075, + "grad_norm": 15.315136989936061, + "learning_rate": 9.837741257003469e-06, + "loss": 1.3098, + "step": 1031 + }, + { + "epoch": 0.0267977371788927, + "grad_norm": 6.843209859356881, + "learning_rate": 9.83911584989858e-06, + "loss": 1.2724, + "step": 1032 + }, + { + "epoch": 0.026823703978484652, + "grad_norm": 6.574569212286721, + "learning_rate": 9.840489111468645e-06, + "loss": 1.3237, + "step": 1033 + }, + { + "epoch": 0.026849670778076602, + "grad_norm": 6.66978318166563, + "learning_rate": 9.841861044290009e-06, + "loss": 1.248, + "step": 1034 + }, + { + "epoch": 0.026875637577668553, + "grad_norm": 13.064317494670837, + "learning_rate": 9.84323165093154e-06, + "loss": 1.3519, + "step": 1035 + }, + { + "epoch": 0.026901604377260503, + "grad_norm": 5.760771903132589, + "learning_rate": 9.84460093395467e-06, + "loss": 1.3055, + "step": 1036 + }, + { + "epoch": 0.026927571176852454, + "grad_norm": 12.91904655599691, + "learning_rate": 9.845968895913416e-06, + "loss": 1.3158, + "step": 1037 + }, + { + "epoch": 0.026953537976444404, + "grad_norm": 7.414051437771647, + "learning_rate": 9.847335539354406e-06, + "loss": 1.2901, + "step": 1038 + }, + { + "epoch": 0.026979504776036355, + "grad_norm": 8.698376769423797, + "learning_rate": 9.848700866816916e-06, + "loss": 1.2545, + "step": 1039 + }, + { + "epoch": 0.027005471575628305, + "grad_norm": 25.95310265840399, + "learning_rate": 9.85006488083289e-06, + "loss": 1.2896, + "step": 1040 + }, + { + "epoch": 0.027031438375220256, + "grad_norm": 6.316976640622777, + "learning_rate": 9.851427583926974e-06, + "loss": 1.3023, + "step": 1041 + }, + { + "epoch": 0.027057405174812206, + "grad_norm": 8.75734729194845, + "learning_rate": 9.852788978616541e-06, + "loss": 1.3546, + "step": 1042 + }, + { + "epoch": 0.027083371974404153, + "grad_norm": 7.832613025851854, + "learning_rate": 9.854149067411725e-06, + "loss": 1.3085, + "step": 1043 + }, + { + "epoch": 0.027109338773996104, + "grad_norm": 5.786629807901263, + "learning_rate": 9.855507852815431e-06, + "loss": 1.292, + "step": 1044 + }, + { + "epoch": 0.027135305573588054, + "grad_norm": 6.641910204217509, + "learning_rate": 9.856865337323388e-06, + "loss": 1.3013, + "step": 1045 + }, + { + "epoch": 0.027161272373180004, + "grad_norm": 8.89247624275867, + "learning_rate": 9.858221523424158e-06, + "loss": 1.2903, + "step": 1046 + }, + { + "epoch": 0.027187239172771955, + "grad_norm": 7.59512655708994, + "learning_rate": 9.859576413599167e-06, + "loss": 1.2719, + "step": 1047 + }, + { + "epoch": 0.027213205972363905, + "grad_norm": 13.562144472586645, + "learning_rate": 9.860930010322737e-06, + "loss": 1.2693, + "step": 1048 + }, + { + "epoch": 0.027239172771955856, + "grad_norm": 8.082274324975623, + "learning_rate": 9.862282316062114e-06, + "loss": 1.2466, + "step": 1049 + }, + { + "epoch": 0.027265139571547806, + "grad_norm": 6.916387927708562, + "learning_rate": 9.863633333277485e-06, + "loss": 1.317, + "step": 1050 + }, + { + "epoch": 0.027291106371139757, + "grad_norm": 13.18405554746759, + "learning_rate": 9.864983064422014e-06, + "loss": 1.298, + "step": 1051 + }, + { + "epoch": 0.027317073170731707, + "grad_norm": 11.17914017484476, + "learning_rate": 9.866331511941864e-06, + "loss": 1.2495, + "step": 1052 + }, + { + "epoch": 0.027343039970323658, + "grad_norm": 6.638131999874204, + "learning_rate": 9.86767867827623e-06, + "loss": 1.2638, + "step": 1053 + }, + { + "epoch": 0.027369006769915608, + "grad_norm": 5.252718993410029, + "learning_rate": 9.869024565857359e-06, + "loss": 1.3059, + "step": 1054 + }, + { + "epoch": 0.02739497356950756, + "grad_norm": 6.303164084166287, + "learning_rate": 9.870369177110576e-06, + "loss": 1.3186, + "step": 1055 + }, + { + "epoch": 0.02742094036909951, + "grad_norm": 8.735537538633052, + "learning_rate": 9.871712514454312e-06, + "loss": 1.3082, + "step": 1056 + }, + { + "epoch": 0.02744690716869146, + "grad_norm": 5.4769325294594395, + "learning_rate": 9.873054580300139e-06, + "loss": 1.2345, + "step": 1057 + }, + { + "epoch": 0.02747287396828341, + "grad_norm": 6.196905520580886, + "learning_rate": 9.874395377052776e-06, + "loss": 1.2992, + "step": 1058 + }, + { + "epoch": 0.02749884076787536, + "grad_norm": 12.31199435827214, + "learning_rate": 9.875734907110135e-06, + "loss": 1.2388, + "step": 1059 + }, + { + "epoch": 0.02752480756746731, + "grad_norm": 7.614061357704357, + "learning_rate": 9.877073172863336e-06, + "loss": 1.3431, + "step": 1060 + }, + { + "epoch": 0.02755077436705926, + "grad_norm": 4.55036650638245, + "learning_rate": 9.878410176696732e-06, + "loss": 1.2793, + "step": 1061 + }, + { + "epoch": 0.02757674116665121, + "grad_norm": 7.171198395371673, + "learning_rate": 9.879745920987944e-06, + "loss": 1.236, + "step": 1062 + }, + { + "epoch": 0.02760270796624316, + "grad_norm": 12.267846012090597, + "learning_rate": 9.88108040810787e-06, + "loss": 1.2779, + "step": 1063 + }, + { + "epoch": 0.02762867476583511, + "grad_norm": 6.188181521711713, + "learning_rate": 9.88241364042073e-06, + "loss": 1.3212, + "step": 1064 + }, + { + "epoch": 0.02765464156542706, + "grad_norm": 5.5221304851121085, + "learning_rate": 9.883745620284072e-06, + "loss": 1.2987, + "step": 1065 + }, + { + "epoch": 0.02768060836501901, + "grad_norm": 7.4426418450075, + "learning_rate": 9.885076350048816e-06, + "loss": 1.3036, + "step": 1066 + }, + { + "epoch": 0.02770657516461096, + "grad_norm": 7.958376846061374, + "learning_rate": 9.886405832059261e-06, + "loss": 1.3301, + "step": 1067 + }, + { + "epoch": 0.02773254196420291, + "grad_norm": 7.713984079872183, + "learning_rate": 9.887734068653118e-06, + "loss": 1.2916, + "step": 1068 + }, + { + "epoch": 0.027758508763794862, + "grad_norm": 9.180510684500218, + "learning_rate": 9.88906106216154e-06, + "loss": 1.3304, + "step": 1069 + }, + { + "epoch": 0.027784475563386812, + "grad_norm": 4.6742500723660125, + "learning_rate": 9.890386814909139e-06, + "loss": 1.2808, + "step": 1070 + }, + { + "epoch": 0.027810442362978763, + "grad_norm": 6.290173924017107, + "learning_rate": 9.891711329214011e-06, + "loss": 1.3367, + "step": 1071 + }, + { + "epoch": 0.027836409162570713, + "grad_norm": 4.803530173750026, + "learning_rate": 9.89303460738776e-06, + "loss": 1.3495, + "step": 1072 + }, + { + "epoch": 0.027862375962162664, + "grad_norm": 4.918166152834282, + "learning_rate": 9.894356651735526e-06, + "loss": 1.2903, + "step": 1073 + }, + { + "epoch": 0.027888342761754614, + "grad_norm": 7.036266927612349, + "learning_rate": 9.895677464556009e-06, + "loss": 1.2709, + "step": 1074 + }, + { + "epoch": 0.027914309561346565, + "grad_norm": 8.30111993794338, + "learning_rate": 9.896997048141487e-06, + "loss": 1.346, + "step": 1075 + }, + { + "epoch": 0.027940276360938515, + "grad_norm": 5.7953790343562055, + "learning_rate": 9.89831540477785e-06, + "loss": 1.2898, + "step": 1076 + }, + { + "epoch": 0.027966243160530466, + "grad_norm": 8.365064506578461, + "learning_rate": 9.899632536744611e-06, + "loss": 1.2642, + "step": 1077 + }, + { + "epoch": 0.027992209960122416, + "grad_norm": 5.868753562008038, + "learning_rate": 9.900948446314938e-06, + "loss": 1.2714, + "step": 1078 + }, + { + "epoch": 0.028018176759714367, + "grad_norm": 30.79777673400813, + "learning_rate": 9.90226313575568e-06, + "loss": 1.2541, + "step": 1079 + }, + { + "epoch": 0.028044143559306317, + "grad_norm": 5.131079626565473, + "learning_rate": 9.903576607327382e-06, + "loss": 1.3216, + "step": 1080 + }, + { + "epoch": 0.028070110358898264, + "grad_norm": 7.652401077343971, + "learning_rate": 9.904888863284312e-06, + "loss": 1.356, + "step": 1081 + }, + { + "epoch": 0.028096077158490215, + "grad_norm": 5.2639260650010895, + "learning_rate": 9.906199905874489e-06, + "loss": 1.3191, + "step": 1082 + }, + { + "epoch": 0.028122043958082165, + "grad_norm": 10.588405645159488, + "learning_rate": 9.907509737339701e-06, + "loss": 1.2601, + "step": 1083 + }, + { + "epoch": 0.028148010757674116, + "grad_norm": 6.939560867619823, + "learning_rate": 9.908818359915526e-06, + "loss": 1.2829, + "step": 1084 + }, + { + "epoch": 0.028173977557266066, + "grad_norm": 8.117869373278994, + "learning_rate": 9.910125775831356e-06, + "loss": 1.2739, + "step": 1085 + }, + { + "epoch": 0.028199944356858016, + "grad_norm": 8.365096884449724, + "learning_rate": 9.911431987310429e-06, + "loss": 1.2724, + "step": 1086 + }, + { + "epoch": 0.028225911156449967, + "grad_norm": 16.42663354284076, + "learning_rate": 9.912736996569839e-06, + "loss": 1.301, + "step": 1087 + }, + { + "epoch": 0.028251877956041917, + "grad_norm": 11.20570367267512, + "learning_rate": 9.914040805820566e-06, + "loss": 1.3151, + "step": 1088 + }, + { + "epoch": 0.028277844755633868, + "grad_norm": 6.506893024311525, + "learning_rate": 9.915343417267494e-06, + "loss": 1.2834, + "step": 1089 + }, + { + "epoch": 0.02830381155522582, + "grad_norm": 5.713663476035565, + "learning_rate": 9.91664483310944e-06, + "loss": 1.3185, + "step": 1090 + }, + { + "epoch": 0.02832977835481777, + "grad_norm": 8.63059705381795, + "learning_rate": 9.917945055539166e-06, + "loss": 1.2822, + "step": 1091 + }, + { + "epoch": 0.02835574515440972, + "grad_norm": 11.057373928219873, + "learning_rate": 9.919244086743413e-06, + "loss": 1.3149, + "step": 1092 + }, + { + "epoch": 0.02838171195400167, + "grad_norm": 5.1049395332902625, + "learning_rate": 9.920541928902911e-06, + "loss": 1.3154, + "step": 1093 + }, + { + "epoch": 0.02840767875359362, + "grad_norm": 14.220859647753146, + "learning_rate": 9.921838584192414e-06, + "loss": 1.2625, + "step": 1094 + }, + { + "epoch": 0.02843364555318557, + "grad_norm": 6.253690316774968, + "learning_rate": 9.923134054780712e-06, + "loss": 1.2559, + "step": 1095 + }, + { + "epoch": 0.02845961235277752, + "grad_norm": 5.732942906474517, + "learning_rate": 9.924428342830654e-06, + "loss": 1.2729, + "step": 1096 + }, + { + "epoch": 0.02848557915236947, + "grad_norm": 12.446745689998023, + "learning_rate": 9.925721450499174e-06, + "loss": 1.3189, + "step": 1097 + }, + { + "epoch": 0.028511545951961422, + "grad_norm": 11.539834070979625, + "learning_rate": 9.927013379937308e-06, + "loss": 1.3065, + "step": 1098 + }, + { + "epoch": 0.028537512751553373, + "grad_norm": 6.907638203152674, + "learning_rate": 9.928304133290221e-06, + "loss": 1.2648, + "step": 1099 + }, + { + "epoch": 0.02856347955114532, + "grad_norm": 6.9482031386156775, + "learning_rate": 9.929593712697221e-06, + "loss": 1.3044, + "step": 1100 + }, + { + "epoch": 0.02858944635073727, + "grad_norm": 7.367737378405153, + "learning_rate": 9.930882120291786e-06, + "loss": 1.2944, + "step": 1101 + }, + { + "epoch": 0.02861541315032922, + "grad_norm": 8.556297115979206, + "learning_rate": 9.932169358201583e-06, + "loss": 1.3347, + "step": 1102 + }, + { + "epoch": 0.02864137994992117, + "grad_norm": 13.650762716929272, + "learning_rate": 9.933455428548492e-06, + "loss": 1.297, + "step": 1103 + }, + { + "epoch": 0.02866734674951312, + "grad_norm": 8.254009312980543, + "learning_rate": 9.934740333448617e-06, + "loss": 1.2905, + "step": 1104 + }, + { + "epoch": 0.028693313549105072, + "grad_norm": 7.556745625562825, + "learning_rate": 9.936024075012323e-06, + "loss": 1.3579, + "step": 1105 + }, + { + "epoch": 0.028719280348697022, + "grad_norm": 5.325919223300342, + "learning_rate": 9.937306655344245e-06, + "loss": 1.2271, + "step": 1106 + }, + { + "epoch": 0.028745247148288973, + "grad_norm": 5.61053471809934, + "learning_rate": 9.938588076543308e-06, + "loss": 1.2147, + "step": 1107 + }, + { + "epoch": 0.028771213947880923, + "grad_norm": 7.589505740270034, + "learning_rate": 9.939868340702757e-06, + "loss": 1.2103, + "step": 1108 + }, + { + "epoch": 0.028797180747472874, + "grad_norm": 6.291160243118104, + "learning_rate": 9.941147449910168e-06, + "loss": 1.2973, + "step": 1109 + }, + { + "epoch": 0.028823147547064824, + "grad_norm": 6.017742626463795, + "learning_rate": 9.942425406247476e-06, + "loss": 1.2366, + "step": 1110 + }, + { + "epoch": 0.028849114346656775, + "grad_norm": 7.604563356965189, + "learning_rate": 9.943702211790988e-06, + "loss": 1.3273, + "step": 1111 + }, + { + "epoch": 0.028875081146248725, + "grad_norm": 5.610686337515092, + "learning_rate": 9.944977868611409e-06, + "loss": 1.3184, + "step": 1112 + }, + { + "epoch": 0.028901047945840676, + "grad_norm": 6.508801590033893, + "learning_rate": 9.94625237877386e-06, + "loss": 1.2517, + "step": 1113 + }, + { + "epoch": 0.028927014745432626, + "grad_norm": 9.339764513727046, + "learning_rate": 9.947525744337894e-06, + "loss": 1.3482, + "step": 1114 + }, + { + "epoch": 0.028952981545024577, + "grad_norm": 5.796246188625047, + "learning_rate": 9.94879796735753e-06, + "loss": 1.2674, + "step": 1115 + }, + { + "epoch": 0.028978948344616527, + "grad_norm": 6.687083578959679, + "learning_rate": 9.950069049881252e-06, + "loss": 1.2835, + "step": 1116 + }, + { + "epoch": 0.029004915144208478, + "grad_norm": 8.015183345532403, + "learning_rate": 9.951338993952048e-06, + "loss": 1.264, + "step": 1117 + }, + { + "epoch": 0.029030881943800428, + "grad_norm": 8.818356563191688, + "learning_rate": 9.952607801607415e-06, + "loss": 1.2409, + "step": 1118 + }, + { + "epoch": 0.02905684874339238, + "grad_norm": 8.316739256191019, + "learning_rate": 9.953875474879389e-06, + "loss": 1.2932, + "step": 1119 + }, + { + "epoch": 0.029082815542984326, + "grad_norm": 5.2969860864204135, + "learning_rate": 9.955142015794563e-06, + "loss": 1.3126, + "step": 1120 + }, + { + "epoch": 0.029108782342576276, + "grad_norm": 8.004558694887704, + "learning_rate": 9.956407426374096e-06, + "loss": 1.3043, + "step": 1121 + }, + { + "epoch": 0.029134749142168227, + "grad_norm": 5.215054054162986, + "learning_rate": 9.957671708633747e-06, + "loss": 1.3041, + "step": 1122 + }, + { + "epoch": 0.029160715941760177, + "grad_norm": 4.984566137780617, + "learning_rate": 9.958934864583887e-06, + "loss": 1.2883, + "step": 1123 + }, + { + "epoch": 0.029186682741352128, + "grad_norm": 5.8515963992676285, + "learning_rate": 9.960196896229518e-06, + "loss": 1.2986, + "step": 1124 + }, + { + "epoch": 0.029212649540944078, + "grad_norm": 4.76953464917211, + "learning_rate": 9.96145780557029e-06, + "loss": 1.307, + "step": 1125 + }, + { + "epoch": 0.02923861634053603, + "grad_norm": 10.915810665514108, + "learning_rate": 9.962717594600529e-06, + "loss": 1.2524, + "step": 1126 + }, + { + "epoch": 0.02926458314012798, + "grad_norm": 8.335292687543234, + "learning_rate": 9.963976265309245e-06, + "loss": 1.318, + "step": 1127 + }, + { + "epoch": 0.02929054993971993, + "grad_norm": 7.8148951406637766, + "learning_rate": 9.965233819680155e-06, + "loss": 1.2543, + "step": 1128 + }, + { + "epoch": 0.02931651673931188, + "grad_norm": 62.18826357885485, + "learning_rate": 9.966490259691707e-06, + "loss": 1.2998, + "step": 1129 + }, + { + "epoch": 0.02934248353890383, + "grad_norm": 5.6153817915432, + "learning_rate": 9.967745587317093e-06, + "loss": 1.2744, + "step": 1130 + }, + { + "epoch": 0.02936845033849578, + "grad_norm": 5.683617912319774, + "learning_rate": 9.968999804524268e-06, + "loss": 1.3293, + "step": 1131 + }, + { + "epoch": 0.02939441713808773, + "grad_norm": 5.148275416229453, + "learning_rate": 9.970252913275966e-06, + "loss": 1.3069, + "step": 1132 + }, + { + "epoch": 0.029420383937679682, + "grad_norm": 5.18055112609583, + "learning_rate": 9.971504915529725e-06, + "loss": 1.2646, + "step": 1133 + }, + { + "epoch": 0.029446350737271632, + "grad_norm": 5.603866039659951, + "learning_rate": 9.972755813237904e-06, + "loss": 1.2675, + "step": 1134 + }, + { + "epoch": 0.029472317536863583, + "grad_norm": 13.454907306988348, + "learning_rate": 9.974005608347697e-06, + "loss": 1.2712, + "step": 1135 + }, + { + "epoch": 0.029498284336455533, + "grad_norm": 7.329232658825036, + "learning_rate": 9.975254302801151e-06, + "loss": 1.3403, + "step": 1136 + }, + { + "epoch": 0.029524251136047484, + "grad_norm": 4.739095720073457, + "learning_rate": 9.976501898535191e-06, + "loss": 1.2899, + "step": 1137 + }, + { + "epoch": 0.029550217935639434, + "grad_norm": 14.136039808233456, + "learning_rate": 9.977748397481631e-06, + "loss": 1.2836, + "step": 1138 + }, + { + "epoch": 0.02957618473523138, + "grad_norm": 5.581011369992656, + "learning_rate": 9.978993801567193e-06, + "loss": 1.2641, + "step": 1139 + }, + { + "epoch": 0.02960215153482333, + "grad_norm": 4.833082696839791, + "learning_rate": 9.980238112713535e-06, + "loss": 1.2963, + "step": 1140 + }, + { + "epoch": 0.029628118334415282, + "grad_norm": 5.94084250078561, + "learning_rate": 9.981481332837249e-06, + "loss": 1.2688, + "step": 1141 + }, + { + "epoch": 0.029654085134007233, + "grad_norm": 6.349469092094617, + "learning_rate": 9.982723463849895e-06, + "loss": 1.3008, + "step": 1142 + }, + { + "epoch": 0.029680051933599183, + "grad_norm": 7.966106380195767, + "learning_rate": 9.983964507658016e-06, + "loss": 1.2562, + "step": 1143 + }, + { + "epoch": 0.029706018733191134, + "grad_norm": 19.006271130475756, + "learning_rate": 9.98520446616315e-06, + "loss": 1.3013, + "step": 1144 + }, + { + "epoch": 0.029731985532783084, + "grad_norm": 5.407256633557118, + "learning_rate": 9.986443341261849e-06, + "loss": 1.3047, + "step": 1145 + }, + { + "epoch": 0.029757952332375034, + "grad_norm": 6.161438590174236, + "learning_rate": 9.987681134845703e-06, + "loss": 1.2815, + "step": 1146 + }, + { + "epoch": 0.029783919131966985, + "grad_norm": 11.131652900134036, + "learning_rate": 9.988917848801347e-06, + "loss": 1.3504, + "step": 1147 + }, + { + "epoch": 0.029809885931558935, + "grad_norm": 8.078324276736927, + "learning_rate": 9.990153485010489e-06, + "loss": 1.2852, + "step": 1148 + }, + { + "epoch": 0.029835852731150886, + "grad_norm": 5.826540488674399, + "learning_rate": 9.991388045349918e-06, + "loss": 1.2674, + "step": 1149 + }, + { + "epoch": 0.029861819530742836, + "grad_norm": 6.355159627738942, + "learning_rate": 9.992621531691526e-06, + "loss": 1.3169, + "step": 1150 + }, + { + "epoch": 0.029887786330334787, + "grad_norm": 6.400457973783588, + "learning_rate": 9.993853945902325e-06, + "loss": 1.2569, + "step": 1151 + }, + { + "epoch": 0.029913753129926737, + "grad_norm": 9.598019522785188, + "learning_rate": 9.99508528984446e-06, + "loss": 1.2778, + "step": 1152 + }, + { + "epoch": 0.029939719929518688, + "grad_norm": 5.07157141716493, + "learning_rate": 9.996315565375231e-06, + "loss": 1.2805, + "step": 1153 + }, + { + "epoch": 0.029965686729110638, + "grad_norm": 8.507910693734308, + "learning_rate": 9.997544774347108e-06, + "loss": 1.2873, + "step": 1154 + }, + { + "epoch": 0.02999165352870259, + "grad_norm": 4.759716936879111, + "learning_rate": 9.998772918607745e-06, + "loss": 1.2622, + "step": 1155 + }, + { + "epoch": 0.03001762032829454, + "grad_norm": 15.711628993981854, + "learning_rate": 1e-05, + "loss": 1.2975, + "step": 1156 + }, + { + "epoch": 0.03004358712788649, + "grad_norm": 5.704958769827323, + "learning_rate": 1e-05, + "loss": 1.2686, + "step": 1157 + }, + { + "epoch": 0.030069553927478437, + "grad_norm": 11.405167703361354, + "learning_rate": 1e-05, + "loss": 1.2969, + "step": 1158 + }, + { + "epoch": 0.030095520727070387, + "grad_norm": 5.8038110905721965, + "learning_rate": 1e-05, + "loss": 1.3172, + "step": 1159 + }, + { + "epoch": 0.030121487526662338, + "grad_norm": 5.209212307829466, + "learning_rate": 1e-05, + "loss": 1.2935, + "step": 1160 + }, + { + "epoch": 0.030147454326254288, + "grad_norm": 5.014955089668461, + "learning_rate": 1e-05, + "loss": 1.338, + "step": 1161 + }, + { + "epoch": 0.03017342112584624, + "grad_norm": 8.799306963593825, + "learning_rate": 1e-05, + "loss": 1.3115, + "step": 1162 + }, + { + "epoch": 0.03019938792543819, + "grad_norm": 4.537021092554795, + "learning_rate": 1e-05, + "loss": 1.2824, + "step": 1163 + }, + { + "epoch": 0.03022535472503014, + "grad_norm": 6.4757861510718575, + "learning_rate": 1e-05, + "loss": 1.3057, + "step": 1164 + }, + { + "epoch": 0.03025132152462209, + "grad_norm": 5.831442208510787, + "learning_rate": 1e-05, + "loss": 1.282, + "step": 1165 + }, + { + "epoch": 0.03027728832421404, + "grad_norm": 8.033116461436856, + "learning_rate": 1e-05, + "loss": 1.303, + "step": 1166 + }, + { + "epoch": 0.03030325512380599, + "grad_norm": 7.5971513830489705, + "learning_rate": 1e-05, + "loss": 1.2837, + "step": 1167 + }, + { + "epoch": 0.03032922192339794, + "grad_norm": 7.725414287235593, + "learning_rate": 1e-05, + "loss": 1.26, + "step": 1168 + }, + { + "epoch": 0.030355188722989892, + "grad_norm": 17.4647926392956, + "learning_rate": 1e-05, + "loss": 1.2975, + "step": 1169 + }, + { + "epoch": 0.030381155522581842, + "grad_norm": 10.514839947943068, + "learning_rate": 1e-05, + "loss": 1.3081, + "step": 1170 + }, + { + "epoch": 0.030407122322173793, + "grad_norm": 4.668977278931304, + "learning_rate": 1e-05, + "loss": 1.2813, + "step": 1171 + }, + { + "epoch": 0.030433089121765743, + "grad_norm": 9.247054997297619, + "learning_rate": 1e-05, + "loss": 1.3499, + "step": 1172 + }, + { + "epoch": 0.030459055921357694, + "grad_norm": 7.986505571289259, + "learning_rate": 1e-05, + "loss": 1.2754, + "step": 1173 + }, + { + "epoch": 0.030485022720949644, + "grad_norm": 6.241017409791946, + "learning_rate": 1e-05, + "loss": 1.2616, + "step": 1174 + }, + { + "epoch": 0.030510989520541595, + "grad_norm": 6.649446404151962, + "learning_rate": 1e-05, + "loss": 1.296, + "step": 1175 + }, + { + "epoch": 0.030536956320133545, + "grad_norm": 8.567301719470455, + "learning_rate": 1e-05, + "loss": 1.301, + "step": 1176 + }, + { + "epoch": 0.030562923119725492, + "grad_norm": 9.475958670478116, + "learning_rate": 1e-05, + "loss": 1.2707, + "step": 1177 + }, + { + "epoch": 0.030588889919317443, + "grad_norm": 11.780495535979894, + "learning_rate": 1e-05, + "loss": 1.259, + "step": 1178 + }, + { + "epoch": 0.030614856718909393, + "grad_norm": 7.469332748088021, + "learning_rate": 1e-05, + "loss": 1.2914, + "step": 1179 + }, + { + "epoch": 0.030640823518501344, + "grad_norm": 10.573077982656883, + "learning_rate": 1e-05, + "loss": 1.3129, + "step": 1180 + }, + { + "epoch": 0.030666790318093294, + "grad_norm": 8.475428801634878, + "learning_rate": 1e-05, + "loss": 1.3162, + "step": 1181 + }, + { + "epoch": 0.030692757117685245, + "grad_norm": 5.347192202176903, + "learning_rate": 1e-05, + "loss": 1.2702, + "step": 1182 + }, + { + "epoch": 0.030718723917277195, + "grad_norm": 5.672581147932818, + "learning_rate": 1e-05, + "loss": 1.3463, + "step": 1183 + }, + { + "epoch": 0.030744690716869146, + "grad_norm": 6.454212605079143, + "learning_rate": 1e-05, + "loss": 1.3169, + "step": 1184 + }, + { + "epoch": 0.030770657516461096, + "grad_norm": 12.429250463718388, + "learning_rate": 1e-05, + "loss": 1.2956, + "step": 1185 + }, + { + "epoch": 0.030796624316053046, + "grad_norm": 7.954239383110013, + "learning_rate": 1e-05, + "loss": 1.2694, + "step": 1186 + }, + { + "epoch": 0.030822591115644997, + "grad_norm": 7.826045873782921, + "learning_rate": 1e-05, + "loss": 1.2545, + "step": 1187 + }, + { + "epoch": 0.030848557915236947, + "grad_norm": 5.073646988933162, + "learning_rate": 1e-05, + "loss": 1.2706, + "step": 1188 + }, + { + "epoch": 0.030874524714828898, + "grad_norm": 9.106322102467423, + "learning_rate": 1e-05, + "loss": 1.3607, + "step": 1189 + }, + { + "epoch": 0.03090049151442085, + "grad_norm": 13.130781371784154, + "learning_rate": 1e-05, + "loss": 1.3482, + "step": 1190 + }, + { + "epoch": 0.0309264583140128, + "grad_norm": 11.873580687179455, + "learning_rate": 1e-05, + "loss": 1.2454, + "step": 1191 + }, + { + "epoch": 0.03095242511360475, + "grad_norm": 5.982146403427537, + "learning_rate": 1e-05, + "loss": 1.3366, + "step": 1192 + }, + { + "epoch": 0.0309783919131967, + "grad_norm": 8.986530397025325, + "learning_rate": 1e-05, + "loss": 1.2897, + "step": 1193 + }, + { + "epoch": 0.03100435871278865, + "grad_norm": 8.614994160743723, + "learning_rate": 1e-05, + "loss": 1.2946, + "step": 1194 + }, + { + "epoch": 0.0310303255123806, + "grad_norm": 6.8083867245669705, + "learning_rate": 1e-05, + "loss": 1.2818, + "step": 1195 + }, + { + "epoch": 0.031056292311972548, + "grad_norm": 7.190492753292799, + "learning_rate": 1e-05, + "loss": 1.283, + "step": 1196 + }, + { + "epoch": 0.031082259111564498, + "grad_norm": 7.464468395112331, + "learning_rate": 1e-05, + "loss": 1.2915, + "step": 1197 + }, + { + "epoch": 0.03110822591115645, + "grad_norm": 6.7693475433094275, + "learning_rate": 1e-05, + "loss": 1.2938, + "step": 1198 + }, + { + "epoch": 0.0311341927107484, + "grad_norm": 10.502012150836077, + "learning_rate": 1e-05, + "loss": 1.2797, + "step": 1199 + }, + { + "epoch": 0.03116015951034035, + "grad_norm": 43.00080480487763, + "learning_rate": 1e-05, + "loss": 1.2769, + "step": 1200 + }, + { + "epoch": 0.0311861263099323, + "grad_norm": 7.406563410183718, + "learning_rate": 1e-05, + "loss": 1.3208, + "step": 1201 + }, + { + "epoch": 0.03121209310952425, + "grad_norm": 9.641912488735297, + "learning_rate": 1e-05, + "loss": 1.2822, + "step": 1202 + }, + { + "epoch": 0.0312380599091162, + "grad_norm": 9.141513879056168, + "learning_rate": 1e-05, + "loss": 1.2793, + "step": 1203 + }, + { + "epoch": 0.031264026708708155, + "grad_norm": 6.081075310231813, + "learning_rate": 1e-05, + "loss": 1.3334, + "step": 1204 + }, + { + "epoch": 0.031289993508300105, + "grad_norm": 9.175164277895366, + "learning_rate": 1e-05, + "loss": 1.2577, + "step": 1205 + }, + { + "epoch": 0.03131596030789205, + "grad_norm": 8.852813928527189, + "learning_rate": 1e-05, + "loss": 1.2783, + "step": 1206 + }, + { + "epoch": 0.031341927107484, + "grad_norm": 8.815681478433714, + "learning_rate": 1e-05, + "loss": 1.3009, + "step": 1207 + }, + { + "epoch": 0.03136789390707595, + "grad_norm": 13.404310392906268, + "learning_rate": 1e-05, + "loss": 1.294, + "step": 1208 + }, + { + "epoch": 0.0313938607066679, + "grad_norm": 9.442077739102237, + "learning_rate": 1e-05, + "loss": 1.3175, + "step": 1209 + }, + { + "epoch": 0.03141982750625985, + "grad_norm": 7.9859395444565475, + "learning_rate": 1e-05, + "loss": 1.2515, + "step": 1210 + }, + { + "epoch": 0.0314457943058518, + "grad_norm": 8.829228533074552, + "learning_rate": 1e-05, + "loss": 1.347, + "step": 1211 + }, + { + "epoch": 0.03147176110544375, + "grad_norm": 5.398500389693376, + "learning_rate": 1e-05, + "loss": 1.3454, + "step": 1212 + }, + { + "epoch": 0.0314977279050357, + "grad_norm": 16.644715973071417, + "learning_rate": 1e-05, + "loss": 1.2729, + "step": 1213 + }, + { + "epoch": 0.03152369470462765, + "grad_norm": 7.646571180062837, + "learning_rate": 1e-05, + "loss": 1.2489, + "step": 1214 + }, + { + "epoch": 0.0315496615042196, + "grad_norm": 5.815216414146542, + "learning_rate": 1e-05, + "loss": 1.2655, + "step": 1215 + }, + { + "epoch": 0.031575628303811554, + "grad_norm": 4.933497871198202, + "learning_rate": 1e-05, + "loss": 1.2444, + "step": 1216 + }, + { + "epoch": 0.031601595103403504, + "grad_norm": 5.812606974612045, + "learning_rate": 1e-05, + "loss": 1.2935, + "step": 1217 + }, + { + "epoch": 0.031627561902995455, + "grad_norm": 5.931935162056789, + "learning_rate": 1e-05, + "loss": 1.2773, + "step": 1218 + }, + { + "epoch": 0.031653528702587405, + "grad_norm": 10.779004026133869, + "learning_rate": 1e-05, + "loss": 1.2677, + "step": 1219 + }, + { + "epoch": 0.031679495502179356, + "grad_norm": 8.341190245517366, + "learning_rate": 1e-05, + "loss": 1.3194, + "step": 1220 + }, + { + "epoch": 0.031705462301771306, + "grad_norm": 7.731110964907242, + "learning_rate": 1e-05, + "loss": 1.285, + "step": 1221 + }, + { + "epoch": 0.03173142910136326, + "grad_norm": 7.713904956725646, + "learning_rate": 1e-05, + "loss": 1.2747, + "step": 1222 + }, + { + "epoch": 0.03175739590095521, + "grad_norm": 8.559153550268292, + "learning_rate": 1e-05, + "loss": 1.2487, + "step": 1223 + }, + { + "epoch": 0.03178336270054716, + "grad_norm": 12.238991910023996, + "learning_rate": 1e-05, + "loss": 1.34, + "step": 1224 + }, + { + "epoch": 0.03180932950013911, + "grad_norm": 7.613851432920012, + "learning_rate": 1e-05, + "loss": 1.2953, + "step": 1225 + }, + { + "epoch": 0.03183529629973106, + "grad_norm": 7.6095532155561445, + "learning_rate": 1e-05, + "loss": 1.3063, + "step": 1226 + }, + { + "epoch": 0.03186126309932301, + "grad_norm": 9.693610251702616, + "learning_rate": 1e-05, + "loss": 1.2799, + "step": 1227 + }, + { + "epoch": 0.03188722989891496, + "grad_norm": 8.430606993410603, + "learning_rate": 1e-05, + "loss": 1.2884, + "step": 1228 + }, + { + "epoch": 0.03191319669850691, + "grad_norm": 19.016820589468452, + "learning_rate": 1e-05, + "loss": 1.2867, + "step": 1229 + }, + { + "epoch": 0.03193916349809886, + "grad_norm": 8.350690034233791, + "learning_rate": 1e-05, + "loss": 1.3015, + "step": 1230 + }, + { + "epoch": 0.03196513029769081, + "grad_norm": 4.953980958547024, + "learning_rate": 1e-05, + "loss": 1.2724, + "step": 1231 + }, + { + "epoch": 0.03199109709728276, + "grad_norm": 6.962629838208837, + "learning_rate": 1e-05, + "loss": 1.2381, + "step": 1232 + }, + { + "epoch": 0.03201706389687471, + "grad_norm": 8.829409561548632, + "learning_rate": 1e-05, + "loss": 1.3128, + "step": 1233 + }, + { + "epoch": 0.03204303069646666, + "grad_norm": 11.243175555955322, + "learning_rate": 1e-05, + "loss": 1.2739, + "step": 1234 + }, + { + "epoch": 0.03206899749605861, + "grad_norm": 7.7195048715993755, + "learning_rate": 1e-05, + "loss": 1.311, + "step": 1235 + }, + { + "epoch": 0.03209496429565056, + "grad_norm": 5.294238489305202, + "learning_rate": 1e-05, + "loss": 1.2405, + "step": 1236 + }, + { + "epoch": 0.032120931095242514, + "grad_norm": 7.307060370361224, + "learning_rate": 1e-05, + "loss": 1.2904, + "step": 1237 + }, + { + "epoch": 0.032146897894834464, + "grad_norm": 6.055877214262894, + "learning_rate": 1e-05, + "loss": 1.2839, + "step": 1238 + }, + { + "epoch": 0.032172864694426415, + "grad_norm": 7.8844218733124105, + "learning_rate": 1e-05, + "loss": 1.3695, + "step": 1239 + }, + { + "epoch": 0.032198831494018365, + "grad_norm": 8.69614205597602, + "learning_rate": 1e-05, + "loss": 1.2393, + "step": 1240 + }, + { + "epoch": 0.032224798293610316, + "grad_norm": 11.98923963666711, + "learning_rate": 1e-05, + "loss": 1.2746, + "step": 1241 + }, + { + "epoch": 0.032250765093202266, + "grad_norm": 5.17861177024725, + "learning_rate": 1e-05, + "loss": 1.2874, + "step": 1242 + }, + { + "epoch": 0.032276731892794217, + "grad_norm": 6.099947081789708, + "learning_rate": 1e-05, + "loss": 1.2706, + "step": 1243 + }, + { + "epoch": 0.03230269869238616, + "grad_norm": 12.756239019049964, + "learning_rate": 1e-05, + "loss": 1.2524, + "step": 1244 + }, + { + "epoch": 0.03232866549197811, + "grad_norm": 5.6104349394098305, + "learning_rate": 1e-05, + "loss": 1.2634, + "step": 1245 + }, + { + "epoch": 0.03235463229157006, + "grad_norm": 4.573522552231114, + "learning_rate": 1e-05, + "loss": 1.3003, + "step": 1246 + }, + { + "epoch": 0.03238059909116201, + "grad_norm": 5.285020373980933, + "learning_rate": 1e-05, + "loss": 1.2939, + "step": 1247 + }, + { + "epoch": 0.03240656589075396, + "grad_norm": 6.6978491448004345, + "learning_rate": 1e-05, + "loss": 1.2356, + "step": 1248 + }, + { + "epoch": 0.03243253269034591, + "grad_norm": 9.586709884454176, + "learning_rate": 1e-05, + "loss": 1.2865, + "step": 1249 + }, + { + "epoch": 0.03245849948993786, + "grad_norm": 5.087433250795417, + "learning_rate": 1e-05, + "loss": 1.3387, + "step": 1250 + }, + { + "epoch": 0.03248446628952981, + "grad_norm": 12.804145880666873, + "learning_rate": 1e-05, + "loss": 1.2915, + "step": 1251 + }, + { + "epoch": 0.032510433089121764, + "grad_norm": 5.876258877550928, + "learning_rate": 1e-05, + "loss": 1.3534, + "step": 1252 + }, + { + "epoch": 0.032536399888713714, + "grad_norm": 6.525342103605894, + "learning_rate": 1e-05, + "loss": 1.3328, + "step": 1253 + }, + { + "epoch": 0.032562366688305665, + "grad_norm": 6.913313485647102, + "learning_rate": 1e-05, + "loss": 1.2654, + "step": 1254 + }, + { + "epoch": 0.032588333487897615, + "grad_norm": 10.85038760850887, + "learning_rate": 1e-05, + "loss": 1.2485, + "step": 1255 + }, + { + "epoch": 0.032614300287489566, + "grad_norm": 34.537960674293764, + "learning_rate": 1e-05, + "loss": 1.2917, + "step": 1256 + }, + { + "epoch": 0.032640267087081516, + "grad_norm": 7.020000495149861, + "learning_rate": 1e-05, + "loss": 1.2845, + "step": 1257 + }, + { + "epoch": 0.03266623388667347, + "grad_norm": 6.901514147052997, + "learning_rate": 1e-05, + "loss": 1.2777, + "step": 1258 + }, + { + "epoch": 0.03269220068626542, + "grad_norm": 7.074747184846409, + "learning_rate": 1e-05, + "loss": 1.2814, + "step": 1259 + }, + { + "epoch": 0.03271816748585737, + "grad_norm": 5.774886964955056, + "learning_rate": 1e-05, + "loss": 1.303, + "step": 1260 + }, + { + "epoch": 0.03274413428544932, + "grad_norm": 5.152112005544501, + "learning_rate": 1e-05, + "loss": 1.2726, + "step": 1261 + }, + { + "epoch": 0.03277010108504127, + "grad_norm": 10.593549788392021, + "learning_rate": 1e-05, + "loss": 1.2408, + "step": 1262 + }, + { + "epoch": 0.03279606788463322, + "grad_norm": 5.882130293278905, + "learning_rate": 1e-05, + "loss": 1.2529, + "step": 1263 + }, + { + "epoch": 0.03282203468422517, + "grad_norm": 6.598534172735621, + "learning_rate": 1e-05, + "loss": 1.2716, + "step": 1264 + }, + { + "epoch": 0.03284800148381712, + "grad_norm": 9.357861531820744, + "learning_rate": 1e-05, + "loss": 1.325, + "step": 1265 + }, + { + "epoch": 0.03287396828340907, + "grad_norm": 5.509338773455769, + "learning_rate": 1e-05, + "loss": 1.275, + "step": 1266 + }, + { + "epoch": 0.03289993508300102, + "grad_norm": 6.328281428911475, + "learning_rate": 1e-05, + "loss": 1.2722, + "step": 1267 + }, + { + "epoch": 0.03292590188259297, + "grad_norm": 9.011185794410562, + "learning_rate": 1e-05, + "loss": 1.2919, + "step": 1268 + }, + { + "epoch": 0.03295186868218492, + "grad_norm": 5.396542516029994, + "learning_rate": 1e-05, + "loss": 1.2356, + "step": 1269 + }, + { + "epoch": 0.03297783548177687, + "grad_norm": 8.003985365946221, + "learning_rate": 1e-05, + "loss": 1.2617, + "step": 1270 + }, + { + "epoch": 0.03300380228136882, + "grad_norm": 7.115819537854547, + "learning_rate": 1e-05, + "loss": 1.3067, + "step": 1271 + }, + { + "epoch": 0.03302976908096077, + "grad_norm": 46.00386362850407, + "learning_rate": 1e-05, + "loss": 1.2708, + "step": 1272 + }, + { + "epoch": 0.033055735880552724, + "grad_norm": 7.253735073454857, + "learning_rate": 1e-05, + "loss": 1.3231, + "step": 1273 + }, + { + "epoch": 0.033081702680144674, + "grad_norm": 6.602340814880723, + "learning_rate": 1e-05, + "loss": 1.2554, + "step": 1274 + }, + { + "epoch": 0.033107669479736625, + "grad_norm": 36.37727689662961, + "learning_rate": 1e-05, + "loss": 1.3138, + "step": 1275 + }, + { + "epoch": 0.033133636279328575, + "grad_norm": 9.262308128521518, + "learning_rate": 1e-05, + "loss": 1.2304, + "step": 1276 + }, + { + "epoch": 0.033159603078920526, + "grad_norm": 19.235772758412416, + "learning_rate": 1e-05, + "loss": 1.2874, + "step": 1277 + }, + { + "epoch": 0.033185569878512476, + "grad_norm": 6.468657119747601, + "learning_rate": 1e-05, + "loss": 1.2965, + "step": 1278 + }, + { + "epoch": 0.03321153667810443, + "grad_norm": 6.397727097792017, + "learning_rate": 1e-05, + "loss": 1.2928, + "step": 1279 + }, + { + "epoch": 0.03323750347769638, + "grad_norm": 5.877681363778691, + "learning_rate": 1e-05, + "loss": 1.3142, + "step": 1280 + }, + { + "epoch": 0.03326347027728833, + "grad_norm": 19.19895925880082, + "learning_rate": 1e-05, + "loss": 1.295, + "step": 1281 + }, + { + "epoch": 0.03328943707688027, + "grad_norm": 15.210322559154214, + "learning_rate": 1e-05, + "loss": 1.2607, + "step": 1282 + }, + { + "epoch": 0.03331540387647222, + "grad_norm": 7.170934278836788, + "learning_rate": 1e-05, + "loss": 1.3358, + "step": 1283 + }, + { + "epoch": 0.03334137067606417, + "grad_norm": 6.645033290943255, + "learning_rate": 1e-05, + "loss": 1.262, + "step": 1284 + }, + { + "epoch": 0.03336733747565612, + "grad_norm": 10.013773016421636, + "learning_rate": 1e-05, + "loss": 1.2818, + "step": 1285 + }, + { + "epoch": 0.03339330427524807, + "grad_norm": 6.534796668977759, + "learning_rate": 1e-05, + "loss": 1.3251, + "step": 1286 + }, + { + "epoch": 0.03341927107484002, + "grad_norm": 6.92043149892748, + "learning_rate": 1e-05, + "loss": 1.3102, + "step": 1287 + }, + { + "epoch": 0.033445237874431974, + "grad_norm": 27.43758786823914, + "learning_rate": 1e-05, + "loss": 1.2839, + "step": 1288 + }, + { + "epoch": 0.033471204674023924, + "grad_norm": 7.785404674128468, + "learning_rate": 1e-05, + "loss": 1.2603, + "step": 1289 + }, + { + "epoch": 0.033497171473615875, + "grad_norm": 10.682396813663603, + "learning_rate": 1e-05, + "loss": 1.3102, + "step": 1290 + }, + { + "epoch": 0.033523138273207825, + "grad_norm": 8.896157418868501, + "learning_rate": 1e-05, + "loss": 1.2859, + "step": 1291 + }, + { + "epoch": 0.033549105072799776, + "grad_norm": 25.194561519815764, + "learning_rate": 1e-05, + "loss": 1.2386, + "step": 1292 + }, + { + "epoch": 0.033575071872391726, + "grad_norm": 7.877894520518539, + "learning_rate": 1e-05, + "loss": 1.278, + "step": 1293 + }, + { + "epoch": 0.03360103867198368, + "grad_norm": 8.789514745309198, + "learning_rate": 1e-05, + "loss": 1.272, + "step": 1294 + }, + { + "epoch": 0.03362700547157563, + "grad_norm": 7.074288311536793, + "learning_rate": 1e-05, + "loss": 1.2737, + "step": 1295 + }, + { + "epoch": 0.03365297227116758, + "grad_norm": 7.906679164411286, + "learning_rate": 1e-05, + "loss": 1.3085, + "step": 1296 + }, + { + "epoch": 0.03367893907075953, + "grad_norm": 8.254794952600134, + "learning_rate": 1e-05, + "loss": 1.2795, + "step": 1297 + }, + { + "epoch": 0.03370490587035148, + "grad_norm": 6.320909092381442, + "learning_rate": 1e-05, + "loss": 1.3044, + "step": 1298 + }, + { + "epoch": 0.03373087266994343, + "grad_norm": 6.757360895314919, + "learning_rate": 1e-05, + "loss": 1.2782, + "step": 1299 + }, + { + "epoch": 0.03375683946953538, + "grad_norm": 13.089433418292561, + "learning_rate": 1e-05, + "loss": 1.2691, + "step": 1300 + }, + { + "epoch": 0.03378280626912733, + "grad_norm": 6.344175728479478, + "learning_rate": 1e-05, + "loss": 1.2876, + "step": 1301 + }, + { + "epoch": 0.03380877306871928, + "grad_norm": 6.69918145261119, + "learning_rate": 1e-05, + "loss": 1.3131, + "step": 1302 + }, + { + "epoch": 0.03383473986831123, + "grad_norm": 8.060469497372123, + "learning_rate": 1e-05, + "loss": 1.2954, + "step": 1303 + }, + { + "epoch": 0.03386070666790318, + "grad_norm": 6.988018135418475, + "learning_rate": 1e-05, + "loss": 1.2601, + "step": 1304 + }, + { + "epoch": 0.03388667346749513, + "grad_norm": 9.599174432224546, + "learning_rate": 1e-05, + "loss": 1.3192, + "step": 1305 + }, + { + "epoch": 0.03391264026708708, + "grad_norm": 7.41644281056642, + "learning_rate": 1e-05, + "loss": 1.2779, + "step": 1306 + }, + { + "epoch": 0.03393860706667903, + "grad_norm": 6.4957437518461605, + "learning_rate": 1e-05, + "loss": 1.2446, + "step": 1307 + }, + { + "epoch": 0.03396457386627098, + "grad_norm": 31.36272637174182, + "learning_rate": 1e-05, + "loss": 1.3113, + "step": 1308 + }, + { + "epoch": 0.033990540665862934, + "grad_norm": 6.295854433715658, + "learning_rate": 1e-05, + "loss": 1.2611, + "step": 1309 + }, + { + "epoch": 0.034016507465454884, + "grad_norm": 7.062798417800024, + "learning_rate": 1e-05, + "loss": 1.296, + "step": 1310 + }, + { + "epoch": 0.034042474265046835, + "grad_norm": 8.709148411426119, + "learning_rate": 1e-05, + "loss": 1.2964, + "step": 1311 + }, + { + "epoch": 0.034068441064638785, + "grad_norm": 9.029187240776652, + "learning_rate": 1e-05, + "loss": 1.3176, + "step": 1312 + }, + { + "epoch": 0.034094407864230736, + "grad_norm": 6.8283831026018476, + "learning_rate": 1e-05, + "loss": 1.2983, + "step": 1313 + }, + { + "epoch": 0.034120374663822686, + "grad_norm": 6.415306574690381, + "learning_rate": 1e-05, + "loss": 1.2572, + "step": 1314 + }, + { + "epoch": 0.03414634146341464, + "grad_norm": 10.424017873191003, + "learning_rate": 1e-05, + "loss": 1.2812, + "step": 1315 + }, + { + "epoch": 0.03417230826300659, + "grad_norm": 8.390611815486615, + "learning_rate": 1e-05, + "loss": 1.2989, + "step": 1316 + }, + { + "epoch": 0.03419827506259854, + "grad_norm": 5.811672356860467, + "learning_rate": 1e-05, + "loss": 1.2801, + "step": 1317 + }, + { + "epoch": 0.03422424186219049, + "grad_norm": 19.12354307920851, + "learning_rate": 1e-05, + "loss": 1.2185, + "step": 1318 + }, + { + "epoch": 0.03425020866178244, + "grad_norm": 7.2307876689694535, + "learning_rate": 1e-05, + "loss": 1.2727, + "step": 1319 + }, + { + "epoch": 0.03427617546137438, + "grad_norm": 10.61963257694615, + "learning_rate": 1e-05, + "loss": 1.2888, + "step": 1320 + }, + { + "epoch": 0.03430214226096633, + "grad_norm": 44.449469862910064, + "learning_rate": 1e-05, + "loss": 1.279, + "step": 1321 + }, + { + "epoch": 0.03432810906055828, + "grad_norm": 8.987658622095314, + "learning_rate": 1e-05, + "loss": 1.3048, + "step": 1322 + }, + { + "epoch": 0.034354075860150234, + "grad_norm": 66.08941696789472, + "learning_rate": 1e-05, + "loss": 1.2854, + "step": 1323 + }, + { + "epoch": 0.034380042659742184, + "grad_norm": 5.887929770299614, + "learning_rate": 1e-05, + "loss": 1.3075, + "step": 1324 + }, + { + "epoch": 0.034406009459334135, + "grad_norm": 6.373643020118117, + "learning_rate": 1e-05, + "loss": 1.313, + "step": 1325 + }, + { + "epoch": 0.034431976258926085, + "grad_norm": 8.354395786355715, + "learning_rate": 1e-05, + "loss": 1.3093, + "step": 1326 + }, + { + "epoch": 0.034457943058518035, + "grad_norm": 7.288814168381656, + "learning_rate": 1e-05, + "loss": 1.2571, + "step": 1327 + }, + { + "epoch": 0.034483909858109986, + "grad_norm": 9.323224133299922, + "learning_rate": 1e-05, + "loss": 1.2884, + "step": 1328 + }, + { + "epoch": 0.034509876657701936, + "grad_norm": 13.656385198486937, + "learning_rate": 1e-05, + "loss": 1.3294, + "step": 1329 + }, + { + "epoch": 0.03453584345729389, + "grad_norm": 7.513107290492215, + "learning_rate": 1e-05, + "loss": 1.3025, + "step": 1330 + }, + { + "epoch": 0.03456181025688584, + "grad_norm": 6.667980191847242, + "learning_rate": 1e-05, + "loss": 1.2485, + "step": 1331 + }, + { + "epoch": 0.03458777705647779, + "grad_norm": 7.939111260705651, + "learning_rate": 1e-05, + "loss": 1.2629, + "step": 1332 + }, + { + "epoch": 0.03461374385606974, + "grad_norm": 6.163674462001961, + "learning_rate": 1e-05, + "loss": 1.3, + "step": 1333 + }, + { + "epoch": 0.03463971065566169, + "grad_norm": 6.864736490658737, + "learning_rate": 1e-05, + "loss": 1.3206, + "step": 1334 + }, + { + "epoch": 0.03466567745525364, + "grad_norm": 4.228491718321565, + "learning_rate": 1e-05, + "loss": 1.258, + "step": 1335 + }, + { + "epoch": 0.03469164425484559, + "grad_norm": 5.717942811459134, + "learning_rate": 1e-05, + "loss": 1.264, + "step": 1336 + }, + { + "epoch": 0.03471761105443754, + "grad_norm": 6.029926688537816, + "learning_rate": 1e-05, + "loss": 1.2607, + "step": 1337 + }, + { + "epoch": 0.03474357785402949, + "grad_norm": 6.445534738119308, + "learning_rate": 1e-05, + "loss": 1.2639, + "step": 1338 + }, + { + "epoch": 0.03476954465362144, + "grad_norm": 8.714780717499526, + "learning_rate": 1e-05, + "loss": 1.3112, + "step": 1339 + }, + { + "epoch": 0.03479551145321339, + "grad_norm": 7.917055970374158, + "learning_rate": 1e-05, + "loss": 1.3202, + "step": 1340 + }, + { + "epoch": 0.03482147825280534, + "grad_norm": 6.418837671321907, + "learning_rate": 1e-05, + "loss": 1.3004, + "step": 1341 + }, + { + "epoch": 0.03484744505239729, + "grad_norm": 5.3144664042235785, + "learning_rate": 1e-05, + "loss": 1.2741, + "step": 1342 + }, + { + "epoch": 0.03487341185198924, + "grad_norm": 7.45483839385641, + "learning_rate": 1e-05, + "loss": 1.2517, + "step": 1343 + }, + { + "epoch": 0.034899378651581194, + "grad_norm": 5.368975079295825, + "learning_rate": 1e-05, + "loss": 1.2748, + "step": 1344 + }, + { + "epoch": 0.034925345451173144, + "grad_norm": 5.83749495512195, + "learning_rate": 1e-05, + "loss": 1.2812, + "step": 1345 + }, + { + "epoch": 0.034951312250765094, + "grad_norm": 5.948602836120789, + "learning_rate": 1e-05, + "loss": 1.2683, + "step": 1346 + }, + { + "epoch": 0.034977279050357045, + "grad_norm": 5.544056006827126, + "learning_rate": 1e-05, + "loss": 1.2989, + "step": 1347 + }, + { + "epoch": 0.035003245849948995, + "grad_norm": 5.2454461829246695, + "learning_rate": 1e-05, + "loss": 1.2333, + "step": 1348 + }, + { + "epoch": 0.035029212649540946, + "grad_norm": 6.047484374576109, + "learning_rate": 1e-05, + "loss": 1.2236, + "step": 1349 + }, + { + "epoch": 0.035055179449132896, + "grad_norm": 14.190101418646886, + "learning_rate": 1e-05, + "loss": 1.324, + "step": 1350 + }, + { + "epoch": 0.03508114624872485, + "grad_norm": 5.513903643729548, + "learning_rate": 1e-05, + "loss": 1.2941, + "step": 1351 + }, + { + "epoch": 0.0351071130483168, + "grad_norm": 8.444230171742086, + "learning_rate": 1e-05, + "loss": 1.2822, + "step": 1352 + }, + { + "epoch": 0.03513307984790875, + "grad_norm": 7.544519408272912, + "learning_rate": 1e-05, + "loss": 1.3309, + "step": 1353 + }, + { + "epoch": 0.0351590466475007, + "grad_norm": 23.154551167018845, + "learning_rate": 1e-05, + "loss": 1.311, + "step": 1354 + }, + { + "epoch": 0.03518501344709265, + "grad_norm": 10.89652060875182, + "learning_rate": 1e-05, + "loss": 1.316, + "step": 1355 + }, + { + "epoch": 0.0352109802466846, + "grad_norm": 8.023164590356185, + "learning_rate": 1e-05, + "loss": 1.2967, + "step": 1356 + }, + { + "epoch": 0.03523694704627655, + "grad_norm": 15.070377300275407, + "learning_rate": 1e-05, + "loss": 1.2911, + "step": 1357 + }, + { + "epoch": 0.0352629138458685, + "grad_norm": 5.098410130438144, + "learning_rate": 1e-05, + "loss": 1.267, + "step": 1358 + }, + { + "epoch": 0.035288880645460444, + "grad_norm": 31.215906206499415, + "learning_rate": 1e-05, + "loss": 1.3349, + "step": 1359 + }, + { + "epoch": 0.035314847445052394, + "grad_norm": 6.3400459186804445, + "learning_rate": 1e-05, + "loss": 1.2551, + "step": 1360 + }, + { + "epoch": 0.035340814244644345, + "grad_norm": 10.067378975969902, + "learning_rate": 1e-05, + "loss": 1.2932, + "step": 1361 + }, + { + "epoch": 0.035366781044236295, + "grad_norm": 6.16290512146121, + "learning_rate": 1e-05, + "loss": 1.2801, + "step": 1362 + }, + { + "epoch": 0.035392747843828246, + "grad_norm": 7.364503276413202, + "learning_rate": 1e-05, + "loss": 1.2806, + "step": 1363 + }, + { + "epoch": 0.035418714643420196, + "grad_norm": 6.001834271114117, + "learning_rate": 1e-05, + "loss": 1.2725, + "step": 1364 + }, + { + "epoch": 0.035444681443012147, + "grad_norm": 6.286827230388621, + "learning_rate": 1e-05, + "loss": 1.306, + "step": 1365 + }, + { + "epoch": 0.0354706482426041, + "grad_norm": 14.887464414052767, + "learning_rate": 1e-05, + "loss": 1.2591, + "step": 1366 + }, + { + "epoch": 0.03549661504219605, + "grad_norm": 6.4016519560277, + "learning_rate": 1e-05, + "loss": 1.2636, + "step": 1367 + }, + { + "epoch": 0.035522581841788, + "grad_norm": 6.615003132952744, + "learning_rate": 1e-05, + "loss": 1.3254, + "step": 1368 + }, + { + "epoch": 0.03554854864137995, + "grad_norm": 6.365226883721428, + "learning_rate": 1e-05, + "loss": 1.2617, + "step": 1369 + }, + { + "epoch": 0.0355745154409719, + "grad_norm": 5.387531310519334, + "learning_rate": 1e-05, + "loss": 1.2957, + "step": 1370 + }, + { + "epoch": 0.03560048224056385, + "grad_norm": 8.066719784312165, + "learning_rate": 1e-05, + "loss": 1.243, + "step": 1371 + }, + { + "epoch": 0.0356264490401558, + "grad_norm": 6.08541158771488, + "learning_rate": 1e-05, + "loss": 1.3317, + "step": 1372 + }, + { + "epoch": 0.03565241583974775, + "grad_norm": 6.01246936814085, + "learning_rate": 1e-05, + "loss": 1.2753, + "step": 1373 + }, + { + "epoch": 0.0356783826393397, + "grad_norm": 7.101684972128753, + "learning_rate": 1e-05, + "loss": 1.31, + "step": 1374 + }, + { + "epoch": 0.03570434943893165, + "grad_norm": 6.467450144914387, + "learning_rate": 1e-05, + "loss": 1.2733, + "step": 1375 + }, + { + "epoch": 0.0357303162385236, + "grad_norm": 6.2344759595913635, + "learning_rate": 1e-05, + "loss": 1.2791, + "step": 1376 + }, + { + "epoch": 0.03575628303811555, + "grad_norm": 15.06135675040884, + "learning_rate": 1e-05, + "loss": 1.308, + "step": 1377 + }, + { + "epoch": 0.0357822498377075, + "grad_norm": 6.847727834498194, + "learning_rate": 1e-05, + "loss": 1.2951, + "step": 1378 + }, + { + "epoch": 0.03580821663729945, + "grad_norm": 16.621913996358654, + "learning_rate": 1e-05, + "loss": 1.2541, + "step": 1379 + }, + { + "epoch": 0.035834183436891404, + "grad_norm": 5.994597864097815, + "learning_rate": 1e-05, + "loss": 1.2488, + "step": 1380 + }, + { + "epoch": 0.035860150236483354, + "grad_norm": 6.220765673743623, + "learning_rate": 1e-05, + "loss": 1.2848, + "step": 1381 + }, + { + "epoch": 0.035886117036075305, + "grad_norm": 9.416822404051864, + "learning_rate": 1e-05, + "loss": 1.2891, + "step": 1382 + }, + { + "epoch": 0.035912083835667255, + "grad_norm": 22.42883099309451, + "learning_rate": 1e-05, + "loss": 1.2797, + "step": 1383 + }, + { + "epoch": 0.035938050635259206, + "grad_norm": 7.98351042782689, + "learning_rate": 1e-05, + "loss": 1.2823, + "step": 1384 + }, + { + "epoch": 0.035964017434851156, + "grad_norm": 6.81895742686551, + "learning_rate": 1e-05, + "loss": 1.2781, + "step": 1385 + }, + { + "epoch": 0.035989984234443106, + "grad_norm": 7.810165422668291, + "learning_rate": 1e-05, + "loss": 1.1889, + "step": 1386 + }, + { + "epoch": 0.03601595103403506, + "grad_norm": 5.895842571273773, + "learning_rate": 1e-05, + "loss": 1.2769, + "step": 1387 + }, + { + "epoch": 0.03604191783362701, + "grad_norm": 6.526722776323092, + "learning_rate": 1e-05, + "loss": 1.2288, + "step": 1388 + }, + { + "epoch": 0.03606788463321896, + "grad_norm": 12.14416625074695, + "learning_rate": 1e-05, + "loss": 1.3107, + "step": 1389 + }, + { + "epoch": 0.03609385143281091, + "grad_norm": 7.416112329004734, + "learning_rate": 1e-05, + "loss": 1.2244, + "step": 1390 + }, + { + "epoch": 0.03611981823240286, + "grad_norm": 4.8878074061509444, + "learning_rate": 1e-05, + "loss": 1.2709, + "step": 1391 + }, + { + "epoch": 0.03614578503199481, + "grad_norm": 7.941081523119879, + "learning_rate": 1e-05, + "loss": 1.2672, + "step": 1392 + }, + { + "epoch": 0.03617175183158676, + "grad_norm": 9.067869082360211, + "learning_rate": 1e-05, + "loss": 1.2382, + "step": 1393 + }, + { + "epoch": 0.03619771863117871, + "grad_norm": 7.985763518651806, + "learning_rate": 1e-05, + "loss": 1.3221, + "step": 1394 + }, + { + "epoch": 0.03622368543077066, + "grad_norm": 6.887612443341868, + "learning_rate": 1e-05, + "loss": 1.2917, + "step": 1395 + }, + { + "epoch": 0.03624965223036261, + "grad_norm": 14.971962219761341, + "learning_rate": 1e-05, + "loss": 1.2713, + "step": 1396 + }, + { + "epoch": 0.036275619029954555, + "grad_norm": 5.199144278321832, + "learning_rate": 1e-05, + "loss": 1.2148, + "step": 1397 + }, + { + "epoch": 0.036301585829546505, + "grad_norm": 5.342294238157496, + "learning_rate": 1e-05, + "loss": 1.3212, + "step": 1398 + }, + { + "epoch": 0.036327552629138456, + "grad_norm": 10.385855637088989, + "learning_rate": 1e-05, + "loss": 1.2486, + "step": 1399 + }, + { + "epoch": 0.036353519428730406, + "grad_norm": 5.433797342968647, + "learning_rate": 1e-05, + "loss": 1.2774, + "step": 1400 + }, + { + "epoch": 0.03637948622832236, + "grad_norm": 5.802861251137336, + "learning_rate": 1e-05, + "loss": 1.2568, + "step": 1401 + }, + { + "epoch": 0.03640545302791431, + "grad_norm": 5.069784128081459, + "learning_rate": 1e-05, + "loss": 1.3167, + "step": 1402 + }, + { + "epoch": 0.03643141982750626, + "grad_norm": 4.982687924712737, + "learning_rate": 1e-05, + "loss": 1.2826, + "step": 1403 + }, + { + "epoch": 0.03645738662709821, + "grad_norm": 5.4853659057972965, + "learning_rate": 1e-05, + "loss": 1.231, + "step": 1404 + }, + { + "epoch": 0.03648335342669016, + "grad_norm": 9.900111080278618, + "learning_rate": 1e-05, + "loss": 1.3256, + "step": 1405 + }, + { + "epoch": 0.03650932022628211, + "grad_norm": 5.926785086674145, + "learning_rate": 1e-05, + "loss": 1.2859, + "step": 1406 + }, + { + "epoch": 0.03653528702587406, + "grad_norm": 7.438520345500167, + "learning_rate": 1e-05, + "loss": 1.2069, + "step": 1407 + }, + { + "epoch": 0.03656125382546601, + "grad_norm": 22.967430606237688, + "learning_rate": 1e-05, + "loss": 1.2574, + "step": 1408 + }, + { + "epoch": 0.03658722062505796, + "grad_norm": 7.870748431863742, + "learning_rate": 1e-05, + "loss": 1.2345, + "step": 1409 + }, + { + "epoch": 0.03661318742464991, + "grad_norm": 26.012742661275045, + "learning_rate": 1e-05, + "loss": 1.2422, + "step": 1410 + }, + { + "epoch": 0.03663915422424186, + "grad_norm": 5.980456949592375, + "learning_rate": 1e-05, + "loss": 1.2297, + "step": 1411 + }, + { + "epoch": 0.03666512102383381, + "grad_norm": 9.626887148853351, + "learning_rate": 1e-05, + "loss": 1.286, + "step": 1412 + }, + { + "epoch": 0.03669108782342576, + "grad_norm": 4.69479230134919, + "learning_rate": 1e-05, + "loss": 1.3, + "step": 1413 + }, + { + "epoch": 0.03671705462301771, + "grad_norm": 33.555599703840414, + "learning_rate": 1e-05, + "loss": 1.2996, + "step": 1414 + }, + { + "epoch": 0.03674302142260966, + "grad_norm": 8.228200142673405, + "learning_rate": 1e-05, + "loss": 1.265, + "step": 1415 + }, + { + "epoch": 0.036768988222201614, + "grad_norm": 7.377671356020894, + "learning_rate": 1e-05, + "loss": 1.2975, + "step": 1416 + }, + { + "epoch": 0.036794955021793564, + "grad_norm": 5.1031794417234355, + "learning_rate": 1e-05, + "loss": 1.2498, + "step": 1417 + }, + { + "epoch": 0.036820921821385515, + "grad_norm": 6.134105531484336, + "learning_rate": 1e-05, + "loss": 1.279, + "step": 1418 + }, + { + "epoch": 0.036846888620977465, + "grad_norm": 24.025969124736935, + "learning_rate": 1e-05, + "loss": 1.2745, + "step": 1419 + }, + { + "epoch": 0.036872855420569416, + "grad_norm": 6.122576798642336, + "learning_rate": 1e-05, + "loss": 1.2846, + "step": 1420 + }, + { + "epoch": 0.036898822220161366, + "grad_norm": 6.7004810445241425, + "learning_rate": 1e-05, + "loss": 1.3152, + "step": 1421 + }, + { + "epoch": 0.03692478901975332, + "grad_norm": 5.777547444537081, + "learning_rate": 1e-05, + "loss": 1.2586, + "step": 1422 + }, + { + "epoch": 0.03695075581934527, + "grad_norm": 9.132151688293828, + "learning_rate": 1e-05, + "loss": 1.3147, + "step": 1423 + }, + { + "epoch": 0.03697672261893722, + "grad_norm": 14.582955201105877, + "learning_rate": 1e-05, + "loss": 1.3026, + "step": 1424 + }, + { + "epoch": 0.03700268941852917, + "grad_norm": 5.062321224117199, + "learning_rate": 1e-05, + "loss": 1.2416, + "step": 1425 + }, + { + "epoch": 0.03702865621812112, + "grad_norm": 10.010121373251504, + "learning_rate": 1e-05, + "loss": 1.2497, + "step": 1426 + }, + { + "epoch": 0.03705462301771307, + "grad_norm": 6.21851260244895, + "learning_rate": 1e-05, + "loss": 1.286, + "step": 1427 + }, + { + "epoch": 0.03708058981730502, + "grad_norm": 5.777751792131986, + "learning_rate": 1e-05, + "loss": 1.2638, + "step": 1428 + }, + { + "epoch": 0.03710655661689697, + "grad_norm": 7.849138696155418, + "learning_rate": 1e-05, + "loss": 1.2964, + "step": 1429 + }, + { + "epoch": 0.03713252341648892, + "grad_norm": 21.727579421544664, + "learning_rate": 1e-05, + "loss": 1.3052, + "step": 1430 + }, + { + "epoch": 0.03715849021608087, + "grad_norm": 7.168280604592141, + "learning_rate": 1e-05, + "loss": 1.2897, + "step": 1431 + }, + { + "epoch": 0.03718445701567282, + "grad_norm": 6.860893620155457, + "learning_rate": 1e-05, + "loss": 1.2532, + "step": 1432 + }, + { + "epoch": 0.03721042381526477, + "grad_norm": 5.91605366858022, + "learning_rate": 1e-05, + "loss": 1.2872, + "step": 1433 + }, + { + "epoch": 0.03723639061485672, + "grad_norm": 13.10571499861797, + "learning_rate": 1e-05, + "loss": 1.2533, + "step": 1434 + }, + { + "epoch": 0.037262357414448666, + "grad_norm": 5.526346738150228, + "learning_rate": 1e-05, + "loss": 1.259, + "step": 1435 + }, + { + "epoch": 0.037288324214040616, + "grad_norm": 4.281932247404471, + "learning_rate": 1e-05, + "loss": 1.2192, + "step": 1436 + }, + { + "epoch": 0.03731429101363257, + "grad_norm": 11.853483925697027, + "learning_rate": 1e-05, + "loss": 1.2821, + "step": 1437 + }, + { + "epoch": 0.03734025781322452, + "grad_norm": 4.771696032132735, + "learning_rate": 1e-05, + "loss": 1.2678, + "step": 1438 + }, + { + "epoch": 0.03736622461281647, + "grad_norm": 5.6152729278567834, + "learning_rate": 1e-05, + "loss": 1.3128, + "step": 1439 + }, + { + "epoch": 0.03739219141240842, + "grad_norm": 5.658687508930889, + "learning_rate": 1e-05, + "loss": 1.2888, + "step": 1440 + }, + { + "epoch": 0.03741815821200037, + "grad_norm": 10.231299484217153, + "learning_rate": 1e-05, + "loss": 1.3061, + "step": 1441 + }, + { + "epoch": 0.03744412501159232, + "grad_norm": 7.139787286747414, + "learning_rate": 1e-05, + "loss": 1.3298, + "step": 1442 + }, + { + "epoch": 0.03747009181118427, + "grad_norm": 7.868600333974291, + "learning_rate": 1e-05, + "loss": 1.3014, + "step": 1443 + }, + { + "epoch": 0.03749605861077622, + "grad_norm": 4.475182019160143, + "learning_rate": 1e-05, + "loss": 1.2558, + "step": 1444 + }, + { + "epoch": 0.03752202541036817, + "grad_norm": 7.16359949643762, + "learning_rate": 1e-05, + "loss": 1.2767, + "step": 1445 + }, + { + "epoch": 0.03754799220996012, + "grad_norm": 4.6829685942504815, + "learning_rate": 1e-05, + "loss": 1.2677, + "step": 1446 + }, + { + "epoch": 0.03757395900955207, + "grad_norm": 6.6403541419945, + "learning_rate": 1e-05, + "loss": 1.2828, + "step": 1447 + }, + { + "epoch": 0.03759992580914402, + "grad_norm": 8.121197559010646, + "learning_rate": 1e-05, + "loss": 1.2488, + "step": 1448 + }, + { + "epoch": 0.03762589260873597, + "grad_norm": 3.6836650756532414, + "learning_rate": 1e-05, + "loss": 1.3311, + "step": 1449 + }, + { + "epoch": 0.03765185940832792, + "grad_norm": 6.9600940540141165, + "learning_rate": 1e-05, + "loss": 1.3126, + "step": 1450 + }, + { + "epoch": 0.03767782620791987, + "grad_norm": 7.134632225261937, + "learning_rate": 1e-05, + "loss": 1.2938, + "step": 1451 + }, + { + "epoch": 0.037703793007511824, + "grad_norm": 5.951107454217293, + "learning_rate": 1e-05, + "loss": 1.2828, + "step": 1452 + }, + { + "epoch": 0.037729759807103774, + "grad_norm": 42.6941381071329, + "learning_rate": 1e-05, + "loss": 1.3023, + "step": 1453 + }, + { + "epoch": 0.037755726606695725, + "grad_norm": 5.864211220282229, + "learning_rate": 1e-05, + "loss": 1.2938, + "step": 1454 + }, + { + "epoch": 0.037781693406287675, + "grad_norm": 8.051684318285716, + "learning_rate": 1e-05, + "loss": 1.2994, + "step": 1455 + }, + { + "epoch": 0.037807660205879626, + "grad_norm": 7.452089761616238, + "learning_rate": 1e-05, + "loss": 1.3164, + "step": 1456 + }, + { + "epoch": 0.037833627005471576, + "grad_norm": 13.097605572337091, + "learning_rate": 1e-05, + "loss": 1.2758, + "step": 1457 + }, + { + "epoch": 0.03785959380506353, + "grad_norm": 8.633665293002826, + "learning_rate": 1e-05, + "loss": 1.3111, + "step": 1458 + }, + { + "epoch": 0.03788556060465548, + "grad_norm": 9.917684508161065, + "learning_rate": 1e-05, + "loss": 1.2751, + "step": 1459 + }, + { + "epoch": 0.03791152740424743, + "grad_norm": 7.626841088325643, + "learning_rate": 1e-05, + "loss": 1.2691, + "step": 1460 + }, + { + "epoch": 0.03793749420383938, + "grad_norm": 5.0988231273437545, + "learning_rate": 1e-05, + "loss": 1.2727, + "step": 1461 + }, + { + "epoch": 0.03796346100343133, + "grad_norm": 6.954166497226147, + "learning_rate": 1e-05, + "loss": 1.2484, + "step": 1462 + }, + { + "epoch": 0.03798942780302328, + "grad_norm": 6.744684740394798, + "learning_rate": 1e-05, + "loss": 1.2766, + "step": 1463 + }, + { + "epoch": 0.03801539460261523, + "grad_norm": 46.278876120752976, + "learning_rate": 1e-05, + "loss": 1.2363, + "step": 1464 + }, + { + "epoch": 0.03804136140220718, + "grad_norm": 8.444910031026316, + "learning_rate": 1e-05, + "loss": 1.3348, + "step": 1465 + }, + { + "epoch": 0.03806732820179913, + "grad_norm": 6.044633474541921, + "learning_rate": 1e-05, + "loss": 1.3226, + "step": 1466 + }, + { + "epoch": 0.03809329500139108, + "grad_norm": 19.769739258068913, + "learning_rate": 1e-05, + "loss": 1.3162, + "step": 1467 + }, + { + "epoch": 0.03811926180098303, + "grad_norm": 6.8684058730582125, + "learning_rate": 1e-05, + "loss": 1.2503, + "step": 1468 + }, + { + "epoch": 0.03814522860057498, + "grad_norm": 15.07984172950479, + "learning_rate": 1e-05, + "loss": 1.232, + "step": 1469 + }, + { + "epoch": 0.03817119540016693, + "grad_norm": 6.185448508749752, + "learning_rate": 1e-05, + "loss": 1.2444, + "step": 1470 + }, + { + "epoch": 0.03819716219975888, + "grad_norm": 5.5655541768352315, + "learning_rate": 1e-05, + "loss": 1.2807, + "step": 1471 + }, + { + "epoch": 0.03822312899935083, + "grad_norm": 5.199508739154117, + "learning_rate": 1e-05, + "loss": 1.2854, + "step": 1472 + }, + { + "epoch": 0.03824909579894278, + "grad_norm": 27.045202775428816, + "learning_rate": 1e-05, + "loss": 1.3158, + "step": 1473 + }, + { + "epoch": 0.03827506259853473, + "grad_norm": 5.68296180174457, + "learning_rate": 1e-05, + "loss": 1.2989, + "step": 1474 + }, + { + "epoch": 0.03830102939812668, + "grad_norm": 5.807861446146656, + "learning_rate": 1e-05, + "loss": 1.2668, + "step": 1475 + }, + { + "epoch": 0.03832699619771863, + "grad_norm": 12.462547040446891, + "learning_rate": 1e-05, + "loss": 1.3103, + "step": 1476 + }, + { + "epoch": 0.03835296299731058, + "grad_norm": 6.59137209279478, + "learning_rate": 1e-05, + "loss": 1.2854, + "step": 1477 + }, + { + "epoch": 0.03837892979690253, + "grad_norm": 7.160386135408781, + "learning_rate": 1e-05, + "loss": 1.2857, + "step": 1478 + }, + { + "epoch": 0.03840489659649448, + "grad_norm": 5.580823400582679, + "learning_rate": 1e-05, + "loss": 1.2698, + "step": 1479 + }, + { + "epoch": 0.03843086339608643, + "grad_norm": 7.634270207688742, + "learning_rate": 1e-05, + "loss": 1.3161, + "step": 1480 + }, + { + "epoch": 0.03845683019567838, + "grad_norm": 6.093433234956069, + "learning_rate": 1e-05, + "loss": 1.2266, + "step": 1481 + }, + { + "epoch": 0.03848279699527033, + "grad_norm": 15.427247136920858, + "learning_rate": 1e-05, + "loss": 1.2883, + "step": 1482 + }, + { + "epoch": 0.03850876379486228, + "grad_norm": 5.614115207763378, + "learning_rate": 1e-05, + "loss": 1.3175, + "step": 1483 + }, + { + "epoch": 0.03853473059445423, + "grad_norm": 5.689081999031058, + "learning_rate": 1e-05, + "loss": 1.296, + "step": 1484 + }, + { + "epoch": 0.03856069739404618, + "grad_norm": 5.808291316416336, + "learning_rate": 1e-05, + "loss": 1.3124, + "step": 1485 + }, + { + "epoch": 0.03858666419363813, + "grad_norm": 10.792207760789763, + "learning_rate": 1e-05, + "loss": 1.2818, + "step": 1486 + }, + { + "epoch": 0.03861263099323008, + "grad_norm": 6.335518443293532, + "learning_rate": 1e-05, + "loss": 1.2715, + "step": 1487 + }, + { + "epoch": 0.038638597792822034, + "grad_norm": 5.887893164722799, + "learning_rate": 1e-05, + "loss": 1.2594, + "step": 1488 + }, + { + "epoch": 0.038664564592413984, + "grad_norm": 3.993757980410344, + "learning_rate": 1e-05, + "loss": 1.3022, + "step": 1489 + }, + { + "epoch": 0.038690531392005935, + "grad_norm": 5.856556288160858, + "learning_rate": 1e-05, + "loss": 1.253, + "step": 1490 + }, + { + "epoch": 0.038716498191597885, + "grad_norm": 9.759506286265722, + "learning_rate": 1e-05, + "loss": 1.2971, + "step": 1491 + }, + { + "epoch": 0.038742464991189836, + "grad_norm": 34.172771621753185, + "learning_rate": 1e-05, + "loss": 1.2788, + "step": 1492 + }, + { + "epoch": 0.038768431790781786, + "grad_norm": 5.482639831624491, + "learning_rate": 1e-05, + "loss": 1.3126, + "step": 1493 + }, + { + "epoch": 0.03879439859037374, + "grad_norm": 5.40406799383187, + "learning_rate": 1e-05, + "loss": 1.2687, + "step": 1494 + }, + { + "epoch": 0.03882036538996569, + "grad_norm": 6.3338955412171725, + "learning_rate": 1e-05, + "loss": 1.2425, + "step": 1495 + }, + { + "epoch": 0.03884633218955764, + "grad_norm": 8.60599535951128, + "learning_rate": 1e-05, + "loss": 1.2556, + "step": 1496 + }, + { + "epoch": 0.03887229898914959, + "grad_norm": 6.157586122159124, + "learning_rate": 1e-05, + "loss": 1.3066, + "step": 1497 + }, + { + "epoch": 0.03889826578874154, + "grad_norm": 6.147750021126885, + "learning_rate": 1e-05, + "loss": 1.2236, + "step": 1498 + }, + { + "epoch": 0.03892423258833349, + "grad_norm": 6.727390712387615, + "learning_rate": 1e-05, + "loss": 1.2583, + "step": 1499 + }, + { + "epoch": 0.03895019938792544, + "grad_norm": 6.72073519454351, + "learning_rate": 1e-05, + "loss": 1.2894, + "step": 1500 + }, + { + "epoch": 0.03897616618751739, + "grad_norm": 8.904263673098331, + "learning_rate": 1e-05, + "loss": 1.3204, + "step": 1501 + }, + { + "epoch": 0.03900213298710934, + "grad_norm": 6.464127957918814, + "learning_rate": 1e-05, + "loss": 1.2664, + "step": 1502 + }, + { + "epoch": 0.03902809978670129, + "grad_norm": 8.704574382090135, + "learning_rate": 1e-05, + "loss": 1.2541, + "step": 1503 + }, + { + "epoch": 0.03905406658629324, + "grad_norm": 230.90646780677235, + "learning_rate": 1e-05, + "loss": 1.2958, + "step": 1504 + }, + { + "epoch": 0.03908003338588519, + "grad_norm": 8.7469568682809, + "learning_rate": 1e-05, + "loss": 1.266, + "step": 1505 + }, + { + "epoch": 0.03910600018547714, + "grad_norm": 8.314097315545643, + "learning_rate": 1e-05, + "loss": 1.2709, + "step": 1506 + }, + { + "epoch": 0.03913196698506909, + "grad_norm": 8.83228780448917, + "learning_rate": 1e-05, + "loss": 1.2937, + "step": 1507 + }, + { + "epoch": 0.03915793378466104, + "grad_norm": 10.28455353045531, + "learning_rate": 1e-05, + "loss": 1.3232, + "step": 1508 + }, + { + "epoch": 0.039183900584252994, + "grad_norm": 5.705399402007346, + "learning_rate": 1e-05, + "loss": 1.303, + "step": 1509 + }, + { + "epoch": 0.039209867383844944, + "grad_norm": 6.962077277618459, + "learning_rate": 1e-05, + "loss": 1.2473, + "step": 1510 + }, + { + "epoch": 0.039235834183436895, + "grad_norm": 6.118213241148117, + "learning_rate": 1e-05, + "loss": 1.2921, + "step": 1511 + }, + { + "epoch": 0.03926180098302884, + "grad_norm": 19.792221329261764, + "learning_rate": 1e-05, + "loss": 1.2777, + "step": 1512 + }, + { + "epoch": 0.03928776778262079, + "grad_norm": 6.994925430696589, + "learning_rate": 1e-05, + "loss": 1.2686, + "step": 1513 + }, + { + "epoch": 0.03931373458221274, + "grad_norm": 4.883692303549341, + "learning_rate": 1e-05, + "loss": 1.2317, + "step": 1514 + }, + { + "epoch": 0.03933970138180469, + "grad_norm": 6.984074050598138, + "learning_rate": 1e-05, + "loss": 1.2781, + "step": 1515 + }, + { + "epoch": 0.03936566818139664, + "grad_norm": 6.60307108779902, + "learning_rate": 1e-05, + "loss": 1.3001, + "step": 1516 + }, + { + "epoch": 0.03939163498098859, + "grad_norm": 6.929970168893084, + "learning_rate": 1e-05, + "loss": 1.3534, + "step": 1517 + }, + { + "epoch": 0.03941760178058054, + "grad_norm": 11.377211806340256, + "learning_rate": 1e-05, + "loss": 1.2973, + "step": 1518 + }, + { + "epoch": 0.03944356858017249, + "grad_norm": 5.198329972811036, + "learning_rate": 1e-05, + "loss": 1.2875, + "step": 1519 + }, + { + "epoch": 0.03946953537976444, + "grad_norm": 6.934549890060232, + "learning_rate": 1e-05, + "loss": 1.2579, + "step": 1520 + }, + { + "epoch": 0.03949550217935639, + "grad_norm": 6.806673410520037, + "learning_rate": 1e-05, + "loss": 1.2698, + "step": 1521 + }, + { + "epoch": 0.03952146897894834, + "grad_norm": 6.016655375433953, + "learning_rate": 1e-05, + "loss": 1.2438, + "step": 1522 + }, + { + "epoch": 0.039547435778540294, + "grad_norm": 6.917758380571187, + "learning_rate": 1e-05, + "loss": 1.2748, + "step": 1523 + }, + { + "epoch": 0.039573402578132244, + "grad_norm": 6.925806586600582, + "learning_rate": 1e-05, + "loss": 1.249, + "step": 1524 + }, + { + "epoch": 0.039599369377724195, + "grad_norm": 5.201284822046263, + "learning_rate": 1e-05, + "loss": 1.2219, + "step": 1525 + }, + { + "epoch": 0.039625336177316145, + "grad_norm": 8.457430886920168, + "learning_rate": 1e-05, + "loss": 1.2354, + "step": 1526 + }, + { + "epoch": 0.039651302976908095, + "grad_norm": 8.749286731531141, + "learning_rate": 1e-05, + "loss": 1.2678, + "step": 1527 + }, + { + "epoch": 0.039677269776500046, + "grad_norm": 4.15235477420926, + "learning_rate": 1e-05, + "loss": 1.2707, + "step": 1528 + }, + { + "epoch": 0.039703236576091996, + "grad_norm": 15.424367160273444, + "learning_rate": 1e-05, + "loss": 1.284, + "step": 1529 + }, + { + "epoch": 0.03972920337568395, + "grad_norm": 7.951138051466379, + "learning_rate": 1e-05, + "loss": 1.3092, + "step": 1530 + }, + { + "epoch": 0.0397551701752759, + "grad_norm": 11.134053179222466, + "learning_rate": 1e-05, + "loss": 1.2855, + "step": 1531 + }, + { + "epoch": 0.03978113697486785, + "grad_norm": 11.399294219003483, + "learning_rate": 1e-05, + "loss": 1.298, + "step": 1532 + }, + { + "epoch": 0.0398071037744598, + "grad_norm": 5.643768364590224, + "learning_rate": 1e-05, + "loss": 1.3036, + "step": 1533 + }, + { + "epoch": 0.03983307057405175, + "grad_norm": 5.508199648405931, + "learning_rate": 1e-05, + "loss": 1.2639, + "step": 1534 + }, + { + "epoch": 0.0398590373736437, + "grad_norm": 6.3387302024801455, + "learning_rate": 1e-05, + "loss": 1.3057, + "step": 1535 + }, + { + "epoch": 0.03988500417323565, + "grad_norm": 9.329324769734393, + "learning_rate": 1e-05, + "loss": 1.2625, + "step": 1536 + }, + { + "epoch": 0.0399109709728276, + "grad_norm": 4.171614620927037, + "learning_rate": 1e-05, + "loss": 1.31, + "step": 1537 + }, + { + "epoch": 0.03993693777241955, + "grad_norm": 5.82453050808805, + "learning_rate": 1e-05, + "loss": 1.2761, + "step": 1538 + }, + { + "epoch": 0.0399629045720115, + "grad_norm": 5.530596333473725, + "learning_rate": 1e-05, + "loss": 1.2582, + "step": 1539 + }, + { + "epoch": 0.03998887137160345, + "grad_norm": 5.391156598844686, + "learning_rate": 1e-05, + "loss": 1.2772, + "step": 1540 + }, + { + "epoch": 0.0400148381711954, + "grad_norm": 5.12743952305847, + "learning_rate": 1e-05, + "loss": 1.2514, + "step": 1541 + }, + { + "epoch": 0.04004080497078735, + "grad_norm": 4.692946461497138, + "learning_rate": 1e-05, + "loss": 1.2877, + "step": 1542 + }, + { + "epoch": 0.0400667717703793, + "grad_norm": 5.411136072994641, + "learning_rate": 1e-05, + "loss": 1.2847, + "step": 1543 + }, + { + "epoch": 0.040092738569971254, + "grad_norm": 5.260227595289775, + "learning_rate": 1e-05, + "loss": 1.2832, + "step": 1544 + }, + { + "epoch": 0.040118705369563204, + "grad_norm": 5.673134908168437, + "learning_rate": 1e-05, + "loss": 1.2988, + "step": 1545 + }, + { + "epoch": 0.040144672169155154, + "grad_norm": 4.314192287837383, + "learning_rate": 1e-05, + "loss": 1.3112, + "step": 1546 + }, + { + "epoch": 0.040170638968747105, + "grad_norm": 7.771581073474804, + "learning_rate": 1e-05, + "loss": 1.2592, + "step": 1547 + }, + { + "epoch": 0.040196605768339055, + "grad_norm": 4.3489550683776415, + "learning_rate": 1e-05, + "loss": 1.3008, + "step": 1548 + }, + { + "epoch": 0.040222572567931006, + "grad_norm": 9.45802792855695, + "learning_rate": 1e-05, + "loss": 1.3124, + "step": 1549 + }, + { + "epoch": 0.04024853936752295, + "grad_norm": 7.97514344089384, + "learning_rate": 1e-05, + "loss": 1.2442, + "step": 1550 + }, + { + "epoch": 0.0402745061671149, + "grad_norm": 4.9961943448507435, + "learning_rate": 1e-05, + "loss": 1.2687, + "step": 1551 + }, + { + "epoch": 0.04030047296670685, + "grad_norm": 5.075194842727754, + "learning_rate": 1e-05, + "loss": 1.2991, + "step": 1552 + }, + { + "epoch": 0.0403264397662988, + "grad_norm": 6.7905830813859485, + "learning_rate": 1e-05, + "loss": 1.277, + "step": 1553 + }, + { + "epoch": 0.04035240656589075, + "grad_norm": 9.504327691796346, + "learning_rate": 1e-05, + "loss": 1.2381, + "step": 1554 + }, + { + "epoch": 0.0403783733654827, + "grad_norm": 5.473604151491644, + "learning_rate": 1e-05, + "loss": 1.3012, + "step": 1555 + }, + { + "epoch": 0.04040434016507465, + "grad_norm": 5.153298938006412, + "learning_rate": 1e-05, + "loss": 1.2892, + "step": 1556 + }, + { + "epoch": 0.0404303069646666, + "grad_norm": 5.396366147083034, + "learning_rate": 1e-05, + "loss": 1.2443, + "step": 1557 + }, + { + "epoch": 0.04045627376425855, + "grad_norm": 6.357134450107851, + "learning_rate": 1e-05, + "loss": 1.2963, + "step": 1558 + }, + { + "epoch": 0.040482240563850504, + "grad_norm": 4.983854356318523, + "learning_rate": 1e-05, + "loss": 1.3173, + "step": 1559 + }, + { + "epoch": 0.040508207363442454, + "grad_norm": 6.12650918857113, + "learning_rate": 1e-05, + "loss": 1.2828, + "step": 1560 + }, + { + "epoch": 0.040534174163034405, + "grad_norm": 6.496121349723389, + "learning_rate": 1e-05, + "loss": 1.2582, + "step": 1561 + }, + { + "epoch": 0.040560140962626355, + "grad_norm": 5.658301892799517, + "learning_rate": 1e-05, + "loss": 1.2937, + "step": 1562 + }, + { + "epoch": 0.040586107762218306, + "grad_norm": 52.26523262184552, + "learning_rate": 1e-05, + "loss": 1.2989, + "step": 1563 + }, + { + "epoch": 0.040612074561810256, + "grad_norm": 12.123319843722804, + "learning_rate": 1e-05, + "loss": 1.2935, + "step": 1564 + }, + { + "epoch": 0.040638041361402207, + "grad_norm": 8.749335127501967, + "learning_rate": 1e-05, + "loss": 1.3102, + "step": 1565 + }, + { + "epoch": 0.04066400816099416, + "grad_norm": 4.830442989403451, + "learning_rate": 1e-05, + "loss": 1.2955, + "step": 1566 + }, + { + "epoch": 0.04068997496058611, + "grad_norm": 7.663544088051034, + "learning_rate": 1e-05, + "loss": 1.312, + "step": 1567 + }, + { + "epoch": 0.04071594176017806, + "grad_norm": 41.410934286920224, + "learning_rate": 1e-05, + "loss": 1.274, + "step": 1568 + }, + { + "epoch": 0.04074190855977001, + "grad_norm": 5.6033888304210775, + "learning_rate": 1e-05, + "loss": 1.2719, + "step": 1569 + }, + { + "epoch": 0.04076787535936196, + "grad_norm": 5.278123821938238, + "learning_rate": 1e-05, + "loss": 1.2583, + "step": 1570 + }, + { + "epoch": 0.04079384215895391, + "grad_norm": 6.682631387100397, + "learning_rate": 1e-05, + "loss": 1.2896, + "step": 1571 + }, + { + "epoch": 0.04081980895854586, + "grad_norm": 4.811918743750767, + "learning_rate": 1e-05, + "loss": 1.2536, + "step": 1572 + }, + { + "epoch": 0.04084577575813781, + "grad_norm": 6.676931426115066, + "learning_rate": 1e-05, + "loss": 1.2612, + "step": 1573 + }, + { + "epoch": 0.04087174255772976, + "grad_norm": 5.849217758698327, + "learning_rate": 1e-05, + "loss": 1.3477, + "step": 1574 + }, + { + "epoch": 0.04089770935732171, + "grad_norm": 5.183037366522977, + "learning_rate": 1e-05, + "loss": 1.2544, + "step": 1575 + }, + { + "epoch": 0.04092367615691366, + "grad_norm": 4.2552906092234535, + "learning_rate": 1e-05, + "loss": 1.2379, + "step": 1576 + }, + { + "epoch": 0.04094964295650561, + "grad_norm": 6.9336341878222205, + "learning_rate": 1e-05, + "loss": 1.3269, + "step": 1577 + }, + { + "epoch": 0.04097560975609756, + "grad_norm": 7.0298327564732555, + "learning_rate": 1e-05, + "loss": 1.2885, + "step": 1578 + }, + { + "epoch": 0.04100157655568951, + "grad_norm": 4.745809011198867, + "learning_rate": 1e-05, + "loss": 1.3265, + "step": 1579 + }, + { + "epoch": 0.041027543355281464, + "grad_norm": 5.210202647128565, + "learning_rate": 1e-05, + "loss": 1.3042, + "step": 1580 + }, + { + "epoch": 0.041053510154873414, + "grad_norm": 7.041838273284174, + "learning_rate": 1e-05, + "loss": 1.3106, + "step": 1581 + }, + { + "epoch": 0.041079476954465365, + "grad_norm": 6.236160825641183, + "learning_rate": 1e-05, + "loss": 1.3374, + "step": 1582 + }, + { + "epoch": 0.041105443754057315, + "grad_norm": 6.4219905350959, + "learning_rate": 1e-05, + "loss": 1.2489, + "step": 1583 + }, + { + "epoch": 0.041131410553649266, + "grad_norm": 7.682454011570299, + "learning_rate": 1e-05, + "loss": 1.3087, + "step": 1584 + }, + { + "epoch": 0.041157377353241216, + "grad_norm": 7.7177128268727495, + "learning_rate": 1e-05, + "loss": 1.286, + "step": 1585 + }, + { + "epoch": 0.041183344152833166, + "grad_norm": 10.128778576735844, + "learning_rate": 1e-05, + "loss": 1.2398, + "step": 1586 + }, + { + "epoch": 0.04120931095242512, + "grad_norm": 6.953697746775856, + "learning_rate": 1e-05, + "loss": 1.3156, + "step": 1587 + }, + { + "epoch": 0.04123527775201706, + "grad_norm": 4.240081208009193, + "learning_rate": 1e-05, + "loss": 1.2764, + "step": 1588 + }, + { + "epoch": 0.04126124455160901, + "grad_norm": 4.732164326333753, + "learning_rate": 1e-05, + "loss": 1.2564, + "step": 1589 + }, + { + "epoch": 0.04128721135120096, + "grad_norm": 6.117078368478966, + "learning_rate": 1e-05, + "loss": 1.32, + "step": 1590 + }, + { + "epoch": 0.04131317815079291, + "grad_norm": 6.6958135820187294, + "learning_rate": 1e-05, + "loss": 1.3126, + "step": 1591 + }, + { + "epoch": 0.04133914495038486, + "grad_norm": 6.139776350787976, + "learning_rate": 1e-05, + "loss": 1.2992, + "step": 1592 + }, + { + "epoch": 0.04136511174997681, + "grad_norm": 10.18764191511104, + "learning_rate": 1e-05, + "loss": 1.2713, + "step": 1593 + }, + { + "epoch": 0.04139107854956876, + "grad_norm": 9.943031163250918, + "learning_rate": 1e-05, + "loss": 1.2872, + "step": 1594 + }, + { + "epoch": 0.041417045349160714, + "grad_norm": 8.348140632881503, + "learning_rate": 1e-05, + "loss": 1.2591, + "step": 1595 + }, + { + "epoch": 0.041443012148752664, + "grad_norm": 5.839804556724667, + "learning_rate": 1e-05, + "loss": 1.2527, + "step": 1596 + }, + { + "epoch": 0.041468978948344615, + "grad_norm": 5.574327046786556, + "learning_rate": 1e-05, + "loss": 1.2672, + "step": 1597 + }, + { + "epoch": 0.041494945747936565, + "grad_norm": 4.743062021981196, + "learning_rate": 1e-05, + "loss": 1.2872, + "step": 1598 + }, + { + "epoch": 0.041520912547528516, + "grad_norm": 14.210874686330392, + "learning_rate": 1e-05, + "loss": 1.2772, + "step": 1599 + }, + { + "epoch": 0.041546879347120466, + "grad_norm": 6.198361814097883, + "learning_rate": 1e-05, + "loss": 1.3522, + "step": 1600 + }, + { + "epoch": 0.04157284614671242, + "grad_norm": 8.900735854685433, + "learning_rate": 1e-05, + "loss": 1.2899, + "step": 1601 + }, + { + "epoch": 0.04159881294630437, + "grad_norm": 6.409308768627204, + "learning_rate": 1e-05, + "loss": 1.3042, + "step": 1602 + }, + { + "epoch": 0.04162477974589632, + "grad_norm": 4.7726616611347, + "learning_rate": 1e-05, + "loss": 1.286, + "step": 1603 + }, + { + "epoch": 0.04165074654548827, + "grad_norm": 6.194499978736966, + "learning_rate": 1e-05, + "loss": 1.2659, + "step": 1604 + }, + { + "epoch": 0.04167671334508022, + "grad_norm": 6.592966910049068, + "learning_rate": 1e-05, + "loss": 1.2591, + "step": 1605 + }, + { + "epoch": 0.04170268014467217, + "grad_norm": 5.141668799269536, + "learning_rate": 1e-05, + "loss": 1.2737, + "step": 1606 + }, + { + "epoch": 0.04172864694426412, + "grad_norm": 5.3240494788641435, + "learning_rate": 1e-05, + "loss": 1.2803, + "step": 1607 + }, + { + "epoch": 0.04175461374385607, + "grad_norm": 5.044879908724987, + "learning_rate": 1e-05, + "loss": 1.2347, + "step": 1608 + }, + { + "epoch": 0.04178058054344802, + "grad_norm": 4.262929434972354, + "learning_rate": 1e-05, + "loss": 1.2958, + "step": 1609 + }, + { + "epoch": 0.04180654734303997, + "grad_norm": 13.298172238608407, + "learning_rate": 1e-05, + "loss": 1.2524, + "step": 1610 + }, + { + "epoch": 0.04183251414263192, + "grad_norm": 4.6723336597468, + "learning_rate": 1e-05, + "loss": 1.3352, + "step": 1611 + }, + { + "epoch": 0.04185848094222387, + "grad_norm": 5.443286841887574, + "learning_rate": 1e-05, + "loss": 1.325, + "step": 1612 + }, + { + "epoch": 0.04188444774181582, + "grad_norm": 6.777347690020469, + "learning_rate": 1e-05, + "loss": 1.2707, + "step": 1613 + }, + { + "epoch": 0.04191041454140777, + "grad_norm": 5.986055700877157, + "learning_rate": 1e-05, + "loss": 1.2717, + "step": 1614 + }, + { + "epoch": 0.04193638134099972, + "grad_norm": 4.1133304024247455, + "learning_rate": 1e-05, + "loss": 1.3131, + "step": 1615 + }, + { + "epoch": 0.041962348140591674, + "grad_norm": 4.975403751111287, + "learning_rate": 1e-05, + "loss": 1.2994, + "step": 1616 + }, + { + "epoch": 0.041988314940183624, + "grad_norm": 6.154702548812207, + "learning_rate": 1e-05, + "loss": 1.2708, + "step": 1617 + }, + { + "epoch": 0.042014281739775575, + "grad_norm": 4.536578603464845, + "learning_rate": 1e-05, + "loss": 1.2784, + "step": 1618 + }, + { + "epoch": 0.042040248539367525, + "grad_norm": 4.834343815550682, + "learning_rate": 1e-05, + "loss": 1.2852, + "step": 1619 + }, + { + "epoch": 0.042066215338959476, + "grad_norm": 4.982393355158967, + "learning_rate": 1e-05, + "loss": 1.3138, + "step": 1620 + }, + { + "epoch": 0.042092182138551426, + "grad_norm": 4.873503626515143, + "learning_rate": 1e-05, + "loss": 1.2717, + "step": 1621 + }, + { + "epoch": 0.04211814893814338, + "grad_norm": 4.504399374362234, + "learning_rate": 1e-05, + "loss": 1.2838, + "step": 1622 + }, + { + "epoch": 0.04214411573773533, + "grad_norm": 4.407870812961715, + "learning_rate": 1e-05, + "loss": 1.2536, + "step": 1623 + }, + { + "epoch": 0.04217008253732728, + "grad_norm": 5.306418494614822, + "learning_rate": 1e-05, + "loss": 1.2655, + "step": 1624 + }, + { + "epoch": 0.04219604933691923, + "grad_norm": 3.6244643243082963, + "learning_rate": 1e-05, + "loss": 1.2685, + "step": 1625 + }, + { + "epoch": 0.04222201613651118, + "grad_norm": 3.928382125851428, + "learning_rate": 1e-05, + "loss": 1.304, + "step": 1626 + }, + { + "epoch": 0.04224798293610312, + "grad_norm": 3.887207245919555, + "learning_rate": 1e-05, + "loss": 1.3044, + "step": 1627 + }, + { + "epoch": 0.04227394973569507, + "grad_norm": 4.514326601895249, + "learning_rate": 1e-05, + "loss": 1.2872, + "step": 1628 + }, + { + "epoch": 0.04229991653528702, + "grad_norm": 6.216234781571489, + "learning_rate": 1e-05, + "loss": 1.2877, + "step": 1629 + }, + { + "epoch": 0.04232588333487897, + "grad_norm": 5.146570726860949, + "learning_rate": 1e-05, + "loss": 1.2476, + "step": 1630 + }, + { + "epoch": 0.042351850134470924, + "grad_norm": 3.9483104230979045, + "learning_rate": 1e-05, + "loss": 1.2398, + "step": 1631 + }, + { + "epoch": 0.042377816934062874, + "grad_norm": 15.242038837531044, + "learning_rate": 1e-05, + "loss": 1.237, + "step": 1632 + }, + { + "epoch": 0.042403783733654825, + "grad_norm": 13.170678280081928, + "learning_rate": 1e-05, + "loss": 1.359, + "step": 1633 + }, + { + "epoch": 0.042429750533246775, + "grad_norm": 5.698616649676542, + "learning_rate": 1e-05, + "loss": 1.2588, + "step": 1634 + }, + { + "epoch": 0.042455717332838726, + "grad_norm": 5.083832242069257, + "learning_rate": 1e-05, + "loss": 1.2782, + "step": 1635 + }, + { + "epoch": 0.042481684132430676, + "grad_norm": 6.053980238429501, + "learning_rate": 1e-05, + "loss": 1.284, + "step": 1636 + }, + { + "epoch": 0.04250765093202263, + "grad_norm": 5.387022899204414, + "learning_rate": 1e-05, + "loss": 1.2932, + "step": 1637 + }, + { + "epoch": 0.04253361773161458, + "grad_norm": 4.038909023770518, + "learning_rate": 1e-05, + "loss": 1.2293, + "step": 1638 + }, + { + "epoch": 0.04255958453120653, + "grad_norm": 4.139760974489145, + "learning_rate": 1e-05, + "loss": 1.3216, + "step": 1639 + }, + { + "epoch": 0.04258555133079848, + "grad_norm": 5.4035219575095725, + "learning_rate": 1e-05, + "loss": 1.2863, + "step": 1640 + }, + { + "epoch": 0.04261151813039043, + "grad_norm": 6.2970761250448115, + "learning_rate": 1e-05, + "loss": 1.3261, + "step": 1641 + }, + { + "epoch": 0.04263748492998238, + "grad_norm": 5.329534528779885, + "learning_rate": 1e-05, + "loss": 1.3137, + "step": 1642 + }, + { + "epoch": 0.04266345172957433, + "grad_norm": 7.443329393881498, + "learning_rate": 1e-05, + "loss": 1.2892, + "step": 1643 + }, + { + "epoch": 0.04268941852916628, + "grad_norm": 5.738344071409867, + "learning_rate": 1e-05, + "loss": 1.2635, + "step": 1644 + }, + { + "epoch": 0.04271538532875823, + "grad_norm": 4.77996868841538, + "learning_rate": 1e-05, + "loss": 1.3205, + "step": 1645 + }, + { + "epoch": 0.04274135212835018, + "grad_norm": 4.77479306716307, + "learning_rate": 1e-05, + "loss": 1.2774, + "step": 1646 + }, + { + "epoch": 0.04276731892794213, + "grad_norm": 4.2632685709536355, + "learning_rate": 1e-05, + "loss": 1.2909, + "step": 1647 + }, + { + "epoch": 0.04279328572753408, + "grad_norm": 7.572750994003929, + "learning_rate": 1e-05, + "loss": 1.2704, + "step": 1648 + }, + { + "epoch": 0.04281925252712603, + "grad_norm": 5.083416901615475, + "learning_rate": 1e-05, + "loss": 1.2552, + "step": 1649 + }, + { + "epoch": 0.04284521932671798, + "grad_norm": 5.0316762802842545, + "learning_rate": 1e-05, + "loss": 1.3323, + "step": 1650 + }, + { + "epoch": 0.04287118612630993, + "grad_norm": 4.884612949308346, + "learning_rate": 1e-05, + "loss": 1.2424, + "step": 1651 + }, + { + "epoch": 0.042897152925901884, + "grad_norm": 5.171474475852562, + "learning_rate": 1e-05, + "loss": 1.2861, + "step": 1652 + }, + { + "epoch": 0.042923119725493834, + "grad_norm": 4.924495713804442, + "learning_rate": 1e-05, + "loss": 1.3101, + "step": 1653 + }, + { + "epoch": 0.042949086525085785, + "grad_norm": 7.149377446314682, + "learning_rate": 1e-05, + "loss": 1.3096, + "step": 1654 + }, + { + "epoch": 0.042975053324677735, + "grad_norm": 5.797001675355309, + "learning_rate": 1e-05, + "loss": 1.2918, + "step": 1655 + }, + { + "epoch": 0.043001020124269686, + "grad_norm": 3.939321550735651, + "learning_rate": 1e-05, + "loss": 1.2736, + "step": 1656 + }, + { + "epoch": 0.043026986923861636, + "grad_norm": 4.004895790449284, + "learning_rate": 1e-05, + "loss": 1.2812, + "step": 1657 + }, + { + "epoch": 0.04305295372345359, + "grad_norm": 6.800438799444506, + "learning_rate": 1e-05, + "loss": 1.2928, + "step": 1658 + }, + { + "epoch": 0.04307892052304554, + "grad_norm": 5.086438132402267, + "learning_rate": 1e-05, + "loss": 1.3017, + "step": 1659 + }, + { + "epoch": 0.04310488732263749, + "grad_norm": 5.3678727913236735, + "learning_rate": 1e-05, + "loss": 1.2706, + "step": 1660 + }, + { + "epoch": 0.04313085412222944, + "grad_norm": 4.224940752144444, + "learning_rate": 1e-05, + "loss": 1.2924, + "step": 1661 + }, + { + "epoch": 0.04315682092182139, + "grad_norm": 4.507970321246173, + "learning_rate": 1e-05, + "loss": 1.2744, + "step": 1662 + }, + { + "epoch": 0.04318278772141334, + "grad_norm": 9.29240717880775, + "learning_rate": 1e-05, + "loss": 1.2754, + "step": 1663 + }, + { + "epoch": 0.04320875452100529, + "grad_norm": 5.16085960798301, + "learning_rate": 1e-05, + "loss": 1.2728, + "step": 1664 + }, + { + "epoch": 0.04323472132059723, + "grad_norm": 4.185714796634459, + "learning_rate": 1e-05, + "loss": 1.2811, + "step": 1665 + }, + { + "epoch": 0.043260688120189184, + "grad_norm": 5.423334194684249, + "learning_rate": 1e-05, + "loss": 1.3203, + "step": 1666 + }, + { + "epoch": 0.043286654919781134, + "grad_norm": 4.431959534993998, + "learning_rate": 1e-05, + "loss": 1.3058, + "step": 1667 + }, + { + "epoch": 0.043312621719373084, + "grad_norm": 4.223844941385418, + "learning_rate": 1e-05, + "loss": 1.262, + "step": 1668 + }, + { + "epoch": 0.043338588518965035, + "grad_norm": 4.236600509195045, + "learning_rate": 1e-05, + "loss": 1.2575, + "step": 1669 + }, + { + "epoch": 0.043364555318556985, + "grad_norm": 5.682279433514941, + "learning_rate": 1e-05, + "loss": 1.239, + "step": 1670 + }, + { + "epoch": 0.043390522118148936, + "grad_norm": 5.437041690848293, + "learning_rate": 1e-05, + "loss": 1.2809, + "step": 1671 + }, + { + "epoch": 0.043416488917740886, + "grad_norm": 6.959859471272626, + "learning_rate": 1e-05, + "loss": 1.2526, + "step": 1672 + }, + { + "epoch": 0.04344245571733284, + "grad_norm": 7.1744513112454555, + "learning_rate": 1e-05, + "loss": 1.2431, + "step": 1673 + }, + { + "epoch": 0.04346842251692479, + "grad_norm": 4.71719698581203, + "learning_rate": 1e-05, + "loss": 1.3192, + "step": 1674 + }, + { + "epoch": 0.04349438931651674, + "grad_norm": 4.903291624312361, + "learning_rate": 1e-05, + "loss": 1.2535, + "step": 1675 + }, + { + "epoch": 0.04352035611610869, + "grad_norm": 4.336343282321383, + "learning_rate": 1e-05, + "loss": 1.3066, + "step": 1676 + }, + { + "epoch": 0.04354632291570064, + "grad_norm": 5.827510034265312, + "learning_rate": 1e-05, + "loss": 1.2613, + "step": 1677 + }, + { + "epoch": 0.04357228971529259, + "grad_norm": 6.486585125516829, + "learning_rate": 1e-05, + "loss": 1.3346, + "step": 1678 + }, + { + "epoch": 0.04359825651488454, + "grad_norm": 7.0430120804268865, + "learning_rate": 1e-05, + "loss": 1.2722, + "step": 1679 + }, + { + "epoch": 0.04362422331447649, + "grad_norm": 5.577999166998078, + "learning_rate": 1e-05, + "loss": 1.2594, + "step": 1680 + }, + { + "epoch": 0.04365019011406844, + "grad_norm": 7.111682425172254, + "learning_rate": 1e-05, + "loss": 1.2974, + "step": 1681 + }, + { + "epoch": 0.04367615691366039, + "grad_norm": 5.432952382281181, + "learning_rate": 1e-05, + "loss": 1.311, + "step": 1682 + }, + { + "epoch": 0.04370212371325234, + "grad_norm": 18.59346359417025, + "learning_rate": 1e-05, + "loss": 1.3342, + "step": 1683 + }, + { + "epoch": 0.04372809051284429, + "grad_norm": 5.355081601273581, + "learning_rate": 1e-05, + "loss": 1.269, + "step": 1684 + }, + { + "epoch": 0.04375405731243624, + "grad_norm": 5.601445706265851, + "learning_rate": 1e-05, + "loss": 1.2881, + "step": 1685 + }, + { + "epoch": 0.04378002411202819, + "grad_norm": 4.83086152285268, + "learning_rate": 1e-05, + "loss": 1.2405, + "step": 1686 + }, + { + "epoch": 0.04380599091162014, + "grad_norm": 9.684307138279701, + "learning_rate": 1e-05, + "loss": 1.2805, + "step": 1687 + }, + { + "epoch": 0.043831957711212094, + "grad_norm": 5.304341772664668, + "learning_rate": 1e-05, + "loss": 1.2806, + "step": 1688 + }, + { + "epoch": 0.043857924510804044, + "grad_norm": 7.197955678419139, + "learning_rate": 1e-05, + "loss": 1.2525, + "step": 1689 + }, + { + "epoch": 0.043883891310395995, + "grad_norm": 3.992233485126849, + "learning_rate": 1e-05, + "loss": 1.2641, + "step": 1690 + }, + { + "epoch": 0.043909858109987945, + "grad_norm": 6.153373859695, + "learning_rate": 1e-05, + "loss": 1.2583, + "step": 1691 + }, + { + "epoch": 0.043935824909579896, + "grad_norm": 5.620913568277048, + "learning_rate": 1e-05, + "loss": 1.321, + "step": 1692 + }, + { + "epoch": 0.043961791709171846, + "grad_norm": 4.874449283108671, + "learning_rate": 1e-05, + "loss": 1.2803, + "step": 1693 + }, + { + "epoch": 0.0439877585087638, + "grad_norm": 6.551026405110059, + "learning_rate": 1e-05, + "loss": 1.2923, + "step": 1694 + }, + { + "epoch": 0.04401372530835575, + "grad_norm": 7.427440090804518, + "learning_rate": 1e-05, + "loss": 1.2772, + "step": 1695 + }, + { + "epoch": 0.0440396921079477, + "grad_norm": 6.21151020600519, + "learning_rate": 1e-05, + "loss": 1.2523, + "step": 1696 + }, + { + "epoch": 0.04406565890753965, + "grad_norm": 5.006516977875523, + "learning_rate": 1e-05, + "loss": 1.246, + "step": 1697 + }, + { + "epoch": 0.0440916257071316, + "grad_norm": 28.619785150325313, + "learning_rate": 1e-05, + "loss": 1.2812, + "step": 1698 + }, + { + "epoch": 0.04411759250672355, + "grad_norm": 5.047571092910756, + "learning_rate": 1e-05, + "loss": 1.2984, + "step": 1699 + }, + { + "epoch": 0.0441435593063155, + "grad_norm": 3.9835839776597997, + "learning_rate": 1e-05, + "loss": 1.335, + "step": 1700 + }, + { + "epoch": 0.04416952610590745, + "grad_norm": 10.743939487812346, + "learning_rate": 1e-05, + "loss": 1.2598, + "step": 1701 + }, + { + "epoch": 0.0441954929054994, + "grad_norm": 4.076693579604946, + "learning_rate": 1e-05, + "loss": 1.2902, + "step": 1702 + }, + { + "epoch": 0.044221459705091344, + "grad_norm": 10.20258510966323, + "learning_rate": 1e-05, + "loss": 1.2388, + "step": 1703 + }, + { + "epoch": 0.044247426504683295, + "grad_norm": 6.25949778824429, + "learning_rate": 1e-05, + "loss": 1.2606, + "step": 1704 + }, + { + "epoch": 0.044273393304275245, + "grad_norm": 5.311398560279364, + "learning_rate": 1e-05, + "loss": 1.2303, + "step": 1705 + }, + { + "epoch": 0.044299360103867196, + "grad_norm": 5.193179594775138, + "learning_rate": 1e-05, + "loss": 1.2751, + "step": 1706 + }, + { + "epoch": 0.044325326903459146, + "grad_norm": 3.894491947323938, + "learning_rate": 1e-05, + "loss": 1.3187, + "step": 1707 + }, + { + "epoch": 0.044351293703051096, + "grad_norm": 6.438040108846421, + "learning_rate": 1e-05, + "loss": 1.2699, + "step": 1708 + }, + { + "epoch": 0.04437726050264305, + "grad_norm": 5.312434117525766, + "learning_rate": 1e-05, + "loss": 1.2379, + "step": 1709 + }, + { + "epoch": 0.044403227302235, + "grad_norm": 4.215068878298872, + "learning_rate": 1e-05, + "loss": 1.3315, + "step": 1710 + }, + { + "epoch": 0.04442919410182695, + "grad_norm": 5.926571717218644, + "learning_rate": 1e-05, + "loss": 1.252, + "step": 1711 + }, + { + "epoch": 0.0444551609014189, + "grad_norm": 4.055077918041368, + "learning_rate": 1e-05, + "loss": 1.282, + "step": 1712 + }, + { + "epoch": 0.04448112770101085, + "grad_norm": 5.243997139273066, + "learning_rate": 1e-05, + "loss": 1.2574, + "step": 1713 + }, + { + "epoch": 0.0445070945006028, + "grad_norm": 7.081498510540348, + "learning_rate": 1e-05, + "loss": 1.2568, + "step": 1714 + }, + { + "epoch": 0.04453306130019475, + "grad_norm": 12.43062658496835, + "learning_rate": 1e-05, + "loss": 1.2932, + "step": 1715 + }, + { + "epoch": 0.0445590280997867, + "grad_norm": 5.874260064461875, + "learning_rate": 1e-05, + "loss": 1.3023, + "step": 1716 + }, + { + "epoch": 0.04458499489937865, + "grad_norm": 4.308355800390538, + "learning_rate": 1e-05, + "loss": 1.2274, + "step": 1717 + }, + { + "epoch": 0.0446109616989706, + "grad_norm": 7.035175228146251, + "learning_rate": 1e-05, + "loss": 1.2622, + "step": 1718 + }, + { + "epoch": 0.04463692849856255, + "grad_norm": 6.220330579013227, + "learning_rate": 1e-05, + "loss": 1.2967, + "step": 1719 + }, + { + "epoch": 0.0446628952981545, + "grad_norm": 8.537030954727332, + "learning_rate": 1e-05, + "loss": 1.3113, + "step": 1720 + }, + { + "epoch": 0.04468886209774645, + "grad_norm": 4.851262130899062, + "learning_rate": 1e-05, + "loss": 1.2964, + "step": 1721 + }, + { + "epoch": 0.0447148288973384, + "grad_norm": 6.260762827649841, + "learning_rate": 1e-05, + "loss": 1.2717, + "step": 1722 + }, + { + "epoch": 0.044740795696930354, + "grad_norm": 7.610477062215085, + "learning_rate": 1e-05, + "loss": 1.3381, + "step": 1723 + }, + { + "epoch": 0.044766762496522304, + "grad_norm": 7.12826061930869, + "learning_rate": 1e-05, + "loss": 1.3166, + "step": 1724 + }, + { + "epoch": 0.044792729296114255, + "grad_norm": 5.59232094694219, + "learning_rate": 1e-05, + "loss": 1.283, + "step": 1725 + }, + { + "epoch": 0.044818696095706205, + "grad_norm": 5.701685391934015, + "learning_rate": 1e-05, + "loss": 1.196, + "step": 1726 + }, + { + "epoch": 0.044844662895298155, + "grad_norm": 4.278295458120855, + "learning_rate": 1e-05, + "loss": 1.3376, + "step": 1727 + }, + { + "epoch": 0.044870629694890106, + "grad_norm": 7.83739612273229, + "learning_rate": 1e-05, + "loss": 1.2236, + "step": 1728 + }, + { + "epoch": 0.044896596494482056, + "grad_norm": 4.977450448771007, + "learning_rate": 1e-05, + "loss": 1.3144, + "step": 1729 + }, + { + "epoch": 0.04492256329407401, + "grad_norm": 12.213635713368365, + "learning_rate": 1e-05, + "loss": 1.3284, + "step": 1730 + }, + { + "epoch": 0.04494853009366596, + "grad_norm": 3.9156131680272526, + "learning_rate": 1e-05, + "loss": 1.2689, + "step": 1731 + }, + { + "epoch": 0.04497449689325791, + "grad_norm": 5.54414700330562, + "learning_rate": 1e-05, + "loss": 1.2966, + "step": 1732 + }, + { + "epoch": 0.04500046369284986, + "grad_norm": 6.31626613558262, + "learning_rate": 1e-05, + "loss": 1.233, + "step": 1733 + }, + { + "epoch": 0.04502643049244181, + "grad_norm": 97.40470102233259, + "learning_rate": 1e-05, + "loss": 1.2751, + "step": 1734 + }, + { + "epoch": 0.04505239729203376, + "grad_norm": 4.071198526255572, + "learning_rate": 1e-05, + "loss": 1.279, + "step": 1735 + }, + { + "epoch": 0.04507836409162571, + "grad_norm": 5.295535117856197, + "learning_rate": 1e-05, + "loss": 1.2556, + "step": 1736 + }, + { + "epoch": 0.04510433089121766, + "grad_norm": 4.778193874623676, + "learning_rate": 1e-05, + "loss": 1.2667, + "step": 1737 + }, + { + "epoch": 0.04513029769080961, + "grad_norm": 4.2079562531289545, + "learning_rate": 1e-05, + "loss": 1.2674, + "step": 1738 + }, + { + "epoch": 0.04515626449040156, + "grad_norm": 5.085732920292595, + "learning_rate": 1e-05, + "loss": 1.2806, + "step": 1739 + }, + { + "epoch": 0.04518223128999351, + "grad_norm": 4.883208772982395, + "learning_rate": 1e-05, + "loss": 1.2387, + "step": 1740 + }, + { + "epoch": 0.045208198089585455, + "grad_norm": 7.106041028012495, + "learning_rate": 1e-05, + "loss": 1.2669, + "step": 1741 + }, + { + "epoch": 0.045234164889177406, + "grad_norm": 7.31190005718685, + "learning_rate": 1e-05, + "loss": 1.2782, + "step": 1742 + }, + { + "epoch": 0.045260131688769356, + "grad_norm": 4.661723561617678, + "learning_rate": 1e-05, + "loss": 1.2991, + "step": 1743 + }, + { + "epoch": 0.04528609848836131, + "grad_norm": 5.102467011303496, + "learning_rate": 1e-05, + "loss": 1.2571, + "step": 1744 + }, + { + "epoch": 0.04531206528795326, + "grad_norm": 4.392407472924515, + "learning_rate": 1e-05, + "loss": 1.2885, + "step": 1745 + }, + { + "epoch": 0.04533803208754521, + "grad_norm": 5.921224029915704, + "learning_rate": 1e-05, + "loss": 1.2506, + "step": 1746 + }, + { + "epoch": 0.04536399888713716, + "grad_norm": 4.802391434037606, + "learning_rate": 1e-05, + "loss": 1.3194, + "step": 1747 + }, + { + "epoch": 0.04538996568672911, + "grad_norm": 6.968113344131262, + "learning_rate": 1e-05, + "loss": 1.2678, + "step": 1748 + }, + { + "epoch": 0.04541593248632106, + "grad_norm": 5.2038778957435765, + "learning_rate": 1e-05, + "loss": 1.245, + "step": 1749 + }, + { + "epoch": 0.04544189928591301, + "grad_norm": 8.2055646483792, + "learning_rate": 1e-05, + "loss": 1.3378, + "step": 1750 + }, + { + "epoch": 0.04546786608550496, + "grad_norm": 3.964737433149686, + "learning_rate": 1e-05, + "loss": 1.224, + "step": 1751 + }, + { + "epoch": 0.04549383288509691, + "grad_norm": 5.784562435235587, + "learning_rate": 1e-05, + "loss": 1.2872, + "step": 1752 + }, + { + "epoch": 0.04551979968468886, + "grad_norm": 4.953625193469228, + "learning_rate": 1e-05, + "loss": 1.3006, + "step": 1753 + }, + { + "epoch": 0.04554576648428081, + "grad_norm": 6.041612015400269, + "learning_rate": 1e-05, + "loss": 1.2465, + "step": 1754 + }, + { + "epoch": 0.04557173328387276, + "grad_norm": 4.820537642255311, + "learning_rate": 1e-05, + "loss": 1.2843, + "step": 1755 + }, + { + "epoch": 0.04559770008346471, + "grad_norm": 7.602786115627001, + "learning_rate": 1e-05, + "loss": 1.2397, + "step": 1756 + }, + { + "epoch": 0.04562366688305666, + "grad_norm": 5.025139361674266, + "learning_rate": 1e-05, + "loss": 1.3117, + "step": 1757 + }, + { + "epoch": 0.04564963368264861, + "grad_norm": 6.568494692703662, + "learning_rate": 1e-05, + "loss": 1.299, + "step": 1758 + }, + { + "epoch": 0.045675600482240564, + "grad_norm": 6.397223836074355, + "learning_rate": 1e-05, + "loss": 1.2817, + "step": 1759 + }, + { + "epoch": 0.045701567281832514, + "grad_norm": 4.921705261210047, + "learning_rate": 1e-05, + "loss": 1.3263, + "step": 1760 + }, + { + "epoch": 0.045727534081424465, + "grad_norm": 49.52906393046536, + "learning_rate": 1e-05, + "loss": 1.289, + "step": 1761 + }, + { + "epoch": 0.045753500881016415, + "grad_norm": 18.17337966685407, + "learning_rate": 1e-05, + "loss": 1.2657, + "step": 1762 + }, + { + "epoch": 0.045779467680608366, + "grad_norm": 5.047517812318628, + "learning_rate": 1e-05, + "loss": 1.2537, + "step": 1763 + }, + { + "epoch": 0.045805434480200316, + "grad_norm": 204.66814794014724, + "learning_rate": 1e-05, + "loss": 1.2836, + "step": 1764 + }, + { + "epoch": 0.045831401279792267, + "grad_norm": 4.594737044189691, + "learning_rate": 1e-05, + "loss": 1.2608, + "step": 1765 + }, + { + "epoch": 0.04585736807938422, + "grad_norm": 4.2036338980230425, + "learning_rate": 1e-05, + "loss": 1.27, + "step": 1766 + }, + { + "epoch": 0.04588333487897617, + "grad_norm": 5.421232245875536, + "learning_rate": 1e-05, + "loss": 1.3085, + "step": 1767 + }, + { + "epoch": 0.04590930167856812, + "grad_norm": 5.879850941539807, + "learning_rate": 1e-05, + "loss": 1.3296, + "step": 1768 + }, + { + "epoch": 0.04593526847816007, + "grad_norm": 5.284169129569244, + "learning_rate": 1e-05, + "loss": 1.3006, + "step": 1769 + }, + { + "epoch": 0.04596123527775202, + "grad_norm": 5.305541923376726, + "learning_rate": 1e-05, + "loss": 1.31, + "step": 1770 + }, + { + "epoch": 0.04598720207734397, + "grad_norm": 5.61944441360337, + "learning_rate": 1e-05, + "loss": 1.239, + "step": 1771 + }, + { + "epoch": 0.04601316887693592, + "grad_norm": 8.563844895377883, + "learning_rate": 1e-05, + "loss": 1.2828, + "step": 1772 + }, + { + "epoch": 0.04603913567652787, + "grad_norm": 4.971019298638021, + "learning_rate": 1e-05, + "loss": 1.301, + "step": 1773 + }, + { + "epoch": 0.04606510247611982, + "grad_norm": 4.260758862669352, + "learning_rate": 1e-05, + "loss": 1.2879, + "step": 1774 + }, + { + "epoch": 0.04609106927571177, + "grad_norm": 5.546360561531825, + "learning_rate": 1e-05, + "loss": 1.2729, + "step": 1775 + }, + { + "epoch": 0.04611703607530372, + "grad_norm": 11.194927115771742, + "learning_rate": 1e-05, + "loss": 1.3144, + "step": 1776 + }, + { + "epoch": 0.04614300287489567, + "grad_norm": 5.836035874442788, + "learning_rate": 1e-05, + "loss": 1.3117, + "step": 1777 + }, + { + "epoch": 0.04616896967448762, + "grad_norm": 4.776319966446448, + "learning_rate": 1e-05, + "loss": 1.2601, + "step": 1778 + }, + { + "epoch": 0.04619493647407957, + "grad_norm": 13.615019616411628, + "learning_rate": 1e-05, + "loss": 1.257, + "step": 1779 + }, + { + "epoch": 0.04622090327367152, + "grad_norm": 5.752946844870231, + "learning_rate": 1e-05, + "loss": 1.2526, + "step": 1780 + }, + { + "epoch": 0.04624687007326347, + "grad_norm": 4.758484340393844, + "learning_rate": 1e-05, + "loss": 1.3009, + "step": 1781 + }, + { + "epoch": 0.04627283687285542, + "grad_norm": 4.339254783955204, + "learning_rate": 1e-05, + "loss": 1.2821, + "step": 1782 + }, + { + "epoch": 0.04629880367244737, + "grad_norm": 4.017213737197356, + "learning_rate": 1e-05, + "loss": 1.2881, + "step": 1783 + }, + { + "epoch": 0.04632477047203932, + "grad_norm": 5.674581423207643, + "learning_rate": 1e-05, + "loss": 1.3225, + "step": 1784 + }, + { + "epoch": 0.04635073727163127, + "grad_norm": 6.976242621971047, + "learning_rate": 1e-05, + "loss": 1.2717, + "step": 1785 + }, + { + "epoch": 0.04637670407122322, + "grad_norm": 4.486924566496487, + "learning_rate": 1e-05, + "loss": 1.2615, + "step": 1786 + }, + { + "epoch": 0.04640267087081517, + "grad_norm": 8.150372605494699, + "learning_rate": 1e-05, + "loss": 1.2535, + "step": 1787 + }, + { + "epoch": 0.04642863767040712, + "grad_norm": 5.320427936752436, + "learning_rate": 1e-05, + "loss": 1.2731, + "step": 1788 + }, + { + "epoch": 0.04645460446999907, + "grad_norm": 8.125498594978168, + "learning_rate": 1e-05, + "loss": 1.2726, + "step": 1789 + }, + { + "epoch": 0.04648057126959102, + "grad_norm": 7.847599133767387, + "learning_rate": 1e-05, + "loss": 1.2425, + "step": 1790 + }, + { + "epoch": 0.04650653806918297, + "grad_norm": 5.199338892946492, + "learning_rate": 1e-05, + "loss": 1.2908, + "step": 1791 + }, + { + "epoch": 0.04653250486877492, + "grad_norm": 9.04706010645581, + "learning_rate": 1e-05, + "loss": 1.2479, + "step": 1792 + }, + { + "epoch": 0.04655847166836687, + "grad_norm": 124.93831681198726, + "learning_rate": 1e-05, + "loss": 1.2702, + "step": 1793 + }, + { + "epoch": 0.04658443846795882, + "grad_norm": 4.697535440860414, + "learning_rate": 1e-05, + "loss": 1.227, + "step": 1794 + }, + { + "epoch": 0.046610405267550774, + "grad_norm": 5.690453066629724, + "learning_rate": 1e-05, + "loss": 1.2747, + "step": 1795 + }, + { + "epoch": 0.046636372067142724, + "grad_norm": 4.914016699160566, + "learning_rate": 1e-05, + "loss": 1.2579, + "step": 1796 + }, + { + "epoch": 0.046662338866734675, + "grad_norm": 5.16579630914425, + "learning_rate": 1e-05, + "loss": 1.2663, + "step": 1797 + }, + { + "epoch": 0.046688305666326625, + "grad_norm": 4.638359158867013, + "learning_rate": 1e-05, + "loss": 1.2529, + "step": 1798 + }, + { + "epoch": 0.046714272465918576, + "grad_norm": 5.034115372255264, + "learning_rate": 1e-05, + "loss": 1.2965, + "step": 1799 + }, + { + "epoch": 0.046740239265510526, + "grad_norm": 5.7570823506688535, + "learning_rate": 1e-05, + "loss": 1.2264, + "step": 1800 + }, + { + "epoch": 0.04676620606510248, + "grad_norm": 11.771405141439413, + "learning_rate": 1e-05, + "loss": 1.3194, + "step": 1801 + }, + { + "epoch": 0.04679217286469443, + "grad_norm": 5.711536297801452, + "learning_rate": 1e-05, + "loss": 1.2243, + "step": 1802 + }, + { + "epoch": 0.04681813966428638, + "grad_norm": 4.504145725894398, + "learning_rate": 1e-05, + "loss": 1.281, + "step": 1803 + }, + { + "epoch": 0.04684410646387833, + "grad_norm": 17.22289058796688, + "learning_rate": 1e-05, + "loss": 1.2737, + "step": 1804 + }, + { + "epoch": 0.04687007326347028, + "grad_norm": 5.557855617387043, + "learning_rate": 1e-05, + "loss": 1.3096, + "step": 1805 + }, + { + "epoch": 0.04689604006306223, + "grad_norm": 9.79374311358046, + "learning_rate": 1e-05, + "loss": 1.3068, + "step": 1806 + }, + { + "epoch": 0.04692200686265418, + "grad_norm": 4.7254048088285305, + "learning_rate": 1e-05, + "loss": 1.2739, + "step": 1807 + }, + { + "epoch": 0.04694797366224613, + "grad_norm": 5.9310971721740895, + "learning_rate": 1e-05, + "loss": 1.2751, + "step": 1808 + }, + { + "epoch": 0.04697394046183808, + "grad_norm": 6.024679290973257, + "learning_rate": 1e-05, + "loss": 1.3386, + "step": 1809 + }, + { + "epoch": 0.04699990726143003, + "grad_norm": 11.9871566707384, + "learning_rate": 1e-05, + "loss": 1.3071, + "step": 1810 + }, + { + "epoch": 0.04702587406102198, + "grad_norm": 6.128457475403744, + "learning_rate": 1e-05, + "loss": 1.2229, + "step": 1811 + }, + { + "epoch": 0.04705184086061393, + "grad_norm": 11.25616184604657, + "learning_rate": 1e-05, + "loss": 1.213, + "step": 1812 + }, + { + "epoch": 0.04707780766020588, + "grad_norm": 7.776655564260045, + "learning_rate": 1e-05, + "loss": 1.294, + "step": 1813 + }, + { + "epoch": 0.04710377445979783, + "grad_norm": 6.407509879924348, + "learning_rate": 1e-05, + "loss": 1.2663, + "step": 1814 + }, + { + "epoch": 0.04712974125938978, + "grad_norm": 8.986304140711853, + "learning_rate": 1e-05, + "loss": 1.2766, + "step": 1815 + }, + { + "epoch": 0.047155708058981734, + "grad_norm": 5.143992888972231, + "learning_rate": 1e-05, + "loss": 1.3125, + "step": 1816 + }, + { + "epoch": 0.047181674858573684, + "grad_norm": 10.601973411113656, + "learning_rate": 1e-05, + "loss": 1.2638, + "step": 1817 + }, + { + "epoch": 0.04720764165816563, + "grad_norm": 10.412734935074694, + "learning_rate": 1e-05, + "loss": 1.2632, + "step": 1818 + }, + { + "epoch": 0.04723360845775758, + "grad_norm": 5.695544579569308, + "learning_rate": 1e-05, + "loss": 1.3273, + "step": 1819 + }, + { + "epoch": 0.04725957525734953, + "grad_norm": 7.505539946120632, + "learning_rate": 1e-05, + "loss": 1.3043, + "step": 1820 + }, + { + "epoch": 0.04728554205694148, + "grad_norm": 16.054332987265948, + "learning_rate": 1e-05, + "loss": 1.2895, + "step": 1821 + }, + { + "epoch": 0.04731150885653343, + "grad_norm": 7.460750621432476, + "learning_rate": 1e-05, + "loss": 1.272, + "step": 1822 + }, + { + "epoch": 0.04733747565612538, + "grad_norm": 6.99519837409696, + "learning_rate": 1e-05, + "loss": 1.2453, + "step": 1823 + }, + { + "epoch": 0.04736344245571733, + "grad_norm": 5.553610738913908, + "learning_rate": 1e-05, + "loss": 1.3091, + "step": 1824 + }, + { + "epoch": 0.04738940925530928, + "grad_norm": 6.370998248536985, + "learning_rate": 1e-05, + "loss": 1.2821, + "step": 1825 + }, + { + "epoch": 0.04741537605490123, + "grad_norm": 10.909058654621877, + "learning_rate": 1e-05, + "loss": 1.251, + "step": 1826 + }, + { + "epoch": 0.04744134285449318, + "grad_norm": 8.838521861089944, + "learning_rate": 1e-05, + "loss": 1.2356, + "step": 1827 + }, + { + "epoch": 0.04746730965408513, + "grad_norm": 5.671735777604333, + "learning_rate": 1e-05, + "loss": 1.2714, + "step": 1828 + }, + { + "epoch": 0.04749327645367708, + "grad_norm": 9.667104601531584, + "learning_rate": 1e-05, + "loss": 1.2714, + "step": 1829 + }, + { + "epoch": 0.04751924325326903, + "grad_norm": 6.224651354482697, + "learning_rate": 1e-05, + "loss": 1.241, + "step": 1830 + }, + { + "epoch": 0.047545210052860984, + "grad_norm": 7.9197597952743335, + "learning_rate": 1e-05, + "loss": 1.2736, + "step": 1831 + }, + { + "epoch": 0.047571176852452934, + "grad_norm": 9.063242072931661, + "learning_rate": 1e-05, + "loss": 1.2873, + "step": 1832 + }, + { + "epoch": 0.047597143652044885, + "grad_norm": 26.38246779460924, + "learning_rate": 1e-05, + "loss": 1.3016, + "step": 1833 + }, + { + "epoch": 0.047623110451636835, + "grad_norm": 6.020588518949119, + "learning_rate": 1e-05, + "loss": 1.2515, + "step": 1834 + }, + { + "epoch": 0.047649077251228786, + "grad_norm": 9.365844732131038, + "learning_rate": 1e-05, + "loss": 1.2558, + "step": 1835 + }, + { + "epoch": 0.047675044050820736, + "grad_norm": 6.30715595312507, + "learning_rate": 1e-05, + "loss": 1.3198, + "step": 1836 + }, + { + "epoch": 0.04770101085041269, + "grad_norm": 6.036441762362213, + "learning_rate": 1e-05, + "loss": 1.2297, + "step": 1837 + }, + { + "epoch": 0.04772697765000464, + "grad_norm": 25.37793673354007, + "learning_rate": 1e-05, + "loss": 1.2563, + "step": 1838 + }, + { + "epoch": 0.04775294444959659, + "grad_norm": 6.673528319008572, + "learning_rate": 1e-05, + "loss": 1.2497, + "step": 1839 + }, + { + "epoch": 0.04777891124918854, + "grad_norm": 12.81490307113777, + "learning_rate": 1e-05, + "loss": 1.259, + "step": 1840 + }, + { + "epoch": 0.04780487804878049, + "grad_norm": 6.533374203664399, + "learning_rate": 1e-05, + "loss": 1.2667, + "step": 1841 + }, + { + "epoch": 0.04783084484837244, + "grad_norm": 5.227597714299449, + "learning_rate": 1e-05, + "loss": 1.273, + "step": 1842 + }, + { + "epoch": 0.04785681164796439, + "grad_norm": 5.442485759753275, + "learning_rate": 1e-05, + "loss": 1.2969, + "step": 1843 + }, + { + "epoch": 0.04788277844755634, + "grad_norm": 4.487993754051065, + "learning_rate": 1e-05, + "loss": 1.3241, + "step": 1844 + }, + { + "epoch": 0.04790874524714829, + "grad_norm": 29.77195019737013, + "learning_rate": 1e-05, + "loss": 1.2436, + "step": 1845 + }, + { + "epoch": 0.04793471204674024, + "grad_norm": 5.450985901577286, + "learning_rate": 1e-05, + "loss": 1.2334, + "step": 1846 + }, + { + "epoch": 0.04796067884633219, + "grad_norm": 5.4091210652232755, + "learning_rate": 1e-05, + "loss": 1.3121, + "step": 1847 + }, + { + "epoch": 0.04798664564592414, + "grad_norm": 5.012766371044418, + "learning_rate": 1e-05, + "loss": 1.2827, + "step": 1848 + }, + { + "epoch": 0.04801261244551609, + "grad_norm": 4.986513067895636, + "learning_rate": 1e-05, + "loss": 1.2933, + "step": 1849 + }, + { + "epoch": 0.04803857924510804, + "grad_norm": 6.9523310861709025, + "learning_rate": 1e-05, + "loss": 1.2813, + "step": 1850 + }, + { + "epoch": 0.04806454604469999, + "grad_norm": 5.75233611044583, + "learning_rate": 1e-05, + "loss": 1.2959, + "step": 1851 + }, + { + "epoch": 0.048090512844291944, + "grad_norm": 6.400184998222677, + "learning_rate": 1e-05, + "loss": 1.2793, + "step": 1852 + }, + { + "epoch": 0.048116479643883894, + "grad_norm": 6.288724333396645, + "learning_rate": 1e-05, + "loss": 1.2899, + "step": 1853 + }, + { + "epoch": 0.048142446443475845, + "grad_norm": 8.156633506908081, + "learning_rate": 1e-05, + "loss": 1.2779, + "step": 1854 + }, + { + "epoch": 0.048168413243067795, + "grad_norm": 21.957389487937768, + "learning_rate": 1e-05, + "loss": 1.2915, + "step": 1855 + }, + { + "epoch": 0.04819438004265974, + "grad_norm": 7.501821423928569, + "learning_rate": 1e-05, + "loss": 1.2449, + "step": 1856 + }, + { + "epoch": 0.04822034684225169, + "grad_norm": 10.132804970602445, + "learning_rate": 1e-05, + "loss": 1.2915, + "step": 1857 + }, + { + "epoch": 0.04824631364184364, + "grad_norm": 8.386159720976142, + "learning_rate": 1e-05, + "loss": 1.3394, + "step": 1858 + }, + { + "epoch": 0.04827228044143559, + "grad_norm": 5.933303474322564, + "learning_rate": 1e-05, + "loss": 1.2832, + "step": 1859 + }, + { + "epoch": 0.04829824724102754, + "grad_norm": 6.02826581269407, + "learning_rate": 1e-05, + "loss": 1.2808, + "step": 1860 + }, + { + "epoch": 0.04832421404061949, + "grad_norm": 5.5473966111480575, + "learning_rate": 1e-05, + "loss": 1.2992, + "step": 1861 + }, + { + "epoch": 0.04835018084021144, + "grad_norm": 15.058370823576883, + "learning_rate": 1e-05, + "loss": 1.2718, + "step": 1862 + }, + { + "epoch": 0.04837614763980339, + "grad_norm": 7.9987296049415715, + "learning_rate": 1e-05, + "loss": 1.2713, + "step": 1863 + }, + { + "epoch": 0.04840211443939534, + "grad_norm": 5.040207370548831, + "learning_rate": 1e-05, + "loss": 1.2685, + "step": 1864 + }, + { + "epoch": 0.04842808123898729, + "grad_norm": 8.871560303258693, + "learning_rate": 1e-05, + "loss": 1.3142, + "step": 1865 + }, + { + "epoch": 0.048454048038579244, + "grad_norm": 5.335537494729656, + "learning_rate": 1e-05, + "loss": 1.3385, + "step": 1866 + }, + { + "epoch": 0.048480014838171194, + "grad_norm": 5.930859194905484, + "learning_rate": 1e-05, + "loss": 1.3099, + "step": 1867 + }, + { + "epoch": 0.048505981637763144, + "grad_norm": 19.378854675777976, + "learning_rate": 1e-05, + "loss": 1.2721, + "step": 1868 + }, + { + "epoch": 0.048531948437355095, + "grad_norm": 7.510099858459487, + "learning_rate": 1e-05, + "loss": 1.2844, + "step": 1869 + }, + { + "epoch": 0.048557915236947045, + "grad_norm": 8.108119860116501, + "learning_rate": 1e-05, + "loss": 1.2596, + "step": 1870 + }, + { + "epoch": 0.048583882036538996, + "grad_norm": 11.019529172943797, + "learning_rate": 1e-05, + "loss": 1.268, + "step": 1871 + }, + { + "epoch": 0.048609848836130946, + "grad_norm": 5.517152223606601, + "learning_rate": 1e-05, + "loss": 1.2922, + "step": 1872 + }, + { + "epoch": 0.0486358156357229, + "grad_norm": 6.498840815545552, + "learning_rate": 1e-05, + "loss": 1.3098, + "step": 1873 + }, + { + "epoch": 0.04866178243531485, + "grad_norm": 9.526329095980355, + "learning_rate": 1e-05, + "loss": 1.2861, + "step": 1874 + }, + { + "epoch": 0.0486877492349068, + "grad_norm": 8.134909822026302, + "learning_rate": 1e-05, + "loss": 1.262, + "step": 1875 + }, + { + "epoch": 0.04871371603449875, + "grad_norm": 16.934608803919765, + "learning_rate": 1e-05, + "loss": 1.2457, + "step": 1876 + }, + { + "epoch": 0.0487396828340907, + "grad_norm": 7.840316689770792, + "learning_rate": 1e-05, + "loss": 1.333, + "step": 1877 + }, + { + "epoch": 0.04876564963368265, + "grad_norm": 8.89527318353774, + "learning_rate": 1e-05, + "loss": 1.2377, + "step": 1878 + }, + { + "epoch": 0.0487916164332746, + "grad_norm": 23.277465348839126, + "learning_rate": 1e-05, + "loss": 1.2792, + "step": 1879 + }, + { + "epoch": 0.04881758323286655, + "grad_norm": 47.42240269475757, + "learning_rate": 1e-05, + "loss": 1.2559, + "step": 1880 + }, + { + "epoch": 0.0488435500324585, + "grad_norm": 5.750630219793713, + "learning_rate": 1e-05, + "loss": 1.3052, + "step": 1881 + }, + { + "epoch": 0.04886951683205045, + "grad_norm": 7.704687985211529, + "learning_rate": 1e-05, + "loss": 1.2874, + "step": 1882 + }, + { + "epoch": 0.0488954836316424, + "grad_norm": 15.410966688923487, + "learning_rate": 1e-05, + "loss": 1.3248, + "step": 1883 + }, + { + "epoch": 0.04892145043123435, + "grad_norm": 5.416523427781068, + "learning_rate": 1e-05, + "loss": 1.2823, + "step": 1884 + }, + { + "epoch": 0.0489474172308263, + "grad_norm": 7.031083168593701, + "learning_rate": 1e-05, + "loss": 1.3193, + "step": 1885 + }, + { + "epoch": 0.04897338403041825, + "grad_norm": 5.31595221156562, + "learning_rate": 1e-05, + "loss": 1.2714, + "step": 1886 + }, + { + "epoch": 0.048999350830010203, + "grad_norm": 5.84494349077905, + "learning_rate": 1e-05, + "loss": 1.3204, + "step": 1887 + }, + { + "epoch": 0.049025317629602154, + "grad_norm": 7.520574450770372, + "learning_rate": 1e-05, + "loss": 1.2886, + "step": 1888 + }, + { + "epoch": 0.049051284429194104, + "grad_norm": 7.305471319076385, + "learning_rate": 1e-05, + "loss": 1.2647, + "step": 1889 + }, + { + "epoch": 0.049077251228786055, + "grad_norm": 5.645390492018515, + "learning_rate": 1e-05, + "loss": 1.2263, + "step": 1890 + }, + { + "epoch": 0.049103218028378005, + "grad_norm": 5.741811145723473, + "learning_rate": 1e-05, + "loss": 1.296, + "step": 1891 + }, + { + "epoch": 0.049129184827969956, + "grad_norm": 5.444795646931189, + "learning_rate": 1e-05, + "loss": 1.2392, + "step": 1892 + }, + { + "epoch": 0.049155151627561906, + "grad_norm": 9.381855810276782, + "learning_rate": 1e-05, + "loss": 1.2825, + "step": 1893 + }, + { + "epoch": 0.04918111842715385, + "grad_norm": 5.961162117941997, + "learning_rate": 1e-05, + "loss": 1.2588, + "step": 1894 + }, + { + "epoch": 0.0492070852267458, + "grad_norm": 10.247577660129764, + "learning_rate": 1e-05, + "loss": 1.3389, + "step": 1895 + }, + { + "epoch": 0.04923305202633775, + "grad_norm": 7.182911345204422, + "learning_rate": 1e-05, + "loss": 1.3139, + "step": 1896 + }, + { + "epoch": 0.0492590188259297, + "grad_norm": 5.600676611488091, + "learning_rate": 1e-05, + "loss": 1.2358, + "step": 1897 + }, + { + "epoch": 0.04928498562552165, + "grad_norm": 10.935695651950535, + "learning_rate": 1e-05, + "loss": 1.2852, + "step": 1898 + }, + { + "epoch": 0.0493109524251136, + "grad_norm": 16.173164765899156, + "learning_rate": 1e-05, + "loss": 1.2505, + "step": 1899 + }, + { + "epoch": 0.04933691922470555, + "grad_norm": 6.5173668305576244, + "learning_rate": 1e-05, + "loss": 1.2505, + "step": 1900 + }, + { + "epoch": 0.0493628860242975, + "grad_norm": 6.450296959953242, + "learning_rate": 1e-05, + "loss": 1.2519, + "step": 1901 + }, + { + "epoch": 0.049388852823889454, + "grad_norm": 7.104233846784936, + "learning_rate": 1e-05, + "loss": 1.269, + "step": 1902 + }, + { + "epoch": 0.049414819623481404, + "grad_norm": 8.57153333872391, + "learning_rate": 1e-05, + "loss": 1.3144, + "step": 1903 + }, + { + "epoch": 0.049440786423073355, + "grad_norm": 6.667142024576866, + "learning_rate": 1e-05, + "loss": 1.3283, + "step": 1904 + }, + { + "epoch": 0.049466753222665305, + "grad_norm": 7.236017621543435, + "learning_rate": 1e-05, + "loss": 1.2739, + "step": 1905 + }, + { + "epoch": 0.049492720022257256, + "grad_norm": 7.5192151922214325, + "learning_rate": 1e-05, + "loss": 1.2498, + "step": 1906 + }, + { + "epoch": 0.049518686821849206, + "grad_norm": 8.165164212446124, + "learning_rate": 1e-05, + "loss": 1.2889, + "step": 1907 + }, + { + "epoch": 0.049544653621441156, + "grad_norm": 5.052787789326772, + "learning_rate": 1e-05, + "loss": 1.2799, + "step": 1908 + }, + { + "epoch": 0.04957062042103311, + "grad_norm": 5.886304976230281, + "learning_rate": 1e-05, + "loss": 1.3188, + "step": 1909 + }, + { + "epoch": 0.04959658722062506, + "grad_norm": 5.355468215775411, + "learning_rate": 1e-05, + "loss": 1.2952, + "step": 1910 + }, + { + "epoch": 0.04962255402021701, + "grad_norm": 9.477256856389019, + "learning_rate": 1e-05, + "loss": 1.2978, + "step": 1911 + }, + { + "epoch": 0.04964852081980896, + "grad_norm": 24.494144907826954, + "learning_rate": 1e-05, + "loss": 1.2791, + "step": 1912 + }, + { + "epoch": 0.04967448761940091, + "grad_norm": 9.894700884353217, + "learning_rate": 1e-05, + "loss": 1.2989, + "step": 1913 + }, + { + "epoch": 0.04970045441899286, + "grad_norm": 6.699718116410232, + "learning_rate": 1e-05, + "loss": 1.2645, + "step": 1914 + }, + { + "epoch": 0.04972642121858481, + "grad_norm": 7.481172963752022, + "learning_rate": 1e-05, + "loss": 1.2843, + "step": 1915 + }, + { + "epoch": 0.04975238801817676, + "grad_norm": 7.556064105527804, + "learning_rate": 1e-05, + "loss": 1.2682, + "step": 1916 + }, + { + "epoch": 0.04977835481776871, + "grad_norm": 5.629437878802716, + "learning_rate": 1e-05, + "loss": 1.2963, + "step": 1917 + }, + { + "epoch": 0.04980432161736066, + "grad_norm": 13.11422606999891, + "learning_rate": 1e-05, + "loss": 1.2631, + "step": 1918 + }, + { + "epoch": 0.04983028841695261, + "grad_norm": 6.860488001868329, + "learning_rate": 1e-05, + "loss": 1.3416, + "step": 1919 + }, + { + "epoch": 0.04985625521654456, + "grad_norm": 9.015231066052406, + "learning_rate": 1e-05, + "loss": 1.2571, + "step": 1920 + }, + { + "epoch": 0.04988222201613651, + "grad_norm": 8.789088107601584, + "learning_rate": 1e-05, + "loss": 1.3522, + "step": 1921 + }, + { + "epoch": 0.04990818881572846, + "grad_norm": 6.331732129251189, + "learning_rate": 1e-05, + "loss": 1.2859, + "step": 1922 + }, + { + "epoch": 0.049934155615320414, + "grad_norm": 16.76210427268717, + "learning_rate": 1e-05, + "loss": 1.2846, + "step": 1923 + }, + { + "epoch": 0.049960122414912364, + "grad_norm": 8.683869186336823, + "learning_rate": 1e-05, + "loss": 1.2787, + "step": 1924 + }, + { + "epoch": 0.049986089214504315, + "grad_norm": 7.674416426052781, + "learning_rate": 1e-05, + "loss": 1.317, + "step": 1925 + }, + { + "epoch": 0.050012056014096265, + "grad_norm": 35.80945599519595, + "learning_rate": 1e-05, + "loss": 1.3138, + "step": 1926 + }, + { + "epoch": 0.050038022813688215, + "grad_norm": 5.267645835794885, + "learning_rate": 1e-05, + "loss": 1.3132, + "step": 1927 + }, + { + "epoch": 0.050063989613280166, + "grad_norm": 7.4721245584010285, + "learning_rate": 1e-05, + "loss": 1.2943, + "step": 1928 + }, + { + "epoch": 0.050089956412872116, + "grad_norm": 41.92557871265017, + "learning_rate": 1e-05, + "loss": 1.3037, + "step": 1929 + }, + { + "epoch": 0.05011592321246407, + "grad_norm": 6.851926891767524, + "learning_rate": 1e-05, + "loss": 1.2423, + "step": 1930 + }, + { + "epoch": 0.05014189001205602, + "grad_norm": 22.15501996541571, + "learning_rate": 1e-05, + "loss": 1.2962, + "step": 1931 + }, + { + "epoch": 0.05016785681164797, + "grad_norm": 6.080687622538312, + "learning_rate": 1e-05, + "loss": 1.2635, + "step": 1932 + }, + { + "epoch": 0.05019382361123991, + "grad_norm": 5.450659251805865, + "learning_rate": 1e-05, + "loss": 1.3056, + "step": 1933 + }, + { + "epoch": 0.05021979041083186, + "grad_norm": 7.361712884197228, + "learning_rate": 1e-05, + "loss": 1.2682, + "step": 1934 + }, + { + "epoch": 0.05024575721042381, + "grad_norm": 7.013675138365991, + "learning_rate": 1e-05, + "loss": 1.284, + "step": 1935 + }, + { + "epoch": 0.05027172401001576, + "grad_norm": 5.417715098322971, + "learning_rate": 1e-05, + "loss": 1.278, + "step": 1936 + }, + { + "epoch": 0.05029769080960771, + "grad_norm": 6.167276335125485, + "learning_rate": 1e-05, + "loss": 1.2989, + "step": 1937 + }, + { + "epoch": 0.050323657609199664, + "grad_norm": 7.897196173165571, + "learning_rate": 1e-05, + "loss": 1.3121, + "step": 1938 + }, + { + "epoch": 0.050349624408791614, + "grad_norm": 7.286270179164323, + "learning_rate": 1e-05, + "loss": 1.2444, + "step": 1939 + }, + { + "epoch": 0.050375591208383565, + "grad_norm": 7.1853008057044505, + "learning_rate": 1e-05, + "loss": 1.2674, + "step": 1940 + }, + { + "epoch": 0.050401558007975515, + "grad_norm": 6.271557669982008, + "learning_rate": 1e-05, + "loss": 1.3054, + "step": 1941 + }, + { + "epoch": 0.050427524807567466, + "grad_norm": 5.917300588828061, + "learning_rate": 1e-05, + "loss": 1.2937, + "step": 1942 + }, + { + "epoch": 0.050453491607159416, + "grad_norm": 7.138916078244428, + "learning_rate": 1e-05, + "loss": 1.2365, + "step": 1943 + }, + { + "epoch": 0.05047945840675137, + "grad_norm": 7.104732934602221, + "learning_rate": 1e-05, + "loss": 1.3306, + "step": 1944 + }, + { + "epoch": 0.05050542520634332, + "grad_norm": 5.576205225108238, + "learning_rate": 1e-05, + "loss": 1.3042, + "step": 1945 + }, + { + "epoch": 0.05053139200593527, + "grad_norm": 8.66255833990095, + "learning_rate": 1e-05, + "loss": 1.2376, + "step": 1946 + }, + { + "epoch": 0.05055735880552722, + "grad_norm": 18.0399261175364, + "learning_rate": 1e-05, + "loss": 1.2949, + "step": 1947 + }, + { + "epoch": 0.05058332560511917, + "grad_norm": 7.159300840968631, + "learning_rate": 1e-05, + "loss": 1.2639, + "step": 1948 + }, + { + "epoch": 0.05060929240471112, + "grad_norm": 17.2515010457084, + "learning_rate": 1e-05, + "loss": 1.2841, + "step": 1949 + }, + { + "epoch": 0.05063525920430307, + "grad_norm": 8.019264391413625, + "learning_rate": 1e-05, + "loss": 1.3107, + "step": 1950 + }, + { + "epoch": 0.05066122600389502, + "grad_norm": 7.399443126035868, + "learning_rate": 1e-05, + "loss": 1.2958, + "step": 1951 + }, + { + "epoch": 0.05068719280348697, + "grad_norm": 7.1413207935928735, + "learning_rate": 1e-05, + "loss": 1.2988, + "step": 1952 + }, + { + "epoch": 0.05071315960307892, + "grad_norm": 10.667791386753354, + "learning_rate": 1e-05, + "loss": 1.293, + "step": 1953 + }, + { + "epoch": 0.05073912640267087, + "grad_norm": 18.724929528301136, + "learning_rate": 1e-05, + "loss": 1.2211, + "step": 1954 + }, + { + "epoch": 0.05076509320226282, + "grad_norm": 7.579356677452445, + "learning_rate": 1e-05, + "loss": 1.3303, + "step": 1955 + }, + { + "epoch": 0.05079106000185477, + "grad_norm": 10.801832990157996, + "learning_rate": 1e-05, + "loss": 1.2363, + "step": 1956 + }, + { + "epoch": 0.05081702680144672, + "grad_norm": 8.379782606414185, + "learning_rate": 1e-05, + "loss": 1.2587, + "step": 1957 + }, + { + "epoch": 0.05084299360103867, + "grad_norm": 5.610315270634359, + "learning_rate": 1e-05, + "loss": 1.3072, + "step": 1958 + }, + { + "epoch": 0.050868960400630624, + "grad_norm": 7.233287886141325, + "learning_rate": 1e-05, + "loss": 1.2657, + "step": 1959 + }, + { + "epoch": 0.050894927200222574, + "grad_norm": 8.832279166412116, + "learning_rate": 1e-05, + "loss": 1.2828, + "step": 1960 + }, + { + "epoch": 0.050920893999814525, + "grad_norm": 6.818479940659199, + "learning_rate": 1e-05, + "loss": 1.2806, + "step": 1961 + }, + { + "epoch": 0.050946860799406475, + "grad_norm": 5.51974342192168, + "learning_rate": 1e-05, + "loss": 1.296, + "step": 1962 + }, + { + "epoch": 0.050972827598998426, + "grad_norm": 7.283832665035375, + "learning_rate": 1e-05, + "loss": 1.2683, + "step": 1963 + }, + { + "epoch": 0.050998794398590376, + "grad_norm": 9.847875499592222, + "learning_rate": 1e-05, + "loss": 1.307, + "step": 1964 + }, + { + "epoch": 0.051024761198182327, + "grad_norm": 7.043855075001682, + "learning_rate": 1e-05, + "loss": 1.2745, + "step": 1965 + }, + { + "epoch": 0.05105072799777428, + "grad_norm": 7.726839467514404, + "learning_rate": 1e-05, + "loss": 1.2952, + "step": 1966 + }, + { + "epoch": 0.05107669479736623, + "grad_norm": 8.155228474104435, + "learning_rate": 1e-05, + "loss": 1.2697, + "step": 1967 + }, + { + "epoch": 0.05110266159695818, + "grad_norm": 5.484086510007219, + "learning_rate": 1e-05, + "loss": 1.2675, + "step": 1968 + }, + { + "epoch": 0.05112862839655013, + "grad_norm": 7.460525133810408, + "learning_rate": 1e-05, + "loss": 1.2511, + "step": 1969 + }, + { + "epoch": 0.05115459519614208, + "grad_norm": 6.515627049141615, + "learning_rate": 1e-05, + "loss": 1.2067, + "step": 1970 + }, + { + "epoch": 0.05118056199573402, + "grad_norm": 8.996293894321584, + "learning_rate": 1e-05, + "loss": 1.2794, + "step": 1971 + }, + { + "epoch": 0.05120652879532597, + "grad_norm": 5.22397137605732, + "learning_rate": 1e-05, + "loss": 1.2795, + "step": 1972 + }, + { + "epoch": 0.05123249559491792, + "grad_norm": 7.822681841688889, + "learning_rate": 1e-05, + "loss": 1.2888, + "step": 1973 + }, + { + "epoch": 0.051258462394509874, + "grad_norm": 8.473564327922082, + "learning_rate": 1e-05, + "loss": 1.2845, + "step": 1974 + }, + { + "epoch": 0.051284429194101824, + "grad_norm": 5.960116708624079, + "learning_rate": 1e-05, + "loss": 1.271, + "step": 1975 + }, + { + "epoch": 0.051310395993693775, + "grad_norm": 15.030822620938263, + "learning_rate": 1e-05, + "loss": 1.2761, + "step": 1976 + }, + { + "epoch": 0.051336362793285725, + "grad_norm": 7.275435979304941, + "learning_rate": 1e-05, + "loss": 1.31, + "step": 1977 + }, + { + "epoch": 0.051362329592877676, + "grad_norm": 6.196647278972719, + "learning_rate": 1e-05, + "loss": 1.2901, + "step": 1978 + }, + { + "epoch": 0.051388296392469626, + "grad_norm": 12.669756729933813, + "learning_rate": 1e-05, + "loss": 1.268, + "step": 1979 + }, + { + "epoch": 0.05141426319206158, + "grad_norm": 8.004492928565622, + "learning_rate": 1e-05, + "loss": 1.2793, + "step": 1980 + }, + { + "epoch": 0.05144022999165353, + "grad_norm": 6.340285684785605, + "learning_rate": 1e-05, + "loss": 1.2903, + "step": 1981 + }, + { + "epoch": 0.05146619679124548, + "grad_norm": 5.165443870475055, + "learning_rate": 1e-05, + "loss": 1.3074, + "step": 1982 + }, + { + "epoch": 0.05149216359083743, + "grad_norm": 6.186666104930545, + "learning_rate": 1e-05, + "loss": 1.3231, + "step": 1983 + }, + { + "epoch": 0.05151813039042938, + "grad_norm": 5.85508338927477, + "learning_rate": 1e-05, + "loss": 1.2739, + "step": 1984 + }, + { + "epoch": 0.05154409719002133, + "grad_norm": 30.165300819570096, + "learning_rate": 1e-05, + "loss": 1.2715, + "step": 1985 + }, + { + "epoch": 0.05157006398961328, + "grad_norm": 9.737936136960416, + "learning_rate": 1e-05, + "loss": 1.3253, + "step": 1986 + }, + { + "epoch": 0.05159603078920523, + "grad_norm": 9.045412989730494, + "learning_rate": 1e-05, + "loss": 1.2976, + "step": 1987 + }, + { + "epoch": 0.05162199758879718, + "grad_norm": 8.579616396138164, + "learning_rate": 1e-05, + "loss": 1.2561, + "step": 1988 + }, + { + "epoch": 0.05164796438838913, + "grad_norm": 12.90841989079796, + "learning_rate": 1e-05, + "loss": 1.3233, + "step": 1989 + }, + { + "epoch": 0.05167393118798108, + "grad_norm": 4.557910439763412, + "learning_rate": 1e-05, + "loss": 1.2706, + "step": 1990 + }, + { + "epoch": 0.05169989798757303, + "grad_norm": 10.965495004261165, + "learning_rate": 1e-05, + "loss": 1.3033, + "step": 1991 + }, + { + "epoch": 0.05172586478716498, + "grad_norm": 7.669509056722117, + "learning_rate": 1e-05, + "loss": 1.3076, + "step": 1992 + }, + { + "epoch": 0.05175183158675693, + "grad_norm": 5.631973690903509, + "learning_rate": 1e-05, + "loss": 1.3025, + "step": 1993 + }, + { + "epoch": 0.05177779838634888, + "grad_norm": 5.176555624559179, + "learning_rate": 1e-05, + "loss": 1.3025, + "step": 1994 + }, + { + "epoch": 0.051803765185940834, + "grad_norm": 23.682636013230375, + "learning_rate": 1e-05, + "loss": 1.2575, + "step": 1995 + }, + { + "epoch": 0.051829731985532784, + "grad_norm": 5.681461356315868, + "learning_rate": 1e-05, + "loss": 1.2854, + "step": 1996 + }, + { + "epoch": 0.051855698785124735, + "grad_norm": 8.652945538468344, + "learning_rate": 1e-05, + "loss": 1.2497, + "step": 1997 + }, + { + "epoch": 0.051881665584716685, + "grad_norm": 6.179629473890025, + "learning_rate": 1e-05, + "loss": 1.3119, + "step": 1998 + }, + { + "epoch": 0.051907632384308636, + "grad_norm": 8.440798870444022, + "learning_rate": 1e-05, + "loss": 1.2387, + "step": 1999 + }, + { + "epoch": 0.051933599183900586, + "grad_norm": 6.495810112005312, + "learning_rate": 1e-05, + "loss": 1.2576, + "step": 2000 + }, + { + "epoch": 0.05195956598349254, + "grad_norm": 7.046141186588436, + "learning_rate": 1e-05, + "loss": 1.2812, + "step": 2001 + }, + { + "epoch": 0.05198553278308449, + "grad_norm": 7.456441912935188, + "learning_rate": 1e-05, + "loss": 1.2388, + "step": 2002 + }, + { + "epoch": 0.05201149958267644, + "grad_norm": 4.34248783792869, + "learning_rate": 1e-05, + "loss": 1.2659, + "step": 2003 + }, + { + "epoch": 0.05203746638226839, + "grad_norm": 8.520546707020825, + "learning_rate": 1e-05, + "loss": 1.2979, + "step": 2004 + }, + { + "epoch": 0.05206343318186034, + "grad_norm": 14.374311613102936, + "learning_rate": 1e-05, + "loss": 1.2371, + "step": 2005 + }, + { + "epoch": 0.05208939998145229, + "grad_norm": 5.714403512285487, + "learning_rate": 1e-05, + "loss": 1.2518, + "step": 2006 + }, + { + "epoch": 0.05211536678104424, + "grad_norm": 4.628496240412189, + "learning_rate": 1e-05, + "loss": 1.3026, + "step": 2007 + }, + { + "epoch": 0.05214133358063619, + "grad_norm": 7.541762237224286, + "learning_rate": 1e-05, + "loss": 1.3003, + "step": 2008 + }, + { + "epoch": 0.052167300380228133, + "grad_norm": 18.71537982234214, + "learning_rate": 1e-05, + "loss": 1.2937, + "step": 2009 + }, + { + "epoch": 0.052193267179820084, + "grad_norm": 10.948203119703843, + "learning_rate": 1e-05, + "loss": 1.2741, + "step": 2010 + }, + { + "epoch": 0.052219233979412034, + "grad_norm": 7.997555836193165, + "learning_rate": 1e-05, + "loss": 1.273, + "step": 2011 + }, + { + "epoch": 0.052245200779003985, + "grad_norm": 10.929727637412899, + "learning_rate": 1e-05, + "loss": 1.2791, + "step": 2012 + }, + { + "epoch": 0.052271167578595935, + "grad_norm": 4.554758061538012, + "learning_rate": 1e-05, + "loss": 1.2714, + "step": 2013 + }, + { + "epoch": 0.052297134378187886, + "grad_norm": 12.979098610166343, + "learning_rate": 1e-05, + "loss": 1.2251, + "step": 2014 + }, + { + "epoch": 0.052323101177779836, + "grad_norm": 9.99226194447933, + "learning_rate": 1e-05, + "loss": 1.2636, + "step": 2015 + }, + { + "epoch": 0.05234906797737179, + "grad_norm": 4.777096808251354, + "learning_rate": 1e-05, + "loss": 1.2594, + "step": 2016 + }, + { + "epoch": 0.05237503477696374, + "grad_norm": 10.711208170358281, + "learning_rate": 1e-05, + "loss": 1.2861, + "step": 2017 + }, + { + "epoch": 0.05240100157655569, + "grad_norm": 7.09856933363848, + "learning_rate": 1e-05, + "loss": 1.2603, + "step": 2018 + }, + { + "epoch": 0.05242696837614764, + "grad_norm": 5.477157669269089, + "learning_rate": 1e-05, + "loss": 1.2434, + "step": 2019 + }, + { + "epoch": 0.05245293517573959, + "grad_norm": 6.302235936190437, + "learning_rate": 1e-05, + "loss": 1.3104, + "step": 2020 + }, + { + "epoch": 0.05247890197533154, + "grad_norm": 8.077370820542944, + "learning_rate": 1e-05, + "loss": 1.2296, + "step": 2021 + }, + { + "epoch": 0.05250486877492349, + "grad_norm": 7.6612368166339095, + "learning_rate": 1e-05, + "loss": 1.3311, + "step": 2022 + }, + { + "epoch": 0.05253083557451544, + "grad_norm": 4.752078956603103, + "learning_rate": 1e-05, + "loss": 1.3173, + "step": 2023 + }, + { + "epoch": 0.05255680237410739, + "grad_norm": 13.29327178797987, + "learning_rate": 1e-05, + "loss": 1.2686, + "step": 2024 + }, + { + "epoch": 0.05258276917369934, + "grad_norm": 8.032413621017133, + "learning_rate": 1e-05, + "loss": 1.2665, + "step": 2025 + }, + { + "epoch": 0.05260873597329129, + "grad_norm": 6.623783503712416, + "learning_rate": 1e-05, + "loss": 1.3128, + "step": 2026 + }, + { + "epoch": 0.05263470277288324, + "grad_norm": 5.490555717539195, + "learning_rate": 1e-05, + "loss": 1.2268, + "step": 2027 + }, + { + "epoch": 0.05266066957247519, + "grad_norm": 5.759088382636912, + "learning_rate": 1e-05, + "loss": 1.3001, + "step": 2028 + }, + { + "epoch": 0.05268663637206714, + "grad_norm": 22.238250716319282, + "learning_rate": 1e-05, + "loss": 1.2768, + "step": 2029 + }, + { + "epoch": 0.05271260317165909, + "grad_norm": 6.731680417268364, + "learning_rate": 1e-05, + "loss": 1.2657, + "step": 2030 + }, + { + "epoch": 0.052738569971251044, + "grad_norm": 6.079375699767681, + "learning_rate": 1e-05, + "loss": 1.2514, + "step": 2031 + }, + { + "epoch": 0.052764536770842994, + "grad_norm": 5.470313497203833, + "learning_rate": 1e-05, + "loss": 1.2676, + "step": 2032 + }, + { + "epoch": 0.052790503570434945, + "grad_norm": 6.055924457819661, + "learning_rate": 1e-05, + "loss": 1.3347, + "step": 2033 + }, + { + "epoch": 0.052816470370026895, + "grad_norm": 5.400938567082161, + "learning_rate": 1e-05, + "loss": 1.2915, + "step": 2034 + }, + { + "epoch": 0.052842437169618846, + "grad_norm": 5.827353582387971, + "learning_rate": 1e-05, + "loss": 1.2935, + "step": 2035 + }, + { + "epoch": 0.052868403969210796, + "grad_norm": 26.79458467799802, + "learning_rate": 1e-05, + "loss": 1.3218, + "step": 2036 + }, + { + "epoch": 0.05289437076880275, + "grad_norm": 4.962234543779694, + "learning_rate": 1e-05, + "loss": 1.2613, + "step": 2037 + }, + { + "epoch": 0.0529203375683947, + "grad_norm": 4.818645566211883, + "learning_rate": 1e-05, + "loss": 1.3221, + "step": 2038 + }, + { + "epoch": 0.05294630436798665, + "grad_norm": 8.225762570171828, + "learning_rate": 1e-05, + "loss": 1.2549, + "step": 2039 + }, + { + "epoch": 0.0529722711675786, + "grad_norm": 7.9022978905305505, + "learning_rate": 1e-05, + "loss": 1.3048, + "step": 2040 + }, + { + "epoch": 0.05299823796717055, + "grad_norm": 4.242486932629864, + "learning_rate": 1e-05, + "loss": 1.2713, + "step": 2041 + }, + { + "epoch": 0.0530242047667625, + "grad_norm": 4.380756405875824, + "learning_rate": 1e-05, + "loss": 1.29, + "step": 2042 + }, + { + "epoch": 0.05305017156635445, + "grad_norm": 8.17119996763834, + "learning_rate": 1e-05, + "loss": 1.2984, + "step": 2043 + }, + { + "epoch": 0.0530761383659464, + "grad_norm": 7.342557509979307, + "learning_rate": 1e-05, + "loss": 1.264, + "step": 2044 + }, + { + "epoch": 0.05310210516553835, + "grad_norm": 9.825722502333917, + "learning_rate": 1e-05, + "loss": 1.3002, + "step": 2045 + }, + { + "epoch": 0.0531280719651303, + "grad_norm": 13.629743677697192, + "learning_rate": 1e-05, + "loss": 1.3062, + "step": 2046 + }, + { + "epoch": 0.053154038764722245, + "grad_norm": 5.649298173643902, + "learning_rate": 1e-05, + "loss": 1.2867, + "step": 2047 + }, + { + "epoch": 0.053180005564314195, + "grad_norm": 5.946873062782954, + "learning_rate": 1e-05, + "loss": 1.2522, + "step": 2048 + }, + { + "epoch": 0.053205972363906145, + "grad_norm": 7.704384226913788, + "learning_rate": 1e-05, + "loss": 1.2725, + "step": 2049 + }, + { + "epoch": 0.053231939163498096, + "grad_norm": 22.801952024186026, + "learning_rate": 1e-05, + "loss": 1.2824, + "step": 2050 + }, + { + "epoch": 0.053257905963090046, + "grad_norm": 22.67654801944346, + "learning_rate": 1e-05, + "loss": 1.2636, + "step": 2051 + }, + { + "epoch": 0.053283872762682, + "grad_norm": 4.729420099175723, + "learning_rate": 1e-05, + "loss": 1.2399, + "step": 2052 + }, + { + "epoch": 0.05330983956227395, + "grad_norm": 5.689237559746514, + "learning_rate": 1e-05, + "loss": 1.2896, + "step": 2053 + }, + { + "epoch": 0.0533358063618659, + "grad_norm": 5.507604543669438, + "learning_rate": 1e-05, + "loss": 1.22, + "step": 2054 + }, + { + "epoch": 0.05336177316145785, + "grad_norm": 6.31127917882317, + "learning_rate": 1e-05, + "loss": 1.2642, + "step": 2055 + }, + { + "epoch": 0.0533877399610498, + "grad_norm": 5.66283829712148, + "learning_rate": 1e-05, + "loss": 1.279, + "step": 2056 + }, + { + "epoch": 0.05341370676064175, + "grad_norm": 7.938103119772173, + "learning_rate": 1e-05, + "loss": 1.3028, + "step": 2057 + }, + { + "epoch": 0.0534396735602337, + "grad_norm": 13.230785328918381, + "learning_rate": 1e-05, + "loss": 1.323, + "step": 2058 + }, + { + "epoch": 0.05346564035982565, + "grad_norm": 5.974034391770983, + "learning_rate": 1e-05, + "loss": 1.2399, + "step": 2059 + }, + { + "epoch": 0.0534916071594176, + "grad_norm": 9.088733665966833, + "learning_rate": 1e-05, + "loss": 1.2475, + "step": 2060 + }, + { + "epoch": 0.05351757395900955, + "grad_norm": 13.003263210670488, + "learning_rate": 1e-05, + "loss": 1.3097, + "step": 2061 + }, + { + "epoch": 0.0535435407586015, + "grad_norm": 10.033568307484483, + "learning_rate": 1e-05, + "loss": 1.2909, + "step": 2062 + }, + { + "epoch": 0.05356950755819345, + "grad_norm": 5.878188221828239, + "learning_rate": 1e-05, + "loss": 1.3137, + "step": 2063 + }, + { + "epoch": 0.0535954743577854, + "grad_norm": 4.478091632854469, + "learning_rate": 1e-05, + "loss": 1.286, + "step": 2064 + }, + { + "epoch": 0.05362144115737735, + "grad_norm": 6.691927291082938, + "learning_rate": 1e-05, + "loss": 1.2652, + "step": 2065 + }, + { + "epoch": 0.053647407956969304, + "grad_norm": 5.414715792170938, + "learning_rate": 1e-05, + "loss": 1.2798, + "step": 2066 + }, + { + "epoch": 0.053673374756561254, + "grad_norm": 7.73023459844664, + "learning_rate": 1e-05, + "loss": 1.2454, + "step": 2067 + }, + { + "epoch": 0.053699341556153204, + "grad_norm": 4.404931919641302, + "learning_rate": 1e-05, + "loss": 1.2073, + "step": 2068 + }, + { + "epoch": 0.053725308355745155, + "grad_norm": 15.537032158238356, + "learning_rate": 1e-05, + "loss": 1.2803, + "step": 2069 + }, + { + "epoch": 0.053751275155337105, + "grad_norm": 7.3226790048190145, + "learning_rate": 1e-05, + "loss": 1.3282, + "step": 2070 + }, + { + "epoch": 0.053777241954929056, + "grad_norm": 5.24153144471979, + "learning_rate": 1e-05, + "loss": 1.2942, + "step": 2071 + }, + { + "epoch": 0.053803208754521006, + "grad_norm": 4.78247453554177, + "learning_rate": 1e-05, + "loss": 1.3086, + "step": 2072 + }, + { + "epoch": 0.05382917555411296, + "grad_norm": 6.1759103451482025, + "learning_rate": 1e-05, + "loss": 1.2815, + "step": 2073 + }, + { + "epoch": 0.05385514235370491, + "grad_norm": 12.911274890076092, + "learning_rate": 1e-05, + "loss": 1.2536, + "step": 2074 + }, + { + "epoch": 0.05388110915329686, + "grad_norm": 10.679697858558436, + "learning_rate": 1e-05, + "loss": 1.2737, + "step": 2075 + }, + { + "epoch": 0.05390707595288881, + "grad_norm": 6.272036651368, + "learning_rate": 1e-05, + "loss": 1.2901, + "step": 2076 + }, + { + "epoch": 0.05393304275248076, + "grad_norm": 7.375676528159764, + "learning_rate": 1e-05, + "loss": 1.3133, + "step": 2077 + }, + { + "epoch": 0.05395900955207271, + "grad_norm": 5.128315853121752, + "learning_rate": 1e-05, + "loss": 1.2867, + "step": 2078 + }, + { + "epoch": 0.05398497635166466, + "grad_norm": 7.685301427545087, + "learning_rate": 1e-05, + "loss": 1.3115, + "step": 2079 + }, + { + "epoch": 0.05401094315125661, + "grad_norm": 4.5336594062296225, + "learning_rate": 1e-05, + "loss": 1.2903, + "step": 2080 + }, + { + "epoch": 0.05403690995084856, + "grad_norm": 25.991695545013748, + "learning_rate": 1e-05, + "loss": 1.2875, + "step": 2081 + }, + { + "epoch": 0.05406287675044051, + "grad_norm": 5.064863618631977, + "learning_rate": 1e-05, + "loss": 1.2662, + "step": 2082 + }, + { + "epoch": 0.05408884355003246, + "grad_norm": 15.072720554273756, + "learning_rate": 1e-05, + "loss": 1.3109, + "step": 2083 + }, + { + "epoch": 0.05411481034962441, + "grad_norm": 7.18625158995025, + "learning_rate": 1e-05, + "loss": 1.2089, + "step": 2084 + }, + { + "epoch": 0.05414077714921636, + "grad_norm": 5.055108596737036, + "learning_rate": 1e-05, + "loss": 1.2243, + "step": 2085 + }, + { + "epoch": 0.054166743948808306, + "grad_norm": 5.966859208451006, + "learning_rate": 1e-05, + "loss": 1.2946, + "step": 2086 + }, + { + "epoch": 0.05419271074840026, + "grad_norm": 5.1205149308188025, + "learning_rate": 1e-05, + "loss": 1.2797, + "step": 2087 + }, + { + "epoch": 0.05421867754799221, + "grad_norm": 11.006872977670907, + "learning_rate": 1e-05, + "loss": 1.2464, + "step": 2088 + }, + { + "epoch": 0.05424464434758416, + "grad_norm": 5.368394385607203, + "learning_rate": 1e-05, + "loss": 1.2736, + "step": 2089 + }, + { + "epoch": 0.05427061114717611, + "grad_norm": 9.175768777928944, + "learning_rate": 1e-05, + "loss": 1.2752, + "step": 2090 + }, + { + "epoch": 0.05429657794676806, + "grad_norm": 6.976119861465484, + "learning_rate": 1e-05, + "loss": 1.2977, + "step": 2091 + }, + { + "epoch": 0.05432254474636001, + "grad_norm": 6.2409437561385825, + "learning_rate": 1e-05, + "loss": 1.2971, + "step": 2092 + }, + { + "epoch": 0.05434851154595196, + "grad_norm": 4.760066959326874, + "learning_rate": 1e-05, + "loss": 1.2807, + "step": 2093 + }, + { + "epoch": 0.05437447834554391, + "grad_norm": 5.799261624921496, + "learning_rate": 1e-05, + "loss": 1.2916, + "step": 2094 + }, + { + "epoch": 0.05440044514513586, + "grad_norm": 6.47698008775146, + "learning_rate": 1e-05, + "loss": 1.2256, + "step": 2095 + }, + { + "epoch": 0.05442641194472781, + "grad_norm": 4.166591084112651, + "learning_rate": 1e-05, + "loss": 1.3049, + "step": 2096 + }, + { + "epoch": 0.05445237874431976, + "grad_norm": 6.544880492632528, + "learning_rate": 1e-05, + "loss": 1.2374, + "step": 2097 + }, + { + "epoch": 0.05447834554391171, + "grad_norm": 6.999009198503899, + "learning_rate": 1e-05, + "loss": 1.2905, + "step": 2098 + }, + { + "epoch": 0.05450431234350366, + "grad_norm": 13.416292426218963, + "learning_rate": 1e-05, + "loss": 1.2269, + "step": 2099 + }, + { + "epoch": 0.05453027914309561, + "grad_norm": 6.49773161994507, + "learning_rate": 1e-05, + "loss": 1.2274, + "step": 2100 + }, + { + "epoch": 0.05455624594268756, + "grad_norm": 7.09213682141028, + "learning_rate": 1e-05, + "loss": 1.3246, + "step": 2101 + }, + { + "epoch": 0.054582212742279514, + "grad_norm": 4.555797094956069, + "learning_rate": 1e-05, + "loss": 1.2481, + "step": 2102 + }, + { + "epoch": 0.054608179541871464, + "grad_norm": 20.566335395770636, + "learning_rate": 1e-05, + "loss": 1.2446, + "step": 2103 + }, + { + "epoch": 0.054634146341463415, + "grad_norm": 4.633022258570866, + "learning_rate": 1e-05, + "loss": 1.3527, + "step": 2104 + }, + { + "epoch": 0.054660113141055365, + "grad_norm": 4.598408315853067, + "learning_rate": 1e-05, + "loss": 1.2975, + "step": 2105 + }, + { + "epoch": 0.054686079940647316, + "grad_norm": 6.220679209137701, + "learning_rate": 1e-05, + "loss": 1.2744, + "step": 2106 + }, + { + "epoch": 0.054712046740239266, + "grad_norm": 4.389019112459098, + "learning_rate": 1e-05, + "loss": 1.2431, + "step": 2107 + }, + { + "epoch": 0.054738013539831216, + "grad_norm": 5.0000915518962, + "learning_rate": 1e-05, + "loss": 1.2645, + "step": 2108 + }, + { + "epoch": 0.05476398033942317, + "grad_norm": 13.169168612829171, + "learning_rate": 1e-05, + "loss": 1.2784, + "step": 2109 + }, + { + "epoch": 0.05478994713901512, + "grad_norm": 5.106315606768518, + "learning_rate": 1e-05, + "loss": 1.2647, + "step": 2110 + }, + { + "epoch": 0.05481591393860707, + "grad_norm": 10.188652862658621, + "learning_rate": 1e-05, + "loss": 1.2566, + "step": 2111 + }, + { + "epoch": 0.05484188073819902, + "grad_norm": 6.262976031761104, + "learning_rate": 1e-05, + "loss": 1.2669, + "step": 2112 + }, + { + "epoch": 0.05486784753779097, + "grad_norm": 6.569535326267466, + "learning_rate": 1e-05, + "loss": 1.2529, + "step": 2113 + }, + { + "epoch": 0.05489381433738292, + "grad_norm": 4.998011575136676, + "learning_rate": 1e-05, + "loss": 1.2523, + "step": 2114 + }, + { + "epoch": 0.05491978113697487, + "grad_norm": 4.542393734097836, + "learning_rate": 1e-05, + "loss": 1.2817, + "step": 2115 + }, + { + "epoch": 0.05494574793656682, + "grad_norm": 6.673422854976175, + "learning_rate": 1e-05, + "loss": 1.3104, + "step": 2116 + }, + { + "epoch": 0.05497171473615877, + "grad_norm": 4.23693747598605, + "learning_rate": 1e-05, + "loss": 1.3252, + "step": 2117 + }, + { + "epoch": 0.05499768153575072, + "grad_norm": 6.227500394976591, + "learning_rate": 1e-05, + "loss": 1.2458, + "step": 2118 + }, + { + "epoch": 0.05502364833534267, + "grad_norm": 4.115792372883921, + "learning_rate": 1e-05, + "loss": 1.275, + "step": 2119 + }, + { + "epoch": 0.05504961513493462, + "grad_norm": 6.5387740168406605, + "learning_rate": 1e-05, + "loss": 1.2348, + "step": 2120 + }, + { + "epoch": 0.05507558193452657, + "grad_norm": 5.744558662215298, + "learning_rate": 1e-05, + "loss": 1.2587, + "step": 2121 + }, + { + "epoch": 0.05510154873411852, + "grad_norm": 5.862946502665297, + "learning_rate": 1e-05, + "loss": 1.2482, + "step": 2122 + }, + { + "epoch": 0.055127515533710474, + "grad_norm": 9.37712866458531, + "learning_rate": 1e-05, + "loss": 1.2675, + "step": 2123 + }, + { + "epoch": 0.05515348233330242, + "grad_norm": 5.385608229436723, + "learning_rate": 1e-05, + "loss": 1.3648, + "step": 2124 + }, + { + "epoch": 0.05517944913289437, + "grad_norm": 5.969059321991458, + "learning_rate": 1e-05, + "loss": 1.3041, + "step": 2125 + }, + { + "epoch": 0.05520541593248632, + "grad_norm": 5.965984242944374, + "learning_rate": 1e-05, + "loss": 1.2962, + "step": 2126 + }, + { + "epoch": 0.05523138273207827, + "grad_norm": 6.208389316110147, + "learning_rate": 1e-05, + "loss": 1.2368, + "step": 2127 + }, + { + "epoch": 0.05525734953167022, + "grad_norm": 5.409067643467444, + "learning_rate": 1e-05, + "loss": 1.2486, + "step": 2128 + }, + { + "epoch": 0.05528331633126217, + "grad_norm": 14.814548270335534, + "learning_rate": 1e-05, + "loss": 1.2862, + "step": 2129 + }, + { + "epoch": 0.05530928313085412, + "grad_norm": 4.713593551884091, + "learning_rate": 1e-05, + "loss": 1.2756, + "step": 2130 + }, + { + "epoch": 0.05533524993044607, + "grad_norm": 3.6869992708226667, + "learning_rate": 1e-05, + "loss": 1.2583, + "step": 2131 + }, + { + "epoch": 0.05536121673003802, + "grad_norm": 4.121926289572622, + "learning_rate": 1e-05, + "loss": 1.2651, + "step": 2132 + }, + { + "epoch": 0.05538718352962997, + "grad_norm": 5.174222016020315, + "learning_rate": 1e-05, + "loss": 1.3267, + "step": 2133 + }, + { + "epoch": 0.05541315032922192, + "grad_norm": 6.186118231722982, + "learning_rate": 1e-05, + "loss": 1.2551, + "step": 2134 + }, + { + "epoch": 0.05543911712881387, + "grad_norm": 4.754959127769943, + "learning_rate": 1e-05, + "loss": 1.3184, + "step": 2135 + }, + { + "epoch": 0.05546508392840582, + "grad_norm": 4.800845612109245, + "learning_rate": 1e-05, + "loss": 1.2883, + "step": 2136 + }, + { + "epoch": 0.05549105072799777, + "grad_norm": 4.335998644698378, + "learning_rate": 1e-05, + "loss": 1.2957, + "step": 2137 + }, + { + "epoch": 0.055517017527589724, + "grad_norm": 4.035827639949442, + "learning_rate": 1e-05, + "loss": 1.3026, + "step": 2138 + }, + { + "epoch": 0.055542984327181674, + "grad_norm": 35.0505790368763, + "learning_rate": 1e-05, + "loss": 1.2665, + "step": 2139 + }, + { + "epoch": 0.055568951126773625, + "grad_norm": 6.336224983363116, + "learning_rate": 1e-05, + "loss": 1.2826, + "step": 2140 + }, + { + "epoch": 0.055594917926365575, + "grad_norm": 5.337151988206336, + "learning_rate": 1e-05, + "loss": 1.3027, + "step": 2141 + }, + { + "epoch": 0.055620884725957526, + "grad_norm": 5.174014107229659, + "learning_rate": 1e-05, + "loss": 1.3132, + "step": 2142 + }, + { + "epoch": 0.055646851525549476, + "grad_norm": 5.529698903335321, + "learning_rate": 1e-05, + "loss": 1.2826, + "step": 2143 + }, + { + "epoch": 0.05567281832514143, + "grad_norm": 5.778593205514163, + "learning_rate": 1e-05, + "loss": 1.2556, + "step": 2144 + }, + { + "epoch": 0.05569878512473338, + "grad_norm": 4.094997303348161, + "learning_rate": 1e-05, + "loss": 1.2601, + "step": 2145 + }, + { + "epoch": 0.05572475192432533, + "grad_norm": 5.944040330280242, + "learning_rate": 1e-05, + "loss": 1.2366, + "step": 2146 + }, + { + "epoch": 0.05575071872391728, + "grad_norm": 4.008053302037369, + "learning_rate": 1e-05, + "loss": 1.278, + "step": 2147 + }, + { + "epoch": 0.05577668552350923, + "grad_norm": 10.329469717323986, + "learning_rate": 1e-05, + "loss": 1.2944, + "step": 2148 + }, + { + "epoch": 0.05580265232310118, + "grad_norm": 7.471299251134303, + "learning_rate": 1e-05, + "loss": 1.2926, + "step": 2149 + }, + { + "epoch": 0.05582861912269313, + "grad_norm": 7.075287711288479, + "learning_rate": 1e-05, + "loss": 1.2304, + "step": 2150 + }, + { + "epoch": 0.05585458592228508, + "grad_norm": 6.370971603672762, + "learning_rate": 1e-05, + "loss": 1.3261, + "step": 2151 + }, + { + "epoch": 0.05588055272187703, + "grad_norm": 5.365230997735104, + "learning_rate": 1e-05, + "loss": 1.2352, + "step": 2152 + }, + { + "epoch": 0.05590651952146898, + "grad_norm": 8.354356974450097, + "learning_rate": 1e-05, + "loss": 1.259, + "step": 2153 + }, + { + "epoch": 0.05593248632106093, + "grad_norm": 7.047410086132291, + "learning_rate": 1e-05, + "loss": 1.3208, + "step": 2154 + }, + { + "epoch": 0.05595845312065288, + "grad_norm": 3.594881360639084, + "learning_rate": 1e-05, + "loss": 1.2432, + "step": 2155 + }, + { + "epoch": 0.05598441992024483, + "grad_norm": 5.987167941703173, + "learning_rate": 1e-05, + "loss": 1.264, + "step": 2156 + }, + { + "epoch": 0.05601038671983678, + "grad_norm": 5.1282182218922445, + "learning_rate": 1e-05, + "loss": 1.2351, + "step": 2157 + }, + { + "epoch": 0.05603635351942873, + "grad_norm": 4.871282089006325, + "learning_rate": 1e-05, + "loss": 1.2649, + "step": 2158 + }, + { + "epoch": 0.056062320319020684, + "grad_norm": 4.268168937936228, + "learning_rate": 1e-05, + "loss": 1.2908, + "step": 2159 + }, + { + "epoch": 0.056088287118612634, + "grad_norm": 8.865644655148726, + "learning_rate": 1e-05, + "loss": 1.2912, + "step": 2160 + }, + { + "epoch": 0.056114253918204585, + "grad_norm": 3.755803766895284, + "learning_rate": 1e-05, + "loss": 1.2846, + "step": 2161 + }, + { + "epoch": 0.05614022071779653, + "grad_norm": 5.348464938112044, + "learning_rate": 1e-05, + "loss": 1.2557, + "step": 2162 + }, + { + "epoch": 0.05616618751738848, + "grad_norm": 4.541330844234099, + "learning_rate": 1e-05, + "loss": 1.305, + "step": 2163 + }, + { + "epoch": 0.05619215431698043, + "grad_norm": 4.779272530715481, + "learning_rate": 1e-05, + "loss": 1.2729, + "step": 2164 + }, + { + "epoch": 0.05621812111657238, + "grad_norm": 5.777653745803437, + "learning_rate": 1e-05, + "loss": 1.3027, + "step": 2165 + }, + { + "epoch": 0.05624408791616433, + "grad_norm": 3.5768387660714707, + "learning_rate": 1e-05, + "loss": 1.2992, + "step": 2166 + }, + { + "epoch": 0.05627005471575628, + "grad_norm": 18.361585638185975, + "learning_rate": 1e-05, + "loss": 1.289, + "step": 2167 + }, + { + "epoch": 0.05629602151534823, + "grad_norm": 3.4707422807037895, + "learning_rate": 1e-05, + "loss": 1.2707, + "step": 2168 + }, + { + "epoch": 0.05632198831494018, + "grad_norm": 7.198915919176787, + "learning_rate": 1e-05, + "loss": 1.2865, + "step": 2169 + }, + { + "epoch": 0.05634795511453213, + "grad_norm": 4.832790848553686, + "learning_rate": 1e-05, + "loss": 1.2972, + "step": 2170 + }, + { + "epoch": 0.05637392191412408, + "grad_norm": 5.55848222997319, + "learning_rate": 1e-05, + "loss": 1.2218, + "step": 2171 + }, + { + "epoch": 0.05639988871371603, + "grad_norm": 4.259008397133711, + "learning_rate": 1e-05, + "loss": 1.2774, + "step": 2172 + }, + { + "epoch": 0.05642585551330798, + "grad_norm": 8.800960696406245, + "learning_rate": 1e-05, + "loss": 1.2735, + "step": 2173 + }, + { + "epoch": 0.056451822312899934, + "grad_norm": 5.55142549645806, + "learning_rate": 1e-05, + "loss": 1.2745, + "step": 2174 + }, + { + "epoch": 0.056477789112491884, + "grad_norm": 5.666962484978381, + "learning_rate": 1e-05, + "loss": 1.2785, + "step": 2175 + }, + { + "epoch": 0.056503755912083835, + "grad_norm": 6.940272421643194, + "learning_rate": 1e-05, + "loss": 1.2657, + "step": 2176 + }, + { + "epoch": 0.056529722711675785, + "grad_norm": 5.993852645106484, + "learning_rate": 1e-05, + "loss": 1.3195, + "step": 2177 + }, + { + "epoch": 0.056555689511267736, + "grad_norm": 11.020369656380916, + "learning_rate": 1e-05, + "loss": 1.316, + "step": 2178 + }, + { + "epoch": 0.056581656310859686, + "grad_norm": 12.137440385289038, + "learning_rate": 1e-05, + "loss": 1.2974, + "step": 2179 + }, + { + "epoch": 0.05660762311045164, + "grad_norm": 8.669186396983545, + "learning_rate": 1e-05, + "loss": 1.2825, + "step": 2180 + }, + { + "epoch": 0.05663358991004359, + "grad_norm": 4.489921090976794, + "learning_rate": 1e-05, + "loss": 1.2374, + "step": 2181 + }, + { + "epoch": 0.05665955670963554, + "grad_norm": 4.422105762943952, + "learning_rate": 1e-05, + "loss": 1.3015, + "step": 2182 + }, + { + "epoch": 0.05668552350922749, + "grad_norm": 4.772096136398991, + "learning_rate": 1e-05, + "loss": 1.2127, + "step": 2183 + }, + { + "epoch": 0.05671149030881944, + "grad_norm": 4.87855576171785, + "learning_rate": 1e-05, + "loss": 1.2883, + "step": 2184 + }, + { + "epoch": 0.05673745710841139, + "grad_norm": 3.3675232474268757, + "learning_rate": 1e-05, + "loss": 1.2552, + "step": 2185 + }, + { + "epoch": 0.05676342390800334, + "grad_norm": 3.7218685377498866, + "learning_rate": 1e-05, + "loss": 1.2536, + "step": 2186 + }, + { + "epoch": 0.05678939070759529, + "grad_norm": 7.93535242539576, + "learning_rate": 1e-05, + "loss": 1.2875, + "step": 2187 + }, + { + "epoch": 0.05681535750718724, + "grad_norm": 6.73312245910591, + "learning_rate": 1e-05, + "loss": 1.2614, + "step": 2188 + }, + { + "epoch": 0.05684132430677919, + "grad_norm": 5.113366111010388, + "learning_rate": 1e-05, + "loss": 1.2775, + "step": 2189 + }, + { + "epoch": 0.05686729110637114, + "grad_norm": 5.067644119509288, + "learning_rate": 1e-05, + "loss": 1.2602, + "step": 2190 + }, + { + "epoch": 0.05689325790596309, + "grad_norm": 4.2003030485993795, + "learning_rate": 1e-05, + "loss": 1.3285, + "step": 2191 + }, + { + "epoch": 0.05691922470555504, + "grad_norm": 6.851777128818337, + "learning_rate": 1e-05, + "loss": 1.3004, + "step": 2192 + }, + { + "epoch": 0.05694519150514699, + "grad_norm": 6.649632849632535, + "learning_rate": 1e-05, + "loss": 1.2637, + "step": 2193 + }, + { + "epoch": 0.05697115830473894, + "grad_norm": 4.188660190476685, + "learning_rate": 1e-05, + "loss": 1.3015, + "step": 2194 + }, + { + "epoch": 0.056997125104330894, + "grad_norm": 6.152378472124242, + "learning_rate": 1e-05, + "loss": 1.3007, + "step": 2195 + }, + { + "epoch": 0.057023091903922844, + "grad_norm": 5.104731605042667, + "learning_rate": 1e-05, + "loss": 1.277, + "step": 2196 + }, + { + "epoch": 0.057049058703514795, + "grad_norm": 5.544410179421333, + "learning_rate": 1e-05, + "loss": 1.2874, + "step": 2197 + }, + { + "epoch": 0.057075025503106745, + "grad_norm": 8.906708715571817, + "learning_rate": 1e-05, + "loss": 1.2889, + "step": 2198 + }, + { + "epoch": 0.057100992302698696, + "grad_norm": 4.571346746291415, + "learning_rate": 1e-05, + "loss": 1.2735, + "step": 2199 + }, + { + "epoch": 0.05712695910229064, + "grad_norm": 7.231198365234028, + "learning_rate": 1e-05, + "loss": 1.2703, + "step": 2200 + }, + { + "epoch": 0.05715292590188259, + "grad_norm": 9.609493298306385, + "learning_rate": 1e-05, + "loss": 1.3214, + "step": 2201 + }, + { + "epoch": 0.05717889270147454, + "grad_norm": 5.203608493278328, + "learning_rate": 1e-05, + "loss": 1.2238, + "step": 2202 + }, + { + "epoch": 0.05720485950106649, + "grad_norm": 7.999355767059512, + "learning_rate": 1e-05, + "loss": 1.2675, + "step": 2203 + }, + { + "epoch": 0.05723082630065844, + "grad_norm": 3.4900898276412433, + "learning_rate": 1e-05, + "loss": 1.2546, + "step": 2204 + }, + { + "epoch": 0.05725679310025039, + "grad_norm": 10.636050233766072, + "learning_rate": 1e-05, + "loss": 1.2796, + "step": 2205 + }, + { + "epoch": 0.05728275989984234, + "grad_norm": 4.9339260252195585, + "learning_rate": 1e-05, + "loss": 1.2555, + "step": 2206 + }, + { + "epoch": 0.05730872669943429, + "grad_norm": 3.7491404501959393, + "learning_rate": 1e-05, + "loss": 1.2375, + "step": 2207 + }, + { + "epoch": 0.05733469349902624, + "grad_norm": 4.585608501428043, + "learning_rate": 1e-05, + "loss": 1.3062, + "step": 2208 + }, + { + "epoch": 0.057360660298618193, + "grad_norm": 9.671027372868787, + "learning_rate": 1e-05, + "loss": 1.2685, + "step": 2209 + }, + { + "epoch": 0.057386627098210144, + "grad_norm": 4.139463326503779, + "learning_rate": 1e-05, + "loss": 1.2923, + "step": 2210 + }, + { + "epoch": 0.057412593897802094, + "grad_norm": 6.2472048803555875, + "learning_rate": 1e-05, + "loss": 1.2659, + "step": 2211 + }, + { + "epoch": 0.057438560697394045, + "grad_norm": 5.961385767569097, + "learning_rate": 1e-05, + "loss": 1.3522, + "step": 2212 + }, + { + "epoch": 0.057464527496985995, + "grad_norm": 6.200505316046137, + "learning_rate": 1e-05, + "loss": 1.2501, + "step": 2213 + }, + { + "epoch": 0.057490494296577946, + "grad_norm": 4.782215694066933, + "learning_rate": 1e-05, + "loss": 1.268, + "step": 2214 + }, + { + "epoch": 0.057516461096169896, + "grad_norm": 15.263829198845336, + "learning_rate": 1e-05, + "loss": 1.2844, + "step": 2215 + }, + { + "epoch": 0.05754242789576185, + "grad_norm": 4.446217959258784, + "learning_rate": 1e-05, + "loss": 1.3012, + "step": 2216 + }, + { + "epoch": 0.0575683946953538, + "grad_norm": 8.893557636381553, + "learning_rate": 1e-05, + "loss": 1.2937, + "step": 2217 + }, + { + "epoch": 0.05759436149494575, + "grad_norm": 4.722301031215497, + "learning_rate": 1e-05, + "loss": 1.3221, + "step": 2218 + }, + { + "epoch": 0.0576203282945377, + "grad_norm": 69.40036839104314, + "learning_rate": 1e-05, + "loss": 1.2627, + "step": 2219 + }, + { + "epoch": 0.05764629509412965, + "grad_norm": 11.842111761080202, + "learning_rate": 1e-05, + "loss": 1.3275, + "step": 2220 + }, + { + "epoch": 0.0576722618937216, + "grad_norm": 5.6307592789945655, + "learning_rate": 1e-05, + "loss": 1.2236, + "step": 2221 + }, + { + "epoch": 0.05769822869331355, + "grad_norm": 4.808233772891092, + "learning_rate": 1e-05, + "loss": 1.2675, + "step": 2222 + }, + { + "epoch": 0.0577241954929055, + "grad_norm": 5.280035477432357, + "learning_rate": 1e-05, + "loss": 1.2851, + "step": 2223 + }, + { + "epoch": 0.05775016229249745, + "grad_norm": 4.034780922464708, + "learning_rate": 1e-05, + "loss": 1.3052, + "step": 2224 + }, + { + "epoch": 0.0577761290920894, + "grad_norm": 22.051805708598234, + "learning_rate": 1e-05, + "loss": 1.238, + "step": 2225 + }, + { + "epoch": 0.05780209589168135, + "grad_norm": 5.450711741037744, + "learning_rate": 1e-05, + "loss": 1.305, + "step": 2226 + }, + { + "epoch": 0.0578280626912733, + "grad_norm": 4.3696083633599825, + "learning_rate": 1e-05, + "loss": 1.2256, + "step": 2227 + }, + { + "epoch": 0.05785402949086525, + "grad_norm": 5.9180728512914875, + "learning_rate": 1e-05, + "loss": 1.3062, + "step": 2228 + }, + { + "epoch": 0.0578799962904572, + "grad_norm": 4.355681355166953, + "learning_rate": 1e-05, + "loss": 1.2549, + "step": 2229 + }, + { + "epoch": 0.05790596309004915, + "grad_norm": 5.2809939350598265, + "learning_rate": 1e-05, + "loss": 1.3111, + "step": 2230 + }, + { + "epoch": 0.057931929889641104, + "grad_norm": 4.68576851972901, + "learning_rate": 1e-05, + "loss": 1.2601, + "step": 2231 + }, + { + "epoch": 0.057957896689233054, + "grad_norm": 5.849538944692865, + "learning_rate": 1e-05, + "loss": 1.2685, + "step": 2232 + }, + { + "epoch": 0.057983863488825005, + "grad_norm": 4.7136321956589, + "learning_rate": 1e-05, + "loss": 1.236, + "step": 2233 + }, + { + "epoch": 0.058009830288416955, + "grad_norm": 6.7464675137122105, + "learning_rate": 1e-05, + "loss": 1.2762, + "step": 2234 + }, + { + "epoch": 0.058035797088008906, + "grad_norm": 7.565072623015849, + "learning_rate": 1e-05, + "loss": 1.2586, + "step": 2235 + }, + { + "epoch": 0.058061763887600856, + "grad_norm": 4.461146361477116, + "learning_rate": 1e-05, + "loss": 1.2456, + "step": 2236 + }, + { + "epoch": 0.05808773068719281, + "grad_norm": 7.212871056702403, + "learning_rate": 1e-05, + "loss": 1.3032, + "step": 2237 + }, + { + "epoch": 0.05811369748678476, + "grad_norm": 5.0855455848181785, + "learning_rate": 1e-05, + "loss": 1.2937, + "step": 2238 + }, + { + "epoch": 0.0581396642863767, + "grad_norm": 5.552641458233755, + "learning_rate": 1e-05, + "loss": 1.1913, + "step": 2239 + }, + { + "epoch": 0.05816563108596865, + "grad_norm": 4.985394890696877, + "learning_rate": 1e-05, + "loss": 1.2628, + "step": 2240 + }, + { + "epoch": 0.0581915978855606, + "grad_norm": 7.772797064541372, + "learning_rate": 1e-05, + "loss": 1.2624, + "step": 2241 + }, + { + "epoch": 0.05821756468515255, + "grad_norm": 6.325479590285557, + "learning_rate": 1e-05, + "loss": 1.3003, + "step": 2242 + }, + { + "epoch": 0.0582435314847445, + "grad_norm": 4.493274112805452, + "learning_rate": 1e-05, + "loss": 1.254, + "step": 2243 + }, + { + "epoch": 0.05826949828433645, + "grad_norm": 7.631556959172469, + "learning_rate": 1e-05, + "loss": 1.257, + "step": 2244 + }, + { + "epoch": 0.058295465083928404, + "grad_norm": 4.900515175031385, + "learning_rate": 1e-05, + "loss": 1.2488, + "step": 2245 + }, + { + "epoch": 0.058321431883520354, + "grad_norm": 5.470481903495315, + "learning_rate": 1e-05, + "loss": 1.2562, + "step": 2246 + }, + { + "epoch": 0.058347398683112305, + "grad_norm": 9.602108771456754, + "learning_rate": 1e-05, + "loss": 1.2884, + "step": 2247 + }, + { + "epoch": 0.058373365482704255, + "grad_norm": 5.73692925965819, + "learning_rate": 1e-05, + "loss": 1.2438, + "step": 2248 + }, + { + "epoch": 0.058399332282296205, + "grad_norm": 4.675415942614111, + "learning_rate": 1e-05, + "loss": 1.2639, + "step": 2249 + }, + { + "epoch": 0.058425299081888156, + "grad_norm": 6.000421191372898, + "learning_rate": 1e-05, + "loss": 1.2668, + "step": 2250 + }, + { + "epoch": 0.058451265881480106, + "grad_norm": 8.864863234754297, + "learning_rate": 1e-05, + "loss": 1.2745, + "step": 2251 + }, + { + "epoch": 0.05847723268107206, + "grad_norm": 5.2239365074890936, + "learning_rate": 1e-05, + "loss": 1.2346, + "step": 2252 + }, + { + "epoch": 0.05850319948066401, + "grad_norm": 15.63008511003299, + "learning_rate": 1e-05, + "loss": 1.3045, + "step": 2253 + }, + { + "epoch": 0.05852916628025596, + "grad_norm": 5.586878301182305, + "learning_rate": 1e-05, + "loss": 1.3128, + "step": 2254 + }, + { + "epoch": 0.05855513307984791, + "grad_norm": 4.378777970843502, + "learning_rate": 1e-05, + "loss": 1.2557, + "step": 2255 + }, + { + "epoch": 0.05858109987943986, + "grad_norm": 4.814717364072209, + "learning_rate": 1e-05, + "loss": 1.2226, + "step": 2256 + }, + { + "epoch": 0.05860706667903181, + "grad_norm": 5.569631569695033, + "learning_rate": 1e-05, + "loss": 1.3007, + "step": 2257 + }, + { + "epoch": 0.05863303347862376, + "grad_norm": 4.385526146809592, + "learning_rate": 1e-05, + "loss": 1.2576, + "step": 2258 + }, + { + "epoch": 0.05865900027821571, + "grad_norm": 6.980871768496637, + "learning_rate": 1e-05, + "loss": 1.3105, + "step": 2259 + }, + { + "epoch": 0.05868496707780766, + "grad_norm": 9.49819848898373, + "learning_rate": 1e-05, + "loss": 1.2507, + "step": 2260 + }, + { + "epoch": 0.05871093387739961, + "grad_norm": 4.672287530451429, + "learning_rate": 1e-05, + "loss": 1.301, + "step": 2261 + }, + { + "epoch": 0.05873690067699156, + "grad_norm": 4.359510850327818, + "learning_rate": 1e-05, + "loss": 1.2886, + "step": 2262 + }, + { + "epoch": 0.05876286747658351, + "grad_norm": 3.3740703573980735, + "learning_rate": 1e-05, + "loss": 1.23, + "step": 2263 + }, + { + "epoch": 0.05878883427617546, + "grad_norm": 4.869521559569915, + "learning_rate": 1e-05, + "loss": 1.2585, + "step": 2264 + }, + { + "epoch": 0.05881480107576741, + "grad_norm": 7.054705073777253, + "learning_rate": 1e-05, + "loss": 1.2642, + "step": 2265 + }, + { + "epoch": 0.058840767875359364, + "grad_norm": 4.815428202158309, + "learning_rate": 1e-05, + "loss": 1.3374, + "step": 2266 + }, + { + "epoch": 0.058866734674951314, + "grad_norm": 5.0615052552791795, + "learning_rate": 1e-05, + "loss": 1.3182, + "step": 2267 + }, + { + "epoch": 0.058892701474543264, + "grad_norm": 9.374995117186229, + "learning_rate": 1e-05, + "loss": 1.271, + "step": 2268 + }, + { + "epoch": 0.058918668274135215, + "grad_norm": 5.995192190975889, + "learning_rate": 1e-05, + "loss": 1.2276, + "step": 2269 + }, + { + "epoch": 0.058944635073727165, + "grad_norm": 4.8726258977104004, + "learning_rate": 1e-05, + "loss": 1.2936, + "step": 2270 + }, + { + "epoch": 0.058970601873319116, + "grad_norm": 7.894905968577811, + "learning_rate": 1e-05, + "loss": 1.3267, + "step": 2271 + }, + { + "epoch": 0.058996568672911066, + "grad_norm": 6.047839499347125, + "learning_rate": 1e-05, + "loss": 1.312, + "step": 2272 + }, + { + "epoch": 0.05902253547250302, + "grad_norm": 4.055405981246452, + "learning_rate": 1e-05, + "loss": 1.2886, + "step": 2273 + }, + { + "epoch": 0.05904850227209497, + "grad_norm": 7.335206081665194, + "learning_rate": 1e-05, + "loss": 1.2233, + "step": 2274 + }, + { + "epoch": 0.05907446907168692, + "grad_norm": 6.307132365081162, + "learning_rate": 1e-05, + "loss": 1.272, + "step": 2275 + }, + { + "epoch": 0.05910043587127887, + "grad_norm": 5.223199831764427, + "learning_rate": 1e-05, + "loss": 1.2711, + "step": 2276 + }, + { + "epoch": 0.05912640267087081, + "grad_norm": 6.519824387695325, + "learning_rate": 1e-05, + "loss": 1.2662, + "step": 2277 + }, + { + "epoch": 0.05915236947046276, + "grad_norm": 4.013487725778614, + "learning_rate": 1e-05, + "loss": 1.2456, + "step": 2278 + }, + { + "epoch": 0.05917833627005471, + "grad_norm": 5.140042257469226, + "learning_rate": 1e-05, + "loss": 1.2603, + "step": 2279 + }, + { + "epoch": 0.05920430306964666, + "grad_norm": 5.604289605400596, + "learning_rate": 1e-05, + "loss": 1.3014, + "step": 2280 + }, + { + "epoch": 0.059230269869238614, + "grad_norm": 5.87634424292663, + "learning_rate": 1e-05, + "loss": 1.2502, + "step": 2281 + }, + { + "epoch": 0.059256236668830564, + "grad_norm": 7.078107484370574, + "learning_rate": 1e-05, + "loss": 1.2847, + "step": 2282 + }, + { + "epoch": 0.059282203468422515, + "grad_norm": 5.8454509034503435, + "learning_rate": 1e-05, + "loss": 1.2851, + "step": 2283 + }, + { + "epoch": 0.059308170268014465, + "grad_norm": 3.5700297504542475, + "learning_rate": 1e-05, + "loss": 1.2849, + "step": 2284 + }, + { + "epoch": 0.059334137067606416, + "grad_norm": 3.975838406032593, + "learning_rate": 1e-05, + "loss": 1.2671, + "step": 2285 + }, + { + "epoch": 0.059360103867198366, + "grad_norm": 5.211100013863885, + "learning_rate": 1e-05, + "loss": 1.2815, + "step": 2286 + }, + { + "epoch": 0.05938607066679032, + "grad_norm": 6.059373501663849, + "learning_rate": 1e-05, + "loss": 1.274, + "step": 2287 + }, + { + "epoch": 0.05941203746638227, + "grad_norm": 5.482858127459584, + "learning_rate": 1e-05, + "loss": 1.2821, + "step": 2288 + }, + { + "epoch": 0.05943800426597422, + "grad_norm": 4.306768838440112, + "learning_rate": 1e-05, + "loss": 1.2563, + "step": 2289 + }, + { + "epoch": 0.05946397106556617, + "grad_norm": 8.027105189581695, + "learning_rate": 1e-05, + "loss": 1.276, + "step": 2290 + }, + { + "epoch": 0.05948993786515812, + "grad_norm": 4.690850649799628, + "learning_rate": 1e-05, + "loss": 1.2661, + "step": 2291 + }, + { + "epoch": 0.05951590466475007, + "grad_norm": 5.879866512106828, + "learning_rate": 1e-05, + "loss": 1.2925, + "step": 2292 + }, + { + "epoch": 0.05954187146434202, + "grad_norm": 4.18191565901341, + "learning_rate": 1e-05, + "loss": 1.2901, + "step": 2293 + }, + { + "epoch": 0.05956783826393397, + "grad_norm": 4.967081617562996, + "learning_rate": 1e-05, + "loss": 1.2466, + "step": 2294 + }, + { + "epoch": 0.05959380506352592, + "grad_norm": 3.703019667789273, + "learning_rate": 1e-05, + "loss": 1.2684, + "step": 2295 + }, + { + "epoch": 0.05961977186311787, + "grad_norm": 5.5009622599061805, + "learning_rate": 1e-05, + "loss": 1.2646, + "step": 2296 + }, + { + "epoch": 0.05964573866270982, + "grad_norm": 4.558686635461142, + "learning_rate": 1e-05, + "loss": 1.2973, + "step": 2297 + }, + { + "epoch": 0.05967170546230177, + "grad_norm": 5.982609660271007, + "learning_rate": 1e-05, + "loss": 1.2571, + "step": 2298 + }, + { + "epoch": 0.05969767226189372, + "grad_norm": 4.908390040088963, + "learning_rate": 1e-05, + "loss": 1.2843, + "step": 2299 + }, + { + "epoch": 0.05972363906148567, + "grad_norm": 4.481130669147532, + "learning_rate": 1e-05, + "loss": 1.2808, + "step": 2300 + }, + { + "epoch": 0.05974960586107762, + "grad_norm": 3.5730724393439486, + "learning_rate": 1e-05, + "loss": 1.2727, + "step": 2301 + }, + { + "epoch": 0.059775572660669574, + "grad_norm": 4.033735351914181, + "learning_rate": 1e-05, + "loss": 1.2149, + "step": 2302 + }, + { + "epoch": 0.059801539460261524, + "grad_norm": 5.6942849403103555, + "learning_rate": 1e-05, + "loss": 1.2904, + "step": 2303 + }, + { + "epoch": 0.059827506259853475, + "grad_norm": 5.170881193166477, + "learning_rate": 1e-05, + "loss": 1.2998, + "step": 2304 + }, + { + "epoch": 0.059853473059445425, + "grad_norm": 4.312084233241544, + "learning_rate": 1e-05, + "loss": 1.2832, + "step": 2305 + }, + { + "epoch": 0.059879439859037376, + "grad_norm": 5.0482606668532926, + "learning_rate": 1e-05, + "loss": 1.2782, + "step": 2306 + }, + { + "epoch": 0.059905406658629326, + "grad_norm": 5.90229786237126, + "learning_rate": 1e-05, + "loss": 1.2849, + "step": 2307 + }, + { + "epoch": 0.059931373458221276, + "grad_norm": 5.5927958607078425, + "learning_rate": 1e-05, + "loss": 1.2635, + "step": 2308 + }, + { + "epoch": 0.05995734025781323, + "grad_norm": 4.610571340709148, + "learning_rate": 1e-05, + "loss": 1.3089, + "step": 2309 + }, + { + "epoch": 0.05998330705740518, + "grad_norm": 4.61690142111566, + "learning_rate": 1e-05, + "loss": 1.2388, + "step": 2310 + }, + { + "epoch": 0.06000927385699713, + "grad_norm": 3.5911934425051095, + "learning_rate": 1e-05, + "loss": 1.2509, + "step": 2311 + }, + { + "epoch": 0.06003524065658908, + "grad_norm": 3.837104076128251, + "learning_rate": 1e-05, + "loss": 1.2507, + "step": 2312 + }, + { + "epoch": 0.06006120745618103, + "grad_norm": 7.398707416126371, + "learning_rate": 1e-05, + "loss": 1.3222, + "step": 2313 + }, + { + "epoch": 0.06008717425577298, + "grad_norm": 8.422570251189981, + "learning_rate": 1e-05, + "loss": 1.2872, + "step": 2314 + }, + { + "epoch": 0.06011314105536492, + "grad_norm": 5.58957520592272, + "learning_rate": 1e-05, + "loss": 1.2286, + "step": 2315 + }, + { + "epoch": 0.06013910785495687, + "grad_norm": 8.196603620591027, + "learning_rate": 1e-05, + "loss": 1.2503, + "step": 2316 + }, + { + "epoch": 0.060165074654548824, + "grad_norm": 3.792909662982741, + "learning_rate": 1e-05, + "loss": 1.2227, + "step": 2317 + }, + { + "epoch": 0.060191041454140774, + "grad_norm": 10.296830172868091, + "learning_rate": 1e-05, + "loss": 1.318, + "step": 2318 + }, + { + "epoch": 0.060217008253732725, + "grad_norm": 3.391007924760509, + "learning_rate": 1e-05, + "loss": 1.2827, + "step": 2319 + }, + { + "epoch": 0.060242975053324675, + "grad_norm": 5.1377492356928345, + "learning_rate": 1e-05, + "loss": 1.2667, + "step": 2320 + }, + { + "epoch": 0.060268941852916626, + "grad_norm": 5.820544838588359, + "learning_rate": 1e-05, + "loss": 1.3185, + "step": 2321 + }, + { + "epoch": 0.060294908652508576, + "grad_norm": 4.2896003316450155, + "learning_rate": 1e-05, + "loss": 1.2375, + "step": 2322 + }, + { + "epoch": 0.06032087545210053, + "grad_norm": 5.153583923965199, + "learning_rate": 1e-05, + "loss": 1.2615, + "step": 2323 + }, + { + "epoch": 0.06034684225169248, + "grad_norm": 4.386900492951555, + "learning_rate": 1e-05, + "loss": 1.318, + "step": 2324 + }, + { + "epoch": 0.06037280905128443, + "grad_norm": 4.069480415564269, + "learning_rate": 1e-05, + "loss": 1.2687, + "step": 2325 + }, + { + "epoch": 0.06039877585087638, + "grad_norm": 17.40857115358337, + "learning_rate": 1e-05, + "loss": 1.2491, + "step": 2326 + }, + { + "epoch": 0.06042474265046833, + "grad_norm": 5.829656077688534, + "learning_rate": 1e-05, + "loss": 1.2656, + "step": 2327 + }, + { + "epoch": 0.06045070945006028, + "grad_norm": 4.522283060871091, + "learning_rate": 1e-05, + "loss": 1.3002, + "step": 2328 + }, + { + "epoch": 0.06047667624965223, + "grad_norm": 5.980114089745043, + "learning_rate": 1e-05, + "loss": 1.2239, + "step": 2329 + }, + { + "epoch": 0.06050264304924418, + "grad_norm": 4.810146796827644, + "learning_rate": 1e-05, + "loss": 1.2547, + "step": 2330 + }, + { + "epoch": 0.06052860984883613, + "grad_norm": 11.007556054056828, + "learning_rate": 1e-05, + "loss": 1.3219, + "step": 2331 + }, + { + "epoch": 0.06055457664842808, + "grad_norm": 6.996604095707453, + "learning_rate": 1e-05, + "loss": 1.2398, + "step": 2332 + }, + { + "epoch": 0.06058054344802003, + "grad_norm": 5.058995670567847, + "learning_rate": 1e-05, + "loss": 1.2517, + "step": 2333 + }, + { + "epoch": 0.06060651024761198, + "grad_norm": 3.9608783905614793, + "learning_rate": 1e-05, + "loss": 1.2239, + "step": 2334 + }, + { + "epoch": 0.06063247704720393, + "grad_norm": 5.562423191183183, + "learning_rate": 1e-05, + "loss": 1.2859, + "step": 2335 + }, + { + "epoch": 0.06065844384679588, + "grad_norm": 4.2421016227783275, + "learning_rate": 1e-05, + "loss": 1.2571, + "step": 2336 + }, + { + "epoch": 0.06068441064638783, + "grad_norm": 5.233945914374233, + "learning_rate": 1e-05, + "loss": 1.3211, + "step": 2337 + }, + { + "epoch": 0.060710377445979784, + "grad_norm": 4.192722764395131, + "learning_rate": 1e-05, + "loss": 1.257, + "step": 2338 + }, + { + "epoch": 0.060736344245571734, + "grad_norm": 4.360898425078545, + "learning_rate": 1e-05, + "loss": 1.2642, + "step": 2339 + }, + { + "epoch": 0.060762311045163685, + "grad_norm": 8.406717393779047, + "learning_rate": 1e-05, + "loss": 1.2874, + "step": 2340 + }, + { + "epoch": 0.060788277844755635, + "grad_norm": 7.630034988743064, + "learning_rate": 1e-05, + "loss": 1.2741, + "step": 2341 + }, + { + "epoch": 0.060814244644347586, + "grad_norm": 7.679103825257347, + "learning_rate": 1e-05, + "loss": 1.289, + "step": 2342 + }, + { + "epoch": 0.060840211443939536, + "grad_norm": 5.099223186198182, + "learning_rate": 1e-05, + "loss": 1.291, + "step": 2343 + }, + { + "epoch": 0.06086617824353149, + "grad_norm": 6.703955254343968, + "learning_rate": 1e-05, + "loss": 1.2082, + "step": 2344 + }, + { + "epoch": 0.06089214504312344, + "grad_norm": 4.421470165255128, + "learning_rate": 1e-05, + "loss": 1.2677, + "step": 2345 + }, + { + "epoch": 0.06091811184271539, + "grad_norm": 29.27368352912358, + "learning_rate": 1e-05, + "loss": 1.3272, + "step": 2346 + }, + { + "epoch": 0.06094407864230734, + "grad_norm": 13.99062469461972, + "learning_rate": 1e-05, + "loss": 1.2286, + "step": 2347 + }, + { + "epoch": 0.06097004544189929, + "grad_norm": 11.135336884848275, + "learning_rate": 1e-05, + "loss": 1.2831, + "step": 2348 + }, + { + "epoch": 0.06099601224149124, + "grad_norm": 4.593455817735412, + "learning_rate": 1e-05, + "loss": 1.2783, + "step": 2349 + }, + { + "epoch": 0.06102197904108319, + "grad_norm": 5.591833031231344, + "learning_rate": 1e-05, + "loss": 1.2858, + "step": 2350 + }, + { + "epoch": 0.06104794584067514, + "grad_norm": 4.562972579937204, + "learning_rate": 1e-05, + "loss": 1.3029, + "step": 2351 + }, + { + "epoch": 0.06107391264026709, + "grad_norm": 5.7692075934311395, + "learning_rate": 1e-05, + "loss": 1.2867, + "step": 2352 + }, + { + "epoch": 0.061099879439859034, + "grad_norm": 4.496195774594501, + "learning_rate": 1e-05, + "loss": 1.2627, + "step": 2353 + }, + { + "epoch": 0.061125846239450984, + "grad_norm": 4.125158364695042, + "learning_rate": 1e-05, + "loss": 1.232, + "step": 2354 + }, + { + "epoch": 0.061151813039042935, + "grad_norm": 5.99980131456109, + "learning_rate": 1e-05, + "loss": 1.3245, + "step": 2355 + }, + { + "epoch": 0.061177779838634885, + "grad_norm": 5.883929345449037, + "learning_rate": 1e-05, + "loss": 1.2908, + "step": 2356 + }, + { + "epoch": 0.061203746638226836, + "grad_norm": 5.733753305770623, + "learning_rate": 1e-05, + "loss": 1.2483, + "step": 2357 + }, + { + "epoch": 0.061229713437818786, + "grad_norm": 6.471845540376124, + "learning_rate": 1e-05, + "loss": 1.2507, + "step": 2358 + }, + { + "epoch": 0.06125568023741074, + "grad_norm": 8.720831072339674, + "learning_rate": 1e-05, + "loss": 1.2356, + "step": 2359 + }, + { + "epoch": 0.06128164703700269, + "grad_norm": 10.187484647587167, + "learning_rate": 1e-05, + "loss": 1.2384, + "step": 2360 + }, + { + "epoch": 0.06130761383659464, + "grad_norm": 5.269673316573175, + "learning_rate": 1e-05, + "loss": 1.3271, + "step": 2361 + }, + { + "epoch": 0.06133358063618659, + "grad_norm": 7.854279157269766, + "learning_rate": 1e-05, + "loss": 1.2268, + "step": 2362 + }, + { + "epoch": 0.06135954743577854, + "grad_norm": 7.170798360091696, + "learning_rate": 1e-05, + "loss": 1.3302, + "step": 2363 + }, + { + "epoch": 0.06138551423537049, + "grad_norm": 5.243731071065172, + "learning_rate": 1e-05, + "loss": 1.2672, + "step": 2364 + }, + { + "epoch": 0.06141148103496244, + "grad_norm": 6.351582619798238, + "learning_rate": 1e-05, + "loss": 1.2832, + "step": 2365 + }, + { + "epoch": 0.06143744783455439, + "grad_norm": 5.625596078442974, + "learning_rate": 1e-05, + "loss": 1.3066, + "step": 2366 + }, + { + "epoch": 0.06146341463414634, + "grad_norm": 10.092257457253858, + "learning_rate": 1e-05, + "loss": 1.2362, + "step": 2367 + }, + { + "epoch": 0.06148938143373829, + "grad_norm": 4.564177531327587, + "learning_rate": 1e-05, + "loss": 1.2659, + "step": 2368 + }, + { + "epoch": 0.06151534823333024, + "grad_norm": 4.757610849200641, + "learning_rate": 1e-05, + "loss": 1.2618, + "step": 2369 + }, + { + "epoch": 0.06154131503292219, + "grad_norm": 4.820485611192089, + "learning_rate": 1e-05, + "loss": 1.231, + "step": 2370 + }, + { + "epoch": 0.06156728183251414, + "grad_norm": 13.483105861661022, + "learning_rate": 1e-05, + "loss": 1.2476, + "step": 2371 + }, + { + "epoch": 0.06159324863210609, + "grad_norm": 6.318840845739752, + "learning_rate": 1e-05, + "loss": 1.267, + "step": 2372 + }, + { + "epoch": 0.06161921543169804, + "grad_norm": 4.479383501822152, + "learning_rate": 1e-05, + "loss": 1.2457, + "step": 2373 + }, + { + "epoch": 0.061645182231289994, + "grad_norm": 5.766277658796017, + "learning_rate": 1e-05, + "loss": 1.2576, + "step": 2374 + }, + { + "epoch": 0.061671149030881944, + "grad_norm": 17.622583534755996, + "learning_rate": 1e-05, + "loss": 1.2938, + "step": 2375 + }, + { + "epoch": 0.061697115830473895, + "grad_norm": 5.87116392173634, + "learning_rate": 1e-05, + "loss": 1.2926, + "step": 2376 + }, + { + "epoch": 0.061723082630065845, + "grad_norm": 4.11004633188237, + "learning_rate": 1e-05, + "loss": 1.2301, + "step": 2377 + }, + { + "epoch": 0.061749049429657796, + "grad_norm": 4.531143292453149, + "learning_rate": 1e-05, + "loss": 1.2525, + "step": 2378 + }, + { + "epoch": 0.061775016229249746, + "grad_norm": 5.552112094978807, + "learning_rate": 1e-05, + "loss": 1.3019, + "step": 2379 + }, + { + "epoch": 0.0618009830288417, + "grad_norm": 5.113914036773858, + "learning_rate": 1e-05, + "loss": 1.2293, + "step": 2380 + }, + { + "epoch": 0.06182694982843365, + "grad_norm": 8.645997955103184, + "learning_rate": 1e-05, + "loss": 1.2842, + "step": 2381 + }, + { + "epoch": 0.0618529166280256, + "grad_norm": 4.93911504286104, + "learning_rate": 1e-05, + "loss": 1.2578, + "step": 2382 + }, + { + "epoch": 0.06187888342761755, + "grad_norm": 27.218570607904816, + "learning_rate": 1e-05, + "loss": 1.2295, + "step": 2383 + }, + { + "epoch": 0.0619048502272095, + "grad_norm": 4.052791556449013, + "learning_rate": 1e-05, + "loss": 1.2902, + "step": 2384 + }, + { + "epoch": 0.06193081702680145, + "grad_norm": 4.377808023941025, + "learning_rate": 1e-05, + "loss": 1.2656, + "step": 2385 + }, + { + "epoch": 0.0619567838263934, + "grad_norm": 6.40507116869324, + "learning_rate": 1e-05, + "loss": 1.2725, + "step": 2386 + }, + { + "epoch": 0.06198275062598535, + "grad_norm": 5.835721135756609, + "learning_rate": 1e-05, + "loss": 1.2517, + "step": 2387 + }, + { + "epoch": 0.0620087174255773, + "grad_norm": 6.206867770424246, + "learning_rate": 1e-05, + "loss": 1.2628, + "step": 2388 + }, + { + "epoch": 0.06203468422516925, + "grad_norm": 14.363630226757287, + "learning_rate": 1e-05, + "loss": 1.2454, + "step": 2389 + }, + { + "epoch": 0.0620606510247612, + "grad_norm": 14.298023348690108, + "learning_rate": 1e-05, + "loss": 1.2761, + "step": 2390 + }, + { + "epoch": 0.06208661782435315, + "grad_norm": 5.42564204060076, + "learning_rate": 1e-05, + "loss": 1.2832, + "step": 2391 + }, + { + "epoch": 0.062112584623945095, + "grad_norm": 4.921363828242293, + "learning_rate": 1e-05, + "loss": 1.2937, + "step": 2392 + }, + { + "epoch": 0.062138551423537046, + "grad_norm": 6.756107251885113, + "learning_rate": 1e-05, + "loss": 1.2753, + "step": 2393 + }, + { + "epoch": 0.062164518223128996, + "grad_norm": 7.333335992061248, + "learning_rate": 1e-05, + "loss": 1.2617, + "step": 2394 + }, + { + "epoch": 0.06219048502272095, + "grad_norm": 4.554392888534089, + "learning_rate": 1e-05, + "loss": 1.2748, + "step": 2395 + }, + { + "epoch": 0.0622164518223129, + "grad_norm": 5.8122066300712785, + "learning_rate": 1e-05, + "loss": 1.2613, + "step": 2396 + }, + { + "epoch": 0.06224241862190485, + "grad_norm": 5.566744781372458, + "learning_rate": 1e-05, + "loss": 1.2107, + "step": 2397 + }, + { + "epoch": 0.0622683854214968, + "grad_norm": 6.209860424002757, + "learning_rate": 1e-05, + "loss": 1.2391, + "step": 2398 + }, + { + "epoch": 0.06229435222108875, + "grad_norm": 4.790100084824099, + "learning_rate": 1e-05, + "loss": 1.296, + "step": 2399 + }, + { + "epoch": 0.0623203190206807, + "grad_norm": 5.795231917894031, + "learning_rate": 1e-05, + "loss": 1.266, + "step": 2400 + }, + { + "epoch": 0.06234628582027265, + "grad_norm": 4.638302822490153, + "learning_rate": 1e-05, + "loss": 1.2883, + "step": 2401 + }, + { + "epoch": 0.0623722526198646, + "grad_norm": 8.631182679860565, + "learning_rate": 1e-05, + "loss": 1.3021, + "step": 2402 + }, + { + "epoch": 0.06239821941945655, + "grad_norm": 8.760564040049287, + "learning_rate": 1e-05, + "loss": 1.261, + "step": 2403 + }, + { + "epoch": 0.0624241862190485, + "grad_norm": 4.840639161396429, + "learning_rate": 1e-05, + "loss": 1.2414, + "step": 2404 + }, + { + "epoch": 0.06245015301864045, + "grad_norm": 5.1279569329820385, + "learning_rate": 1e-05, + "loss": 1.2772, + "step": 2405 + }, + { + "epoch": 0.0624761198182324, + "grad_norm": 5.678684655060632, + "learning_rate": 1e-05, + "loss": 1.2692, + "step": 2406 + }, + { + "epoch": 0.06250208661782436, + "grad_norm": 4.825627245675839, + "learning_rate": 1e-05, + "loss": 1.2971, + "step": 2407 + }, + { + "epoch": 0.06252805341741631, + "grad_norm": 5.78503524993734, + "learning_rate": 1e-05, + "loss": 1.2478, + "step": 2408 + }, + { + "epoch": 0.06255402021700826, + "grad_norm": 16.04872722392067, + "learning_rate": 1e-05, + "loss": 1.3423, + "step": 2409 + }, + { + "epoch": 0.06257998701660021, + "grad_norm": 6.0309754936721, + "learning_rate": 1e-05, + "loss": 1.2735, + "step": 2410 + }, + { + "epoch": 0.06260595381619215, + "grad_norm": 7.2252120782157885, + "learning_rate": 1e-05, + "loss": 1.2821, + "step": 2411 + }, + { + "epoch": 0.0626319206157841, + "grad_norm": 5.411781259789478, + "learning_rate": 1e-05, + "loss": 1.2899, + "step": 2412 + }, + { + "epoch": 0.06265788741537605, + "grad_norm": 14.740615704331127, + "learning_rate": 1e-05, + "loss": 1.2833, + "step": 2413 + }, + { + "epoch": 0.062683854214968, + "grad_norm": 4.519632323510584, + "learning_rate": 1e-05, + "loss": 1.235, + "step": 2414 + }, + { + "epoch": 0.06270982101455995, + "grad_norm": 8.766193337468824, + "learning_rate": 1e-05, + "loss": 1.284, + "step": 2415 + }, + { + "epoch": 0.0627357878141519, + "grad_norm": 18.094236588070746, + "learning_rate": 1e-05, + "loss": 1.3056, + "step": 2416 + }, + { + "epoch": 0.06276175461374385, + "grad_norm": 5.119959888301128, + "learning_rate": 1e-05, + "loss": 1.2788, + "step": 2417 + }, + { + "epoch": 0.0627877214133358, + "grad_norm": 9.485931419605611, + "learning_rate": 1e-05, + "loss": 1.2154, + "step": 2418 + }, + { + "epoch": 0.06281368821292775, + "grad_norm": 6.65136768933066, + "learning_rate": 1e-05, + "loss": 1.2265, + "step": 2419 + }, + { + "epoch": 0.0628396550125197, + "grad_norm": 11.583281446015942, + "learning_rate": 1e-05, + "loss": 1.2925, + "step": 2420 + }, + { + "epoch": 0.06286562181211165, + "grad_norm": 5.037199495122474, + "learning_rate": 1e-05, + "loss": 1.3161, + "step": 2421 + }, + { + "epoch": 0.0628915886117036, + "grad_norm": 5.7831724527428765, + "learning_rate": 1e-05, + "loss": 1.2741, + "step": 2422 + }, + { + "epoch": 0.06291755541129555, + "grad_norm": 6.232944869450192, + "learning_rate": 1e-05, + "loss": 1.196, + "step": 2423 + }, + { + "epoch": 0.0629435222108875, + "grad_norm": 14.144310099072028, + "learning_rate": 1e-05, + "loss": 1.3002, + "step": 2424 + }, + { + "epoch": 0.06296948901047945, + "grad_norm": 9.419777910404664, + "learning_rate": 1e-05, + "loss": 1.2673, + "step": 2425 + }, + { + "epoch": 0.0629954558100714, + "grad_norm": 5.816909921946451, + "learning_rate": 1e-05, + "loss": 1.2411, + "step": 2426 + }, + { + "epoch": 0.06302142260966336, + "grad_norm": 6.003491974972124, + "learning_rate": 1e-05, + "loss": 1.2135, + "step": 2427 + }, + { + "epoch": 0.0630473894092553, + "grad_norm": 5.325338848956108, + "learning_rate": 1e-05, + "loss": 1.2624, + "step": 2428 + }, + { + "epoch": 0.06307335620884726, + "grad_norm": 5.85853639832243, + "learning_rate": 1e-05, + "loss": 1.2358, + "step": 2429 + }, + { + "epoch": 0.0630993230084392, + "grad_norm": 7.052223763882506, + "learning_rate": 1e-05, + "loss": 1.2491, + "step": 2430 + }, + { + "epoch": 0.06312528980803116, + "grad_norm": 4.519471744287391, + "learning_rate": 1e-05, + "loss": 1.2834, + "step": 2431 + }, + { + "epoch": 0.06315125660762311, + "grad_norm": 13.222929742371003, + "learning_rate": 1e-05, + "loss": 1.2627, + "step": 2432 + }, + { + "epoch": 0.06317722340721506, + "grad_norm": 4.700788050388757, + "learning_rate": 1e-05, + "loss": 1.2684, + "step": 2433 + }, + { + "epoch": 0.06320319020680701, + "grad_norm": 70.05764925224084, + "learning_rate": 1e-05, + "loss": 1.249, + "step": 2434 + }, + { + "epoch": 0.06322915700639896, + "grad_norm": 8.646876358787615, + "learning_rate": 1e-05, + "loss": 1.247, + "step": 2435 + }, + { + "epoch": 0.06325512380599091, + "grad_norm": 7.776399502417003, + "learning_rate": 1e-05, + "loss": 1.3043, + "step": 2436 + }, + { + "epoch": 0.06328109060558286, + "grad_norm": 5.187900573707142, + "learning_rate": 1e-05, + "loss": 1.2888, + "step": 2437 + }, + { + "epoch": 0.06330705740517481, + "grad_norm": 37.29423528310458, + "learning_rate": 1e-05, + "loss": 1.2951, + "step": 2438 + }, + { + "epoch": 0.06333302420476676, + "grad_norm": 5.269953637977809, + "learning_rate": 1e-05, + "loss": 1.2319, + "step": 2439 + }, + { + "epoch": 0.06335899100435871, + "grad_norm": 5.206527193344818, + "learning_rate": 1e-05, + "loss": 1.2614, + "step": 2440 + }, + { + "epoch": 0.06338495780395066, + "grad_norm": 14.83880828514346, + "learning_rate": 1e-05, + "loss": 1.2317, + "step": 2441 + }, + { + "epoch": 0.06341092460354261, + "grad_norm": 5.899781252151827, + "learning_rate": 1e-05, + "loss": 1.2523, + "step": 2442 + }, + { + "epoch": 0.06343689140313456, + "grad_norm": 5.204420140458789, + "learning_rate": 1e-05, + "loss": 1.2632, + "step": 2443 + }, + { + "epoch": 0.06346285820272651, + "grad_norm": 25.611673602227388, + "learning_rate": 1e-05, + "loss": 1.2492, + "step": 2444 + }, + { + "epoch": 0.06348882500231846, + "grad_norm": 4.868622349447836, + "learning_rate": 1e-05, + "loss": 1.267, + "step": 2445 + }, + { + "epoch": 0.06351479180191041, + "grad_norm": 8.573095295846194, + "learning_rate": 1e-05, + "loss": 1.2564, + "step": 2446 + }, + { + "epoch": 0.06354075860150236, + "grad_norm": 5.359331047389468, + "learning_rate": 1e-05, + "loss": 1.2214, + "step": 2447 + }, + { + "epoch": 0.06356672540109432, + "grad_norm": 8.276726635354505, + "learning_rate": 1e-05, + "loss": 1.2655, + "step": 2448 + }, + { + "epoch": 0.06359269220068627, + "grad_norm": 5.462679387339805, + "learning_rate": 1e-05, + "loss": 1.2378, + "step": 2449 + }, + { + "epoch": 0.06361865900027822, + "grad_norm": 5.7255192283818515, + "learning_rate": 1e-05, + "loss": 1.285, + "step": 2450 + }, + { + "epoch": 0.06364462579987017, + "grad_norm": 6.575737381696224, + "learning_rate": 1e-05, + "loss": 1.221, + "step": 2451 + }, + { + "epoch": 0.06367059259946212, + "grad_norm": 8.947392464242288, + "learning_rate": 1e-05, + "loss": 1.2426, + "step": 2452 + }, + { + "epoch": 0.06369655939905407, + "grad_norm": 5.775686855672246, + "learning_rate": 1e-05, + "loss": 1.256, + "step": 2453 + }, + { + "epoch": 0.06372252619864602, + "grad_norm": 4.9214144052218165, + "learning_rate": 1e-05, + "loss": 1.2753, + "step": 2454 + }, + { + "epoch": 0.06374849299823797, + "grad_norm": 4.92028085232657, + "learning_rate": 1e-05, + "loss": 1.2482, + "step": 2455 + }, + { + "epoch": 0.06377445979782992, + "grad_norm": 6.694098808824433, + "learning_rate": 1e-05, + "loss": 1.2416, + "step": 2456 + }, + { + "epoch": 0.06380042659742187, + "grad_norm": 6.041624643464664, + "learning_rate": 1e-05, + "loss": 1.2862, + "step": 2457 + }, + { + "epoch": 0.06382639339701382, + "grad_norm": 4.5401030248976255, + "learning_rate": 1e-05, + "loss": 1.2577, + "step": 2458 + }, + { + "epoch": 0.06385236019660577, + "grad_norm": 4.944557650600044, + "learning_rate": 1e-05, + "loss": 1.2577, + "step": 2459 + }, + { + "epoch": 0.06387832699619772, + "grad_norm": 8.2366706210654, + "learning_rate": 1e-05, + "loss": 1.2814, + "step": 2460 + }, + { + "epoch": 0.06390429379578967, + "grad_norm": 6.268752689684274, + "learning_rate": 1e-05, + "loss": 1.2272, + "step": 2461 + }, + { + "epoch": 0.06393026059538162, + "grad_norm": 25.811878113725324, + "learning_rate": 1e-05, + "loss": 1.3314, + "step": 2462 + }, + { + "epoch": 0.06395622739497357, + "grad_norm": 14.414316033508742, + "learning_rate": 1e-05, + "loss": 1.2915, + "step": 2463 + }, + { + "epoch": 0.06398219419456552, + "grad_norm": 5.793412565180442, + "learning_rate": 1e-05, + "loss": 1.2305, + "step": 2464 + }, + { + "epoch": 0.06400816099415747, + "grad_norm": 6.357787587887418, + "learning_rate": 1e-05, + "loss": 1.2503, + "step": 2465 + }, + { + "epoch": 0.06403412779374942, + "grad_norm": 4.96144344122835, + "learning_rate": 1e-05, + "loss": 1.202, + "step": 2466 + }, + { + "epoch": 0.06406009459334137, + "grad_norm": 4.879330594474093, + "learning_rate": 1e-05, + "loss": 1.2457, + "step": 2467 + }, + { + "epoch": 0.06408606139293332, + "grad_norm": 44.75981910430381, + "learning_rate": 1e-05, + "loss": 1.3408, + "step": 2468 + }, + { + "epoch": 0.06411202819252527, + "grad_norm": 5.090042543664392, + "learning_rate": 1e-05, + "loss": 1.2666, + "step": 2469 + }, + { + "epoch": 0.06413799499211723, + "grad_norm": 9.107434231240278, + "learning_rate": 1e-05, + "loss": 1.309, + "step": 2470 + }, + { + "epoch": 0.06416396179170918, + "grad_norm": 9.879893189134735, + "learning_rate": 1e-05, + "loss": 1.2608, + "step": 2471 + }, + { + "epoch": 0.06418992859130113, + "grad_norm": 8.763963728912804, + "learning_rate": 1e-05, + "loss": 1.2445, + "step": 2472 + }, + { + "epoch": 0.06421589539089308, + "grad_norm": 8.362937781217767, + "learning_rate": 1e-05, + "loss": 1.2714, + "step": 2473 + }, + { + "epoch": 0.06424186219048503, + "grad_norm": 4.924716480440242, + "learning_rate": 1e-05, + "loss": 1.2809, + "step": 2474 + }, + { + "epoch": 0.06426782899007698, + "grad_norm": 11.569242254054835, + "learning_rate": 1e-05, + "loss": 1.2651, + "step": 2475 + }, + { + "epoch": 0.06429379578966893, + "grad_norm": 7.808826040426222, + "learning_rate": 1e-05, + "loss": 1.3415, + "step": 2476 + }, + { + "epoch": 0.06431976258926088, + "grad_norm": 4.889807294547262, + "learning_rate": 1e-05, + "loss": 1.2746, + "step": 2477 + }, + { + "epoch": 0.06434572938885283, + "grad_norm": 4.2868384203954, + "learning_rate": 1e-05, + "loss": 1.2245, + "step": 2478 + }, + { + "epoch": 0.06437169618844478, + "grad_norm": 4.963206719875152, + "learning_rate": 1e-05, + "loss": 1.2316, + "step": 2479 + }, + { + "epoch": 0.06439766298803673, + "grad_norm": 5.672394867919339, + "learning_rate": 1e-05, + "loss": 1.2817, + "step": 2480 + }, + { + "epoch": 0.06442362978762868, + "grad_norm": 7.77861597297862, + "learning_rate": 1e-05, + "loss": 1.2879, + "step": 2481 + }, + { + "epoch": 0.06444959658722063, + "grad_norm": 5.846854787574463, + "learning_rate": 1e-05, + "loss": 1.2617, + "step": 2482 + }, + { + "epoch": 0.06447556338681258, + "grad_norm": 7.592776389167433, + "learning_rate": 1e-05, + "loss": 1.3314, + "step": 2483 + }, + { + "epoch": 0.06450153018640453, + "grad_norm": 5.571760206168599, + "learning_rate": 1e-05, + "loss": 1.2591, + "step": 2484 + }, + { + "epoch": 0.06452749698599648, + "grad_norm": 8.873515260103465, + "learning_rate": 1e-05, + "loss": 1.3137, + "step": 2485 + }, + { + "epoch": 0.06455346378558843, + "grad_norm": 8.714019914276433, + "learning_rate": 1e-05, + "loss": 1.2995, + "step": 2486 + }, + { + "epoch": 0.06457943058518037, + "grad_norm": 3.811346285996562, + "learning_rate": 1e-05, + "loss": 1.2365, + "step": 2487 + }, + { + "epoch": 0.06460539738477232, + "grad_norm": 3.951087517057025, + "learning_rate": 1e-05, + "loss": 1.2565, + "step": 2488 + }, + { + "epoch": 0.06463136418436427, + "grad_norm": 15.182586754297102, + "learning_rate": 1e-05, + "loss": 1.268, + "step": 2489 + }, + { + "epoch": 0.06465733098395622, + "grad_norm": 5.869053204943787, + "learning_rate": 1e-05, + "loss": 1.2398, + "step": 2490 + }, + { + "epoch": 0.06468329778354817, + "grad_norm": 5.190526975364103, + "learning_rate": 1e-05, + "loss": 1.2623, + "step": 2491 + }, + { + "epoch": 0.06470926458314012, + "grad_norm": 6.0872154216320835, + "learning_rate": 1e-05, + "loss": 1.2451, + "step": 2492 + }, + { + "epoch": 0.06473523138273207, + "grad_norm": 7.180258378504222, + "learning_rate": 1e-05, + "loss": 1.2374, + "step": 2493 + }, + { + "epoch": 0.06476119818232402, + "grad_norm": 5.917391808717004, + "learning_rate": 1e-05, + "loss": 1.2522, + "step": 2494 + }, + { + "epoch": 0.06478716498191597, + "grad_norm": 8.226501290726704, + "learning_rate": 1e-05, + "loss": 1.2987, + "step": 2495 + }, + { + "epoch": 0.06481313178150792, + "grad_norm": 7.542543925763397, + "learning_rate": 1e-05, + "loss": 1.275, + "step": 2496 + }, + { + "epoch": 0.06483909858109987, + "grad_norm": 4.957615883262002, + "learning_rate": 1e-05, + "loss": 1.2683, + "step": 2497 + }, + { + "epoch": 0.06486506538069182, + "grad_norm": 5.438057004965819, + "learning_rate": 1e-05, + "loss": 1.2966, + "step": 2498 + }, + { + "epoch": 0.06489103218028378, + "grad_norm": 5.068472459299618, + "learning_rate": 1e-05, + "loss": 1.2668, + "step": 2499 + }, + { + "epoch": 0.06491699897987573, + "grad_norm": 5.6915989452408535, + "learning_rate": 1e-05, + "loss": 1.2619, + "step": 2500 + }, + { + "epoch": 0.06494296577946768, + "grad_norm": 5.259070961426489, + "learning_rate": 1e-05, + "loss": 1.3037, + "step": 2501 + }, + { + "epoch": 0.06496893257905963, + "grad_norm": 5.093719482330457, + "learning_rate": 1e-05, + "loss": 1.2622, + "step": 2502 + }, + { + "epoch": 0.06499489937865158, + "grad_norm": 5.272519089586004, + "learning_rate": 1e-05, + "loss": 1.2424, + "step": 2503 + }, + { + "epoch": 0.06502086617824353, + "grad_norm": 4.832780784505333, + "learning_rate": 1e-05, + "loss": 1.2574, + "step": 2504 + }, + { + "epoch": 0.06504683297783548, + "grad_norm": 5.478216729535863, + "learning_rate": 1e-05, + "loss": 1.3022, + "step": 2505 + }, + { + "epoch": 0.06507279977742743, + "grad_norm": 6.097127281768676, + "learning_rate": 1e-05, + "loss": 1.2472, + "step": 2506 + }, + { + "epoch": 0.06509876657701938, + "grad_norm": 4.497172845091993, + "learning_rate": 1e-05, + "loss": 1.2913, + "step": 2507 + }, + { + "epoch": 0.06512473337661133, + "grad_norm": 4.963889185563837, + "learning_rate": 1e-05, + "loss": 1.2734, + "step": 2508 + }, + { + "epoch": 0.06515070017620328, + "grad_norm": 8.497146576290469, + "learning_rate": 1e-05, + "loss": 1.2363, + "step": 2509 + }, + { + "epoch": 0.06517666697579523, + "grad_norm": 4.645616512057319, + "learning_rate": 1e-05, + "loss": 1.2416, + "step": 2510 + }, + { + "epoch": 0.06520263377538718, + "grad_norm": 5.296234542251512, + "learning_rate": 1e-05, + "loss": 1.2677, + "step": 2511 + }, + { + "epoch": 0.06522860057497913, + "grad_norm": 10.231099264051684, + "learning_rate": 1e-05, + "loss": 1.3021, + "step": 2512 + }, + { + "epoch": 0.06525456737457108, + "grad_norm": 13.193992426753418, + "learning_rate": 1e-05, + "loss": 1.2731, + "step": 2513 + }, + { + "epoch": 0.06528053417416303, + "grad_norm": 7.107852043264921, + "learning_rate": 1e-05, + "loss": 1.3051, + "step": 2514 + }, + { + "epoch": 0.06530650097375498, + "grad_norm": 4.817642634863131, + "learning_rate": 1e-05, + "loss": 1.2578, + "step": 2515 + }, + { + "epoch": 0.06533246777334693, + "grad_norm": 4.597555647197448, + "learning_rate": 1e-05, + "loss": 1.3073, + "step": 2516 + }, + { + "epoch": 0.06535843457293888, + "grad_norm": 9.546560129320632, + "learning_rate": 1e-05, + "loss": 1.2573, + "step": 2517 + }, + { + "epoch": 0.06538440137253083, + "grad_norm": 5.578513081191213, + "learning_rate": 1e-05, + "loss": 1.2897, + "step": 2518 + }, + { + "epoch": 0.06541036817212278, + "grad_norm": 5.022972548163537, + "learning_rate": 1e-05, + "loss": 1.2889, + "step": 2519 + }, + { + "epoch": 0.06543633497171474, + "grad_norm": 9.85036282718299, + "learning_rate": 1e-05, + "loss": 1.2764, + "step": 2520 + }, + { + "epoch": 0.06546230177130669, + "grad_norm": 6.798775102583392, + "learning_rate": 1e-05, + "loss": 1.2685, + "step": 2521 + }, + { + "epoch": 0.06548826857089864, + "grad_norm": 4.703684827366752, + "learning_rate": 1e-05, + "loss": 1.2742, + "step": 2522 + }, + { + "epoch": 0.06551423537049059, + "grad_norm": 5.921979998863015, + "learning_rate": 1e-05, + "loss": 1.2474, + "step": 2523 + }, + { + "epoch": 0.06554020217008254, + "grad_norm": 12.215350290304562, + "learning_rate": 1e-05, + "loss": 1.2387, + "step": 2524 + }, + { + "epoch": 0.06556616896967449, + "grad_norm": 3.8385583815950453, + "learning_rate": 1e-05, + "loss": 1.256, + "step": 2525 + }, + { + "epoch": 0.06559213576926644, + "grad_norm": 4.287449934864514, + "learning_rate": 1e-05, + "loss": 1.2187, + "step": 2526 + }, + { + "epoch": 0.06561810256885839, + "grad_norm": 5.211221712931912, + "learning_rate": 1e-05, + "loss": 1.2322, + "step": 2527 + }, + { + "epoch": 0.06564406936845034, + "grad_norm": 6.034185930289901, + "learning_rate": 1e-05, + "loss": 1.2968, + "step": 2528 + }, + { + "epoch": 0.06567003616804229, + "grad_norm": 8.570049438333964, + "learning_rate": 1e-05, + "loss": 1.2391, + "step": 2529 + }, + { + "epoch": 0.06569600296763424, + "grad_norm": 4.829377737142015, + "learning_rate": 1e-05, + "loss": 1.2622, + "step": 2530 + }, + { + "epoch": 0.06572196976722619, + "grad_norm": 16.10721239879097, + "learning_rate": 1e-05, + "loss": 1.2105, + "step": 2531 + }, + { + "epoch": 0.06574793656681814, + "grad_norm": 5.2746177426362015, + "learning_rate": 1e-05, + "loss": 1.2737, + "step": 2532 + }, + { + "epoch": 0.06577390336641009, + "grad_norm": 4.920359157075746, + "learning_rate": 1e-05, + "loss": 1.2571, + "step": 2533 + }, + { + "epoch": 0.06579987016600204, + "grad_norm": 5.603331984671651, + "learning_rate": 1e-05, + "loss": 1.2773, + "step": 2534 + }, + { + "epoch": 0.06582583696559399, + "grad_norm": 6.038499971274006, + "learning_rate": 1e-05, + "loss": 1.2721, + "step": 2535 + }, + { + "epoch": 0.06585180376518594, + "grad_norm": 5.400549380295378, + "learning_rate": 1e-05, + "loss": 1.3496, + "step": 2536 + }, + { + "epoch": 0.0658777705647779, + "grad_norm": 6.300680747416058, + "learning_rate": 1e-05, + "loss": 1.3086, + "step": 2537 + }, + { + "epoch": 0.06590373736436984, + "grad_norm": 9.80704225022784, + "learning_rate": 1e-05, + "loss": 1.2931, + "step": 2538 + }, + { + "epoch": 0.0659297041639618, + "grad_norm": 10.670507295913744, + "learning_rate": 1e-05, + "loss": 1.2624, + "step": 2539 + }, + { + "epoch": 0.06595567096355374, + "grad_norm": 5.603912328836188, + "learning_rate": 1e-05, + "loss": 1.2486, + "step": 2540 + }, + { + "epoch": 0.0659816377631457, + "grad_norm": 5.217851572881418, + "learning_rate": 1e-05, + "loss": 1.298, + "step": 2541 + }, + { + "epoch": 0.06600760456273765, + "grad_norm": 9.806244040415056, + "learning_rate": 1e-05, + "loss": 1.2602, + "step": 2542 + }, + { + "epoch": 0.0660335713623296, + "grad_norm": 9.49969883491585, + "learning_rate": 1e-05, + "loss": 1.2182, + "step": 2543 + }, + { + "epoch": 0.06605953816192155, + "grad_norm": 7.126954998372596, + "learning_rate": 1e-05, + "loss": 1.2469, + "step": 2544 + }, + { + "epoch": 0.0660855049615135, + "grad_norm": 7.601666880064448, + "learning_rate": 1e-05, + "loss": 1.3032, + "step": 2545 + }, + { + "epoch": 0.06611147176110545, + "grad_norm": 6.532993211780935, + "learning_rate": 1e-05, + "loss": 1.2437, + "step": 2546 + }, + { + "epoch": 0.0661374385606974, + "grad_norm": 5.384807067645202, + "learning_rate": 1e-05, + "loss": 1.2852, + "step": 2547 + }, + { + "epoch": 0.06616340536028935, + "grad_norm": 4.868269945406945, + "learning_rate": 1e-05, + "loss": 1.239, + "step": 2548 + }, + { + "epoch": 0.0661893721598813, + "grad_norm": 7.0488029600028765, + "learning_rate": 1e-05, + "loss": 1.255, + "step": 2549 + }, + { + "epoch": 0.06621533895947325, + "grad_norm": 5.763744347329016, + "learning_rate": 1e-05, + "loss": 1.253, + "step": 2550 + }, + { + "epoch": 0.0662413057590652, + "grad_norm": 6.191904985334651, + "learning_rate": 1e-05, + "loss": 1.2482, + "step": 2551 + }, + { + "epoch": 0.06626727255865715, + "grad_norm": 6.557217460941655, + "learning_rate": 1e-05, + "loss": 1.2784, + "step": 2552 + }, + { + "epoch": 0.0662932393582491, + "grad_norm": 13.520668879107888, + "learning_rate": 1e-05, + "loss": 1.2545, + "step": 2553 + }, + { + "epoch": 0.06631920615784105, + "grad_norm": 5.902085869821456, + "learning_rate": 1e-05, + "loss": 1.2696, + "step": 2554 + }, + { + "epoch": 0.066345172957433, + "grad_norm": 5.197445814355746, + "learning_rate": 1e-05, + "loss": 1.3069, + "step": 2555 + }, + { + "epoch": 0.06637113975702495, + "grad_norm": 8.648218277906972, + "learning_rate": 1e-05, + "loss": 1.2675, + "step": 2556 + }, + { + "epoch": 0.0663971065566169, + "grad_norm": 5.420457814155467, + "learning_rate": 1e-05, + "loss": 1.3153, + "step": 2557 + }, + { + "epoch": 0.06642307335620885, + "grad_norm": 6.679660658754551, + "learning_rate": 1e-05, + "loss": 1.2886, + "step": 2558 + }, + { + "epoch": 0.0664490401558008, + "grad_norm": 72.77638924837677, + "learning_rate": 1e-05, + "loss": 1.3631, + "step": 2559 + }, + { + "epoch": 0.06647500695539275, + "grad_norm": 7.582650569599296, + "learning_rate": 1e-05, + "loss": 1.2705, + "step": 2560 + }, + { + "epoch": 0.0665009737549847, + "grad_norm": 7.364343994510469, + "learning_rate": 1e-05, + "loss": 1.2698, + "step": 2561 + }, + { + "epoch": 0.06652694055457666, + "grad_norm": 9.957642684247535, + "learning_rate": 1e-05, + "loss": 1.2731, + "step": 2562 + }, + { + "epoch": 0.0665529073541686, + "grad_norm": 4.611558920526792, + "learning_rate": 1e-05, + "loss": 1.228, + "step": 2563 + }, + { + "epoch": 0.06657887415376054, + "grad_norm": 5.4795970481801675, + "learning_rate": 1e-05, + "loss": 1.262, + "step": 2564 + }, + { + "epoch": 0.06660484095335249, + "grad_norm": 5.694567303249672, + "learning_rate": 1e-05, + "loss": 1.2901, + "step": 2565 + }, + { + "epoch": 0.06663080775294444, + "grad_norm": 7.063433290104041, + "learning_rate": 1e-05, + "loss": 1.2724, + "step": 2566 + }, + { + "epoch": 0.0666567745525364, + "grad_norm": 35.5400766163839, + "learning_rate": 1e-05, + "loss": 1.2519, + "step": 2567 + }, + { + "epoch": 0.06668274135212834, + "grad_norm": 6.850158756393385, + "learning_rate": 1e-05, + "loss": 1.2637, + "step": 2568 + }, + { + "epoch": 0.0667087081517203, + "grad_norm": 12.468182773164623, + "learning_rate": 1e-05, + "loss": 1.3001, + "step": 2569 + }, + { + "epoch": 0.06673467495131225, + "grad_norm": 15.164095706107384, + "learning_rate": 1e-05, + "loss": 1.2508, + "step": 2570 + }, + { + "epoch": 0.0667606417509042, + "grad_norm": 5.098429209836579, + "learning_rate": 1e-05, + "loss": 1.3053, + "step": 2571 + }, + { + "epoch": 0.06678660855049615, + "grad_norm": 6.857357623506314, + "learning_rate": 1e-05, + "loss": 1.3059, + "step": 2572 + }, + { + "epoch": 0.0668125753500881, + "grad_norm": 6.0452383423824525, + "learning_rate": 1e-05, + "loss": 1.2654, + "step": 2573 + }, + { + "epoch": 0.06683854214968005, + "grad_norm": 6.100134375921311, + "learning_rate": 1e-05, + "loss": 1.263, + "step": 2574 + }, + { + "epoch": 0.066864508949272, + "grad_norm": 5.454193622628383, + "learning_rate": 1e-05, + "loss": 1.2886, + "step": 2575 + }, + { + "epoch": 0.06689047574886395, + "grad_norm": 4.937342580869212, + "learning_rate": 1e-05, + "loss": 1.2666, + "step": 2576 + }, + { + "epoch": 0.0669164425484559, + "grad_norm": 6.476389331596276, + "learning_rate": 1e-05, + "loss": 1.3008, + "step": 2577 + }, + { + "epoch": 0.06694240934804785, + "grad_norm": 5.074496337251829, + "learning_rate": 1e-05, + "loss": 1.3119, + "step": 2578 + }, + { + "epoch": 0.0669683761476398, + "grad_norm": 5.092718101778712, + "learning_rate": 1e-05, + "loss": 1.2994, + "step": 2579 + }, + { + "epoch": 0.06699434294723175, + "grad_norm": 15.30089842707876, + "learning_rate": 1e-05, + "loss": 1.2673, + "step": 2580 + }, + { + "epoch": 0.0670203097468237, + "grad_norm": 13.01638582897834, + "learning_rate": 1e-05, + "loss": 1.2953, + "step": 2581 + }, + { + "epoch": 0.06704627654641565, + "grad_norm": 6.026400657342034, + "learning_rate": 1e-05, + "loss": 1.3015, + "step": 2582 + }, + { + "epoch": 0.0670722433460076, + "grad_norm": 6.320320949855291, + "learning_rate": 1e-05, + "loss": 1.2628, + "step": 2583 + }, + { + "epoch": 0.06709821014559955, + "grad_norm": 8.936494784079432, + "learning_rate": 1e-05, + "loss": 1.2475, + "step": 2584 + }, + { + "epoch": 0.0671241769451915, + "grad_norm": 5.174050786781035, + "learning_rate": 1e-05, + "loss": 1.2567, + "step": 2585 + }, + { + "epoch": 0.06715014374478345, + "grad_norm": 4.762257671448749, + "learning_rate": 1e-05, + "loss": 1.2007, + "step": 2586 + }, + { + "epoch": 0.0671761105443754, + "grad_norm": 4.622624379986683, + "learning_rate": 1e-05, + "loss": 1.3135, + "step": 2587 + }, + { + "epoch": 0.06720207734396735, + "grad_norm": 5.479968961084827, + "learning_rate": 1e-05, + "loss": 1.2437, + "step": 2588 + }, + { + "epoch": 0.0672280441435593, + "grad_norm": 5.319320250085873, + "learning_rate": 1e-05, + "loss": 1.241, + "step": 2589 + }, + { + "epoch": 0.06725401094315125, + "grad_norm": 5.588375130548364, + "learning_rate": 1e-05, + "loss": 1.2915, + "step": 2590 + }, + { + "epoch": 0.0672799777427432, + "grad_norm": 7.456199155955498, + "learning_rate": 1e-05, + "loss": 1.2839, + "step": 2591 + }, + { + "epoch": 0.06730594454233516, + "grad_norm": 7.4276257531335865, + "learning_rate": 1e-05, + "loss": 1.3281, + "step": 2592 + }, + { + "epoch": 0.0673319113419271, + "grad_norm": 5.738420852199072, + "learning_rate": 1e-05, + "loss": 1.2469, + "step": 2593 + }, + { + "epoch": 0.06735787814151906, + "grad_norm": 4.71108345577432, + "learning_rate": 1e-05, + "loss": 1.2501, + "step": 2594 + }, + { + "epoch": 0.067383844941111, + "grad_norm": 5.750282280668754, + "learning_rate": 1e-05, + "loss": 1.3136, + "step": 2595 + }, + { + "epoch": 0.06740981174070296, + "grad_norm": 4.563776961042957, + "learning_rate": 1e-05, + "loss": 1.2596, + "step": 2596 + }, + { + "epoch": 0.06743577854029491, + "grad_norm": 5.786980505547088, + "learning_rate": 1e-05, + "loss": 1.2559, + "step": 2597 + }, + { + "epoch": 0.06746174533988686, + "grad_norm": 3.919825288032145, + "learning_rate": 1e-05, + "loss": 1.2527, + "step": 2598 + }, + { + "epoch": 0.06748771213947881, + "grad_norm": 22.818955101276273, + "learning_rate": 1e-05, + "loss": 1.2527, + "step": 2599 + }, + { + "epoch": 0.06751367893907076, + "grad_norm": 4.837043302993996, + "learning_rate": 1e-05, + "loss": 1.2533, + "step": 2600 + }, + { + "epoch": 0.06753964573866271, + "grad_norm": 6.981204548604982, + "learning_rate": 1e-05, + "loss": 1.27, + "step": 2601 + }, + { + "epoch": 0.06756561253825466, + "grad_norm": 5.624259900041762, + "learning_rate": 1e-05, + "loss": 1.2964, + "step": 2602 + }, + { + "epoch": 0.06759157933784661, + "grad_norm": 4.506454924475349, + "learning_rate": 1e-05, + "loss": 1.284, + "step": 2603 + }, + { + "epoch": 0.06761754613743856, + "grad_norm": 10.5882921591335, + "learning_rate": 1e-05, + "loss": 1.2885, + "step": 2604 + }, + { + "epoch": 0.06764351293703051, + "grad_norm": 4.625276093361275, + "learning_rate": 1e-05, + "loss": 1.3171, + "step": 2605 + }, + { + "epoch": 0.06766947973662246, + "grad_norm": 10.979402504595146, + "learning_rate": 1e-05, + "loss": 1.3131, + "step": 2606 + }, + { + "epoch": 0.06769544653621441, + "grad_norm": 6.486503968558939, + "learning_rate": 1e-05, + "loss": 1.3188, + "step": 2607 + }, + { + "epoch": 0.06772141333580636, + "grad_norm": 5.333187617854478, + "learning_rate": 1e-05, + "loss": 1.2822, + "step": 2608 + }, + { + "epoch": 0.06774738013539831, + "grad_norm": 5.746746635559162, + "learning_rate": 1e-05, + "loss": 1.2665, + "step": 2609 + }, + { + "epoch": 0.06777334693499026, + "grad_norm": 8.152707320744124, + "learning_rate": 1e-05, + "loss": 1.212, + "step": 2610 + }, + { + "epoch": 0.06779931373458221, + "grad_norm": 8.288756455677955, + "learning_rate": 1e-05, + "loss": 1.2595, + "step": 2611 + }, + { + "epoch": 0.06782528053417416, + "grad_norm": 12.370377727344767, + "learning_rate": 1e-05, + "loss": 1.2954, + "step": 2612 + }, + { + "epoch": 0.06785124733376612, + "grad_norm": 6.118063599659018, + "learning_rate": 1e-05, + "loss": 1.2745, + "step": 2613 + }, + { + "epoch": 0.06787721413335807, + "grad_norm": 6.504083084589267, + "learning_rate": 1e-05, + "loss": 1.285, + "step": 2614 + }, + { + "epoch": 0.06790318093295002, + "grad_norm": 6.5999723607264835, + "learning_rate": 1e-05, + "loss": 1.2789, + "step": 2615 + }, + { + "epoch": 0.06792914773254197, + "grad_norm": 5.04490618494705, + "learning_rate": 1e-05, + "loss": 1.2798, + "step": 2616 + }, + { + "epoch": 0.06795511453213392, + "grad_norm": 7.488604980857661, + "learning_rate": 1e-05, + "loss": 1.2331, + "step": 2617 + }, + { + "epoch": 0.06798108133172587, + "grad_norm": 7.725964591533468, + "learning_rate": 1e-05, + "loss": 1.3457, + "step": 2618 + }, + { + "epoch": 0.06800704813131782, + "grad_norm": 6.397135880465924, + "learning_rate": 1e-05, + "loss": 1.2719, + "step": 2619 + }, + { + "epoch": 0.06803301493090977, + "grad_norm": 5.864527357070318, + "learning_rate": 1e-05, + "loss": 1.2978, + "step": 2620 + }, + { + "epoch": 0.06805898173050172, + "grad_norm": 6.24425303406013, + "learning_rate": 1e-05, + "loss": 1.2823, + "step": 2621 + }, + { + "epoch": 0.06808494853009367, + "grad_norm": 6.498011651730504, + "learning_rate": 1e-05, + "loss": 1.2569, + "step": 2622 + }, + { + "epoch": 0.06811091532968562, + "grad_norm": 19.458062877313868, + "learning_rate": 1e-05, + "loss": 1.2521, + "step": 2623 + }, + { + "epoch": 0.06813688212927757, + "grad_norm": 8.352781515483938, + "learning_rate": 1e-05, + "loss": 1.2164, + "step": 2624 + }, + { + "epoch": 0.06816284892886952, + "grad_norm": 5.715295634261776, + "learning_rate": 1e-05, + "loss": 1.2695, + "step": 2625 + }, + { + "epoch": 0.06818881572846147, + "grad_norm": 14.824964469879957, + "learning_rate": 1e-05, + "loss": 1.233, + "step": 2626 + }, + { + "epoch": 0.06821478252805342, + "grad_norm": 24.396665051293382, + "learning_rate": 1e-05, + "loss": 1.2633, + "step": 2627 + }, + { + "epoch": 0.06824074932764537, + "grad_norm": 6.504992401798863, + "learning_rate": 1e-05, + "loss": 1.2673, + "step": 2628 + }, + { + "epoch": 0.06826671612723732, + "grad_norm": 11.604934931765357, + "learning_rate": 1e-05, + "loss": 1.1967, + "step": 2629 + }, + { + "epoch": 0.06829268292682927, + "grad_norm": 5.525578063827632, + "learning_rate": 1e-05, + "loss": 1.234, + "step": 2630 + }, + { + "epoch": 0.06831864972642122, + "grad_norm": 11.561048261903524, + "learning_rate": 1e-05, + "loss": 1.2812, + "step": 2631 + }, + { + "epoch": 0.06834461652601317, + "grad_norm": 10.102989101941331, + "learning_rate": 1e-05, + "loss": 1.274, + "step": 2632 + }, + { + "epoch": 0.06837058332560512, + "grad_norm": 12.19903863339655, + "learning_rate": 1e-05, + "loss": 1.26, + "step": 2633 + }, + { + "epoch": 0.06839655012519708, + "grad_norm": 5.105523106957671, + "learning_rate": 1e-05, + "loss": 1.2713, + "step": 2634 + }, + { + "epoch": 0.06842251692478903, + "grad_norm": 6.967880673776236, + "learning_rate": 1e-05, + "loss": 1.2677, + "step": 2635 + }, + { + "epoch": 0.06844848372438098, + "grad_norm": 21.96816794436967, + "learning_rate": 1e-05, + "loss": 1.2631, + "step": 2636 + }, + { + "epoch": 0.06847445052397293, + "grad_norm": 6.950122244878143, + "learning_rate": 1e-05, + "loss": 1.269, + "step": 2637 + }, + { + "epoch": 0.06850041732356488, + "grad_norm": 6.516154828015966, + "learning_rate": 1e-05, + "loss": 1.3109, + "step": 2638 + }, + { + "epoch": 0.06852638412315683, + "grad_norm": 5.698071056368789, + "learning_rate": 1e-05, + "loss": 1.2174, + "step": 2639 + }, + { + "epoch": 0.06855235092274876, + "grad_norm": 37.10807070076282, + "learning_rate": 1e-05, + "loss": 1.3047, + "step": 2640 + }, + { + "epoch": 0.06857831772234071, + "grad_norm": 10.042384162901406, + "learning_rate": 1e-05, + "loss": 1.2987, + "step": 2641 + }, + { + "epoch": 0.06860428452193267, + "grad_norm": 5.6708283943119255, + "learning_rate": 1e-05, + "loss": 1.271, + "step": 2642 + }, + { + "epoch": 0.06863025132152462, + "grad_norm": 6.435780119540312, + "learning_rate": 1e-05, + "loss": 1.2651, + "step": 2643 + }, + { + "epoch": 0.06865621812111657, + "grad_norm": 5.46919920030146, + "learning_rate": 1e-05, + "loss": 1.2681, + "step": 2644 + }, + { + "epoch": 0.06868218492070852, + "grad_norm": 11.138117910727185, + "learning_rate": 1e-05, + "loss": 1.2273, + "step": 2645 + }, + { + "epoch": 0.06870815172030047, + "grad_norm": 6.564315399477569, + "learning_rate": 1e-05, + "loss": 1.2793, + "step": 2646 + }, + { + "epoch": 0.06873411851989242, + "grad_norm": 5.330705571522291, + "learning_rate": 1e-05, + "loss": 1.2011, + "step": 2647 + }, + { + "epoch": 0.06876008531948437, + "grad_norm": 7.187088929736124, + "learning_rate": 1e-05, + "loss": 1.2689, + "step": 2648 + }, + { + "epoch": 0.06878605211907632, + "grad_norm": 72.40215987045035, + "learning_rate": 1e-05, + "loss": 1.3284, + "step": 2649 + }, + { + "epoch": 0.06881201891866827, + "grad_norm": 8.922208923397497, + "learning_rate": 1e-05, + "loss": 1.2733, + "step": 2650 + }, + { + "epoch": 0.06883798571826022, + "grad_norm": 6.009293035224953, + "learning_rate": 1e-05, + "loss": 1.2875, + "step": 2651 + }, + { + "epoch": 0.06886395251785217, + "grad_norm": 5.936998526828885, + "learning_rate": 1e-05, + "loss": 1.3207, + "step": 2652 + }, + { + "epoch": 0.06888991931744412, + "grad_norm": 11.850013546191493, + "learning_rate": 1e-05, + "loss": 1.3048, + "step": 2653 + }, + { + "epoch": 0.06891588611703607, + "grad_norm": 9.428428483680353, + "learning_rate": 1e-05, + "loss": 1.2953, + "step": 2654 + }, + { + "epoch": 0.06894185291662802, + "grad_norm": 8.656458001787957, + "learning_rate": 1e-05, + "loss": 1.2534, + "step": 2655 + }, + { + "epoch": 0.06896781971621997, + "grad_norm": 6.863145908330465, + "learning_rate": 1e-05, + "loss": 1.2613, + "step": 2656 + }, + { + "epoch": 0.06899378651581192, + "grad_norm": 5.5814209504981935, + "learning_rate": 1e-05, + "loss": 1.2519, + "step": 2657 + }, + { + "epoch": 0.06901975331540387, + "grad_norm": 6.96455181577601, + "learning_rate": 1e-05, + "loss": 1.2783, + "step": 2658 + }, + { + "epoch": 0.06904572011499582, + "grad_norm": 6.905116475840352, + "learning_rate": 1e-05, + "loss": 1.2657, + "step": 2659 + }, + { + "epoch": 0.06907168691458777, + "grad_norm": 7.055005984124688, + "learning_rate": 1e-05, + "loss": 1.3016, + "step": 2660 + }, + { + "epoch": 0.06909765371417972, + "grad_norm": 24.063104953840252, + "learning_rate": 1e-05, + "loss": 1.2652, + "step": 2661 + }, + { + "epoch": 0.06912362051377167, + "grad_norm": 7.302310463007207, + "learning_rate": 1e-05, + "loss": 1.2699, + "step": 2662 + }, + { + "epoch": 0.06914958731336363, + "grad_norm": 19.09246811879458, + "learning_rate": 1e-05, + "loss": 1.238, + "step": 2663 + }, + { + "epoch": 0.06917555411295558, + "grad_norm": 8.18424305672926, + "learning_rate": 1e-05, + "loss": 1.2584, + "step": 2664 + }, + { + "epoch": 0.06920152091254753, + "grad_norm": 8.453074456872924, + "learning_rate": 1e-05, + "loss": 1.2449, + "step": 2665 + }, + { + "epoch": 0.06922748771213948, + "grad_norm": 5.853116123337336, + "learning_rate": 1e-05, + "loss": 1.2452, + "step": 2666 + }, + { + "epoch": 0.06925345451173143, + "grad_norm": 18.86646284840262, + "learning_rate": 1e-05, + "loss": 1.2828, + "step": 2667 + }, + { + "epoch": 0.06927942131132338, + "grad_norm": 7.559240726510046, + "learning_rate": 1e-05, + "loss": 1.3108, + "step": 2668 + }, + { + "epoch": 0.06930538811091533, + "grad_norm": 8.390028039023795, + "learning_rate": 1e-05, + "loss": 1.2445, + "step": 2669 + }, + { + "epoch": 0.06933135491050728, + "grad_norm": 13.31825039023637, + "learning_rate": 1e-05, + "loss": 1.2486, + "step": 2670 + }, + { + "epoch": 0.06935732171009923, + "grad_norm": 13.478119814911766, + "learning_rate": 1e-05, + "loss": 1.2778, + "step": 2671 + }, + { + "epoch": 0.06938328850969118, + "grad_norm": 4.829952745858046, + "learning_rate": 1e-05, + "loss": 1.2068, + "step": 2672 + }, + { + "epoch": 0.06940925530928313, + "grad_norm": 7.682609925655173, + "learning_rate": 1e-05, + "loss": 1.2464, + "step": 2673 + }, + { + "epoch": 0.06943522210887508, + "grad_norm": 7.7213597383956145, + "learning_rate": 1e-05, + "loss": 1.2345, + "step": 2674 + }, + { + "epoch": 0.06946118890846703, + "grad_norm": 7.793905361557331, + "learning_rate": 1e-05, + "loss": 1.2748, + "step": 2675 + }, + { + "epoch": 0.06948715570805898, + "grad_norm": 9.118624615607942, + "learning_rate": 1e-05, + "loss": 1.2394, + "step": 2676 + }, + { + "epoch": 0.06951312250765093, + "grad_norm": 9.151850662266819, + "learning_rate": 1e-05, + "loss": 1.2526, + "step": 2677 + }, + { + "epoch": 0.06953908930724288, + "grad_norm": 5.405007809501885, + "learning_rate": 1e-05, + "loss": 1.2545, + "step": 2678 + }, + { + "epoch": 0.06956505610683483, + "grad_norm": 14.500813428490508, + "learning_rate": 1e-05, + "loss": 1.2686, + "step": 2679 + }, + { + "epoch": 0.06959102290642678, + "grad_norm": 6.0784535454062745, + "learning_rate": 1e-05, + "loss": 1.2349, + "step": 2680 + }, + { + "epoch": 0.06961698970601873, + "grad_norm": 7.912010295834221, + "learning_rate": 1e-05, + "loss": 1.2407, + "step": 2681 + }, + { + "epoch": 0.06964295650561068, + "grad_norm": 6.585397356538621, + "learning_rate": 1e-05, + "loss": 1.2756, + "step": 2682 + }, + { + "epoch": 0.06966892330520263, + "grad_norm": 28.01388314614347, + "learning_rate": 1e-05, + "loss": 1.294, + "step": 2683 + }, + { + "epoch": 0.06969489010479459, + "grad_norm": 7.312171080927199, + "learning_rate": 1e-05, + "loss": 1.2836, + "step": 2684 + }, + { + "epoch": 0.06972085690438654, + "grad_norm": 7.771132421660966, + "learning_rate": 1e-05, + "loss": 1.2645, + "step": 2685 + }, + { + "epoch": 0.06974682370397849, + "grad_norm": 5.542141466188859, + "learning_rate": 1e-05, + "loss": 1.3142, + "step": 2686 + }, + { + "epoch": 0.06977279050357044, + "grad_norm": 17.017919632641554, + "learning_rate": 1e-05, + "loss": 1.287, + "step": 2687 + }, + { + "epoch": 0.06979875730316239, + "grad_norm": 5.263379079286936, + "learning_rate": 1e-05, + "loss": 1.2858, + "step": 2688 + }, + { + "epoch": 0.06982472410275434, + "grad_norm": 7.175379875308377, + "learning_rate": 1e-05, + "loss": 1.2748, + "step": 2689 + }, + { + "epoch": 0.06985069090234629, + "grad_norm": 10.672483368228653, + "learning_rate": 1e-05, + "loss": 1.2561, + "step": 2690 + }, + { + "epoch": 0.06987665770193824, + "grad_norm": 8.834762697514106, + "learning_rate": 1e-05, + "loss": 1.2822, + "step": 2691 + }, + { + "epoch": 0.06990262450153019, + "grad_norm": 11.17436700517343, + "learning_rate": 1e-05, + "loss": 1.3203, + "step": 2692 + }, + { + "epoch": 0.06992859130112214, + "grad_norm": 5.866800308150337, + "learning_rate": 1e-05, + "loss": 1.2759, + "step": 2693 + }, + { + "epoch": 0.06995455810071409, + "grad_norm": 9.316386186430178, + "learning_rate": 1e-05, + "loss": 1.2579, + "step": 2694 + }, + { + "epoch": 0.06998052490030604, + "grad_norm": 9.053985594828788, + "learning_rate": 1e-05, + "loss": 1.2573, + "step": 2695 + }, + { + "epoch": 0.07000649169989799, + "grad_norm": 19.062101366221242, + "learning_rate": 1e-05, + "loss": 1.2667, + "step": 2696 + }, + { + "epoch": 0.07003245849948994, + "grad_norm": 7.28323167021424, + "learning_rate": 1e-05, + "loss": 1.2508, + "step": 2697 + }, + { + "epoch": 0.07005842529908189, + "grad_norm": 6.467527412206053, + "learning_rate": 1e-05, + "loss": 1.2879, + "step": 2698 + }, + { + "epoch": 0.07008439209867384, + "grad_norm": 10.283552010641381, + "learning_rate": 1e-05, + "loss": 1.293, + "step": 2699 + }, + { + "epoch": 0.07011035889826579, + "grad_norm": 6.987790222146105, + "learning_rate": 1e-05, + "loss": 1.2184, + "step": 2700 + }, + { + "epoch": 0.07013632569785774, + "grad_norm": 6.649827894694173, + "learning_rate": 1e-05, + "loss": 1.2631, + "step": 2701 + }, + { + "epoch": 0.0701622924974497, + "grad_norm": 9.295977586019943, + "learning_rate": 1e-05, + "loss": 1.2066, + "step": 2702 + }, + { + "epoch": 0.07018825929704164, + "grad_norm": 7.549177556999794, + "learning_rate": 1e-05, + "loss": 1.2771, + "step": 2703 + }, + { + "epoch": 0.0702142260966336, + "grad_norm": 14.703934455667385, + "learning_rate": 1e-05, + "loss": 1.2463, + "step": 2704 + }, + { + "epoch": 0.07024019289622555, + "grad_norm": 6.959305868503462, + "learning_rate": 1e-05, + "loss": 1.2394, + "step": 2705 + }, + { + "epoch": 0.0702661596958175, + "grad_norm": 13.804043075295345, + "learning_rate": 1e-05, + "loss": 1.2683, + "step": 2706 + }, + { + "epoch": 0.07029212649540945, + "grad_norm": 13.841163410834715, + "learning_rate": 1e-05, + "loss": 1.2961, + "step": 2707 + }, + { + "epoch": 0.0703180932950014, + "grad_norm": 11.2523946968445, + "learning_rate": 1e-05, + "loss": 1.2374, + "step": 2708 + }, + { + "epoch": 0.07034406009459335, + "grad_norm": 5.629736876767487, + "learning_rate": 1e-05, + "loss": 1.2437, + "step": 2709 + }, + { + "epoch": 0.0703700268941853, + "grad_norm": 12.144860430968127, + "learning_rate": 1e-05, + "loss": 1.2947, + "step": 2710 + }, + { + "epoch": 0.07039599369377725, + "grad_norm": 4.283496350898315, + "learning_rate": 1e-05, + "loss": 1.2285, + "step": 2711 + }, + { + "epoch": 0.0704219604933692, + "grad_norm": 6.992125578025394, + "learning_rate": 1e-05, + "loss": 1.2701, + "step": 2712 + }, + { + "epoch": 0.07044792729296115, + "grad_norm": 7.4307084049020276, + "learning_rate": 1e-05, + "loss": 1.281, + "step": 2713 + }, + { + "epoch": 0.0704738940925531, + "grad_norm": 19.64980891088702, + "learning_rate": 1e-05, + "loss": 1.2913, + "step": 2714 + }, + { + "epoch": 0.07049986089214505, + "grad_norm": 6.650250377100474, + "learning_rate": 1e-05, + "loss": 1.2574, + "step": 2715 + }, + { + "epoch": 0.070525827691737, + "grad_norm": 6.688222970832264, + "learning_rate": 1e-05, + "loss": 1.2699, + "step": 2716 + }, + { + "epoch": 0.07055179449132894, + "grad_norm": 7.611832051655973, + "learning_rate": 1e-05, + "loss": 1.2712, + "step": 2717 + }, + { + "epoch": 0.07057776129092089, + "grad_norm": 7.236134127797696, + "learning_rate": 1e-05, + "loss": 1.2999, + "step": 2718 + }, + { + "epoch": 0.07060372809051284, + "grad_norm": 15.065085541349045, + "learning_rate": 1e-05, + "loss": 1.2737, + "step": 2719 + }, + { + "epoch": 0.07062969489010479, + "grad_norm": 9.070811281774652, + "learning_rate": 1e-05, + "loss": 1.2705, + "step": 2720 + }, + { + "epoch": 0.07065566168969674, + "grad_norm": 6.248107318400978, + "learning_rate": 1e-05, + "loss": 1.2589, + "step": 2721 + }, + { + "epoch": 0.07068162848928869, + "grad_norm": 6.8717030335854234, + "learning_rate": 1e-05, + "loss": 1.2971, + "step": 2722 + }, + { + "epoch": 0.07070759528888064, + "grad_norm": 6.349181307183184, + "learning_rate": 1e-05, + "loss": 1.2969, + "step": 2723 + }, + { + "epoch": 0.07073356208847259, + "grad_norm": 13.359152844461347, + "learning_rate": 1e-05, + "loss": 1.2989, + "step": 2724 + }, + { + "epoch": 0.07075952888806454, + "grad_norm": 16.310805894130908, + "learning_rate": 1e-05, + "loss": 1.2565, + "step": 2725 + }, + { + "epoch": 0.07078549568765649, + "grad_norm": 7.202769372067751, + "learning_rate": 1e-05, + "loss": 1.2455, + "step": 2726 + }, + { + "epoch": 0.07081146248724844, + "grad_norm": 9.449211079601062, + "learning_rate": 1e-05, + "loss": 1.2518, + "step": 2727 + }, + { + "epoch": 0.07083742928684039, + "grad_norm": 10.852061315816737, + "learning_rate": 1e-05, + "loss": 1.2475, + "step": 2728 + }, + { + "epoch": 0.07086339608643234, + "grad_norm": 5.844965354483907, + "learning_rate": 1e-05, + "loss": 1.2375, + "step": 2729 + }, + { + "epoch": 0.07088936288602429, + "grad_norm": 9.948946136945352, + "learning_rate": 1e-05, + "loss": 1.2863, + "step": 2730 + }, + { + "epoch": 0.07091532968561624, + "grad_norm": 4.9117979523547035, + "learning_rate": 1e-05, + "loss": 1.2346, + "step": 2731 + }, + { + "epoch": 0.0709412964852082, + "grad_norm": 10.99985122580056, + "learning_rate": 1e-05, + "loss": 1.2779, + "step": 2732 + }, + { + "epoch": 0.07096726328480014, + "grad_norm": 6.669435148785647, + "learning_rate": 1e-05, + "loss": 1.2456, + "step": 2733 + }, + { + "epoch": 0.0709932300843921, + "grad_norm": 13.035512688975954, + "learning_rate": 1e-05, + "loss": 1.2672, + "step": 2734 + }, + { + "epoch": 0.07101919688398405, + "grad_norm": 8.433267619765308, + "learning_rate": 1e-05, + "loss": 1.2549, + "step": 2735 + }, + { + "epoch": 0.071045163683576, + "grad_norm": 22.472251905692286, + "learning_rate": 1e-05, + "loss": 1.2823, + "step": 2736 + }, + { + "epoch": 0.07107113048316795, + "grad_norm": 9.040532288535813, + "learning_rate": 1e-05, + "loss": 1.2502, + "step": 2737 + }, + { + "epoch": 0.0710970972827599, + "grad_norm": 8.496326325839672, + "learning_rate": 1e-05, + "loss": 1.2846, + "step": 2738 + }, + { + "epoch": 0.07112306408235185, + "grad_norm": 8.010331158787226, + "learning_rate": 1e-05, + "loss": 1.2596, + "step": 2739 + }, + { + "epoch": 0.0711490308819438, + "grad_norm": 11.173687981171925, + "learning_rate": 1e-05, + "loss": 1.2947, + "step": 2740 + }, + { + "epoch": 0.07117499768153575, + "grad_norm": 9.331745103398651, + "learning_rate": 1e-05, + "loss": 1.2635, + "step": 2741 + }, + { + "epoch": 0.0712009644811277, + "grad_norm": 7.452808941231618, + "learning_rate": 1e-05, + "loss": 1.2618, + "step": 2742 + }, + { + "epoch": 0.07122693128071965, + "grad_norm": 5.517182819080339, + "learning_rate": 1e-05, + "loss": 1.2606, + "step": 2743 + }, + { + "epoch": 0.0712528980803116, + "grad_norm": 10.83479119297241, + "learning_rate": 1e-05, + "loss": 1.2458, + "step": 2744 + }, + { + "epoch": 0.07127886487990355, + "grad_norm": 7.713168578018988, + "learning_rate": 1e-05, + "loss": 1.3356, + "step": 2745 + }, + { + "epoch": 0.0713048316794955, + "grad_norm": 10.528484400171205, + "learning_rate": 1e-05, + "loss": 1.25, + "step": 2746 + }, + { + "epoch": 0.07133079847908745, + "grad_norm": 7.879556215831401, + "learning_rate": 1e-05, + "loss": 1.282, + "step": 2747 + }, + { + "epoch": 0.0713567652786794, + "grad_norm": 5.951406796622687, + "learning_rate": 1e-05, + "loss": 1.3102, + "step": 2748 + }, + { + "epoch": 0.07138273207827135, + "grad_norm": 15.628833025946937, + "learning_rate": 1e-05, + "loss": 1.2715, + "step": 2749 + }, + { + "epoch": 0.0714086988778633, + "grad_norm": 6.890383623181613, + "learning_rate": 1e-05, + "loss": 1.2632, + "step": 2750 + }, + { + "epoch": 0.07143466567745525, + "grad_norm": 12.515481745629902, + "learning_rate": 1e-05, + "loss": 1.2547, + "step": 2751 + }, + { + "epoch": 0.0714606324770472, + "grad_norm": 6.968745347094159, + "learning_rate": 1e-05, + "loss": 1.2737, + "step": 2752 + }, + { + "epoch": 0.07148659927663915, + "grad_norm": 6.809214508420166, + "learning_rate": 1e-05, + "loss": 1.2352, + "step": 2753 + }, + { + "epoch": 0.0715125660762311, + "grad_norm": 7.8264655450111125, + "learning_rate": 1e-05, + "loss": 1.2278, + "step": 2754 + }, + { + "epoch": 0.07153853287582305, + "grad_norm": 6.72867578501329, + "learning_rate": 1e-05, + "loss": 1.2907, + "step": 2755 + }, + { + "epoch": 0.071564499675415, + "grad_norm": 8.428701376663707, + "learning_rate": 1e-05, + "loss": 1.2895, + "step": 2756 + }, + { + "epoch": 0.07159046647500696, + "grad_norm": 7.536165145165998, + "learning_rate": 1e-05, + "loss": 1.2846, + "step": 2757 + }, + { + "epoch": 0.0716164332745989, + "grad_norm": 7.9518929297225505, + "learning_rate": 1e-05, + "loss": 1.2567, + "step": 2758 + }, + { + "epoch": 0.07164240007419086, + "grad_norm": 6.223200906503427, + "learning_rate": 1e-05, + "loss": 1.2932, + "step": 2759 + }, + { + "epoch": 0.07166836687378281, + "grad_norm": 7.593418694949373, + "learning_rate": 1e-05, + "loss": 1.2685, + "step": 2760 + }, + { + "epoch": 0.07169433367337476, + "grad_norm": 10.56630119229196, + "learning_rate": 1e-05, + "loss": 1.2346, + "step": 2761 + }, + { + "epoch": 0.07172030047296671, + "grad_norm": 12.212624343660742, + "learning_rate": 1e-05, + "loss": 1.2822, + "step": 2762 + }, + { + "epoch": 0.07174626727255866, + "grad_norm": 6.627118041990837, + "learning_rate": 1e-05, + "loss": 1.3062, + "step": 2763 + }, + { + "epoch": 0.07177223407215061, + "grad_norm": 7.937883262692812, + "learning_rate": 1e-05, + "loss": 1.2469, + "step": 2764 + }, + { + "epoch": 0.07179820087174256, + "grad_norm": 5.596444104517639, + "learning_rate": 1e-05, + "loss": 1.2523, + "step": 2765 + }, + { + "epoch": 0.07182416767133451, + "grad_norm": 6.727538708108026, + "learning_rate": 1e-05, + "loss": 1.2264, + "step": 2766 + }, + { + "epoch": 0.07185013447092646, + "grad_norm": 6.802321923955739, + "learning_rate": 1e-05, + "loss": 1.2659, + "step": 2767 + }, + { + "epoch": 0.07187610127051841, + "grad_norm": 5.156344881051755, + "learning_rate": 1e-05, + "loss": 1.3262, + "step": 2768 + }, + { + "epoch": 0.07190206807011036, + "grad_norm": 6.393864563692126, + "learning_rate": 1e-05, + "loss": 1.2514, + "step": 2769 + }, + { + "epoch": 0.07192803486970231, + "grad_norm": 9.955128514501174, + "learning_rate": 1e-05, + "loss": 1.2133, + "step": 2770 + }, + { + "epoch": 0.07195400166929426, + "grad_norm": 7.939214326052058, + "learning_rate": 1e-05, + "loss": 1.3213, + "step": 2771 + }, + { + "epoch": 0.07197996846888621, + "grad_norm": 5.7947596060234945, + "learning_rate": 1e-05, + "loss": 1.2872, + "step": 2772 + }, + { + "epoch": 0.07200593526847816, + "grad_norm": 9.669383971544422, + "learning_rate": 1e-05, + "loss": 1.2864, + "step": 2773 + }, + { + "epoch": 0.07203190206807011, + "grad_norm": 9.11516385800075, + "learning_rate": 1e-05, + "loss": 1.2475, + "step": 2774 + }, + { + "epoch": 0.07205786886766206, + "grad_norm": 5.34449629006162, + "learning_rate": 1e-05, + "loss": 1.2854, + "step": 2775 + }, + { + "epoch": 0.07208383566725401, + "grad_norm": 22.606716461051803, + "learning_rate": 1e-05, + "loss": 1.2757, + "step": 2776 + }, + { + "epoch": 0.07210980246684597, + "grad_norm": 7.712322817951218, + "learning_rate": 1e-05, + "loss": 1.2692, + "step": 2777 + }, + { + "epoch": 0.07213576926643792, + "grad_norm": 5.438129958569312, + "learning_rate": 1e-05, + "loss": 1.2586, + "step": 2778 + }, + { + "epoch": 0.07216173606602987, + "grad_norm": 9.735994501242422, + "learning_rate": 1e-05, + "loss": 1.2679, + "step": 2779 + }, + { + "epoch": 0.07218770286562182, + "grad_norm": 8.746487593717308, + "learning_rate": 1e-05, + "loss": 1.2673, + "step": 2780 + }, + { + "epoch": 0.07221366966521377, + "grad_norm": 7.353336391823007, + "learning_rate": 1e-05, + "loss": 1.2125, + "step": 2781 + }, + { + "epoch": 0.07223963646480572, + "grad_norm": 8.775019597306063, + "learning_rate": 1e-05, + "loss": 1.2527, + "step": 2782 + }, + { + "epoch": 0.07226560326439767, + "grad_norm": 8.747860892265589, + "learning_rate": 1e-05, + "loss": 1.3038, + "step": 2783 + }, + { + "epoch": 0.07229157006398962, + "grad_norm": 8.785186645400954, + "learning_rate": 1e-05, + "loss": 1.2914, + "step": 2784 + }, + { + "epoch": 0.07231753686358157, + "grad_norm": 6.284866430643167, + "learning_rate": 1e-05, + "loss": 1.2726, + "step": 2785 + }, + { + "epoch": 0.07234350366317352, + "grad_norm": 21.87060363187665, + "learning_rate": 1e-05, + "loss": 1.253, + "step": 2786 + }, + { + "epoch": 0.07236947046276547, + "grad_norm": 4.99028482728899, + "learning_rate": 1e-05, + "loss": 1.2889, + "step": 2787 + }, + { + "epoch": 0.07239543726235742, + "grad_norm": 7.633013157466752, + "learning_rate": 1e-05, + "loss": 1.2878, + "step": 2788 + }, + { + "epoch": 0.07242140406194937, + "grad_norm": 5.480619443424491, + "learning_rate": 1e-05, + "loss": 1.2781, + "step": 2789 + }, + { + "epoch": 0.07244737086154132, + "grad_norm": 5.406939512173746, + "learning_rate": 1e-05, + "loss": 1.2305, + "step": 2790 + }, + { + "epoch": 0.07247333766113327, + "grad_norm": 6.479874485354424, + "learning_rate": 1e-05, + "loss": 1.2771, + "step": 2791 + }, + { + "epoch": 0.07249930446072522, + "grad_norm": 12.97142585816784, + "learning_rate": 1e-05, + "loss": 1.2589, + "step": 2792 + }, + { + "epoch": 0.07252527126031716, + "grad_norm": 5.303075351606316, + "learning_rate": 1e-05, + "loss": 1.2993, + "step": 2793 + }, + { + "epoch": 0.07255123805990911, + "grad_norm": 5.963164747117887, + "learning_rate": 1e-05, + "loss": 1.2398, + "step": 2794 + }, + { + "epoch": 0.07257720485950106, + "grad_norm": 5.354036836744658, + "learning_rate": 1e-05, + "loss": 1.2828, + "step": 2795 + }, + { + "epoch": 0.07260317165909301, + "grad_norm": 7.398509427036189, + "learning_rate": 1e-05, + "loss": 1.2291, + "step": 2796 + }, + { + "epoch": 0.07262913845868496, + "grad_norm": 9.652891099540584, + "learning_rate": 1e-05, + "loss": 1.2281, + "step": 2797 + }, + { + "epoch": 0.07265510525827691, + "grad_norm": 16.53639784273713, + "learning_rate": 1e-05, + "loss": 1.2864, + "step": 2798 + }, + { + "epoch": 0.07268107205786886, + "grad_norm": 18.75592191638667, + "learning_rate": 1e-05, + "loss": 1.2624, + "step": 2799 + }, + { + "epoch": 0.07270703885746081, + "grad_norm": 7.854241759517169, + "learning_rate": 1e-05, + "loss": 1.2867, + "step": 2800 + }, + { + "epoch": 0.07273300565705276, + "grad_norm": 6.543222942823528, + "learning_rate": 1e-05, + "loss": 1.2549, + "step": 2801 + }, + { + "epoch": 0.07275897245664471, + "grad_norm": 13.918586522389765, + "learning_rate": 1e-05, + "loss": 1.2753, + "step": 2802 + }, + { + "epoch": 0.07278493925623666, + "grad_norm": 6.105739594820259, + "learning_rate": 1e-05, + "loss": 1.2244, + "step": 2803 + }, + { + "epoch": 0.07281090605582861, + "grad_norm": 5.194803267453815, + "learning_rate": 1e-05, + "loss": 1.2561, + "step": 2804 + }, + { + "epoch": 0.07283687285542056, + "grad_norm": 7.6164232389029936, + "learning_rate": 1e-05, + "loss": 1.2524, + "step": 2805 + }, + { + "epoch": 0.07286283965501252, + "grad_norm": 6.4289995383961, + "learning_rate": 1e-05, + "loss": 1.2715, + "step": 2806 + }, + { + "epoch": 0.07288880645460447, + "grad_norm": 8.599607378846578, + "learning_rate": 1e-05, + "loss": 1.2757, + "step": 2807 + }, + { + "epoch": 0.07291477325419642, + "grad_norm": 11.624052275716664, + "learning_rate": 1e-05, + "loss": 1.3039, + "step": 2808 + }, + { + "epoch": 0.07294074005378837, + "grad_norm": 6.791917417925078, + "learning_rate": 1e-05, + "loss": 1.2512, + "step": 2809 + }, + { + "epoch": 0.07296670685338032, + "grad_norm": 18.576255556699593, + "learning_rate": 1e-05, + "loss": 1.2855, + "step": 2810 + }, + { + "epoch": 0.07299267365297227, + "grad_norm": 6.991283984920974, + "learning_rate": 1e-05, + "loss": 1.2983, + "step": 2811 + }, + { + "epoch": 0.07301864045256422, + "grad_norm": 6.457221478973415, + "learning_rate": 1e-05, + "loss": 1.2287, + "step": 2812 + }, + { + "epoch": 0.07304460725215617, + "grad_norm": 6.373600525675316, + "learning_rate": 1e-05, + "loss": 1.2661, + "step": 2813 + }, + { + "epoch": 0.07307057405174812, + "grad_norm": 13.906242319169227, + "learning_rate": 1e-05, + "loss": 1.2779, + "step": 2814 + }, + { + "epoch": 0.07309654085134007, + "grad_norm": 5.255561789001969, + "learning_rate": 1e-05, + "loss": 1.3044, + "step": 2815 + }, + { + "epoch": 0.07312250765093202, + "grad_norm": 6.800698233267003, + "learning_rate": 1e-05, + "loss": 1.2516, + "step": 2816 + }, + { + "epoch": 0.07314847445052397, + "grad_norm": 5.559426304975551, + "learning_rate": 1e-05, + "loss": 1.2959, + "step": 2817 + }, + { + "epoch": 0.07317444125011592, + "grad_norm": 5.258908660577114, + "learning_rate": 1e-05, + "loss": 1.3151, + "step": 2818 + }, + { + "epoch": 0.07320040804970787, + "grad_norm": 8.06987951101289, + "learning_rate": 1e-05, + "loss": 1.2898, + "step": 2819 + }, + { + "epoch": 0.07322637484929982, + "grad_norm": 6.691071884432754, + "learning_rate": 1e-05, + "loss": 1.282, + "step": 2820 + }, + { + "epoch": 0.07325234164889177, + "grad_norm": 6.071114623343508, + "learning_rate": 1e-05, + "loss": 1.2245, + "step": 2821 + }, + { + "epoch": 0.07327830844848372, + "grad_norm": 16.55110715176537, + "learning_rate": 1e-05, + "loss": 1.3083, + "step": 2822 + }, + { + "epoch": 0.07330427524807567, + "grad_norm": 7.712809512266575, + "learning_rate": 1e-05, + "loss": 1.2785, + "step": 2823 + }, + { + "epoch": 0.07333024204766762, + "grad_norm": 6.055188991328275, + "learning_rate": 1e-05, + "loss": 1.2154, + "step": 2824 + }, + { + "epoch": 0.07335620884725957, + "grad_norm": 18.38436879169749, + "learning_rate": 1e-05, + "loss": 1.2091, + "step": 2825 + }, + { + "epoch": 0.07338217564685152, + "grad_norm": 6.682180981055079, + "learning_rate": 1e-05, + "loss": 1.2057, + "step": 2826 + }, + { + "epoch": 0.07340814244644348, + "grad_norm": 6.2008000011528255, + "learning_rate": 1e-05, + "loss": 1.2722, + "step": 2827 + }, + { + "epoch": 0.07343410924603543, + "grad_norm": 9.521939292017114, + "learning_rate": 1e-05, + "loss": 1.2776, + "step": 2828 + }, + { + "epoch": 0.07346007604562738, + "grad_norm": 7.754968096767695, + "learning_rate": 1e-05, + "loss": 1.2762, + "step": 2829 + }, + { + "epoch": 0.07348604284521933, + "grad_norm": 6.7437466512677045, + "learning_rate": 1e-05, + "loss": 1.2638, + "step": 2830 + }, + { + "epoch": 0.07351200964481128, + "grad_norm": 12.250617186899644, + "learning_rate": 1e-05, + "loss": 1.2298, + "step": 2831 + }, + { + "epoch": 0.07353797644440323, + "grad_norm": 5.726098180997071, + "learning_rate": 1e-05, + "loss": 1.2618, + "step": 2832 + }, + { + "epoch": 0.07356394324399518, + "grad_norm": 10.65848748217658, + "learning_rate": 1e-05, + "loss": 1.2476, + "step": 2833 + }, + { + "epoch": 0.07358991004358713, + "grad_norm": 7.505585688560242, + "learning_rate": 1e-05, + "loss": 1.2535, + "step": 2834 + }, + { + "epoch": 0.07361587684317908, + "grad_norm": 6.258453755377851, + "learning_rate": 1e-05, + "loss": 1.3277, + "step": 2835 + }, + { + "epoch": 0.07364184364277103, + "grad_norm": 7.528873052679224, + "learning_rate": 1e-05, + "loss": 1.3025, + "step": 2836 + }, + { + "epoch": 0.07366781044236298, + "grad_norm": 12.931321295001927, + "learning_rate": 1e-05, + "loss": 1.2961, + "step": 2837 + }, + { + "epoch": 0.07369377724195493, + "grad_norm": 23.373374515257073, + "learning_rate": 1e-05, + "loss": 1.2486, + "step": 2838 + }, + { + "epoch": 0.07371974404154688, + "grad_norm": 13.443276374734696, + "learning_rate": 1e-05, + "loss": 1.2755, + "step": 2839 + }, + { + "epoch": 0.07374571084113883, + "grad_norm": 4.8922186744205955, + "learning_rate": 1e-05, + "loss": 1.2793, + "step": 2840 + }, + { + "epoch": 0.07377167764073078, + "grad_norm": 4.623834617811927, + "learning_rate": 1e-05, + "loss": 1.3093, + "step": 2841 + }, + { + "epoch": 0.07379764444032273, + "grad_norm": 5.172229766202825, + "learning_rate": 1e-05, + "loss": 1.2584, + "step": 2842 + }, + { + "epoch": 0.07382361123991468, + "grad_norm": 8.215711604924614, + "learning_rate": 1e-05, + "loss": 1.2695, + "step": 2843 + }, + { + "epoch": 0.07384957803950663, + "grad_norm": 8.868551141016232, + "learning_rate": 1e-05, + "loss": 1.2819, + "step": 2844 + }, + { + "epoch": 0.07387554483909858, + "grad_norm": 6.288927538827511, + "learning_rate": 1e-05, + "loss": 1.2419, + "step": 2845 + }, + { + "epoch": 0.07390151163869053, + "grad_norm": 6.123123504305185, + "learning_rate": 1e-05, + "loss": 1.258, + "step": 2846 + }, + { + "epoch": 0.07392747843828248, + "grad_norm": 11.949817633733671, + "learning_rate": 1e-05, + "loss": 1.3015, + "step": 2847 + }, + { + "epoch": 0.07395344523787444, + "grad_norm": 6.763401996582993, + "learning_rate": 1e-05, + "loss": 1.2894, + "step": 2848 + }, + { + "epoch": 0.07397941203746639, + "grad_norm": 5.102075618158096, + "learning_rate": 1e-05, + "loss": 1.2329, + "step": 2849 + }, + { + "epoch": 0.07400537883705834, + "grad_norm": 4.915311378713157, + "learning_rate": 1e-05, + "loss": 1.2681, + "step": 2850 + }, + { + "epoch": 0.07403134563665029, + "grad_norm": 8.946086040517157, + "learning_rate": 1e-05, + "loss": 1.2795, + "step": 2851 + }, + { + "epoch": 0.07405731243624224, + "grad_norm": 5.913334225923685, + "learning_rate": 1e-05, + "loss": 1.2094, + "step": 2852 + }, + { + "epoch": 0.07408327923583419, + "grad_norm": 6.920940534146878, + "learning_rate": 1e-05, + "loss": 1.312, + "step": 2853 + }, + { + "epoch": 0.07410924603542614, + "grad_norm": 6.423011252314858, + "learning_rate": 1e-05, + "loss": 1.3052, + "step": 2854 + }, + { + "epoch": 0.07413521283501809, + "grad_norm": 8.234337012187043, + "learning_rate": 1e-05, + "loss": 1.2914, + "step": 2855 + }, + { + "epoch": 0.07416117963461004, + "grad_norm": 5.666563631504775, + "learning_rate": 1e-05, + "loss": 1.2594, + "step": 2856 + }, + { + "epoch": 0.07418714643420199, + "grad_norm": 7.384993942351236, + "learning_rate": 1e-05, + "loss": 1.2868, + "step": 2857 + }, + { + "epoch": 0.07421311323379394, + "grad_norm": 8.619365676234217, + "learning_rate": 1e-05, + "loss": 1.3268, + "step": 2858 + }, + { + "epoch": 0.07423908003338589, + "grad_norm": 5.805571662032779, + "learning_rate": 1e-05, + "loss": 1.2596, + "step": 2859 + }, + { + "epoch": 0.07426504683297784, + "grad_norm": 8.222961507101052, + "learning_rate": 1e-05, + "loss": 1.2512, + "step": 2860 + }, + { + "epoch": 0.07429101363256979, + "grad_norm": 6.617049142591337, + "learning_rate": 1e-05, + "loss": 1.2626, + "step": 2861 + }, + { + "epoch": 0.07431698043216174, + "grad_norm": 8.661594324808718, + "learning_rate": 1e-05, + "loss": 1.2613, + "step": 2862 + }, + { + "epoch": 0.07434294723175369, + "grad_norm": 15.400021421739616, + "learning_rate": 1e-05, + "loss": 1.2127, + "step": 2863 + }, + { + "epoch": 0.07436891403134564, + "grad_norm": 6.603906700561176, + "learning_rate": 1e-05, + "loss": 1.2351, + "step": 2864 + }, + { + "epoch": 0.0743948808309376, + "grad_norm": 5.330131980593163, + "learning_rate": 1e-05, + "loss": 1.2965, + "step": 2865 + }, + { + "epoch": 0.07442084763052954, + "grad_norm": 6.6160102166278, + "learning_rate": 1e-05, + "loss": 1.3175, + "step": 2866 + }, + { + "epoch": 0.0744468144301215, + "grad_norm": 6.764492442283381, + "learning_rate": 1e-05, + "loss": 1.2674, + "step": 2867 + }, + { + "epoch": 0.07447278122971344, + "grad_norm": 6.924784787135592, + "learning_rate": 1e-05, + "loss": 1.2852, + "step": 2868 + }, + { + "epoch": 0.0744987480293054, + "grad_norm": 6.632507389091807, + "learning_rate": 1e-05, + "loss": 1.2767, + "step": 2869 + }, + { + "epoch": 0.07452471482889733, + "grad_norm": 6.452710006707574, + "learning_rate": 1e-05, + "loss": 1.2358, + "step": 2870 + }, + { + "epoch": 0.07455068162848928, + "grad_norm": 10.378497166265248, + "learning_rate": 1e-05, + "loss": 1.2775, + "step": 2871 + }, + { + "epoch": 0.07457664842808123, + "grad_norm": 7.177570948394608, + "learning_rate": 1e-05, + "loss": 1.2747, + "step": 2872 + }, + { + "epoch": 0.07460261522767318, + "grad_norm": 6.306959988236241, + "learning_rate": 1e-05, + "loss": 1.2456, + "step": 2873 + }, + { + "epoch": 0.07462858202726513, + "grad_norm": 7.335572189969099, + "learning_rate": 1e-05, + "loss": 1.2939, + "step": 2874 + }, + { + "epoch": 0.07465454882685708, + "grad_norm": 12.922990854556584, + "learning_rate": 1e-05, + "loss": 1.2418, + "step": 2875 + }, + { + "epoch": 0.07468051562644903, + "grad_norm": 6.1390158216760655, + "learning_rate": 1e-05, + "loss": 1.2592, + "step": 2876 + }, + { + "epoch": 0.07470648242604098, + "grad_norm": 11.603835004704534, + "learning_rate": 1e-05, + "loss": 1.2405, + "step": 2877 + }, + { + "epoch": 0.07473244922563294, + "grad_norm": 5.450205723790674, + "learning_rate": 1e-05, + "loss": 1.2575, + "step": 2878 + }, + { + "epoch": 0.07475841602522489, + "grad_norm": 6.230157711671887, + "learning_rate": 1e-05, + "loss": 1.2809, + "step": 2879 + }, + { + "epoch": 0.07478438282481684, + "grad_norm": 5.816498396369649, + "learning_rate": 1e-05, + "loss": 1.2659, + "step": 2880 + }, + { + "epoch": 0.07481034962440879, + "grad_norm": 6.158007066039355, + "learning_rate": 1e-05, + "loss": 1.2553, + "step": 2881 + }, + { + "epoch": 0.07483631642400074, + "grad_norm": 6.747765630445234, + "learning_rate": 1e-05, + "loss": 1.2879, + "step": 2882 + }, + { + "epoch": 0.07486228322359269, + "grad_norm": 7.064815985462556, + "learning_rate": 1e-05, + "loss": 1.2425, + "step": 2883 + }, + { + "epoch": 0.07488825002318464, + "grad_norm": 8.065605097926822, + "learning_rate": 1e-05, + "loss": 1.2296, + "step": 2884 + }, + { + "epoch": 0.07491421682277659, + "grad_norm": 10.560145369598413, + "learning_rate": 1e-05, + "loss": 1.3026, + "step": 2885 + }, + { + "epoch": 0.07494018362236854, + "grad_norm": 5.722675914535326, + "learning_rate": 1e-05, + "loss": 1.2291, + "step": 2886 + }, + { + "epoch": 0.07496615042196049, + "grad_norm": 5.779811040832097, + "learning_rate": 1e-05, + "loss": 1.3019, + "step": 2887 + }, + { + "epoch": 0.07499211722155244, + "grad_norm": 7.224889745168018, + "learning_rate": 1e-05, + "loss": 1.2895, + "step": 2888 + }, + { + "epoch": 0.07501808402114439, + "grad_norm": 38.04916734624015, + "learning_rate": 1e-05, + "loss": 1.2643, + "step": 2889 + }, + { + "epoch": 0.07504405082073634, + "grad_norm": 7.131413835520768, + "learning_rate": 1e-05, + "loss": 1.261, + "step": 2890 + }, + { + "epoch": 0.07507001762032829, + "grad_norm": 12.835460800262611, + "learning_rate": 1e-05, + "loss": 1.3027, + "step": 2891 + }, + { + "epoch": 0.07509598441992024, + "grad_norm": 8.73020352295673, + "learning_rate": 1e-05, + "loss": 1.2777, + "step": 2892 + }, + { + "epoch": 0.07512195121951219, + "grad_norm": 9.151696019813077, + "learning_rate": 1e-05, + "loss": 1.2877, + "step": 2893 + }, + { + "epoch": 0.07514791801910414, + "grad_norm": 30.85837913741228, + "learning_rate": 1e-05, + "loss": 1.2831, + "step": 2894 + }, + { + "epoch": 0.0751738848186961, + "grad_norm": 13.678525168352397, + "learning_rate": 1e-05, + "loss": 1.2446, + "step": 2895 + }, + { + "epoch": 0.07519985161828804, + "grad_norm": 10.039027161294806, + "learning_rate": 1e-05, + "loss": 1.2804, + "step": 2896 + }, + { + "epoch": 0.07522581841788, + "grad_norm": 6.142615696924044, + "learning_rate": 1e-05, + "loss": 1.3226, + "step": 2897 + }, + { + "epoch": 0.07525178521747194, + "grad_norm": 8.698107932418631, + "learning_rate": 1e-05, + "loss": 1.2314, + "step": 2898 + }, + { + "epoch": 0.0752777520170639, + "grad_norm": 6.244344364931483, + "learning_rate": 1e-05, + "loss": 1.2163, + "step": 2899 + }, + { + "epoch": 0.07530371881665585, + "grad_norm": 5.4200214658642905, + "learning_rate": 1e-05, + "loss": 1.2451, + "step": 2900 + }, + { + "epoch": 0.0753296856162478, + "grad_norm": 7.924193274415571, + "learning_rate": 1e-05, + "loss": 1.239, + "step": 2901 + }, + { + "epoch": 0.07535565241583975, + "grad_norm": 6.8783841213862065, + "learning_rate": 1e-05, + "loss": 1.2434, + "step": 2902 + }, + { + "epoch": 0.0753816192154317, + "grad_norm": 8.597749868157956, + "learning_rate": 1e-05, + "loss": 1.2752, + "step": 2903 + }, + { + "epoch": 0.07540758601502365, + "grad_norm": 8.821815274854282, + "learning_rate": 1e-05, + "loss": 1.2427, + "step": 2904 + }, + { + "epoch": 0.0754335528146156, + "grad_norm": 9.888600315450638, + "learning_rate": 1e-05, + "loss": 1.2777, + "step": 2905 + }, + { + "epoch": 0.07545951961420755, + "grad_norm": 7.109709813544607, + "learning_rate": 1e-05, + "loss": 1.2826, + "step": 2906 + }, + { + "epoch": 0.0754854864137995, + "grad_norm": 8.228821823685152, + "learning_rate": 1e-05, + "loss": 1.2761, + "step": 2907 + }, + { + "epoch": 0.07551145321339145, + "grad_norm": 8.005357855494735, + "learning_rate": 1e-05, + "loss": 1.2809, + "step": 2908 + }, + { + "epoch": 0.0755374200129834, + "grad_norm": 7.444518552319558, + "learning_rate": 1e-05, + "loss": 1.2652, + "step": 2909 + }, + { + "epoch": 0.07556338681257535, + "grad_norm": 13.362778828979643, + "learning_rate": 1e-05, + "loss": 1.2982, + "step": 2910 + }, + { + "epoch": 0.0755893536121673, + "grad_norm": 45.3487785805748, + "learning_rate": 1e-05, + "loss": 1.273, + "step": 2911 + }, + { + "epoch": 0.07561532041175925, + "grad_norm": 7.099275159774373, + "learning_rate": 1e-05, + "loss": 1.2432, + "step": 2912 + }, + { + "epoch": 0.0756412872113512, + "grad_norm": 21.083686835701613, + "learning_rate": 1e-05, + "loss": 1.2453, + "step": 2913 + }, + { + "epoch": 0.07566725401094315, + "grad_norm": 8.944801176024724, + "learning_rate": 1e-05, + "loss": 1.2782, + "step": 2914 + }, + { + "epoch": 0.0756932208105351, + "grad_norm": 9.109879599151416, + "learning_rate": 1e-05, + "loss": 1.2283, + "step": 2915 + }, + { + "epoch": 0.07571918761012705, + "grad_norm": 10.037052936027836, + "learning_rate": 1e-05, + "loss": 1.269, + "step": 2916 + }, + { + "epoch": 0.075745154409719, + "grad_norm": 8.299652037163261, + "learning_rate": 1e-05, + "loss": 1.2354, + "step": 2917 + }, + { + "epoch": 0.07577112120931095, + "grad_norm": 8.5627147341345, + "learning_rate": 1e-05, + "loss": 1.2436, + "step": 2918 + }, + { + "epoch": 0.0757970880089029, + "grad_norm": 37.73112854823504, + "learning_rate": 1e-05, + "loss": 1.3093, + "step": 2919 + }, + { + "epoch": 0.07582305480849486, + "grad_norm": 6.82168715563929, + "learning_rate": 1e-05, + "loss": 1.2332, + "step": 2920 + }, + { + "epoch": 0.0758490216080868, + "grad_norm": 8.85199431454441, + "learning_rate": 1e-05, + "loss": 1.2314, + "step": 2921 + }, + { + "epoch": 0.07587498840767876, + "grad_norm": 9.69565676954744, + "learning_rate": 1e-05, + "loss": 1.2885, + "step": 2922 + }, + { + "epoch": 0.0759009552072707, + "grad_norm": 14.319718452115305, + "learning_rate": 1e-05, + "loss": 1.3014, + "step": 2923 + }, + { + "epoch": 0.07592692200686266, + "grad_norm": 14.907148226180107, + "learning_rate": 1e-05, + "loss": 1.2406, + "step": 2924 + }, + { + "epoch": 0.07595288880645461, + "grad_norm": 7.836740469747965, + "learning_rate": 1e-05, + "loss": 1.2774, + "step": 2925 + }, + { + "epoch": 0.07597885560604656, + "grad_norm": 10.574481015287502, + "learning_rate": 1e-05, + "loss": 1.2802, + "step": 2926 + }, + { + "epoch": 0.07600482240563851, + "grad_norm": 13.562620944701228, + "learning_rate": 1e-05, + "loss": 1.2815, + "step": 2927 + }, + { + "epoch": 0.07603078920523046, + "grad_norm": 8.254922958559808, + "learning_rate": 1e-05, + "loss": 1.2989, + "step": 2928 + }, + { + "epoch": 0.07605675600482241, + "grad_norm": 10.407373499434826, + "learning_rate": 1e-05, + "loss": 1.2761, + "step": 2929 + }, + { + "epoch": 0.07608272280441436, + "grad_norm": 26.152378465716318, + "learning_rate": 1e-05, + "loss": 1.2908, + "step": 2930 + }, + { + "epoch": 0.07610868960400631, + "grad_norm": 38.78489639636236, + "learning_rate": 1e-05, + "loss": 1.2945, + "step": 2931 + }, + { + "epoch": 0.07613465640359826, + "grad_norm": 15.260154970114623, + "learning_rate": 1e-05, + "loss": 1.2342, + "step": 2932 + }, + { + "epoch": 0.07616062320319021, + "grad_norm": 6.988407617900916, + "learning_rate": 1e-05, + "loss": 1.2246, + "step": 2933 + }, + { + "epoch": 0.07618659000278216, + "grad_norm": 5.899541041739927, + "learning_rate": 1e-05, + "loss": 1.2922, + "step": 2934 + }, + { + "epoch": 0.07621255680237411, + "grad_norm": 34.51293172210509, + "learning_rate": 1e-05, + "loss": 1.28, + "step": 2935 + }, + { + "epoch": 0.07623852360196606, + "grad_norm": 8.867966564757241, + "learning_rate": 1e-05, + "loss": 1.2832, + "step": 2936 + }, + { + "epoch": 0.07626449040155801, + "grad_norm": 8.27756772496645, + "learning_rate": 1e-05, + "loss": 1.288, + "step": 2937 + }, + { + "epoch": 0.07629045720114996, + "grad_norm": 8.705081849828124, + "learning_rate": 1e-05, + "loss": 1.2926, + "step": 2938 + }, + { + "epoch": 0.07631642400074191, + "grad_norm": 10.711148338552162, + "learning_rate": 1e-05, + "loss": 1.3032, + "step": 2939 + }, + { + "epoch": 0.07634239080033386, + "grad_norm": 6.2628188853442825, + "learning_rate": 1e-05, + "loss": 1.3123, + "step": 2940 + }, + { + "epoch": 0.07636835759992582, + "grad_norm": 5.1842999986922464, + "learning_rate": 1e-05, + "loss": 1.2877, + "step": 2941 + }, + { + "epoch": 0.07639432439951777, + "grad_norm": 8.298085040057812, + "learning_rate": 1e-05, + "loss": 1.2489, + "step": 2942 + }, + { + "epoch": 0.07642029119910972, + "grad_norm": 6.152320808488972, + "learning_rate": 1e-05, + "loss": 1.2716, + "step": 2943 + }, + { + "epoch": 0.07644625799870167, + "grad_norm": 10.965025005738237, + "learning_rate": 1e-05, + "loss": 1.2546, + "step": 2944 + }, + { + "epoch": 0.07647222479829362, + "grad_norm": 9.236989950054204, + "learning_rate": 1e-05, + "loss": 1.2608, + "step": 2945 + }, + { + "epoch": 0.07649819159788555, + "grad_norm": 9.049999376159983, + "learning_rate": 1e-05, + "loss": 1.286, + "step": 2946 + }, + { + "epoch": 0.0765241583974775, + "grad_norm": 8.921413215939012, + "learning_rate": 1e-05, + "loss": 1.2565, + "step": 2947 + }, + { + "epoch": 0.07655012519706945, + "grad_norm": 13.528425454874116, + "learning_rate": 1e-05, + "loss": 1.2219, + "step": 2948 + }, + { + "epoch": 0.0765760919966614, + "grad_norm": 6.471192714178969, + "learning_rate": 1e-05, + "loss": 1.1972, + "step": 2949 + }, + { + "epoch": 0.07660205879625336, + "grad_norm": 7.6430829858962595, + "learning_rate": 1e-05, + "loss": 1.2498, + "step": 2950 + }, + { + "epoch": 0.0766280255958453, + "grad_norm": 7.976855176437603, + "learning_rate": 1e-05, + "loss": 1.2889, + "step": 2951 + }, + { + "epoch": 0.07665399239543726, + "grad_norm": 9.10515495424008, + "learning_rate": 1e-05, + "loss": 1.2631, + "step": 2952 + }, + { + "epoch": 0.07667995919502921, + "grad_norm": 35.90384973482997, + "learning_rate": 1e-05, + "loss": 1.272, + "step": 2953 + }, + { + "epoch": 0.07670592599462116, + "grad_norm": 7.4907573969936605, + "learning_rate": 1e-05, + "loss": 1.3077, + "step": 2954 + }, + { + "epoch": 0.07673189279421311, + "grad_norm": 6.531633721719784, + "learning_rate": 1e-05, + "loss": 1.2806, + "step": 2955 + }, + { + "epoch": 0.07675785959380506, + "grad_norm": 5.831561918594399, + "learning_rate": 1e-05, + "loss": 1.2882, + "step": 2956 + }, + { + "epoch": 0.07678382639339701, + "grad_norm": 9.075821129506943, + "learning_rate": 1e-05, + "loss": 1.3052, + "step": 2957 + }, + { + "epoch": 0.07680979319298896, + "grad_norm": 24.453687842963355, + "learning_rate": 1e-05, + "loss": 1.2477, + "step": 2958 + }, + { + "epoch": 0.07683575999258091, + "grad_norm": 12.188183574579776, + "learning_rate": 1e-05, + "loss": 1.2498, + "step": 2959 + }, + { + "epoch": 0.07686172679217286, + "grad_norm": 5.0290497891367325, + "learning_rate": 1e-05, + "loss": 1.2377, + "step": 2960 + }, + { + "epoch": 0.07688769359176481, + "grad_norm": 5.356776194059007, + "learning_rate": 1e-05, + "loss": 1.293, + "step": 2961 + }, + { + "epoch": 0.07691366039135676, + "grad_norm": 5.591319147369362, + "learning_rate": 1e-05, + "loss": 1.2823, + "step": 2962 + }, + { + "epoch": 0.07693962719094871, + "grad_norm": 20.578618836614286, + "learning_rate": 1e-05, + "loss": 1.2785, + "step": 2963 + }, + { + "epoch": 0.07696559399054066, + "grad_norm": 6.242676678755557, + "learning_rate": 1e-05, + "loss": 1.2885, + "step": 2964 + }, + { + "epoch": 0.07699156079013261, + "grad_norm": 121.15022473410852, + "learning_rate": 1e-05, + "loss": 1.2736, + "step": 2965 + }, + { + "epoch": 0.07701752758972456, + "grad_norm": 5.782789200718059, + "learning_rate": 1e-05, + "loss": 1.2856, + "step": 2966 + }, + { + "epoch": 0.07704349438931651, + "grad_norm": 5.645898778707353, + "learning_rate": 1e-05, + "loss": 1.2901, + "step": 2967 + }, + { + "epoch": 0.07706946118890846, + "grad_norm": 5.105292412792551, + "learning_rate": 1e-05, + "loss": 1.3347, + "step": 2968 + }, + { + "epoch": 0.07709542798850041, + "grad_norm": 40.61714918131487, + "learning_rate": 1e-05, + "loss": 1.2803, + "step": 2969 + }, + { + "epoch": 0.07712139478809237, + "grad_norm": 6.51956284631349, + "learning_rate": 1e-05, + "loss": 1.2568, + "step": 2970 + }, + { + "epoch": 0.07714736158768432, + "grad_norm": 6.571693012295352, + "learning_rate": 1e-05, + "loss": 1.2461, + "step": 2971 + }, + { + "epoch": 0.07717332838727627, + "grad_norm": 8.125304231083058, + "learning_rate": 1e-05, + "loss": 1.2443, + "step": 2972 + }, + { + "epoch": 0.07719929518686822, + "grad_norm": 8.948037930881831, + "learning_rate": 1e-05, + "loss": 1.2754, + "step": 2973 + }, + { + "epoch": 0.07722526198646017, + "grad_norm": 6.231599194567035, + "learning_rate": 1e-05, + "loss": 1.2825, + "step": 2974 + }, + { + "epoch": 0.07725122878605212, + "grad_norm": 8.665823039540832, + "learning_rate": 1e-05, + "loss": 1.2556, + "step": 2975 + }, + { + "epoch": 0.07727719558564407, + "grad_norm": 5.347243388520772, + "learning_rate": 1e-05, + "loss": 1.2763, + "step": 2976 + }, + { + "epoch": 0.07730316238523602, + "grad_norm": 6.327827804556718, + "learning_rate": 1e-05, + "loss": 1.2389, + "step": 2977 + }, + { + "epoch": 0.07732912918482797, + "grad_norm": 9.572148261433647, + "learning_rate": 1e-05, + "loss": 1.2771, + "step": 2978 + }, + { + "epoch": 0.07735509598441992, + "grad_norm": 5.294963840724973, + "learning_rate": 1e-05, + "loss": 1.263, + "step": 2979 + }, + { + "epoch": 0.07738106278401187, + "grad_norm": 5.310773860770588, + "learning_rate": 1e-05, + "loss": 1.2533, + "step": 2980 + }, + { + "epoch": 0.07740702958360382, + "grad_norm": 6.566151211770853, + "learning_rate": 1e-05, + "loss": 1.2566, + "step": 2981 + }, + { + "epoch": 0.07743299638319577, + "grad_norm": 6.017794289772098, + "learning_rate": 1e-05, + "loss": 1.2725, + "step": 2982 + }, + { + "epoch": 0.07745896318278772, + "grad_norm": 6.413429665796999, + "learning_rate": 1e-05, + "loss": 1.26, + "step": 2983 + }, + { + "epoch": 0.07748492998237967, + "grad_norm": 5.6680808079248495, + "learning_rate": 1e-05, + "loss": 1.313, + "step": 2984 + }, + { + "epoch": 0.07751089678197162, + "grad_norm": 7.9203962943603266, + "learning_rate": 1e-05, + "loss": 1.3101, + "step": 2985 + }, + { + "epoch": 0.07753686358156357, + "grad_norm": 7.1697683793328535, + "learning_rate": 1e-05, + "loss": 1.2567, + "step": 2986 + }, + { + "epoch": 0.07756283038115552, + "grad_norm": 4.714962481352479, + "learning_rate": 1e-05, + "loss": 1.302, + "step": 2987 + }, + { + "epoch": 0.07758879718074747, + "grad_norm": 4.9813866341684685, + "learning_rate": 1e-05, + "loss": 1.2418, + "step": 2988 + }, + { + "epoch": 0.07761476398033942, + "grad_norm": 12.428408880063412, + "learning_rate": 1e-05, + "loss": 1.2637, + "step": 2989 + }, + { + "epoch": 0.07764073077993137, + "grad_norm": 9.871974046867791, + "learning_rate": 1e-05, + "loss": 1.2833, + "step": 2990 + }, + { + "epoch": 0.07766669757952332, + "grad_norm": 7.858978200796571, + "learning_rate": 1e-05, + "loss": 1.3028, + "step": 2991 + }, + { + "epoch": 0.07769266437911528, + "grad_norm": 6.68428233092686, + "learning_rate": 1e-05, + "loss": 1.2763, + "step": 2992 + }, + { + "epoch": 0.07771863117870723, + "grad_norm": 4.8345041939573905, + "learning_rate": 1e-05, + "loss": 1.2575, + "step": 2993 + }, + { + "epoch": 0.07774459797829918, + "grad_norm": 9.35937866822236, + "learning_rate": 1e-05, + "loss": 1.2701, + "step": 2994 + }, + { + "epoch": 0.07777056477789113, + "grad_norm": 8.598345717186861, + "learning_rate": 1e-05, + "loss": 1.2939, + "step": 2995 + }, + { + "epoch": 0.07779653157748308, + "grad_norm": 6.871540100624618, + "learning_rate": 1e-05, + "loss": 1.213, + "step": 2996 + }, + { + "epoch": 0.07782249837707503, + "grad_norm": 11.327631320061673, + "learning_rate": 1e-05, + "loss": 1.2512, + "step": 2997 + }, + { + "epoch": 0.07784846517666698, + "grad_norm": 10.476363325963721, + "learning_rate": 1e-05, + "loss": 1.181, + "step": 2998 + }, + { + "epoch": 0.07787443197625893, + "grad_norm": 5.525124126438141, + "learning_rate": 1e-05, + "loss": 1.2392, + "step": 2999 + }, + { + "epoch": 0.07790039877585088, + "grad_norm": 8.017536017322435, + "learning_rate": 1e-05, + "loss": 1.2553, + "step": 3000 + } + ], + "logging_steps": 1.0, + "max_steps": 38510, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 50, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.2253061281352253e+18, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}