{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 6807, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00044072278536800354, "grad_norm": 0.24354467550812778, "learning_rate": 2.936857562408223e-07, "loss": 2.1339, "step": 1 }, { "epoch": 0.0022036139268400176, "grad_norm": 0.27040776216996587, "learning_rate": 1.4684287812041115e-06, "loss": 2.3535, "step": 5 }, { "epoch": 0.004407227853680035, "grad_norm": 0.23978713638500837, "learning_rate": 2.936857562408223e-06, "loss": 2.0659, "step": 10 }, { "epoch": 0.006610841780520053, "grad_norm": 0.2562898248528363, "learning_rate": 4.4052863436123355e-06, "loss": 2.0742, "step": 15 }, { "epoch": 0.00881445570736007, "grad_norm": 0.2671940707327157, "learning_rate": 5.873715124816446e-06, "loss": 2.0397, "step": 20 }, { "epoch": 0.011018069634200088, "grad_norm": 0.2840547143688825, "learning_rate": 7.3421439060205585e-06, "loss": 2.3034, "step": 25 }, { "epoch": 0.013221683561040106, "grad_norm": 0.2739646963997833, "learning_rate": 8.810572687224671e-06, "loss": 2.0168, "step": 30 }, { "epoch": 0.015425297487880123, "grad_norm": 0.36046684768779125, "learning_rate": 1.0279001468428782e-05, "loss": 2.1073, "step": 35 }, { "epoch": 0.01762891141472014, "grad_norm": 0.29214076993878346, "learning_rate": 1.1747430249632892e-05, "loss": 1.9053, "step": 40 }, { "epoch": 0.01983252534156016, "grad_norm": 0.4413639637555924, "learning_rate": 1.3215859030837005e-05, "loss": 2.272, "step": 45 }, { "epoch": 0.022036139268400177, "grad_norm": 0.3645215252816923, "learning_rate": 1.4684287812041117e-05, "loss": 2.0033, "step": 50 }, { "epoch": 0.024239753195240195, "grad_norm": 0.26241461964383006, "learning_rate": 1.615271659324523e-05, "loss": 2.0168, "step": 55 }, { "epoch": 0.026443367122080213, "grad_norm": 0.3108898378679939, "learning_rate": 1.7621145374449342e-05, "loss": 2.1111, "step": 60 }, { "epoch": 0.02864698104892023, "grad_norm": 0.2662326669897319, "learning_rate": 1.9089574155653454e-05, "loss": 1.8806, "step": 65 }, { "epoch": 0.030850594975760245, "grad_norm": 0.16963519927101947, "learning_rate": 2.0558002936857563e-05, "loss": 2.1257, "step": 70 }, { "epoch": 0.03305420890260027, "grad_norm": 0.23310716143802493, "learning_rate": 2.2026431718061676e-05, "loss": 1.8759, "step": 75 }, { "epoch": 0.03525782282944028, "grad_norm": 0.2575993518919339, "learning_rate": 2.3494860499265785e-05, "loss": 1.9053, "step": 80 }, { "epoch": 0.0374614367562803, "grad_norm": 0.2746073869104198, "learning_rate": 2.4963289280469897e-05, "loss": 1.7651, "step": 85 }, { "epoch": 0.03966505068312032, "grad_norm": 0.2279037029835098, "learning_rate": 2.643171806167401e-05, "loss": 1.8627, "step": 90 }, { "epoch": 0.04186866460996033, "grad_norm": 0.27296809706241965, "learning_rate": 2.7900146842878122e-05, "loss": 1.9057, "step": 95 }, { "epoch": 0.044072278536800354, "grad_norm": 0.2670117144218064, "learning_rate": 2.9368575624082234e-05, "loss": 1.8911, "step": 100 }, { "epoch": 0.04627589246364037, "grad_norm": 0.23511621219550868, "learning_rate": 3.0837004405286347e-05, "loss": 2.024, "step": 105 }, { "epoch": 0.04847950639048039, "grad_norm": 0.24414856722360948, "learning_rate": 3.230543318649046e-05, "loss": 1.9756, "step": 110 }, { "epoch": 0.050683120317320404, "grad_norm": 0.25077238640232496, "learning_rate": 3.377386196769457e-05, "loss": 1.7627, "step": 115 }, { "epoch": 0.052886734244160426, "grad_norm": 0.30541323094735956, "learning_rate": 3.5242290748898684e-05, "loss": 1.9455, "step": 120 }, { "epoch": 0.05509034817100044, "grad_norm": 0.2593735959377697, "learning_rate": 3.6710719530102796e-05, "loss": 1.9408, "step": 125 }, { "epoch": 0.05729396209784046, "grad_norm": 0.2356435173973878, "learning_rate": 3.817914831130691e-05, "loss": 1.9746, "step": 130 }, { "epoch": 0.059497576024680476, "grad_norm": 0.2556575372905199, "learning_rate": 3.9647577092511014e-05, "loss": 1.9133, "step": 135 }, { "epoch": 0.06170118995152049, "grad_norm": 0.23671492386855494, "learning_rate": 4.1116005873715127e-05, "loss": 1.7228, "step": 140 }, { "epoch": 0.0639048038783605, "grad_norm": 0.2522128793647996, "learning_rate": 4.258443465491924e-05, "loss": 1.9416, "step": 145 }, { "epoch": 0.06610841780520053, "grad_norm": 0.259439737149894, "learning_rate": 4.405286343612335e-05, "loss": 2.0295, "step": 150 }, { "epoch": 0.06831203173204055, "grad_norm": 0.2632402024665515, "learning_rate": 4.5521292217327464e-05, "loss": 1.7906, "step": 155 }, { "epoch": 0.07051564565888056, "grad_norm": 0.28869506877919027, "learning_rate": 4.698972099853157e-05, "loss": 1.73, "step": 160 }, { "epoch": 0.07271925958572058, "grad_norm": 0.2586750438465526, "learning_rate": 4.845814977973568e-05, "loss": 1.8038, "step": 165 }, { "epoch": 0.0749228735125606, "grad_norm": 0.2506125082680772, "learning_rate": 4.9926578560939794e-05, "loss": 1.8558, "step": 170 }, { "epoch": 0.07712648743940062, "grad_norm": 0.2876243539987618, "learning_rate": 5.1395007342143906e-05, "loss": 1.9784, "step": 175 }, { "epoch": 0.07933010136624064, "grad_norm": 0.23288536708042795, "learning_rate": 5.286343612334802e-05, "loss": 1.7534, "step": 180 }, { "epoch": 0.08153371529308065, "grad_norm": 0.25062968804040664, "learning_rate": 5.433186490455213e-05, "loss": 1.8032, "step": 185 }, { "epoch": 0.08373732921992066, "grad_norm": 0.3358849445093002, "learning_rate": 5.5800293685756244e-05, "loss": 1.9325, "step": 190 }, { "epoch": 0.08594094314676069, "grad_norm": 0.25234625121100573, "learning_rate": 5.7268722466960356e-05, "loss": 1.8855, "step": 195 }, { "epoch": 0.08814455707360071, "grad_norm": 0.3365993483895123, "learning_rate": 5.873715124816447e-05, "loss": 1.8118, "step": 200 }, { "epoch": 0.09034817100044072, "grad_norm": 0.311599846916865, "learning_rate": 6.020558002936858e-05, "loss": 2.0721, "step": 205 }, { "epoch": 0.09255178492728074, "grad_norm": 0.3363293627242514, "learning_rate": 6.167400881057269e-05, "loss": 1.804, "step": 210 }, { "epoch": 0.09475539885412076, "grad_norm": 0.3414057038117249, "learning_rate": 6.31424375917768e-05, "loss": 1.871, "step": 215 }, { "epoch": 0.09695901278096078, "grad_norm": 0.3051539417193126, "learning_rate": 6.461086637298092e-05, "loss": 2.0932, "step": 220 }, { "epoch": 0.0991626267078008, "grad_norm": 0.2859223655623353, "learning_rate": 6.607929515418503e-05, "loss": 1.8641, "step": 225 }, { "epoch": 0.10136624063464081, "grad_norm": 0.30959368455808256, "learning_rate": 6.754772393538914e-05, "loss": 1.8437, "step": 230 }, { "epoch": 0.10356985456148082, "grad_norm": 0.3544528067642694, "learning_rate": 6.901615271659326e-05, "loss": 1.9002, "step": 235 }, { "epoch": 0.10577346848832085, "grad_norm": 0.34632827275641465, "learning_rate": 7.048458149779737e-05, "loss": 1.8398, "step": 240 }, { "epoch": 0.10797708241516087, "grad_norm": 0.597473977518495, "learning_rate": 7.195301027900148e-05, "loss": 1.9357, "step": 245 }, { "epoch": 0.11018069634200088, "grad_norm": 0.35953743798846133, "learning_rate": 7.342143906020559e-05, "loss": 1.944, "step": 250 }, { "epoch": 0.1123843102688409, "grad_norm": 0.32165430097730124, "learning_rate": 7.48898678414097e-05, "loss": 1.8974, "step": 255 }, { "epoch": 0.11458792419568092, "grad_norm": 0.2616578656705798, "learning_rate": 7.635829662261382e-05, "loss": 1.6988, "step": 260 }, { "epoch": 0.11679153812252094, "grad_norm": 0.37507900741140277, "learning_rate": 7.782672540381793e-05, "loss": 1.9061, "step": 265 }, { "epoch": 0.11899515204936095, "grad_norm": 0.3505810167092092, "learning_rate": 7.929515418502203e-05, "loss": 1.7829, "step": 270 }, { "epoch": 0.12119876597620097, "grad_norm": 0.2905187770243884, "learning_rate": 8.076358296622614e-05, "loss": 1.896, "step": 275 }, { "epoch": 0.12340237990304098, "grad_norm": 0.31667074390598227, "learning_rate": 8.223201174743025e-05, "loss": 1.8549, "step": 280 }, { "epoch": 0.125605993829881, "grad_norm": 0.33256810111534474, "learning_rate": 8.370044052863437e-05, "loss": 1.9184, "step": 285 }, { "epoch": 0.127809607756721, "grad_norm": 0.41094013265883617, "learning_rate": 8.516886930983848e-05, "loss": 1.9407, "step": 290 }, { "epoch": 0.13001322168356105, "grad_norm": 0.3268179907564926, "learning_rate": 8.663729809104259e-05, "loss": 1.858, "step": 295 }, { "epoch": 0.13221683561040107, "grad_norm": 0.4082234533551937, "learning_rate": 8.81057268722467e-05, "loss": 1.8172, "step": 300 }, { "epoch": 0.13442044953724108, "grad_norm": 0.3303954200472522, "learning_rate": 8.957415565345081e-05, "loss": 1.7879, "step": 305 }, { "epoch": 0.1366240634640811, "grad_norm": 0.3785434588284624, "learning_rate": 9.104258443465493e-05, "loss": 1.7931, "step": 310 }, { "epoch": 0.1388276773909211, "grad_norm": 0.3603897034728447, "learning_rate": 9.251101321585903e-05, "loss": 1.7793, "step": 315 }, { "epoch": 0.14103129131776113, "grad_norm": 0.2967904662976913, "learning_rate": 9.397944199706314e-05, "loss": 1.7092, "step": 320 }, { "epoch": 0.14323490524460114, "grad_norm": 0.2976559195453983, "learning_rate": 9.544787077826725e-05, "loss": 1.8302, "step": 325 }, { "epoch": 0.14543851917144116, "grad_norm": 0.34525724393698276, "learning_rate": 9.691629955947136e-05, "loss": 1.9488, "step": 330 }, { "epoch": 0.14764213309828117, "grad_norm": 0.3473487500731534, "learning_rate": 9.838472834067548e-05, "loss": 1.8169, "step": 335 }, { "epoch": 0.1498457470251212, "grad_norm": 0.35776737697830296, "learning_rate": 9.985315712187959e-05, "loss": 1.7562, "step": 340 }, { "epoch": 0.15204936095196123, "grad_norm": 0.35067754753032443, "learning_rate": 0.00010132158590308371, "loss": 1.9597, "step": 345 }, { "epoch": 0.15425297487880124, "grad_norm": 0.31111407532049806, "learning_rate": 0.00010279001468428781, "loss": 1.6977, "step": 350 }, { "epoch": 0.15645658880564126, "grad_norm": 0.350081092372353, "learning_rate": 0.00010425844346549194, "loss": 1.7585, "step": 355 }, { "epoch": 0.15866020273248127, "grad_norm": 0.3445787779571548, "learning_rate": 0.00010572687224669604, "loss": 1.7956, "step": 360 }, { "epoch": 0.16086381665932128, "grad_norm": 0.2992117867602888, "learning_rate": 0.00010719530102790014, "loss": 1.7752, "step": 365 }, { "epoch": 0.1630674305861613, "grad_norm": 0.3754242349038873, "learning_rate": 0.00010866372980910426, "loss": 1.8151, "step": 370 }, { "epoch": 0.16527104451300131, "grad_norm": 0.36786720092684805, "learning_rate": 0.00011013215859030836, "loss": 1.8172, "step": 375 }, { "epoch": 0.16747465843984133, "grad_norm": 0.3572694028170277, "learning_rate": 0.00011160058737151249, "loss": 1.8668, "step": 380 }, { "epoch": 0.16967827236668137, "grad_norm": 0.34400588996372305, "learning_rate": 0.00011306901615271659, "loss": 1.9685, "step": 385 }, { "epoch": 0.17188188629352139, "grad_norm": 0.3484332919699417, "learning_rate": 0.00011453744493392071, "loss": 1.999, "step": 390 }, { "epoch": 0.1740855002203614, "grad_norm": 0.3167720096524829, "learning_rate": 0.00011600587371512481, "loss": 1.7002, "step": 395 }, { "epoch": 0.17628911414720141, "grad_norm": 0.30865191358312394, "learning_rate": 0.00011747430249632894, "loss": 1.7808, "step": 400 }, { "epoch": 0.17849272807404143, "grad_norm": 0.3635805195870158, "learning_rate": 0.00011894273127753304, "loss": 1.8711, "step": 405 }, { "epoch": 0.18069634200088144, "grad_norm": 0.3113426681083048, "learning_rate": 0.00012041116005873716, "loss": 1.7559, "step": 410 }, { "epoch": 0.18289995592772146, "grad_norm": 0.3500754402872819, "learning_rate": 0.00012187958883994126, "loss": 2.0623, "step": 415 }, { "epoch": 0.18510356985456147, "grad_norm": 0.3002611868025747, "learning_rate": 0.00012334801762114539, "loss": 1.6155, "step": 420 }, { "epoch": 0.1873071837814015, "grad_norm": 0.30728017118253237, "learning_rate": 0.00012481644640234947, "loss": 1.7865, "step": 425 }, { "epoch": 0.18951079770824153, "grad_norm": 0.30435853108098426, "learning_rate": 0.0001262848751835536, "loss": 1.8101, "step": 430 }, { "epoch": 0.19171441163508154, "grad_norm": 0.33728305221904875, "learning_rate": 0.0001277533039647577, "loss": 1.7445, "step": 435 }, { "epoch": 0.19391802556192156, "grad_norm": 0.3269188152820616, "learning_rate": 0.00012922173274596184, "loss": 1.8443, "step": 440 }, { "epoch": 0.19612163948876157, "grad_norm": 0.30824761799725153, "learning_rate": 0.00013069016152716592, "loss": 1.8385, "step": 445 }, { "epoch": 0.1983252534156016, "grad_norm": 0.2692779779339547, "learning_rate": 0.00013215859030837006, "loss": 1.7256, "step": 450 }, { "epoch": 0.2005288673424416, "grad_norm": 0.29633902366885784, "learning_rate": 0.00013362701908957415, "loss": 1.8496, "step": 455 }, { "epoch": 0.20273248126928162, "grad_norm": 0.31921033190555953, "learning_rate": 0.00013509544787077829, "loss": 1.9118, "step": 460 }, { "epoch": 0.20493609519612163, "grad_norm": 0.38557464681376724, "learning_rate": 0.00013656387665198237, "loss": 1.7992, "step": 465 }, { "epoch": 0.20713970912296165, "grad_norm": 0.28175106471056965, "learning_rate": 0.0001380323054331865, "loss": 1.8768, "step": 470 }, { "epoch": 0.2093433230498017, "grad_norm": 0.31789045276124145, "learning_rate": 0.0001395007342143906, "loss": 1.7628, "step": 475 }, { "epoch": 0.2115469369766417, "grad_norm": 0.2412861323934925, "learning_rate": 0.00014096916299559473, "loss": 1.7988, "step": 480 }, { "epoch": 0.21375055090348172, "grad_norm": 0.24858109179997467, "learning_rate": 0.00014243759177679882, "loss": 1.8673, "step": 485 }, { "epoch": 0.21595416483032173, "grad_norm": 0.30755969433310765, "learning_rate": 0.00014390602055800296, "loss": 1.7105, "step": 490 }, { "epoch": 0.21815777875716175, "grad_norm": 0.2249398504391689, "learning_rate": 0.00014537444933920705, "loss": 1.7837, "step": 495 }, { "epoch": 0.22036139268400176, "grad_norm": 0.3335603893570301, "learning_rate": 0.00014684287812041118, "loss": 1.8333, "step": 500 }, { "epoch": 0.22256500661084178, "grad_norm": 0.3038176355961198, "learning_rate": 0.00014831130690161527, "loss": 1.8275, "step": 505 }, { "epoch": 0.2247686205376818, "grad_norm": 0.30793332193972345, "learning_rate": 0.0001497797356828194, "loss": 1.8177, "step": 510 }, { "epoch": 0.2269722344645218, "grad_norm": 0.3072544700688745, "learning_rate": 0.0001512481644640235, "loss": 1.9828, "step": 515 }, { "epoch": 0.22917584839136185, "grad_norm": 0.34999895392965585, "learning_rate": 0.00015271659324522763, "loss": 1.9117, "step": 520 }, { "epoch": 0.23137946231820186, "grad_norm": 0.32911766969763456, "learning_rate": 0.00015418502202643172, "loss": 1.6697, "step": 525 }, { "epoch": 0.23358307624504188, "grad_norm": 0.2617607674469746, "learning_rate": 0.00015565345080763586, "loss": 1.6296, "step": 530 }, { "epoch": 0.2357866901718819, "grad_norm": 0.43182139308340584, "learning_rate": 0.00015712187958883994, "loss": 1.9794, "step": 535 }, { "epoch": 0.2379903040987219, "grad_norm": 0.3127396374267501, "learning_rate": 0.00015859030837004406, "loss": 1.7436, "step": 540 }, { "epoch": 0.24019391802556192, "grad_norm": 0.21415922561789702, "learning_rate": 0.00016005873715124817, "loss": 1.7742, "step": 545 }, { "epoch": 0.24239753195240193, "grad_norm": 0.31852170762507637, "learning_rate": 0.00016152716593245228, "loss": 1.7969, "step": 550 }, { "epoch": 0.24460114587924195, "grad_norm": 0.29928941424884764, "learning_rate": 0.0001629955947136564, "loss": 1.8178, "step": 555 }, { "epoch": 0.24680475980608196, "grad_norm": 0.2145836014852306, "learning_rate": 0.0001644640234948605, "loss": 1.6149, "step": 560 }, { "epoch": 0.249008373732922, "grad_norm": 0.2765877443446801, "learning_rate": 0.00016593245227606462, "loss": 1.859, "step": 565 }, { "epoch": 0.251211987659762, "grad_norm": 0.26872348364190873, "learning_rate": 0.00016740088105726873, "loss": 1.8757, "step": 570 }, { "epoch": 0.25341560158660204, "grad_norm": 0.3149952835357651, "learning_rate": 0.00016886930983847284, "loss": 1.869, "step": 575 }, { "epoch": 0.255619215513442, "grad_norm": 0.2304679465113612, "learning_rate": 0.00017033773861967696, "loss": 1.7151, "step": 580 }, { "epoch": 0.25782282944028206, "grad_norm": 0.24264862789474126, "learning_rate": 0.00017180616740088107, "loss": 2.0695, "step": 585 }, { "epoch": 0.2600264433671221, "grad_norm": 0.2598067252093709, "learning_rate": 0.00017327459618208518, "loss": 1.839, "step": 590 }, { "epoch": 0.2622300572939621, "grad_norm": 0.26705538323533523, "learning_rate": 0.0001747430249632893, "loss": 1.8008, "step": 595 }, { "epoch": 0.26443367122080214, "grad_norm": 0.28479994342274373, "learning_rate": 0.0001762114537444934, "loss": 1.7284, "step": 600 }, { "epoch": 0.2666372851476421, "grad_norm": 0.28497317726898896, "learning_rate": 0.00017767988252569752, "loss": 1.9214, "step": 605 }, { "epoch": 0.26884089907448216, "grad_norm": 0.33639476801612694, "learning_rate": 0.00017914831130690163, "loss": 1.7516, "step": 610 }, { "epoch": 0.27104451300132215, "grad_norm": 0.25526934350033054, "learning_rate": 0.00018061674008810574, "loss": 1.7688, "step": 615 }, { "epoch": 0.2732481269281622, "grad_norm": 0.28869758623973935, "learning_rate": 0.00018208516886930985, "loss": 1.7843, "step": 620 }, { "epoch": 0.2754517408550022, "grad_norm": 0.2782854068624019, "learning_rate": 0.00018355359765051397, "loss": 1.8081, "step": 625 }, { "epoch": 0.2776553547818422, "grad_norm": 0.28975241284668296, "learning_rate": 0.00018502202643171805, "loss": 2.0243, "step": 630 }, { "epoch": 0.27985896870868227, "grad_norm": 0.3308791482681284, "learning_rate": 0.0001864904552129222, "loss": 1.9301, "step": 635 }, { "epoch": 0.28206258263552225, "grad_norm": 0.2725800531393519, "learning_rate": 0.00018795888399412628, "loss": 1.8126, "step": 640 }, { "epoch": 0.2842661965623623, "grad_norm": 0.2347435274751105, "learning_rate": 0.00018942731277533042, "loss": 1.9526, "step": 645 }, { "epoch": 0.2864698104892023, "grad_norm": 0.2594593727241724, "learning_rate": 0.0001908957415565345, "loss": 1.7362, "step": 650 }, { "epoch": 0.2886734244160423, "grad_norm": 0.20428739622605385, "learning_rate": 0.00019236417033773864, "loss": 1.9143, "step": 655 }, { "epoch": 0.2908770383428823, "grad_norm": 0.35231366470248277, "learning_rate": 0.00019383259911894273, "loss": 1.882, "step": 660 }, { "epoch": 0.29308065226972235, "grad_norm": 0.28303949235132847, "learning_rate": 0.00019530102790014687, "loss": 1.9956, "step": 665 }, { "epoch": 0.29528426619656234, "grad_norm": 0.28380137271318484, "learning_rate": 0.00019676945668135095, "loss": 1.9298, "step": 670 }, { "epoch": 0.2974878801234024, "grad_norm": 0.29197603939206423, "learning_rate": 0.0001982378854625551, "loss": 1.9936, "step": 675 }, { "epoch": 0.2996914940502424, "grad_norm": 0.3192279737304851, "learning_rate": 0.00019970631424375918, "loss": 1.856, "step": 680 }, { "epoch": 0.3018951079770824, "grad_norm": 0.31555391042987874, "learning_rate": 0.00019999978960491256, "loss": 1.9492, "step": 685 }, { "epoch": 0.30409872190392245, "grad_norm": 0.26468213545329267, "learning_rate": 0.0001999989348763872, "loss": 1.867, "step": 690 }, { "epoch": 0.30630233583076244, "grad_norm": 0.2591553540599883, "learning_rate": 0.0001999974226703463, "loss": 1.7565, "step": 695 }, { "epoch": 0.3085059497576025, "grad_norm": 0.33857027825097774, "learning_rate": 0.00019999525299673244, "loss": 1.8407, "step": 700 }, { "epoch": 0.31070956368444247, "grad_norm": 0.27554046406812405, "learning_rate": 0.0001999924258698108, "loss": 1.8449, "step": 705 }, { "epoch": 0.3129131776112825, "grad_norm": 0.2906036501277061, "learning_rate": 0.0001999889413081694, "loss": 1.9425, "step": 710 }, { "epoch": 0.3151167915381225, "grad_norm": 0.24516499633235853, "learning_rate": 0.00019998479933471862, "loss": 1.8373, "step": 715 }, { "epoch": 0.31732040546496254, "grad_norm": 0.261325768775238, "learning_rate": 0.0001999799999766913, "loss": 1.8655, "step": 720 }, { "epoch": 0.3195240193918026, "grad_norm": 0.26292913191780404, "learning_rate": 0.00019997454326564252, "loss": 1.8011, "step": 725 }, { "epoch": 0.32172763331864257, "grad_norm": 0.24941214673161538, "learning_rate": 0.0001999684292374493, "loss": 1.7063, "step": 730 }, { "epoch": 0.3239312472454826, "grad_norm": 0.285047883815678, "learning_rate": 0.00019996165793231038, "loss": 1.9537, "step": 735 }, { "epoch": 0.3261348611723226, "grad_norm": 0.2853612548244496, "learning_rate": 0.0001999542293947461, "loss": 1.7641, "step": 740 }, { "epoch": 0.32833847509916264, "grad_norm": 0.2600200330765211, "learning_rate": 0.00019994614367359792, "loss": 1.8886, "step": 745 }, { "epoch": 0.33054208902600263, "grad_norm": 0.36999348537839605, "learning_rate": 0.00019993740082202818, "loss": 1.798, "step": 750 }, { "epoch": 0.33274570295284267, "grad_norm": 0.24647791783433215, "learning_rate": 0.00019992800089751984, "loss": 1.8922, "step": 755 }, { "epoch": 0.33494931687968266, "grad_norm": 0.23414809097297196, "learning_rate": 0.0001999179439618759, "loss": 1.6675, "step": 760 }, { "epoch": 0.3371529308065227, "grad_norm": 0.24171495213943817, "learning_rate": 0.00019990723008121917, "loss": 1.5054, "step": 765 }, { "epoch": 0.33935654473336274, "grad_norm": 0.2462163957540956, "learning_rate": 0.00019989585932599172, "loss": 1.8441, "step": 770 }, { "epoch": 0.34156015866020273, "grad_norm": 0.24192478872496498, "learning_rate": 0.00019988383177095459, "loss": 1.8535, "step": 775 }, { "epoch": 0.34376377258704277, "grad_norm": 0.21554799200025276, "learning_rate": 0.000199871147495187, "loss": 1.7299, "step": 780 }, { "epoch": 0.34596738651388276, "grad_norm": 0.29240426650240625, "learning_rate": 0.00019985780658208618, "loss": 1.9846, "step": 785 }, { "epoch": 0.3481710004407228, "grad_norm": 0.32582827846868756, "learning_rate": 0.00019984380911936648, "loss": 1.5922, "step": 790 }, { "epoch": 0.3503746143675628, "grad_norm": 0.3253623043496682, "learning_rate": 0.00019982915519905912, "loss": 1.7138, "step": 795 }, { "epoch": 0.35257822829440283, "grad_norm": 0.3012253411690825, "learning_rate": 0.00019981384491751133, "loss": 1.9526, "step": 800 }, { "epoch": 0.3547818422212428, "grad_norm": 0.32373710599771705, "learning_rate": 0.00019979787837538587, "loss": 2.0799, "step": 805 }, { "epoch": 0.35698545614808286, "grad_norm": 0.2736459891711902, "learning_rate": 0.00019978125567766023, "loss": 1.8422, "step": 810 }, { "epoch": 0.3591890700749229, "grad_norm": 0.28970333253517644, "learning_rate": 0.00019976397693362614, "loss": 1.8309, "step": 815 }, { "epoch": 0.3613926840017629, "grad_norm": 0.271249926943477, "learning_rate": 0.0001997460422568886, "loss": 1.6581, "step": 820 }, { "epoch": 0.36359629792860293, "grad_norm": 0.25582927997161137, "learning_rate": 0.00019972745176536537, "loss": 1.9441, "step": 825 }, { "epoch": 0.3657999118554429, "grad_norm": 0.27987061803178176, "learning_rate": 0.00019970820558128604, "loss": 1.8015, "step": 830 }, { "epoch": 0.36800352578228296, "grad_norm": 0.2956257318754398, "learning_rate": 0.0001996883038311913, "loss": 1.7853, "step": 835 }, { "epoch": 0.37020713970912295, "grad_norm": 0.25727871315384043, "learning_rate": 0.00019966774664593206, "loss": 1.7594, "step": 840 }, { "epoch": 0.372410753635963, "grad_norm": 0.3069722747545403, "learning_rate": 0.00019964653416066868, "loss": 1.9102, "step": 845 }, { "epoch": 0.374614367562803, "grad_norm": 0.362312858514461, "learning_rate": 0.0001996246665148699, "loss": 1.8419, "step": 850 }, { "epoch": 0.376817981489643, "grad_norm": 0.30043599837768675, "learning_rate": 0.00019960214385231217, "loss": 1.9281, "step": 855 }, { "epoch": 0.37902159541648306, "grad_norm": 0.25742900200071334, "learning_rate": 0.00019957896632107845, "loss": 1.8382, "step": 860 }, { "epoch": 0.38122520934332305, "grad_norm": 0.22904042108335546, "learning_rate": 0.00019955513407355743, "loss": 1.585, "step": 865 }, { "epoch": 0.3834288232701631, "grad_norm": 0.35248924984262536, "learning_rate": 0.0001995306472664425, "loss": 1.8779, "step": 870 }, { "epoch": 0.3856324371970031, "grad_norm": 0.21394107686214808, "learning_rate": 0.00019950550606073056, "loss": 1.6203, "step": 875 }, { "epoch": 0.3878360511238431, "grad_norm": 0.23210191109497874, "learning_rate": 0.00019947971062172118, "loss": 1.7579, "step": 880 }, { "epoch": 0.3900396650506831, "grad_norm": 0.280197585305159, "learning_rate": 0.00019945326111901542, "loss": 1.8697, "step": 885 }, { "epoch": 0.39224327897752315, "grad_norm": 0.2007938963214883, "learning_rate": 0.00019942615772651455, "loss": 1.6718, "step": 890 }, { "epoch": 0.39444689290436313, "grad_norm": 0.28396842059026434, "learning_rate": 0.0001993984006224193, "loss": 1.8261, "step": 895 }, { "epoch": 0.3966505068312032, "grad_norm": 0.2970710682132302, "learning_rate": 0.00019936998998922826, "loss": 1.8988, "step": 900 }, { "epoch": 0.3988541207580432, "grad_norm": 0.4794659287163716, "learning_rate": 0.00019934092601373694, "loss": 1.8387, "step": 905 }, { "epoch": 0.4010577346848832, "grad_norm": 0.2810211439045603, "learning_rate": 0.00019931120888703652, "loss": 1.7516, "step": 910 }, { "epoch": 0.40326134861172325, "grad_norm": 0.28984779802870314, "learning_rate": 0.0001992808388045125, "loss": 1.9212, "step": 915 }, { "epoch": 0.40546496253856323, "grad_norm": 0.26294811280148017, "learning_rate": 0.00019924981596584345, "loss": 1.8798, "step": 920 }, { "epoch": 0.4076685764654033, "grad_norm": 0.27546864810479915, "learning_rate": 0.00019921814057499978, "loss": 1.7595, "step": 925 }, { "epoch": 0.40987219039224326, "grad_norm": 0.2506019656377615, "learning_rate": 0.0001991858128402422, "loss": 1.7625, "step": 930 }, { "epoch": 0.4120758043190833, "grad_norm": 0.35470941281384444, "learning_rate": 0.0001991528329741206, "loss": 1.8644, "step": 935 }, { "epoch": 0.4142794182459233, "grad_norm": 0.30456048523132856, "learning_rate": 0.00019911920119347254, "loss": 1.8427, "step": 940 }, { "epoch": 0.41648303217276333, "grad_norm": 0.21180617855630868, "learning_rate": 0.0001990849177194217, "loss": 1.947, "step": 945 }, { "epoch": 0.4186866460996034, "grad_norm": 0.27175217869149726, "learning_rate": 0.00019904998277737668, "loss": 1.5794, "step": 950 }, { "epoch": 0.42089026002644336, "grad_norm": 0.25340188979799705, "learning_rate": 0.00019901439659702924, "loss": 1.655, "step": 955 }, { "epoch": 0.4230938739532834, "grad_norm": 0.32039092541913483, "learning_rate": 0.00019897815941235307, "loss": 1.9448, "step": 960 }, { "epoch": 0.4252974878801234, "grad_norm": 0.28203099326236186, "learning_rate": 0.00019894127146160204, "loss": 1.73, "step": 965 }, { "epoch": 0.42750110180696343, "grad_norm": 0.790698073004524, "learning_rate": 0.00019890373298730868, "loss": 2.0466, "step": 970 }, { "epoch": 0.4297047157338034, "grad_norm": 0.23107801878483133, "learning_rate": 0.0001988655442362827, "loss": 1.5805, "step": 975 }, { "epoch": 0.43190832966064346, "grad_norm": 0.2396293478769625, "learning_rate": 0.00019882670545960914, "loss": 1.7482, "step": 980 }, { "epoch": 0.43411194358748345, "grad_norm": 0.29159214942433453, "learning_rate": 0.00019878721691264704, "loss": 1.9851, "step": 985 }, { "epoch": 0.4363155575143235, "grad_norm": 0.2825006648123125, "learning_rate": 0.00019874707885502745, "loss": 1.7534, "step": 990 }, { "epoch": 0.43851917144116354, "grad_norm": 0.28034425756379244, "learning_rate": 0.00019870629155065186, "loss": 1.7489, "step": 995 }, { "epoch": 0.4407227853680035, "grad_norm": 0.3181059895374247, "learning_rate": 0.0001986648552676905, "loss": 1.8798, "step": 1000 }, { "epoch": 0.44292639929484356, "grad_norm": 0.20367681253555586, "learning_rate": 0.0001986227702785805, "loss": 1.8065, "step": 1005 }, { "epoch": 0.44513001322168355, "grad_norm": 0.25873539379499144, "learning_rate": 0.0001985800368600242, "loss": 1.687, "step": 1010 }, { "epoch": 0.4473336271485236, "grad_norm": 0.2342074611961214, "learning_rate": 0.0001985366552929871, "loss": 1.9431, "step": 1015 }, { "epoch": 0.4495372410753636, "grad_norm": 0.48215863294928724, "learning_rate": 0.00019849262586269642, "loss": 1.8137, "step": 1020 }, { "epoch": 0.4517408550022036, "grad_norm": 0.27724392120608005, "learning_rate": 0.00019844794885863877, "loss": 1.8311, "step": 1025 }, { "epoch": 0.4539444689290436, "grad_norm": 0.22655546528476625, "learning_rate": 0.00019840262457455855, "loss": 1.6968, "step": 1030 }, { "epoch": 0.45614808285588365, "grad_norm": 0.24804519754733267, "learning_rate": 0.00019835665330845595, "loss": 1.844, "step": 1035 }, { "epoch": 0.4583516967827237, "grad_norm": 0.3053870909241666, "learning_rate": 0.00019831003536258487, "loss": 1.6674, "step": 1040 }, { "epoch": 0.4605553107095637, "grad_norm": 0.2677054196356507, "learning_rate": 0.00019826277104345109, "loss": 1.994, "step": 1045 }, { "epoch": 0.4627589246364037, "grad_norm": 0.2332936772139347, "learning_rate": 0.0001982148606618102, "loss": 1.7963, "step": 1050 }, { "epoch": 0.4649625385632437, "grad_norm": 0.290304703836439, "learning_rate": 0.00019816630453266555, "loss": 1.8278, "step": 1055 }, { "epoch": 0.46716615249008375, "grad_norm": 0.31440944408786714, "learning_rate": 0.0001981171029752662, "loss": 1.7053, "step": 1060 }, { "epoch": 0.46936976641692374, "grad_norm": 0.31837742235142563, "learning_rate": 0.00019806725631310476, "loss": 1.8377, "step": 1065 }, { "epoch": 0.4715733803437638, "grad_norm": 0.31835369363351046, "learning_rate": 0.00019801676487391529, "loss": 1.7635, "step": 1070 }, { "epoch": 0.47377699427060377, "grad_norm": 0.32008261758346795, "learning_rate": 0.0001979656289896712, "loss": 1.8322, "step": 1075 }, { "epoch": 0.4759806081974438, "grad_norm": 0.2436306554476478, "learning_rate": 0.000197913848996583, "loss": 1.7057, "step": 1080 }, { "epoch": 0.47818422212428385, "grad_norm": 0.24420764132400904, "learning_rate": 0.00019786142523509615, "loss": 1.7756, "step": 1085 }, { "epoch": 0.48038783605112384, "grad_norm": 0.24156752527695532, "learning_rate": 0.00019780835804988876, "loss": 1.7788, "step": 1090 }, { "epoch": 0.4825914499779639, "grad_norm": 0.3341728235303617, "learning_rate": 0.00019775464778986934, "loss": 1.9, "step": 1095 }, { "epoch": 0.48479506390480387, "grad_norm": 0.23920551280340352, "learning_rate": 0.00019770029480817454, "loss": 1.8956, "step": 1100 }, { "epoch": 0.4869986778316439, "grad_norm": 0.2463312946508512, "learning_rate": 0.00019764529946216682, "loss": 1.5345, "step": 1105 }, { "epoch": 0.4892022917584839, "grad_norm": 0.26366246938923554, "learning_rate": 0.00019758966211343206, "loss": 1.7621, "step": 1110 }, { "epoch": 0.49140590568532394, "grad_norm": 0.3123943209000793, "learning_rate": 0.00019753338312777718, "loss": 1.815, "step": 1115 }, { "epoch": 0.4936095196121639, "grad_norm": 0.2551443475914167, "learning_rate": 0.00019747646287522784, "loss": 1.8611, "step": 1120 }, { "epoch": 0.49581313353900397, "grad_norm": 0.2332525414967985, "learning_rate": 0.0001974189017300259, "loss": 1.5384, "step": 1125 }, { "epoch": 0.498016747465844, "grad_norm": 0.2861706387721841, "learning_rate": 0.00019736070007062692, "loss": 2.029, "step": 1130 }, { "epoch": 0.500220361392684, "grad_norm": 0.4583730418300565, "learning_rate": 0.00019730185827969784, "loss": 1.6826, "step": 1135 }, { "epoch": 0.502423975319524, "grad_norm": 0.1931367375627417, "learning_rate": 0.00019724237674411432, "loss": 1.6877, "step": 1140 }, { "epoch": 0.5046275892463641, "grad_norm": 0.3082508866109686, "learning_rate": 0.00019718225585495824, "loss": 1.8148, "step": 1145 }, { "epoch": 0.5068312031732041, "grad_norm": 0.18106453992956664, "learning_rate": 0.00019712149600751517, "loss": 1.6556, "step": 1150 }, { "epoch": 0.5090348171000441, "grad_norm": 0.23000297672552616, "learning_rate": 0.00019706009760127164, "loss": 2.0152, "step": 1155 }, { "epoch": 0.511238431026884, "grad_norm": 0.29387352965288377, "learning_rate": 0.00019699806103991272, "loss": 1.7962, "step": 1160 }, { "epoch": 0.5134420449537241, "grad_norm": 0.23060861917057546, "learning_rate": 0.00019693538673131917, "loss": 1.8123, "step": 1165 }, { "epoch": 0.5156456588805641, "grad_norm": 0.25329490761690543, "learning_rate": 0.00019687207508756486, "loss": 1.7052, "step": 1170 }, { "epoch": 0.5178492728074041, "grad_norm": 0.3356944723720528, "learning_rate": 0.00019680812652491408, "loss": 1.7985, "step": 1175 }, { "epoch": 0.5200528867342442, "grad_norm": 0.31492967037029806, "learning_rate": 0.0001967435414638187, "loss": 1.7971, "step": 1180 }, { "epoch": 0.5222565006610842, "grad_norm": 0.2915314540239125, "learning_rate": 0.00019667832032891554, "loss": 1.9571, "step": 1185 }, { "epoch": 0.5244601145879242, "grad_norm": 0.28065572919249093, "learning_rate": 0.00019661246354902342, "loss": 1.9185, "step": 1190 }, { "epoch": 0.5266637285147642, "grad_norm": 0.22888871376259906, "learning_rate": 0.00019654597155714044, "loss": 1.7367, "step": 1195 }, { "epoch": 0.5288673424416043, "grad_norm": 0.23069615225682083, "learning_rate": 0.00019647884479044123, "loss": 1.7333, "step": 1200 }, { "epoch": 0.5310709563684443, "grad_norm": 0.25845750908289505, "learning_rate": 0.00019641108369027385, "loss": 1.5907, "step": 1205 }, { "epoch": 0.5332745702952842, "grad_norm": 0.3069614304651339, "learning_rate": 0.00019634268870215703, "loss": 1.9282, "step": 1210 }, { "epoch": 0.5354781842221242, "grad_norm": 0.3005722535622308, "learning_rate": 0.00019627366027577726, "loss": 1.6378, "step": 1215 }, { "epoch": 0.5376817981489643, "grad_norm": 0.270244780455395, "learning_rate": 0.00019620399886498578, "loss": 1.6499, "step": 1220 }, { "epoch": 0.5398854120758043, "grad_norm": 0.2796374681033303, "learning_rate": 0.0001961337049277955, "loss": 1.7962, "step": 1225 }, { "epoch": 0.5420890260026443, "grad_norm": 0.2608673103569785, "learning_rate": 0.00019606277892637823, "loss": 1.6946, "step": 1230 }, { "epoch": 0.5442926399294844, "grad_norm": 0.35197474426731085, "learning_rate": 0.00019599122132706146, "loss": 1.9751, "step": 1235 }, { "epoch": 0.5464962538563244, "grad_norm": 0.26530267535084306, "learning_rate": 0.0001959190326003253, "loss": 1.7257, "step": 1240 }, { "epoch": 0.5486998677831644, "grad_norm": 0.2540337035705083, "learning_rate": 0.00019584621322079942, "loss": 1.8693, "step": 1245 }, { "epoch": 0.5509034817100044, "grad_norm": 0.2530867344602867, "learning_rate": 0.00019577276366726003, "loss": 1.6761, "step": 1250 }, { "epoch": 0.5531070956368445, "grad_norm": 0.2549205571103006, "learning_rate": 0.00019569868442262655, "loss": 1.8729, "step": 1255 }, { "epoch": 0.5553107095636844, "grad_norm": 0.2871173332255517, "learning_rate": 0.00019562397597395857, "loss": 1.8347, "step": 1260 }, { "epoch": 0.5575143234905244, "grad_norm": 0.28369117286467893, "learning_rate": 0.0001955486388124525, "loss": 1.8547, "step": 1265 }, { "epoch": 0.5597179374173645, "grad_norm": 0.2852722364768899, "learning_rate": 0.00019547267343343857, "loss": 1.6552, "step": 1270 }, { "epoch": 0.5619215513442045, "grad_norm": 0.34585189639791736, "learning_rate": 0.0001953960803363774, "loss": 1.6727, "step": 1275 }, { "epoch": 0.5641251652710445, "grad_norm": 0.2767216642372831, "learning_rate": 0.00019531886002485674, "loss": 1.8886, "step": 1280 }, { "epoch": 0.5663287791978845, "grad_norm": 0.20036600537602942, "learning_rate": 0.00019524101300658813, "loss": 1.8685, "step": 1285 }, { "epoch": 0.5685323931247246, "grad_norm": 0.261316473836974, "learning_rate": 0.0001951625397934037, "loss": 1.782, "step": 1290 }, { "epoch": 0.5707360070515646, "grad_norm": 0.4236076570583356, "learning_rate": 0.0001950834409012527, "loss": 1.8318, "step": 1295 }, { "epoch": 0.5729396209784046, "grad_norm": 0.28925306291568, "learning_rate": 0.00019500371685019806, "loss": 1.6012, "step": 1300 }, { "epoch": 0.5751432349052445, "grad_norm": 0.2945702905493002, "learning_rate": 0.0001949233681644131, "loss": 1.9158, "step": 1305 }, { "epoch": 0.5773468488320846, "grad_norm": 0.27644920036772963, "learning_rate": 0.00019484239537217798, "loss": 1.8232, "step": 1310 }, { "epoch": 0.5795504627589246, "grad_norm": 0.24910322311175717, "learning_rate": 0.00019476079900587626, "loss": 2.0731, "step": 1315 }, { "epoch": 0.5817540766857646, "grad_norm": 0.2746087574116961, "learning_rate": 0.00019467857960199142, "loss": 1.8429, "step": 1320 }, { "epoch": 0.5839576906126047, "grad_norm": 0.30262920301398166, "learning_rate": 0.00019459573770110335, "loss": 1.7647, "step": 1325 }, { "epoch": 0.5861613045394447, "grad_norm": 0.2570054670626858, "learning_rate": 0.0001945122738478847, "loss": 1.7511, "step": 1330 }, { "epoch": 0.5883649184662847, "grad_norm": 0.2506934757685486, "learning_rate": 0.00019442818859109737, "loss": 1.9036, "step": 1335 }, { "epoch": 0.5905685323931247, "grad_norm": 0.28984253571835894, "learning_rate": 0.00019434348248358892, "loss": 1.7763, "step": 1340 }, { "epoch": 0.5927721463199648, "grad_norm": 0.28120650964379307, "learning_rate": 0.00019425815608228888, "loss": 1.8062, "step": 1345 }, { "epoch": 0.5949757602468048, "grad_norm": 0.26796794914129696, "learning_rate": 0.00019417220994820514, "loss": 1.7886, "step": 1350 }, { "epoch": 0.5971793741736448, "grad_norm": 0.27855604023858827, "learning_rate": 0.00019408564464642024, "loss": 1.6226, "step": 1355 }, { "epoch": 0.5993829881004848, "grad_norm": 0.23996517928921976, "learning_rate": 0.00019399846074608757, "loss": 1.8206, "step": 1360 }, { "epoch": 0.6015866020273248, "grad_norm": 0.2718103628870133, "learning_rate": 0.00019391065882042786, "loss": 1.622, "step": 1365 }, { "epoch": 0.6037902159541648, "grad_norm": 0.40326599118637163, "learning_rate": 0.00019382223944672516, "loss": 1.8595, "step": 1370 }, { "epoch": 0.6059938298810048, "grad_norm": 0.26188146699351833, "learning_rate": 0.00019373320320632313, "loss": 1.719, "step": 1375 }, { "epoch": 0.6081974438078449, "grad_norm": 0.26353376595142103, "learning_rate": 0.00019364355068462126, "loss": 1.7599, "step": 1380 }, { "epoch": 0.6104010577346849, "grad_norm": 0.27506310894040453, "learning_rate": 0.00019355328247107106, "loss": 1.6895, "step": 1385 }, { "epoch": 0.6126046716615249, "grad_norm": 0.23293714707305346, "learning_rate": 0.00019346239915917204, "loss": 1.9199, "step": 1390 }, { "epoch": 0.6148082855883649, "grad_norm": 0.2728024319863431, "learning_rate": 0.00019337090134646787, "loss": 1.6137, "step": 1395 }, { "epoch": 0.617011899515205, "grad_norm": 0.2580080135345881, "learning_rate": 0.00019327878963454253, "loss": 1.9251, "step": 1400 }, { "epoch": 0.619215513442045, "grad_norm": 0.18292171734629373, "learning_rate": 0.00019318606462901625, "loss": 1.6127, "step": 1405 }, { "epoch": 0.6214191273688849, "grad_norm": 0.2540991707755816, "learning_rate": 0.0001930927269395416, "loss": 1.716, "step": 1410 }, { "epoch": 0.623622741295725, "grad_norm": 0.2222252074542422, "learning_rate": 0.00019299877717979944, "loss": 1.649, "step": 1415 }, { "epoch": 0.625826355222565, "grad_norm": 0.23656133580313402, "learning_rate": 0.00019290421596749487, "loss": 1.7321, "step": 1420 }, { "epoch": 0.628029969149405, "grad_norm": 0.2770878731693278, "learning_rate": 0.00019280904392435328, "loss": 1.7982, "step": 1425 }, { "epoch": 0.630233583076245, "grad_norm": 0.38713553349288304, "learning_rate": 0.00019271326167611606, "loss": 1.757, "step": 1430 }, { "epoch": 0.6324371970030851, "grad_norm": 0.21826448168974197, "learning_rate": 0.00019261686985253668, "loss": 1.6568, "step": 1435 }, { "epoch": 0.6346408109299251, "grad_norm": 0.28392855935252787, "learning_rate": 0.00019251986908737646, "loss": 1.6995, "step": 1440 }, { "epoch": 0.6368444248567651, "grad_norm": 0.3335345636849735, "learning_rate": 0.00019242226001840043, "loss": 1.6445, "step": 1445 }, { "epoch": 0.6390480387836052, "grad_norm": 0.23367317521663916, "learning_rate": 0.0001923240432873731, "loss": 1.7995, "step": 1450 }, { "epoch": 0.6412516527104452, "grad_norm": 0.26851942754392183, "learning_rate": 0.00019222521954005424, "loss": 1.8078, "step": 1455 }, { "epoch": 0.6434552666372851, "grad_norm": 0.28393938915488615, "learning_rate": 0.00019212578942619474, "loss": 1.7108, "step": 1460 }, { "epoch": 0.6456588805641251, "grad_norm": 0.2481400915816766, "learning_rate": 0.00019202575359953213, "loss": 1.7509, "step": 1465 }, { "epoch": 0.6478624944909652, "grad_norm": 0.22200215432596082, "learning_rate": 0.00019192511271778656, "loss": 1.6549, "step": 1470 }, { "epoch": 0.6500661084178052, "grad_norm": 0.24388551441860462, "learning_rate": 0.00019182386744265623, "loss": 1.9977, "step": 1475 }, { "epoch": 0.6522697223446452, "grad_norm": 0.28231311563927136, "learning_rate": 0.00019172201843981314, "loss": 1.7473, "step": 1480 }, { "epoch": 0.6544733362714852, "grad_norm": 0.2911468005732327, "learning_rate": 0.00019161956637889872, "loss": 1.8572, "step": 1485 }, { "epoch": 0.6566769501983253, "grad_norm": 0.2809891282005259, "learning_rate": 0.0001915165119335194, "loss": 1.6363, "step": 1490 }, { "epoch": 0.6588805641251653, "grad_norm": 0.257092783900919, "learning_rate": 0.0001914128557812422, "loss": 1.6894, "step": 1495 }, { "epoch": 0.6610841780520053, "grad_norm": 0.25492249438744313, "learning_rate": 0.00019130859860359026, "loss": 1.9549, "step": 1500 }, { "epoch": 0.6632877919788454, "grad_norm": 0.29656417671091373, "learning_rate": 0.00019120374108603843, "loss": 1.882, "step": 1505 }, { "epoch": 0.6654914059056853, "grad_norm": 0.26795937749586346, "learning_rate": 0.0001910982839180086, "loss": 1.7532, "step": 1510 }, { "epoch": 0.6676950198325253, "grad_norm": 0.30019068058678056, "learning_rate": 0.0001909922277928654, "loss": 1.9185, "step": 1515 }, { "epoch": 0.6698986337593653, "grad_norm": 0.2908491561313871, "learning_rate": 0.00019088557340791136, "loss": 1.8659, "step": 1520 }, { "epoch": 0.6721022476862054, "grad_norm": 0.2664246491795904, "learning_rate": 0.00019077832146438257, "loss": 1.618, "step": 1525 }, { "epoch": 0.6743058616130454, "grad_norm": 0.2670150987011056, "learning_rate": 0.00019067047266744396, "loss": 1.82, "step": 1530 }, { "epoch": 0.6765094755398854, "grad_norm": 0.26014739460054054, "learning_rate": 0.0001905620277261847, "loss": 1.8267, "step": 1535 }, { "epoch": 0.6787130894667255, "grad_norm": 0.39761269473071104, "learning_rate": 0.00019045298735361345, "loss": 1.7682, "step": 1540 }, { "epoch": 0.6809167033935655, "grad_norm": 0.25574765659361764, "learning_rate": 0.0001903433522666538, "loss": 1.8709, "step": 1545 }, { "epoch": 0.6831203173204055, "grad_norm": 0.2633667660696579, "learning_rate": 0.00019023312318613945, "loss": 1.7439, "step": 1550 }, { "epoch": 0.6853239312472454, "grad_norm": 0.33365458542367915, "learning_rate": 0.00019012230083680954, "loss": 1.8991, "step": 1555 }, { "epoch": 0.6875275451740855, "grad_norm": 0.26233949374382926, "learning_rate": 0.0001900108859473039, "loss": 1.7681, "step": 1560 }, { "epoch": 0.6897311591009255, "grad_norm": 0.27517789876709425, "learning_rate": 0.00018989887925015814, "loss": 1.8164, "step": 1565 }, { "epoch": 0.6919347730277655, "grad_norm": 0.22352922332415057, "learning_rate": 0.00018978628148179897, "loss": 1.6674, "step": 1570 }, { "epoch": 0.6941383869546055, "grad_norm": 0.2705784713837994, "learning_rate": 0.0001896730933825393, "loss": 1.883, "step": 1575 }, { "epoch": 0.6963420008814456, "grad_norm": 0.2333580239686404, "learning_rate": 0.00018955931569657333, "loss": 1.6889, "step": 1580 }, { "epoch": 0.6985456148082856, "grad_norm": 0.29573837903023903, "learning_rate": 0.00018944494917197172, "loss": 1.8473, "step": 1585 }, { "epoch": 0.7007492287351256, "grad_norm": 0.24394341802838163, "learning_rate": 0.00018932999456067675, "loss": 1.844, "step": 1590 }, { "epoch": 0.7029528426619657, "grad_norm": 0.2454925329534158, "learning_rate": 0.0001892144526184971, "loss": 1.7824, "step": 1595 }, { "epoch": 0.7051564565888057, "grad_norm": 0.23532373440650545, "learning_rate": 0.00018909832410510315, "loss": 1.8537, "step": 1600 }, { "epoch": 0.7073600705156456, "grad_norm": 0.2602720785746293, "learning_rate": 0.00018898160978402198, "loss": 1.8717, "step": 1605 }, { "epoch": 0.7095636844424856, "grad_norm": 0.23292617880883168, "learning_rate": 0.00018886431042263208, "loss": 1.703, "step": 1610 }, { "epoch": 0.7117672983693257, "grad_norm": 0.32473590479932374, "learning_rate": 0.0001887464267921587, "loss": 1.7075, "step": 1615 }, { "epoch": 0.7139709122961657, "grad_norm": 0.21131060430922607, "learning_rate": 0.00018862795966766833, "loss": 1.6993, "step": 1620 }, { "epoch": 0.7161745262230057, "grad_norm": 0.2945260951578129, "learning_rate": 0.0001885089098280641, "loss": 1.7315, "step": 1625 }, { "epoch": 0.7183781401498458, "grad_norm": 0.2988718316963067, "learning_rate": 0.0001883892780560802, "loss": 1.7079, "step": 1630 }, { "epoch": 0.7205817540766858, "grad_norm": 0.2671861990469183, "learning_rate": 0.00018826906513827704, "loss": 1.8816, "step": 1635 }, { "epoch": 0.7227853680035258, "grad_norm": 0.3539590759381266, "learning_rate": 0.00018814827186503595, "loss": 1.7559, "step": 1640 }, { "epoch": 0.7249889819303658, "grad_norm": 0.2098483762554333, "learning_rate": 0.00018802689903055396, "loss": 1.8296, "step": 1645 }, { "epoch": 0.7271925958572059, "grad_norm": 0.24342758941079398, "learning_rate": 0.0001879049474328387, "loss": 1.8845, "step": 1650 }, { "epoch": 0.7293962097840458, "grad_norm": 0.30213591141833546, "learning_rate": 0.00018778241787370303, "loss": 1.7739, "step": 1655 }, { "epoch": 0.7315998237108858, "grad_norm": 0.2796229215884198, "learning_rate": 0.00018765931115875985, "loss": 1.7238, "step": 1660 }, { "epoch": 0.7338034376377258, "grad_norm": 0.2818539631618944, "learning_rate": 0.00018753562809741673, "loss": 1.7833, "step": 1665 }, { "epoch": 0.7360070515645659, "grad_norm": 0.2534113717910187, "learning_rate": 0.00018741136950287067, "loss": 1.781, "step": 1670 }, { "epoch": 0.7382106654914059, "grad_norm": 0.25817838823410116, "learning_rate": 0.0001872865361921027, "loss": 1.6845, "step": 1675 }, { "epoch": 0.7404142794182459, "grad_norm": 0.25172352057271447, "learning_rate": 0.00018716112898587247, "loss": 1.9169, "step": 1680 }, { "epoch": 0.742617893345086, "grad_norm": 0.257829203989891, "learning_rate": 0.000187035148708713, "loss": 1.7977, "step": 1685 }, { "epoch": 0.744821507271926, "grad_norm": 0.21408758544817338, "learning_rate": 0.00018690859618892506, "loss": 1.6934, "step": 1690 }, { "epoch": 0.747025121198766, "grad_norm": 0.32305529721640136, "learning_rate": 0.0001867814722585719, "loss": 1.811, "step": 1695 }, { "epoch": 0.749228735125606, "grad_norm": 0.27366670016595385, "learning_rate": 0.0001866537777534737, "loss": 1.6083, "step": 1700 }, { "epoch": 0.751432349052446, "grad_norm": 0.2978907514821807, "learning_rate": 0.00018652551351320198, "loss": 1.7621, "step": 1705 }, { "epoch": 0.753635962979286, "grad_norm": 0.2682049213061407, "learning_rate": 0.00018639668038107437, "loss": 1.8008, "step": 1710 }, { "epoch": 0.755839576906126, "grad_norm": 0.2422117187271709, "learning_rate": 0.0001862672792041487, "loss": 1.9899, "step": 1715 }, { "epoch": 0.7580431908329661, "grad_norm": 0.27006478070278994, "learning_rate": 0.0001861373108332177, "loss": 1.7577, "step": 1720 }, { "epoch": 0.7602468047598061, "grad_norm": 0.31176719361390687, "learning_rate": 0.0001860067761228033, "loss": 1.7494, "step": 1725 }, { "epoch": 0.7624504186866461, "grad_norm": 0.2965711368621697, "learning_rate": 0.00018587567593115098, "loss": 1.9554, "step": 1730 }, { "epoch": 0.7646540326134861, "grad_norm": 0.26922434332877426, "learning_rate": 0.0001857440111202242, "loss": 1.7415, "step": 1735 }, { "epoch": 0.7668576465403262, "grad_norm": 0.25558172886185604, "learning_rate": 0.00018561178255569879, "loss": 1.7389, "step": 1740 }, { "epoch": 0.7690612604671662, "grad_norm": 0.24853091753480136, "learning_rate": 0.000185478991106957, "loss": 1.9658, "step": 1745 }, { "epoch": 0.7712648743940062, "grad_norm": 0.29324751667885357, "learning_rate": 0.00018534563764708206, "loss": 1.8161, "step": 1750 }, { "epoch": 0.7734684883208461, "grad_norm": 0.2517432199347208, "learning_rate": 0.00018521172305285236, "loss": 1.6512, "step": 1755 }, { "epoch": 0.7756721022476862, "grad_norm": 0.25142671678387746, "learning_rate": 0.00018507724820473556, "loss": 1.7221, "step": 1760 }, { "epoch": 0.7778757161745262, "grad_norm": 0.2851848328431096, "learning_rate": 0.00018494221398688307, "loss": 1.9137, "step": 1765 }, { "epoch": 0.7800793301013662, "grad_norm": 0.27922286690055653, "learning_rate": 0.00018480662128712389, "loss": 1.7529, "step": 1770 }, { "epoch": 0.7822829440282063, "grad_norm": 0.24188014554995038, "learning_rate": 0.00018467047099695905, "loss": 1.7036, "step": 1775 }, { "epoch": 0.7844865579550463, "grad_norm": 0.21832286860860883, "learning_rate": 0.00018453376401155562, "loss": 1.8127, "step": 1780 }, { "epoch": 0.7866901718818863, "grad_norm": 0.26225286533808284, "learning_rate": 0.00018439650122974087, "loss": 1.7398, "step": 1785 }, { "epoch": 0.7888937858087263, "grad_norm": 0.2585197080539731, "learning_rate": 0.0001842586835539964, "loss": 1.8645, "step": 1790 }, { "epoch": 0.7910973997355664, "grad_norm": 0.24265804391287996, "learning_rate": 0.00018412031189045196, "loss": 1.7356, "step": 1795 }, { "epoch": 0.7933010136624064, "grad_norm": 0.2590383769624523, "learning_rate": 0.00018398138714887993, "loss": 1.6518, "step": 1800 }, { "epoch": 0.7955046275892463, "grad_norm": 0.23488722647726504, "learning_rate": 0.00018384191024268894, "loss": 1.8054, "step": 1805 }, { "epoch": 0.7977082415160864, "grad_norm": 0.23606528165986015, "learning_rate": 0.00018370188208891803, "loss": 1.6994, "step": 1810 }, { "epoch": 0.7999118554429264, "grad_norm": 0.29081075890767155, "learning_rate": 0.00018356130360823068, "loss": 1.987, "step": 1815 }, { "epoch": 0.8021154693697664, "grad_norm": 0.24366570601075402, "learning_rate": 0.00018342017572490858, "loss": 1.5363, "step": 1820 }, { "epoch": 0.8043190832966064, "grad_norm": 0.2677025615653137, "learning_rate": 0.0001832784993668458, "loss": 1.7781, "step": 1825 }, { "epoch": 0.8065226972234465, "grad_norm": 0.22776397402134876, "learning_rate": 0.0001831362754655424, "loss": 1.8064, "step": 1830 }, { "epoch": 0.8087263111502865, "grad_norm": 0.21327160658864644, "learning_rate": 0.0001829935049560985, "loss": 1.5278, "step": 1835 }, { "epoch": 0.8109299250771265, "grad_norm": 0.4311681362534179, "learning_rate": 0.0001828501887772081, "loss": 1.9316, "step": 1840 }, { "epoch": 0.8131335390039665, "grad_norm": 0.3127936675415962, "learning_rate": 0.00018270632787115295, "loss": 1.9393, "step": 1845 }, { "epoch": 0.8153371529308066, "grad_norm": 0.2872735828645766, "learning_rate": 0.0001825619231837962, "loss": 1.8913, "step": 1850 }, { "epoch": 0.8175407668576465, "grad_norm": 0.23659304750850696, "learning_rate": 0.0001824169756645763, "loss": 1.79, "step": 1855 }, { "epoch": 0.8197443807844865, "grad_norm": 0.22502963250778782, "learning_rate": 0.00018227148626650072, "loss": 1.7616, "step": 1860 }, { "epoch": 0.8219479947113266, "grad_norm": 0.22975193518596113, "learning_rate": 0.00018212545594613978, "loss": 1.7862, "step": 1865 }, { "epoch": 0.8241516086381666, "grad_norm": 0.22813094138391649, "learning_rate": 0.00018197888566362023, "loss": 1.6909, "step": 1870 }, { "epoch": 0.8263552225650066, "grad_norm": 0.3170639776475332, "learning_rate": 0.00018183177638261895, "loss": 1.8876, "step": 1875 }, { "epoch": 0.8285588364918466, "grad_norm": 0.25433638493270494, "learning_rate": 0.00018168412907035672, "loss": 1.7447, "step": 1880 }, { "epoch": 0.8307624504186867, "grad_norm": 0.2527115922098035, "learning_rate": 0.00018153594469759175, "loss": 1.7288, "step": 1885 }, { "epoch": 0.8329660643455267, "grad_norm": 0.24448176152726306, "learning_rate": 0.00018138722423861333, "loss": 1.8385, "step": 1890 }, { "epoch": 0.8351696782723667, "grad_norm": 0.2797043743850419, "learning_rate": 0.00018123796867123548, "loss": 1.8353, "step": 1895 }, { "epoch": 0.8373732921992068, "grad_norm": 0.26948384291872024, "learning_rate": 0.00018108817897679043, "loss": 1.6995, "step": 1900 }, { "epoch": 0.8395769061260467, "grad_norm": 0.23050209288787957, "learning_rate": 0.00018093785614012228, "loss": 1.7752, "step": 1905 }, { "epoch": 0.8417805200528867, "grad_norm": 0.28186670258249874, "learning_rate": 0.0001807870011495803, "loss": 1.8608, "step": 1910 }, { "epoch": 0.8439841339797267, "grad_norm": 0.29105838653881044, "learning_rate": 0.00018063561499701282, "loss": 1.8997, "step": 1915 }, { "epoch": 0.8461877479065668, "grad_norm": 0.29426186229081736, "learning_rate": 0.00018048369867776029, "loss": 1.6416, "step": 1920 }, { "epoch": 0.8483913618334068, "grad_norm": 0.21168547827802603, "learning_rate": 0.00018033125319064902, "loss": 1.8158, "step": 1925 }, { "epoch": 0.8505949757602468, "grad_norm": 0.35188920795977724, "learning_rate": 0.00018017827953798444, "loss": 1.7531, "step": 1930 }, { "epoch": 0.8527985896870868, "grad_norm": 0.2979347408479712, "learning_rate": 0.0001800247787255447, "loss": 1.9657, "step": 1935 }, { "epoch": 0.8550022036139269, "grad_norm": 0.3102746795792285, "learning_rate": 0.00017987075176257382, "loss": 1.5273, "step": 1940 }, { "epoch": 0.8572058175407669, "grad_norm": 0.2141203314519693, "learning_rate": 0.00017971619966177524, "loss": 1.7978, "step": 1945 }, { "epoch": 0.8594094314676068, "grad_norm": 0.3010217472666534, "learning_rate": 0.00017956112343930512, "loss": 1.8066, "step": 1950 }, { "epoch": 0.8616130453944469, "grad_norm": 0.2685511500450377, "learning_rate": 0.00017940552411476566, "loss": 1.8096, "step": 1955 }, { "epoch": 0.8638166593212869, "grad_norm": 0.3648269512255727, "learning_rate": 0.00017924940271119827, "loss": 1.9212, "step": 1960 }, { "epoch": 0.8660202732481269, "grad_norm": 0.23673592232409962, "learning_rate": 0.00017909276025507696, "loss": 1.9925, "step": 1965 }, { "epoch": 0.8682238871749669, "grad_norm": 0.22051543871837528, "learning_rate": 0.00017893559777630173, "loss": 1.8895, "step": 1970 }, { "epoch": 0.870427501101807, "grad_norm": 0.25128906373836973, "learning_rate": 0.00017877791630819149, "loss": 1.7637, "step": 1975 }, { "epoch": 0.872631115028647, "grad_norm": 0.2982588862018805, "learning_rate": 0.00017861971688747747, "loss": 1.865, "step": 1980 }, { "epoch": 0.874834728955487, "grad_norm": 0.2927080940082808, "learning_rate": 0.00017846100055429642, "loss": 1.742, "step": 1985 }, { "epoch": 0.8770383428823271, "grad_norm": 0.30412581475879524, "learning_rate": 0.00017830176835218368, "loss": 1.6706, "step": 1990 }, { "epoch": 0.8792419568091671, "grad_norm": 0.2902194881075315, "learning_rate": 0.0001781420213280662, "loss": 1.8014, "step": 1995 }, { "epoch": 0.881445570736007, "grad_norm": 0.3378104042442177, "learning_rate": 0.00017798176053225606, "loss": 1.8318, "step": 2000 }, { "epoch": 0.883649184662847, "grad_norm": 0.2465500138547731, "learning_rate": 0.0001778209870184431, "loss": 1.6756, "step": 2005 }, { "epoch": 0.8858527985896871, "grad_norm": 0.27205493094420413, "learning_rate": 0.00017765970184368835, "loss": 1.7398, "step": 2010 }, { "epoch": 0.8880564125165271, "grad_norm": 0.28881718844442433, "learning_rate": 0.0001774979060684168, "loss": 1.8652, "step": 2015 }, { "epoch": 0.8902600264433671, "grad_norm": 0.2930186193777016, "learning_rate": 0.0001773356007564107, "loss": 1.7748, "step": 2020 }, { "epoch": 0.8924636403702071, "grad_norm": 0.28319545050901546, "learning_rate": 0.0001771727869748023, "loss": 1.7198, "step": 2025 }, { "epoch": 0.8946672542970472, "grad_norm": 0.26485391899745814, "learning_rate": 0.000177009465794067, "loss": 1.7109, "step": 2030 }, { "epoch": 0.8968708682238872, "grad_norm": 0.2367084302014588, "learning_rate": 0.0001768456382880163, "loss": 1.773, "step": 2035 }, { "epoch": 0.8990744821507272, "grad_norm": 0.28707015291179266, "learning_rate": 0.00017668130553379063, "loss": 1.8698, "step": 2040 }, { "epoch": 0.9012780960775673, "grad_norm": 0.26521748140973644, "learning_rate": 0.00017651646861185252, "loss": 1.5433, "step": 2045 }, { "epoch": 0.9034817100044072, "grad_norm": 0.25186929013428017, "learning_rate": 0.0001763511286059791, "loss": 1.7003, "step": 2050 }, { "epoch": 0.9056853239312472, "grad_norm": 0.2646341657457682, "learning_rate": 0.0001761852866032554, "loss": 1.8017, "step": 2055 }, { "epoch": 0.9078889378580872, "grad_norm": 0.27426709503702656, "learning_rate": 0.0001760189436940669, "loss": 1.717, "step": 2060 }, { "epoch": 0.9100925517849273, "grad_norm": 0.2909513812455515, "learning_rate": 0.00017585210097209242, "loss": 1.8286, "step": 2065 }, { "epoch": 0.9122961657117673, "grad_norm": 0.3195261689470233, "learning_rate": 0.00017568475953429706, "loss": 1.9248, "step": 2070 }, { "epoch": 0.9144997796386073, "grad_norm": 0.2559454772629871, "learning_rate": 0.00017551692048092487, "loss": 1.9235, "step": 2075 }, { "epoch": 0.9167033935654474, "grad_norm": 0.2787085951940099, "learning_rate": 0.00017534858491549167, "loss": 1.5563, "step": 2080 }, { "epoch": 0.9189070074922874, "grad_norm": 0.27430346968172775, "learning_rate": 0.00017517975394477765, "loss": 1.7408, "step": 2085 }, { "epoch": 0.9211106214191274, "grad_norm": 0.3057530869061833, "learning_rate": 0.00017501042867882043, "loss": 1.8029, "step": 2090 }, { "epoch": 0.9233142353459673, "grad_norm": 0.28792933063158205, "learning_rate": 0.0001748406102309073, "loss": 1.7174, "step": 2095 }, { "epoch": 0.9255178492728074, "grad_norm": 0.2926998767200136, "learning_rate": 0.00017467029971756837, "loss": 1.7753, "step": 2100 }, { "epoch": 0.9277214631996474, "grad_norm": 0.3202640628143745, "learning_rate": 0.00017449949825856881, "loss": 1.6815, "step": 2105 }, { "epoch": 0.9299250771264874, "grad_norm": 0.2898987783743886, "learning_rate": 0.00017432820697690183, "loss": 1.5471, "step": 2110 }, { "epoch": 0.9321286910533274, "grad_norm": 0.25429358796976326, "learning_rate": 0.00017415642699878108, "loss": 1.801, "step": 2115 }, { "epoch": 0.9343323049801675, "grad_norm": 0.2906871474145296, "learning_rate": 0.00017398415945363326, "loss": 1.7255, "step": 2120 }, { "epoch": 0.9365359189070075, "grad_norm": 0.24941394640440692, "learning_rate": 0.00017381140547409091, "loss": 1.7265, "step": 2125 }, { "epoch": 0.9387395328338475, "grad_norm": 0.2680278555985492, "learning_rate": 0.00017363816619598462, "loss": 1.8507, "step": 2130 }, { "epoch": 0.9409431467606876, "grad_norm": 0.24326339969260075, "learning_rate": 0.00017346444275833587, "loss": 1.8278, "step": 2135 }, { "epoch": 0.9431467606875276, "grad_norm": 0.2881192799230667, "learning_rate": 0.00017329023630334935, "loss": 1.6301, "step": 2140 }, { "epoch": 0.9453503746143676, "grad_norm": 0.22324975257691354, "learning_rate": 0.00017311554797640552, "loss": 1.8182, "step": 2145 }, { "epoch": 0.9475539885412075, "grad_norm": 0.38310870823903137, "learning_rate": 0.0001729403789260531, "loss": 1.6758, "step": 2150 }, { "epoch": 0.9497576024680476, "grad_norm": 0.3082503271893105, "learning_rate": 0.0001727647303040015, "loss": 1.717, "step": 2155 }, { "epoch": 0.9519612163948876, "grad_norm": 0.2910152870088472, "learning_rate": 0.00017258860326511318, "loss": 1.6762, "step": 2160 }, { "epoch": 0.9541648303217276, "grad_norm": 0.2751619266213339, "learning_rate": 0.00017241199896739614, "loss": 1.5402, "step": 2165 }, { "epoch": 0.9563684442485677, "grad_norm": 0.2633670365064433, "learning_rate": 0.00017223491857199636, "loss": 1.6089, "step": 2170 }, { "epoch": 0.9585720581754077, "grad_norm": 0.26253468282607706, "learning_rate": 0.00017205736324318999, "loss": 1.8698, "step": 2175 }, { "epoch": 0.9607756721022477, "grad_norm": 0.2623810146620032, "learning_rate": 0.0001718793341483758, "loss": 1.7996, "step": 2180 }, { "epoch": 0.9629792860290877, "grad_norm": 0.2541467303967672, "learning_rate": 0.00017170083245806757, "loss": 1.7066, "step": 2185 }, { "epoch": 0.9651828999559278, "grad_norm": 0.2450837271489877, "learning_rate": 0.00017152185934588623, "loss": 1.9326, "step": 2190 }, { "epoch": 0.9673865138827678, "grad_norm": 0.3277904173770177, "learning_rate": 0.00017134241598855236, "loss": 1.941, "step": 2195 }, { "epoch": 0.9695901278096077, "grad_norm": 0.26036896345598354, "learning_rate": 0.0001711625035658782, "loss": 1.7543, "step": 2200 }, { "epoch": 0.9717937417364477, "grad_norm": 0.294563677729738, "learning_rate": 0.00017098212326076008, "loss": 1.6402, "step": 2205 }, { "epoch": 0.9739973556632878, "grad_norm": 0.3220804392466043, "learning_rate": 0.0001708012762591706, "loss": 1.7078, "step": 2210 }, { "epoch": 0.9762009695901278, "grad_norm": 0.2811686235748074, "learning_rate": 0.00017061996375015078, "loss": 1.9067, "step": 2215 }, { "epoch": 0.9784045835169678, "grad_norm": 0.23864027527398188, "learning_rate": 0.00017043818692580228, "loss": 1.9146, "step": 2220 }, { "epoch": 0.9806081974438079, "grad_norm": 0.2954583734368435, "learning_rate": 0.00017025594698127965, "loss": 1.791, "step": 2225 }, { "epoch": 0.9828118113706479, "grad_norm": 0.2510099463630807, "learning_rate": 0.00017007324511478223, "loss": 1.5883, "step": 2230 }, { "epoch": 0.9850154252974879, "grad_norm": 0.27874972856939056, "learning_rate": 0.00016989008252754655, "loss": 1.7865, "step": 2235 }, { "epoch": 0.9872190392243279, "grad_norm": 0.2611091563705049, "learning_rate": 0.00016970646042383826, "loss": 1.8104, "step": 2240 }, { "epoch": 0.989422653151168, "grad_norm": 0.28511909006315234, "learning_rate": 0.00016952238001094428, "loss": 1.6686, "step": 2245 }, { "epoch": 0.9916262670780079, "grad_norm": 0.3381628494565006, "learning_rate": 0.00016933784249916476, "loss": 1.9412, "step": 2250 }, { "epoch": 0.9938298810048479, "grad_norm": 0.28520119611042416, "learning_rate": 0.00016915284910180533, "loss": 1.6889, "step": 2255 }, { "epoch": 0.996033494931688, "grad_norm": 0.2883852687543481, "learning_rate": 0.00016896740103516895, "loss": 1.8003, "step": 2260 }, { "epoch": 0.998237108858528, "grad_norm": 0.24451797378236065, "learning_rate": 0.0001687814995185479, "loss": 1.8752, "step": 2265 }, { "epoch": 1.000440722785368, "grad_norm": 0.33508649606500057, "learning_rate": 0.00016859514577421592, "loss": 1.8299, "step": 2270 }, { "epoch": 1.002644336712208, "grad_norm": 0.2636718535549577, "learning_rate": 0.00016840834102741997, "loss": 1.706, "step": 2275 }, { "epoch": 1.004847950639048, "grad_norm": 0.27046399540212357, "learning_rate": 0.00016822108650637238, "loss": 1.6819, "step": 2280 }, { "epoch": 1.007051564565888, "grad_norm": 0.30877609431387876, "learning_rate": 0.00016803338344224266, "loss": 1.7218, "step": 2285 }, { "epoch": 1.0092551784927282, "grad_norm": 0.23492423335336904, "learning_rate": 0.00016784523306914934, "loss": 1.57, "step": 2290 }, { "epoch": 1.0114587924195682, "grad_norm": 0.2795118641324036, "learning_rate": 0.00016765663662415204, "loss": 1.7023, "step": 2295 }, { "epoch": 1.0136624063464081, "grad_norm": 0.34933397332267874, "learning_rate": 0.00016746759534724316, "loss": 1.9401, "step": 2300 }, { "epoch": 1.0158660202732481, "grad_norm": 0.2928363020374088, "learning_rate": 0.00016727811048133985, "loss": 1.7873, "step": 2305 }, { "epoch": 1.0180696342000881, "grad_norm": 0.2765841121495588, "learning_rate": 0.00016708818327227574, "loss": 1.6925, "step": 2310 }, { "epoch": 1.020273248126928, "grad_norm": 0.3407816261608907, "learning_rate": 0.00016689781496879283, "loss": 1.7583, "step": 2315 }, { "epoch": 1.022476862053768, "grad_norm": 0.3837729171353601, "learning_rate": 0.00016670700682253328, "loss": 1.7058, "step": 2320 }, { "epoch": 1.0246804759806083, "grad_norm": 0.2833490201778371, "learning_rate": 0.00016651576008803112, "loss": 1.7306, "step": 2325 }, { "epoch": 1.0268840899074483, "grad_norm": 0.24052446562056756, "learning_rate": 0.00016632407602270398, "loss": 1.6612, "step": 2330 }, { "epoch": 1.0290877038342883, "grad_norm": 0.2706767115179465, "learning_rate": 0.00016613195588684488, "loss": 1.5943, "step": 2335 }, { "epoch": 1.0312913177611283, "grad_norm": 0.2694153548940927, "learning_rate": 0.00016593940094361407, "loss": 1.7072, "step": 2340 }, { "epoch": 1.0334949316879682, "grad_norm": 0.2675415111083724, "learning_rate": 0.0001657464124590304, "loss": 1.7392, "step": 2345 }, { "epoch": 1.0356985456148082, "grad_norm": 0.2783138545253644, "learning_rate": 0.00016555299170196332, "loss": 1.7264, "step": 2350 }, { "epoch": 1.0379021595416482, "grad_norm": 0.29818023025744944, "learning_rate": 0.00016535913994412436, "loss": 1.6038, "step": 2355 }, { "epoch": 1.0401057734684884, "grad_norm": 0.30521070526088534, "learning_rate": 0.00016516485846005882, "loss": 1.652, "step": 2360 }, { "epoch": 1.0423093873953284, "grad_norm": 0.24990284589728196, "learning_rate": 0.00016497014852713738, "loss": 1.5745, "step": 2365 }, { "epoch": 1.0445130013221684, "grad_norm": 0.36544530123344093, "learning_rate": 0.0001647750114255477, "loss": 1.8678, "step": 2370 }, { "epoch": 1.0467166152490084, "grad_norm": 0.2984936690547218, "learning_rate": 0.000164579448438286, "loss": 1.8263, "step": 2375 }, { "epoch": 1.0489202291758484, "grad_norm": 0.3138432337464576, "learning_rate": 0.00016438346085114865, "loss": 1.7807, "step": 2380 }, { "epoch": 1.0511238431026884, "grad_norm": 0.409332779624452, "learning_rate": 0.00016418704995272373, "loss": 1.8612, "step": 2385 }, { "epoch": 1.0533274570295283, "grad_norm": 0.27134598526523496, "learning_rate": 0.00016399021703438247, "loss": 1.7323, "step": 2390 }, { "epoch": 1.0555310709563686, "grad_norm": 0.2852569861948395, "learning_rate": 0.0001637929633902708, "loss": 1.7619, "step": 2395 }, { "epoch": 1.0577346848832085, "grad_norm": 0.2913405836362002, "learning_rate": 0.00016359529031730093, "loss": 1.8196, "step": 2400 }, { "epoch": 1.0599382988100485, "grad_norm": 0.3029325708755559, "learning_rate": 0.00016339719911514272, "loss": 1.7579, "step": 2405 }, { "epoch": 1.0621419127368885, "grad_norm": 0.2738995593865038, "learning_rate": 0.00016319869108621512, "loss": 1.8309, "step": 2410 }, { "epoch": 1.0643455266637285, "grad_norm": 0.26503420605103173, "learning_rate": 0.00016299976753567772, "loss": 1.708, "step": 2415 }, { "epoch": 1.0665491405905685, "grad_norm": 0.32283303116470646, "learning_rate": 0.00016280042977142204, "loss": 1.6915, "step": 2420 }, { "epoch": 1.0687527545174085, "grad_norm": 0.24107839513422735, "learning_rate": 0.00016260067910406304, "loss": 1.5685, "step": 2425 }, { "epoch": 1.0709563684442487, "grad_norm": 0.29608435429464214, "learning_rate": 0.00016240051684693042, "loss": 1.7239, "step": 2430 }, { "epoch": 1.0731599823710887, "grad_norm": 0.29986988301001716, "learning_rate": 0.00016219994431606005, "loss": 1.6816, "step": 2435 }, { "epoch": 1.0753635962979287, "grad_norm": 0.24843757070636935, "learning_rate": 0.00016199896283018527, "loss": 1.5677, "step": 2440 }, { "epoch": 1.0775672102247686, "grad_norm": 0.2983161128962509, "learning_rate": 0.00016179757371072824, "loss": 1.7859, "step": 2445 }, { "epoch": 1.0797708241516086, "grad_norm": 0.2787284156730912, "learning_rate": 0.00016159577828179123, "loss": 1.562, "step": 2450 }, { "epoch": 1.0819744380784486, "grad_norm": 0.3269743373279123, "learning_rate": 0.0001613935778701479, "loss": 1.8759, "step": 2455 }, { "epoch": 1.0841780520052886, "grad_norm": 0.2802831939826356, "learning_rate": 0.0001611909738052347, "loss": 1.7401, "step": 2460 }, { "epoch": 1.0863816659321286, "grad_norm": 0.23846388238067243, "learning_rate": 0.000160987967419142, "loss": 1.727, "step": 2465 }, { "epoch": 1.0885852798589688, "grad_norm": 0.2659078946356941, "learning_rate": 0.00016078456004660536, "loss": 1.6454, "step": 2470 }, { "epoch": 1.0907888937858088, "grad_norm": 0.29111397323788823, "learning_rate": 0.00016058075302499673, "loss": 1.7724, "step": 2475 }, { "epoch": 1.0929925077126488, "grad_norm": 0.35649745669814514, "learning_rate": 0.00016037654769431576, "loss": 1.6527, "step": 2480 }, { "epoch": 1.0951961216394888, "grad_norm": 0.28097148721584675, "learning_rate": 0.00016017194539718086, "loss": 1.7563, "step": 2485 }, { "epoch": 1.0973997355663287, "grad_norm": 0.39237884989816185, "learning_rate": 0.0001599669474788205, "loss": 1.8656, "step": 2490 }, { "epoch": 1.0996033494931687, "grad_norm": 0.2633010799321111, "learning_rate": 0.00015976155528706415, "loss": 1.7129, "step": 2495 }, { "epoch": 1.1018069634200087, "grad_norm": 0.2719172135219212, "learning_rate": 0.0001595557701723338, "loss": 1.5688, "step": 2500 }, { "epoch": 1.104010577346849, "grad_norm": 0.2731378192802766, "learning_rate": 0.00015934959348763467, "loss": 1.7727, "step": 2505 }, { "epoch": 1.106214191273689, "grad_norm": 0.31808171740983593, "learning_rate": 0.00015914302658854657, "loss": 1.8461, "step": 2510 }, { "epoch": 1.108417805200529, "grad_norm": 0.2869427974730242, "learning_rate": 0.00015893607083321477, "loss": 1.8664, "step": 2515 }, { "epoch": 1.110621419127369, "grad_norm": 0.2187585282755932, "learning_rate": 0.00015872872758234148, "loss": 1.6029, "step": 2520 }, { "epoch": 1.1128250330542089, "grad_norm": 0.2667619562054791, "learning_rate": 0.00015852099819917639, "loss": 1.8981, "step": 2525 }, { "epoch": 1.1150286469810489, "grad_norm": 0.3372014415136272, "learning_rate": 0.00015831288404950802, "loss": 1.7639, "step": 2530 }, { "epoch": 1.1172322609078889, "grad_norm": 0.3048158336194716, "learning_rate": 0.0001581043865016547, "loss": 1.6387, "step": 2535 }, { "epoch": 1.1194358748347288, "grad_norm": 0.2971544124275263, "learning_rate": 0.00015789550692645556, "loss": 1.7692, "step": 2540 }, { "epoch": 1.121639488761569, "grad_norm": 0.2518439420968902, "learning_rate": 0.00015768624669726145, "loss": 1.5533, "step": 2545 }, { "epoch": 1.123843102688409, "grad_norm": 0.28947857100850594, "learning_rate": 0.00015747660718992598, "loss": 1.6443, "step": 2550 }, { "epoch": 1.126046716615249, "grad_norm": 0.27744579032481387, "learning_rate": 0.00015726658978279642, "loss": 1.7146, "step": 2555 }, { "epoch": 1.128250330542089, "grad_norm": 0.28434343297956843, "learning_rate": 0.00015705619585670478, "loss": 1.8686, "step": 2560 }, { "epoch": 1.130453944468929, "grad_norm": 0.26116640553101533, "learning_rate": 0.00015684542679495847, "loss": 1.7831, "step": 2565 }, { "epoch": 1.132657558395769, "grad_norm": 0.31380511545232626, "learning_rate": 0.00015663428398333157, "loss": 1.6778, "step": 2570 }, { "epoch": 1.134861172322609, "grad_norm": 0.2747114435356579, "learning_rate": 0.0001564227688100552, "loss": 1.6324, "step": 2575 }, { "epoch": 1.1370647862494492, "grad_norm": 0.2678076483406117, "learning_rate": 0.00015621088266580904, "loss": 1.4946, "step": 2580 }, { "epoch": 1.1392684001762892, "grad_norm": 0.33262666113926864, "learning_rate": 0.00015599862694371157, "loss": 1.882, "step": 2585 }, { "epoch": 1.1414720141031292, "grad_norm": 0.30861283361311576, "learning_rate": 0.00015578600303931136, "loss": 1.6738, "step": 2590 }, { "epoch": 1.1436756280299691, "grad_norm": 0.30927635505724477, "learning_rate": 0.00015557301235057767, "loss": 1.7005, "step": 2595 }, { "epoch": 1.1458792419568091, "grad_norm": 0.28688292761286854, "learning_rate": 0.00015535965627789126, "loss": 1.6462, "step": 2600 }, { "epoch": 1.1480828558836491, "grad_norm": 0.27171604870426835, "learning_rate": 0.00015514593622403532, "loss": 1.585, "step": 2605 }, { "epoch": 1.150286469810489, "grad_norm": 0.3092542662127764, "learning_rate": 0.0001549318535941861, "loss": 1.9096, "step": 2610 }, { "epoch": 1.1524900837373293, "grad_norm": 0.2716117253477657, "learning_rate": 0.00015471740979590377, "loss": 1.7352, "step": 2615 }, { "epoch": 1.1546936976641693, "grad_norm": 0.3066935084879198, "learning_rate": 0.0001545026062391231, "loss": 1.8141, "step": 2620 }, { "epoch": 1.1568973115910093, "grad_norm": 0.346363671435922, "learning_rate": 0.00015428744433614415, "loss": 1.5573, "step": 2625 }, { "epoch": 1.1591009255178493, "grad_norm": 0.29157425354464905, "learning_rate": 0.00015407192550162318, "loss": 1.5464, "step": 2630 }, { "epoch": 1.1613045394446893, "grad_norm": 0.2806374114105095, "learning_rate": 0.0001538560511525632, "loss": 1.6386, "step": 2635 }, { "epoch": 1.1635081533715292, "grad_norm": 0.2982356223064332, "learning_rate": 0.0001536398227083046, "loss": 1.7813, "step": 2640 }, { "epoch": 1.1657117672983692, "grad_norm": 0.33202427377524757, "learning_rate": 0.00015342324159051587, "loss": 1.9532, "step": 2645 }, { "epoch": 1.1679153812252094, "grad_norm": 0.33897582151852124, "learning_rate": 0.00015320630922318444, "loss": 1.7746, "step": 2650 }, { "epoch": 1.1701189951520494, "grad_norm": 0.3321654404402584, "learning_rate": 0.00015298902703260692, "loss": 2.0143, "step": 2655 }, { "epoch": 1.1723226090788894, "grad_norm": 0.21833177585011845, "learning_rate": 0.0001527713964473802, "loss": 1.6702, "step": 2660 }, { "epoch": 1.1745262230057294, "grad_norm": 0.26595318714741284, "learning_rate": 0.00015255341889839157, "loss": 1.683, "step": 2665 }, { "epoch": 1.1767298369325694, "grad_norm": 0.2880553889858748, "learning_rate": 0.00015233509581880973, "loss": 1.6248, "step": 2670 }, { "epoch": 1.1789334508594094, "grad_norm": 0.2966201251141622, "learning_rate": 0.0001521164286440751, "loss": 1.5788, "step": 2675 }, { "epoch": 1.1811370647862494, "grad_norm": 0.32819026600076084, "learning_rate": 0.00015189741881189054, "loss": 1.6132, "step": 2680 }, { "epoch": 1.1833406787130896, "grad_norm": 0.25598509051489227, "learning_rate": 0.00015167806776221178, "loss": 1.7194, "step": 2685 }, { "epoch": 1.1855442926399296, "grad_norm": 0.32216524743892266, "learning_rate": 0.000151458376937238, "loss": 1.6499, "step": 2690 }, { "epoch": 1.1877479065667695, "grad_norm": 0.25107175983686675, "learning_rate": 0.00015123834778140233, "loss": 1.6059, "step": 2695 }, { "epoch": 1.1899515204936095, "grad_norm": 0.26100544568346645, "learning_rate": 0.00015101798174136247, "loss": 1.7677, "step": 2700 }, { "epoch": 1.1921551344204495, "grad_norm": 0.28164900490660866, "learning_rate": 0.000150797280265991, "loss": 1.6141, "step": 2705 }, { "epoch": 1.1943587483472895, "grad_norm": 0.25606567943961145, "learning_rate": 0.00015057624480636594, "loss": 1.6868, "step": 2710 }, { "epoch": 1.1965623622741295, "grad_norm": 0.30536561643444826, "learning_rate": 0.0001503548768157612, "loss": 1.515, "step": 2715 }, { "epoch": 1.1987659762009697, "grad_norm": 0.2495814195185963, "learning_rate": 0.00015013317774963708, "loss": 1.5754, "step": 2720 }, { "epoch": 1.2009695901278097, "grad_norm": 0.3292329862356163, "learning_rate": 0.00014991114906563055, "loss": 1.6599, "step": 2725 }, { "epoch": 1.2031732040546497, "grad_norm": 0.28736997494401945, "learning_rate": 0.00014968879222354597, "loss": 1.6939, "step": 2730 }, { "epoch": 1.2053768179814897, "grad_norm": 0.28066545139959265, "learning_rate": 0.00014946610868534502, "loss": 1.5954, "step": 2735 }, { "epoch": 1.2075804319083296, "grad_norm": 0.27963157670324124, "learning_rate": 0.00014924309991513757, "loss": 1.6816, "step": 2740 }, { "epoch": 1.2097840458351696, "grad_norm": 0.3074722093066757, "learning_rate": 0.0001490197673791717, "loss": 1.5102, "step": 2745 }, { "epoch": 1.2119876597620096, "grad_norm": 0.5175039724428968, "learning_rate": 0.00014879611254582428, "loss": 1.8587, "step": 2750 }, { "epoch": 1.2141912736888498, "grad_norm": 0.4607644456138267, "learning_rate": 0.00014857213688559124, "loss": 1.8861, "step": 2755 }, { "epoch": 1.2163948876156898, "grad_norm": 0.203471979843397, "learning_rate": 0.00014834784187107785, "loss": 1.5549, "step": 2760 }, { "epoch": 1.2185985015425298, "grad_norm": 0.27836500512738993, "learning_rate": 0.00014812322897698912, "loss": 1.6677, "step": 2765 }, { "epoch": 1.2208021154693698, "grad_norm": 0.2681432382916657, "learning_rate": 0.00014789829968012, "loss": 1.8601, "step": 2770 }, { "epoch": 1.2230057293962098, "grad_norm": 0.24822408823990583, "learning_rate": 0.00014767305545934588, "loss": 1.9008, "step": 2775 }, { "epoch": 1.2252093433230498, "grad_norm": 0.26961514947075566, "learning_rate": 0.00014744749779561258, "loss": 1.7573, "step": 2780 }, { "epoch": 1.2274129572498897, "grad_norm": 0.40650037835816966, "learning_rate": 0.0001472216281719269, "loss": 1.6177, "step": 2785 }, { "epoch": 1.22961657117673, "grad_norm": 0.27436209441284687, "learning_rate": 0.0001469954480733465, "loss": 1.6021, "step": 2790 }, { "epoch": 1.23182018510357, "grad_norm": 0.2563256590060921, "learning_rate": 0.00014676895898697062, "loss": 1.9842, "step": 2795 }, { "epoch": 1.23402379903041, "grad_norm": 0.35591941682342815, "learning_rate": 0.00014654216240192995, "loss": 1.6028, "step": 2800 }, { "epoch": 1.23622741295725, "grad_norm": 0.33349441263276575, "learning_rate": 0.00014631505980937688, "loss": 1.852, "step": 2805 }, { "epoch": 1.23843102688409, "grad_norm": 0.31528740240587627, "learning_rate": 0.0001460876527024758, "loss": 1.5587, "step": 2810 }, { "epoch": 1.24063464081093, "grad_norm": 0.25877213395041015, "learning_rate": 0.00014585994257639324, "loss": 1.5482, "step": 2815 }, { "epoch": 1.2428382547377699, "grad_norm": 0.2910006006185105, "learning_rate": 0.00014563193092828803, "loss": 1.6998, "step": 2820 }, { "epoch": 1.24504186866461, "grad_norm": 0.37486054420050446, "learning_rate": 0.00014540361925730147, "loss": 1.8516, "step": 2825 }, { "epoch": 1.24724548259145, "grad_norm": 0.34335913235176224, "learning_rate": 0.00014517500906454742, "loss": 1.6384, "step": 2830 }, { "epoch": 1.24944909651829, "grad_norm": 0.2930060418053195, "learning_rate": 0.00014494610185310252, "loss": 1.8508, "step": 2835 }, { "epoch": 1.25165271044513, "grad_norm": 0.27045205482633095, "learning_rate": 0.00014471689912799626, "loss": 1.5935, "step": 2840 }, { "epoch": 1.25385632437197, "grad_norm": 0.30870157000744336, "learning_rate": 0.00014448740239620108, "loss": 1.8287, "step": 2845 }, { "epoch": 1.25605993829881, "grad_norm": 0.27087387170107313, "learning_rate": 0.00014425761316662241, "loss": 1.9209, "step": 2850 }, { "epoch": 1.25826355222565, "grad_norm": 0.32362518237545235, "learning_rate": 0.0001440275329500889, "loss": 1.7297, "step": 2855 }, { "epoch": 1.2604671661524902, "grad_norm": 0.3473479960796504, "learning_rate": 0.00014379716325934236, "loss": 1.7847, "step": 2860 }, { "epoch": 1.26267078007933, "grad_norm": 0.2985625193084418, "learning_rate": 0.0001435665056090278, "loss": 1.7811, "step": 2865 }, { "epoch": 1.2648743940061702, "grad_norm": 0.2549368072217001, "learning_rate": 0.00014333556151568364, "loss": 1.8424, "step": 2870 }, { "epoch": 1.2670780079330102, "grad_norm": 0.3031064940015495, "learning_rate": 0.00014310433249773146, "loss": 1.8502, "step": 2875 }, { "epoch": 1.2692816218598502, "grad_norm": 0.2398504180714546, "learning_rate": 0.00014287282007546627, "loss": 1.648, "step": 2880 }, { "epoch": 1.2714852357866901, "grad_norm": 0.33873410733492354, "learning_rate": 0.00014264102577104645, "loss": 1.6617, "step": 2885 }, { "epoch": 1.2736888497135301, "grad_norm": 0.2655552963090036, "learning_rate": 0.00014240895110848365, "loss": 1.7205, "step": 2890 }, { "epoch": 1.2758924636403703, "grad_norm": 0.30714278749605195, "learning_rate": 0.0001421765976136328, "loss": 1.7343, "step": 2895 }, { "epoch": 1.27809607756721, "grad_norm": 0.31410869299454564, "learning_rate": 0.0001419439668141822, "loss": 1.7369, "step": 2900 }, { "epoch": 1.2802996914940503, "grad_norm": 0.29231072094243427, "learning_rate": 0.0001417110602396434, "loss": 1.5317, "step": 2905 }, { "epoch": 1.2825033054208903, "grad_norm": 0.26156390622111436, "learning_rate": 0.00014147787942134089, "loss": 1.4907, "step": 2910 }, { "epoch": 1.2847069193477303, "grad_norm": 0.2432071100976, "learning_rate": 0.00014124442589240265, "loss": 1.7181, "step": 2915 }, { "epoch": 1.2869105332745703, "grad_norm": 0.28813004538064096, "learning_rate": 0.00014101070118774936, "loss": 1.7243, "step": 2920 }, { "epoch": 1.2891141472014103, "grad_norm": 0.29339284132217475, "learning_rate": 0.00014077670684408485, "loss": 1.7679, "step": 2925 }, { "epoch": 1.2913177611282503, "grad_norm": 0.2588877094640161, "learning_rate": 0.00014054244439988566, "loss": 1.72, "step": 2930 }, { "epoch": 1.2935213750550902, "grad_norm": 0.3011003249431555, "learning_rate": 0.0001403079153953911, "loss": 1.879, "step": 2935 }, { "epoch": 1.2957249889819304, "grad_norm": 0.29129639306491034, "learning_rate": 0.00014007312137259307, "loss": 1.7124, "step": 2940 }, { "epoch": 1.2979286029087704, "grad_norm": 0.30248111079542994, "learning_rate": 0.00013983806387522592, "loss": 1.5669, "step": 2945 }, { "epoch": 1.3001322168356104, "grad_norm": 0.39466981086485026, "learning_rate": 0.00013960274444875628, "loss": 1.7579, "step": 2950 }, { "epoch": 1.3023358307624504, "grad_norm": 0.30172725615707247, "learning_rate": 0.000139367164640373, "loss": 1.8218, "step": 2955 }, { "epoch": 1.3045394446892904, "grad_norm": 0.286453848416204, "learning_rate": 0.00013913132599897683, "loss": 1.9354, "step": 2960 }, { "epoch": 1.3067430586161304, "grad_norm": 0.3902213358624171, "learning_rate": 0.00013889523007517028, "loss": 1.7235, "step": 2965 }, { "epoch": 1.3089466725429704, "grad_norm": 0.2589601846721671, "learning_rate": 0.00013865887842124755, "loss": 1.5088, "step": 2970 }, { "epoch": 1.3111502864698106, "grad_norm": 0.2838598126846581, "learning_rate": 0.0001384222725911842, "loss": 1.6694, "step": 2975 }, { "epoch": 1.3133539003966506, "grad_norm": 0.2985112357585295, "learning_rate": 0.00013818541414062683, "loss": 1.8195, "step": 2980 }, { "epoch": 1.3155575143234906, "grad_norm": 0.344496826298518, "learning_rate": 0.0001379483046268832, "loss": 1.7105, "step": 2985 }, { "epoch": 1.3177611282503305, "grad_norm": 0.2832321651335367, "learning_rate": 0.00013771094560891155, "loss": 1.6398, "step": 2990 }, { "epoch": 1.3199647421771705, "grad_norm": 0.3166967200246273, "learning_rate": 0.00013747333864731073, "loss": 1.8804, "step": 2995 }, { "epoch": 1.3221683561040105, "grad_norm": 0.29889345173462817, "learning_rate": 0.00013723548530430974, "loss": 1.5327, "step": 3000 }, { "epoch": 1.3243719700308505, "grad_norm": 0.2989561597186618, "learning_rate": 0.00013699738714375748, "loss": 1.8312, "step": 3005 }, { "epoch": 1.3265755839576907, "grad_norm": 0.27767734489817053, "learning_rate": 0.00013675904573111247, "loss": 1.7797, "step": 3010 }, { "epoch": 1.3287791978845307, "grad_norm": 0.3263812983049982, "learning_rate": 0.00013652046263343262, "loss": 1.7061, "step": 3015 }, { "epoch": 1.3309828118113707, "grad_norm": 0.24589188706441673, "learning_rate": 0.00013628163941936485, "loss": 1.7644, "step": 3020 }, { "epoch": 1.3331864257382107, "grad_norm": 0.277800302556096, "learning_rate": 0.00013604257765913484, "loss": 1.8151, "step": 3025 }, { "epoch": 1.3353900396650507, "grad_norm": 0.23353363973323982, "learning_rate": 0.0001358032789245366, "loss": 1.7236, "step": 3030 }, { "epoch": 1.3375936535918906, "grad_norm": 0.34178889147907426, "learning_rate": 0.00013556374478892232, "loss": 1.7669, "step": 3035 }, { "epoch": 1.3397972675187306, "grad_norm": 0.28321134962453065, "learning_rate": 0.00013532397682719185, "loss": 1.6165, "step": 3040 }, { "epoch": 1.3420008814455708, "grad_norm": 0.3160168756862356, "learning_rate": 0.00013508397661578242, "loss": 1.8131, "step": 3045 }, { "epoch": 1.3442044953724108, "grad_norm": 0.2800833181613442, "learning_rate": 0.0001348437457326582, "loss": 1.9182, "step": 3050 }, { "epoch": 1.3464081092992508, "grad_norm": 0.2679976561310916, "learning_rate": 0.00013460328575730019, "loss": 1.8312, "step": 3055 }, { "epoch": 1.3486117232260908, "grad_norm": 0.2899158112172882, "learning_rate": 0.00013436259827069534, "loss": 1.8217, "step": 3060 }, { "epoch": 1.3508153371529308, "grad_norm": 0.26525449374755994, "learning_rate": 0.00013412168485532676, "loss": 1.7636, "step": 3065 }, { "epoch": 1.3530189510797708, "grad_norm": 0.2643909185128004, "learning_rate": 0.00013388054709516272, "loss": 1.6257, "step": 3070 }, { "epoch": 1.3552225650066108, "grad_norm": 0.3217683133395989, "learning_rate": 0.0001336391865756468, "loss": 1.8385, "step": 3075 }, { "epoch": 1.357426178933451, "grad_norm": 0.30256367251501726, "learning_rate": 0.00013339760488368695, "loss": 1.5994, "step": 3080 }, { "epoch": 1.359629792860291, "grad_norm": 0.33945781722722157, "learning_rate": 0.00013315580360764542, "loss": 1.6502, "step": 3085 }, { "epoch": 1.361833406787131, "grad_norm": 0.2343801676740979, "learning_rate": 0.00013291378433732818, "loss": 1.7302, "step": 3090 }, { "epoch": 1.364037020713971, "grad_norm": 0.3789024872984378, "learning_rate": 0.00013267154866397447, "loss": 1.9092, "step": 3095 }, { "epoch": 1.366240634640811, "grad_norm": 0.27745717625968813, "learning_rate": 0.00013242909818024628, "loss": 1.6587, "step": 3100 }, { "epoch": 1.368444248567651, "grad_norm": 0.26534498515018917, "learning_rate": 0.0001321864344802181, "loss": 1.7184, "step": 3105 }, { "epoch": 1.3706478624944909, "grad_norm": 0.25912670390615655, "learning_rate": 0.00013194355915936611, "loss": 1.7708, "step": 3110 }, { "epoch": 1.372851476421331, "grad_norm": 0.23803080715278022, "learning_rate": 0.000131700473814558, "loss": 1.7224, "step": 3115 }, { "epoch": 1.375055090348171, "grad_norm": 0.30746363012665606, "learning_rate": 0.00013145718004404223, "loss": 1.754, "step": 3120 }, { "epoch": 1.377258704275011, "grad_norm": 0.2672636477164184, "learning_rate": 0.00013121367944743777, "loss": 1.6989, "step": 3125 }, { "epoch": 1.379462318201851, "grad_norm": 0.28413083026015534, "learning_rate": 0.0001309699736257232, "loss": 1.6421, "step": 3130 }, { "epoch": 1.381665932128691, "grad_norm": 0.3193377255035563, "learning_rate": 0.00013072606418122667, "loss": 1.8467, "step": 3135 }, { "epoch": 1.383869546055531, "grad_norm": 0.2833583691387121, "learning_rate": 0.00013048195271761498, "loss": 1.6013, "step": 3140 }, { "epoch": 1.386073159982371, "grad_norm": 0.25616581325290944, "learning_rate": 0.00013023764083988323, "loss": 1.7542, "step": 3145 }, { "epoch": 1.3882767739092112, "grad_norm": 0.34179796686003233, "learning_rate": 0.0001299931301543442, "loss": 1.6674, "step": 3150 }, { "epoch": 1.390480387836051, "grad_norm": 0.2612422912889042, "learning_rate": 0.00012974842226861773, "loss": 1.3979, "step": 3155 }, { "epoch": 1.3926840017628912, "grad_norm": 0.30556335921079647, "learning_rate": 0.0001295035187916204, "loss": 1.7775, "step": 3160 }, { "epoch": 1.3948876156897312, "grad_norm": 0.34927377227523054, "learning_rate": 0.00012925842133355454, "loss": 1.7384, "step": 3165 }, { "epoch": 1.3970912296165712, "grad_norm": 0.35826503954646516, "learning_rate": 0.00012901313150589806, "loss": 1.8279, "step": 3170 }, { "epoch": 1.3992948435434112, "grad_norm": 0.3558470664702752, "learning_rate": 0.0001287676509213936, "loss": 1.6467, "step": 3175 }, { "epoch": 1.4014984574702511, "grad_norm": 0.2897608843662268, "learning_rate": 0.00012852198119403798, "loss": 1.6509, "step": 3180 }, { "epoch": 1.4037020713970914, "grad_norm": 0.32428607006990234, "learning_rate": 0.00012827612393907163, "loss": 1.7118, "step": 3185 }, { "epoch": 1.4059056853239311, "grad_norm": 0.29054441869310144, "learning_rate": 0.0001280300807729679, "loss": 1.6328, "step": 3190 }, { "epoch": 1.4081092992507713, "grad_norm": 0.2672079347199706, "learning_rate": 0.0001277838533134226, "loss": 1.7875, "step": 3195 }, { "epoch": 1.4103129131776113, "grad_norm": 0.3317588448314954, "learning_rate": 0.00012753744317934307, "loss": 1.9754, "step": 3200 }, { "epoch": 1.4125165271044513, "grad_norm": 0.30976057441678767, "learning_rate": 0.0001272908519908379, "loss": 1.7292, "step": 3205 }, { "epoch": 1.4147201410312913, "grad_norm": 0.2588899590749228, "learning_rate": 0.00012704408136920585, "loss": 1.661, "step": 3210 }, { "epoch": 1.4169237549581313, "grad_norm": 0.34042631761749226, "learning_rate": 0.0001267971329369256, "loss": 1.7688, "step": 3215 }, { "epoch": 1.4191273688849715, "grad_norm": 0.2613604846991468, "learning_rate": 0.00012655000831764495, "loss": 1.7979, "step": 3220 }, { "epoch": 1.4213309828118112, "grad_norm": 0.30387612038339795, "learning_rate": 0.00012630270913616985, "loss": 1.6008, "step": 3225 }, { "epoch": 1.4235345967386515, "grad_norm": 0.3245461529092582, "learning_rate": 0.00012605523701845431, "loss": 1.7394, "step": 3230 }, { "epoch": 1.4257382106654914, "grad_norm": 0.23964868596701627, "learning_rate": 0.00012580759359158905, "loss": 1.5526, "step": 3235 }, { "epoch": 1.4279418245923314, "grad_norm": 0.281150127445612, "learning_rate": 0.00012555978048379133, "loss": 1.6581, "step": 3240 }, { "epoch": 1.4301454385191714, "grad_norm": 0.31208534783523834, "learning_rate": 0.00012531179932439397, "loss": 1.6698, "step": 3245 }, { "epoch": 1.4323490524460114, "grad_norm": 0.3197645332854783, "learning_rate": 0.00012506365174383467, "loss": 1.8493, "step": 3250 }, { "epoch": 1.4345526663728516, "grad_norm": 0.29747281096057276, "learning_rate": 0.0001248153393736454, "loss": 1.923, "step": 3255 }, { "epoch": 1.4367562802996914, "grad_norm": 0.2706957926203667, "learning_rate": 0.00012456686384644148, "loss": 1.7219, "step": 3260 }, { "epoch": 1.4389598942265316, "grad_norm": 0.3015008988665459, "learning_rate": 0.00012431822679591112, "loss": 1.6334, "step": 3265 }, { "epoch": 1.4411635081533716, "grad_norm": 0.28824055515626146, "learning_rate": 0.00012406942985680437, "loss": 1.7096, "step": 3270 }, { "epoch": 1.4433671220802116, "grad_norm": 0.28792375010811966, "learning_rate": 0.00012382047466492262, "loss": 1.6993, "step": 3275 }, { "epoch": 1.4455707360070515, "grad_norm": 0.2778794221727809, "learning_rate": 0.0001235713628571077, "loss": 1.699, "step": 3280 }, { "epoch": 1.4477743499338915, "grad_norm": 0.3173174516544841, "learning_rate": 0.00012332209607123117, "loss": 1.6214, "step": 3285 }, { "epoch": 1.4499779638607315, "grad_norm": 0.30655650928697775, "learning_rate": 0.0001230726759461836, "loss": 1.7923, "step": 3290 }, { "epoch": 1.4521815777875715, "grad_norm": 0.2517175305379352, "learning_rate": 0.00012282310412186365, "loss": 1.8434, "step": 3295 }, { "epoch": 1.4543851917144117, "grad_norm": 0.24920356879351888, "learning_rate": 0.0001225733822391675, "loss": 1.6146, "step": 3300 }, { "epoch": 1.4565888056412517, "grad_norm": 0.2950517285567546, "learning_rate": 0.00012232351193997774, "loss": 1.6819, "step": 3305 }, { "epoch": 1.4587924195680917, "grad_norm": 0.3058013022960617, "learning_rate": 0.000122073494867153, "loss": 1.579, "step": 3310 }, { "epoch": 1.4609960334949317, "grad_norm": 0.46566643268469327, "learning_rate": 0.00012182333266451684, "loss": 1.6713, "step": 3315 }, { "epoch": 1.4631996474217717, "grad_norm": 0.3269081430168815, "learning_rate": 0.00012157302697684695, "loss": 1.6608, "step": 3320 }, { "epoch": 1.4654032613486117, "grad_norm": 0.2472677464376836, "learning_rate": 0.00012132257944986454, "loss": 1.6504, "step": 3325 }, { "epoch": 1.4676068752754516, "grad_norm": 0.3189526409165166, "learning_rate": 0.00012107199173022327, "loss": 1.6308, "step": 3330 }, { "epoch": 1.4698104892022918, "grad_norm": 0.2439726953786154, "learning_rate": 0.00012082126546549864, "loss": 1.6694, "step": 3335 }, { "epoch": 1.4720141031291318, "grad_norm": 0.28416189235796196, "learning_rate": 0.000120570402304177, "loss": 1.9048, "step": 3340 }, { "epoch": 1.4742177170559718, "grad_norm": 0.25976522560441834, "learning_rate": 0.00012031940389564478, "loss": 1.7083, "step": 3345 }, { "epoch": 1.4764213309828118, "grad_norm": 0.2874680056323443, "learning_rate": 0.00012006827189017773, "loss": 1.7914, "step": 3350 }, { "epoch": 1.4786249449096518, "grad_norm": 0.35212840385267163, "learning_rate": 0.00011981700793892982, "loss": 1.8617, "step": 3355 }, { "epoch": 1.4808285588364918, "grad_norm": 0.3258646795205973, "learning_rate": 0.00011956561369392274, "loss": 1.8569, "step": 3360 }, { "epoch": 1.4830321727633318, "grad_norm": 0.25990120954046436, "learning_rate": 0.0001193140908080346, "loss": 1.7778, "step": 3365 }, { "epoch": 1.485235786690172, "grad_norm": 0.2916499249746569, "learning_rate": 0.00011906244093498955, "loss": 1.7442, "step": 3370 }, { "epoch": 1.487439400617012, "grad_norm": 0.38118475856684764, "learning_rate": 0.00011881066572934644, "loss": 1.6281, "step": 3375 }, { "epoch": 1.489643014543852, "grad_norm": 0.33602332943649665, "learning_rate": 0.00011855876684648837, "loss": 1.6655, "step": 3380 }, { "epoch": 1.491846628470692, "grad_norm": 0.3091891353046593, "learning_rate": 0.00011830674594261145, "loss": 1.818, "step": 3385 }, { "epoch": 1.494050242397532, "grad_norm": 0.2675107541956203, "learning_rate": 0.0001180546046747141, "loss": 1.9917, "step": 3390 }, { "epoch": 1.496253856324372, "grad_norm": 0.24171145502227592, "learning_rate": 0.00011780234470058613, "loss": 1.747, "step": 3395 }, { "epoch": 1.498457470251212, "grad_norm": 0.31043683691075824, "learning_rate": 0.0001175499676787978, "loss": 1.7863, "step": 3400 }, { "epoch": 1.500661084178052, "grad_norm": 0.25938236019167105, "learning_rate": 0.000117297475268689, "loss": 1.6216, "step": 3405 }, { "epoch": 1.5028646981048919, "grad_norm": 0.2889407665894309, "learning_rate": 0.00011704486913035819, "loss": 1.7023, "step": 3410 }, { "epoch": 1.505068312031732, "grad_norm": 0.2595635779433101, "learning_rate": 0.00011679215092465163, "loss": 1.6651, "step": 3415 }, { "epoch": 1.507271925958572, "grad_norm": 0.3147360748379521, "learning_rate": 0.00011653932231315245, "loss": 1.6855, "step": 3420 }, { "epoch": 1.509475539885412, "grad_norm": 0.2456935617451673, "learning_rate": 0.00011628638495816955, "loss": 1.6982, "step": 3425 }, { "epoch": 1.511679153812252, "grad_norm": 0.2644287001406921, "learning_rate": 0.00011603334052272696, "loss": 1.7438, "step": 3430 }, { "epoch": 1.513882767739092, "grad_norm": 0.3098487244790836, "learning_rate": 0.0001157801906705526, "loss": 1.7459, "step": 3435 }, { "epoch": 1.5160863816659322, "grad_norm": 0.2635907160016382, "learning_rate": 0.00011552693706606758, "loss": 1.5969, "step": 3440 }, { "epoch": 1.518289995592772, "grad_norm": 0.293285003547433, "learning_rate": 0.00011527358137437516, "loss": 1.7899, "step": 3445 }, { "epoch": 1.5204936095196122, "grad_norm": 0.3154800148422279, "learning_rate": 0.00011502012526124978, "loss": 1.7859, "step": 3450 }, { "epoch": 1.5226972234464522, "grad_norm": 0.3029474817652635, "learning_rate": 0.00011476657039312613, "loss": 1.8433, "step": 3455 }, { "epoch": 1.5249008373732922, "grad_norm": 0.3394061586641444, "learning_rate": 0.00011451291843708824, "loss": 1.8191, "step": 3460 }, { "epoch": 1.5271044513001322, "grad_norm": 0.28588437415991474, "learning_rate": 0.00011425917106085844, "loss": 1.6528, "step": 3465 }, { "epoch": 1.5293080652269722, "grad_norm": 0.2624098766559022, "learning_rate": 0.00011400532993278643, "loss": 1.8208, "step": 3470 }, { "epoch": 1.5315116791538124, "grad_norm": 0.29050819773398057, "learning_rate": 0.00011375139672183834, "loss": 1.763, "step": 3475 }, { "epoch": 1.5337152930806521, "grad_norm": 0.28595831334373306, "learning_rate": 0.00011349737309758572, "loss": 1.6389, "step": 3480 }, { "epoch": 1.5359189070074923, "grad_norm": 0.310106838673627, "learning_rate": 0.00011324326073019458, "loss": 1.7008, "step": 3485 }, { "epoch": 1.5381225209343323, "grad_norm": 0.3425497639069633, "learning_rate": 0.0001129890612904144, "loss": 1.7975, "step": 3490 }, { "epoch": 1.5403261348611723, "grad_norm": 0.3049155101860015, "learning_rate": 0.0001127347764495671, "loss": 1.6302, "step": 3495 }, { "epoch": 1.5425297487880123, "grad_norm": 0.3083723366063809, "learning_rate": 0.00011248040787953622, "loss": 1.8779, "step": 3500 }, { "epoch": 1.5447333627148523, "grad_norm": 0.2714184317474351, "learning_rate": 0.00011222595725275562, "loss": 1.6655, "step": 3505 }, { "epoch": 1.5469369766416925, "grad_norm": 0.3020528286222207, "learning_rate": 0.00011197142624219887, "loss": 1.5374, "step": 3510 }, { "epoch": 1.5491405905685323, "grad_norm": 0.270325291856936, "learning_rate": 0.00011171681652136793, "loss": 1.7442, "step": 3515 }, { "epoch": 1.5513442044953725, "grad_norm": 0.2718748140775875, "learning_rate": 0.00011146212976428232, "loss": 1.793, "step": 3520 }, { "epoch": 1.5535478184222125, "grad_norm": 0.2684213723870114, "learning_rate": 0.00011120736764546799, "loss": 1.5847, "step": 3525 }, { "epoch": 1.5557514323490524, "grad_norm": 0.3293563453835575, "learning_rate": 0.00011095253183994645, "loss": 1.5808, "step": 3530 }, { "epoch": 1.5579550462758924, "grad_norm": 0.32122911366332685, "learning_rate": 0.0001106976240232237, "loss": 1.7343, "step": 3535 }, { "epoch": 1.5601586602027324, "grad_norm": 0.31939212525307864, "learning_rate": 0.0001104426458712791, "loss": 1.7123, "step": 3540 }, { "epoch": 1.5623622741295726, "grad_norm": 0.2676344457188956, "learning_rate": 0.00011018759906055463, "loss": 1.4029, "step": 3545 }, { "epoch": 1.5645658880564124, "grad_norm": 0.33141673784681086, "learning_rate": 0.00010993248526794347, "loss": 1.8105, "step": 3550 }, { "epoch": 1.5667695019832526, "grad_norm": 0.2715916392293134, "learning_rate": 0.00010967730617077938, "loss": 1.73, "step": 3555 }, { "epoch": 1.5689731159100926, "grad_norm": 0.2615038650065928, "learning_rate": 0.00010942206344682541, "loss": 1.7547, "step": 3560 }, { "epoch": 1.5711767298369326, "grad_norm": 0.2662856667093564, "learning_rate": 0.00010916675877426296, "loss": 1.6934, "step": 3565 }, { "epoch": 1.5733803437637726, "grad_norm": 0.20249417258651908, "learning_rate": 0.00010891139383168072, "loss": 1.7876, "step": 3570 }, { "epoch": 1.5755839576906125, "grad_norm": 0.30704028921333987, "learning_rate": 0.00010865597029806365, "loss": 1.7228, "step": 3575 }, { "epoch": 1.5777875716174528, "grad_norm": 0.32055060377455574, "learning_rate": 0.00010840048985278195, "loss": 1.7169, "step": 3580 }, { "epoch": 1.5799911855442925, "grad_norm": 0.31196324685842286, "learning_rate": 0.00010814495417557997, "loss": 1.875, "step": 3585 }, { "epoch": 1.5821947994711327, "grad_norm": 0.3005168077317045, "learning_rate": 0.00010788936494656523, "loss": 1.8862, "step": 3590 }, { "epoch": 1.5843984133979727, "grad_norm": 0.2856041438770126, "learning_rate": 0.00010763372384619738, "loss": 1.6419, "step": 3595 }, { "epoch": 1.5866020273248127, "grad_norm": 0.27459499931453724, "learning_rate": 0.00010737803255527702, "loss": 1.7495, "step": 3600 }, { "epoch": 1.5888056412516527, "grad_norm": 0.23142425230470423, "learning_rate": 0.00010712229275493489, "loss": 1.7615, "step": 3605 }, { "epoch": 1.5910092551784927, "grad_norm": 0.26704359594443944, "learning_rate": 0.00010686650612662048, "loss": 1.7043, "step": 3610 }, { "epoch": 1.5932128691053329, "grad_norm": 0.2644858878666521, "learning_rate": 0.00010661067435209135, "loss": 1.8665, "step": 3615 }, { "epoch": 1.5954164830321726, "grad_norm": 0.306618968505366, "learning_rate": 0.00010635479911340176, "loss": 1.8191, "step": 3620 }, { "epoch": 1.5976200969590129, "grad_norm": 0.32629807608957256, "learning_rate": 0.00010609888209289183, "loss": 1.781, "step": 3625 }, { "epoch": 1.5998237108858528, "grad_norm": 0.2701943161629176, "learning_rate": 0.00010584292497317633, "loss": 1.6162, "step": 3630 }, { "epoch": 1.6020273248126928, "grad_norm": 0.40336230029221404, "learning_rate": 0.00010558692943713373, "loss": 1.722, "step": 3635 }, { "epoch": 1.6042309387395328, "grad_norm": 0.2767732465609723, "learning_rate": 0.000105330897167895, "loss": 1.6427, "step": 3640 }, { "epoch": 1.6064345526663728, "grad_norm": 0.28371049131821663, "learning_rate": 0.00010507482984883268, "loss": 1.6872, "step": 3645 }, { "epoch": 1.608638166593213, "grad_norm": 0.2975217481512648, "learning_rate": 0.00010481872916354978, "loss": 1.6807, "step": 3650 }, { "epoch": 1.6108417805200528, "grad_norm": 0.30522288055794994, "learning_rate": 0.00010456259679586862, "loss": 1.6253, "step": 3655 }, { "epoch": 1.613045394446893, "grad_norm": 0.2893887034344458, "learning_rate": 0.00010430643442981986, "loss": 1.6465, "step": 3660 }, { "epoch": 1.615249008373733, "grad_norm": 0.28628624467329145, "learning_rate": 0.0001040502437496315, "loss": 1.6428, "step": 3665 }, { "epoch": 1.617452622300573, "grad_norm": 0.31683631498188874, "learning_rate": 0.00010379402643971746, "loss": 1.7033, "step": 3670 }, { "epoch": 1.619656236227413, "grad_norm": 0.2627512390977551, "learning_rate": 0.00010353778418466697, "loss": 1.8805, "step": 3675 }, { "epoch": 1.621859850154253, "grad_norm": 0.2976244892863047, "learning_rate": 0.00010328151866923316, "loss": 1.8013, "step": 3680 }, { "epoch": 1.6240634640810931, "grad_norm": 0.31860629887164105, "learning_rate": 0.00010302523157832216, "loss": 1.648, "step": 3685 }, { "epoch": 1.626267078007933, "grad_norm": 0.331987729083311, "learning_rate": 0.00010276892459698182, "loss": 1.6325, "step": 3690 }, { "epoch": 1.6284706919347731, "grad_norm": 0.2818827039809831, "learning_rate": 0.00010251259941039098, "loss": 1.7213, "step": 3695 }, { "epoch": 1.6306743058616129, "grad_norm": 0.2511192996283762, "learning_rate": 0.00010225625770384797, "loss": 1.5629, "step": 3700 }, { "epoch": 1.632877919788453, "grad_norm": 0.29741607234777423, "learning_rate": 0.00010199990116275988, "loss": 1.7834, "step": 3705 }, { "epoch": 1.635081533715293, "grad_norm": 0.30036469948366823, "learning_rate": 0.00010174353147263125, "loss": 1.4849, "step": 3710 }, { "epoch": 1.637285147642133, "grad_norm": 0.30522753547462433, "learning_rate": 0.00010148715031905312, "loss": 1.8071, "step": 3715 }, { "epoch": 1.6394887615689733, "grad_norm": 0.2769872314517015, "learning_rate": 0.00010123075938769187, "loss": 1.685, "step": 3720 }, { "epoch": 1.641692375495813, "grad_norm": 0.2709018320510704, "learning_rate": 0.00010097436036427816, "loss": 1.7853, "step": 3725 }, { "epoch": 1.6438959894226532, "grad_norm": 0.3446719925528754, "learning_rate": 0.00010071795493459591, "loss": 1.7783, "step": 3730 }, { "epoch": 1.646099603349493, "grad_norm": 0.3208180036756325, "learning_rate": 0.00010046154478447114, "loss": 1.8982, "step": 3735 }, { "epoch": 1.6483032172763332, "grad_norm": 0.289239117609306, "learning_rate": 0.00010020513159976084, "loss": 1.7313, "step": 3740 }, { "epoch": 1.6505068312031732, "grad_norm": 0.30122205784761724, "learning_rate": 9.994871706634204e-05, "loss": 1.6831, "step": 3745 }, { "epoch": 1.6527104451300132, "grad_norm": 0.2835725442151606, "learning_rate": 9.96923028701006e-05, "loss": 1.6129, "step": 3750 }, { "epoch": 1.6549140590568534, "grad_norm": 0.4849263642027594, "learning_rate": 9.943589069692014e-05, "loss": 1.783, "step": 3755 }, { "epoch": 1.6571176729836932, "grad_norm": 0.34060901148734135, "learning_rate": 9.917948223267105e-05, "loss": 1.642, "step": 3760 }, { "epoch": 1.6593212869105334, "grad_norm": 0.3001957536668968, "learning_rate": 9.892307916319919e-05, "loss": 1.6005, "step": 3765 }, { "epoch": 1.6615249008373731, "grad_norm": 0.3141764748091563, "learning_rate": 9.866668317431514e-05, "loss": 1.8968, "step": 3770 }, { "epoch": 1.6637285147642134, "grad_norm": 0.3066452663914771, "learning_rate": 9.841029595178282e-05, "loss": 1.6288, "step": 3775 }, { "epoch": 1.6659321286910533, "grad_norm": 0.2737312482046674, "learning_rate": 9.815391918130848e-05, "loss": 1.6151, "step": 3780 }, { "epoch": 1.6681357426178933, "grad_norm": 0.337368580313636, "learning_rate": 9.789755454852971e-05, "loss": 1.6298, "step": 3785 }, { "epoch": 1.6703393565447333, "grad_norm": 0.22864544249672591, "learning_rate": 9.764120373900436e-05, "loss": 1.7166, "step": 3790 }, { "epoch": 1.6725429704715733, "grad_norm": 0.3199596986088049, "learning_rate": 9.738486843819919e-05, "loss": 1.4291, "step": 3795 }, { "epoch": 1.6747465843984135, "grad_norm": 0.484113228627448, "learning_rate": 9.712855033147921e-05, "loss": 1.8267, "step": 3800 }, { "epoch": 1.6769501983252533, "grad_norm": 0.29066627384947913, "learning_rate": 9.68722511040962e-05, "loss": 1.7618, "step": 3805 }, { "epoch": 1.6791538122520935, "grad_norm": 0.2756578416564992, "learning_rate": 9.661597244117802e-05, "loss": 1.7626, "step": 3810 }, { "epoch": 1.6813574261789335, "grad_norm": 0.26262162877838774, "learning_rate": 9.635971602771716e-05, "loss": 1.6437, "step": 3815 }, { "epoch": 1.6835610401057735, "grad_norm": 0.3485930873206027, "learning_rate": 9.61034835485598e-05, "loss": 1.9054, "step": 3820 }, { "epoch": 1.6857646540326134, "grad_norm": 0.347720471228885, "learning_rate": 9.584727668839487e-05, "loss": 1.6653, "step": 3825 }, { "epoch": 1.6879682679594534, "grad_norm": 0.259896835339376, "learning_rate": 9.559109713174282e-05, "loss": 1.8298, "step": 3830 }, { "epoch": 1.6901718818862936, "grad_norm": 0.27688333027135853, "learning_rate": 9.533494656294458e-05, "loss": 1.5074, "step": 3835 }, { "epoch": 1.6923754958131334, "grad_norm": 0.3186540043438054, "learning_rate": 9.507882666615049e-05, "loss": 1.666, "step": 3840 }, { "epoch": 1.6945791097399736, "grad_norm": 0.34137430082378867, "learning_rate": 9.482273912530913e-05, "loss": 1.7862, "step": 3845 }, { "epoch": 1.6967827236668136, "grad_norm": 0.31312228735788816, "learning_rate": 9.456668562415657e-05, "loss": 1.6969, "step": 3850 }, { "epoch": 1.6989863375936536, "grad_norm": 0.3211377859121694, "learning_rate": 9.431066784620486e-05, "loss": 1.8653, "step": 3855 }, { "epoch": 1.7011899515204936, "grad_norm": 0.3443229930133863, "learning_rate": 9.405468747473127e-05, "loss": 1.696, "step": 3860 }, { "epoch": 1.7033935654473336, "grad_norm": 0.2900871769643087, "learning_rate": 9.379874619276707e-05, "loss": 1.7654, "step": 3865 }, { "epoch": 1.7055971793741738, "grad_norm": 0.3109956258190758, "learning_rate": 9.354284568308665e-05, "loss": 1.6229, "step": 3870 }, { "epoch": 1.7078007933010135, "grad_norm": 0.29096423007153205, "learning_rate": 9.328698762819623e-05, "loss": 1.7274, "step": 3875 }, { "epoch": 1.7100044072278537, "grad_norm": 0.2674081631314747, "learning_rate": 9.303117371032284e-05, "loss": 1.7598, "step": 3880 }, { "epoch": 1.7122080211546937, "grad_norm": 0.3370488411342742, "learning_rate": 9.277540561140342e-05, "loss": 1.7854, "step": 3885 }, { "epoch": 1.7144116350815337, "grad_norm": 0.2236164025203487, "learning_rate": 9.251968501307365e-05, "loss": 1.735, "step": 3890 }, { "epoch": 1.7166152490083737, "grad_norm": 0.31253584348251195, "learning_rate": 9.226401359665686e-05, "loss": 1.5912, "step": 3895 }, { "epoch": 1.7188188629352137, "grad_norm": 0.28523913611604496, "learning_rate": 9.2008393043153e-05, "loss": 1.7252, "step": 3900 }, { "epoch": 1.721022476862054, "grad_norm": 0.3641271132364838, "learning_rate": 9.17528250332277e-05, "loss": 1.7001, "step": 3905 }, { "epoch": 1.7232260907888937, "grad_norm": 0.34337087912187736, "learning_rate": 9.149731124720104e-05, "loss": 1.941, "step": 3910 }, { "epoch": 1.7254297047157339, "grad_norm": 0.3388940885148476, "learning_rate": 9.124185336503656e-05, "loss": 1.751, "step": 3915 }, { "epoch": 1.7276333186425739, "grad_norm": 0.3980423482061792, "learning_rate": 9.098645306633029e-05, "loss": 1.7571, "step": 3920 }, { "epoch": 1.7298369325694138, "grad_norm": 0.28139197000268995, "learning_rate": 9.073111203029972e-05, "loss": 1.5225, "step": 3925 }, { "epoch": 1.7320405464962538, "grad_norm": 0.28174263964005763, "learning_rate": 9.04758319357726e-05, "loss": 1.6746, "step": 3930 }, { "epoch": 1.7342441604230938, "grad_norm": 0.2461518030614674, "learning_rate": 9.0220614461176e-05, "loss": 1.5447, "step": 3935 }, { "epoch": 1.736447774349934, "grad_norm": 0.3297581828947644, "learning_rate": 8.99654612845253e-05, "loss": 1.5802, "step": 3940 }, { "epoch": 1.7386513882767738, "grad_norm": 0.3434764691584567, "learning_rate": 8.971037408341319e-05, "loss": 1.6836, "step": 3945 }, { "epoch": 1.740855002203614, "grad_norm": 0.2967197456557559, "learning_rate": 8.94553545349985e-05, "loss": 1.6141, "step": 3950 }, { "epoch": 1.743058616130454, "grad_norm": 0.3235657879123794, "learning_rate": 8.92004043159953e-05, "loss": 1.8539, "step": 3955 }, { "epoch": 1.745262230057294, "grad_norm": 0.30230482414452203, "learning_rate": 8.894552510266172e-05, "loss": 1.6447, "step": 3960 }, { "epoch": 1.747465843984134, "grad_norm": 0.25937552544500664, "learning_rate": 8.869071857078926e-05, "loss": 1.7132, "step": 3965 }, { "epoch": 1.749669457910974, "grad_norm": 0.3504915271341817, "learning_rate": 8.843598639569134e-05, "loss": 1.616, "step": 3970 }, { "epoch": 1.7518730718378142, "grad_norm": 0.3391539087204084, "learning_rate": 8.818133025219258e-05, "loss": 1.7726, "step": 3975 }, { "epoch": 1.754076685764654, "grad_norm": 0.2945081027363211, "learning_rate": 8.79267518146177e-05, "loss": 1.6354, "step": 3980 }, { "epoch": 1.7562802996914941, "grad_norm": 0.3122456807955038, "learning_rate": 8.767225275678054e-05, "loss": 1.7703, "step": 3985 }, { "epoch": 1.7584839136183341, "grad_norm": 0.2911403528384371, "learning_rate": 8.741783475197301e-05, "loss": 1.6184, "step": 3990 }, { "epoch": 1.760687527545174, "grad_norm": 0.28063561583988544, "learning_rate": 8.716349947295406e-05, "loss": 1.6723, "step": 3995 }, { "epoch": 1.762891141472014, "grad_norm": 0.24899262558923158, "learning_rate": 8.690924859193877e-05, "loss": 1.5999, "step": 4000 }, { "epoch": 1.765094755398854, "grad_norm": 0.21950111519024182, "learning_rate": 8.665508378058737e-05, "loss": 1.59, "step": 4005 }, { "epoch": 1.7672983693256943, "grad_norm": 0.2812929263006604, "learning_rate": 8.640100670999413e-05, "loss": 1.7219, "step": 4010 }, { "epoch": 1.769501983252534, "grad_norm": 0.26800092694380395, "learning_rate": 8.614701905067648e-05, "loss": 1.5878, "step": 4015 }, { "epoch": 1.7717055971793743, "grad_norm": 0.3421307184861735, "learning_rate": 8.589312247256385e-05, "loss": 1.5289, "step": 4020 }, { "epoch": 1.7739092111062142, "grad_norm": 0.2759648702531539, "learning_rate": 8.563931864498709e-05, "loss": 1.7232, "step": 4025 }, { "epoch": 1.7761128250330542, "grad_norm": 0.3083755515457062, "learning_rate": 8.538560923666697e-05, "loss": 1.5333, "step": 4030 }, { "epoch": 1.7783164389598942, "grad_norm": 0.32344968491148207, "learning_rate": 8.51319959157036e-05, "loss": 1.6531, "step": 4035 }, { "epoch": 1.7805200528867342, "grad_norm": 0.2628449279485873, "learning_rate": 8.487848034956527e-05, "loss": 1.7176, "step": 4040 }, { "epoch": 1.7827236668135744, "grad_norm": 0.30081231536696845, "learning_rate": 8.462506420507764e-05, "loss": 1.6087, "step": 4045 }, { "epoch": 1.7849272807404142, "grad_norm": 0.24172711814982975, "learning_rate": 8.437174914841261e-05, "loss": 1.6365, "step": 4050 }, { "epoch": 1.7871308946672544, "grad_norm": 0.3240513311621846, "learning_rate": 8.411853684507744e-05, "loss": 1.6818, "step": 4055 }, { "epoch": 1.7893345085940942, "grad_norm": 0.26629546007810934, "learning_rate": 8.38654289599038e-05, "loss": 1.6165, "step": 4060 }, { "epoch": 1.7915381225209344, "grad_norm": 0.2718697282057903, "learning_rate": 8.36124271570369e-05, "loss": 1.7767, "step": 4065 }, { "epoch": 1.7937417364477743, "grad_norm": 0.2991997948062704, "learning_rate": 8.335953309992442e-05, "loss": 1.6968, "step": 4070 }, { "epoch": 1.7959453503746143, "grad_norm": 0.33154914217795595, "learning_rate": 8.310674845130563e-05, "loss": 1.8523, "step": 4075 }, { "epoch": 1.7981489643014545, "grad_norm": 0.29895299240181794, "learning_rate": 8.285407487320042e-05, "loss": 1.5945, "step": 4080 }, { "epoch": 1.8003525782282943, "grad_norm": 0.2947016318335874, "learning_rate": 8.260151402689848e-05, "loss": 1.625, "step": 4085 }, { "epoch": 1.8025561921551345, "grad_norm": 0.3039906812402064, "learning_rate": 8.234906757294829e-05, "loss": 1.8956, "step": 4090 }, { "epoch": 1.8047598060819743, "grad_norm": 0.311603713861004, "learning_rate": 8.209673717114618e-05, "loss": 1.5808, "step": 4095 }, { "epoch": 1.8069634200088145, "grad_norm": 0.3208961582381735, "learning_rate": 8.184452448052547e-05, "loss": 1.4928, "step": 4100 }, { "epoch": 1.8091670339356545, "grad_norm": 0.27150517803601, "learning_rate": 8.15924311593456e-05, "loss": 1.7155, "step": 4105 }, { "epoch": 1.8113706478624945, "grad_norm": 0.31637619824367874, "learning_rate": 8.134045886508108e-05, "loss": 1.4761, "step": 4110 }, { "epoch": 1.8135742617893347, "grad_norm": 0.2536681416637017, "learning_rate": 8.108860925441076e-05, "loss": 1.7682, "step": 4115 }, { "epoch": 1.8157778757161744, "grad_norm": 0.2683657294785762, "learning_rate": 8.083688398320681e-05, "loss": 1.6091, "step": 4120 }, { "epoch": 1.8179814896430146, "grad_norm": 0.3004847316685122, "learning_rate": 8.058528470652396e-05, "loss": 1.7524, "step": 4125 }, { "epoch": 1.8201851035698544, "grad_norm": 0.3033198553489031, "learning_rate": 8.03338130785885e-05, "loss": 1.6975, "step": 4130 }, { "epoch": 1.8223887174966946, "grad_norm": 0.2932969746872377, "learning_rate": 8.008247075278742e-05, "loss": 1.6345, "step": 4135 }, { "epoch": 1.8245923314235346, "grad_norm": 0.29419726566032006, "learning_rate": 7.983125938165758e-05, "loss": 2.0007, "step": 4140 }, { "epoch": 1.8267959453503746, "grad_norm": 0.36994662608619383, "learning_rate": 7.958018061687494e-05, "loss": 1.8041, "step": 4145 }, { "epoch": 1.8289995592772146, "grad_norm": 0.3171521170329957, "learning_rate": 7.932923610924343e-05, "loss": 1.8268, "step": 4150 }, { "epoch": 1.8312031732040546, "grad_norm": 0.29749910928069495, "learning_rate": 7.907842750868441e-05, "loss": 1.7521, "step": 4155 }, { "epoch": 1.8334067871308948, "grad_norm": 0.28978378846287695, "learning_rate": 7.882775646422547e-05, "loss": 1.8141, "step": 4160 }, { "epoch": 1.8356104010577345, "grad_norm": 0.32370603539130244, "learning_rate": 7.857722462399009e-05, "loss": 1.5852, "step": 4165 }, { "epoch": 1.8378140149845748, "grad_norm": 0.3000864365709902, "learning_rate": 7.832683363518621e-05, "loss": 1.5174, "step": 4170 }, { "epoch": 1.8400176289114147, "grad_norm": 0.30867413513998754, "learning_rate": 7.807658514409587e-05, "loss": 1.7091, "step": 4175 }, { "epoch": 1.8422212428382547, "grad_norm": 0.2941777962805528, "learning_rate": 7.782648079606412e-05, "loss": 1.9314, "step": 4180 }, { "epoch": 1.8444248567650947, "grad_norm": 0.26789438468880916, "learning_rate": 7.757652223548836e-05, "loss": 1.5959, "step": 4185 }, { "epoch": 1.8466284706919347, "grad_norm": 0.2852613225045537, "learning_rate": 7.732671110580746e-05, "loss": 1.4776, "step": 4190 }, { "epoch": 1.848832084618775, "grad_norm": 0.28204667104600045, "learning_rate": 7.707704904949085e-05, "loss": 1.6044, "step": 4195 }, { "epoch": 1.8510356985456147, "grad_norm": 0.3433461472883869, "learning_rate": 7.682753770802791e-05, "loss": 1.8343, "step": 4200 }, { "epoch": 1.8532393124724549, "grad_norm": 0.27355747329642977, "learning_rate": 7.657817872191713e-05, "loss": 1.6496, "step": 4205 }, { "epoch": 1.8554429263992949, "grad_norm": 0.2860370187153179, "learning_rate": 7.632897373065522e-05, "loss": 1.5719, "step": 4210 }, { "epoch": 1.8576465403261349, "grad_norm": 0.2618172945255669, "learning_rate": 7.607992437272642e-05, "loss": 1.6911, "step": 4215 }, { "epoch": 1.8598501542529748, "grad_norm": 0.2570942578849571, "learning_rate": 7.583103228559164e-05, "loss": 1.7205, "step": 4220 }, { "epoch": 1.8620537681798148, "grad_norm": 0.30089829678159197, "learning_rate": 7.558229910567794e-05, "loss": 1.7114, "step": 4225 }, { "epoch": 1.864257382106655, "grad_norm": 0.3120407343609943, "learning_rate": 7.533372646836736e-05, "loss": 1.6438, "step": 4230 }, { "epoch": 1.8664609960334948, "grad_norm": 0.29920830893784145, "learning_rate": 7.508531600798657e-05, "loss": 1.8773, "step": 4235 }, { "epoch": 1.868664609960335, "grad_norm": 0.28932016728007587, "learning_rate": 7.483706935779584e-05, "loss": 1.5398, "step": 4240 }, { "epoch": 1.870868223887175, "grad_norm": 0.3323660400642421, "learning_rate": 7.458898814997852e-05, "loss": 1.8427, "step": 4245 }, { "epoch": 1.873071837814015, "grad_norm": 0.24905133812180258, "learning_rate": 7.434107401563016e-05, "loss": 1.5713, "step": 4250 }, { "epoch": 1.875275451740855, "grad_norm": 0.3177914709078092, "learning_rate": 7.409332858474772e-05, "loss": 1.8163, "step": 4255 }, { "epoch": 1.877479065667695, "grad_norm": 0.2731160161178577, "learning_rate": 7.384575348621909e-05, "loss": 1.502, "step": 4260 }, { "epoch": 1.8796826795945352, "grad_norm": 0.318150552351912, "learning_rate": 7.359835034781227e-05, "loss": 1.709, "step": 4265 }, { "epoch": 1.881886293521375, "grad_norm": 0.2992460408245423, "learning_rate": 7.335112079616456e-05, "loss": 1.6948, "step": 4270 }, { "epoch": 1.8840899074482151, "grad_norm": 0.26242651145469986, "learning_rate": 7.31040664567719e-05, "loss": 1.5288, "step": 4275 }, { "epoch": 1.8862935213750551, "grad_norm": 0.32400746012853343, "learning_rate": 7.285718895397848e-05, "loss": 1.7299, "step": 4280 }, { "epoch": 1.8884971353018951, "grad_norm": 0.30579297267422695, "learning_rate": 7.261048991096558e-05, "loss": 1.8361, "step": 4285 }, { "epoch": 1.890700749228735, "grad_norm": 0.28173595807802665, "learning_rate": 7.236397094974119e-05, "loss": 1.737, "step": 4290 }, { "epoch": 1.892904363155575, "grad_norm": 0.3062487579152163, "learning_rate": 7.211763369112934e-05, "loss": 1.6612, "step": 4295 }, { "epoch": 1.8951079770824153, "grad_norm": 0.29979139920156694, "learning_rate": 7.18714797547594e-05, "loss": 1.762, "step": 4300 }, { "epoch": 1.897311591009255, "grad_norm": 0.3021721281737111, "learning_rate": 7.162551075905538e-05, "loss": 1.8317, "step": 4305 }, { "epoch": 1.8995152049360953, "grad_norm": 0.2809347403792435, "learning_rate": 7.137972832122532e-05, "loss": 1.7406, "step": 4310 }, { "epoch": 1.9017188188629353, "grad_norm": 0.31981826220168086, "learning_rate": 7.113413405725069e-05, "loss": 1.8273, "step": 4315 }, { "epoch": 1.9039224327897752, "grad_norm": 0.27829653919403746, "learning_rate": 7.088872958187578e-05, "loss": 1.6196, "step": 4320 }, { "epoch": 1.9061260467166152, "grad_norm": 0.2612389626891207, "learning_rate": 7.064351650859704e-05, "loss": 1.7173, "step": 4325 }, { "epoch": 1.9083296606434552, "grad_norm": 0.3351698187645255, "learning_rate": 7.039849644965246e-05, "loss": 1.5561, "step": 4330 }, { "epoch": 1.9105332745702954, "grad_norm": 0.276151473793176, "learning_rate": 7.015367101601091e-05, "loss": 1.5952, "step": 4335 }, { "epoch": 1.9127368884971352, "grad_norm": 0.2866569066490694, "learning_rate": 6.990904181736187e-05, "loss": 1.7386, "step": 4340 }, { "epoch": 1.9149405024239754, "grad_norm": 0.30680340268949835, "learning_rate": 6.96646104621043e-05, "loss": 1.793, "step": 4345 }, { "epoch": 1.9171441163508154, "grad_norm": 0.2853878652168025, "learning_rate": 6.942037855733661e-05, "loss": 1.8032, "step": 4350 }, { "epoch": 1.9193477302776554, "grad_norm": 0.3335656030865567, "learning_rate": 6.917634770884571e-05, "loss": 1.8019, "step": 4355 }, { "epoch": 1.9215513442044954, "grad_norm": 0.26469328077404297, "learning_rate": 6.893251952109668e-05, "loss": 1.7769, "step": 4360 }, { "epoch": 1.9237549581313353, "grad_norm": 0.3611778018097071, "learning_rate": 6.868889559722213e-05, "loss": 1.845, "step": 4365 }, { "epoch": 1.9259585720581756, "grad_norm": 0.2607913972126291, "learning_rate": 6.84454775390116e-05, "loss": 1.6583, "step": 4370 }, { "epoch": 1.9281621859850153, "grad_norm": 0.2952126078040435, "learning_rate": 6.820226694690112e-05, "loss": 1.4631, "step": 4375 }, { "epoch": 1.9303657999118555, "grad_norm": 0.28021793360716346, "learning_rate": 6.795926541996273e-05, "loss": 1.7197, "step": 4380 }, { "epoch": 1.9325694138386955, "grad_norm": 0.2431287772236666, "learning_rate": 6.771647455589384e-05, "loss": 1.7528, "step": 4385 }, { "epoch": 1.9347730277655355, "grad_norm": 0.24277132942171645, "learning_rate": 6.74738959510068e-05, "loss": 1.52, "step": 4390 }, { "epoch": 1.9369766416923755, "grad_norm": 0.2616438959393551, "learning_rate": 6.723153120021833e-05, "loss": 1.6386, "step": 4395 }, { "epoch": 1.9391802556192155, "grad_norm": 0.3044048108764881, "learning_rate": 6.698938189703918e-05, "loss": 1.653, "step": 4400 }, { "epoch": 1.9413838695460557, "grad_norm": 0.2952474841434869, "learning_rate": 6.674744963356357e-05, "loss": 1.6325, "step": 4405 }, { "epoch": 1.9435874834728954, "grad_norm": 0.3104804592935538, "learning_rate": 6.65057360004586e-05, "loss": 1.7827, "step": 4410 }, { "epoch": 1.9457910973997357, "grad_norm": 0.27421700814052996, "learning_rate": 6.626424258695403e-05, "loss": 1.6614, "step": 4415 }, { "epoch": 1.9479947113265754, "grad_norm": 0.28977404672357854, "learning_rate": 6.60229709808317e-05, "loss": 1.8225, "step": 4420 }, { "epoch": 1.9501983252534156, "grad_norm": 0.30875392634058607, "learning_rate": 6.578192276841501e-05, "loss": 1.7437, "step": 4425 }, { "epoch": 1.9524019391802556, "grad_norm": 0.307292209389746, "learning_rate": 6.554109953455864e-05, "loss": 1.7637, "step": 4430 }, { "epoch": 1.9546055531070956, "grad_norm": 0.34674701377289235, "learning_rate": 6.53005028626381e-05, "loss": 1.6782, "step": 4435 }, { "epoch": 1.9568091670339358, "grad_norm": 0.2997355914742966, "learning_rate": 6.506013433453926e-05, "loss": 1.7479, "step": 4440 }, { "epoch": 1.9590127809607756, "grad_norm": 0.33789074614600445, "learning_rate": 6.4819995530648e-05, "loss": 1.6811, "step": 4445 }, { "epoch": 1.9612163948876158, "grad_norm": 0.3008656237866022, "learning_rate": 6.45800880298397e-05, "loss": 1.6704, "step": 4450 }, { "epoch": 1.9634200088144556, "grad_norm": 0.31248757394845805, "learning_rate": 6.434041340946909e-05, "loss": 1.6695, "step": 4455 }, { "epoch": 1.9656236227412958, "grad_norm": 0.3086687965739077, "learning_rate": 6.41009732453597e-05, "loss": 1.5949, "step": 4460 }, { "epoch": 1.9678272366681357, "grad_norm": 0.2536971137620243, "learning_rate": 6.386176911179353e-05, "loss": 1.6463, "step": 4465 }, { "epoch": 1.9700308505949757, "grad_norm": 0.3287438241265034, "learning_rate": 6.362280258150074e-05, "loss": 1.6429, "step": 4470 }, { "epoch": 1.972234464521816, "grad_norm": 0.2565655853570647, "learning_rate": 6.33840752256492e-05, "loss": 1.5546, "step": 4475 }, { "epoch": 1.9744380784486557, "grad_norm": 0.3055708289742853, "learning_rate": 6.314558861383442e-05, "loss": 1.7164, "step": 4480 }, { "epoch": 1.976641692375496, "grad_norm": 0.26286511562137227, "learning_rate": 6.29073443140689e-05, "loss": 1.7841, "step": 4485 }, { "epoch": 1.9788453063023357, "grad_norm": 0.2842174046049182, "learning_rate": 6.266934389277204e-05, "loss": 1.7053, "step": 4490 }, { "epoch": 1.981048920229176, "grad_norm": 0.34820565260475556, "learning_rate": 6.24315889147597e-05, "loss": 1.7649, "step": 4495 }, { "epoch": 1.9832525341560159, "grad_norm": 0.3290871426082691, "learning_rate": 6.219408094323415e-05, "loss": 1.6402, "step": 4500 }, { "epoch": 1.9854561480828559, "grad_norm": 0.25604694542787715, "learning_rate": 6.195682153977351e-05, "loss": 1.5192, "step": 4505 }, { "epoch": 1.9876597620096959, "grad_norm": 0.28478382829773047, "learning_rate": 6.17198122643216e-05, "loss": 1.649, "step": 4510 }, { "epoch": 1.9898633759365358, "grad_norm": 0.2776833563766362, "learning_rate": 6.148305467517768e-05, "loss": 1.7351, "step": 4515 }, { "epoch": 1.992066989863376, "grad_norm": 0.32063858764742265, "learning_rate": 6.124655032898631e-05, "loss": 1.8315, "step": 4520 }, { "epoch": 1.9942706037902158, "grad_norm": 0.29572004437320715, "learning_rate": 6.1010300780726925e-05, "loss": 1.7337, "step": 4525 }, { "epoch": 1.996474217717056, "grad_norm": 0.3096259639483799, "learning_rate": 6.077430758370376e-05, "loss": 1.759, "step": 4530 }, { "epoch": 1.998677831643896, "grad_norm": 0.31442409226477874, "learning_rate": 6.053857228953546e-05, "loss": 1.7822, "step": 4535 }, { "epoch": 2.000881445570736, "grad_norm": 0.2377883462279695, "learning_rate": 6.03030964481452e-05, "loss": 1.4966, "step": 4540 }, { "epoch": 2.003085059497576, "grad_norm": 0.2900063226146335, "learning_rate": 6.0067881607750134e-05, "loss": 1.7189, "step": 4545 }, { "epoch": 2.005288673424416, "grad_norm": 0.30629874279137115, "learning_rate": 5.983292931485142e-05, "loss": 1.5509, "step": 4550 }, { "epoch": 2.007492287351256, "grad_norm": 0.2682166094278232, "learning_rate": 5.9598241114223986e-05, "loss": 1.5487, "step": 4555 }, { "epoch": 2.009695901278096, "grad_norm": 0.3321377020430881, "learning_rate": 5.936381854890646e-05, "loss": 1.8033, "step": 4560 }, { "epoch": 2.011899515204936, "grad_norm": 0.31810161747738896, "learning_rate": 5.912966316019093e-05, "loss": 1.621, "step": 4565 }, { "epoch": 2.014103129131776, "grad_norm": 0.24729310525461057, "learning_rate": 5.8895776487612765e-05, "loss": 1.5993, "step": 4570 }, { "epoch": 2.016306743058616, "grad_norm": 0.30363634336485235, "learning_rate": 5.8662160068940655e-05, "loss": 1.6749, "step": 4575 }, { "epoch": 2.0185103569854563, "grad_norm": 0.2851113038101266, "learning_rate": 5.84288154401664e-05, "loss": 1.6956, "step": 4580 }, { "epoch": 2.020713970912296, "grad_norm": 0.345208802987356, "learning_rate": 5.81957441354948e-05, "loss": 1.5851, "step": 4585 }, { "epoch": 2.0229175848391363, "grad_norm": 0.33771374182942, "learning_rate": 5.796294768733362e-05, "loss": 1.5908, "step": 4590 }, { "epoch": 2.025121198765976, "grad_norm": 0.29219932563018164, "learning_rate": 5.773042762628342e-05, "loss": 1.677, "step": 4595 }, { "epoch": 2.0273248126928163, "grad_norm": 0.26977700042523883, "learning_rate": 5.749818548112762e-05, "loss": 1.6073, "step": 4600 }, { "epoch": 2.029528426619656, "grad_norm": 0.30654875158910516, "learning_rate": 5.726622277882243e-05, "loss": 1.7622, "step": 4605 }, { "epoch": 2.0317320405464963, "grad_norm": 0.3217700066222481, "learning_rate": 5.703454104448665e-05, "loss": 1.804, "step": 4610 }, { "epoch": 2.0339356544733365, "grad_norm": 0.30632620282091305, "learning_rate": 5.680314180139178e-05, "loss": 1.7833, "step": 4615 }, { "epoch": 2.0361392684001762, "grad_norm": 0.3023215763344122, "learning_rate": 5.657202657095206e-05, "loss": 1.7969, "step": 4620 }, { "epoch": 2.0383428823270164, "grad_norm": 0.3078282400161021, "learning_rate": 5.6341196872714394e-05, "loss": 1.6958, "step": 4625 }, { "epoch": 2.040546496253856, "grad_norm": 0.27870283271610047, "learning_rate": 5.611065422434828e-05, "loss": 1.5725, "step": 4630 }, { "epoch": 2.0427501101806964, "grad_norm": 0.3114689828172716, "learning_rate": 5.588040014163585e-05, "loss": 1.562, "step": 4635 }, { "epoch": 2.044953724107536, "grad_norm": 0.2912835457860766, "learning_rate": 5.565043613846219e-05, "loss": 1.7486, "step": 4640 }, { "epoch": 2.0471573380343764, "grad_norm": 0.29628653391558857, "learning_rate": 5.542076372680498e-05, "loss": 1.7084, "step": 4645 }, { "epoch": 2.0493609519612166, "grad_norm": 0.29948576765849355, "learning_rate": 5.519138441672471e-05, "loss": 1.6903, "step": 4650 }, { "epoch": 2.0515645658880564, "grad_norm": 0.26477213023267704, "learning_rate": 5.496229971635487e-05, "loss": 1.6743, "step": 4655 }, { "epoch": 2.0537681798148966, "grad_norm": 0.36060338821204513, "learning_rate": 5.473351113189194e-05, "loss": 1.8093, "step": 4660 }, { "epoch": 2.0559717937417363, "grad_norm": 0.31416116481556966, "learning_rate": 5.4505020167585396e-05, "loss": 1.6409, "step": 4665 }, { "epoch": 2.0581754076685765, "grad_norm": 0.23960293640658495, "learning_rate": 5.4276828325727934e-05, "loss": 1.5688, "step": 4670 }, { "epoch": 2.0603790215954163, "grad_norm": 0.33607407453371085, "learning_rate": 5.4048937106645613e-05, "loss": 1.5812, "step": 4675 }, { "epoch": 2.0625826355222565, "grad_norm": 0.3362219928372315, "learning_rate": 5.3821348008687967e-05, "loss": 1.7184, "step": 4680 }, { "epoch": 2.0647862494490967, "grad_norm": 0.31128427440833256, "learning_rate": 5.3594062528218025e-05, "loss": 1.6606, "step": 4685 }, { "epoch": 2.0669898633759365, "grad_norm": 0.28720777906212147, "learning_rate": 5.336708215960258e-05, "loss": 1.6505, "step": 4690 }, { "epoch": 2.0691934773027767, "grad_norm": 0.35035039806258184, "learning_rate": 5.314040839520253e-05, "loss": 1.7716, "step": 4695 }, { "epoch": 2.0713970912296165, "grad_norm": 0.2931752659797425, "learning_rate": 5.291404272536275e-05, "loss": 1.6877, "step": 4700 }, { "epoch": 2.0736007051564567, "grad_norm": 0.2873712201718596, "learning_rate": 5.268798663840243e-05, "loss": 1.6062, "step": 4705 }, { "epoch": 2.0758043190832964, "grad_norm": 0.26676705722923527, "learning_rate": 5.2462241620605366e-05, "loss": 1.6592, "step": 4710 }, { "epoch": 2.0780079330101366, "grad_norm": 0.3509576273254501, "learning_rate": 5.223680915621014e-05, "loss": 1.671, "step": 4715 }, { "epoch": 2.080211546936977, "grad_norm": 0.39209953129493824, "learning_rate": 5.2011690727400285e-05, "loss": 1.6385, "step": 4720 }, { "epoch": 2.0824151608638166, "grad_norm": 0.3402308135526598, "learning_rate": 5.178688781429455e-05, "loss": 1.6095, "step": 4725 }, { "epoch": 2.084618774790657, "grad_norm": 0.29891446961728113, "learning_rate": 5.1562401894937365e-05, "loss": 1.6653, "step": 4730 }, { "epoch": 2.0868223887174966, "grad_norm": 0.2466567525935987, "learning_rate": 5.133823444528889e-05, "loss": 1.6558, "step": 4735 }, { "epoch": 2.089026002644337, "grad_norm": 0.3676958248067488, "learning_rate": 5.111438693921536e-05, "loss": 1.6279, "step": 4740 }, { "epoch": 2.0912296165711766, "grad_norm": 0.343788030408807, "learning_rate": 5.089086084847954e-05, "loss": 1.6951, "step": 4745 }, { "epoch": 2.0934332304980168, "grad_norm": 0.3319476261884821, "learning_rate": 5.066765764273078e-05, "loss": 1.617, "step": 4750 }, { "epoch": 2.095636844424857, "grad_norm": 0.33586220021618074, "learning_rate": 5.044477878949571e-05, "loss": 1.6601, "step": 4755 }, { "epoch": 2.0978404583516967, "grad_norm": 0.27982003995931076, "learning_rate": 5.0222225754168175e-05, "loss": 1.6063, "step": 4760 }, { "epoch": 2.100044072278537, "grad_norm": 0.3330827204915635, "learning_rate": 5.000000000000002e-05, "loss": 1.7513, "step": 4765 }, { "epoch": 2.1022476862053767, "grad_norm": 0.34805614239612326, "learning_rate": 4.97781029880911e-05, "loss": 1.5524, "step": 4770 }, { "epoch": 2.104451300132217, "grad_norm": 0.287355630557104, "learning_rate": 4.955653617737995e-05, "loss": 1.6138, "step": 4775 }, { "epoch": 2.1066549140590567, "grad_norm": 0.29898227350779466, "learning_rate": 4.9335301024634094e-05, "loss": 1.6648, "step": 4780 }, { "epoch": 2.108858527985897, "grad_norm": 0.26174467755819714, "learning_rate": 4.911439898444036e-05, "loss": 1.594, "step": 4785 }, { "epoch": 2.111062141912737, "grad_norm": 0.35692436121383353, "learning_rate": 4.889383150919543e-05, "loss": 1.5403, "step": 4790 }, { "epoch": 2.113265755839577, "grad_norm": 0.2926942813060104, "learning_rate": 4.867360004909635e-05, "loss": 1.754, "step": 4795 }, { "epoch": 2.115469369766417, "grad_norm": 0.3290438654965286, "learning_rate": 4.845370605213091e-05, "loss": 1.5578, "step": 4800 }, { "epoch": 2.117672983693257, "grad_norm": 0.3373040566376742, "learning_rate": 4.823415096406806e-05, "loss": 1.6939, "step": 4805 }, { "epoch": 2.119876597620097, "grad_norm": 0.34391563105196393, "learning_rate": 4.801493622844847e-05, "loss": 1.7067, "step": 4810 }, { "epoch": 2.122080211546937, "grad_norm": 0.3002969326243971, "learning_rate": 4.779606328657513e-05, "loss": 1.716, "step": 4815 }, { "epoch": 2.124283825473777, "grad_norm": 0.3388857319061881, "learning_rate": 4.75775335775038e-05, "loss": 1.843, "step": 4820 }, { "epoch": 2.1264874394006172, "grad_norm": 0.4156596099940998, "learning_rate": 4.735934853803339e-05, "loss": 1.7106, "step": 4825 }, { "epoch": 2.128691053327457, "grad_norm": 0.2940085091207592, "learning_rate": 4.71415096026968e-05, "loss": 1.6581, "step": 4830 }, { "epoch": 2.130894667254297, "grad_norm": 0.306112268047742, "learning_rate": 4.692401820375134e-05, "loss": 1.7315, "step": 4835 }, { "epoch": 2.133098281181137, "grad_norm": 0.29996735672531655, "learning_rate": 4.6706875771169265e-05, "loss": 1.649, "step": 4840 }, { "epoch": 2.135301895107977, "grad_norm": 0.27463359410451954, "learning_rate": 4.64900837326284e-05, "loss": 1.6608, "step": 4845 }, { "epoch": 2.137505509034817, "grad_norm": 0.3165826117114464, "learning_rate": 4.627364351350288e-05, "loss": 1.6793, "step": 4850 }, { "epoch": 2.139709122961657, "grad_norm": 0.3469777212157657, "learning_rate": 4.605755653685366e-05, "loss": 1.7027, "step": 4855 }, { "epoch": 2.1419127368884974, "grad_norm": 0.28698621030093346, "learning_rate": 4.584182422341915e-05, "loss": 1.6516, "step": 4860 }, { "epoch": 2.144116350815337, "grad_norm": 0.3294649962724426, "learning_rate": 4.562644799160585e-05, "loss": 1.7214, "step": 4865 }, { "epoch": 2.1463199647421773, "grad_norm": 0.3185351974138239, "learning_rate": 4.541142925747919e-05, "loss": 1.6362, "step": 4870 }, { "epoch": 2.148523578669017, "grad_norm": 0.3126109058989699, "learning_rate": 4.519676943475408e-05, "loss": 1.7064, "step": 4875 }, { "epoch": 2.1507271925958573, "grad_norm": 0.5380540251828319, "learning_rate": 4.4982469934785574e-05, "loss": 1.6943, "step": 4880 }, { "epoch": 2.152930806522697, "grad_norm": 0.3435754281589444, "learning_rate": 4.4768532166559763e-05, "loss": 1.715, "step": 4885 }, { "epoch": 2.1551344204495373, "grad_norm": 0.28522234926734513, "learning_rate": 4.455495753668428e-05, "loss": 1.3429, "step": 4890 }, { "epoch": 2.157338034376377, "grad_norm": 0.29145572840002815, "learning_rate": 4.4341747449379335e-05, "loss": 1.5995, "step": 4895 }, { "epoch": 2.1595416483032173, "grad_norm": 0.47296843361442037, "learning_rate": 4.412890330646815e-05, "loss": 1.8911, "step": 4900 }, { "epoch": 2.1617452622300575, "grad_norm": 0.3101777212408002, "learning_rate": 4.391642650736811e-05, "loss": 1.5388, "step": 4905 }, { "epoch": 2.1639488761568972, "grad_norm": 0.27664700170021256, "learning_rate": 4.370431844908119e-05, "loss": 1.5866, "step": 4910 }, { "epoch": 2.1661524900837374, "grad_norm": 0.29560765708014347, "learning_rate": 4.349258052618509e-05, "loss": 1.7198, "step": 4915 }, { "epoch": 2.168356104010577, "grad_norm": 0.3363838001346494, "learning_rate": 4.328121413082388e-05, "loss": 1.6872, "step": 4920 }, { "epoch": 2.1705597179374174, "grad_norm": 0.2850203542162651, "learning_rate": 4.307022065269887e-05, "loss": 1.6207, "step": 4925 }, { "epoch": 2.172763331864257, "grad_norm": 0.32285248455004484, "learning_rate": 4.285960147905946e-05, "loss": 1.6117, "step": 4930 }, { "epoch": 2.1749669457910974, "grad_norm": 0.27430213811378995, "learning_rate": 4.264935799469417e-05, "loss": 1.6949, "step": 4935 }, { "epoch": 2.1771705597179376, "grad_norm": 0.46396848778851696, "learning_rate": 4.2439491581921373e-05, "loss": 1.6883, "step": 4940 }, { "epoch": 2.1793741736447774, "grad_norm": 0.31449151532095687, "learning_rate": 4.223000362058023e-05, "loss": 1.6213, "step": 4945 }, { "epoch": 2.1815777875716176, "grad_norm": 0.27602615724586665, "learning_rate": 4.202089548802157e-05, "loss": 1.6365, "step": 4950 }, { "epoch": 2.1837814014984573, "grad_norm": 0.2989766105095882, "learning_rate": 4.181216855909913e-05, "loss": 1.6936, "step": 4955 }, { "epoch": 2.1859850154252976, "grad_norm": 0.29722828765908, "learning_rate": 4.16038242061601e-05, "loss": 1.5737, "step": 4960 }, { "epoch": 2.1881886293521373, "grad_norm": 0.2573098385617781, "learning_rate": 4.139586379903629e-05, "loss": 1.6852, "step": 4965 }, { "epoch": 2.1903922432789775, "grad_norm": 0.2951816124750441, "learning_rate": 4.1188288705035226e-05, "loss": 1.645, "step": 4970 }, { "epoch": 2.1925958572058177, "grad_norm": 0.33288427202987686, "learning_rate": 4.098110028893105e-05, "loss": 1.5257, "step": 4975 }, { "epoch": 2.1947994711326575, "grad_norm": 0.3489398250177744, "learning_rate": 4.077429991295549e-05, "loss": 1.6671, "step": 4980 }, { "epoch": 2.1970030850594977, "grad_norm": 0.2899374350308383, "learning_rate": 4.056788893678898e-05, "loss": 1.6132, "step": 4985 }, { "epoch": 2.1992066989863375, "grad_norm": 0.3500814252732699, "learning_rate": 4.036186871755173e-05, "loss": 1.5695, "step": 4990 }, { "epoch": 2.2014103129131777, "grad_norm": 0.3687163067932294, "learning_rate": 4.015624060979486e-05, "loss": 1.5143, "step": 4995 }, { "epoch": 2.2036139268400174, "grad_norm": 0.33423795880289425, "learning_rate": 3.995100596549128e-05, "loss": 1.6156, "step": 5000 }, { "epoch": 2.2058175407668577, "grad_norm": 0.2868033623703877, "learning_rate": 3.9746166134026995e-05, "loss": 1.5561, "step": 5005 }, { "epoch": 2.208021154693698, "grad_norm": 0.34323096599301134, "learning_rate": 3.9541722462192196e-05, "loss": 1.4618, "step": 5010 }, { "epoch": 2.2102247686205376, "grad_norm": 0.27958346490895863, "learning_rate": 3.9337676294172424e-05, "loss": 1.7082, "step": 5015 }, { "epoch": 2.212428382547378, "grad_norm": 0.3257321749343037, "learning_rate": 3.913402897153957e-05, "loss": 1.6946, "step": 5020 }, { "epoch": 2.2146319964742176, "grad_norm": 0.3365177337266649, "learning_rate": 3.893078183324329e-05, "loss": 1.6428, "step": 5025 }, { "epoch": 2.216835610401058, "grad_norm": 0.34401308894946486, "learning_rate": 3.8727936215602077e-05, "loss": 1.5488, "step": 5030 }, { "epoch": 2.2190392243278976, "grad_norm": 0.2845220746832339, "learning_rate": 3.852549345229445e-05, "loss": 1.6519, "step": 5035 }, { "epoch": 2.221242838254738, "grad_norm": 0.25529471083550676, "learning_rate": 3.832345487435019e-05, "loss": 1.8166, "step": 5040 }, { "epoch": 2.223446452181578, "grad_norm": 0.3591343853259783, "learning_rate": 3.812182181014169e-05, "loss": 1.7223, "step": 5045 }, { "epoch": 2.2256500661084178, "grad_norm": 0.3054726611280714, "learning_rate": 3.792059558537518e-05, "loss": 1.8144, "step": 5050 }, { "epoch": 2.227853680035258, "grad_norm": 0.30167111375727146, "learning_rate": 3.7719777523081864e-05, "loss": 1.5961, "step": 5055 }, { "epoch": 2.2300572939620977, "grad_norm": 0.2916710789608964, "learning_rate": 3.751936894360949e-05, "loss": 1.7809, "step": 5060 }, { "epoch": 2.232260907888938, "grad_norm": 0.3732355683913339, "learning_rate": 3.731937116461336e-05, "loss": 1.6552, "step": 5065 }, { "epoch": 2.2344645218157777, "grad_norm": 0.35016354658091353, "learning_rate": 3.7119785501047977e-05, "loss": 1.649, "step": 5070 }, { "epoch": 2.236668135742618, "grad_norm": 0.34880893384696754, "learning_rate": 3.6920613265158124e-05, "loss": 1.5914, "step": 5075 }, { "epoch": 2.2388717496694577, "grad_norm": 0.3026190037167945, "learning_rate": 3.672185576647047e-05, "loss": 1.5736, "step": 5080 }, { "epoch": 2.241075363596298, "grad_norm": 0.3484793101368692, "learning_rate": 3.652351431178473e-05, "loss": 1.7296, "step": 5085 }, { "epoch": 2.243278977523138, "grad_norm": 0.2587698411979413, "learning_rate": 3.6325590205165314e-05, "loss": 1.7112, "step": 5090 }, { "epoch": 2.245482591449978, "grad_norm": 0.25882732375987005, "learning_rate": 3.612808474793261e-05, "loss": 1.7223, "step": 5095 }, { "epoch": 2.247686205376818, "grad_norm": 0.27481092620324143, "learning_rate": 3.593099923865438e-05, "loss": 1.5473, "step": 5100 }, { "epoch": 2.249889819303658, "grad_norm": 0.3411799767876264, "learning_rate": 3.573433497313731e-05, "loss": 1.6459, "step": 5105 }, { "epoch": 2.252093433230498, "grad_norm": 0.2887170405054378, "learning_rate": 3.5538093244418525e-05, "loss": 1.6195, "step": 5110 }, { "epoch": 2.254297047157338, "grad_norm": 0.3884613083691619, "learning_rate": 3.5342275342757046e-05, "loss": 1.8638, "step": 5115 }, { "epoch": 2.256500661084178, "grad_norm": 0.28793406477711025, "learning_rate": 3.5146882555625226e-05, "loss": 1.5124, "step": 5120 }, { "epoch": 2.2587042750110182, "grad_norm": 0.30122690590343093, "learning_rate": 3.495191616770034e-05, "loss": 1.7147, "step": 5125 }, { "epoch": 2.260907888937858, "grad_norm": 0.27840602962105204, "learning_rate": 3.475737746085631e-05, "loss": 1.5467, "step": 5130 }, { "epoch": 2.263111502864698, "grad_norm": 0.3165143945785378, "learning_rate": 3.456326771415498e-05, "loss": 1.6215, "step": 5135 }, { "epoch": 2.265315116791538, "grad_norm": 0.315000953366212, "learning_rate": 3.436958820383783e-05, "loss": 1.5548, "step": 5140 }, { "epoch": 2.267518730718378, "grad_norm": 0.29091408470720853, "learning_rate": 3.417634020331769e-05, "loss": 1.786, "step": 5145 }, { "epoch": 2.269722344645218, "grad_norm": 0.335006684656636, "learning_rate": 3.398352498317029e-05, "loss": 1.6015, "step": 5150 }, { "epoch": 2.271925958572058, "grad_norm": 0.3244552469471718, "learning_rate": 3.379114381112581e-05, "loss": 1.653, "step": 5155 }, { "epoch": 2.2741295724988984, "grad_norm": 0.32402505918566016, "learning_rate": 3.359919795206065e-05, "loss": 1.5578, "step": 5160 }, { "epoch": 2.276333186425738, "grad_norm": 0.3417399295841799, "learning_rate": 3.3407688667989124e-05, "loss": 1.8143, "step": 5165 }, { "epoch": 2.2785368003525783, "grad_norm": 0.3293628623372523, "learning_rate": 3.321661721805519e-05, "loss": 1.62, "step": 5170 }, { "epoch": 2.280740414279418, "grad_norm": 0.3594982936202251, "learning_rate": 3.302598485852401e-05, "loss": 1.5937, "step": 5175 }, { "epoch": 2.2829440282062583, "grad_norm": 0.2986899594367847, "learning_rate": 3.283579284277378e-05, "loss": 1.5761, "step": 5180 }, { "epoch": 2.285147642133098, "grad_norm": 0.2633442789610725, "learning_rate": 3.2646042421287625e-05, "loss": 1.7272, "step": 5185 }, { "epoch": 2.2873512560599383, "grad_norm": 0.3664081223341677, "learning_rate": 3.245673484164521e-05, "loss": 1.6607, "step": 5190 }, { "epoch": 2.2895548699867785, "grad_norm": 0.36095917628329555, "learning_rate": 3.2267871348514475e-05, "loss": 1.6644, "step": 5195 }, { "epoch": 2.2917584839136182, "grad_norm": 0.3155817104820564, "learning_rate": 3.207945318364376e-05, "loss": 1.7833, "step": 5200 }, { "epoch": 2.2939620978404585, "grad_norm": 0.5525894084347752, "learning_rate": 3.1891481585853224e-05, "loss": 1.7846, "step": 5205 }, { "epoch": 2.2961657117672982, "grad_norm": 0.3183611553729405, "learning_rate": 3.1703957791027104e-05, "loss": 1.8015, "step": 5210 }, { "epoch": 2.2983693256941384, "grad_norm": 0.3341156230287187, "learning_rate": 3.151688303210525e-05, "loss": 1.4901, "step": 5215 }, { "epoch": 2.300572939620978, "grad_norm": 0.29108487309353126, "learning_rate": 3.133025853907531e-05, "loss": 1.6021, "step": 5220 }, { "epoch": 2.3027765535478184, "grad_norm": 0.31003713411448125, "learning_rate": 3.114408553896437e-05, "loss": 1.6835, "step": 5225 }, { "epoch": 2.3049801674746586, "grad_norm": 0.31639971696298724, "learning_rate": 3.09583652558311e-05, "loss": 1.7131, "step": 5230 }, { "epoch": 2.3071837814014984, "grad_norm": 0.320459066996447, "learning_rate": 3.077309891075766e-05, "loss": 1.7207, "step": 5235 }, { "epoch": 2.3093873953283386, "grad_norm": 0.3658345444037042, "learning_rate": 3.058828772184155e-05, "loss": 1.637, "step": 5240 }, { "epoch": 2.3115910092551784, "grad_norm": 0.3119356619189155, "learning_rate": 3.0403932904187694e-05, "loss": 1.7374, "step": 5245 }, { "epoch": 2.3137946231820186, "grad_norm": 0.34062681606017264, "learning_rate": 3.0220035669900493e-05, "loss": 1.3662, "step": 5250 }, { "epoch": 2.3159982371088583, "grad_norm": 0.3231089104173347, "learning_rate": 3.0036597228075847e-05, "loss": 1.7862, "step": 5255 }, { "epoch": 2.3182018510356985, "grad_norm": 0.2845067164587982, "learning_rate": 2.985361878479307e-05, "loss": 1.6374, "step": 5260 }, { "epoch": 2.3204054649625387, "grad_norm": 0.3158916323796938, "learning_rate": 2.9671101543107037e-05, "loss": 1.7791, "step": 5265 }, { "epoch": 2.3226090788893785, "grad_norm": 0.3473567438345015, "learning_rate": 2.9489046703040478e-05, "loss": 1.6438, "step": 5270 }, { "epoch": 2.3248126928162187, "grad_norm": 0.32956604729846517, "learning_rate": 2.9307455461575728e-05, "loss": 1.5174, "step": 5275 }, { "epoch": 2.3270163067430585, "grad_norm": 0.31139915651994976, "learning_rate": 2.9126329012647048e-05, "loss": 1.6661, "step": 5280 }, { "epoch": 2.3292199206698987, "grad_norm": 0.3916001778917859, "learning_rate": 2.894566854713283e-05, "loss": 1.7324, "step": 5285 }, { "epoch": 2.3314235345967385, "grad_norm": 0.33538535373990935, "learning_rate": 2.8765475252847696e-05, "loss": 1.8397, "step": 5290 }, { "epoch": 2.3336271485235787, "grad_norm": 0.4045876334606375, "learning_rate": 2.8585750314534633e-05, "loss": 1.761, "step": 5295 }, { "epoch": 2.335830762450419, "grad_norm": 0.34476500843118907, "learning_rate": 2.8406494913857264e-05, "loss": 1.7239, "step": 5300 }, { "epoch": 2.3380343763772586, "grad_norm": 0.3377854881404091, "learning_rate": 2.8227710229392102e-05, "loss": 1.6767, "step": 5305 }, { "epoch": 2.340237990304099, "grad_norm": 0.2943468051471504, "learning_rate": 2.8049397436620817e-05, "loss": 1.7027, "step": 5310 }, { "epoch": 2.3424416042309386, "grad_norm": 0.3325646272609782, "learning_rate": 2.7871557707922356e-05, "loss": 1.7092, "step": 5315 }, { "epoch": 2.344645218157779, "grad_norm": 0.3194408987322575, "learning_rate": 2.769419221256546e-05, "loss": 1.5551, "step": 5320 }, { "epoch": 2.3468488320846186, "grad_norm": 0.3154791851100986, "learning_rate": 2.751730211670075e-05, "loss": 1.5952, "step": 5325 }, { "epoch": 2.349052446011459, "grad_norm": 0.37152524303673207, "learning_rate": 2.7340888583353263e-05, "loss": 1.6328, "step": 5330 }, { "epoch": 2.351256059938299, "grad_norm": 0.364737719498444, "learning_rate": 2.716495277241463e-05, "loss": 1.5125, "step": 5335 }, { "epoch": 2.3534596738651388, "grad_norm": 0.3524349639748066, "learning_rate": 2.6989495840635615e-05, "loss": 1.6589, "step": 5340 }, { "epoch": 2.355663287791979, "grad_norm": 0.27640673317944237, "learning_rate": 2.6814518941618326e-05, "loss": 1.5661, "step": 5345 }, { "epoch": 2.3578669017188187, "grad_norm": 0.31837314971643976, "learning_rate": 2.6640023225808852e-05, "loss": 1.7214, "step": 5350 }, { "epoch": 2.360070515645659, "grad_norm": 0.29946292415188364, "learning_rate": 2.6466009840489436e-05, "loss": 1.4745, "step": 5355 }, { "epoch": 2.3622741295724987, "grad_norm": 0.3368911389398718, "learning_rate": 2.629247992977122e-05, "loss": 1.6371, "step": 5360 }, { "epoch": 2.364477743499339, "grad_norm": 0.3078861715002911, "learning_rate": 2.6119434634586427e-05, "loss": 1.6562, "step": 5365 }, { "epoch": 2.366681357426179, "grad_norm": 0.39044087680489714, "learning_rate": 2.5946875092681134e-05, "loss": 1.7854, "step": 5370 }, { "epoch": 2.368884971353019, "grad_norm": 0.3905475645799491, "learning_rate": 2.5774802438607627e-05, "loss": 1.7027, "step": 5375 }, { "epoch": 2.371088585279859, "grad_norm": 0.3344158542218199, "learning_rate": 2.5603217803716938e-05, "loss": 1.6856, "step": 5380 }, { "epoch": 2.373292199206699, "grad_norm": 0.3171235587081351, "learning_rate": 2.5432122316151463e-05, "loss": 1.6338, "step": 5385 }, { "epoch": 2.375495813133539, "grad_norm": 0.31814287079371045, "learning_rate": 2.5261517100837563e-05, "loss": 1.6072, "step": 5390 }, { "epoch": 2.377699427060379, "grad_norm": 0.3715801248464969, "learning_rate": 2.509140327947814e-05, "loss": 1.7025, "step": 5395 }, { "epoch": 2.379903040987219, "grad_norm": 0.30689235196019754, "learning_rate": 2.4921781970545178e-05, "loss": 1.704, "step": 5400 }, { "epoch": 2.3821066549140593, "grad_norm": 0.2783210103898173, "learning_rate": 2.4752654289272568e-05, "loss": 1.8138, "step": 5405 }, { "epoch": 2.384310268840899, "grad_norm": 0.4115102536349836, "learning_rate": 2.4584021347648645e-05, "loss": 1.7562, "step": 5410 }, { "epoch": 2.3865138827677392, "grad_norm": 0.320397115810562, "learning_rate": 2.441588425440886e-05, "loss": 1.7002, "step": 5415 }, { "epoch": 2.388717496694579, "grad_norm": 0.32551710279990625, "learning_rate": 2.424824411502856e-05, "loss": 1.6053, "step": 5420 }, { "epoch": 2.390921110621419, "grad_norm": 0.37060732842102345, "learning_rate": 2.408110203171572e-05, "loss": 1.6564, "step": 5425 }, { "epoch": 2.393124724548259, "grad_norm": 0.33585884001747096, "learning_rate": 2.3914459103403696e-05, "loss": 1.7012, "step": 5430 }, { "epoch": 2.395328338475099, "grad_norm": 0.297871745937653, "learning_rate": 2.374831642574392e-05, "loss": 1.7399, "step": 5435 }, { "epoch": 2.3975319524019394, "grad_norm": 0.28026347006814506, "learning_rate": 2.3582675091098717e-05, "loss": 1.6698, "step": 5440 }, { "epoch": 2.399735566328779, "grad_norm": 0.3092926951118008, "learning_rate": 2.3417536188534327e-05, "loss": 1.7019, "step": 5445 }, { "epoch": 2.4019391802556194, "grad_norm": 0.29956315407231204, "learning_rate": 2.3252900803813415e-05, "loss": 1.7835, "step": 5450 }, { "epoch": 2.404142794182459, "grad_norm": 0.3472861996721381, "learning_rate": 2.3088770019388116e-05, "loss": 1.7523, "step": 5455 }, { "epoch": 2.4063464081092993, "grad_norm": 0.3346967409970768, "learning_rate": 2.292514491439297e-05, "loss": 1.543, "step": 5460 }, { "epoch": 2.408550022036139, "grad_norm": 0.30667047442127315, "learning_rate": 2.2762026564637717e-05, "loss": 1.7131, "step": 5465 }, { "epoch": 2.4107536359629793, "grad_norm": 0.32172911896887835, "learning_rate": 2.259941604260024e-05, "loss": 1.4888, "step": 5470 }, { "epoch": 2.4129572498898195, "grad_norm": 0.33661859838509534, "learning_rate": 2.2437314417419518e-05, "loss": 1.6434, "step": 5475 }, { "epoch": 2.4151608638166593, "grad_norm": 0.2766235524609549, "learning_rate": 2.2275722754888662e-05, "loss": 1.4497, "step": 5480 }, { "epoch": 2.4173644777434995, "grad_norm": 0.31531488102987404, "learning_rate": 2.211464211744787e-05, "loss": 1.7619, "step": 5485 }, { "epoch": 2.4195680916703393, "grad_norm": 0.4045821730130733, "learning_rate": 2.195407356417737e-05, "loss": 1.5253, "step": 5490 }, { "epoch": 2.4217717055971795, "grad_norm": 0.3077739063590653, "learning_rate": 2.1794018150790507e-05, "loss": 1.3737, "step": 5495 }, { "epoch": 2.4239753195240192, "grad_norm": 0.26933598773812595, "learning_rate": 2.1634476929626868e-05, "loss": 1.5562, "step": 5500 }, { "epoch": 2.4261789334508594, "grad_norm": 0.3180561976663089, "learning_rate": 2.1475450949645325e-05, "loss": 1.8415, "step": 5505 }, { "epoch": 2.4283825473776997, "grad_norm": 0.326460873305707, "learning_rate": 2.1316941256417024e-05, "loss": 1.6886, "step": 5510 }, { "epoch": 2.4305861613045394, "grad_norm": 0.2651305009582311, "learning_rate": 2.115894889211869e-05, "loss": 1.556, "step": 5515 }, { "epoch": 2.4327897752313796, "grad_norm": 0.285018315967615, "learning_rate": 2.100147489552562e-05, "loss": 1.6264, "step": 5520 }, { "epoch": 2.4349933891582194, "grad_norm": 0.3039400323881749, "learning_rate": 2.084452030200502e-05, "loss": 1.68, "step": 5525 }, { "epoch": 2.4371970030850596, "grad_norm": 0.4530331306706194, "learning_rate": 2.068808614350899e-05, "loss": 1.8822, "step": 5530 }, { "epoch": 2.4394006170118994, "grad_norm": 0.32124417856720616, "learning_rate": 2.0532173448567936e-05, "loss": 1.6896, "step": 5535 }, { "epoch": 2.4416042309387396, "grad_norm": 0.3526640578616687, "learning_rate": 2.037678324228366e-05, "loss": 1.3874, "step": 5540 }, { "epoch": 2.44380784486558, "grad_norm": 0.3025841716329574, "learning_rate": 2.022191654632274e-05, "loss": 1.5026, "step": 5545 }, { "epoch": 2.4460114587924195, "grad_norm": 0.3451825131993311, "learning_rate": 2.0067574378909726e-05, "loss": 1.6466, "step": 5550 }, { "epoch": 2.4482150727192598, "grad_norm": 0.2946116685019073, "learning_rate": 1.9913757754820483e-05, "loss": 1.7246, "step": 5555 }, { "epoch": 2.4504186866460995, "grad_norm": 0.3604250683804044, "learning_rate": 1.976046768537544e-05, "loss": 1.5967, "step": 5560 }, { "epoch": 2.4526223005729397, "grad_norm": 0.356529608888727, "learning_rate": 1.9607705178433124e-05, "loss": 1.8344, "step": 5565 }, { "epoch": 2.4548259144997795, "grad_norm": 0.3062637993405146, "learning_rate": 1.9455471238383394e-05, "loss": 1.727, "step": 5570 }, { "epoch": 2.4570295284266197, "grad_norm": 0.2913261589880019, "learning_rate": 1.9303766866140794e-05, "loss": 1.6422, "step": 5575 }, { "epoch": 2.45923314235346, "grad_norm": 0.28539967100565233, "learning_rate": 1.9152593059138036e-05, "loss": 1.5191, "step": 5580 }, { "epoch": 2.4614367562802997, "grad_norm": 0.38278133166473177, "learning_rate": 1.9001950811319624e-05, "loss": 1.747, "step": 5585 }, { "epoch": 2.46364037020714, "grad_norm": 0.3213863995857726, "learning_rate": 1.885184111313494e-05, "loss": 1.6493, "step": 5590 }, { "epoch": 2.4658439841339796, "grad_norm": 0.3515687604082846, "learning_rate": 1.870226495153199e-05, "loss": 1.6207, "step": 5595 }, { "epoch": 2.46804759806082, "grad_norm": 0.2818899743965069, "learning_rate": 1.8553223309950907e-05, "loss": 1.5783, "step": 5600 }, { "epoch": 2.4702512119876596, "grad_norm": 0.32339197111018464, "learning_rate": 1.8404717168317444e-05, "loss": 1.6023, "step": 5605 }, { "epoch": 2.4724548259145, "grad_norm": 0.32235437648153625, "learning_rate": 1.8256747503036465e-05, "loss": 1.5901, "step": 5610 }, { "epoch": 2.47465843984134, "grad_norm": 0.2996816683579224, "learning_rate": 1.8109315286985575e-05, "loss": 1.7065, "step": 5615 }, { "epoch": 2.47686205376818, "grad_norm": 0.34309522162413447, "learning_rate": 1.7962421489508797e-05, "loss": 1.7226, "step": 5620 }, { "epoch": 2.47906566769502, "grad_norm": 0.29849313259138993, "learning_rate": 1.7816067076410138e-05, "loss": 1.7579, "step": 5625 }, { "epoch": 2.48126928162186, "grad_norm": 0.3368071136503966, "learning_rate": 1.7670253009947146e-05, "loss": 1.5962, "step": 5630 }, { "epoch": 2.4834728955487, "grad_norm": 0.3175198598873115, "learning_rate": 1.7524980248824806e-05, "loss": 1.7556, "step": 5635 }, { "epoch": 2.4856765094755398, "grad_norm": 0.30889458937514946, "learning_rate": 1.738024974818896e-05, "loss": 1.7268, "step": 5640 }, { "epoch": 2.48788012340238, "grad_norm": 0.30454000273861387, "learning_rate": 1.7236062459620306e-05, "loss": 1.6084, "step": 5645 }, { "epoch": 2.49008373732922, "grad_norm": 0.30973625136809374, "learning_rate": 1.7092419331127894e-05, "loss": 1.581, "step": 5650 }, { "epoch": 2.49228735125606, "grad_norm": 0.3312428081002817, "learning_rate": 1.6949321307143096e-05, "loss": 1.6826, "step": 5655 }, { "epoch": 2.4944909651829, "grad_norm": 0.29438754562772956, "learning_rate": 1.6806769328513226e-05, "loss": 1.6531, "step": 5660 }, { "epoch": 2.49669457910974, "grad_norm": 0.3072379816792179, "learning_rate": 1.666476433249552e-05, "loss": 1.6036, "step": 5665 }, { "epoch": 2.49889819303658, "grad_norm": 0.2653188527319854, "learning_rate": 1.6523307252750787e-05, "loss": 1.6677, "step": 5670 }, { "epoch": 2.50110180696342, "grad_norm": 0.31853001670175285, "learning_rate": 1.6382399019337493e-05, "loss": 1.7512, "step": 5675 }, { "epoch": 2.50330542089026, "grad_norm": 0.3541014940940505, "learning_rate": 1.6242040558705386e-05, "loss": 1.4784, "step": 5680 }, { "epoch": 2.5055090348171003, "grad_norm": 0.3022737220008577, "learning_rate": 1.6102232793689652e-05, "loss": 1.552, "step": 5685 }, { "epoch": 2.50771264874394, "grad_norm": 0.2865580516371412, "learning_rate": 1.5962976643504734e-05, "loss": 1.6162, "step": 5690 }, { "epoch": 2.50991626267078, "grad_norm": 0.33229646654342737, "learning_rate": 1.5824273023738223e-05, "loss": 1.7025, "step": 5695 }, { "epoch": 2.51211987659762, "grad_norm": 0.25315774233489335, "learning_rate": 1.5686122846344932e-05, "loss": 1.6556, "step": 5700 }, { "epoch": 2.5143234905244602, "grad_norm": 0.32926703632524384, "learning_rate": 1.55485270196409e-05, "loss": 1.7055, "step": 5705 }, { "epoch": 2.5165271044513, "grad_norm": 0.38254621782613324, "learning_rate": 1.541148644829743e-05, "loss": 1.8189, "step": 5710 }, { "epoch": 2.51873071837814, "grad_norm": 0.3262138673835723, "learning_rate": 1.5275002033335016e-05, "loss": 1.6328, "step": 5715 }, { "epoch": 2.5209343323049804, "grad_norm": 0.36898284620242594, "learning_rate": 1.5139074672117514e-05, "loss": 1.7229, "step": 5720 }, { "epoch": 2.52313794623182, "grad_norm": 0.29383774220673775, "learning_rate": 1.500370525834639e-05, "loss": 1.7057, "step": 5725 }, { "epoch": 2.52534156015866, "grad_norm": 0.35255942435868104, "learning_rate": 1.4868894682054535e-05, "loss": 1.703, "step": 5730 }, { "epoch": 2.5275451740855, "grad_norm": 0.3224975998851544, "learning_rate": 1.473464382960057e-05, "loss": 1.6255, "step": 5735 }, { "epoch": 2.5297487880123404, "grad_norm": 0.2728863606551014, "learning_rate": 1.4600953583663114e-05, "loss": 1.5348, "step": 5740 }, { "epoch": 2.53195240193918, "grad_norm": 0.3292693963632781, "learning_rate": 1.4467824823234843e-05, "loss": 1.6536, "step": 5745 }, { "epoch": 2.5341560158660204, "grad_norm": 0.3530164649944846, "learning_rate": 1.4335258423616737e-05, "loss": 1.631, "step": 5750 }, { "epoch": 2.5363596297928606, "grad_norm": 0.3107181811257503, "learning_rate": 1.4203255256412318e-05, "loss": 1.5969, "step": 5755 }, { "epoch": 2.5385632437197003, "grad_norm": 0.3700803583018722, "learning_rate": 1.407181618952199e-05, "loss": 1.7883, "step": 5760 }, { "epoch": 2.54076685764654, "grad_norm": 0.30243816752307084, "learning_rate": 1.394094208713732e-05, "loss": 1.652, "step": 5765 }, { "epoch": 2.5429704715733803, "grad_norm": 0.3413616757398393, "learning_rate": 1.3810633809735196e-05, "loss": 1.7507, "step": 5770 }, { "epoch": 2.5451740855002205, "grad_norm": 0.3120087119800215, "learning_rate": 1.3680892214072405e-05, "loss": 1.7198, "step": 5775 }, { "epoch": 2.5473776994270603, "grad_norm": 0.3518504932831635, "learning_rate": 1.3551718153179871e-05, "loss": 1.8579, "step": 5780 }, { "epoch": 2.5495813133539005, "grad_norm": 0.34752155952619, "learning_rate": 1.3423112476357036e-05, "loss": 1.5468, "step": 5785 }, { "epoch": 2.5517849272807407, "grad_norm": 0.3260308226099618, "learning_rate": 1.3295076029166265e-05, "loss": 1.5258, "step": 5790 }, { "epoch": 2.5539885412075805, "grad_norm": 0.35584838362058835, "learning_rate": 1.3167609653427426e-05, "loss": 1.7373, "step": 5795 }, { "epoch": 2.55619215513442, "grad_norm": 0.32184918024972864, "learning_rate": 1.304071418721221e-05, "loss": 1.6741, "step": 5800 }, { "epoch": 2.5583957690612604, "grad_norm": 0.315994062331127, "learning_rate": 1.2914390464838655e-05, "loss": 1.7156, "step": 5805 }, { "epoch": 2.5605993829881006, "grad_norm": 0.23924524979523612, "learning_rate": 1.2788639316865635e-05, "loss": 1.738, "step": 5810 }, { "epoch": 2.5628029969149404, "grad_norm": 0.3098121698841723, "learning_rate": 1.266346157008753e-05, "loss": 1.5198, "step": 5815 }, { "epoch": 2.5650066108417806, "grad_norm": 0.3262487762949137, "learning_rate": 1.2538858047528646e-05, "loss": 1.587, "step": 5820 }, { "epoch": 2.5672102247686204, "grad_norm": 0.33166973277913664, "learning_rate": 1.2414829568437825e-05, "loss": 1.5043, "step": 5825 }, { "epoch": 2.5694138386954606, "grad_norm": 0.3022890966007592, "learning_rate": 1.2291376948283139e-05, "loss": 1.605, "step": 5830 }, { "epoch": 2.5716174526223003, "grad_norm": 0.3389630036906691, "learning_rate": 1.2168500998746435e-05, "loss": 1.6955, "step": 5835 }, { "epoch": 2.5738210665491406, "grad_norm": 0.3796380833169501, "learning_rate": 1.2046202527718076e-05, "loss": 1.6275, "step": 5840 }, { "epoch": 2.5760246804759808, "grad_norm": 0.3238245323234113, "learning_rate": 1.1924482339291554e-05, "loss": 1.7289, "step": 5845 }, { "epoch": 2.5782282944028205, "grad_norm": 0.30218042381327415, "learning_rate": 1.1803341233758291e-05, "loss": 1.5412, "step": 5850 }, { "epoch": 2.5804319083296607, "grad_norm": 0.33206646260037787, "learning_rate": 1.1682780007602268e-05, "loss": 1.8011, "step": 5855 }, { "epoch": 2.5826355222565005, "grad_norm": 0.3538395424898534, "learning_rate": 1.1562799453494899e-05, "loss": 1.7862, "step": 5860 }, { "epoch": 2.5848391361833407, "grad_norm": 0.35528233062631853, "learning_rate": 1.144340036028978e-05, "loss": 1.685, "step": 5865 }, { "epoch": 2.5870427501101805, "grad_norm": 0.33412660972869573, "learning_rate": 1.132458351301744e-05, "loss": 1.7575, "step": 5870 }, { "epoch": 2.5892463640370207, "grad_norm": 0.3175660326892495, "learning_rate": 1.1206349692880236e-05, "loss": 1.5658, "step": 5875 }, { "epoch": 2.591449977963861, "grad_norm": 0.30007433696171515, "learning_rate": 1.1088699677247238e-05, "loss": 1.586, "step": 5880 }, { "epoch": 2.5936535918907007, "grad_norm": 0.3180669505120074, "learning_rate": 1.097163423964912e-05, "loss": 1.7819, "step": 5885 }, { "epoch": 2.595857205817541, "grad_norm": 0.25188231699393177, "learning_rate": 1.0855154149772994e-05, "loss": 1.512, "step": 5890 }, { "epoch": 2.5980608197443806, "grad_norm": 0.2725825366458783, "learning_rate": 1.0739260173457355e-05, "loss": 1.7237, "step": 5895 }, { "epoch": 2.600264433671221, "grad_norm": 0.4353012547028696, "learning_rate": 1.0623953072687265e-05, "loss": 1.5664, "step": 5900 }, { "epoch": 2.6024680475980606, "grad_norm": 0.3208919789812601, "learning_rate": 1.0509233605588997e-05, "loss": 1.7184, "step": 5905 }, { "epoch": 2.604671661524901, "grad_norm": 0.2811764278977997, "learning_rate": 1.0395102526425282e-05, "loss": 1.6933, "step": 5910 }, { "epoch": 2.606875275451741, "grad_norm": 0.3283193021777617, "learning_rate": 1.0281560585590311e-05, "loss": 1.709, "step": 5915 }, { "epoch": 2.609078889378581, "grad_norm": 0.31601773043700576, "learning_rate": 1.0168608529604783e-05, "loss": 1.5517, "step": 5920 }, { "epoch": 2.611282503305421, "grad_norm": 0.26388766629051863, "learning_rate": 1.0056247101110972e-05, "loss": 1.5716, "step": 5925 }, { "epoch": 2.6134861172322608, "grad_norm": 0.3170004934043696, "learning_rate": 9.944477038867838e-06, "loss": 1.3933, "step": 5930 }, { "epoch": 2.615689731159101, "grad_norm": 0.3026100677449506, "learning_rate": 9.833299077746261e-06, "loss": 1.538, "step": 5935 }, { "epoch": 2.6178933450859407, "grad_norm": 0.32599901382842245, "learning_rate": 9.72271394872416e-06, "loss": 1.722, "step": 5940 }, { "epoch": 2.620096959012781, "grad_norm": 0.31252235277224677, "learning_rate": 9.612722378881578e-06, "loss": 1.5777, "step": 5945 }, { "epoch": 2.622300572939621, "grad_norm": 0.27949230064797415, "learning_rate": 9.503325091396098e-06, "loss": 1.7781, "step": 5950 }, { "epoch": 2.624504186866461, "grad_norm": 0.3031767326868999, "learning_rate": 9.394522805537931e-06, "loss": 1.6123, "step": 5955 }, { "epoch": 2.626707800793301, "grad_norm": 0.3596489809565846, "learning_rate": 9.286316236665271e-06, "loss": 1.8234, "step": 5960 }, { "epoch": 2.628911414720141, "grad_norm": 0.37067976899261396, "learning_rate": 9.178706096219547e-06, "loss": 1.5176, "step": 5965 }, { "epoch": 2.631115028646981, "grad_norm": 0.30187017500054564, "learning_rate": 9.0716930917208e-06, "loss": 1.5401, "step": 5970 }, { "epoch": 2.633318642573821, "grad_norm": 0.338495742270805, "learning_rate": 8.965277926762916e-06, "loss": 1.5802, "step": 5975 }, { "epoch": 2.635522256500661, "grad_norm": 0.33527970219407616, "learning_rate": 8.859461301009186e-06, "loss": 1.6115, "step": 5980 }, { "epoch": 2.6377258704275013, "grad_norm": 0.3284526374151953, "learning_rate": 8.754243910187498e-06, "loss": 1.7051, "step": 5985 }, { "epoch": 2.639929484354341, "grad_norm": 0.2820385948307506, "learning_rate": 8.649626446085945e-06, "loss": 1.4949, "step": 5990 }, { "epoch": 2.6421330982811813, "grad_norm": 0.3357888909663005, "learning_rate": 8.545609596548121e-06, "loss": 1.5265, "step": 5995 }, { "epoch": 2.644336712208021, "grad_norm": 0.24829105724152342, "learning_rate": 8.442194045468733e-06, "loss": 1.4418, "step": 6000 }, { "epoch": 2.6465403261348612, "grad_norm": 0.3445954692738449, "learning_rate": 8.339380472789016e-06, "loss": 1.471, "step": 6005 }, { "epoch": 2.648743940061701, "grad_norm": 0.34193200011238856, "learning_rate": 8.237169554492297e-06, "loss": 1.4714, "step": 6010 }, { "epoch": 2.650947553988541, "grad_norm": 0.2994838498414857, "learning_rate": 8.135561962599514e-06, "loss": 1.5747, "step": 6015 }, { "epoch": 2.6531511679153814, "grad_norm": 0.2510845629901009, "learning_rate": 8.034558365164868e-06, "loss": 1.6476, "step": 6020 }, { "epoch": 2.655354781842221, "grad_norm": 0.3375793364015892, "learning_rate": 7.934159426271403e-06, "loss": 1.6063, "step": 6025 }, { "epoch": 2.6575583957690614, "grad_norm": 0.33573003764748477, "learning_rate": 7.834365806026578e-06, "loss": 1.5814, "step": 6030 }, { "epoch": 2.659762009695901, "grad_norm": 0.38295818181236296, "learning_rate": 7.735178160557943e-06, "loss": 1.7642, "step": 6035 }, { "epoch": 2.6619656236227414, "grad_norm": 0.28003247319118707, "learning_rate": 7.636597142009017e-06, "loss": 1.7946, "step": 6040 }, { "epoch": 2.664169237549581, "grad_norm": 0.367119348898622, "learning_rate": 7.538623398534661e-06, "loss": 1.5553, "step": 6045 }, { "epoch": 2.6663728514764213, "grad_norm": 0.30087617582351794, "learning_rate": 7.441257574297089e-06, "loss": 1.56, "step": 6050 }, { "epoch": 2.6685764654032615, "grad_norm": 0.2679029740855203, "learning_rate": 7.344500309461511e-06, "loss": 1.5622, "step": 6055 }, { "epoch": 2.6707800793301013, "grad_norm": 0.38643212881302264, "learning_rate": 7.248352240192002e-06, "loss": 1.6344, "step": 6060 }, { "epoch": 2.6729836932569415, "grad_norm": 0.3036596598041859, "learning_rate": 7.15281399864719e-06, "loss": 1.6126, "step": 6065 }, { "epoch": 2.6751873071837813, "grad_norm": 0.3409000984647797, "learning_rate": 7.057886212976239e-06, "loss": 1.7453, "step": 6070 }, { "epoch": 2.6773909211106215, "grad_norm": 0.3123434626912612, "learning_rate": 6.963569507314627e-06, "loss": 1.6624, "step": 6075 }, { "epoch": 2.6795945350374613, "grad_norm": 0.3272418793360404, "learning_rate": 6.8698645017801325e-06, "loss": 1.8614, "step": 6080 }, { "epoch": 2.6817981489643015, "grad_norm": 0.2432770598979605, "learning_rate": 6.776771812468618e-06, "loss": 1.6761, "step": 6085 }, { "epoch": 2.6840017628911417, "grad_norm": 0.29638907037530365, "learning_rate": 6.684292051450147e-06, "loss": 1.5734, "step": 6090 }, { "epoch": 2.6862053768179814, "grad_norm": 0.2927874947198909, "learning_rate": 6.592425826764781e-06, "loss": 1.6527, "step": 6095 }, { "epoch": 2.6884089907448216, "grad_norm": 0.3262506822619773, "learning_rate": 6.501173742418753e-06, "loss": 1.7488, "step": 6100 }, { "epoch": 2.6906126046716614, "grad_norm": 0.29388683856815984, "learning_rate": 6.410536398380385e-06, "loss": 1.7391, "step": 6105 }, { "epoch": 2.6928162185985016, "grad_norm": 0.42564932659573224, "learning_rate": 6.320514390576193e-06, "loss": 1.5618, "step": 6110 }, { "epoch": 2.6950198325253414, "grad_norm": 0.33120895230745345, "learning_rate": 6.231108310886924e-06, "loss": 1.5172, "step": 6115 }, { "epoch": 2.6972234464521816, "grad_norm": 0.2970159546645706, "learning_rate": 6.142318747143716e-06, "loss": 1.5319, "step": 6120 }, { "epoch": 2.699427060379022, "grad_norm": 0.3125793002150719, "learning_rate": 6.054146283124218e-06, "loss": 1.6401, "step": 6125 }, { "epoch": 2.7016306743058616, "grad_norm": 0.30146872233206856, "learning_rate": 5.966591498548724e-06, "loss": 1.7384, "step": 6130 }, { "epoch": 2.7038342882327018, "grad_norm": 0.3887993733574576, "learning_rate": 5.8796549690763645e-06, "loss": 1.8019, "step": 6135 }, { "epoch": 2.7060379021595415, "grad_norm": 0.3033022818548619, "learning_rate": 5.79333726630138e-06, "loss": 1.5844, "step": 6140 }, { "epoch": 2.7082415160863818, "grad_norm": 0.368025241100897, "learning_rate": 5.7076389577493175e-06, "loss": 1.8454, "step": 6145 }, { "epoch": 2.7104451300132215, "grad_norm": 0.298047160015892, "learning_rate": 5.622560606873262e-06, "loss": 1.6045, "step": 6150 }, { "epoch": 2.7126487439400617, "grad_norm": 0.30605420571316905, "learning_rate": 5.538102773050235e-06, "loss": 1.696, "step": 6155 }, { "epoch": 2.714852357866902, "grad_norm": 0.3454395184873584, "learning_rate": 5.454266011577369e-06, "loss": 1.6258, "step": 6160 }, { "epoch": 2.7170559717937417, "grad_norm": 0.30863792610727064, "learning_rate": 5.371050873668437e-06, "loss": 1.5895, "step": 6165 }, { "epoch": 2.719259585720582, "grad_norm": 0.32671612304128317, "learning_rate": 5.2884579064500615e-06, "loss": 1.751, "step": 6170 }, { "epoch": 2.7214631996474217, "grad_norm": 0.24404249683796003, "learning_rate": 5.206487652958214e-06, "loss": 1.5318, "step": 6175 }, { "epoch": 2.723666813574262, "grad_norm": 0.31345216172176077, "learning_rate": 5.125140652134652e-06, "loss": 1.6814, "step": 6180 }, { "epoch": 2.7258704275011016, "grad_norm": 0.3301202829233535, "learning_rate": 5.044417438823279e-06, "loss": 1.6688, "step": 6185 }, { "epoch": 2.728074041427942, "grad_norm": 0.3180153311679256, "learning_rate": 4.964318543766733e-06, "loss": 1.8152, "step": 6190 }, { "epoch": 2.730277655354782, "grad_norm": 0.3477053424267945, "learning_rate": 4.884844493602847e-06, "loss": 1.6068, "step": 6195 }, { "epoch": 2.732481269281622, "grad_norm": 0.3829450747175287, "learning_rate": 4.805995810861219e-06, "loss": 1.5436, "step": 6200 }, { "epoch": 2.734684883208462, "grad_norm": 0.35144134365971347, "learning_rate": 4.727773013959702e-06, "loss": 1.7733, "step": 6205 }, { "epoch": 2.736888497135302, "grad_norm": 0.3511048030633929, "learning_rate": 4.650176617201074e-06, "loss": 1.7483, "step": 6210 }, { "epoch": 2.739092111062142, "grad_norm": 0.3146009823620477, "learning_rate": 4.573207130769663e-06, "loss": 1.6416, "step": 6215 }, { "epoch": 2.7412957249889818, "grad_norm": 0.2961123189548949, "learning_rate": 4.496865060727917e-06, "loss": 1.5871, "step": 6220 }, { "epoch": 2.743499338915822, "grad_norm": 0.2827248539547275, "learning_rate": 4.421150909013094e-06, "loss": 1.6537, "step": 6225 }, { "epoch": 2.745702952842662, "grad_norm": 0.3230949462447901, "learning_rate": 4.346065173434055e-06, "loss": 1.5128, "step": 6230 }, { "epoch": 2.747906566769502, "grad_norm": 0.27203611650594783, "learning_rate": 4.271608347667888e-06, "loss": 1.6916, "step": 6235 }, { "epoch": 2.750110180696342, "grad_norm": 0.316718312569841, "learning_rate": 4.197780921256678e-06, "loss": 1.7967, "step": 6240 }, { "epoch": 2.752313794623182, "grad_norm": 0.31564921665656825, "learning_rate": 4.1245833796043184e-06, "loss": 1.5092, "step": 6245 }, { "epoch": 2.754517408550022, "grad_norm": 0.3023941557144956, "learning_rate": 4.052016203973319e-06, "loss": 1.6864, "step": 6250 }, { "epoch": 2.756721022476862, "grad_norm": 0.31013607515444674, "learning_rate": 3.9800798714816566e-06, "loss": 1.7096, "step": 6255 }, { "epoch": 2.758924636403702, "grad_norm": 0.34766875505420175, "learning_rate": 3.908774855099529e-06, "loss": 1.6837, "step": 6260 }, { "epoch": 2.7611282503305423, "grad_norm": 0.2943152497920515, "learning_rate": 3.838101623646429e-06, "loss": 1.6478, "step": 6265 }, { "epoch": 2.763331864257382, "grad_norm": 0.32690567367741863, "learning_rate": 3.768060641787874e-06, "loss": 1.8321, "step": 6270 }, { "epoch": 2.765535478184222, "grad_norm": 0.35174727110880194, "learning_rate": 3.698652370032496e-06, "loss": 1.7583, "step": 6275 }, { "epoch": 2.767739092111062, "grad_norm": 0.30775438684619605, "learning_rate": 3.6298772647289204e-06, "loss": 1.7887, "step": 6280 }, { "epoch": 2.7699427060379023, "grad_norm": 0.2974347187890948, "learning_rate": 3.561735778062847e-06, "loss": 1.5669, "step": 6285 }, { "epoch": 2.772146319964742, "grad_norm": 0.32631762135645986, "learning_rate": 3.4942283580539747e-06, "loss": 1.5496, "step": 6290 }, { "epoch": 2.7743499338915822, "grad_norm": 0.3078445012000501, "learning_rate": 3.427355448553149e-06, "loss": 1.4473, "step": 6295 }, { "epoch": 2.7765535478184225, "grad_norm": 0.31874633296757016, "learning_rate": 3.3611174892393848e-06, "loss": 1.7297, "step": 6300 }, { "epoch": 2.778757161745262, "grad_norm": 0.3877251343286691, "learning_rate": 3.2955149156170373e-06, "loss": 1.7889, "step": 6305 }, { "epoch": 2.780960775672102, "grad_norm": 0.3050428005282347, "learning_rate": 3.230548159012836e-06, "loss": 1.7297, "step": 6310 }, { "epoch": 2.783164389598942, "grad_norm": 0.3297939270880444, "learning_rate": 3.1662176465731776e-06, "loss": 1.7542, "step": 6315 }, { "epoch": 2.7853680035257824, "grad_norm": 0.35972869738440505, "learning_rate": 3.1025238012612146e-06, "loss": 1.6169, "step": 6320 }, { "epoch": 2.787571617452622, "grad_norm": 0.3384226865100565, "learning_rate": 3.039467041854105e-06, "loss": 1.6362, "step": 6325 }, { "epoch": 2.7897752313794624, "grad_norm": 0.3670760451609635, "learning_rate": 2.97704778294029e-06, "loss": 1.6398, "step": 6330 }, { "epoch": 2.7919788453063026, "grad_norm": 0.2854714227749342, "learning_rate": 2.9152664349167415e-06, "loss": 1.4325, "step": 6335 }, { "epoch": 2.7941824592331423, "grad_norm": 0.3042468904319354, "learning_rate": 2.854123403986253e-06, "loss": 1.6423, "step": 6340 }, { "epoch": 2.796386073159982, "grad_norm": 0.33954395775276786, "learning_rate": 2.793619092154787e-06, "loss": 1.6785, "step": 6345 }, { "epoch": 2.7985896870868223, "grad_norm": 0.41052304755954266, "learning_rate": 2.7337538972287967e-06, "loss": 1.7808, "step": 6350 }, { "epoch": 2.8007933010136625, "grad_norm": 0.2735021459128279, "learning_rate": 2.674528212812721e-06, "loss": 1.579, "step": 6355 }, { "epoch": 2.8029969149405023, "grad_norm": 0.42116340427498267, "learning_rate": 2.6159424283062507e-06, "loss": 1.665, "step": 6360 }, { "epoch": 2.8052005288673425, "grad_norm": 0.40395544915333126, "learning_rate": 2.557996928901829e-06, "loss": 1.6685, "step": 6365 }, { "epoch": 2.8074041427941827, "grad_norm": 0.41455342226161546, "learning_rate": 2.5006920955821465e-06, "loss": 1.7578, "step": 6370 }, { "epoch": 2.8096077567210225, "grad_norm": 0.32321328330549354, "learning_rate": 2.4440283051176405e-06, "loss": 1.7026, "step": 6375 }, { "epoch": 2.8118113706478622, "grad_norm": 0.2956835138059502, "learning_rate": 2.388005930063941e-06, "loss": 1.8632, "step": 6380 }, { "epoch": 2.8140149845747024, "grad_norm": 0.3268794222372435, "learning_rate": 2.3326253387594753e-06, "loss": 1.6233, "step": 6385 }, { "epoch": 2.8162185985015427, "grad_norm": 0.3626797930342101, "learning_rate": 2.277886895323078e-06, "loss": 1.74, "step": 6390 }, { "epoch": 2.8184222124283824, "grad_norm": 0.33358450270104334, "learning_rate": 2.2237909596515396e-06, "loss": 1.4655, "step": 6395 }, { "epoch": 2.8206258263552226, "grad_norm": 0.29838126441173135, "learning_rate": 2.1703378874172507e-06, "loss": 1.4969, "step": 6400 }, { "epoch": 2.822829440282063, "grad_norm": 0.34601838203569607, "learning_rate": 2.117528030065907e-06, "loss": 1.6886, "step": 6405 }, { "epoch": 2.8250330542089026, "grad_norm": 0.3110562652655748, "learning_rate": 2.0653617348141084e-06, "loss": 1.4905, "step": 6410 }, { "epoch": 2.8272366681357424, "grad_norm": 0.3274079081069346, "learning_rate": 2.013839344647217e-06, "loss": 1.6808, "step": 6415 }, { "epoch": 2.8294402820625826, "grad_norm": 0.34133335820544286, "learning_rate": 1.962961198316937e-06, "loss": 1.7414, "step": 6420 }, { "epoch": 2.831643895989423, "grad_norm": 0.3234135423197038, "learning_rate": 1.912727630339217e-06, "loss": 1.4927, "step": 6425 }, { "epoch": 2.8338475099162626, "grad_norm": 0.31099973650003665, "learning_rate": 1.8631389709919843e-06, "loss": 1.5605, "step": 6430 }, { "epoch": 2.8360511238431028, "grad_norm": 0.37303627703284004, "learning_rate": 1.8141955463129912e-06, "loss": 1.6712, "step": 6435 }, { "epoch": 2.838254737769943, "grad_norm": 0.3614598619091711, "learning_rate": 1.7658976780976944e-06, "loss": 1.7914, "step": 6440 }, { "epoch": 2.8404583516967827, "grad_norm": 0.30969623455642703, "learning_rate": 1.7182456838971016e-06, "loss": 1.5793, "step": 6445 }, { "epoch": 2.8426619656236225, "grad_norm": 0.26680083024728835, "learning_rate": 1.6712398770156734e-06, "loss": 1.5423, "step": 6450 }, { "epoch": 2.8448655795504627, "grad_norm": 0.38620184585663864, "learning_rate": 1.6248805665093348e-06, "loss": 1.7361, "step": 6455 }, { "epoch": 2.847069193477303, "grad_norm": 0.29245866381267194, "learning_rate": 1.5791680571833667e-06, "loss": 1.4591, "step": 6460 }, { "epoch": 2.8492728074041427, "grad_norm": 0.36925991583350115, "learning_rate": 1.5341026495904409e-06, "loss": 1.5466, "step": 6465 }, { "epoch": 2.851476421330983, "grad_norm": 0.31621129108601936, "learning_rate": 1.4896846400286323e-06, "loss": 1.5198, "step": 6470 }, { "epoch": 2.853680035257823, "grad_norm": 0.3918256317744239, "learning_rate": 1.4459143205394876e-06, "loss": 1.8413, "step": 6475 }, { "epoch": 2.855883649184663, "grad_norm": 0.3333872096143664, "learning_rate": 1.4027919789060818e-06, "loss": 1.6091, "step": 6480 }, { "epoch": 2.8580872631115026, "grad_norm": 0.350383265332482, "learning_rate": 1.36031789865112e-06, "loss": 1.77, "step": 6485 }, { "epoch": 2.860290877038343, "grad_norm": 0.32052781126282354, "learning_rate": 1.3184923590351062e-06, "loss": 1.6178, "step": 6490 }, { "epoch": 2.862494490965183, "grad_norm": 0.33800257995746813, "learning_rate": 1.27731563505451e-06, "loss": 1.6759, "step": 6495 }, { "epoch": 2.864698104892023, "grad_norm": 0.32607540150243636, "learning_rate": 1.236787997439892e-06, "loss": 1.5576, "step": 6500 }, { "epoch": 2.866901718818863, "grad_norm": 0.32271754224482374, "learning_rate": 1.196909712654204e-06, "loss": 1.5769, "step": 6505 }, { "epoch": 2.8691053327457032, "grad_norm": 0.32279106125213675, "learning_rate": 1.1576810428910012e-06, "loss": 1.4904, "step": 6510 }, { "epoch": 2.871308946672543, "grad_norm": 0.33356280248337183, "learning_rate": 1.1191022460727007e-06, "loss": 1.5742, "step": 6515 }, { "epoch": 2.8735125605993828, "grad_norm": 0.3151589561132343, "learning_rate": 1.0811735758489372e-06, "loss": 1.6439, "step": 6520 }, { "epoch": 2.875716174526223, "grad_norm": 0.3317166675779296, "learning_rate": 1.04389528159482e-06, "loss": 1.4983, "step": 6525 }, { "epoch": 2.877919788453063, "grad_norm": 0.3198594908928924, "learning_rate": 1.0072676084093902e-06, "loss": 1.6749, "step": 6530 }, { "epoch": 2.880123402379903, "grad_norm": 0.2937946173545102, "learning_rate": 9.712907971139218e-07, "loss": 1.7593, "step": 6535 }, { "epoch": 2.882327016306743, "grad_norm": 0.3493222702605276, "learning_rate": 9.359650842503565e-07, "loss": 1.737, "step": 6540 }, { "epoch": 2.884530630233583, "grad_norm": 0.3038129813169421, "learning_rate": 9.012907020798156e-07, "loss": 1.6078, "step": 6545 }, { "epoch": 2.886734244160423, "grad_norm": 0.3031344362125413, "learning_rate": 8.672678785809796e-07, "loss": 1.6788, "step": 6550 }, { "epoch": 2.888937858087263, "grad_norm": 0.31295231167033, "learning_rate": 8.338968374486555e-07, "loss": 1.734, "step": 6555 }, { "epoch": 2.891141472014103, "grad_norm": 0.30464503391760195, "learning_rate": 8.011777980922564e-07, "loss": 1.6216, "step": 6560 }, { "epoch": 2.8933450859409433, "grad_norm": 0.31057203787361704, "learning_rate": 7.691109756344128e-07, "loss": 1.6683, "step": 6565 }, { "epoch": 2.895548699867783, "grad_norm": 0.3406431085406218, "learning_rate": 7.376965809095193e-07, "loss": 1.7457, "step": 6570 }, { "epoch": 2.8977523137946233, "grad_norm": 0.26328954802317356, "learning_rate": 7.06934820462346e-07, "loss": 1.6027, "step": 6575 }, { "epoch": 2.899955927721463, "grad_norm": 0.3427927943612674, "learning_rate": 6.768258965467289e-07, "loss": 1.7368, "step": 6580 }, { "epoch": 2.9021595416483033, "grad_norm": 0.3099689467854988, "learning_rate": 6.473700071241484e-07, "loss": 1.7899, "step": 6585 }, { "epoch": 2.904363155575143, "grad_norm": 0.38502367800150844, "learning_rate": 6.185673458625418e-07, "loss": 1.732, "step": 6590 }, { "epoch": 2.9065667695019832, "grad_norm": 0.3501339480817073, "learning_rate": 5.904181021349375e-07, "loss": 1.6615, "step": 6595 }, { "epoch": 2.9087703834288234, "grad_norm": 0.28836814250088966, "learning_rate": 5.629224610182671e-07, "loss": 1.5576, "step": 6600 }, { "epoch": 2.910973997355663, "grad_norm": 0.2999170211226527, "learning_rate": 5.360806032920995e-07, "loss": 1.6333, "step": 6605 }, { "epoch": 2.9131776112825034, "grad_norm": 0.3369079119860257, "learning_rate": 5.09892705437498e-07, "loss": 1.4815, "step": 6610 }, { "epoch": 2.915381225209343, "grad_norm": 0.30167843329688654, "learning_rate": 4.843589396358427e-07, "loss": 1.7719, "step": 6615 }, { "epoch": 2.9175848391361834, "grad_norm": 0.3352146963211379, "learning_rate": 4.5947947376767663e-07, "loss": 1.8039, "step": 6620 }, { "epoch": 2.919788453063023, "grad_norm": 0.31153697039563194, "learning_rate": 4.3525447141165023e-07, "loss": 1.4281, "step": 6625 }, { "epoch": 2.9219920669898634, "grad_norm": 0.3021129933679619, "learning_rate": 4.116840918434006e-07, "loss": 1.7845, "step": 6630 }, { "epoch": 2.9241956809167036, "grad_norm": 0.3038448101025625, "learning_rate": 3.887684900345301e-07, "loss": 1.6785, "step": 6635 }, { "epoch": 2.9263992948435433, "grad_norm": 0.29571689111577726, "learning_rate": 3.665078166515623e-07, "loss": 1.5903, "step": 6640 }, { "epoch": 2.9286029087703835, "grad_norm": 0.3420405614774203, "learning_rate": 3.449022180549766e-07, "loss": 1.7721, "step": 6645 }, { "epoch": 2.9308065226972233, "grad_norm": 0.28585767479477975, "learning_rate": 3.2395183629824186e-07, "loss": 1.6843, "step": 6650 }, { "epoch": 2.9330101366240635, "grad_norm": 0.2974354225013896, "learning_rate": 3.0365680912688434e-07, "loss": 1.557, "step": 6655 }, { "epoch": 2.9352137505509033, "grad_norm": 0.3294613096996469, "learning_rate": 2.840172699775656e-07, "loss": 1.5854, "step": 6660 }, { "epoch": 2.9374173644777435, "grad_norm": 0.2754088803888805, "learning_rate": 2.650333479771949e-07, "loss": 1.5721, "step": 6665 }, { "epoch": 2.9396209784045837, "grad_norm": 0.2983838481391387, "learning_rate": 2.467051679421406e-07, "loss": 1.6993, "step": 6670 }, { "epoch": 2.9418245923314235, "grad_norm": 0.31155141036980233, "learning_rate": 2.290328503773309e-07, "loss": 1.687, "step": 6675 }, { "epoch": 2.9440282062582637, "grad_norm": 0.29451801544591716, "learning_rate": 2.1201651147554347e-07, "loss": 1.7157, "step": 6680 }, { "epoch": 2.9462318201851034, "grad_norm": 0.3316660290657273, "learning_rate": 1.956562631165504e-07, "loss": 1.6958, "step": 6685 }, { "epoch": 2.9484354341119436, "grad_norm": 0.37499704382786475, "learning_rate": 1.7995221286645215e-07, "loss": 1.7061, "step": 6690 }, { "epoch": 2.9506390480387834, "grad_norm": 0.21043643461218126, "learning_rate": 1.6490446397696702e-07, "loss": 1.3901, "step": 6695 }, { "epoch": 2.9528426619656236, "grad_norm": 0.3260933032444566, "learning_rate": 1.5051311538469837e-07, "loss": 1.6567, "step": 6700 }, { "epoch": 2.955046275892464, "grad_norm": 0.3164050584909926, "learning_rate": 1.367782617105351e-07, "loss": 1.6661, "step": 6705 }, { "epoch": 2.9572498898193036, "grad_norm": 0.3495591454896855, "learning_rate": 1.2369999325901881e-07, "loss": 1.6197, "step": 6710 }, { "epoch": 2.959453503746144, "grad_norm": 0.3351019168716464, "learning_rate": 1.1127839601774437e-07, "loss": 1.6162, "step": 6715 }, { "epoch": 2.9616571176729836, "grad_norm": 0.34641497402663024, "learning_rate": 9.951355165678244e-08, "loss": 1.7908, "step": 6720 }, { "epoch": 2.9638607315998238, "grad_norm": 0.37716614486780947, "learning_rate": 8.840553752815783e-08, "loss": 1.6302, "step": 6725 }, { "epoch": 2.9660643455266635, "grad_norm": 0.3039829114023131, "learning_rate": 7.79544266653609e-08, "loss": 1.7006, "step": 6730 }, { "epoch": 2.9682679594535037, "grad_norm": 0.31609676417425303, "learning_rate": 6.816028778281469e-08, "loss": 1.6702, "step": 6735 }, { "epoch": 2.970471573380344, "grad_norm": 0.37777667052819425, "learning_rate": 5.902318527547523e-08, "loss": 1.4444, "step": 6740 }, { "epoch": 2.9726751873071837, "grad_norm": 0.36388069806115425, "learning_rate": 5.0543179218365265e-08, "loss": 1.5438, "step": 6745 }, { "epoch": 2.974878801234024, "grad_norm": 0.362171321908763, "learning_rate": 4.272032536621895e-08, "loss": 1.7638, "step": 6750 }, { "epoch": 2.9770824151608637, "grad_norm": 0.3560152063618098, "learning_rate": 3.5554675153082195e-08, "loss": 1.6422, "step": 6755 }, { "epoch": 2.979286029087704, "grad_norm": 0.2934857454303967, "learning_rate": 2.9046275692012904e-08, "loss": 1.5529, "step": 6760 }, { "epoch": 2.9814896430145437, "grad_norm": 0.31059161987377404, "learning_rate": 2.3195169774714586e-08, "loss": 1.5975, "step": 6765 }, { "epoch": 2.983693256941384, "grad_norm": 0.4364323493515609, "learning_rate": 1.8001395871303228e-08, "loss": 1.7395, "step": 6770 }, { "epoch": 2.985896870868224, "grad_norm": 0.3199314191111005, "learning_rate": 1.3464988130051925e-08, "loss": 1.5351, "step": 6775 }, { "epoch": 2.988100484795064, "grad_norm": 0.2946914049828253, "learning_rate": 9.585976377124439e-09, "loss": 1.519, "step": 6780 }, { "epoch": 2.990304098721904, "grad_norm": 0.2758574818813751, "learning_rate": 6.364386116419762e-09, "loss": 1.4758, "step": 6785 }, { "epoch": 2.992507712648744, "grad_norm": 0.329124521705272, "learning_rate": 3.800238529416688e-09, "loss": 1.8557, "step": 6790 }, { "epoch": 2.994711326575584, "grad_norm": 0.3394680059091016, "learning_rate": 1.8935504749628684e-09, "loss": 1.7226, "step": 6795 }, { "epoch": 2.996914940502424, "grad_norm": 0.27640804395611795, "learning_rate": 6.443344892637093e-10, "loss": 1.6186, "step": 6800 }, { "epoch": 2.999118554429264, "grad_norm": 0.3138822790633703, "learning_rate": 5.259878569363608e-11, "loss": 1.6958, "step": 6805 }, { "epoch": 3.0, "step": 6807, "total_flos": 1.620091928969216e+16, "train_loss": 1.72514724640379, "train_runtime": 38823.0414, "train_samples_per_second": 0.701, "train_steps_per_second": 0.175 } ], "logging_steps": 5, "max_steps": 6807, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.620091928969216e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }