{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.999537251272559, "eval_steps": 100, "global_step": 900, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.005552984729291994, "grad_norm": 5.378993948282789, "learning_rate": 1.111111111111111e-06, "loss": 1.3618, "step": 5 }, { "epoch": 0.011105969458583989, "grad_norm": 4.007932344194533, "learning_rate": 2.222222222222222e-06, "loss": 1.3497, "step": 10 }, { "epoch": 0.016658954187875982, "grad_norm": 2.870467768621527, "learning_rate": 3.3333333333333333e-06, "loss": 1.3292, "step": 15 }, { "epoch": 0.022211938917167977, "grad_norm": 2.9129656624638147, "learning_rate": 4.444444444444444e-06, "loss": 1.3086, "step": 20 }, { "epoch": 0.027764923646459973, "grad_norm": 2.0216771653661767, "learning_rate": 5.555555555555557e-06, "loss": 1.1997, "step": 25 }, { "epoch": 0.033317908375751965, "grad_norm": 1.7768511495820114, "learning_rate": 6.666666666666667e-06, "loss": 1.1742, "step": 30 }, { "epoch": 0.03887089310504396, "grad_norm": 1.4804724166332308, "learning_rate": 7.77777777777778e-06, "loss": 1.1244, "step": 35 }, { "epoch": 0.044423877834335955, "grad_norm": 1.5180593968939422, "learning_rate": 8.888888888888888e-06, "loss": 1.0946, "step": 40 }, { "epoch": 0.04997686256362795, "grad_norm": 1.4879318494570208, "learning_rate": 1e-05, "loss": 1.1032, "step": 45 }, { "epoch": 0.055529847292919945, "grad_norm": 1.449589161585785, "learning_rate": 1.1111111111111113e-05, "loss": 1.0751, "step": 50 }, { "epoch": 0.06108283202221194, "grad_norm": 1.3300428861282843, "learning_rate": 1.2222222222222224e-05, "loss": 1.0704, "step": 55 }, { "epoch": 0.06663581675150393, "grad_norm": 1.5071782473179958, "learning_rate": 1.3333333333333333e-05, "loss": 1.0561, "step": 60 }, { "epoch": 0.07218880148079593, "grad_norm": 1.3957616523065333, "learning_rate": 1.4444444444444446e-05, "loss": 1.0527, "step": 65 }, { "epoch": 0.07774178621008793, "grad_norm": 1.3254556004203786, "learning_rate": 1.555555555555556e-05, "loss": 1.0107, "step": 70 }, { "epoch": 0.08329477093937991, "grad_norm": 1.5821507688192253, "learning_rate": 1.6666666666666667e-05, "loss": 1.035, "step": 75 }, { "epoch": 0.08884775566867191, "grad_norm": 1.5770964291038414, "learning_rate": 1.7777777777777777e-05, "loss": 1.0285, "step": 80 }, { "epoch": 0.09440074039796391, "grad_norm": 1.5156588992406825, "learning_rate": 1.888888888888889e-05, "loss": 1.0462, "step": 85 }, { "epoch": 0.0999537251272559, "grad_norm": 1.8368814946604881, "learning_rate": 2e-05, "loss": 1.0321, "step": 90 }, { "epoch": 0.10550670985654789, "grad_norm": 1.6154848357095026, "learning_rate": 1.9998119704485016e-05, "loss": 0.9962, "step": 95 }, { "epoch": 0.11105969458583989, "grad_norm": 1.6218644276479086, "learning_rate": 1.9992479525042305e-05, "loss": 1.0216, "step": 100 }, { "epoch": 0.11105969458583989, "eval_loss": 1.0477724075317383, "eval_runtime": 14.0681, "eval_samples_per_second": 18.197, "eval_steps_per_second": 4.549, "step": 100 }, { "epoch": 0.11661267931513189, "grad_norm": 1.4399255902202484, "learning_rate": 1.9983081582712684e-05, "loss": 1.0344, "step": 105 }, { "epoch": 0.12216566404442387, "grad_norm": 1.8069324642784876, "learning_rate": 1.996992941167792e-05, "loss": 1.0204, "step": 110 }, { "epoch": 0.12771864877371586, "grad_norm": 2.1003495945173865, "learning_rate": 1.9953027957931658e-05, "loss": 1.0169, "step": 115 }, { "epoch": 0.13327163350300786, "grad_norm": 1.9735058045205416, "learning_rate": 1.9932383577419432e-05, "loss": 1.0039, "step": 120 }, { "epoch": 0.13882461823229986, "grad_norm": 1.4822914643295109, "learning_rate": 1.9908004033648452e-05, "loss": 1.0025, "step": 125 }, { "epoch": 0.14437760296159186, "grad_norm": 1.5730749273183444, "learning_rate": 1.9879898494768093e-05, "loss": 1.0014, "step": 130 }, { "epoch": 0.14993058769088385, "grad_norm": 1.4317675236519043, "learning_rate": 1.9848077530122083e-05, "loss": 1.0055, "step": 135 }, { "epoch": 0.15548357242017585, "grad_norm": 1.4305310143482486, "learning_rate": 1.9812553106273848e-05, "loss": 0.9772, "step": 140 }, { "epoch": 0.16103655714946785, "grad_norm": 1.835206300273569, "learning_rate": 1.9773338582506357e-05, "loss": 1.0041, "step": 145 }, { "epoch": 0.16658954187875982, "grad_norm": 1.346307644251197, "learning_rate": 1.973044870579824e-05, "loss": 0.9836, "step": 150 }, { "epoch": 0.17214252660805182, "grad_norm": 1.3087431013115765, "learning_rate": 1.9683899605278062e-05, "loss": 0.9906, "step": 155 }, { "epoch": 0.17769551133734382, "grad_norm": 1.6304479084212757, "learning_rate": 1.9633708786158803e-05, "loss": 0.9846, "step": 160 }, { "epoch": 0.18324849606663582, "grad_norm": 1.4079238364588627, "learning_rate": 1.957989512315489e-05, "loss": 0.9953, "step": 165 }, { "epoch": 0.18880148079592782, "grad_norm": 1.4160494597971853, "learning_rate": 1.9522478853384154e-05, "loss": 0.9728, "step": 170 }, { "epoch": 0.19435446552521982, "grad_norm": 1.3831279332888735, "learning_rate": 1.946148156875751e-05, "loss": 0.9786, "step": 175 }, { "epoch": 0.1999074502545118, "grad_norm": 1.3223005814385944, "learning_rate": 1.9396926207859085e-05, "loss": 0.9915, "step": 180 }, { "epoch": 0.20546043498380379, "grad_norm": 7.733682148553893, "learning_rate": 1.932883704732001e-05, "loss": 1.0145, "step": 185 }, { "epoch": 0.21101341971309578, "grad_norm": 1.6669211194032283, "learning_rate": 1.9257239692688907e-05, "loss": 0.9862, "step": 190 }, { "epoch": 0.21656640444238778, "grad_norm": 1.4461392109939817, "learning_rate": 1.9182161068802742e-05, "loss": 0.9944, "step": 195 }, { "epoch": 0.22211938917167978, "grad_norm": 1.3816905654957743, "learning_rate": 1.9103629409661468e-05, "loss": 0.99, "step": 200 }, { "epoch": 0.22211938917167978, "eval_loss": 1.0106741189956665, "eval_runtime": 13.0423, "eval_samples_per_second": 19.628, "eval_steps_per_second": 4.907, "step": 200 }, { "epoch": 0.22767237390097178, "grad_norm": 1.3296649623180647, "learning_rate": 1.902167424781038e-05, "loss": 0.9874, "step": 205 }, { "epoch": 0.23322535863026378, "grad_norm": 1.4147575392387848, "learning_rate": 1.8936326403234125e-05, "loss": 0.9906, "step": 210 }, { "epoch": 0.23877834335955575, "grad_norm": 1.5048777388561694, "learning_rate": 1.8847617971766577e-05, "loss": 0.9721, "step": 215 }, { "epoch": 0.24433132808884775, "grad_norm": 1.3281954721580007, "learning_rate": 1.8755582313020912e-05, "loss": 0.9848, "step": 220 }, { "epoch": 0.24988431281813975, "grad_norm": 1.3865529123681517, "learning_rate": 1.866025403784439e-05, "loss": 0.9885, "step": 225 }, { "epoch": 0.2554372975474317, "grad_norm": 1.468591497661801, "learning_rate": 1.8561668995302668e-05, "loss": 0.9713, "step": 230 }, { "epoch": 0.2609902822767237, "grad_norm": 1.3686298836564843, "learning_rate": 1.845986425919841e-05, "loss": 0.9579, "step": 235 }, { "epoch": 0.2665432670060157, "grad_norm": 1.4148255169077197, "learning_rate": 1.8354878114129368e-05, "loss": 0.9506, "step": 240 }, { "epoch": 0.2720962517353077, "grad_norm": 1.5172086489276786, "learning_rate": 1.824675004109107e-05, "loss": 0.99, "step": 245 }, { "epoch": 0.2776492364645997, "grad_norm": 1.3436284741737878, "learning_rate": 1.8135520702629677e-05, "loss": 0.9654, "step": 250 }, { "epoch": 0.2832022211938917, "grad_norm": 1.4414606449378646, "learning_rate": 1.802123192755044e-05, "loss": 0.9668, "step": 255 }, { "epoch": 0.2887552059231837, "grad_norm": 1.4031899462530004, "learning_rate": 1.7903926695187595e-05, "loss": 0.9626, "step": 260 }, { "epoch": 0.2943081906524757, "grad_norm": 1.4639958980701138, "learning_rate": 1.7783649119241603e-05, "loss": 0.9459, "step": 265 }, { "epoch": 0.2998611753817677, "grad_norm": 1.426651386589118, "learning_rate": 1.766044443118978e-05, "loss": 0.9863, "step": 270 }, { "epoch": 0.3054141601110597, "grad_norm": 1.4219614478552796, "learning_rate": 1.7534358963276606e-05, "loss": 0.9719, "step": 275 }, { "epoch": 0.3109671448403517, "grad_norm": 1.3151103765065284, "learning_rate": 1.740544013109005e-05, "loss": 0.9903, "step": 280 }, { "epoch": 0.3165201295696437, "grad_norm": 1.3354926740055781, "learning_rate": 1.7273736415730488e-05, "loss": 0.9681, "step": 285 }, { "epoch": 0.3220731142989357, "grad_norm": 1.3717221395296357, "learning_rate": 1.7139297345578992e-05, "loss": 0.9456, "step": 290 }, { "epoch": 0.32762609902822765, "grad_norm": 1.6411114205659896, "learning_rate": 1.7002173477671685e-05, "loss": 0.9591, "step": 295 }, { "epoch": 0.33317908375751965, "grad_norm": 1.2826425067775913, "learning_rate": 1.686241637868734e-05, "loss": 0.9328, "step": 300 }, { "epoch": 0.33317908375751965, "eval_loss": 0.99033522605896, "eval_runtime": 13.2544, "eval_samples_per_second": 19.314, "eval_steps_per_second": 4.829, "step": 300 }, { "epoch": 0.33873206848681164, "grad_norm": 1.4005676511844571, "learning_rate": 1.6720078605555227e-05, "loss": 0.9803, "step": 305 }, { "epoch": 0.34428505321610364, "grad_norm": 1.3022121862564202, "learning_rate": 1.657521368569064e-05, "loss": 0.9622, "step": 310 }, { "epoch": 0.34983803794539564, "grad_norm": 1.4147938551822972, "learning_rate": 1.6427876096865394e-05, "loss": 0.9733, "step": 315 }, { "epoch": 0.35539102267468764, "grad_norm": 1.4563599805627543, "learning_rate": 1.627812124672099e-05, "loss": 0.9695, "step": 320 }, { "epoch": 0.36094400740397964, "grad_norm": 1.2666931064627047, "learning_rate": 1.6126005451932028e-05, "loss": 0.9512, "step": 325 }, { "epoch": 0.36649699213327164, "grad_norm": 1.3459612867606927, "learning_rate": 1.5971585917027864e-05, "loss": 0.9962, "step": 330 }, { "epoch": 0.37204997686256364, "grad_norm": 1.4053767425847852, "learning_rate": 1.5814920712880267e-05, "loss": 0.9456, "step": 335 }, { "epoch": 0.37760296159185563, "grad_norm": 1.2542630494816203, "learning_rate": 1.5656068754865388e-05, "loss": 0.9625, "step": 340 }, { "epoch": 0.38315594632114763, "grad_norm": 1.2467268401516984, "learning_rate": 1.5495089780708062e-05, "loss": 0.9416, "step": 345 }, { "epoch": 0.38870893105043963, "grad_norm": 1.3642420280912566, "learning_rate": 1.5332044328016916e-05, "loss": 0.9745, "step": 350 }, { "epoch": 0.39426191577973163, "grad_norm": 1.370419991211419, "learning_rate": 1.5166993711518631e-05, "loss": 0.9235, "step": 355 }, { "epoch": 0.3998149005090236, "grad_norm": 1.4175545037228292, "learning_rate": 1.5000000000000002e-05, "loss": 0.9419, "step": 360 }, { "epoch": 0.4053678852383156, "grad_norm": 1.5230125115585145, "learning_rate": 1.4831125992966386e-05, "loss": 0.9482, "step": 365 }, { "epoch": 0.41092086996760757, "grad_norm": 1.3243119971736406, "learning_rate": 1.4660435197025391e-05, "loss": 0.9516, "step": 370 }, { "epoch": 0.41647385469689957, "grad_norm": 1.3168758353005081, "learning_rate": 1.4487991802004625e-05, "loss": 0.9274, "step": 375 }, { "epoch": 0.42202683942619157, "grad_norm": 1.3316158417964403, "learning_rate": 1.4313860656812537e-05, "loss": 0.9279, "step": 380 }, { "epoch": 0.42757982415548357, "grad_norm": 1.2112610899992784, "learning_rate": 1.4138107245051394e-05, "loss": 0.9604, "step": 385 }, { "epoch": 0.43313280888477557, "grad_norm": 1.4256068345744652, "learning_rate": 1.396079766039157e-05, "loss": 0.9504, "step": 390 }, { "epoch": 0.43868579361406757, "grad_norm": 1.48676297842706, "learning_rate": 1.3781998581716427e-05, "loss": 0.9783, "step": 395 }, { "epoch": 0.44423877834335956, "grad_norm": 1.3917645840392392, "learning_rate": 1.3601777248047105e-05, "loss": 0.9428, "step": 400 }, { "epoch": 0.44423877834335956, "eval_loss": 0.975586473941803, "eval_runtime": 13.0584, "eval_samples_per_second": 19.604, "eval_steps_per_second": 4.901, "step": 400 }, { "epoch": 0.44979176307265156, "grad_norm": 1.3230682794396744, "learning_rate": 1.342020143325669e-05, "loss": 0.9477, "step": 405 }, { "epoch": 0.45534474780194356, "grad_norm": 1.2045267312184351, "learning_rate": 1.3237339420583213e-05, "loss": 0.9568, "step": 410 }, { "epoch": 0.46089773253123556, "grad_norm": 1.2839147189555775, "learning_rate": 1.3053259976951134e-05, "loss": 0.9256, "step": 415 }, { "epoch": 0.46645071726052756, "grad_norm": 1.284551507215065, "learning_rate": 1.2868032327110904e-05, "loss": 0.9246, "step": 420 }, { "epoch": 0.4720037019898195, "grad_norm": 1.3021372211969566, "learning_rate": 1.2681726127606374e-05, "loss": 0.9527, "step": 425 }, { "epoch": 0.4775566867191115, "grad_norm": 1.3264954697360052, "learning_rate": 1.2494411440579814e-05, "loss": 0.9659, "step": 430 }, { "epoch": 0.4831096714484035, "grad_norm": 1.2931203310254533, "learning_rate": 1.2306158707424402e-05, "loss": 0.9044, "step": 435 }, { "epoch": 0.4886626561776955, "grad_norm": 1.5033726081614123, "learning_rate": 1.211703872229411e-05, "loss": 0.9135, "step": 440 }, { "epoch": 0.4942156409069875, "grad_norm": 1.2465062347279259, "learning_rate": 1.1927122605480899e-05, "loss": 0.9382, "step": 445 }, { "epoch": 0.4997686256362795, "grad_norm": 1.2206738144949079, "learning_rate": 1.1736481776669307e-05, "loss": 0.9429, "step": 450 }, { "epoch": 0.5053216103655715, "grad_norm": 1.268246804784844, "learning_rate": 1.1545187928078407e-05, "loss": 0.9163, "step": 455 }, { "epoch": 0.5108745950948634, "grad_norm": 1.1842171344447763, "learning_rate": 1.1353312997501313e-05, "loss": 0.9296, "step": 460 }, { "epoch": 0.5164275798241554, "grad_norm": 1.3329158977326516, "learning_rate": 1.1160929141252303e-05, "loss": 0.9536, "step": 465 }, { "epoch": 0.5219805645534474, "grad_norm": 1.3140264320853574, "learning_rate": 1.0968108707031792e-05, "loss": 0.9314, "step": 470 }, { "epoch": 0.5275335492827394, "grad_norm": 1.2112452453436886, "learning_rate": 1.077492420671931e-05, "loss": 0.8858, "step": 475 }, { "epoch": 0.5330865340120314, "grad_norm": 1.244066114064373, "learning_rate": 1.0581448289104759e-05, "loss": 0.938, "step": 480 }, { "epoch": 0.5386395187413234, "grad_norm": 1.2958379438943313, "learning_rate": 1.038775371256817e-05, "loss": 0.9366, "step": 485 }, { "epoch": 0.5441925034706154, "grad_norm": 1.2706792533588414, "learning_rate": 1.0193913317718245e-05, "loss": 0.9809, "step": 490 }, { "epoch": 0.5497454881999074, "grad_norm": 1.2373793128051882, "learning_rate": 1e-05, "loss": 0.9107, "step": 495 }, { "epoch": 0.5552984729291994, "grad_norm": 1.3259462807230906, "learning_rate": 9.806086682281759e-06, "loss": 0.9478, "step": 500 }, { "epoch": 0.5552984729291994, "eval_loss": 0.9645185470581055, "eval_runtime": 14.0973, "eval_samples_per_second": 18.16, "eval_steps_per_second": 4.54, "step": 500 }, { "epoch": 0.5608514576584914, "grad_norm": 1.2818749562353056, "learning_rate": 9.612246287431832e-06, "loss": 0.9157, "step": 505 }, { "epoch": 0.5664044423877834, "grad_norm": 1.2789066996692675, "learning_rate": 9.418551710895243e-06, "loss": 0.9715, "step": 510 }, { "epoch": 0.5719574271170754, "grad_norm": 1.3159690301384173, "learning_rate": 9.225075793280693e-06, "loss": 0.9564, "step": 515 }, { "epoch": 0.5775104118463674, "grad_norm": 1.1820999233185365, "learning_rate": 9.03189129296821e-06, "loss": 0.9333, "step": 520 }, { "epoch": 0.5830633965756594, "grad_norm": 2.9705367947063595, "learning_rate": 8.839070858747697e-06, "loss": 0.9393, "step": 525 }, { "epoch": 0.5886163813049514, "grad_norm": 1.1578484132538005, "learning_rate": 8.646687002498692e-06, "loss": 0.9386, "step": 530 }, { "epoch": 0.5941693660342434, "grad_norm": 1.2495396181990488, "learning_rate": 8.454812071921597e-06, "loss": 0.9202, "step": 535 }, { "epoch": 0.5997223507635354, "grad_norm": 1.18243382811664, "learning_rate": 8.263518223330698e-06, "loss": 0.9222, "step": 540 }, { "epoch": 0.6052753354928274, "grad_norm": 1.2247856999180733, "learning_rate": 8.072877394519103e-06, "loss": 0.9426, "step": 545 }, { "epoch": 0.6108283202221194, "grad_norm": 1.220145252189745, "learning_rate": 7.882961277705897e-06, "loss": 0.9161, "step": 550 }, { "epoch": 0.6163813049514114, "grad_norm": 1.2450054960684753, "learning_rate": 7.6938412925756e-06, "loss": 0.9419, "step": 555 }, { "epoch": 0.6219342896807034, "grad_norm": 1.2450637061923184, "learning_rate": 7.505588559420188e-06, "loss": 0.9471, "step": 560 }, { "epoch": 0.6274872744099954, "grad_norm": 1.2160705969336116, "learning_rate": 7.3182738723936255e-06, "loss": 0.9446, "step": 565 }, { "epoch": 0.6330402591392874, "grad_norm": 1.2857156644799776, "learning_rate": 7.131967672889101e-06, "loss": 0.9473, "step": 570 }, { "epoch": 0.6385932438685794, "grad_norm": 1.1688923012156514, "learning_rate": 6.94674002304887e-06, "loss": 0.9193, "step": 575 }, { "epoch": 0.6441462285978714, "grad_norm": 1.3240345620915759, "learning_rate": 6.762660579416791e-06, "loss": 0.955, "step": 580 }, { "epoch": 0.6496992133271634, "grad_norm": 1.171487612102084, "learning_rate": 6.579798566743314e-06, "loss": 0.9345, "step": 585 }, { "epoch": 0.6552521980564553, "grad_norm": 1.1809918233177483, "learning_rate": 6.3982227519528986e-06, "loss": 0.9317, "step": 590 }, { "epoch": 0.6608051827857473, "grad_norm": 1.1849135550680583, "learning_rate": 6.218001418283577e-06, "loss": 0.9282, "step": 595 }, { "epoch": 0.6663581675150393, "grad_norm": 1.2353428612054926, "learning_rate": 6.039202339608432e-06, "loss": 0.9186, "step": 600 }, { "epoch": 0.6663581675150393, "eval_loss": 0.9549762010574341, "eval_runtime": 13.1058, "eval_samples_per_second": 19.533, "eval_steps_per_second": 4.883, "step": 600 }, { "epoch": 0.6719111522443313, "grad_norm": 1.2260206730700363, "learning_rate": 5.8618927549486095e-06, "loss": 0.91, "step": 605 }, { "epoch": 0.6774641369736233, "grad_norm": 1.2511346515975978, "learning_rate": 5.686139343187468e-06, "loss": 0.9445, "step": 610 }, { "epoch": 0.6830171217029153, "grad_norm": 1.1415920400151276, "learning_rate": 5.512008197995379e-06, "loss": 0.9267, "step": 615 }, { "epoch": 0.6885701064322073, "grad_norm": 1.1718028560057823, "learning_rate": 5.339564802974615e-06, "loss": 0.9173, "step": 620 }, { "epoch": 0.6941230911614993, "grad_norm": 1.1427826499364968, "learning_rate": 5.168874007033615e-06, "loss": 0.9113, "step": 625 }, { "epoch": 0.6996760758907913, "grad_norm": 1.1646711162888848, "learning_rate": 5.000000000000003e-06, "loss": 0.9314, "step": 630 }, { "epoch": 0.7052290606200833, "grad_norm": 1.1666389482954498, "learning_rate": 4.8330062884813714e-06, "loss": 0.949, "step": 635 }, { "epoch": 0.7107820453493753, "grad_norm": 1.2669383225264854, "learning_rate": 4.66795567198309e-06, "loss": 0.9298, "step": 640 }, { "epoch": 0.7163350300786673, "grad_norm": 1.1727436218210467, "learning_rate": 4.504910219291941e-06, "loss": 0.9384, "step": 645 }, { "epoch": 0.7218880148079593, "grad_norm": 1.160141429869342, "learning_rate": 4.343931245134616e-06, "loss": 0.9231, "step": 650 }, { "epoch": 0.7274409995372513, "grad_norm": 1.1571744280380951, "learning_rate": 4.185079287119733e-06, "loss": 0.9379, "step": 655 }, { "epoch": 0.7329939842665433, "grad_norm": 1.2080957370089618, "learning_rate": 4.028414082972141e-06, "loss": 0.9087, "step": 660 }, { "epoch": 0.7385469689958353, "grad_norm": 1.2262375492125892, "learning_rate": 3.873994548067972e-06, "loss": 0.9175, "step": 665 }, { "epoch": 0.7440999537251273, "grad_norm": 1.1868816861234752, "learning_rate": 3.7218787532790167e-06, "loss": 0.915, "step": 670 }, { "epoch": 0.7496529384544193, "grad_norm": 1.2149836180874647, "learning_rate": 3.5721239031346067e-06, "loss": 0.9359, "step": 675 }, { "epoch": 0.7552059231837113, "grad_norm": 1.1401883150941166, "learning_rate": 3.424786314309365e-06, "loss": 0.8996, "step": 680 }, { "epoch": 0.7607589079130033, "grad_norm": 1.1918737445166034, "learning_rate": 3.279921394444776e-06, "loss": 0.9478, "step": 685 }, { "epoch": 0.7663118926422953, "grad_norm": 1.1601543056853199, "learning_rate": 3.1375836213126653e-06, "loss": 0.9144, "step": 690 }, { "epoch": 0.7718648773715873, "grad_norm": 1.1849230333131153, "learning_rate": 2.9978265223283152e-06, "loss": 0.9134, "step": 695 }, { "epoch": 0.7774178621008793, "grad_norm": 1.1922764155084293, "learning_rate": 2.8607026544210115e-06, "loss": 0.9184, "step": 700 }, { "epoch": 0.7774178621008793, "eval_loss": 0.9487817287445068, "eval_runtime": 12.9966, "eval_samples_per_second": 19.697, "eval_steps_per_second": 4.924, "step": 700 }, { "epoch": 0.7829708468301713, "grad_norm": 1.0668112491118575, "learning_rate": 2.726263584269513e-06, "loss": 0.9038, "step": 705 }, { "epoch": 0.7885238315594633, "grad_norm": 1.191610088893989, "learning_rate": 2.594559868909956e-06, "loss": 0.9003, "step": 710 }, { "epoch": 0.7940768162887553, "grad_norm": 1.2465046516243472, "learning_rate": 2.4656410367233928e-06, "loss": 0.9215, "step": 715 }, { "epoch": 0.7996298010180471, "grad_norm": 1.1463175579772371, "learning_rate": 2.339555568810221e-06, "loss": 0.9303, "step": 720 }, { "epoch": 0.8051827857473391, "grad_norm": 1.1279074050372055, "learning_rate": 2.2163508807584e-06, "loss": 0.9294, "step": 725 }, { "epoch": 0.8107357704766311, "grad_norm": 1.1668600336379225, "learning_rate": 2.0960733048124082e-06, "loss": 0.9082, "step": 730 }, { "epoch": 0.8162887552059231, "grad_norm": 1.1280124295582716, "learning_rate": 1.9787680724495617e-06, "loss": 0.898, "step": 735 }, { "epoch": 0.8218417399352151, "grad_norm": 1.2693061690074205, "learning_rate": 1.8644792973703252e-06, "loss": 0.9473, "step": 740 }, { "epoch": 0.8273947246645071, "grad_norm": 1.0935688113313613, "learning_rate": 1.7532499589089324e-06, "loss": 0.9312, "step": 745 }, { "epoch": 0.8329477093937991, "grad_norm": 1.1491375579091732, "learning_rate": 1.6451218858706374e-06, "loss": 0.9295, "step": 750 }, { "epoch": 0.8385006941230911, "grad_norm": 1.1971381545639446, "learning_rate": 1.5401357408015893e-06, "loss": 0.8977, "step": 755 }, { "epoch": 0.8440536788523831, "grad_norm": 1.2118976466189748, "learning_rate": 1.4383310046973365e-06, "loss": 0.9201, "step": 760 }, { "epoch": 0.8496066635816751, "grad_norm": 1.1129578498310642, "learning_rate": 1.339745962155613e-06, "loss": 0.9235, "step": 765 }, { "epoch": 0.8551596483109671, "grad_norm": 1.155034707390975, "learning_rate": 1.2444176869790925e-06, "loss": 0.906, "step": 770 }, { "epoch": 0.8607126330402591, "grad_norm": 1.1211050312494577, "learning_rate": 1.152382028233422e-06, "loss": 0.9027, "step": 775 }, { "epoch": 0.8662656177695511, "grad_norm": 1.161462499770876, "learning_rate": 1.0636735967658785e-06, "loss": 0.902, "step": 780 }, { "epoch": 0.8718186024988431, "grad_norm": 1.2466679714276248, "learning_rate": 9.783257521896228e-07, "loss": 0.9312, "step": 785 }, { "epoch": 0.8773715872281351, "grad_norm": 1.1688203046154668, "learning_rate": 8.963705903385344e-07, "loss": 0.9231, "step": 790 }, { "epoch": 0.8829245719574271, "grad_norm": 1.1446515266652288, "learning_rate": 8.178389311972612e-07, "loss": 0.924, "step": 795 }, { "epoch": 0.8884775566867191, "grad_norm": 1.1971045442791841, "learning_rate": 7.427603073110967e-07, "loss": 0.8564, "step": 800 }, { "epoch": 0.8884775566867191, "eval_loss": 0.9455364346504211, "eval_runtime": 13.3345, "eval_samples_per_second": 19.198, "eval_steps_per_second": 4.8, "step": 800 }, { "epoch": 0.8940305414160111, "grad_norm": 1.158461858290895, "learning_rate": 6.711629526799946e-07, "loss": 0.9399, "step": 805 }, { "epoch": 0.8995835261453031, "grad_norm": 1.1264872944659339, "learning_rate": 6.030737921409169e-07, "loss": 0.8946, "step": 810 }, { "epoch": 0.9051365108745951, "grad_norm": 1.122295252126309, "learning_rate": 5.385184312424973e-07, "loss": 0.9228, "step": 815 }, { "epoch": 0.9106894956038871, "grad_norm": 1.130327163633901, "learning_rate": 4.775211466158469e-07, "loss": 0.9278, "step": 820 }, { "epoch": 0.9162424803331791, "grad_norm": 1.152980775746581, "learning_rate": 4.2010487684511105e-07, "loss": 0.9105, "step": 825 }, { "epoch": 0.9217954650624711, "grad_norm": 1.2630255671771575, "learning_rate": 3.662912138411967e-07, "loss": 0.9103, "step": 830 }, { "epoch": 0.9273484497917631, "grad_norm": 1.1212305446295874, "learning_rate": 3.161003947219421e-07, "loss": 0.8847, "step": 835 }, { "epoch": 0.9329014345210551, "grad_norm": 1.1635748807768116, "learning_rate": 2.6955129420176193e-07, "loss": 0.9139, "step": 840 }, { "epoch": 0.9384544192503471, "grad_norm": 1.101552910147697, "learning_rate": 2.2666141749364434e-07, "loss": 0.9233, "step": 845 }, { "epoch": 0.944007403979639, "grad_norm": 1.1104286718600382, "learning_rate": 1.874468937261531e-07, "loss": 0.8872, "step": 850 }, { "epoch": 0.949560388708931, "grad_norm": 1.1513796874736284, "learning_rate": 1.519224698779198e-07, "loss": 0.938, "step": 855 }, { "epoch": 0.955113373438223, "grad_norm": 1.172996631179348, "learning_rate": 1.201015052319099e-07, "loss": 0.9366, "step": 860 }, { "epoch": 0.960666358167515, "grad_norm": 1.178813337349509, "learning_rate": 9.199596635154684e-08, "loss": 0.9382, "step": 865 }, { "epoch": 0.966219342896807, "grad_norm": 1.123260974533791, "learning_rate": 6.761642258056977e-08, "loss": 0.8969, "step": 870 }, { "epoch": 0.971772327626099, "grad_norm": 1.2440553964560856, "learning_rate": 4.6972042068341714e-08, "loss": 0.9125, "step": 875 }, { "epoch": 0.977325312355391, "grad_norm": 1.100622329056596, "learning_rate": 3.0070588322079765e-08, "loss": 0.9106, "step": 880 }, { "epoch": 0.982878297084683, "grad_norm": 1.224884941200071, "learning_rate": 1.6918417287318245e-08, "loss": 0.9045, "step": 885 }, { "epoch": 0.988431281813975, "grad_norm": 1.1506168883594277, "learning_rate": 7.520474957699586e-09, "loss": 0.9341, "step": 890 }, { "epoch": 0.993984266543267, "grad_norm": 1.3030322318546892, "learning_rate": 1.8802955149865854e-09, "loss": 0.9093, "step": 895 }, { "epoch": 0.999537251272559, "grad_norm": 1.164399511709177, "learning_rate": 0.0, "loss": 0.943, "step": 900 }, { "epoch": 0.999537251272559, "eval_loss": 0.9448966979980469, "eval_runtime": 12.8793, "eval_samples_per_second": 19.877, "eval_steps_per_second": 4.969, "step": 900 }, { "epoch": 0.999537251272559, "step": 900, "total_flos": 37979261239296.0, "train_loss": 0.9636553647783067, "train_runtime": 8704.1469, "train_samples_per_second": 4.965, "train_steps_per_second": 0.103 } ], "logging_steps": 5, "max_steps": 900, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": false, "should_training_stop": false }, "attributes": {} } }, "total_flos": 37979261239296.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }