{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.9676393598144797, "eval_steps": 500, "global_step": 28000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.007027283427908856, "grad_norm": 0.5592142939567566, "learning_rate": 0.0001995548787630315, "loss": 1.0457, "step": 100 }, { "epoch": 0.014054566855817713, "grad_norm": 0.6605722308158875, "learning_rate": 0.00019908633009253836, "loss": 0.8075, "step": 200 }, { "epoch": 0.02108185028372657, "grad_norm": 0.5018069744110107, "learning_rate": 0.00019861778142204523, "loss": 0.7927, "step": 300 }, { "epoch": 0.028109133711635426, "grad_norm": 0.5429800152778625, "learning_rate": 0.00019814923275155208, "loss": 0.7648, "step": 400 }, { "epoch": 0.03513641713954428, "grad_norm": 0.5747570395469666, "learning_rate": 0.00019768068408105893, "loss": 0.7732, "step": 500 }, { "epoch": 0.03513641713954428, "eval_loss": 0.7362164855003357, "eval_runtime": 118.3556, "eval_samples_per_second": 106.873, "eval_steps_per_second": 13.367, "step": 500 }, { "epoch": 0.04216370056745314, "grad_norm": 0.6291443109512329, "learning_rate": 0.00019721213541056578, "loss": 0.7614, "step": 600 }, { "epoch": 0.049190983995361995, "grad_norm": 0.6160371899604797, "learning_rate": 0.00019674358674007263, "loss": 0.7489, "step": 700 }, { "epoch": 0.05621826742327085, "grad_norm": 0.5804228782653809, "learning_rate": 0.00019627503806957949, "loss": 0.7399, "step": 800 }, { "epoch": 0.06324555085117971, "grad_norm": 0.6000151634216309, "learning_rate": 0.00019580648939908634, "loss": 0.7349, "step": 900 }, { "epoch": 0.07027283427908856, "grad_norm": 0.5746223330497742, "learning_rate": 0.00019533794072859319, "loss": 0.7282, "step": 1000 }, { "epoch": 0.07027283427908856, "eval_loss": 0.7145671248435974, "eval_runtime": 118.2812, "eval_samples_per_second": 106.94, "eval_steps_per_second": 13.375, "step": 1000 }, { "epoch": 0.07730011770699742, "grad_norm": 0.6321994066238403, "learning_rate": 0.00019486939205810004, "loss": 0.7241, "step": 1100 }, { "epoch": 0.08432740113490628, "grad_norm": 0.7580232620239258, "learning_rate": 0.00019440084338760689, "loss": 0.7254, "step": 1200 }, { "epoch": 0.09135468456281513, "grad_norm": 0.6181788444519043, "learning_rate": 0.00019393229471711376, "loss": 0.7349, "step": 1300 }, { "epoch": 0.09838196799072399, "grad_norm": 0.6435703635215759, "learning_rate": 0.00019346374604662061, "loss": 0.7082, "step": 1400 }, { "epoch": 0.10540925141863285, "grad_norm": 0.5934786200523376, "learning_rate": 0.00019299519737612746, "loss": 0.7145, "step": 1500 }, { "epoch": 0.10540925141863285, "eval_loss": 0.6975060701370239, "eval_runtime": 118.4288, "eval_samples_per_second": 106.807, "eval_steps_per_second": 13.358, "step": 1500 }, { "epoch": 0.1124365348465417, "grad_norm": 0.4486166536808014, "learning_rate": 0.00019252664870563431, "loss": 0.7226, "step": 1600 }, { "epoch": 0.11946381827445056, "grad_norm": 0.4868922531604767, "learning_rate": 0.00019205810003514116, "loss": 0.6951, "step": 1700 }, { "epoch": 0.12649110170235942, "grad_norm": 0.5752139687538147, "learning_rate": 0.00019158955136464801, "loss": 0.7117, "step": 1800 }, { "epoch": 0.13351838513026826, "grad_norm": 0.5520344972610474, "learning_rate": 0.00019112100269415486, "loss": 0.7111, "step": 1900 }, { "epoch": 0.14054566855817713, "grad_norm": 0.5359894633293152, "learning_rate": 0.00019065245402366172, "loss": 0.7093, "step": 2000 }, { "epoch": 0.14054566855817713, "eval_loss": 0.687991201877594, "eval_runtime": 118.7631, "eval_samples_per_second": 106.506, "eval_steps_per_second": 13.321, "step": 2000 }, { "epoch": 0.14757295198608597, "grad_norm": 0.5415408611297607, "learning_rate": 0.00019018390535316857, "loss": 0.7117, "step": 2100 }, { "epoch": 0.15460023541399484, "grad_norm": 0.6338439583778381, "learning_rate": 0.00018971535668267542, "loss": 0.7253, "step": 2200 }, { "epoch": 0.16162751884190368, "grad_norm": 0.6666418313980103, "learning_rate": 0.00018924680801218227, "loss": 0.7058, "step": 2300 }, { "epoch": 0.16865480226981255, "grad_norm": 0.5344674587249756, "learning_rate": 0.00018877825934168914, "loss": 0.7036, "step": 2400 }, { "epoch": 0.1756820856977214, "grad_norm": 0.5522785186767578, "learning_rate": 0.000188309710671196, "loss": 0.6809, "step": 2500 }, { "epoch": 0.1756820856977214, "eval_loss": 0.679720938205719, "eval_runtime": 118.6143, "eval_samples_per_second": 106.64, "eval_steps_per_second": 13.337, "step": 2500 }, { "epoch": 0.18270936912563027, "grad_norm": 0.7211841344833374, "learning_rate": 0.00018784116200070284, "loss": 0.6972, "step": 2600 }, { "epoch": 0.1897366525535391, "grad_norm": 0.5469601154327393, "learning_rate": 0.00018737261333020967, "loss": 0.6996, "step": 2700 }, { "epoch": 0.19676393598144798, "grad_norm": 0.5906969904899597, "learning_rate": 0.00018690406465971652, "loss": 0.6828, "step": 2800 }, { "epoch": 0.20379121940935682, "grad_norm": 0.703484833240509, "learning_rate": 0.0001864355159892234, "loss": 0.6906, "step": 2900 }, { "epoch": 0.2108185028372657, "grad_norm": 0.5544711947441101, "learning_rate": 0.00018596696731873024, "loss": 0.6891, "step": 3000 }, { "epoch": 0.2108185028372657, "eval_loss": 0.6728695034980774, "eval_runtime": 118.909, "eval_samples_per_second": 106.375, "eval_steps_per_second": 13.304, "step": 3000 }, { "epoch": 0.21784578626517453, "grad_norm": 0.5521793365478516, "learning_rate": 0.0001854984186482371, "loss": 0.6777, "step": 3100 }, { "epoch": 0.2248730696930834, "grad_norm": 0.4634329080581665, "learning_rate": 0.00018502986997774394, "loss": 0.6744, "step": 3200 }, { "epoch": 0.23190035312099225, "grad_norm": 0.7366927266120911, "learning_rate": 0.0001845613213072508, "loss": 0.6642, "step": 3300 }, { "epoch": 0.23892763654890112, "grad_norm": 0.5567039251327515, "learning_rate": 0.00018409277263675767, "loss": 0.6895, "step": 3400 }, { "epoch": 0.24595491997680996, "grad_norm": 0.6940245628356934, "learning_rate": 0.0001836242239662645, "loss": 0.6708, "step": 3500 }, { "epoch": 0.24595491997680996, "eval_loss": 0.6706892848014832, "eval_runtime": 118.4648, "eval_samples_per_second": 106.774, "eval_steps_per_second": 13.354, "step": 3500 }, { "epoch": 0.25298220340471883, "grad_norm": 0.5577242970466614, "learning_rate": 0.00018315567529577135, "loss": 0.679, "step": 3600 }, { "epoch": 0.2600094868326277, "grad_norm": 0.6587842702865601, "learning_rate": 0.0001826871266252782, "loss": 0.6918, "step": 3700 }, { "epoch": 0.2670367702605365, "grad_norm": 0.6200099587440491, "learning_rate": 0.00018221857795478505, "loss": 0.7002, "step": 3800 }, { "epoch": 0.2740640536884454, "grad_norm": 0.5648295879364014, "learning_rate": 0.00018175002928429192, "loss": 0.6791, "step": 3900 }, { "epoch": 0.28109133711635426, "grad_norm": 0.5908897519111633, "learning_rate": 0.00018128148061379877, "loss": 0.6839, "step": 4000 }, { "epoch": 0.28109133711635426, "eval_loss": 0.664506196975708, "eval_runtime": 119.1229, "eval_samples_per_second": 106.184, "eval_steps_per_second": 13.28, "step": 4000 }, { "epoch": 0.2881186205442631, "grad_norm": 0.5778653621673584, "learning_rate": 0.00018081293194330562, "loss": 0.6654, "step": 4100 }, { "epoch": 0.29514590397217194, "grad_norm": 0.700835645198822, "learning_rate": 0.00018034438327281247, "loss": 0.6897, "step": 4200 }, { "epoch": 0.3021731874000808, "grad_norm": 0.5250533819198608, "learning_rate": 0.00017987583460231932, "loss": 0.6934, "step": 4300 }, { "epoch": 0.3092004708279897, "grad_norm": 0.725397527217865, "learning_rate": 0.00017940728593182617, "loss": 0.6781, "step": 4400 }, { "epoch": 0.31622775425589855, "grad_norm": 0.5805392265319824, "learning_rate": 0.00017893873726133302, "loss": 0.6866, "step": 4500 }, { "epoch": 0.31622775425589855, "eval_loss": 0.65822434425354, "eval_runtime": 118.7959, "eval_samples_per_second": 106.477, "eval_steps_per_second": 13.317, "step": 4500 }, { "epoch": 0.32325503768380737, "grad_norm": 0.6944029331207275, "learning_rate": 0.00017847018859083988, "loss": 0.6683, "step": 4600 }, { "epoch": 0.33028232111171624, "grad_norm": 0.5891593098640442, "learning_rate": 0.00017800163992034673, "loss": 0.6817, "step": 4700 }, { "epoch": 0.3373096045396251, "grad_norm": 0.6220216751098633, "learning_rate": 0.00017753309124985358, "loss": 0.6845, "step": 4800 }, { "epoch": 0.344336887967534, "grad_norm": 0.7002474665641785, "learning_rate": 0.00017706454257936045, "loss": 0.6551, "step": 4900 }, { "epoch": 0.3513641713954428, "grad_norm": 0.5766635537147522, "learning_rate": 0.0001765959939088673, "loss": 0.6741, "step": 5000 }, { "epoch": 0.3513641713954428, "eval_loss": 0.6550154685974121, "eval_runtime": 118.5512, "eval_samples_per_second": 106.697, "eval_steps_per_second": 13.344, "step": 5000 }, { "epoch": 0.35839145482335166, "grad_norm": 0.5734873414039612, "learning_rate": 0.00017612744523837415, "loss": 0.6583, "step": 5100 }, { "epoch": 0.36541873825126053, "grad_norm": 0.4953276515007019, "learning_rate": 0.000175658896567881, "loss": 0.6621, "step": 5200 }, { "epoch": 0.37244602167916935, "grad_norm": 0.7837636470794678, "learning_rate": 0.00017519034789738785, "loss": 0.6846, "step": 5300 }, { "epoch": 0.3794733051070782, "grad_norm": 0.718715488910675, "learning_rate": 0.0001747217992268947, "loss": 0.6776, "step": 5400 }, { "epoch": 0.3865005885349871, "grad_norm": 0.5844186544418335, "learning_rate": 0.00017425325055640155, "loss": 0.6627, "step": 5500 }, { "epoch": 0.3865005885349871, "eval_loss": 0.6532958149909973, "eval_runtime": 118.7133, "eval_samples_per_second": 106.551, "eval_steps_per_second": 13.326, "step": 5500 }, { "epoch": 0.39352787196289596, "grad_norm": 0.4426696300506592, "learning_rate": 0.0001737847018859084, "loss": 0.6546, "step": 5600 }, { "epoch": 0.4005551553908048, "grad_norm": 0.5954882502555847, "learning_rate": 0.00017331615321541525, "loss": 0.6448, "step": 5700 }, { "epoch": 0.40758243881871364, "grad_norm": 0.601349413394928, "learning_rate": 0.0001728476045449221, "loss": 0.6637, "step": 5800 }, { "epoch": 0.4146097222466225, "grad_norm": 0.6108406782150269, "learning_rate": 0.00017237905587442898, "loss": 0.6709, "step": 5900 }, { "epoch": 0.4216370056745314, "grad_norm": 0.6442033052444458, "learning_rate": 0.00017191050720393583, "loss": 0.6588, "step": 6000 }, { "epoch": 0.4216370056745314, "eval_loss": 0.6498988270759583, "eval_runtime": 118.8665, "eval_samples_per_second": 106.413, "eval_steps_per_second": 13.309, "step": 6000 }, { "epoch": 0.4286642891024402, "grad_norm": 0.4642776846885681, "learning_rate": 0.00017144195853344268, "loss": 0.6706, "step": 6100 }, { "epoch": 0.43569157253034907, "grad_norm": 0.6703388690948486, "learning_rate": 0.0001709734098629495, "loss": 0.6675, "step": 6200 }, { "epoch": 0.44271885595825794, "grad_norm": 0.6001936793327332, "learning_rate": 0.00017050486119245636, "loss": 0.6545, "step": 6300 }, { "epoch": 0.4497461393861668, "grad_norm": 0.7159720063209534, "learning_rate": 0.00017003631252196323, "loss": 0.6684, "step": 6400 }, { "epoch": 0.4567734228140756, "grad_norm": 0.5892972350120544, "learning_rate": 0.00016956776385147008, "loss": 0.6509, "step": 6500 }, { "epoch": 0.4567734228140756, "eval_loss": 0.6456841230392456, "eval_runtime": 118.8557, "eval_samples_per_second": 106.423, "eval_steps_per_second": 13.31, "step": 6500 }, { "epoch": 0.4638007062419845, "grad_norm": 0.6351083517074585, "learning_rate": 0.00016909921518097693, "loss": 0.634, "step": 6600 }, { "epoch": 0.47082798966989337, "grad_norm": 0.7939039468765259, "learning_rate": 0.00016863066651048378, "loss": 0.6595, "step": 6700 }, { "epoch": 0.47785527309780224, "grad_norm": 0.5831073522567749, "learning_rate": 0.00016816211783999063, "loss": 0.6444, "step": 6800 }, { "epoch": 0.48488255652571105, "grad_norm": 0.6314815282821655, "learning_rate": 0.0001676935691694975, "loss": 0.6622, "step": 6900 }, { "epoch": 0.4919098399536199, "grad_norm": 0.5781182646751404, "learning_rate": 0.00016722502049900433, "loss": 0.6448, "step": 7000 }, { "epoch": 0.4919098399536199, "eval_loss": 0.6440666317939758, "eval_runtime": 118.5556, "eval_samples_per_second": 106.693, "eval_steps_per_second": 13.344, "step": 7000 }, { "epoch": 0.4989371233815288, "grad_norm": 0.7241762280464172, "learning_rate": 0.00016675647182851119, "loss": 0.6727, "step": 7100 }, { "epoch": 0.5059644068094377, "grad_norm": 0.7668181657791138, "learning_rate": 0.00016628792315801804, "loss": 0.6564, "step": 7200 }, { "epoch": 0.5129916902373465, "grad_norm": 0.5727465152740479, "learning_rate": 0.00016581937448752489, "loss": 0.6348, "step": 7300 }, { "epoch": 0.5200189736652554, "grad_norm": 0.6217190027236938, "learning_rate": 0.00016535082581703176, "loss": 0.6434, "step": 7400 }, { "epoch": 0.5270462570931642, "grad_norm": 0.6559625864028931, "learning_rate": 0.0001648822771465386, "loss": 0.6551, "step": 7500 }, { "epoch": 0.5270462570931642, "eval_loss": 0.6428527235984802, "eval_runtime": 118.6739, "eval_samples_per_second": 106.586, "eval_steps_per_second": 13.331, "step": 7500 }, { "epoch": 0.534073540521073, "grad_norm": 0.6792352795600891, "learning_rate": 0.00016441372847604546, "loss": 0.6622, "step": 7600 }, { "epoch": 0.541100823948982, "grad_norm": 0.6426942944526672, "learning_rate": 0.00016394517980555231, "loss": 0.6529, "step": 7700 }, { "epoch": 0.5481281073768908, "grad_norm": 0.8337587118148804, "learning_rate": 0.00016347663113505916, "loss": 0.6506, "step": 7800 }, { "epoch": 0.5551553908047996, "grad_norm": 0.5555398464202881, "learning_rate": 0.00016300808246456601, "loss": 0.6572, "step": 7900 }, { "epoch": 0.5621826742327085, "grad_norm": 0.6205873489379883, "learning_rate": 0.00016253953379407286, "loss": 0.6545, "step": 8000 }, { "epoch": 0.5621826742327085, "eval_loss": 0.639687716960907, "eval_runtime": 118.9175, "eval_samples_per_second": 106.368, "eval_steps_per_second": 13.303, "step": 8000 }, { "epoch": 0.5692099576606173, "grad_norm": 0.7055862545967102, "learning_rate": 0.00016207098512357971, "loss": 0.6428, "step": 8100 }, { "epoch": 0.5762372410885263, "grad_norm": 0.5852298140525818, "learning_rate": 0.00016160243645308656, "loss": 0.6404, "step": 8200 }, { "epoch": 0.5832645245164351, "grad_norm": 0.551387369632721, "learning_rate": 0.00016113388778259341, "loss": 0.6618, "step": 8300 }, { "epoch": 0.5902918079443439, "grad_norm": 0.7349231839179993, "learning_rate": 0.0001606653391121003, "loss": 0.6592, "step": 8400 }, { "epoch": 0.5973190913722528, "grad_norm": 0.502613365650177, "learning_rate": 0.00016019679044160714, "loss": 0.6467, "step": 8500 }, { "epoch": 0.5973190913722528, "eval_loss": 0.6377580165863037, "eval_runtime": 118.7487, "eval_samples_per_second": 106.519, "eval_steps_per_second": 13.322, "step": 8500 }, { "epoch": 0.6043463748001616, "grad_norm": 0.7003266215324402, "learning_rate": 0.000159728241771114, "loss": 0.6531, "step": 8600 }, { "epoch": 0.6113736582280704, "grad_norm": 0.42152824997901917, "learning_rate": 0.00015925969310062084, "loss": 0.647, "step": 8700 }, { "epoch": 0.6184009416559794, "grad_norm": 0.7964949607849121, "learning_rate": 0.0001587911444301277, "loss": 0.6623, "step": 8800 }, { "epoch": 0.6254282250838882, "grad_norm": 0.6723759770393372, "learning_rate": 0.00015832259575963452, "loss": 0.6528, "step": 8900 }, { "epoch": 0.6324555085117971, "grad_norm": 0.6479921936988831, "learning_rate": 0.0001578540470891414, "loss": 0.6615, "step": 9000 }, { "epoch": 0.6324555085117971, "eval_loss": 0.634124755859375, "eval_runtime": 118.6171, "eval_samples_per_second": 106.637, "eval_steps_per_second": 13.337, "step": 9000 }, { "epoch": 0.6394827919397059, "grad_norm": 0.6251150965690613, "learning_rate": 0.00015738549841864824, "loss": 0.6575, "step": 9100 }, { "epoch": 0.6465100753676147, "grad_norm": 0.6354021430015564, "learning_rate": 0.0001569169497481551, "loss": 0.6371, "step": 9200 }, { "epoch": 0.6535373587955237, "grad_norm": 0.6997053027153015, "learning_rate": 0.00015644840107766194, "loss": 0.6724, "step": 9300 }, { "epoch": 0.6605646422234325, "grad_norm": 0.6767123341560364, "learning_rate": 0.0001559798524071688, "loss": 0.6382, "step": 9400 }, { "epoch": 0.6675919256513413, "grad_norm": 0.5579701662063599, "learning_rate": 0.00015551130373667567, "loss": 0.6362, "step": 9500 }, { "epoch": 0.6675919256513413, "eval_loss": 0.6337981224060059, "eval_runtime": 118.7227, "eval_samples_per_second": 106.542, "eval_steps_per_second": 13.325, "step": 9500 }, { "epoch": 0.6746192090792502, "grad_norm": 0.6185320019721985, "learning_rate": 0.00015504275506618252, "loss": 0.638, "step": 9600 }, { "epoch": 0.681646492507159, "grad_norm": 0.5869000554084778, "learning_rate": 0.00015457420639568935, "loss": 0.6557, "step": 9700 }, { "epoch": 0.688673775935068, "grad_norm": 0.6124538779258728, "learning_rate": 0.0001541056577251962, "loss": 0.6394, "step": 9800 }, { "epoch": 0.6957010593629768, "grad_norm": 1.008245587348938, "learning_rate": 0.00015363710905470305, "loss": 0.6516, "step": 9900 }, { "epoch": 0.7027283427908856, "grad_norm": 0.5377674698829651, "learning_rate": 0.00015316856038420992, "loss": 0.6089, "step": 10000 }, { "epoch": 0.7027283427908856, "eval_loss": 0.6323862671852112, "eval_runtime": 118.7128, "eval_samples_per_second": 106.551, "eval_steps_per_second": 13.326, "step": 10000 }, { "epoch": 0.7097556262187945, "grad_norm": 0.6718229651451111, "learning_rate": 0.00015270001171371677, "loss": 0.6463, "step": 10100 }, { "epoch": 0.7167829096467033, "grad_norm": 0.613488495349884, "learning_rate": 0.00015223146304322362, "loss": 0.6513, "step": 10200 }, { "epoch": 0.7238101930746121, "grad_norm": 0.6889612674713135, "learning_rate": 0.00015176291437273047, "loss": 0.644, "step": 10300 }, { "epoch": 0.7308374765025211, "grad_norm": 0.845743715763092, "learning_rate": 0.00015129436570223732, "loss": 0.6299, "step": 10400 }, { "epoch": 0.7378647599304299, "grad_norm": 0.8277881145477295, "learning_rate": 0.00015082581703174417, "loss": 0.6592, "step": 10500 }, { "epoch": 0.7378647599304299, "eval_loss": 0.629069983959198, "eval_runtime": 118.7997, "eval_samples_per_second": 106.473, "eval_steps_per_second": 13.317, "step": 10500 }, { "epoch": 0.7448920433583387, "grad_norm": 0.5254293084144592, "learning_rate": 0.00015035726836125102, "loss": 0.6634, "step": 10600 }, { "epoch": 0.7519193267862476, "grad_norm": 0.7291231155395508, "learning_rate": 0.00014988871969075787, "loss": 0.653, "step": 10700 }, { "epoch": 0.7589466102141564, "grad_norm": 0.5473717451095581, "learning_rate": 0.00014942017102026472, "loss": 0.6477, "step": 10800 }, { "epoch": 0.7659738936420654, "grad_norm": 0.717761218547821, "learning_rate": 0.00014895162234977158, "loss": 0.6481, "step": 10900 }, { "epoch": 0.7730011770699742, "grad_norm": 0.5327322483062744, "learning_rate": 0.00014848307367927845, "loss": 0.6581, "step": 11000 }, { "epoch": 0.7730011770699742, "eval_loss": 0.6284623146057129, "eval_runtime": 119.1069, "eval_samples_per_second": 106.199, "eval_steps_per_second": 13.282, "step": 11000 }, { "epoch": 0.780028460497883, "grad_norm": 0.5596719980239868, "learning_rate": 0.0001480145250087853, "loss": 0.6372, "step": 11100 }, { "epoch": 0.7870557439257919, "grad_norm": 0.56830894947052, "learning_rate": 0.00014754597633829215, "loss": 0.6353, "step": 11200 }, { "epoch": 0.7940830273537007, "grad_norm": 0.6329615712165833, "learning_rate": 0.000147077427667799, "loss": 0.6365, "step": 11300 }, { "epoch": 0.8011103107816095, "grad_norm": 0.8399169445037842, "learning_rate": 0.00014660887899730585, "loss": 0.6212, "step": 11400 }, { "epoch": 0.8081375942095185, "grad_norm": 0.7220659255981445, "learning_rate": 0.0001461403303268127, "loss": 0.6438, "step": 11500 }, { "epoch": 0.8081375942095185, "eval_loss": 0.6267364621162415, "eval_runtime": 118.3852, "eval_samples_per_second": 106.846, "eval_steps_per_second": 13.363, "step": 11500 }, { "epoch": 0.8151648776374273, "grad_norm": 0.5614886283874512, "learning_rate": 0.00014567178165631955, "loss": 0.6262, "step": 11600 }, { "epoch": 0.8221921610653362, "grad_norm": 0.7679696679115295, "learning_rate": 0.0001452032329858264, "loss": 0.6473, "step": 11700 }, { "epoch": 0.829219444493245, "grad_norm": 0.6253560185432434, "learning_rate": 0.00014473468431533325, "loss": 0.6483, "step": 11800 }, { "epoch": 0.8362467279211538, "grad_norm": 0.5834682583808899, "learning_rate": 0.0001442661356448401, "loss": 0.6411, "step": 11900 }, { "epoch": 0.8432740113490628, "grad_norm": 0.4512103497982025, "learning_rate": 0.00014379758697434698, "loss": 0.6321, "step": 12000 }, { "epoch": 0.8432740113490628, "eval_loss": 0.6246777772903442, "eval_runtime": 118.7689, "eval_samples_per_second": 106.501, "eval_steps_per_second": 13.32, "step": 12000 }, { "epoch": 0.8503012947769716, "grad_norm": 0.6617989540100098, "learning_rate": 0.00014332903830385383, "loss": 0.6485, "step": 12100 }, { "epoch": 0.8573285782048804, "grad_norm": 0.6235445737838745, "learning_rate": 0.00014286048963336068, "loss": 0.6254, "step": 12200 }, { "epoch": 0.8643558616327893, "grad_norm": 0.612450122833252, "learning_rate": 0.00014239194096286753, "loss": 0.649, "step": 12300 }, { "epoch": 0.8713831450606981, "grad_norm": 0.7379807829856873, "learning_rate": 0.00014192339229237436, "loss": 0.6284, "step": 12400 }, { "epoch": 0.8784104284886071, "grad_norm": 0.8035106658935547, "learning_rate": 0.00014145484362188123, "loss": 0.6466, "step": 12500 }, { "epoch": 0.8784104284886071, "eval_loss": 0.6244432330131531, "eval_runtime": 118.6995, "eval_samples_per_second": 106.563, "eval_steps_per_second": 13.328, "step": 12500 }, { "epoch": 0.8854377119165159, "grad_norm": 0.6433550715446472, "learning_rate": 0.00014098629495138808, "loss": 0.6324, "step": 12600 }, { "epoch": 0.8924649953444247, "grad_norm": 0.7102698087692261, "learning_rate": 0.00014051774628089493, "loss": 0.6536, "step": 12700 }, { "epoch": 0.8994922787723336, "grad_norm": 0.7628334164619446, "learning_rate": 0.00014004919761040178, "loss": 0.6434, "step": 12800 }, { "epoch": 0.9065195622002424, "grad_norm": 0.5142523050308228, "learning_rate": 0.00013958064893990863, "loss": 0.6406, "step": 12900 }, { "epoch": 0.9135468456281512, "grad_norm": 0.8540221452713013, "learning_rate": 0.0001391121002694155, "loss": 0.6317, "step": 13000 }, { "epoch": 0.9135468456281512, "eval_loss": 0.6225576996803284, "eval_runtime": 118.6623, "eval_samples_per_second": 106.597, "eval_steps_per_second": 13.332, "step": 13000 }, { "epoch": 0.9205741290560602, "grad_norm": 0.6770111918449402, "learning_rate": 0.00013864355159892236, "loss": 0.6381, "step": 13100 }, { "epoch": 0.927601412483969, "grad_norm": 0.7313960194587708, "learning_rate": 0.00013817500292842918, "loss": 0.643, "step": 13200 }, { "epoch": 0.9346286959118779, "grad_norm": 0.8158569931983948, "learning_rate": 0.00013770645425793603, "loss": 0.6381, "step": 13300 }, { "epoch": 0.9416559793397867, "grad_norm": 0.6563596725463867, "learning_rate": 0.00013723790558744289, "loss": 0.6445, "step": 13400 }, { "epoch": 0.9486832627676955, "grad_norm": 0.6007642149925232, "learning_rate": 0.00013676935691694976, "loss": 0.6165, "step": 13500 }, { "epoch": 0.9486832627676955, "eval_loss": 0.6213079690933228, "eval_runtime": 118.8856, "eval_samples_per_second": 106.396, "eval_steps_per_second": 13.307, "step": 13500 }, { "epoch": 0.9557105461956045, "grad_norm": 0.7704166173934937, "learning_rate": 0.0001363008082464566, "loss": 0.6475, "step": 13600 }, { "epoch": 0.9627378296235133, "grad_norm": 0.6467058658599854, "learning_rate": 0.00013583225957596346, "loss": 0.626, "step": 13700 }, { "epoch": 0.9697651130514221, "grad_norm": 0.5320102572441101, "learning_rate": 0.0001353637109054703, "loss": 0.6283, "step": 13800 }, { "epoch": 0.976792396479331, "grad_norm": 0.6444761157035828, "learning_rate": 0.00013489516223497716, "loss": 0.6381, "step": 13900 }, { "epoch": 0.9838196799072398, "grad_norm": 0.7598044872283936, "learning_rate": 0.00013442661356448404, "loss": 0.6305, "step": 14000 }, { "epoch": 0.9838196799072398, "eval_loss": 0.6203290820121765, "eval_runtime": 118.8928, "eval_samples_per_second": 106.39, "eval_steps_per_second": 13.306, "step": 14000 }, { "epoch": 0.9908469633351488, "grad_norm": 0.735137403011322, "learning_rate": 0.00013395806489399086, "loss": 0.6378, "step": 14100 }, { "epoch": 0.9978742467630576, "grad_norm": 0.758840799331665, "learning_rate": 0.00013348951622349771, "loss": 0.6469, "step": 14200 }, { "epoch": 1.0049015301909665, "grad_norm": 0.5422857403755188, "learning_rate": 0.00013302096755300456, "loss": 0.586, "step": 14300 }, { "epoch": 1.0119288136188753, "grad_norm": 0.5595451593399048, "learning_rate": 0.00013255241888251141, "loss": 0.5967, "step": 14400 }, { "epoch": 1.0189560970467841, "grad_norm": 0.6194477081298828, "learning_rate": 0.0001320838702120183, "loss": 0.602, "step": 14500 }, { "epoch": 1.0189560970467841, "eval_loss": 0.6205016374588013, "eval_runtime": 118.9575, "eval_samples_per_second": 106.332, "eval_steps_per_second": 13.299, "step": 14500 }, { "epoch": 1.025983380474693, "grad_norm": 0.710670530796051, "learning_rate": 0.00013161532154152514, "loss": 0.5954, "step": 14600 }, { "epoch": 1.0330106639026018, "grad_norm": 0.613923192024231, "learning_rate": 0.000131146772871032, "loss": 0.5855, "step": 14700 }, { "epoch": 1.0400379473305108, "grad_norm": 0.7411431670188904, "learning_rate": 0.00013067822420053884, "loss": 0.5881, "step": 14800 }, { "epoch": 1.0470652307584196, "grad_norm": 0.7143989205360413, "learning_rate": 0.0001302096755300457, "loss": 0.5891, "step": 14900 }, { "epoch": 1.0540925141863284, "grad_norm": 0.6597478985786438, "learning_rate": 0.00012974112685955254, "loss": 0.5915, "step": 15000 }, { "epoch": 1.0540925141863284, "eval_loss": 0.6205988526344299, "eval_runtime": 118.9337, "eval_samples_per_second": 106.353, "eval_steps_per_second": 13.302, "step": 15000 }, { "epoch": 1.0611197976142372, "grad_norm": 0.5813501477241516, "learning_rate": 0.0001292725781890594, "loss": 0.5877, "step": 15100 }, { "epoch": 1.068147081042146, "grad_norm": 0.5610823631286621, "learning_rate": 0.00012880402951856624, "loss": 0.5911, "step": 15200 }, { "epoch": 1.075174364470055, "grad_norm": 0.5839988589286804, "learning_rate": 0.0001283354808480731, "loss": 0.5903, "step": 15300 }, { "epoch": 1.082201647897964, "grad_norm": 0.6421522498130798, "learning_rate": 0.00012786693217757994, "loss": 0.591, "step": 15400 }, { "epoch": 1.0892289313258727, "grad_norm": 0.6854695081710815, "learning_rate": 0.0001273983835070868, "loss": 0.5998, "step": 15500 }, { "epoch": 1.0892289313258727, "eval_loss": 0.6208451986312866, "eval_runtime": 118.6829, "eval_samples_per_second": 106.578, "eval_steps_per_second": 13.33, "step": 15500 }, { "epoch": 1.0962562147537815, "grad_norm": 0.7333750128746033, "learning_rate": 0.00012692983483659367, "loss": 0.5916, "step": 15600 }, { "epoch": 1.1032834981816904, "grad_norm": 0.5983602404594421, "learning_rate": 0.00012646128616610052, "loss": 0.581, "step": 15700 }, { "epoch": 1.1103107816095992, "grad_norm": 0.5112642645835876, "learning_rate": 0.00012599273749560737, "loss": 0.5984, "step": 15800 }, { "epoch": 1.1173380650375082, "grad_norm": 0.5569522976875305, "learning_rate": 0.0001255241888251142, "loss": 0.5983, "step": 15900 }, { "epoch": 1.124365348465417, "grad_norm": 0.6089062690734863, "learning_rate": 0.00012505564015462105, "loss": 0.6019, "step": 16000 }, { "epoch": 1.124365348465417, "eval_loss": 0.6191478967666626, "eval_runtime": 118.6608, "eval_samples_per_second": 106.598, "eval_steps_per_second": 13.332, "step": 16000 }, { "epoch": 1.1313926318933258, "grad_norm": 0.5710394978523254, "learning_rate": 0.00012458709148412792, "loss": 0.5907, "step": 16100 }, { "epoch": 1.1384199153212347, "grad_norm": 0.7836496233940125, "learning_rate": 0.00012411854281363477, "loss": 0.5818, "step": 16200 }, { "epoch": 1.1454471987491435, "grad_norm": 0.600236177444458, "learning_rate": 0.00012364999414314162, "loss": 0.5834, "step": 16300 }, { "epoch": 1.1524744821770523, "grad_norm": 0.7090241312980652, "learning_rate": 0.00012318144547264847, "loss": 0.5963, "step": 16400 }, { "epoch": 1.1595017656049613, "grad_norm": 0.5439143180847168, "learning_rate": 0.00012271289680215532, "loss": 0.5939, "step": 16500 }, { "epoch": 1.1595017656049613, "eval_loss": 0.6187065243721008, "eval_runtime": 118.7823, "eval_samples_per_second": 106.489, "eval_steps_per_second": 13.318, "step": 16500 }, { "epoch": 1.1665290490328701, "grad_norm": 0.6133089661598206, "learning_rate": 0.0001222443481316622, "loss": 0.583, "step": 16600 }, { "epoch": 1.173556332460779, "grad_norm": 0.6947652101516724, "learning_rate": 0.00012177579946116904, "loss": 0.5851, "step": 16700 }, { "epoch": 1.1805836158886878, "grad_norm": 0.799213707447052, "learning_rate": 0.00012130725079067589, "loss": 0.5815, "step": 16800 }, { "epoch": 1.1876108993165966, "grad_norm": 0.6443912982940674, "learning_rate": 0.00012083870212018274, "loss": 0.5794, "step": 16900 }, { "epoch": 1.1946381827445056, "grad_norm": 0.8435219526290894, "learning_rate": 0.00012037015344968959, "loss": 0.5747, "step": 17000 }, { "epoch": 1.1946381827445056, "eval_loss": 0.6187562942504883, "eval_runtime": 118.6879, "eval_samples_per_second": 106.574, "eval_steps_per_second": 13.329, "step": 17000 }, { "epoch": 1.2016654661724144, "grad_norm": 0.7162328958511353, "learning_rate": 0.00011990160477919645, "loss": 0.596, "step": 17100 }, { "epoch": 1.2086927496003232, "grad_norm": 0.6594322919845581, "learning_rate": 0.0001194330561087033, "loss": 0.6054, "step": 17200 }, { "epoch": 1.215720033028232, "grad_norm": 0.5395209193229675, "learning_rate": 0.00011896450743821015, "loss": 0.5835, "step": 17300 }, { "epoch": 1.2227473164561409, "grad_norm": 0.7208767533302307, "learning_rate": 0.000118495958767717, "loss": 0.595, "step": 17400 }, { "epoch": 1.22977459988405, "grad_norm": 0.6752803921699524, "learning_rate": 0.00011802741009722384, "loss": 0.5937, "step": 17500 }, { "epoch": 1.22977459988405, "eval_loss": 0.6167559623718262, "eval_runtime": 118.8202, "eval_samples_per_second": 106.455, "eval_steps_per_second": 13.314, "step": 17500 }, { "epoch": 1.2368018833119587, "grad_norm": 0.7853017449378967, "learning_rate": 0.00011755886142673072, "loss": 0.6035, "step": 17600 }, { "epoch": 1.2438291667398675, "grad_norm": 0.6420643329620361, "learning_rate": 0.00011709031275623757, "loss": 0.5834, "step": 17700 }, { "epoch": 1.2508564501677764, "grad_norm": 0.8449912667274475, "learning_rate": 0.00011662176408574442, "loss": 0.5964, "step": 17800 }, { "epoch": 1.2578837335956852, "grad_norm": 0.6199436783790588, "learning_rate": 0.00011615321541525125, "loss": 0.5771, "step": 17900 }, { "epoch": 1.2649110170235942, "grad_norm": 1.0329114198684692, "learning_rate": 0.0001156846667447581, "loss": 0.5678, "step": 18000 }, { "epoch": 1.2649110170235942, "eval_loss": 0.6171479821205139, "eval_runtime": 118.9379, "eval_samples_per_second": 106.35, "eval_steps_per_second": 13.301, "step": 18000 }, { "epoch": 1.271938300451503, "grad_norm": 0.7668006420135498, "learning_rate": 0.00011521611807426498, "loss": 0.5895, "step": 18100 }, { "epoch": 1.2789655838794118, "grad_norm": 0.661259651184082, "learning_rate": 0.00011474756940377183, "loss": 0.5951, "step": 18200 }, { "epoch": 1.2859928673073207, "grad_norm": 0.5448057055473328, "learning_rate": 0.00011427902073327867, "loss": 0.5958, "step": 18300 }, { "epoch": 1.2930201507352295, "grad_norm": 0.5419151782989502, "learning_rate": 0.00011381047206278552, "loss": 0.5788, "step": 18400 }, { "epoch": 1.3000474341631385, "grad_norm": 0.595245361328125, "learning_rate": 0.00011334192339229237, "loss": 0.5849, "step": 18500 }, { "epoch": 1.3000474341631385, "eval_loss": 0.6158913969993591, "eval_runtime": 118.7534, "eval_samples_per_second": 106.515, "eval_steps_per_second": 13.322, "step": 18500 }, { "epoch": 1.3070747175910473, "grad_norm": 0.6495450139045715, "learning_rate": 0.00011287337472179925, "loss": 0.5849, "step": 18600 }, { "epoch": 1.3141020010189561, "grad_norm": 0.686590313911438, "learning_rate": 0.00011240482605130608, "loss": 0.5679, "step": 18700 }, { "epoch": 1.321129284446865, "grad_norm": 0.60063636302948, "learning_rate": 0.00011193627738081293, "loss": 0.5805, "step": 18800 }, { "epoch": 1.3281565678747738, "grad_norm": 0.6396400332450867, "learning_rate": 0.00011146772871031978, "loss": 0.5996, "step": 18900 }, { "epoch": 1.3351838513026828, "grad_norm": 0.6631867289543152, "learning_rate": 0.00011099918003982663, "loss": 0.5926, "step": 19000 }, { "epoch": 1.3351838513026828, "eval_loss": 0.6149775981903076, "eval_runtime": 118.2284, "eval_samples_per_second": 106.988, "eval_steps_per_second": 13.381, "step": 19000 }, { "epoch": 1.3422111347305914, "grad_norm": 0.6868234276771545, "learning_rate": 0.0001105306313693335, "loss": 0.5826, "step": 19100 }, { "epoch": 1.3492384181585004, "grad_norm": 0.6282151937484741, "learning_rate": 0.00011006208269884035, "loss": 0.5927, "step": 19200 }, { "epoch": 1.3562657015864092, "grad_norm": 0.6498789191246033, "learning_rate": 0.0001095935340283472, "loss": 0.598, "step": 19300 }, { "epoch": 1.363292985014318, "grad_norm": 0.7654560804367065, "learning_rate": 0.00010912498535785405, "loss": 0.5819, "step": 19400 }, { "epoch": 1.3703202684422269, "grad_norm": 0.6079320907592773, "learning_rate": 0.0001086564366873609, "loss": 0.5941, "step": 19500 }, { "epoch": 1.3703202684422269, "eval_loss": 0.6134491562843323, "eval_runtime": 118.1186, "eval_samples_per_second": 107.087, "eval_steps_per_second": 13.393, "step": 19500 }, { "epoch": 1.3773475518701357, "grad_norm": 0.7858242392539978, "learning_rate": 0.00010818788801686776, "loss": 0.6035, "step": 19600 }, { "epoch": 1.3843748352980447, "grad_norm": 0.46175357699394226, "learning_rate": 0.00010771933934637461, "loss": 0.6054, "step": 19700 }, { "epoch": 1.3914021187259535, "grad_norm": 0.817308783531189, "learning_rate": 0.00010725079067588146, "loss": 0.5844, "step": 19800 }, { "epoch": 1.3984294021538624, "grad_norm": 0.7891727685928345, "learning_rate": 0.00010678224200538831, "loss": 0.5885, "step": 19900 }, { "epoch": 1.4054566855817712, "grad_norm": 0.7089536786079407, "learning_rate": 0.00010631369333489516, "loss": 0.6108, "step": 20000 }, { "epoch": 1.4054566855817712, "eval_loss": 0.6131945252418518, "eval_runtime": 119.3154, "eval_samples_per_second": 106.013, "eval_steps_per_second": 13.259, "step": 20000 }, { "epoch": 1.41248396900968, "grad_norm": 0.6867943406105042, "learning_rate": 0.00010584514466440203, "loss": 0.5863, "step": 20100 }, { "epoch": 1.419511252437589, "grad_norm": 0.9261388182640076, "learning_rate": 0.00010537659599390888, "loss": 0.6027, "step": 20200 }, { "epoch": 1.4265385358654978, "grad_norm": 0.7854331135749817, "learning_rate": 0.00010490804732341573, "loss": 0.5637, "step": 20300 }, { "epoch": 1.4335658192934067, "grad_norm": 0.4610428214073181, "learning_rate": 0.00010443949865292258, "loss": 0.5961, "step": 20400 }, { "epoch": 1.4405931027213155, "grad_norm": 0.651196300983429, "learning_rate": 0.00010397094998242943, "loss": 0.5829, "step": 20500 }, { "epoch": 1.4405931027213155, "eval_loss": 0.6108871698379517, "eval_runtime": 119.1208, "eval_samples_per_second": 106.186, "eval_steps_per_second": 13.281, "step": 20500 }, { "epoch": 1.4476203861492243, "grad_norm": 0.7416488528251648, "learning_rate": 0.00010350240131193629, "loss": 0.5869, "step": 20600 }, { "epoch": 1.4546476695771333, "grad_norm": 0.6142196655273438, "learning_rate": 0.00010303385264144314, "loss": 0.5858, "step": 20700 }, { "epoch": 1.4616749530050421, "grad_norm": 0.644241213798523, "learning_rate": 0.00010256530397094999, "loss": 0.5861, "step": 20800 }, { "epoch": 1.468702236432951, "grad_norm": 0.8656560182571411, "learning_rate": 0.00010209675530045684, "loss": 0.6099, "step": 20900 }, { "epoch": 1.4757295198608598, "grad_norm": 0.5667104721069336, "learning_rate": 0.00010162820662996368, "loss": 0.601, "step": 21000 }, { "epoch": 1.4757295198608598, "eval_loss": 0.6112544536590576, "eval_runtime": 119.0171, "eval_samples_per_second": 106.279, "eval_steps_per_second": 13.292, "step": 21000 }, { "epoch": 1.4827568032887686, "grad_norm": 0.6884378790855408, "learning_rate": 0.00010115965795947056, "loss": 0.5882, "step": 21100 }, { "epoch": 1.4897840867166776, "grad_norm": 0.7136459946632385, "learning_rate": 0.0001006911092889774, "loss": 0.5905, "step": 21200 }, { "epoch": 1.4968113701445864, "grad_norm": 0.8048639297485352, "learning_rate": 0.00010022256061848426, "loss": 0.5875, "step": 21300 }, { "epoch": 1.5038386535724952, "grad_norm": 0.7300230860710144, "learning_rate": 9.975401194799109e-05, "loss": 0.5916, "step": 21400 }, { "epoch": 1.510865937000404, "grad_norm": 0.7058496475219727, "learning_rate": 9.928546327749796e-05, "loss": 0.5768, "step": 21500 }, { "epoch": 1.510865937000404, "eval_loss": 0.6108001470565796, "eval_runtime": 118.8504, "eval_samples_per_second": 106.428, "eval_steps_per_second": 13.311, "step": 21500 }, { "epoch": 1.5178932204283129, "grad_norm": 0.8157733678817749, "learning_rate": 9.881691460700481e-05, "loss": 0.5872, "step": 21600 }, { "epoch": 1.524920503856222, "grad_norm": 0.8618035316467285, "learning_rate": 9.834836593651166e-05, "loss": 0.5861, "step": 21700 }, { "epoch": 1.5319477872841305, "grad_norm": 0.7457069158554077, "learning_rate": 9.787981726601851e-05, "loss": 0.5948, "step": 21800 }, { "epoch": 1.5389750707120395, "grad_norm": 0.7307142615318298, "learning_rate": 9.741126859552536e-05, "loss": 0.6068, "step": 21900 }, { "epoch": 1.5460023541399484, "grad_norm": 0.7692698836326599, "learning_rate": 9.694271992503222e-05, "loss": 0.5732, "step": 22000 }, { "epoch": 1.5460023541399484, "eval_loss": 0.6103039979934692, "eval_runtime": 118.5221, "eval_samples_per_second": 106.723, "eval_steps_per_second": 13.348, "step": 22000 }, { "epoch": 1.5530296375678572, "grad_norm": 0.7781071662902832, "learning_rate": 9.647417125453907e-05, "loss": 0.5915, "step": 22100 }, { "epoch": 1.5600569209957662, "grad_norm": 0.7720737457275391, "learning_rate": 9.600562258404592e-05, "loss": 0.6007, "step": 22200 }, { "epoch": 1.5670842044236748, "grad_norm": 0.632757306098938, "learning_rate": 9.553707391355277e-05, "loss": 0.5838, "step": 22300 }, { "epoch": 1.5741114878515838, "grad_norm": 0.7514855265617371, "learning_rate": 9.506852524305962e-05, "loss": 0.5894, "step": 22400 }, { "epoch": 1.5811387712794926, "grad_norm": 0.5634511113166809, "learning_rate": 9.459997657256649e-05, "loss": 0.5774, "step": 22500 }, { "epoch": 1.5811387712794926, "eval_loss": 0.6085862517356873, "eval_runtime": 119.1435, "eval_samples_per_second": 106.166, "eval_steps_per_second": 13.278, "step": 22500 }, { "epoch": 1.5881660547074015, "grad_norm": 0.8155964612960815, "learning_rate": 9.413142790207334e-05, "loss": 0.582, "step": 22600 }, { "epoch": 1.5951933381353105, "grad_norm": 0.7442721128463745, "learning_rate": 9.366287923158019e-05, "loss": 0.595, "step": 22700 }, { "epoch": 1.602220621563219, "grad_norm": 0.6977400183677673, "learning_rate": 9.319433056108704e-05, "loss": 0.5906, "step": 22800 }, { "epoch": 1.6092479049911281, "grad_norm": 0.7066090703010559, "learning_rate": 9.272578189059389e-05, "loss": 0.5917, "step": 22900 }, { "epoch": 1.616275188419037, "grad_norm": 0.9004433751106262, "learning_rate": 9.225723322010075e-05, "loss": 0.5821, "step": 23000 }, { "epoch": 1.616275188419037, "eval_loss": 0.608232319355011, "eval_runtime": 118.5994, "eval_samples_per_second": 106.653, "eval_steps_per_second": 13.339, "step": 23000 }, { "epoch": 1.6233024718469458, "grad_norm": 0.6980244517326355, "learning_rate": 9.17886845496076e-05, "loss": 0.5986, "step": 23100 }, { "epoch": 1.6303297552748546, "grad_norm": 0.6959982514381409, "learning_rate": 9.132013587911444e-05, "loss": 0.5751, "step": 23200 }, { "epoch": 1.6373570387027634, "grad_norm": 0.6452066898345947, "learning_rate": 9.08515872086213e-05, "loss": 0.5834, "step": 23300 }, { "epoch": 1.6443843221306724, "grad_norm": 0.563113272190094, "learning_rate": 9.038303853812815e-05, "loss": 0.5943, "step": 23400 }, { "epoch": 1.6514116055585812, "grad_norm": 0.6849614977836609, "learning_rate": 8.991448986763502e-05, "loss": 0.581, "step": 23500 }, { "epoch": 1.6514116055585812, "eval_loss": 0.607568621635437, "eval_runtime": 119.0065, "eval_samples_per_second": 106.288, "eval_steps_per_second": 13.293, "step": 23500 }, { "epoch": 1.65843888898649, "grad_norm": 0.8570700287818909, "learning_rate": 8.944594119714185e-05, "loss": 0.592, "step": 23600 }, { "epoch": 1.6654661724143989, "grad_norm": 0.4864564538002014, "learning_rate": 8.89773925266487e-05, "loss": 0.5722, "step": 23700 }, { "epoch": 1.6724934558423077, "grad_norm": 0.6979348063468933, "learning_rate": 8.850884385615557e-05, "loss": 0.578, "step": 23800 }, { "epoch": 1.6795207392702167, "grad_norm": 0.7699964046478271, "learning_rate": 8.804029518566242e-05, "loss": 0.5897, "step": 23900 }, { "epoch": 1.6865480226981253, "grad_norm": 0.8256754875183105, "learning_rate": 8.757174651516927e-05, "loss": 0.5845, "step": 24000 }, { "epoch": 1.6865480226981253, "eval_loss": 0.6071833372116089, "eval_runtime": 119.1197, "eval_samples_per_second": 106.187, "eval_steps_per_second": 13.281, "step": 24000 }, { "epoch": 1.6935753061260344, "grad_norm": 0.6591055393218994, "learning_rate": 8.710319784467612e-05, "loss": 0.5867, "step": 24100 }, { "epoch": 1.7006025895539432, "grad_norm": 0.7720032334327698, "learning_rate": 8.663464917418297e-05, "loss": 0.5905, "step": 24200 }, { "epoch": 1.707629872981852, "grad_norm": 0.5807234048843384, "learning_rate": 8.616610050368983e-05, "loss": 0.5849, "step": 24300 }, { "epoch": 1.714657156409761, "grad_norm": 0.6583465337753296, "learning_rate": 8.569755183319668e-05, "loss": 0.5706, "step": 24400 }, { "epoch": 1.7216844398376696, "grad_norm": 0.7280032634735107, "learning_rate": 8.522900316270352e-05, "loss": 0.5843, "step": 24500 }, { "epoch": 1.7216844398376696, "eval_loss": 0.6051200032234192, "eval_runtime": 119.1844, "eval_samples_per_second": 106.13, "eval_steps_per_second": 13.274, "step": 24500 }, { "epoch": 1.7287117232655786, "grad_norm": 0.7264565825462341, "learning_rate": 8.476045449221038e-05, "loss": 0.5866, "step": 24600 }, { "epoch": 1.7357390066934875, "grad_norm": 0.7795102596282959, "learning_rate": 8.429190582171723e-05, "loss": 0.5807, "step": 24700 }, { "epoch": 1.7427662901213963, "grad_norm": 0.702314019203186, "learning_rate": 8.38233571512241e-05, "loss": 0.5856, "step": 24800 }, { "epoch": 1.7497935735493053, "grad_norm": 0.7014954090118408, "learning_rate": 8.335480848073095e-05, "loss": 0.5812, "step": 24900 }, { "epoch": 1.756820856977214, "grad_norm": 0.6724287867546082, "learning_rate": 8.288625981023778e-05, "loss": 0.5803, "step": 25000 }, { "epoch": 1.756820856977214, "eval_loss": 0.6055319309234619, "eval_runtime": 119.2213, "eval_samples_per_second": 106.097, "eval_steps_per_second": 13.269, "step": 25000 }, { "epoch": 1.763848140405123, "grad_norm": 0.9690128564834595, "learning_rate": 8.241771113974465e-05, "loss": 0.5674, "step": 25100 }, { "epoch": 1.7708754238330318, "grad_norm": 1.0121440887451172, "learning_rate": 8.19491624692515e-05, "loss": 0.5913, "step": 25200 }, { "epoch": 1.7779027072609406, "grad_norm": 0.7207921743392944, "learning_rate": 8.148061379875836e-05, "loss": 0.592, "step": 25300 }, { "epoch": 1.7849299906888496, "grad_norm": 0.6326346397399902, "learning_rate": 8.10120651282652e-05, "loss": 0.592, "step": 25400 }, { "epoch": 1.7919572741167582, "grad_norm": 0.7215606570243835, "learning_rate": 8.054351645777205e-05, "loss": 0.578, "step": 25500 }, { "epoch": 1.7919572741167582, "eval_loss": 0.6045902371406555, "eval_runtime": 118.5573, "eval_samples_per_second": 106.691, "eval_steps_per_second": 13.344, "step": 25500 }, { "epoch": 1.7989845575446672, "grad_norm": 0.6932191848754883, "learning_rate": 8.007496778727891e-05, "loss": 0.5881, "step": 25600 }, { "epoch": 1.806011840972576, "grad_norm": 0.7887512445449829, "learning_rate": 7.960641911678576e-05, "loss": 0.5875, "step": 25700 }, { "epoch": 1.8130391244004849, "grad_norm": 0.5214329957962036, "learning_rate": 7.913787044629261e-05, "loss": 0.5757, "step": 25800 }, { "epoch": 1.8200664078283937, "grad_norm": 0.6298120021820068, "learning_rate": 7.866932177579946e-05, "loss": 0.5634, "step": 25900 }, { "epoch": 1.8270936912563025, "grad_norm": 0.6642977595329285, "learning_rate": 7.820077310530631e-05, "loss": 0.5708, "step": 26000 }, { "epoch": 1.8270936912563025, "eval_loss": 0.6030368804931641, "eval_runtime": 118.8611, "eval_samples_per_second": 106.418, "eval_steps_per_second": 13.31, "step": 26000 }, { "epoch": 1.8341209746842115, "grad_norm": 0.5501639246940613, "learning_rate": 7.773222443481318e-05, "loss": 0.5692, "step": 26100 }, { "epoch": 1.8411482581121204, "grad_norm": 0.6441388726234436, "learning_rate": 7.726367576432003e-05, "loss": 0.5881, "step": 26200 }, { "epoch": 1.8481755415400292, "grad_norm": 0.6169604063034058, "learning_rate": 7.679512709382688e-05, "loss": 0.5762, "step": 26300 }, { "epoch": 1.855202824967938, "grad_norm": 0.6980007290840149, "learning_rate": 7.632657842333373e-05, "loss": 0.5739, "step": 26400 }, { "epoch": 1.8622301083958468, "grad_norm": 0.7939792275428772, "learning_rate": 7.585802975284058e-05, "loss": 0.5746, "step": 26500 }, { "epoch": 1.8622301083958468, "eval_loss": 0.6025614142417908, "eval_runtime": 118.9608, "eval_samples_per_second": 106.329, "eval_steps_per_second": 13.298, "step": 26500 }, { "epoch": 1.8692573918237558, "grad_norm": 0.6231071352958679, "learning_rate": 7.538948108234744e-05, "loss": 0.5634, "step": 26600 }, { "epoch": 1.8762846752516644, "grad_norm": 0.6939712762832642, "learning_rate": 7.492093241185428e-05, "loss": 0.5792, "step": 26700 }, { "epoch": 1.8833119586795735, "grad_norm": 0.6055401563644409, "learning_rate": 7.445238374136114e-05, "loss": 0.5929, "step": 26800 }, { "epoch": 1.8903392421074823, "grad_norm": 0.7465933561325073, "learning_rate": 7.398383507086799e-05, "loss": 0.5817, "step": 26900 }, { "epoch": 1.897366525535391, "grad_norm": 0.6460291743278503, "learning_rate": 7.351528640037484e-05, "loss": 0.5831, "step": 27000 }, { "epoch": 1.897366525535391, "eval_loss": 0.6021212935447693, "eval_runtime": 118.8793, "eval_samples_per_second": 106.402, "eval_steps_per_second": 13.308, "step": 27000 }, { "epoch": 1.9043938089633001, "grad_norm": 0.5837533473968506, "learning_rate": 7.304673772988169e-05, "loss": 0.5807, "step": 27100 }, { "epoch": 1.9114210923912087, "grad_norm": 0.7371869087219238, "learning_rate": 7.257818905938854e-05, "loss": 0.5967, "step": 27200 }, { "epoch": 1.9184483758191178, "grad_norm": 0.8853654861450195, "learning_rate": 7.21096403888954e-05, "loss": 0.5967, "step": 27300 }, { "epoch": 1.9254756592470266, "grad_norm": 0.7515887022018433, "learning_rate": 7.164109171840226e-05, "loss": 0.5708, "step": 27400 }, { "epoch": 1.9325029426749354, "grad_norm": 0.6723042130470276, "learning_rate": 7.11725430479091e-05, "loss": 0.5687, "step": 27500 }, { "epoch": 1.9325029426749354, "eval_loss": 0.6017782092094421, "eval_runtime": 118.9577, "eval_samples_per_second": 106.332, "eval_steps_per_second": 13.299, "step": 27500 }, { "epoch": 1.9395302261028444, "grad_norm": 0.5766080617904663, "learning_rate": 7.070399437741596e-05, "loss": 0.5876, "step": 27600 }, { "epoch": 1.946557509530753, "grad_norm": 0.6501230597496033, "learning_rate": 7.02354457069228e-05, "loss": 0.5726, "step": 27700 }, { "epoch": 1.953584792958662, "grad_norm": 0.7194878458976746, "learning_rate": 6.976689703642966e-05, "loss": 0.5741, "step": 27800 }, { "epoch": 1.9606120763865709, "grad_norm": 0.8090994954109192, "learning_rate": 6.929834836593652e-05, "loss": 0.5764, "step": 27900 }, { "epoch": 1.9676393598144797, "grad_norm": 0.6500638127326965, "learning_rate": 6.882979969544337e-05, "loss": 0.5752, "step": 28000 }, { "epoch": 1.9676393598144797, "eval_loss": 0.6010117530822754, "eval_runtime": 119.0388, "eval_samples_per_second": 106.259, "eval_steps_per_second": 13.29, "step": 28000 } ], "logging_steps": 100, "max_steps": 42690, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 9.547282008374477e+16, "train_batch_size": 2, "trial_name": null, "trial_params": null }