{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 26, "global_step": 253, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.003952569169960474, "grad_norm": 3.0407819747924805, "learning_rate": 3.3333333333333333e-06, "loss": 2.3652, "step": 1 }, { "epoch": 0.007905138339920948, "grad_norm": 3.093733310699463, "learning_rate": 6.666666666666667e-06, "loss": 2.1458, "step": 2 }, { "epoch": 0.011857707509881422, "grad_norm": 3.0263988971710205, "learning_rate": 1e-05, "loss": 2.2628, "step": 3 }, { "epoch": 0.015810276679841896, "grad_norm": 2.6713979244232178, "learning_rate": 9.999605221019082e-06, "loss": 2.0308, "step": 4 }, { "epoch": 0.019762845849802372, "grad_norm": 2.4664857387542725, "learning_rate": 9.9984209464165e-06, "loss": 1.6917, "step": 5 }, { "epoch": 0.023715415019762844, "grad_norm": 2.100984811782837, "learning_rate": 9.996447363202947e-06, "loss": 1.8489, "step": 6 }, { "epoch": 0.02766798418972332, "grad_norm": 1.7837852239608765, "learning_rate": 9.99368478303009e-06, "loss": 1.9894, "step": 7 }, { "epoch": 0.03162055335968379, "grad_norm": 1.6338787078857422, "learning_rate": 9.990133642141359e-06, "loss": 1.8366, "step": 8 }, { "epoch": 0.03557312252964427, "grad_norm": 1.7102547883987427, "learning_rate": 9.98579450130307e-06, "loss": 1.9013, "step": 9 }, { "epoch": 0.039525691699604744, "grad_norm": 1.8172574043273926, "learning_rate": 9.980668045715864e-06, "loss": 2.034, "step": 10 }, { "epoch": 0.043478260869565216, "grad_norm": 1.7970800399780273, "learning_rate": 9.974755084906503e-06, "loss": 1.8565, "step": 11 }, { "epoch": 0.04743083003952569, "grad_norm": 1.9662507772445679, "learning_rate": 9.968056552600043e-06, "loss": 1.8154, "step": 12 }, { "epoch": 0.05138339920948617, "grad_norm": 1.6226023435592651, "learning_rate": 9.960573506572391e-06, "loss": 1.7205, "step": 13 }, { "epoch": 0.05533596837944664, "grad_norm": 1.4996356964111328, "learning_rate": 9.952307128483257e-06, "loss": 1.8481, "step": 14 }, { "epoch": 0.05928853754940711, "grad_norm": 1.2719931602478027, "learning_rate": 9.94325872368957e-06, "loss": 1.5792, "step": 15 }, { "epoch": 0.06324110671936758, "grad_norm": 1.2413394451141357, "learning_rate": 9.93342972103934e-06, "loss": 1.6527, "step": 16 }, { "epoch": 0.06719367588932806, "grad_norm": 1.1712374687194824, "learning_rate": 9.922821672646028e-06, "loss": 1.3883, "step": 17 }, { "epoch": 0.07114624505928854, "grad_norm": 1.1695728302001953, "learning_rate": 9.911436253643445e-06, "loss": 1.5297, "step": 18 }, { "epoch": 0.07509881422924901, "grad_norm": 1.1794648170471191, "learning_rate": 9.899275261921236e-06, "loss": 1.5525, "step": 19 }, { "epoch": 0.07905138339920949, "grad_norm": 1.131502628326416, "learning_rate": 9.886340617840968e-06, "loss": 1.5752, "step": 20 }, { "epoch": 0.08300395256916997, "grad_norm": 1.1554540395736694, "learning_rate": 9.872634363932887e-06, "loss": 1.5944, "step": 21 }, { "epoch": 0.08695652173913043, "grad_norm": 1.1099374294281006, "learning_rate": 9.85815866457337e-06, "loss": 1.4019, "step": 22 }, { "epoch": 0.09090909090909091, "grad_norm": 1.2356349229812622, "learning_rate": 9.842915805643156e-06, "loss": 1.706, "step": 23 }, { "epoch": 0.09486166007905138, "grad_norm": 1.1689324378967285, "learning_rate": 9.82690819416637e-06, "loss": 1.6249, "step": 24 }, { "epoch": 0.09881422924901186, "grad_norm": 1.2375783920288086, "learning_rate": 9.81013835793043e-06, "loss": 1.8374, "step": 25 }, { "epoch": 0.10276679841897234, "grad_norm": 1.1797913312911987, "learning_rate": 9.79260894508688e-06, "loss": 1.6161, "step": 26 }, { "epoch": 0.10276679841897234, "eval_loss": 1.4920315742492676, "eval_runtime": 0.6276, "eval_samples_per_second": 33.462, "eval_steps_per_second": 4.78, "step": 26 }, { "epoch": 0.1067193675889328, "grad_norm": 1.1279163360595703, "learning_rate": 9.774322723733216e-06, "loss": 1.6593, "step": 27 }, { "epoch": 0.11067193675889328, "grad_norm": 1.1455029249191284, "learning_rate": 9.755282581475769e-06, "loss": 1.6982, "step": 28 }, { "epoch": 0.11462450592885376, "grad_norm": 1.0429457426071167, "learning_rate": 9.735491524973723e-06, "loss": 1.4423, "step": 29 }, { "epoch": 0.11857707509881422, "grad_norm": 0.9471132755279541, "learning_rate": 9.714952679464324e-06, "loss": 1.3055, "step": 30 }, { "epoch": 0.1225296442687747, "grad_norm": 1.0985954999923706, "learning_rate": 9.693669288269371e-06, "loss": 1.6302, "step": 31 }, { "epoch": 0.12648221343873517, "grad_norm": 1.11372709274292, "learning_rate": 9.671644712283061e-06, "loss": 1.7975, "step": 32 }, { "epoch": 0.13043478260869565, "grad_norm": 0.9115369915962219, "learning_rate": 9.648882429441258e-06, "loss": 1.2709, "step": 33 }, { "epoch": 0.13438735177865613, "grad_norm": 1.0038697719573975, "learning_rate": 9.62538603417229e-06, "loss": 1.4833, "step": 34 }, { "epoch": 0.1383399209486166, "grad_norm": 1.0109611749649048, "learning_rate": 9.601159236829353e-06, "loss": 1.2978, "step": 35 }, { "epoch": 0.1422924901185771, "grad_norm": 0.966667115688324, "learning_rate": 9.576205863104588e-06, "loss": 1.5622, "step": 36 }, { "epoch": 0.14624505928853754, "grad_norm": 1.0922729969024658, "learning_rate": 9.550529853424979e-06, "loss": 1.7766, "step": 37 }, { "epoch": 0.15019762845849802, "grad_norm": 1.010776162147522, "learning_rate": 9.524135262330098e-06, "loss": 1.5206, "step": 38 }, { "epoch": 0.1541501976284585, "grad_norm": 0.9982693791389465, "learning_rate": 9.497026257831856e-06, "loss": 1.5341, "step": 39 }, { "epoch": 0.15810276679841898, "grad_norm": 0.9735735058784485, "learning_rate": 9.46920712075632e-06, "loss": 1.4157, "step": 40 }, { "epoch": 0.16205533596837945, "grad_norm": 0.9151104092597961, "learning_rate": 9.440682244067724e-06, "loss": 1.2846, "step": 41 }, { "epoch": 0.16600790513833993, "grad_norm": 0.9795275330543518, "learning_rate": 9.411456132174768e-06, "loss": 1.4727, "step": 42 }, { "epoch": 0.16996047430830039, "grad_norm": 1.1112526655197144, "learning_rate": 9.381533400219319e-06, "loss": 1.6619, "step": 43 }, { "epoch": 0.17391304347826086, "grad_norm": 0.9755843877792358, "learning_rate": 9.35091877334763e-06, "loss": 1.4687, "step": 44 }, { "epoch": 0.17786561264822134, "grad_norm": 0.9589361548423767, "learning_rate": 9.319617085964177e-06, "loss": 1.4871, "step": 45 }, { "epoch": 0.18181818181818182, "grad_norm": 1.105637788772583, "learning_rate": 9.287633280968263e-06, "loss": 1.2694, "step": 46 }, { "epoch": 0.1857707509881423, "grad_norm": 0.9662937521934509, "learning_rate": 9.25497240897346e-06, "loss": 1.2148, "step": 47 }, { "epoch": 0.18972332015810275, "grad_norm": 1.0539385080337524, "learning_rate": 9.221639627510076e-06, "loss": 1.5932, "step": 48 }, { "epoch": 0.19367588932806323, "grad_norm": 0.9249448180198669, "learning_rate": 9.18764020021071e-06, "loss": 1.3074, "step": 49 }, { "epoch": 0.1976284584980237, "grad_norm": 1.0217876434326172, "learning_rate": 9.152979495979064e-06, "loss": 1.5327, "step": 50 }, { "epoch": 0.2015810276679842, "grad_norm": 1.070178747177124, "learning_rate": 9.117662988142138e-06, "loss": 1.6561, "step": 51 }, { "epoch": 0.20553359683794467, "grad_norm": 0.9971439242362976, "learning_rate": 9.08169625358592e-06, "loss": 1.3683, "step": 52 }, { "epoch": 0.20553359683794467, "eval_loss": 1.325599193572998, "eval_runtime": 0.6316, "eval_samples_per_second": 33.25, "eval_steps_per_second": 4.75, "step": 52 }, { "epoch": 0.20948616600790515, "grad_norm": 0.9724494814872742, "learning_rate": 9.045084971874738e-06, "loss": 1.5192, "step": 53 }, { "epoch": 0.2134387351778656, "grad_norm": 0.9346773624420166, "learning_rate": 9.007834924354384e-06, "loss": 1.4107, "step": 54 }, { "epoch": 0.21739130434782608, "grad_norm": 0.9584831595420837, "learning_rate": 8.969951993239177e-06, "loss": 1.392, "step": 55 }, { "epoch": 0.22134387351778656, "grad_norm": 0.8788594603538513, "learning_rate": 8.931442160683094e-06, "loss": 1.2833, "step": 56 }, { "epoch": 0.22529644268774704, "grad_norm": 0.9413896203041077, "learning_rate": 8.892311507835118e-06, "loss": 1.26, "step": 57 }, { "epoch": 0.22924901185770752, "grad_norm": 0.944340169429779, "learning_rate": 8.852566213878947e-06, "loss": 1.3223, "step": 58 }, { "epoch": 0.233201581027668, "grad_norm": 0.8935737609863281, "learning_rate": 8.81221255505724e-06, "loss": 1.3826, "step": 59 }, { "epoch": 0.23715415019762845, "grad_norm": 0.9506601095199585, "learning_rate": 8.77125690368052e-06, "loss": 1.4478, "step": 60 }, { "epoch": 0.24110671936758893, "grad_norm": 0.9523410201072693, "learning_rate": 8.729705727120911e-06, "loss": 1.3503, "step": 61 }, { "epoch": 0.2450592885375494, "grad_norm": 0.8316011428833008, "learning_rate": 8.68756558679087e-06, "loss": 1.041, "step": 62 }, { "epoch": 0.2490118577075099, "grad_norm": 0.9092923998832703, "learning_rate": 8.644843137107058e-06, "loss": 1.323, "step": 63 }, { "epoch": 0.25296442687747034, "grad_norm": 0.9320958852767944, "learning_rate": 8.601545124439535e-06, "loss": 1.3562, "step": 64 }, { "epoch": 0.25691699604743085, "grad_norm": 0.9427369832992554, "learning_rate": 8.557678386046429e-06, "loss": 1.5575, "step": 65 }, { "epoch": 0.2608695652173913, "grad_norm": 0.9649813175201416, "learning_rate": 8.513249848994248e-06, "loss": 1.2068, "step": 66 }, { "epoch": 0.2648221343873518, "grad_norm": 0.9272534251213074, "learning_rate": 8.468266529064025e-06, "loss": 1.2409, "step": 67 }, { "epoch": 0.26877470355731226, "grad_norm": 0.952318549156189, "learning_rate": 8.422735529643445e-06, "loss": 1.4648, "step": 68 }, { "epoch": 0.2727272727272727, "grad_norm": 0.7945717573165894, "learning_rate": 8.376664040605122e-06, "loss": 1.142, "step": 69 }, { "epoch": 0.2766798418972332, "grad_norm": 0.8417496085166931, "learning_rate": 8.33005933717126e-06, "loss": 1.179, "step": 70 }, { "epoch": 0.28063241106719367, "grad_norm": 0.8613891005516052, "learning_rate": 8.282928778764783e-06, "loss": 1.258, "step": 71 }, { "epoch": 0.2845849802371542, "grad_norm": 0.8771772980690002, "learning_rate": 8.235279807847223e-06, "loss": 1.3277, "step": 72 }, { "epoch": 0.2885375494071146, "grad_norm": 0.9778826832771301, "learning_rate": 8.18711994874345e-06, "loss": 1.1869, "step": 73 }, { "epoch": 0.2924901185770751, "grad_norm": 0.9443055391311646, "learning_rate": 8.138456806453503e-06, "loss": 1.37, "step": 74 }, { "epoch": 0.2964426877470356, "grad_norm": 0.8966163992881775, "learning_rate": 8.089298065451673e-06, "loss": 1.2171, "step": 75 }, { "epoch": 0.30039525691699603, "grad_norm": 0.9195836782455444, "learning_rate": 8.039651488473028e-06, "loss": 1.3983, "step": 76 }, { "epoch": 0.30434782608695654, "grad_norm": 0.952008843421936, "learning_rate": 7.989524915287595e-06, "loss": 1.4224, "step": 77 }, { "epoch": 0.308300395256917, "grad_norm": 1.0114339590072632, "learning_rate": 7.938926261462366e-06, "loss": 1.409, "step": 78 }, { "epoch": 0.308300395256917, "eval_loss": 1.2486090660095215, "eval_runtime": 0.6312, "eval_samples_per_second": 33.269, "eval_steps_per_second": 4.753, "step": 78 }, { "epoch": 0.31225296442687744, "grad_norm": 0.8975977301597595, "learning_rate": 7.887863517111337e-06, "loss": 1.3565, "step": 79 }, { "epoch": 0.31620553359683795, "grad_norm": 0.9322842955589294, "learning_rate": 7.836344745633785e-06, "loss": 1.336, "step": 80 }, { "epoch": 0.3201581027667984, "grad_norm": 0.934016227722168, "learning_rate": 7.78437808244094e-06, "loss": 1.318, "step": 81 }, { "epoch": 0.3241106719367589, "grad_norm": 0.8984368443489075, "learning_rate": 7.731971733671347e-06, "loss": 1.264, "step": 82 }, { "epoch": 0.32806324110671936, "grad_norm": 0.9175249338150024, "learning_rate": 7.679133974894984e-06, "loss": 1.2758, "step": 83 }, { "epoch": 0.33201581027667987, "grad_norm": 1.0880227088928223, "learning_rate": 7.6258731498064796e-06, "loss": 1.1448, "step": 84 }, { "epoch": 0.3359683794466403, "grad_norm": 0.8632190823554993, "learning_rate": 7.572197668907533e-06, "loss": 1.2644, "step": 85 }, { "epoch": 0.33992094861660077, "grad_norm": 0.8895804286003113, "learning_rate": 7.518116008178805e-06, "loss": 1.2656, "step": 86 }, { "epoch": 0.3438735177865613, "grad_norm": 0.9582334756851196, "learning_rate": 7.463636707741458e-06, "loss": 1.2453, "step": 87 }, { "epoch": 0.34782608695652173, "grad_norm": 0.8999531865119934, "learning_rate": 7.408768370508577e-06, "loss": 1.1815, "step": 88 }, { "epoch": 0.35177865612648224, "grad_norm": 0.8813087940216064, "learning_rate": 7.353519660826665e-06, "loss": 1.2232, "step": 89 }, { "epoch": 0.3557312252964427, "grad_norm": 0.8202010989189148, "learning_rate": 7.297899303107441e-06, "loss": 1.1386, "step": 90 }, { "epoch": 0.35968379446640314, "grad_norm": 0.9317258596420288, "learning_rate": 7.241916080450163e-06, "loss": 1.2962, "step": 91 }, { "epoch": 0.36363636363636365, "grad_norm": 0.8814587593078613, "learning_rate": 7.185578833254665e-06, "loss": 1.3109, "step": 92 }, { "epoch": 0.3675889328063241, "grad_norm": 0.9203696846961975, "learning_rate": 7.128896457825364e-06, "loss": 1.4728, "step": 93 }, { "epoch": 0.3715415019762846, "grad_norm": 0.842402994632721, "learning_rate": 7.071877904966422e-06, "loss": 1.2597, "step": 94 }, { "epoch": 0.37549407114624506, "grad_norm": 0.8974599838256836, "learning_rate": 7.014532178568314e-06, "loss": 1.2041, "step": 95 }, { "epoch": 0.3794466403162055, "grad_norm": 0.8709033727645874, "learning_rate": 6.9568683341860135e-06, "loss": 1.1369, "step": 96 }, { "epoch": 0.383399209486166, "grad_norm": 0.8162269592285156, "learning_rate": 6.898895477609007e-06, "loss": 1.2103, "step": 97 }, { "epoch": 0.38735177865612647, "grad_norm": 0.8966283798217773, "learning_rate": 6.840622763423391e-06, "loss": 1.2609, "step": 98 }, { "epoch": 0.391304347826087, "grad_norm": 0.8919047117233276, "learning_rate": 6.782059393566254e-06, "loss": 1.1807, "step": 99 }, { "epoch": 0.3952569169960474, "grad_norm": 0.9296970367431641, "learning_rate": 6.723214615872585e-06, "loss": 1.3748, "step": 100 }, { "epoch": 0.39920948616600793, "grad_norm": 0.8833282589912415, "learning_rate": 6.664097722614934e-06, "loss": 1.271, "step": 101 }, { "epoch": 0.4031620553359684, "grad_norm": 0.9111071228981018, "learning_rate": 6.604718049036047e-06, "loss": 1.3016, "step": 102 }, { "epoch": 0.40711462450592883, "grad_norm": 0.9141144156455994, "learning_rate": 6.545084971874738e-06, "loss": 1.3072, "step": 103 }, { "epoch": 0.41106719367588934, "grad_norm": 0.8589788675308228, "learning_rate": 6.485207907885175e-06, "loss": 1.1914, "step": 104 }, { "epoch": 0.41106719367588934, "eval_loss": 1.2048839330673218, "eval_runtime": 0.6355, "eval_samples_per_second": 33.045, "eval_steps_per_second": 4.721, "step": 104 }, { "epoch": 0.4150197628458498, "grad_norm": 0.9179502725601196, "learning_rate": 6.425096312349881e-06, "loss": 1.3336, "step": 105 }, { "epoch": 0.4189723320158103, "grad_norm": 0.8490063548088074, "learning_rate": 6.364759677586627e-06, "loss": 1.275, "step": 106 }, { "epoch": 0.42292490118577075, "grad_norm": 0.8626274466514587, "learning_rate": 6.304207531449486e-06, "loss": 1.237, "step": 107 }, { "epoch": 0.4268774703557312, "grad_norm": 0.855128824710846, "learning_rate": 6.243449435824276e-06, "loss": 1.1993, "step": 108 }, { "epoch": 0.4308300395256917, "grad_norm": 0.8959077000617981, "learning_rate": 6.182494985118625e-06, "loss": 1.3898, "step": 109 }, { "epoch": 0.43478260869565216, "grad_norm": 0.9411128163337708, "learning_rate": 6.121353804746907e-06, "loss": 1.1805, "step": 110 }, { "epoch": 0.43873517786561267, "grad_norm": 0.8071564435958862, "learning_rate": 6.060035549610275e-06, "loss": 0.8793, "step": 111 }, { "epoch": 0.4426877470355731, "grad_norm": 0.873630702495575, "learning_rate": 5.9985499025720354e-06, "loss": 1.3551, "step": 112 }, { "epoch": 0.44664031620553357, "grad_norm": 0.7491222620010376, "learning_rate": 5.936906572928625e-06, "loss": 1.0515, "step": 113 }, { "epoch": 0.4505928853754941, "grad_norm": 0.8194934129714966, "learning_rate": 5.8751152948763815e-06, "loss": 1.0884, "step": 114 }, { "epoch": 0.45454545454545453, "grad_norm": 0.9355419874191284, "learning_rate": 5.813185825974419e-06, "loss": 1.4432, "step": 115 }, { "epoch": 0.45849802371541504, "grad_norm": 0.760794460773468, "learning_rate": 5.751127945603786e-06, "loss": 1.0239, "step": 116 }, { "epoch": 0.4624505928853755, "grad_norm": 0.9574670791625977, "learning_rate": 5.68895145342319e-06, "loss": 1.5178, "step": 117 }, { "epoch": 0.466403162055336, "grad_norm": 0.9131107330322266, "learning_rate": 5.626666167821522e-06, "loss": 1.3872, "step": 118 }, { "epoch": 0.47035573122529645, "grad_norm": 0.9126591682434082, "learning_rate": 5.5642819243674085e-06, "loss": 1.3254, "step": 119 }, { "epoch": 0.4743083003952569, "grad_norm": 0.8607676029205322, "learning_rate": 5.5018085742560745e-06, "loss": 1.1636, "step": 120 }, { "epoch": 0.4782608695652174, "grad_norm": 0.8300676345825195, "learning_rate": 5.439255982753717e-06, "loss": 1.0787, "step": 121 }, { "epoch": 0.48221343873517786, "grad_norm": 0.8748376965522766, "learning_rate": 5.376634027639664e-06, "loss": 1.3438, "step": 122 }, { "epoch": 0.48616600790513836, "grad_norm": 0.8148254156112671, "learning_rate": 5.3139525976465675e-06, "loss": 1.2148, "step": 123 }, { "epoch": 0.4901185770750988, "grad_norm": 0.7850709557533264, "learning_rate": 5.251221590898848e-06, "loss": 1.1715, "step": 124 }, { "epoch": 0.49407114624505927, "grad_norm": 0.8794577717781067, "learning_rate": 5.188450913349674e-06, "loss": 1.1467, "step": 125 }, { "epoch": 0.4980237154150198, "grad_norm": 0.8375518321990967, "learning_rate": 5.1256504772166885e-06, "loss": 1.2558, "step": 126 }, { "epoch": 0.5019762845849802, "grad_norm": 0.7941477298736572, "learning_rate": 5.062830199416764e-06, "loss": 1.0128, "step": 127 }, { "epoch": 0.5059288537549407, "grad_norm": 0.9250499606132507, "learning_rate": 5e-06, "loss": 1.2622, "step": 128 }, { "epoch": 0.5098814229249012, "grad_norm": 0.9209011793136597, "learning_rate": 4.937169800583237e-06, "loss": 1.3989, "step": 129 }, { "epoch": 0.5138339920948617, "grad_norm": 0.8219020962715149, "learning_rate": 4.874349522783313e-06, "loss": 1.1264, "step": 130 }, { "epoch": 0.5138339920948617, "eval_loss": 1.1763263940811157, "eval_runtime": 0.6328, "eval_samples_per_second": 33.184, "eval_steps_per_second": 4.741, "step": 130 }, { "epoch": 0.5177865612648221, "grad_norm": 0.9172279238700867, "learning_rate": 4.811549086650327e-06, "loss": 1.2868, "step": 131 }, { "epoch": 0.5217391304347826, "grad_norm": 0.8297377228736877, "learning_rate": 4.748778409101153e-06, "loss": 1.221, "step": 132 }, { "epoch": 0.525691699604743, "grad_norm": 0.8089572787284851, "learning_rate": 4.686047402353433e-06, "loss": 1.2466, "step": 133 }, { "epoch": 0.5296442687747036, "grad_norm": 0.8571174144744873, "learning_rate": 4.6233659723603374e-06, "loss": 1.3095, "step": 134 }, { "epoch": 0.5335968379446641, "grad_norm": 0.9645123481750488, "learning_rate": 4.560744017246284e-06, "loss": 1.4281, "step": 135 }, { "epoch": 0.5375494071146245, "grad_norm": 0.8215659856796265, "learning_rate": 4.4981914257439254e-06, "loss": 1.1636, "step": 136 }, { "epoch": 0.541501976284585, "grad_norm": 0.8466573357582092, "learning_rate": 4.4357180756325915e-06, "loss": 1.2409, "step": 137 }, { "epoch": 0.5454545454545454, "grad_norm": 0.8330338597297668, "learning_rate": 4.373333832178478e-06, "loss": 1.1857, "step": 138 }, { "epoch": 0.549407114624506, "grad_norm": 0.8727869987487793, "learning_rate": 4.31104854657681e-06, "loss": 1.1947, "step": 139 }, { "epoch": 0.5533596837944664, "grad_norm": 0.8822157979011536, "learning_rate": 4.248872054396215e-06, "loss": 1.1367, "step": 140 }, { "epoch": 0.5573122529644269, "grad_norm": 0.876277506351471, "learning_rate": 4.186814174025582e-06, "loss": 1.1754, "step": 141 }, { "epoch": 0.5612648221343873, "grad_norm": 0.8091803789138794, "learning_rate": 4.124884705123619e-06, "loss": 1.1789, "step": 142 }, { "epoch": 0.5652173913043478, "grad_norm": 0.9175235629081726, "learning_rate": 4.063093427071376e-06, "loss": 1.2189, "step": 143 }, { "epoch": 0.5691699604743083, "grad_norm": 0.8601986765861511, "learning_rate": 4.001450097427965e-06, "loss": 1.3021, "step": 144 }, { "epoch": 0.5731225296442688, "grad_norm": 0.821972668170929, "learning_rate": 3.939964450389728e-06, "loss": 1.2536, "step": 145 }, { "epoch": 0.5770750988142292, "grad_norm": 0.8965577483177185, "learning_rate": 3.8786461952530955e-06, "loss": 1.2669, "step": 146 }, { "epoch": 0.5810276679841897, "grad_norm": 0.8901940584182739, "learning_rate": 3.817505014881378e-06, "loss": 1.2246, "step": 147 }, { "epoch": 0.5849802371541502, "grad_norm": 0.8637762069702148, "learning_rate": 3.756550564175727e-06, "loss": 1.2705, "step": 148 }, { "epoch": 0.5889328063241107, "grad_norm": 0.8907751441001892, "learning_rate": 3.695792468550517e-06, "loss": 1.3191, "step": 149 }, { "epoch": 0.5928853754940712, "grad_norm": 0.9000714421272278, "learning_rate": 3.635240322413375e-06, "loss": 1.3388, "step": 150 }, { "epoch": 0.5968379446640316, "grad_norm": 0.8116331696510315, "learning_rate": 3.5749036876501196e-06, "loss": 1.2169, "step": 151 }, { "epoch": 0.6007905138339921, "grad_norm": 0.9376166462898254, "learning_rate": 3.5147920921148267e-06, "loss": 1.4159, "step": 152 }, { "epoch": 0.6047430830039525, "grad_norm": 0.8815961480140686, "learning_rate": 3.4549150281252635e-06, "loss": 1.3918, "step": 153 }, { "epoch": 0.6086956521739131, "grad_norm": 0.8155955672264099, "learning_rate": 3.3952819509639534e-06, "loss": 1.0842, "step": 154 }, { "epoch": 0.6126482213438735, "grad_norm": 0.8554919958114624, "learning_rate": 3.3359022773850673e-06, "loss": 1.3536, "step": 155 }, { "epoch": 0.616600790513834, "grad_norm": 0.8661870956420898, "learning_rate": 3.2767853841274154e-06, "loss": 1.1721, "step": 156 }, { "epoch": 0.616600790513834, "eval_loss": 1.1584229469299316, "eval_runtime": 0.6329, "eval_samples_per_second": 33.182, "eval_steps_per_second": 4.74, "step": 156 }, { "epoch": 0.6205533596837944, "grad_norm": 0.8114671111106873, "learning_rate": 3.217940606433747e-06, "loss": 1.0664, "step": 157 }, { "epoch": 0.6245059288537549, "grad_norm": 0.863220751285553, "learning_rate": 3.1593772365766107e-06, "loss": 1.2761, "step": 158 }, { "epoch": 0.6284584980237155, "grad_norm": 0.8586646318435669, "learning_rate": 3.1011045223909954e-06, "loss": 1.2633, "step": 159 }, { "epoch": 0.6324110671936759, "grad_norm": 0.8601418137550354, "learning_rate": 3.043131665813988e-06, "loss": 1.2919, "step": 160 }, { "epoch": 0.6363636363636364, "grad_norm": 0.8035183548927307, "learning_rate": 2.9854678214316875e-06, "loss": 1.1463, "step": 161 }, { "epoch": 0.6403162055335968, "grad_norm": 0.8140465021133423, "learning_rate": 2.92812209503358e-06, "loss": 1.0966, "step": 162 }, { "epoch": 0.6442687747035574, "grad_norm": 0.8538705706596375, "learning_rate": 2.871103542174637e-06, "loss": 1.2748, "step": 163 }, { "epoch": 0.6482213438735178, "grad_norm": 0.8856031894683838, "learning_rate": 2.814421166745337e-06, "loss": 1.3632, "step": 164 }, { "epoch": 0.6521739130434783, "grad_norm": 0.8147869110107422, "learning_rate": 2.7580839195498397e-06, "loss": 1.1984, "step": 165 }, { "epoch": 0.6561264822134387, "grad_norm": 0.9556436538696289, "learning_rate": 2.7021006968925613e-06, "loss": 1.5325, "step": 166 }, { "epoch": 0.6600790513833992, "grad_norm": 0.8936392664909363, "learning_rate": 2.646480339173337e-06, "loss": 1.4192, "step": 167 }, { "epoch": 0.6640316205533597, "grad_norm": 0.8458386063575745, "learning_rate": 2.5912316294914232e-06, "loss": 1.2821, "step": 168 }, { "epoch": 0.6679841897233202, "grad_norm": 0.9311079382896423, "learning_rate": 2.536363292258543e-06, "loss": 1.3371, "step": 169 }, { "epoch": 0.6719367588932806, "grad_norm": 0.8807729482650757, "learning_rate": 2.4818839918211963e-06, "loss": 1.2789, "step": 170 }, { "epoch": 0.6758893280632411, "grad_norm": 0.8326642513275146, "learning_rate": 2.4278023310924676e-06, "loss": 1.2545, "step": 171 }, { "epoch": 0.6798418972332015, "grad_norm": 0.7561216950416565, "learning_rate": 2.3741268501935212e-06, "loss": 0.9872, "step": 172 }, { "epoch": 0.6837944664031621, "grad_norm": 0.9049252867698669, "learning_rate": 2.320866025105016e-06, "loss": 1.0876, "step": 173 }, { "epoch": 0.6877470355731226, "grad_norm": 0.8050791621208191, "learning_rate": 2.268028266328655e-06, "loss": 1.2257, "step": 174 }, { "epoch": 0.691699604743083, "grad_norm": 0.7893306016921997, "learning_rate": 2.2156219175590623e-06, "loss": 1.1499, "step": 175 }, { "epoch": 0.6956521739130435, "grad_norm": 0.8402442336082458, "learning_rate": 2.1636552543662187e-06, "loss": 1.3324, "step": 176 }, { "epoch": 0.6996047430830039, "grad_norm": 0.847458004951477, "learning_rate": 2.112136482888663e-06, "loss": 1.3478, "step": 177 }, { "epoch": 0.7035573122529645, "grad_norm": 0.8522107601165771, "learning_rate": 2.061073738537635e-06, "loss": 1.2586, "step": 178 }, { "epoch": 0.7075098814229249, "grad_norm": 0.8222721219062805, "learning_rate": 2.0104750847124075e-06, "loss": 1.2455, "step": 179 }, { "epoch": 0.7114624505928854, "grad_norm": 0.8545589447021484, "learning_rate": 1.9603485115269743e-06, "loss": 1.2892, "step": 180 }, { "epoch": 0.7154150197628458, "grad_norm": 0.7990689873695374, "learning_rate": 1.910701934548329e-06, "loss": 1.1216, "step": 181 }, { "epoch": 0.7193675889328063, "grad_norm": 0.8381747007369995, "learning_rate": 1.8615431935464984e-06, "loss": 1.2738, "step": 182 }, { "epoch": 0.7193675889328063, "eval_loss": 1.1467477083206177, "eval_runtime": 0.6308, "eval_samples_per_second": 33.29, "eval_steps_per_second": 4.756, "step": 182 }, { "epoch": 0.7233201581027668, "grad_norm": 0.8528211712837219, "learning_rate": 1.8128800512565514e-06, "loss": 1.3237, "step": 183 }, { "epoch": 0.7272727272727273, "grad_norm": 0.8386929035186768, "learning_rate": 1.7647201921527802e-06, "loss": 1.1454, "step": 184 }, { "epoch": 0.7312252964426877, "grad_norm": 0.9228261709213257, "learning_rate": 1.7170712212352187e-06, "loss": 1.5128, "step": 185 }, { "epoch": 0.7351778656126482, "grad_norm": 0.763982892036438, "learning_rate": 1.6699406628287423e-06, "loss": 0.9779, "step": 186 }, { "epoch": 0.7391304347826086, "grad_norm": 0.9443770051002502, "learning_rate": 1.6233359593948777e-06, "loss": 0.9658, "step": 187 }, { "epoch": 0.7430830039525692, "grad_norm": 0.8337656855583191, "learning_rate": 1.5772644703565564e-06, "loss": 1.2752, "step": 188 }, { "epoch": 0.7470355731225297, "grad_norm": 0.8052669763565063, "learning_rate": 1.531733470935976e-06, "loss": 1.1755, "step": 189 }, { "epoch": 0.7509881422924901, "grad_norm": 0.8237268328666687, "learning_rate": 1.4867501510057548e-06, "loss": 1.1529, "step": 190 }, { "epoch": 0.7549407114624506, "grad_norm": 0.8448726534843445, "learning_rate": 1.4423216139535735e-06, "loss": 1.2361, "step": 191 }, { "epoch": 0.758893280632411, "grad_norm": 0.8243027329444885, "learning_rate": 1.3984548755604655e-06, "loss": 1.2842, "step": 192 }, { "epoch": 0.7628458498023716, "grad_norm": 0.8658849000930786, "learning_rate": 1.3551568628929434e-06, "loss": 1.256, "step": 193 }, { "epoch": 0.766798418972332, "grad_norm": 0.8200253248214722, "learning_rate": 1.312434413209131e-06, "loss": 1.1356, "step": 194 }, { "epoch": 0.7707509881422925, "grad_norm": 0.8177831768989563, "learning_rate": 1.2702942728790897e-06, "loss": 1.151, "step": 195 }, { "epoch": 0.7747035573122529, "grad_norm": 0.8590410351753235, "learning_rate": 1.2287430963194807e-06, "loss": 1.3267, "step": 196 }, { "epoch": 0.7786561264822134, "grad_norm": 0.8247159719467163, "learning_rate": 1.18778744494276e-06, "loss": 1.1943, "step": 197 }, { "epoch": 0.782608695652174, "grad_norm": 0.9330562949180603, "learning_rate": 1.1474337861210543e-06, "loss": 1.4392, "step": 198 }, { "epoch": 0.7865612648221344, "grad_norm": 0.8507537841796875, "learning_rate": 1.1076884921648834e-06, "loss": 1.2678, "step": 199 }, { "epoch": 0.7905138339920948, "grad_norm": 0.8290367126464844, "learning_rate": 1.0685578393169054e-06, "loss": 1.2738, "step": 200 }, { "epoch": 0.7944664031620553, "grad_norm": 0.8783449530601501, "learning_rate": 1.0300480067608232e-06, "loss": 1.4097, "step": 201 }, { "epoch": 0.7984189723320159, "grad_norm": 0.8614851236343384, "learning_rate": 9.921650756456164e-07, "loss": 1.1463, "step": 202 }, { "epoch": 0.8023715415019763, "grad_norm": 0.8946048617362976, "learning_rate": 9.549150281252633e-07, "loss": 1.3288, "step": 203 }, { "epoch": 0.8063241106719368, "grad_norm": 0.8253918290138245, "learning_rate": 9.183037464140804e-07, "loss": 1.1852, "step": 204 }, { "epoch": 0.8102766798418972, "grad_norm": 0.8394426107406616, "learning_rate": 8.823370118578628e-07, "loss": 1.2328, "step": 205 }, { "epoch": 0.8142292490118577, "grad_norm": 0.7891396880149841, "learning_rate": 8.470205040209362e-07, "loss": 1.2409, "step": 206 }, { "epoch": 0.8181818181818182, "grad_norm": 0.8969730138778687, "learning_rate": 8.123597997892918e-07, "loss": 1.4195, "step": 207 }, { "epoch": 0.8221343873517787, "grad_norm": 0.7859252095222473, "learning_rate": 7.783603724899258e-07, "loss": 1.1729, "step": 208 }, { "epoch": 0.8221343873517787, "eval_loss": 1.1414591073989868, "eval_runtime": 0.63, "eval_samples_per_second": 33.334, "eval_steps_per_second": 4.762, "step": 208 }, { "epoch": 0.8260869565217391, "grad_norm": 0.8004709482192993, "learning_rate": 7.450275910265415e-07, "loss": 1.1787, "step": 209 }, { "epoch": 0.8300395256916996, "grad_norm": 0.8863723874092102, "learning_rate": 7.123667190317396e-07, "loss": 1.3486, "step": 210 }, { "epoch": 0.83399209486166, "grad_norm": 0.7677893042564392, "learning_rate": 6.803829140358237e-07, "loss": 1.2067, "step": 211 }, { "epoch": 0.8379446640316206, "grad_norm": 0.7935658097267151, "learning_rate": 6.490812266523716e-07, "loss": 1.0929, "step": 212 }, { "epoch": 0.841897233201581, "grad_norm": 0.7727168798446655, "learning_rate": 6.184665997806832e-07, "loss": 1.0544, "step": 213 }, { "epoch": 0.8458498023715415, "grad_norm": 0.8302650451660156, "learning_rate": 5.885438678252342e-07, "loss": 1.2409, "step": 214 }, { "epoch": 0.849802371541502, "grad_norm": 0.8888508081436157, "learning_rate": 5.593177559322776e-07, "loss": 1.3816, "step": 215 }, { "epoch": 0.8537549407114624, "grad_norm": 0.76105135679245, "learning_rate": 5.307928792436812e-07, "loss": 1.0972, "step": 216 }, { "epoch": 0.857707509881423, "grad_norm": 0.8869984149932861, "learning_rate": 5.029737421681446e-07, "loss": 1.3786, "step": 217 }, { "epoch": 0.8616600790513834, "grad_norm": 0.8762255907058716, "learning_rate": 4.758647376699033e-07, "loss": 1.3592, "step": 218 }, { "epoch": 0.8656126482213439, "grad_norm": 0.8182483911514282, "learning_rate": 4.494701465750217e-07, "loss": 1.3265, "step": 219 }, { "epoch": 0.8695652173913043, "grad_norm": 0.8035054206848145, "learning_rate": 4.237941368954124e-07, "loss": 1.1705, "step": 220 }, { "epoch": 0.8735177865612648, "grad_norm": 0.8334324359893799, "learning_rate": 3.9884076317064813e-07, "loss": 1.2725, "step": 221 }, { "epoch": 0.8774703557312253, "grad_norm": 0.8486202955245972, "learning_rate": 3.7461396582771035e-07, "loss": 1.2205, "step": 222 }, { "epoch": 0.8814229249011858, "grad_norm": 0.8365004658699036, "learning_rate": 3.511175705587433e-07, "loss": 1.2246, "step": 223 }, { "epoch": 0.8853754940711462, "grad_norm": 0.7247976660728455, "learning_rate": 3.283552877169399e-07, "loss": 1.0818, "step": 224 }, { "epoch": 0.8893280632411067, "grad_norm": 0.9177906513214111, "learning_rate": 3.0633071173062966e-07, "loss": 1.328, "step": 225 }, { "epoch": 0.8932806324110671, "grad_norm": 0.8105781078338623, "learning_rate": 2.850473205356774e-07, "loss": 1.0848, "step": 226 }, { "epoch": 0.8972332015810277, "grad_norm": 0.8612114191055298, "learning_rate": 2.6450847502627883e-07, "loss": 1.3192, "step": 227 }, { "epoch": 0.9011857707509882, "grad_norm": 0.8288993239402771, "learning_rate": 2.447174185242324e-07, "loss": 1.2458, "step": 228 }, { "epoch": 0.9051383399209486, "grad_norm": 0.8506494760513306, "learning_rate": 2.2567727626678527e-07, "loss": 1.2568, "step": 229 }, { "epoch": 0.9090909090909091, "grad_norm": 0.9084119200706482, "learning_rate": 2.0739105491312028e-07, "loss": 1.4233, "step": 230 }, { "epoch": 0.9130434782608695, "grad_norm": 0.9303674101829529, "learning_rate": 1.8986164206957037e-07, "loss": 1.3435, "step": 231 }, { "epoch": 0.9169960474308301, "grad_norm": 0.8007397055625916, "learning_rate": 1.7309180583363062e-07, "loss": 1.1919, "step": 232 }, { "epoch": 0.9209486166007905, "grad_norm": 0.7730187177658081, "learning_rate": 1.5708419435684463e-07, "loss": 1.0833, "step": 233 }, { "epoch": 0.924901185770751, "grad_norm": 0.9434295296669006, "learning_rate": 1.4184133542663014e-07, "loss": 1.4446, "step": 234 }, { "epoch": 0.924901185770751, "eval_loss": 1.1398719549179077, "eval_runtime": 0.6371, "eval_samples_per_second": 32.96, "eval_steps_per_second": 4.709, "step": 234 }, { "epoch": 0.9288537549407114, "grad_norm": 0.8378840684890747, "learning_rate": 1.2736563606711384e-07, "loss": 1.3058, "step": 235 }, { "epoch": 0.932806324110672, "grad_norm": 0.7894282341003418, "learning_rate": 1.136593821590326e-07, "loss": 1.1822, "step": 236 }, { "epoch": 0.9367588932806324, "grad_norm": 0.8614757657051086, "learning_rate": 1.007247380787657e-07, "loss": 1.2477, "step": 237 }, { "epoch": 0.9407114624505929, "grad_norm": 0.7437376976013184, "learning_rate": 8.856374635655696e-08, "loss": 0.9579, "step": 238 }, { "epoch": 0.9446640316205533, "grad_norm": 0.8177477717399597, "learning_rate": 7.717832735397335e-08, "loss": 1.2895, "step": 239 }, { "epoch": 0.9486166007905138, "grad_norm": 0.8371565937995911, "learning_rate": 6.657027896065982e-08, "loss": 1.2901, "step": 240 }, { "epoch": 0.9525691699604744, "grad_norm": 0.8062224388122559, "learning_rate": 5.674127631043025e-08, "loss": 1.1185, "step": 241 }, { "epoch": 0.9565217391304348, "grad_norm": 0.782124936580658, "learning_rate": 4.769287151674407e-08, "loss": 1.1315, "step": 242 }, { "epoch": 0.9604743083003953, "grad_norm": 0.8169525265693665, "learning_rate": 3.9426493427611177e-08, "loss": 1.3511, "step": 243 }, { "epoch": 0.9644268774703557, "grad_norm": 0.791191577911377, "learning_rate": 3.194344739995803e-08, "loss": 1.2501, "step": 244 }, { "epoch": 0.9683794466403162, "grad_norm": 0.7813810706138611, "learning_rate": 2.5244915093499134e-08, "loss": 1.1408, "step": 245 }, { "epoch": 0.9723320158102767, "grad_norm": 0.8077353239059448, "learning_rate": 1.9331954284137476e-08, "loss": 1.2169, "step": 246 }, { "epoch": 0.9762845849802372, "grad_norm": 0.8798792362213135, "learning_rate": 1.4205498696930332e-08, "loss": 1.4156, "step": 247 }, { "epoch": 0.9802371541501976, "grad_norm": 0.7928077578544617, "learning_rate": 9.866357858642206e-09, "loss": 1.1922, "step": 248 }, { "epoch": 0.9841897233201581, "grad_norm": 0.8758398294448853, "learning_rate": 6.315216969912663e-09, "loss": 1.4098, "step": 249 }, { "epoch": 0.9881422924901185, "grad_norm": 0.8929165005683899, "learning_rate": 3.5526367970539765e-09, "loss": 1.3601, "step": 250 }, { "epoch": 0.9920948616600791, "grad_norm": 0.9001562595367432, "learning_rate": 1.5790535835003006e-09, "loss": 1.2039, "step": 251 }, { "epoch": 0.9960474308300395, "grad_norm": 0.8405194878578186, "learning_rate": 3.9477898091944135e-10, "loss": 1.1514, "step": 252 }, { "epoch": 1.0, "grad_norm": 0.8833491206169128, "learning_rate": 0.0, "loss": 1.2887, "step": 253 }, { "epoch": 1.0, "step": 253, "total_flos": 8.443588829301965e+16, "train_loss": 1.3333067849219553, "train_runtime": 1520.8803, "train_samples_per_second": 1.329, "train_steps_per_second": 0.166 } ], "logging_steps": 1, "max_steps": 253, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 253, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 8.443588829301965e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }