{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.146509341199607, "eval_steps": 100, "global_step": 400, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.03933136676499508, "grad_norm": 0.481609046459198, "learning_rate": 0.00015, "loss": 2.0722, "step": 5 }, { "epoch": 0.07866273352999016, "grad_norm": 0.15720224380493164, "learning_rate": 0.0003, "loss": 1.4825, "step": 10 }, { "epoch": 0.11799410029498525, "grad_norm": 0.06716315448284149, "learning_rate": 0.00029759999999999997, "loss": 1.3333, "step": 15 }, { "epoch": 0.15732546705998032, "grad_norm": 0.06133478134870529, "learning_rate": 0.00029519999999999997, "loss": 1.2341, "step": 20 }, { "epoch": 0.19665683382497542, "grad_norm": 0.07264667749404907, "learning_rate": 0.00029279999999999996, "loss": 1.1756, "step": 25 }, { "epoch": 0.2359882005899705, "grad_norm": 0.07928217202425003, "learning_rate": 0.00029039999999999996, "loss": 1.1197, "step": 30 }, { "epoch": 0.2753195673549656, "grad_norm": 0.09420346468687057, "learning_rate": 0.00028799999999999995, "loss": 1.0834, "step": 35 }, { "epoch": 0.31465093411996065, "grad_norm": 0.0862259566783905, "learning_rate": 0.00028559999999999995, "loss": 1.044, "step": 40 }, { "epoch": 0.35398230088495575, "grad_norm": 0.09086894243955612, "learning_rate": 0.00028319999999999994, "loss": 1.0205, "step": 45 }, { "epoch": 0.39331366764995085, "grad_norm": 0.08469890058040619, "learning_rate": 0.0002808, "loss": 0.9798, "step": 50 }, { "epoch": 0.4326450344149459, "grad_norm": 0.10012397915124893, "learning_rate": 0.0002784, "loss": 0.9811, "step": 55 }, { "epoch": 0.471976401179941, "grad_norm": 0.08633492887020111, "learning_rate": 0.000276, "loss": 0.9556, "step": 60 }, { "epoch": 0.511307767944936, "grad_norm": 0.09879346191883087, "learning_rate": 0.0002736, "loss": 0.9446, "step": 65 }, { "epoch": 0.5506391347099312, "grad_norm": 0.08795857429504395, "learning_rate": 0.0002712, "loss": 0.9228, "step": 70 }, { "epoch": 0.5899705014749262, "grad_norm": 0.0837111845612526, "learning_rate": 0.0002688, "loss": 0.9279, "step": 75 }, { "epoch": 0.6293018682399213, "grad_norm": 0.08551318198442459, "learning_rate": 0.00026639999999999997, "loss": 0.9267, "step": 80 }, { "epoch": 0.6686332350049164, "grad_norm": 0.08481767773628235, "learning_rate": 0.00026399999999999997, "loss": 0.9082, "step": 85 }, { "epoch": 0.7079646017699115, "grad_norm": 0.100365050137043, "learning_rate": 0.00026159999999999996, "loss": 0.9028, "step": 90 }, { "epoch": 0.7472959685349065, "grad_norm": 0.08463772386312485, "learning_rate": 0.00025919999999999996, "loss": 0.8866, "step": 95 }, { "epoch": 0.7866273352999017, "grad_norm": 0.09628409892320633, "learning_rate": 0.00025679999999999995, "loss": 0.8787, "step": 100 }, { "epoch": 0.7866273352999017, "eval_loss": 0.8853636980056763, "eval_runtime": 24.3719, "eval_samples_per_second": 6.729, "eval_steps_per_second": 0.862, "step": 100 }, { "epoch": 0.8259587020648967, "grad_norm": 0.08835043758153915, "learning_rate": 0.00025439999999999995, "loss": 0.8786, "step": 105 }, { "epoch": 0.8652900688298918, "grad_norm": 0.09190791845321655, "learning_rate": 0.00025199999999999995, "loss": 0.8693, "step": 110 }, { "epoch": 0.904621435594887, "grad_norm": 0.08965795487165451, "learning_rate": 0.00024959999999999994, "loss": 0.8772, "step": 115 }, { "epoch": 0.943952802359882, "grad_norm": 0.09055910259485245, "learning_rate": 0.0002472, "loss": 0.867, "step": 120 }, { "epoch": 0.983284169124877, "grad_norm": 0.09172637015581131, "learning_rate": 0.0002448, "loss": 0.8536, "step": 125 }, { "epoch": 1.022615535889872, "grad_norm": 0.10374542325735092, "learning_rate": 0.00024239999999999998, "loss": 0.9888, "step": 130 }, { "epoch": 1.0619469026548674, "grad_norm": 0.08842068910598755, "learning_rate": 0.00023999999999999998, "loss": 0.8443, "step": 135 }, { "epoch": 1.1012782694198624, "grad_norm": 0.0736837387084961, "learning_rate": 0.0002376, "loss": 0.8457, "step": 140 }, { "epoch": 1.1406096361848574, "grad_norm": 0.07575016468763351, "learning_rate": 0.0002352, "loss": 0.8335, "step": 145 }, { "epoch": 1.1799410029498525, "grad_norm": 0.07092955708503723, "learning_rate": 0.0002328, "loss": 0.8246, "step": 150 }, { "epoch": 1.2192723697148475, "grad_norm": 0.077423095703125, "learning_rate": 0.0002304, "loss": 0.823, "step": 155 }, { "epoch": 1.2586037364798428, "grad_norm": 0.07389391213655472, "learning_rate": 0.00022799999999999999, "loss": 0.819, "step": 160 }, { "epoch": 1.2979351032448379, "grad_norm": 0.08229434490203857, "learning_rate": 0.00022559999999999998, "loss": 0.8181, "step": 165 }, { "epoch": 1.337266470009833, "grad_norm": 0.07665972411632538, "learning_rate": 0.00022319999999999998, "loss": 0.8118, "step": 170 }, { "epoch": 1.376597836774828, "grad_norm": 0.09001573175191879, "learning_rate": 0.00022079999999999997, "loss": 0.8157, "step": 175 }, { "epoch": 1.415929203539823, "grad_norm": 0.07965826243162155, "learning_rate": 0.00021839999999999997, "loss": 0.8111, "step": 180 }, { "epoch": 1.455260570304818, "grad_norm": 0.08642959594726562, "learning_rate": 0.00021599999999999996, "loss": 0.8003, "step": 185 }, { "epoch": 1.494591937069813, "grad_norm": 0.0749087929725647, "learning_rate": 0.00021359999999999996, "loss": 0.7975, "step": 190 }, { "epoch": 1.5339233038348081, "grad_norm": 0.08575734496116638, "learning_rate": 0.00021119999999999996, "loss": 0.7888, "step": 195 }, { "epoch": 1.5732546705998034, "grad_norm": 0.0887129157781601, "learning_rate": 0.00020879999999999998, "loss": 0.7857, "step": 200 }, { "epoch": 1.5732546705998034, "eval_loss": 0.8026237487792969, "eval_runtime": 24.2397, "eval_samples_per_second": 6.766, "eval_steps_per_second": 0.866, "step": 200 }, { "epoch": 1.6125860373647984, "grad_norm": 0.0926935002207756, "learning_rate": 0.00020639999999999998, "loss": 0.7877, "step": 205 }, { "epoch": 1.6519174041297935, "grad_norm": 0.08537031710147858, "learning_rate": 0.000204, "loss": 0.7767, "step": 210 }, { "epoch": 1.6912487708947888, "grad_norm": 0.0766814798116684, "learning_rate": 0.0002016, "loss": 0.785, "step": 215 }, { "epoch": 1.7305801376597838, "grad_norm": 0.08394207805395126, "learning_rate": 0.0001992, "loss": 0.7832, "step": 220 }, { "epoch": 1.7699115044247788, "grad_norm": 0.0813060775399208, "learning_rate": 0.00019679999999999999, "loss": 0.7766, "step": 225 }, { "epoch": 1.809242871189774, "grad_norm": 0.08242856711149216, "learning_rate": 0.00019439999999999998, "loss": 0.7775, "step": 230 }, { "epoch": 1.848574237954769, "grad_norm": 0.07610878348350525, "learning_rate": 0.00019199999999999998, "loss": 0.7736, "step": 235 }, { "epoch": 1.887905604719764, "grad_norm": 0.08326178044080734, "learning_rate": 0.00018959999999999997, "loss": 0.7753, "step": 240 }, { "epoch": 1.927236971484759, "grad_norm": 0.09425383061170578, "learning_rate": 0.0001872, "loss": 0.7577, "step": 245 }, { "epoch": 1.966568338249754, "grad_norm": 0.08694498240947723, "learning_rate": 0.0001848, "loss": 0.7606, "step": 250 }, { "epoch": 2.005899705014749, "grad_norm": 0.22805309295654297, "learning_rate": 0.0001824, "loss": 0.8871, "step": 255 }, { "epoch": 2.045231071779744, "grad_norm": 0.09610473364591599, "learning_rate": 0.00017999999999999998, "loss": 0.7315, "step": 260 }, { "epoch": 2.084562438544739, "grad_norm": 0.09666857868432999, "learning_rate": 0.00017759999999999998, "loss": 0.7315, "step": 265 }, { "epoch": 2.1238938053097347, "grad_norm": 0.09328849613666534, "learning_rate": 0.00017519999999999998, "loss": 0.7344, "step": 270 }, { "epoch": 2.1632251720747298, "grad_norm": 0.08137473464012146, "learning_rate": 0.00017279999999999997, "loss": 0.7347, "step": 275 }, { "epoch": 2.202556538839725, "grad_norm": 0.08166103810071945, "learning_rate": 0.00017039999999999997, "loss": 0.7281, "step": 280 }, { "epoch": 2.24188790560472, "grad_norm": 0.08074019104242325, "learning_rate": 0.000168, "loss": 0.7345, "step": 285 }, { "epoch": 2.281219272369715, "grad_norm": 0.08479057997465134, "learning_rate": 0.0001656, "loss": 0.726, "step": 290 }, { "epoch": 2.32055063913471, "grad_norm": 0.08091601729393005, "learning_rate": 0.0001632, "loss": 0.7184, "step": 295 }, { "epoch": 2.359882005899705, "grad_norm": 0.08470489084720612, "learning_rate": 0.0001608, "loss": 0.7233, "step": 300 }, { "epoch": 2.359882005899705, "eval_loss": 0.7612683176994324, "eval_runtime": 24.27, "eval_samples_per_second": 6.757, "eval_steps_per_second": 0.865, "step": 300 }, { "epoch": 2.3992133726647, "grad_norm": 0.08677177131175995, "learning_rate": 0.0001584, "loss": 0.721, "step": 305 }, { "epoch": 2.438544739429695, "grad_norm": 0.08474377542734146, "learning_rate": 0.000156, "loss": 0.7141, "step": 310 }, { "epoch": 2.47787610619469, "grad_norm": 0.08565227687358856, "learning_rate": 0.0001536, "loss": 0.7173, "step": 315 }, { "epoch": 2.5172074729596856, "grad_norm": 0.08714301139116287, "learning_rate": 0.0001512, "loss": 0.7274, "step": 320 }, { "epoch": 2.5565388397246807, "grad_norm": 0.0934271439909935, "learning_rate": 0.00014879999999999998, "loss": 0.7263, "step": 325 }, { "epoch": 2.5958702064896757, "grad_norm": 0.08581375330686569, "learning_rate": 0.00014639999999999998, "loss": 0.7248, "step": 330 }, { "epoch": 2.6352015732546707, "grad_norm": 0.08378680050373077, "learning_rate": 0.00014399999999999998, "loss": 0.721, "step": 335 }, { "epoch": 2.674532940019666, "grad_norm": 0.08449660986661911, "learning_rate": 0.00014159999999999997, "loss": 0.7156, "step": 340 }, { "epoch": 2.713864306784661, "grad_norm": 0.08646751940250397, "learning_rate": 0.0001392, "loss": 0.7094, "step": 345 }, { "epoch": 2.753195673549656, "grad_norm": 0.08911272883415222, "learning_rate": 0.0001368, "loss": 0.709, "step": 350 }, { "epoch": 2.792527040314651, "grad_norm": 0.0970829427242279, "learning_rate": 0.0001344, "loss": 0.7107, "step": 355 }, { "epoch": 2.831858407079646, "grad_norm": 0.0854572132229805, "learning_rate": 0.00013199999999999998, "loss": 0.7148, "step": 360 }, { "epoch": 2.871189773844641, "grad_norm": 0.08210612088441849, "learning_rate": 0.00012959999999999998, "loss": 0.7132, "step": 365 }, { "epoch": 2.910521140609636, "grad_norm": 0.0925467386841774, "learning_rate": 0.00012719999999999997, "loss": 0.7201, "step": 370 }, { "epoch": 2.949852507374631, "grad_norm": 0.09149914979934692, "learning_rate": 0.00012479999999999997, "loss": 0.7086, "step": 375 }, { "epoch": 2.989183874139626, "grad_norm": 0.0827464610338211, "learning_rate": 0.0001224, "loss": 0.7102, "step": 380 }, { "epoch": 3.0285152409046217, "grad_norm": 0.09861475974321365, "learning_rate": 0.00011999999999999999, "loss": 0.8086, "step": 385 }, { "epoch": 3.0678466076696167, "grad_norm": 0.09810496121644974, "learning_rate": 0.0001176, "loss": 0.6784, "step": 390 }, { "epoch": 3.1071779744346117, "grad_norm": 0.08657824248075485, "learning_rate": 0.0001152, "loss": 0.6818, "step": 395 }, { "epoch": 3.146509341199607, "grad_norm": 0.08861815184354782, "learning_rate": 0.00011279999999999999, "loss": 0.6755, "step": 400 }, { "epoch": 3.146509341199607, "eval_loss": 0.7408613562583923, "eval_runtime": 24.2895, "eval_samples_per_second": 6.752, "eval_steps_per_second": 0.865, "step": 400 } ], "logging_steps": 5, "max_steps": 635, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2.0292300729210634e+19, "train_batch_size": 4, "trial_name": null, "trial_params": null }