{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.9939045038943446, "eval_steps": 500, "global_step": 368, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.005418218760582459, "grad_norm": 11.441351738340396, "learning_rate": 5.405405405405406e-07, "loss": 2.4665, "step": 1 }, { "epoch": 0.027091093802912292, "grad_norm": 11.567929424436132, "learning_rate": 2.702702702702703e-06, "loss": 2.4607, "step": 5 }, { "epoch": 0.054182187605824585, "grad_norm": 2.1647507226825287, "learning_rate": 5.405405405405406e-06, "loss": 2.3924, "step": 10 }, { "epoch": 0.08127328140873688, "grad_norm": 1.0652926230951467, "learning_rate": 8.108108108108109e-06, "loss": 2.2869, "step": 15 }, { "epoch": 0.10836437521164917, "grad_norm": 0.6956953231915263, "learning_rate": 1.0810810810810812e-05, "loss": 2.2241, "step": 20 }, { "epoch": 0.13545546901456146, "grad_norm": 0.6368662485384925, "learning_rate": 1.3513513513513515e-05, "loss": 2.1943, "step": 25 }, { "epoch": 0.16254656281747376, "grad_norm": 0.52626831061368, "learning_rate": 1.6216216216216218e-05, "loss": 2.1796, "step": 30 }, { "epoch": 0.18963765662038604, "grad_norm": 0.5041342654725786, "learning_rate": 1.891891891891892e-05, "loss": 2.1739, "step": 35 }, { "epoch": 0.21672875042329834, "grad_norm": 0.531083346249148, "learning_rate": 1.9995946530314384e-05, "loss": 2.1573, "step": 40 }, { "epoch": 0.24381984422621064, "grad_norm": 0.4970950800288572, "learning_rate": 1.9971187226043746e-05, "loss": 2.1556, "step": 45 }, { "epoch": 0.2709109380291229, "grad_norm": 0.4930331306794567, "learning_rate": 1.9923976226947417e-05, "loss": 2.1516, "step": 50 }, { "epoch": 0.2980020318320352, "grad_norm": 0.4792958893497388, "learning_rate": 1.985441983600819e-05, "loss": 2.1416, "step": 55 }, { "epoch": 0.3250931256349475, "grad_norm": 0.4704908677461612, "learning_rate": 1.9762674670369757e-05, "loss": 2.1375, "step": 60 }, { "epoch": 0.3521842194378598, "grad_norm": 0.46530825034717316, "learning_rate": 1.9648947308688594e-05, "loss": 2.1328, "step": 65 }, { "epoch": 0.37927531324077207, "grad_norm": 0.47535573444890883, "learning_rate": 1.9513493825989664e-05, "loss": 2.1402, "step": 70 }, { "epoch": 0.4063664070436844, "grad_norm": 0.4759377502233077, "learning_rate": 1.9356619217073252e-05, "loss": 2.1332, "step": 75 }, { "epoch": 0.4334575008465967, "grad_norm": 0.4662210394921992, "learning_rate": 1.917867670977126e-05, "loss": 2.1335, "step": 80 }, { "epoch": 0.460548594649509, "grad_norm": 0.4757611076987449, "learning_rate": 1.8980066969599216e-05, "loss": 2.1258, "step": 85 }, { "epoch": 0.4876396884524213, "grad_norm": 0.45382312904649785, "learning_rate": 1.8761237197594945e-05, "loss": 2.1293, "step": 90 }, { "epoch": 0.5147307822553335, "grad_norm": 0.47488848221887603, "learning_rate": 1.852268012337514e-05, "loss": 2.1279, "step": 95 }, { "epoch": 0.5418218760582458, "grad_norm": 0.47098691848945873, "learning_rate": 1.8264932895677195e-05, "loss": 2.126, "step": 100 }, { "epoch": 0.5689129698611581, "grad_norm": 0.47815599863300057, "learning_rate": 1.798857587288445e-05, "loss": 2.1249, "step": 105 }, { "epoch": 0.5960040636640704, "grad_norm": 0.45607203288859494, "learning_rate": 1.769423131625808e-05, "loss": 2.1182, "step": 110 }, { "epoch": 0.6230951574669827, "grad_norm": 0.4566297150188363, "learning_rate": 1.738256198881809e-05, "loss": 2.1182, "step": 115 }, { "epoch": 0.650186251269895, "grad_norm": 0.4563332108229212, "learning_rate": 1.7054269663028232e-05, "loss": 2.1158, "step": 120 }, { "epoch": 0.6772773450728073, "grad_norm": 0.4719747539449419, "learning_rate": 1.6710093540645056e-05, "loss": 2.1124, "step": 125 }, { "epoch": 0.7043684388757196, "grad_norm": 0.45233465694057096, "learning_rate": 1.6350808588288964e-05, "loss": 2.1173, "step": 130 }, { "epoch": 0.731459532678632, "grad_norm": 0.4494520104287378, "learning_rate": 1.597722379248512e-05, "loss": 2.1147, "step": 135 }, { "epoch": 0.7585506264815441, "grad_norm": 0.44217741153395734, "learning_rate": 1.559018033810316e-05, "loss": 2.1158, "step": 140 }, { "epoch": 0.7856417202844564, "grad_norm": 0.43568954148921973, "learning_rate": 1.5190549714297303e-05, "loss": 2.1155, "step": 145 }, { "epoch": 0.8127328140873687, "grad_norm": 0.45615630991925854, "learning_rate": 1.4779231752211546e-05, "loss": 2.1162, "step": 150 }, { "epoch": 0.839823907890281, "grad_norm": 0.4607530478729013, "learning_rate": 1.4357152598868478e-05, "loss": 2.117, "step": 155 }, { "epoch": 0.8669150016931934, "grad_norm": 0.43874353467346544, "learning_rate": 1.3925262631803722e-05, "loss": 2.1047, "step": 160 }, { "epoch": 0.8940060954961057, "grad_norm": 0.4577418545526205, "learning_rate": 1.3484534319141592e-05, "loss": 2.1103, "step": 165 }, { "epoch": 0.921097189299018, "grad_norm": 0.44202574956917495, "learning_rate": 1.303596002993028e-05, "loss": 2.1117, "step": 170 }, { "epoch": 0.9481882831019303, "grad_norm": 0.4461244150173344, "learning_rate": 1.2580549799667034e-05, "loss": 2.1141, "step": 175 }, { "epoch": 0.9752793769048426, "grad_norm": 0.45529168414311305, "learning_rate": 1.2119329056044533e-05, "loss": 2.1046, "step": 180 }, { "epoch": 0.9969522519471724, "eval_loss": 2.1050913333892822, "eval_runtime": 76.5237, "eval_samples_per_second": 17.054, "eval_steps_per_second": 8.533, "step": 184 }, { "epoch": 1.0023704707077548, "grad_norm": 0.4809974366908029, "learning_rate": 1.165333631003928e-05, "loss": 2.0998, "step": 185 }, { "epoch": 1.029461564510667, "grad_norm": 0.48705866202596343, "learning_rate": 1.1183620817540985e-05, "loss": 2.0462, "step": 190 }, { "epoch": 1.0565526583135794, "grad_norm": 0.4868339560578216, "learning_rate": 1.0711240216788036e-05, "loss": 2.0529, "step": 195 }, { "epoch": 1.0836437521164917, "grad_norm": 0.4620568487887428, "learning_rate": 1.0237258146928849e-05, "loss": 2.0494, "step": 200 }, { "epoch": 1.110734845919404, "grad_norm": 0.4641633596897171, "learning_rate": 9.762741853071153e-06, "loss": 2.0467, "step": 205 }, { "epoch": 1.1378259397223163, "grad_norm": 0.48677368374631885, "learning_rate": 9.288759783211967e-06, "loss": 2.0482, "step": 210 }, { "epoch": 1.1649170335252286, "grad_norm": 0.46534164607337153, "learning_rate": 8.81637918245902e-06, "loss": 2.0469, "step": 215 }, { "epoch": 1.1920081273281409, "grad_norm": 0.46946874812594946, "learning_rate": 8.346663689960724e-06, "loss": 2.0445, "step": 220 }, { "epoch": 1.2190992211310532, "grad_norm": 0.4663999382035462, "learning_rate": 7.880670943955467e-06, "loss": 2.0481, "step": 225 }, { "epoch": 1.2461903149339655, "grad_norm": 0.4457099195715021, "learning_rate": 7.419450200332965e-06, "loss": 2.0529, "step": 230 }, { "epoch": 1.2732814087368778, "grad_norm": 0.46304032166442355, "learning_rate": 6.964039970069722e-06, "loss": 2.0467, "step": 235 }, { "epoch": 1.30037250253979, "grad_norm": 0.4449406613479281, "learning_rate": 6.515465680858412e-06, "loss": 2.0486, "step": 240 }, { "epoch": 1.3274635963427024, "grad_norm": 0.4389522489189322, "learning_rate": 6.074737368196279e-06, "loss": 2.0516, "step": 245 }, { "epoch": 1.3545546901456147, "grad_norm": 0.43429376447812934, "learning_rate": 5.642847401131526e-06, "loss": 2.0398, "step": 250 }, { "epoch": 1.381645783948527, "grad_norm": 0.42907764023330297, "learning_rate": 5.220768247788458e-06, "loss": 2.0469, "step": 255 }, { "epoch": 1.4087368777514393, "grad_norm": 0.43261082208262897, "learning_rate": 4.809450285702697e-06, "loss": 2.049, "step": 260 }, { "epoch": 1.4358279715543514, "grad_norm": 0.4495153435008709, "learning_rate": 4.409819661896839e-06, "loss": 2.0453, "step": 265 }, { "epoch": 1.462919065357264, "grad_norm": 0.44801407055899056, "learning_rate": 4.022776207514885e-06, "loss": 2.0401, "step": 270 }, { "epoch": 1.490010159160176, "grad_norm": 0.43760507616111505, "learning_rate": 3.6491914117110405e-06, "loss": 2.047, "step": 275 }, { "epoch": 1.5171012529630885, "grad_norm": 0.4323485633364769, "learning_rate": 3.2899064593549477e-06, "loss": 2.0445, "step": 280 }, { "epoch": 1.5441923467660006, "grad_norm": 0.44598125572196134, "learning_rate": 2.945730336971767e-06, "loss": 2.0405, "step": 285 }, { "epoch": 1.5712834405689131, "grad_norm": 0.44045190794321376, "learning_rate": 2.6174380111819144e-06, "loss": 2.0449, "step": 290 }, { "epoch": 1.5983745343718252, "grad_norm": 0.4315376791223261, "learning_rate": 2.3057686837419246e-06, "loss": 2.0455, "step": 295 }, { "epoch": 1.6254656281747377, "grad_norm": 0.4141056168787056, "learning_rate": 2.011424127115552e-06, "loss": 2.0429, "step": 300 }, { "epoch": 1.6525567219776498, "grad_norm": 0.42561322896204196, "learning_rate": 1.7350671043228072e-06, "loss": 2.0473, "step": 305 }, { "epoch": 1.679647815780562, "grad_norm": 0.4318633569152662, "learning_rate": 1.4773198766248642e-06, "loss": 2.0471, "step": 310 }, { "epoch": 1.7067389095834744, "grad_norm": 0.42654823899186717, "learning_rate": 1.2387628024050557e-06, "loss": 2.0502, "step": 315 }, { "epoch": 1.7338300033863867, "grad_norm": 0.4318585206830447, "learning_rate": 1.0199330304007858e-06, "loss": 2.0374, "step": 320 }, { "epoch": 1.760921097189299, "grad_norm": 0.42587188797302283, "learning_rate": 8.213232902287438e-07, "loss": 2.0508, "step": 325 }, { "epoch": 1.7880121909922113, "grad_norm": 0.42658756639019496, "learning_rate": 6.433807829267491e-07, "loss": 2.0458, "step": 330 }, { "epoch": 1.8151032847951236, "grad_norm": 0.4254891411016103, "learning_rate": 4.865061740103361e-07, "loss": 2.0538, "step": 335 }, { "epoch": 1.842194378598036, "grad_norm": 0.43121146874552274, "learning_rate": 3.510526913114065e-07, "loss": 2.0437, "step": 340 }, { "epoch": 1.8692854724009482, "grad_norm": 0.4148670481064235, "learning_rate": 2.3732532963024468e-07, "loss": 2.0366, "step": 345 }, { "epoch": 1.8963765662038605, "grad_norm": 0.4240449705158572, "learning_rate": 1.4558016399181086e-07, "loss": 2.0462, "step": 350 }, { "epoch": 1.9234676600067728, "grad_norm": 0.41343102256412484, "learning_rate": 7.602377305258479e-08, "loss": 2.0485, "step": 355 }, { "epoch": 1.950558753809685, "grad_norm": 0.4219092082642883, "learning_rate": 2.8812773956256034e-08, "loss": 2.0398, "step": 360 }, { "epoch": 1.9776498476125974, "grad_norm": 0.4203661391482514, "learning_rate": 4.053469685617595e-09, "loss": 2.0438, "step": 365 }, { "epoch": 1.9939045038943446, "eval_loss": 2.097979784011841, "eval_runtime": 77.4795, "eval_samples_per_second": 16.843, "eval_steps_per_second": 8.428, "step": 368 }, { "epoch": 1.9939045038943446, "step": 368, "total_flos": 53334434119680.0, "train_loss": 2.0994615062423376, "train_runtime": 7293.2016, "train_samples_per_second": 6.477, "train_steps_per_second": 0.05 } ], "logging_steps": 5, "max_steps": 368, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 53334434119680.0, "train_batch_size": 4, "trial_name": null, "trial_params": null }