{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 375, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.04, "grad_norm": 0.4841386675834656, "learning_rate": 4.997807075247146e-05, "loss": 0.688, "num_input_tokens_seen": 46776, "step": 5 }, { "epoch": 0.08, "grad_norm": 0.3365215063095093, "learning_rate": 4.991232148123761e-05, "loss": 0.763, "num_input_tokens_seen": 85136, "step": 10 }, { "epoch": 0.12, "grad_norm": 0.5198595523834229, "learning_rate": 4.980286753286195e-05, "loss": 0.6882, "num_input_tokens_seen": 126584, "step": 15 }, { "epoch": 0.16, "grad_norm": 0.40807807445526123, "learning_rate": 4.964990092676263e-05, "loss": 0.6951, "num_input_tokens_seen": 167968, "step": 20 }, { "epoch": 0.2, "grad_norm": 0.439756840467453, "learning_rate": 4.9453690018345144e-05, "loss": 0.5008, "num_input_tokens_seen": 206832, "step": 25 }, { "epoch": 0.24, "grad_norm": 0.45932796597480774, "learning_rate": 4.9214579028215776e-05, "loss": 0.542, "num_input_tokens_seen": 243656, "step": 30 }, { "epoch": 0.28, "grad_norm": 0.3281826376914978, "learning_rate": 4.893298743830168e-05, "loss": 0.5369, "num_input_tokens_seen": 284016, "step": 35 }, { "epoch": 0.32, "grad_norm": 0.2551577389240265, "learning_rate": 4.860940925593703e-05, "loss": 0.4948, "num_input_tokens_seen": 327408, "step": 40 }, { "epoch": 0.36, "grad_norm": 0.5936368107795715, "learning_rate": 4.8244412147206284e-05, "loss": 0.5244, "num_input_tokens_seen": 363376, "step": 45 }, { "epoch": 0.4, "grad_norm": 0.2480110377073288, "learning_rate": 4.783863644106502e-05, "loss": 0.421, "num_input_tokens_seen": 398840, "step": 50 }, { "epoch": 0.44, "grad_norm": 0.4601157307624817, "learning_rate": 4.7392794005985326e-05, "loss": 0.4517, "num_input_tokens_seen": 436136, "step": 55 }, { "epoch": 0.48, "grad_norm": 0.6387960314750671, "learning_rate": 4.690766700109659e-05, "loss": 0.4661, "num_input_tokens_seen": 475856, "step": 60 }, { "epoch": 0.52, "grad_norm": 0.4365377128124237, "learning_rate": 4.638410650401267e-05, "loss": 0.4928, "num_input_tokens_seen": 514496, "step": 65 }, { "epoch": 0.56, "grad_norm": 0.2992519438266754, "learning_rate": 4.5823031017752485e-05, "loss": 0.5424, "num_input_tokens_seen": 554424, "step": 70 }, { "epoch": 0.6, "grad_norm": 0.828787088394165, "learning_rate": 4.522542485937369e-05, "loss": 0.5419, "num_input_tokens_seen": 593264, "step": 75 }, { "epoch": 0.64, "grad_norm": 0.2943418622016907, "learning_rate": 4.4592336433146e-05, "loss": 0.4558, "num_input_tokens_seen": 630264, "step": 80 }, { "epoch": 0.68, "grad_norm": 0.3707512617111206, "learning_rate": 4.3924876391293915e-05, "loss": 0.5656, "num_input_tokens_seen": 668864, "step": 85 }, { "epoch": 0.72, "grad_norm": 0.4809654653072357, "learning_rate": 4.3224215685535294e-05, "loss": 0.4832, "num_input_tokens_seen": 712504, "step": 90 }, { "epoch": 0.76, "grad_norm": 0.516292929649353, "learning_rate": 4.249158351283414e-05, "loss": 0.4626, "num_input_tokens_seen": 748872, "step": 95 }, { "epoch": 0.8, "grad_norm": 0.5604212284088135, "learning_rate": 4.172826515897146e-05, "loss": 0.4837, "num_input_tokens_seen": 788408, "step": 100 }, { "epoch": 0.84, "grad_norm": 0.7355071902275085, "learning_rate": 4.093559974371725e-05, "loss": 0.5144, "num_input_tokens_seen": 828448, "step": 105 }, { "epoch": 0.88, "grad_norm": 0.46999862790107727, "learning_rate": 4.011497787155938e-05, "loss": 0.493, "num_input_tokens_seen": 864680, "step": 110 }, { "epoch": 0.92, "grad_norm": 0.5488362312316895, "learning_rate": 3.92678391921108e-05, "loss": 0.4083, "num_input_tokens_seen": 902568, "step": 115 }, { "epoch": 0.96, "grad_norm": 0.5053719282150269, "learning_rate": 3.8395669874474915e-05, "loss": 0.5172, "num_input_tokens_seen": 944752, "step": 120 }, { "epoch": 1.0, "grad_norm": 0.44234323501586914, "learning_rate": 3.7500000000000003e-05, "loss": 0.5843, "num_input_tokens_seen": 988656, "step": 125 }, { "epoch": 1.04, "grad_norm": 0.4518623352050781, "learning_rate": 3.6582400877996546e-05, "loss": 0.4567, "num_input_tokens_seen": 1028584, "step": 130 }, { "epoch": 1.08, "grad_norm": 0.39792686700820923, "learning_rate": 3.564448228912682e-05, "loss": 0.418, "num_input_tokens_seen": 1065104, "step": 135 }, { "epoch": 1.12, "grad_norm": 0.5141780376434326, "learning_rate": 3.4687889661302576e-05, "loss": 0.3785, "num_input_tokens_seen": 1106744, "step": 140 }, { "epoch": 1.16, "grad_norm": 0.5755515098571777, "learning_rate": 3.3714301183045385e-05, "loss": 0.4097, "num_input_tokens_seen": 1140672, "step": 145 }, { "epoch": 1.2, "grad_norm": 0.4626426696777344, "learning_rate": 3.272542485937369e-05, "loss": 0.4507, "num_input_tokens_seen": 1177280, "step": 150 }, { "epoch": 1.24, "grad_norm": 0.49699896574020386, "learning_rate": 3.172299551538164e-05, "loss": 0.368, "num_input_tokens_seen": 1215744, "step": 155 }, { "epoch": 1.28, "grad_norm": 0.5842998027801514, "learning_rate": 3.0708771752766394e-05, "loss": 0.4301, "num_input_tokens_seen": 1257040, "step": 160 }, { "epoch": 1.32, "grad_norm": 0.46173322200775146, "learning_rate": 2.9684532864643122e-05, "loss": 0.4488, "num_input_tokens_seen": 1295992, "step": 165 }, { "epoch": 1.3599999999999999, "grad_norm": 0.4926168918609619, "learning_rate": 2.8652075714060295e-05, "loss": 0.4075, "num_input_tokens_seen": 1334672, "step": 170 }, { "epoch": 1.4, "grad_norm": 0.5110897421836853, "learning_rate": 2.761321158169134e-05, "loss": 0.4991, "num_input_tokens_seen": 1379280, "step": 175 }, { "epoch": 1.44, "grad_norm": 0.33892518281936646, "learning_rate": 2.656976298823284e-05, "loss": 0.4894, "num_input_tokens_seen": 1419704, "step": 180 }, { "epoch": 1.48, "grad_norm": 0.46542254090309143, "learning_rate": 2.5523560497083926e-05, "loss": 0.4967, "num_input_tokens_seen": 1459256, "step": 185 }, { "epoch": 1.52, "grad_norm": 0.9100029468536377, "learning_rate": 2.447643950291608e-05, "loss": 0.5297, "num_input_tokens_seen": 1503576, "step": 190 }, { "epoch": 1.56, "grad_norm": 0.7347203493118286, "learning_rate": 2.3430237011767167e-05, "loss": 0.3939, "num_input_tokens_seen": 1539904, "step": 195 }, { "epoch": 1.6, "grad_norm": 0.6078445315361023, "learning_rate": 2.238678841830867e-05, "loss": 0.461, "num_input_tokens_seen": 1580560, "step": 200 }, { "epoch": 1.6400000000000001, "grad_norm": 0.5878064632415771, "learning_rate": 2.1347924285939714e-05, "loss": 0.4622, "num_input_tokens_seen": 1623608, "step": 205 }, { "epoch": 1.6800000000000002, "grad_norm": 0.6516324877738953, "learning_rate": 2.031546713535688e-05, "loss": 0.4043, "num_input_tokens_seen": 1660744, "step": 210 }, { "epoch": 1.72, "grad_norm": 0.9136825203895569, "learning_rate": 1.9291228247233605e-05, "loss": 0.428, "num_input_tokens_seen": 1698784, "step": 215 }, { "epoch": 1.76, "grad_norm": 0.751567006111145, "learning_rate": 1.827700448461836e-05, "loss": 0.3779, "num_input_tokens_seen": 1734792, "step": 220 }, { "epoch": 1.8, "grad_norm": 0.8802843689918518, "learning_rate": 1.7274575140626318e-05, "loss": 0.4526, "num_input_tokens_seen": 1776144, "step": 225 }, { "epoch": 1.8399999999999999, "grad_norm": 0.6807656288146973, "learning_rate": 1.6285698816954624e-05, "loss": 0.4627, "num_input_tokens_seen": 1815864, "step": 230 }, { "epoch": 1.88, "grad_norm": 0.6222789883613586, "learning_rate": 1.5312110338697426e-05, "loss": 0.4873, "num_input_tokens_seen": 1856168, "step": 235 }, { "epoch": 1.92, "grad_norm": 0.5224676728248596, "learning_rate": 1.4355517710873184e-05, "loss": 0.3234, "num_input_tokens_seen": 1899120, "step": 240 }, { "epoch": 1.96, "grad_norm": 0.4191940724849701, "learning_rate": 1.3417599122003464e-05, "loss": 0.4438, "num_input_tokens_seen": 1936808, "step": 245 }, { "epoch": 2.0, "grad_norm": 0.6064372062683105, "learning_rate": 1.2500000000000006e-05, "loss": 0.4407, "num_input_tokens_seen": 1977312, "step": 250 }, { "epoch": 2.04, "grad_norm": 0.5188178420066833, "learning_rate": 1.1604330125525079e-05, "loss": 0.4401, "num_input_tokens_seen": 2018336, "step": 255 }, { "epoch": 2.08, "grad_norm": 0.5036932826042175, "learning_rate": 1.0732160807889211e-05, "loss": 0.3771, "num_input_tokens_seen": 2057776, "step": 260 }, { "epoch": 2.12, "grad_norm": 0.5732894539833069, "learning_rate": 9.88502212844063e-06, "loss": 0.4043, "num_input_tokens_seen": 2102312, "step": 265 }, { "epoch": 2.16, "grad_norm": 0.6521058082580566, "learning_rate": 9.064400256282757e-06, "loss": 0.4018, "num_input_tokens_seen": 2137304, "step": 270 }, { "epoch": 2.2, "grad_norm": 0.698835015296936, "learning_rate": 8.271734841028553e-06, "loss": 0.4258, "num_input_tokens_seen": 2176144, "step": 275 }, { "epoch": 2.24, "grad_norm": 0.7500255107879639, "learning_rate": 7.508416487165862e-06, "loss": 0.3912, "num_input_tokens_seen": 2211640, "step": 280 }, { "epoch": 2.2800000000000002, "grad_norm": 0.7223864793777466, "learning_rate": 6.775784314464717e-06, "loss": 0.3458, "num_input_tokens_seen": 2247272, "step": 285 }, { "epoch": 2.32, "grad_norm": 0.6346307992935181, "learning_rate": 6.075123608706093e-06, "loss": 0.4255, "num_input_tokens_seen": 2285776, "step": 290 }, { "epoch": 2.36, "grad_norm": 0.6048356890678406, "learning_rate": 5.4076635668540075e-06, "loss": 0.4222, "num_input_tokens_seen": 2326112, "step": 295 }, { "epoch": 2.4, "grad_norm": 1.0743402242660522, "learning_rate": 4.7745751406263165e-06, "loss": 0.399, "num_input_tokens_seen": 2362816, "step": 300 }, { "epoch": 2.44, "grad_norm": 0.6533491015434265, "learning_rate": 4.176968982247514e-06, "loss": 0.3382, "num_input_tokens_seen": 2402328, "step": 305 }, { "epoch": 2.48, "grad_norm": 0.7778833508491516, "learning_rate": 3.6158934959873353e-06, "loss": 0.4465, "num_input_tokens_seen": 2445600, "step": 310 }, { "epoch": 2.52, "grad_norm": 0.7358531355857849, "learning_rate": 3.092332998903416e-06, "loss": 0.325, "num_input_tokens_seen": 2485840, "step": 315 }, { "epoch": 2.56, "grad_norm": 0.7607108354568481, "learning_rate": 2.6072059940146775e-06, "loss": 0.392, "num_input_tokens_seen": 2524760, "step": 320 }, { "epoch": 2.6, "grad_norm": 0.7286326289176941, "learning_rate": 2.1613635589349756e-06, "loss": 0.3672, "num_input_tokens_seen": 2562040, "step": 325 }, { "epoch": 2.64, "grad_norm": 0.5720709562301636, "learning_rate": 1.7555878527937164e-06, "loss": 0.3554, "num_input_tokens_seen": 2600056, "step": 330 }, { "epoch": 2.68, "grad_norm": 0.9434547424316406, "learning_rate": 1.3905907440629752e-06, "loss": 0.3801, "num_input_tokens_seen": 2633392, "step": 335 }, { "epoch": 2.7199999999999998, "grad_norm": 0.9704461693763733, "learning_rate": 1.067012561698319e-06, "loss": 0.435, "num_input_tokens_seen": 2677016, "step": 340 }, { "epoch": 2.76, "grad_norm": 0.6400224566459656, "learning_rate": 7.854209717842231e-07, "loss": 0.4063, "num_input_tokens_seen": 2719536, "step": 345 }, { "epoch": 2.8, "grad_norm": 0.46599817276000977, "learning_rate": 5.463099816548579e-07, "loss": 0.4894, "num_input_tokens_seen": 2767424, "step": 350 }, { "epoch": 2.84, "grad_norm": 0.4850185215473175, "learning_rate": 3.5009907323737825e-07, "loss": 0.3822, "num_input_tokens_seen": 2805832, "step": 355 }, { "epoch": 2.88, "grad_norm": 0.8774814605712891, "learning_rate": 1.9713246713805588e-07, "loss": 0.4028, "num_input_tokens_seen": 2845912, "step": 360 }, { "epoch": 2.92, "grad_norm": 0.6378630995750427, "learning_rate": 8.767851876239074e-08, "loss": 0.4293, "num_input_tokens_seen": 2888128, "step": 365 }, { "epoch": 2.96, "grad_norm": 0.8157021403312683, "learning_rate": 2.192924752854042e-08, "loss": 0.428, "num_input_tokens_seen": 2925024, "step": 370 }, { "epoch": 3.0, "grad_norm": 0.5457362532615662, "learning_rate": 0.0, "loss": 0.4766, "num_input_tokens_seen": 2965968, "step": 375 }, { "epoch": 3.0, "num_input_tokens_seen": 2965968, "step": 375, "total_flos": 1.2618658999382835e+17, "train_loss": 0.45766232585906985, "train_runtime": 3803.7297, "train_samples_per_second": 1.577, "train_steps_per_second": 0.099 } ], "logging_steps": 5, "max_steps": 375, "num_input_tokens_seen": 2965968, "num_train_epochs": 3, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.2618658999382835e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }