{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 10, "global_step": 98, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.01020408163265306, "grad_norm": 1.5935372114181519, "learning_rate": 1e-05, "loss": 1.1572, "step": 1 }, { "epoch": 0.02040816326530612, "grad_norm": 1.8600682020187378, "learning_rate": 9.997377845227577e-06, "loss": 1.3198, "step": 2 }, { "epoch": 0.030612244897959183, "grad_norm": 1.662075400352478, "learning_rate": 9.98951413118856e-06, "loss": 1.0636, "step": 3 }, { "epoch": 0.04081632653061224, "grad_norm": 1.6947352886199951, "learning_rate": 9.97641710583307e-06, "loss": 1.2853, "step": 4 }, { "epoch": 0.05102040816326531, "grad_norm": 1.231564998626709, "learning_rate": 9.958100506132127e-06, "loss": 1.0573, "step": 5 }, { "epoch": 0.061224489795918366, "grad_norm": 1.550899863243103, "learning_rate": 9.934583543669454e-06, "loss": 1.5486, "step": 6 }, { "epoch": 0.07142857142857142, "grad_norm": 1.017148733139038, "learning_rate": 9.905890884491196e-06, "loss": 0.8672, "step": 7 }, { "epoch": 0.08163265306122448, "grad_norm": 1.2291380167007446, "learning_rate": 9.872052623234632e-06, "loss": 1.132, "step": 8 }, { "epoch": 0.09183673469387756, "grad_norm": 1.1650055646896362, "learning_rate": 9.833104251563058e-06, "loss": 1.2419, "step": 9 }, { "epoch": 0.10204081632653061, "grad_norm": 0.9753865003585815, "learning_rate": 9.789086620939936e-06, "loss": 0.9946, "step": 10 }, { "epoch": 0.10204081632653061, "eval_loss": 0.9807827472686768, "eval_runtime": 0.2604, "eval_samples_per_second": 30.717, "eval_steps_per_second": 3.84, "step": 10 }, { "epoch": 0.11224489795918367, "grad_norm": 1.191615343093872, "learning_rate": 9.740045899781353e-06, "loss": 1.3435, "step": 11 }, { "epoch": 0.12244897959183673, "grad_norm": 0.9799769520759583, "learning_rate": 9.68603352503172e-06, "loss": 0.7463, "step": 12 }, { "epoch": 0.1326530612244898, "grad_norm": 1.0710457563400269, "learning_rate": 9.627106148213521e-06, "loss": 1.329, "step": 13 }, { "epoch": 0.14285714285714285, "grad_norm": 1.071432113647461, "learning_rate": 9.563325576007702e-06, "loss": 1.3815, "step": 14 }, { "epoch": 0.15306122448979592, "grad_norm": 0.9442039132118225, "learning_rate": 9.494758705426978e-06, "loss": 1.0305, "step": 15 }, { "epoch": 0.16326530612244897, "grad_norm": 1.022035837173462, "learning_rate": 9.421477453650118e-06, "loss": 1.3317, "step": 16 }, { "epoch": 0.17346938775510204, "grad_norm": 1.0751738548278809, "learning_rate": 9.343558682590757e-06, "loss": 1.323, "step": 17 }, { "epoch": 0.1836734693877551, "grad_norm": 1.0566898584365845, "learning_rate": 9.261084118279846e-06, "loss": 1.5154, "step": 18 }, { "epoch": 0.19387755102040816, "grad_norm": 1.0009628534317017, "learning_rate": 9.174140265146356e-06, "loss": 1.2371, "step": 19 }, { "epoch": 0.20408163265306123, "grad_norm": 0.9259373545646667, "learning_rate": 9.082818315286054e-06, "loss": 1.042, "step": 20 }, { "epoch": 0.20408163265306123, "eval_loss": 0.9510844945907593, "eval_runtime": 0.2601, "eval_samples_per_second": 30.76, "eval_steps_per_second": 3.845, "step": 20 }, { "epoch": 0.21428571428571427, "grad_norm": 0.9078995585441589, "learning_rate": 8.987214052813605e-06, "loss": 1.07, "step": 21 }, { "epoch": 0.22448979591836735, "grad_norm": 0.9820942282676697, "learning_rate": 8.887427753398249e-06, "loss": 1.2102, "step": 22 }, { "epoch": 0.23469387755102042, "grad_norm": 1.0587691068649292, "learning_rate": 8.783564079088478e-06, "loss": 1.0913, "step": 23 }, { "epoch": 0.24489795918367346, "grad_norm": 0.8137860894203186, "learning_rate": 8.675731968536004e-06, "loss": 0.8923, "step": 24 }, { "epoch": 0.25510204081632654, "grad_norm": 0.9471170902252197, "learning_rate": 8.564044522734147e-06, "loss": 1.1723, "step": 25 }, { "epoch": 0.2653061224489796, "grad_norm": 0.8885685205459595, "learning_rate": 8.448618886390523e-06, "loss": 1.1206, "step": 26 }, { "epoch": 0.2755102040816326, "grad_norm": 0.8613781332969666, "learning_rate": 8.329576125058406e-06, "loss": 1.1261, "step": 27 }, { "epoch": 0.2857142857142857, "grad_norm": 0.9407626986503601, "learning_rate": 8.207041098155701e-06, "loss": 1.2727, "step": 28 }, { "epoch": 0.29591836734693877, "grad_norm": 0.8954651951789856, "learning_rate": 8.081142328004638e-06, "loss": 1.2197, "step": 29 }, { "epoch": 0.30612244897959184, "grad_norm": 0.8864608407020569, "learning_rate": 7.952011865029614e-06, "loss": 1.1236, "step": 30 }, { "epoch": 0.30612244897959184, "eval_loss": 0.9308969974517822, "eval_runtime": 0.2599, "eval_samples_per_second": 30.775, "eval_steps_per_second": 3.847, "step": 30 }, { "epoch": 0.3163265306122449, "grad_norm": 0.8076571226119995, "learning_rate": 7.819785149254534e-06, "loss": 0.9098, "step": 31 }, { "epoch": 0.32653061224489793, "grad_norm": 0.8902921080589294, "learning_rate": 7.68460086824492e-06, "loss": 1.0564, "step": 32 }, { "epoch": 0.336734693877551, "grad_norm": 0.9010928273200989, "learning_rate": 7.546600811643816e-06, "loss": 1.0174, "step": 33 }, { "epoch": 0.3469387755102041, "grad_norm": 0.8672641515731812, "learning_rate": 7.405929722454026e-06, "loss": 1.1875, "step": 34 }, { "epoch": 0.35714285714285715, "grad_norm": 0.7878830432891846, "learning_rate": 7.262735145222696e-06, "loss": 0.9453, "step": 35 }, { "epoch": 0.3673469387755102, "grad_norm": 0.85221266746521, "learning_rate": 7.117167271287453e-06, "loss": 1.0147, "step": 36 }, { "epoch": 0.37755102040816324, "grad_norm": 0.8591777682304382, "learning_rate": 6.969378781246436e-06, "loss": 0.995, "step": 37 }, { "epoch": 0.3877551020408163, "grad_norm": 0.9585978388786316, "learning_rate": 6.819524684817439e-06, "loss": 0.9636, "step": 38 }, { "epoch": 0.3979591836734694, "grad_norm": 0.9561256170272827, "learning_rate": 6.667762158254104e-06, "loss": 1.1511, "step": 39 }, { "epoch": 0.40816326530612246, "grad_norm": 0.9076462388038635, "learning_rate": 6.514250379489754e-06, "loss": 1.238, "step": 40 }, { "epoch": 0.40816326530612246, "eval_loss": 0.9210565090179443, "eval_runtime": 0.26, "eval_samples_per_second": 30.764, "eval_steps_per_second": 3.846, "step": 40 }, { "epoch": 0.41836734693877553, "grad_norm": 0.779888391494751, "learning_rate": 6.3591503611817155e-06, "loss": 0.927, "step": 41 }, { "epoch": 0.42857142857142855, "grad_norm": 0.826004147529602, "learning_rate": 6.202624781831269e-06, "loss": 1.1018, "step": 42 }, { "epoch": 0.4387755102040816, "grad_norm": 0.8641281127929688, "learning_rate": 6.044837815156377e-06, "loss": 0.938, "step": 43 }, { "epoch": 0.4489795918367347, "grad_norm": 0.896023154258728, "learning_rate": 5.885954957896115e-06, "loss": 1.0764, "step": 44 }, { "epoch": 0.45918367346938777, "grad_norm": 0.9072222709655762, "learning_rate": 5.726142856227453e-06, "loss": 1.2437, "step": 45 }, { "epoch": 0.46938775510204084, "grad_norm": 0.7771997451782227, "learning_rate": 5.5655691309764225e-06, "loss": 1.0162, "step": 46 }, { "epoch": 0.47959183673469385, "grad_norm": 0.8864902853965759, "learning_rate": 5.404402201807022e-06, "loss": 1.1398, "step": 47 }, { "epoch": 0.4897959183673469, "grad_norm": 0.702143669128418, "learning_rate": 5.242811110572243e-06, "loss": 0.8257, "step": 48 }, { "epoch": 0.5, "grad_norm": 0.7524011731147766, "learning_rate": 5.080965344012509e-06, "loss": 0.9318, "step": 49 }, { "epoch": 0.5102040816326531, "grad_norm": 0.7233298420906067, "learning_rate": 4.919034655987493e-06, "loss": 0.8178, "step": 50 }, { "epoch": 0.5102040816326531, "eval_loss": 0.9119246006011963, "eval_runtime": 0.2595, "eval_samples_per_second": 30.823, "eval_steps_per_second": 3.853, "step": 50 }, { "epoch": 0.5204081632653061, "grad_norm": 0.8432551622390747, "learning_rate": 4.757188889427761e-06, "loss": 1.0659, "step": 51 }, { "epoch": 0.5306122448979592, "grad_norm": 0.772146999835968, "learning_rate": 4.59559779819298e-06, "loss": 1.0016, "step": 52 }, { "epoch": 0.5408163265306123, "grad_norm": 0.8176573514938354, "learning_rate": 4.434430869023579e-06, "loss": 1.0541, "step": 53 }, { "epoch": 0.5510204081632653, "grad_norm": 0.730097770690918, "learning_rate": 4.27385714377255e-06, "loss": 0.8034, "step": 54 }, { "epoch": 0.5612244897959183, "grad_norm": 0.7190036177635193, "learning_rate": 4.1140450421038865e-06, "loss": 0.8833, "step": 55 }, { "epoch": 0.5714285714285714, "grad_norm": 0.732845664024353, "learning_rate": 3.955162184843625e-06, "loss": 0.8837, "step": 56 }, { "epoch": 0.5816326530612245, "grad_norm": 0.770910918712616, "learning_rate": 3.7973752181687336e-06, "loss": 1.0044, "step": 57 }, { "epoch": 0.5918367346938775, "grad_norm": 0.7635073065757751, "learning_rate": 3.6408496388182857e-06, "loss": 0.9338, "step": 58 }, { "epoch": 0.6020408163265306, "grad_norm": 0.8665493130683899, "learning_rate": 3.4857496205102475e-06, "loss": 1.1868, "step": 59 }, { "epoch": 0.6122448979591837, "grad_norm": 0.7718795537948608, "learning_rate": 3.3322378417458985e-06, "loss": 0.951, "step": 60 }, { "epoch": 0.6122448979591837, "eval_loss": 0.9065544605255127, "eval_runtime": 0.2595, "eval_samples_per_second": 30.831, "eval_steps_per_second": 3.854, "step": 60 }, { "epoch": 0.6224489795918368, "grad_norm": 0.932878851890564, "learning_rate": 3.180475315182563e-06, "loss": 1.2871, "step": 61 }, { "epoch": 0.6326530612244898, "grad_norm": 0.8946093320846558, "learning_rate": 3.0306212187535653e-06, "loss": 0.9965, "step": 62 }, { "epoch": 0.6428571428571429, "grad_norm": 0.7468050718307495, "learning_rate": 2.882832728712551e-06, "loss": 0.8174, "step": 63 }, { "epoch": 0.6530612244897959, "grad_norm": 0.7200383543968201, "learning_rate": 2.7372648547773063e-06, "loss": 0.9641, "step": 64 }, { "epoch": 0.6632653061224489, "grad_norm": 0.7331112623214722, "learning_rate": 2.594070277545975e-06, "loss": 0.8579, "step": 65 }, { "epoch": 0.673469387755102, "grad_norm": 0.713594377040863, "learning_rate": 2.4533991883561868e-06, "loss": 0.872, "step": 66 }, { "epoch": 0.6836734693877551, "grad_norm": 0.8468301296234131, "learning_rate": 2.315399131755081e-06, "loss": 1.1029, "step": 67 }, { "epoch": 0.6938775510204082, "grad_norm": 0.7141689658164978, "learning_rate": 2.1802148507454675e-06, "loss": 0.8727, "step": 68 }, { "epoch": 0.7040816326530612, "grad_norm": 0.7567382454872131, "learning_rate": 2.0479881349703885e-06, "loss": 0.9211, "step": 69 }, { "epoch": 0.7142857142857143, "grad_norm": 0.7020692825317383, "learning_rate": 1.9188576719953635e-06, "loss": 0.7869, "step": 70 }, { "epoch": 0.7142857142857143, "eval_loss": 0.9032285809516907, "eval_runtime": 0.2594, "eval_samples_per_second": 30.835, "eval_steps_per_second": 3.854, "step": 70 }, { "epoch": 0.7244897959183674, "grad_norm": 0.853243350982666, "learning_rate": 1.7929589018443016e-06, "loss": 1.1762, "step": 71 }, { "epoch": 0.7346938775510204, "grad_norm": 0.8287662267684937, "learning_rate": 1.6704238749415958e-06, "loss": 1.0489, "step": 72 }, { "epoch": 0.7448979591836735, "grad_norm": 0.6973975300788879, "learning_rate": 1.5513811136094786e-06, "loss": 0.8249, "step": 73 }, { "epoch": 0.7551020408163265, "grad_norm": 0.8150553703308105, "learning_rate": 1.4359554772658551e-06, "loss": 1.0011, "step": 74 }, { "epoch": 0.7653061224489796, "grad_norm": 0.7974684238433838, "learning_rate": 1.3242680314639995e-06, "loss": 1.0774, "step": 75 }, { "epoch": 0.7755102040816326, "grad_norm": 0.8233014941215515, "learning_rate": 1.2164359209115235e-06, "loss": 1.1492, "step": 76 }, { "epoch": 0.7857142857142857, "grad_norm": 0.8305495977401733, "learning_rate": 1.1125722466017547e-06, "loss": 1.1329, "step": 77 }, { "epoch": 0.7959183673469388, "grad_norm": 0.7523290514945984, "learning_rate": 1.012785947186397e-06, "loss": 0.9375, "step": 78 }, { "epoch": 0.8061224489795918, "grad_norm": 0.8073377013206482, "learning_rate": 9.171816847139447e-07, "loss": 1.0936, "step": 79 }, { "epoch": 0.8163265306122449, "grad_norm": 0.8685086965560913, "learning_rate": 8.258597348536452e-07, "loss": 1.2626, "step": 80 }, { "epoch": 0.8163265306122449, "eval_loss": 0.9012266397476196, "eval_runtime": 0.2595, "eval_samples_per_second": 30.833, "eval_steps_per_second": 3.854, "step": 80 }, { "epoch": 0.826530612244898, "grad_norm": 0.877649188041687, "learning_rate": 7.389158817201541e-07, "loss": 1.3479, "step": 81 }, { "epoch": 0.8367346938775511, "grad_norm": 0.7652468085289001, "learning_rate": 6.564413174092443e-07, "loss": 0.9699, "step": 82 }, { "epoch": 0.8469387755102041, "grad_norm": 0.8436093926429749, "learning_rate": 5.785225463498828e-07, "loss": 1.1757, "step": 83 }, { "epoch": 0.8571428571428571, "grad_norm": 0.8732744455337524, "learning_rate": 5.05241294573024e-07, "loss": 1.1877, "step": 84 }, { "epoch": 0.8673469387755102, "grad_norm": 0.7433239221572876, "learning_rate": 4.3667442399229985e-07, "loss": 0.9878, "step": 85 }, { "epoch": 0.8775510204081632, "grad_norm": 0.8225981593132019, "learning_rate": 3.728938517864794e-07, "loss": 1.145, "step": 86 }, { "epoch": 0.8877551020408163, "grad_norm": 0.8591734170913696, "learning_rate": 3.1396647496828245e-07, "loss": 1.128, "step": 87 }, { "epoch": 0.8979591836734694, "grad_norm": 0.7502849102020264, "learning_rate": 2.599541002186479e-07, "loss": 0.9712, "step": 88 }, { "epoch": 0.9081632653061225, "grad_norm": 0.7429087162017822, "learning_rate": 2.109133790600648e-07, "loss": 0.9278, "step": 89 }, { "epoch": 0.9183673469387755, "grad_norm": 0.8890379667282104, "learning_rate": 1.6689574843694433e-07, "loss": 1.2385, "step": 90 }, { "epoch": 0.9183673469387755, "eval_loss": 0.9008685350418091, "eval_runtime": 0.2594, "eval_samples_per_second": 30.837, "eval_steps_per_second": 3.855, "step": 90 }, { "epoch": 0.9285714285714286, "grad_norm": 0.9237527251243591, "learning_rate": 1.2794737676536993e-07, "loss": 1.0866, "step": 91 }, { "epoch": 0.9387755102040817, "grad_norm": 0.8110918402671814, "learning_rate": 9.410911550880474e-08, "loss": 1.0466, "step": 92 }, { "epoch": 0.9489795918367347, "grad_norm": 0.7134807109832764, "learning_rate": 6.54164563305465e-08, "loss": 0.8466, "step": 93 }, { "epoch": 0.9591836734693877, "grad_norm": 0.8616638779640198, "learning_rate": 4.189949386787462e-08, "loss": 1.1521, "step": 94 }, { "epoch": 0.9693877551020408, "grad_norm": 0.9501765966415405, "learning_rate": 2.358289416693027e-08, "loss": 1.1505, "step": 95 }, { "epoch": 0.9795918367346939, "grad_norm": 0.7652553915977478, "learning_rate": 1.0485868811441757e-08, "loss": 0.9055, "step": 96 }, { "epoch": 0.9897959183673469, "grad_norm": 0.862601637840271, "learning_rate": 2.6221547724253337e-09, "loss": 1.0224, "step": 97 }, { "epoch": 1.0, "grad_norm": 0.712780773639679, "learning_rate": 0.0, "loss": 0.9534, "step": 98 }, { "epoch": 1.0, "step": 98, "total_flos": 3.270639151271117e+16, "train_loss": 1.0707874030483013, "train_runtime": 199.5779, "train_samples_per_second": 3.908, "train_steps_per_second": 0.491 } ], "logging_steps": 1, "max_steps": 98, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 98, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.270639151271117e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }