{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.898550724637681, "eval_steps": 500, "global_step": 136, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.057971014492753624, "grad_norm": 55.32845687866211, "learning_rate": 0.0001985294117647059, "loss": 76.4214, "step": 2 }, { "epoch": 0.11594202898550725, "grad_norm": 59.647117614746094, "learning_rate": 0.00019558823529411764, "loss": 66.9992, "step": 4 }, { "epoch": 0.17391304347826086, "grad_norm": 37.24838638305664, "learning_rate": 0.00019264705882352944, "loss": 53.3101, "step": 6 }, { "epoch": 0.2318840579710145, "grad_norm": 31.579923629760742, "learning_rate": 0.00018970588235294117, "loss": 43.1133, "step": 8 }, { "epoch": 0.2898550724637681, "grad_norm": 33.05708312988281, "learning_rate": 0.00018676470588235297, "loss": 41.4356, "step": 10 }, { "epoch": 0.34782608695652173, "grad_norm": 19.261913299560547, "learning_rate": 0.0001838235294117647, "loss": 40.5694, "step": 12 }, { "epoch": 0.4057971014492754, "grad_norm": 10.231746673583984, "learning_rate": 0.00018088235294117647, "loss": 35.4952, "step": 14 }, { "epoch": 0.463768115942029, "grad_norm": 12.686185836791992, "learning_rate": 0.00017794117647058823, "loss": 35.9178, "step": 16 }, { "epoch": 0.5217391304347826, "grad_norm": 5.934300899505615, "learning_rate": 0.000175, "loss": 36.0789, "step": 18 }, { "epoch": 0.5797101449275363, "grad_norm": 6.5412092208862305, "learning_rate": 0.0001720588235294118, "loss": 36.4717, "step": 20 }, { "epoch": 0.6376811594202898, "grad_norm": 7.47074031829834, "learning_rate": 0.00016911764705882353, "loss": 36.1205, "step": 22 }, { "epoch": 0.6956521739130435, "grad_norm": 7.197345733642578, "learning_rate": 0.00016617647058823532, "loss": 35.8026, "step": 24 }, { "epoch": 0.7536231884057971, "grad_norm": 8.729438781738281, "learning_rate": 0.00016323529411764706, "loss": 36.109, "step": 26 }, { "epoch": 0.8115942028985508, "grad_norm": 10.730180740356445, "learning_rate": 0.00016029411764705885, "loss": 34.3386, "step": 28 }, { "epoch": 0.8695652173913043, "grad_norm": 8.981538772583008, "learning_rate": 0.0001573529411764706, "loss": 33.3324, "step": 30 }, { "epoch": 0.927536231884058, "grad_norm": 13.125741958618164, "learning_rate": 0.00015441176470588238, "loss": 34.8168, "step": 32 }, { "epoch": 0.9855072463768116, "grad_norm": 7.202057361602783, "learning_rate": 0.00015147058823529412, "loss": 33.9379, "step": 34 }, { "epoch": 1.0289855072463767, "grad_norm": 7.689969539642334, "learning_rate": 0.00014852941176470588, "loss": 26.4506, "step": 36 }, { "epoch": 1.0869565217391304, "grad_norm": 12.119438171386719, "learning_rate": 0.00014558823529411765, "loss": 34.1455, "step": 38 }, { "epoch": 1.144927536231884, "grad_norm": 10.431544303894043, "learning_rate": 0.0001426470588235294, "loss": 33.6817, "step": 40 }, { "epoch": 1.2028985507246377, "grad_norm": 6.7350311279296875, "learning_rate": 0.00013970588235294118, "loss": 34.3042, "step": 42 }, { "epoch": 1.2608695652173914, "grad_norm": 7.63714599609375, "learning_rate": 0.00013676470588235294, "loss": 33.6539, "step": 44 }, { "epoch": 1.318840579710145, "grad_norm": 14.94995403289795, "learning_rate": 0.0001338235294117647, "loss": 34.8423, "step": 46 }, { "epoch": 1.3768115942028984, "grad_norm": 5.653194904327393, "learning_rate": 0.00013088235294117647, "loss": 33.6543, "step": 48 }, { "epoch": 1.434782608695652, "grad_norm": 7.0490946769714355, "learning_rate": 0.00012794117647058824, "loss": 34.4564, "step": 50 }, { "epoch": 1.4927536231884058, "grad_norm": 9.5311279296875, "learning_rate": 0.000125, "loss": 34.0708, "step": 52 }, { "epoch": 1.5507246376811594, "grad_norm": 9.313169479370117, "learning_rate": 0.00012205882352941178, "loss": 35.0762, "step": 54 }, { "epoch": 1.608695652173913, "grad_norm": 6.120841026306152, "learning_rate": 0.00011911764705882353, "loss": 33.5708, "step": 56 }, { "epoch": 1.6666666666666665, "grad_norm": 19.58426856994629, "learning_rate": 0.00011617647058823531, "loss": 32.583, "step": 58 }, { "epoch": 1.7246376811594204, "grad_norm": 10.036286354064941, "learning_rate": 0.00011323529411764706, "loss": 35.4897, "step": 60 }, { "epoch": 1.7826086956521738, "grad_norm": 9.211186408996582, "learning_rate": 0.00011029411764705884, "loss": 33.5828, "step": 62 }, { "epoch": 1.8405797101449275, "grad_norm": 7.165535926818848, "learning_rate": 0.00010735294117647059, "loss": 34.0128, "step": 64 }, { "epoch": 1.8985507246376812, "grad_norm": 8.84890079498291, "learning_rate": 0.00010441176470588237, "loss": 34.0061, "step": 66 }, { "epoch": 1.9565217391304348, "grad_norm": 6.122589111328125, "learning_rate": 0.00010147058823529412, "loss": 34.1371, "step": 68 }, { "epoch": 2.0, "grad_norm": 5.6373186111450195, "learning_rate": 9.852941176470589e-05, "loss": 25.4753, "step": 70 }, { "epoch": 2.0579710144927534, "grad_norm": 10.314593315124512, "learning_rate": 9.558823529411765e-05, "loss": 33.9195, "step": 72 }, { "epoch": 2.1159420289855073, "grad_norm": 8.193323135375977, "learning_rate": 9.264705882352942e-05, "loss": 33.1874, "step": 74 }, { "epoch": 2.1739130434782608, "grad_norm": 9.070418357849121, "learning_rate": 8.970588235294118e-05, "loss": 33.3055, "step": 76 }, { "epoch": 2.2318840579710146, "grad_norm": 8.063404083251953, "learning_rate": 8.676470588235295e-05, "loss": 34.1831, "step": 78 }, { "epoch": 2.289855072463768, "grad_norm": 7.3460516929626465, "learning_rate": 8.382352941176471e-05, "loss": 32.5002, "step": 80 }, { "epoch": 2.3478260869565215, "grad_norm": 7.9106974601745605, "learning_rate": 8.088235294117648e-05, "loss": 33.9427, "step": 82 }, { "epoch": 2.4057971014492754, "grad_norm": 6.9293532371521, "learning_rate": 7.794117647058824e-05, "loss": 32.7437, "step": 84 }, { "epoch": 2.463768115942029, "grad_norm": 5.583312511444092, "learning_rate": 7.500000000000001e-05, "loss": 32.4636, "step": 86 }, { "epoch": 2.5217391304347827, "grad_norm": 9.602357864379883, "learning_rate": 7.205882352941177e-05, "loss": 34.4266, "step": 88 }, { "epoch": 2.579710144927536, "grad_norm": 7.29642391204834, "learning_rate": 6.911764705882354e-05, "loss": 33.2233, "step": 90 }, { "epoch": 2.63768115942029, "grad_norm": 7.916087627410889, "learning_rate": 6.61764705882353e-05, "loss": 32.328, "step": 92 }, { "epoch": 2.6956521739130435, "grad_norm": 9.700789451599121, "learning_rate": 6.323529411764705e-05, "loss": 33.3864, "step": 94 }, { "epoch": 2.753623188405797, "grad_norm": 7.925725936889648, "learning_rate": 6.0294117647058825e-05, "loss": 33.0956, "step": 96 }, { "epoch": 2.8115942028985508, "grad_norm": 7.136451721191406, "learning_rate": 5.735294117647059e-05, "loss": 34.077, "step": 98 }, { "epoch": 2.869565217391304, "grad_norm": 7.057907581329346, "learning_rate": 5.441176470588235e-05, "loss": 32.854, "step": 100 }, { "epoch": 2.927536231884058, "grad_norm": 8.482731819152832, "learning_rate": 5.147058823529411e-05, "loss": 31.9702, "step": 102 }, { "epoch": 2.9855072463768115, "grad_norm": 8.195663452148438, "learning_rate": 4.8529411764705885e-05, "loss": 32.0516, "step": 104 }, { "epoch": 3.028985507246377, "grad_norm": 13.711084365844727, "learning_rate": 4.558823529411765e-05, "loss": 23.3778, "step": 106 }, { "epoch": 3.0869565217391304, "grad_norm": 13.546501159667969, "learning_rate": 4.2647058823529415e-05, "loss": 32.14, "step": 108 }, { "epoch": 3.1449275362318843, "grad_norm": 11.532342910766602, "learning_rate": 3.970588235294117e-05, "loss": 32.6737, "step": 110 }, { "epoch": 3.2028985507246377, "grad_norm": 9.859041213989258, "learning_rate": 3.6764705882352945e-05, "loss": 32.2716, "step": 112 }, { "epoch": 3.260869565217391, "grad_norm": 7.993894577026367, "learning_rate": 3.382352941176471e-05, "loss": 32.8686, "step": 114 }, { "epoch": 3.318840579710145, "grad_norm": 5.935787677764893, "learning_rate": 3.0882352941176475e-05, "loss": 33.3099, "step": 116 }, { "epoch": 3.3768115942028984, "grad_norm": 8.535140037536621, "learning_rate": 2.7941176470588236e-05, "loss": 33.1781, "step": 118 }, { "epoch": 3.4347826086956523, "grad_norm": 8.691988945007324, "learning_rate": 2.5e-05, "loss": 31.1977, "step": 120 }, { "epoch": 3.4927536231884058, "grad_norm": 7.092562675476074, "learning_rate": 2.2058823529411766e-05, "loss": 32.9427, "step": 122 }, { "epoch": 3.550724637681159, "grad_norm": 10.397586822509766, "learning_rate": 1.9117647058823528e-05, "loss": 32.8582, "step": 124 }, { "epoch": 3.608695652173913, "grad_norm": 7.606701374053955, "learning_rate": 1.6176470588235296e-05, "loss": 33.5666, "step": 126 }, { "epoch": 3.6666666666666665, "grad_norm": 9.118799209594727, "learning_rate": 1.323529411764706e-05, "loss": 32.2489, "step": 128 }, { "epoch": 3.7246376811594204, "grad_norm": 8.418676376342773, "learning_rate": 1.0294117647058824e-05, "loss": 32.8995, "step": 130 }, { "epoch": 3.782608695652174, "grad_norm": 7.379878044128418, "learning_rate": 7.3529411764705884e-06, "loss": 33.5503, "step": 132 }, { "epoch": 3.8405797101449277, "grad_norm": 7.183637619018555, "learning_rate": 4.411764705882353e-06, "loss": 31.911, "step": 134 }, { "epoch": 3.898550724637681, "grad_norm": 12.431236267089844, "learning_rate": 1.4705882352941177e-06, "loss": 31.6491, "step": 136 }, { "epoch": 3.898550724637681, "step": 136, "total_flos": 29261690983296.0, "train_loss": 35.055379671208996, "train_runtime": 354.4991, "train_samples_per_second": 6.229, "train_steps_per_second": 0.384 } ], "logging_steps": 2, "max_steps": 136, "num_input_tokens_seen": 0, "num_train_epochs": 4, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 29261690983296.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }