{ "best_metric": null, "best_model_checkpoint": null, "epoch": 11.0, "eval_steps": 500, "global_step": 3498, "is_hyper_param_search": true, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.9968553459119497, "grad_norm": 0.5557537078857422, "learning_rate": 0.00032468465370880226, "loss": 0.2366, "step": 317 }, { "epoch": 1.0, "eval_accuracy": 0.8729032258064516, "eval_loss": 0.06576934456825256, "eval_runtime": 2.6516, "eval_samples_per_second": 1169.097, "eval_steps_per_second": 24.513, "step": 318 }, { "epoch": 1.9937106918238994, "grad_norm": 0.34931233525276184, "learning_rate": 0.0002923284653322885, "loss": 0.0546, "step": 634 }, { "epoch": 2.0, "eval_accuracy": 0.9090322580645162, "eval_loss": 0.04315788671374321, "eval_runtime": 2.6079, "eval_samples_per_second": 1188.685, "eval_steps_per_second": 24.924, "step": 636 }, { "epoch": 2.990566037735849, "grad_norm": 0.35186338424682617, "learning_rate": 0.0002599722769557747, "loss": 0.034, "step": 951 }, { "epoch": 3.0, "eval_accuracy": 0.922258064516129, "eval_loss": 0.038206350058317184, "eval_runtime": 2.6441, "eval_samples_per_second": 1172.423, "eval_steps_per_second": 24.583, "step": 954 }, { "epoch": 3.9874213836477987, "grad_norm": 0.393400639295578, "learning_rate": 0.00022761608857926093, "loss": 0.0277, "step": 1268 }, { "epoch": 4.0, "eval_accuracy": 0.9296774193548387, "eval_loss": 0.03332997485995293, "eval_runtime": 2.6838, "eval_samples_per_second": 1155.08, "eval_steps_per_second": 24.219, "step": 1272 }, { "epoch": 4.984276729559748, "grad_norm": 0.08043188601732254, "learning_rate": 0.00019525990020274712, "loss": 0.0223, "step": 1585 }, { "epoch": 5.0, "eval_accuracy": 0.9387096774193548, "eval_loss": 0.028788963332772255, "eval_runtime": 2.6515, "eval_samples_per_second": 1169.143, "eval_steps_per_second": 24.514, "step": 1590 }, { "epoch": 5.981132075471698, "grad_norm": 0.08290695399045944, "learning_rate": 0.00016290371182623337, "loss": 0.0192, "step": 1902 }, { "epoch": 6.0, "eval_accuracy": 0.9393548387096774, "eval_loss": 0.02643631212413311, "eval_runtime": 2.63, "eval_samples_per_second": 1178.689, "eval_steps_per_second": 24.714, "step": 1908 }, { "epoch": 6.977987421383648, "grad_norm": 0.08547654002904892, "learning_rate": 0.0001305475234497196, "loss": 0.0172, "step": 2219 }, { "epoch": 7.0, "eval_accuracy": 0.9458064516129032, "eval_loss": 0.023828405886888504, "eval_runtime": 2.6537, "eval_samples_per_second": 1168.168, "eval_steps_per_second": 24.494, "step": 2226 }, { "epoch": 7.9748427672955975, "grad_norm": 0.0587199442088604, "learning_rate": 9.819133507320583e-05, "loss": 0.0159, "step": 2536 }, { "epoch": 8.0, "eval_accuracy": 0.9448387096774193, "eval_loss": 0.023019853979349136, "eval_runtime": 2.6517, "eval_samples_per_second": 1169.073, "eval_steps_per_second": 24.513, "step": 2544 }, { "epoch": 8.971698113207546, "grad_norm": 0.06312248110771179, "learning_rate": 6.583514669669206e-05, "loss": 0.0149, "step": 2853 }, { "epoch": 9.0, "eval_accuracy": 0.9438709677419355, "eval_loss": 0.02213130332529545, "eval_runtime": 2.6757, "eval_samples_per_second": 1158.577, "eval_steps_per_second": 24.293, "step": 2862 }, { "epoch": 9.968553459119496, "grad_norm": 0.051413584500551224, "learning_rate": 3.3478958320178286e-05, "loss": 0.014, "step": 3170 }, { "epoch": 10.0, "eval_accuracy": 0.9438709677419355, "eval_loss": 0.021344251930713654, "eval_runtime": 2.6694, "eval_samples_per_second": 1161.296, "eval_steps_per_second": 24.35, "step": 3180 }, { "epoch": 10.965408805031446, "grad_norm": 0.0515441969037056, "learning_rate": 1.1227699436645158e-06, "loss": 0.0133, "step": 3487 } ], "logging_steps": 317, "max_steps": 3498, "num_input_tokens_seen": 0, "num_train_epochs": 11, "save_steps": 1000000000.0, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 907333363443108.0, "train_batch_size": 48, "trial_name": null, "trial_params": { "alpha": 0.9860087246286913, "fp16": true, "learning_rate": 0.0003356061431608116, "lr_scheduler": "cosine", "num_train_epochs": 11, "temperature": 20, "warmup_steps": 210, "weight_decay": 0.1094679341539599 } }